hex -> Q
This commit is contained in:
parent
775d9e450f
commit
d20d36922e
|
@ -195,7 +195,7 @@ The results are 4D data (time, nodes, metrics, file system) per job.
|
|||
The distance measures should handle jobs of different lengths and node count.
|
||||
In \cite{Eugen20HPS}, we discussed a variety of options from 1D job-profiles to data reductions to compare time series data and the general workflow and pre-processing in detail.
|
||||
In a nutshell, for each job executed on Mistral, we partition it into 10-minute segments and compute the arithmetic mean of each metric, categorize the value into non-IO (0), HighIO (1), and CriticalIO (4) for values below 99-percentile, up to 99.9-percentile, and above, respectively.
|
||||
After data is reduced across nodes, we quantize the timelines either using binary or hexadecimal representation which is then ready for similarity analysis.
|
||||
After data is reduced across nodes, we quantize the timelines either using binary or quantum hexadecimal representation which is then ready for similarity analysis.
|
||||
By pre-filtering jobs with no I/O activity -- their sum across all dimensions and time series is equal to zero, we are reducing the dataset from about 1 million jobs to about 580k jobs.
|
||||
|
||||
\subsection{Algorithms for Computing Similarity}
|
||||
|
@ -356,8 +356,8 @@ Finally, the quantitative behavior of the 100 most similar jobs is investigated.
|
|||
To measure the performance for computing the similarity to the reference jobs, the algorithms are executed 10 times on a compute node at DKRZ which is equipped with two Intel Xeon E5-2680v3 @2.50GHz and 64GB DDR4 RAM.
|
||||
A boxplot for the runtimes is shown in \Cref{fig:performance}.
|
||||
The runtime is normalized for 100k jobs, i.e., for B-all it takes about 41\,s to process 100k jobs out of the 500k total jobs that this algorithm will process.
|
||||
Generally, the bin algorithms are fastest, while the hex algorithms take often 4-5x as long.
|
||||
Hex\_phases is slow for Job-S and Job-M while it is fast for Job-L, the reason is that just one phase is extracted for Job-L.
|
||||
Generally, the bin algorithms are fastest, while the Q algorithms take often 4-5x as long.
|
||||
Q\_phases is slow for Job-S and Job-M while it is fast for Job-L, the reason is that just one phase is extracted for Job-L.
|
||||
The Levenshtein based algorithms take longer for longer jobs -- proportional to the job length as it applies a sliding window.
|
||||
The KS algorithm is faster than the others by 10x but it operates on the statistics of the time series.
|
||||
Note that the current algorithms are sequential and executed on just one core.
|
||||
|
@ -480,7 +480,7 @@ To understand how the Top\,100 are distributed across users, the data is grouped
|
|||
\Cref{fig:userids} shows the stacked user information, where the lowest stack is the user with the most jobs and the topmost user in the stack has the smallest number of jobs.
|
||||
For Job-S, we can see that about 70-80\% of jobs stem from one user, for the Q-lev and Q-native algorithms, the other jobs stem from a second user while bin includes jobs from additional users (5 in total).
|
||||
For Job-M, jobs from more users are included (13); about 25\% of jobs stem from the same user; here, Q-lev, Q-native, and KS is including more users (29, 33, and 37, respectively) than the other three algorithms.
|
||||
For Job-L, the two hex algorithms include with (12 and 13) a bit more diverse user community than the bin algorithms (9) but Q-phases cover 35 users.
|
||||
For Job-L, the two Q algorithms include with (12 and 13) a bit more diverse user community than the bin algorithms (9) but Q-phases cover 35 users.
|
||||
We didn't include the group analysis in the figure as user count and group id is proportional, at most the number of users is 2x the number of groups.
|
||||
Thus, a user is likely from the same group and the number of groups is similar to the number of unique users.
|
||||
|
||||
|
@ -568,7 +568,7 @@ Bin\_all and B-aggz overlap with at least 99 ranks for all three jobs.
|
|||
While there is some reordering, both algorithms lead to a comparable set.
|
||||
All algorithms have a significant overlap for Job-S.
|
||||
For Job\-M, however, they lead to a different ranking, and Top\,100, particularly KS determines a different set.
|
||||
Generally, Q-lev and Hex\_native are generating more similar results than other algorithms.
|
||||
Generally, Q-lev and Q\_native are generating more similar results than other algorithms.
|
||||
From this analysis, we conclude that one representative from binary quantization is sufficient as it generates very similar results while the other algorithms identify mostly disjoint behavioral aspects and, therefore, should be analyzed individually.
|
||||
|
||||
|
||||
|
|
Loading…
Reference in New Issue