Keywords

1 Introduction

Understanding the characteristics of workloads is essential to promoting the management of an HPC cluster. These known workload characteristics enable better services that have data dependence on workload status, such as cluster planning and scheduling [4, 34]. However, real HPC workloads are rarely open for studying because of the restrictions of privacy and confidentiality. Due to the lack of publicly available real HPC workloads, synthetic workloads are widely used in HPC research. For example, with synthetic workloads, portable I/O analysis of commercially sensitive HPC applications can be conducted [11]. The evaluation of scheduling algorithms also relies on the large number of synthetic workloads available [24, 30]. In particular, learning-based schedulers can be more fully trained on a large number of available synthetic workloads [12, 13]. Therefore, automated HPC workload generation has long been an active research topic among HPC providers and researchers [4, 23].

To accurately synthesize HPC workloads, characterizing workload patterns is essential. Some of these characteristics are intuitive and straightforward, such as the natural fluctuations in workloads throughout the week. Weekday workloads tend to be heavier than weekend workloads. Besides, there are implicit characteristics, such as workload fluctuations due to user habits [9]. Therefore, it is crucial to construct a detailed model and incorporate a broad set of factors to facilitate the extraction of all workload characteristics. With a comprehensive understanding of these workload characteristics, synthetic workloads can then be generated reasonably.

The classic approach to workload generation is statistical modeling [8, 25, 35]. Statistical modeling is a type of modeling method based on the regression of a large number of independent data to obtain the statistical distribution of the objective. For example, Poisson regression is a conventional method to simulate job arrivals [4]. It assumes that a large number of randomly arriving jobs usually follow a Poisson distribution. Then based on the time-dependent features, the regression model will fit a Poisson distribution for the job arrivals in each time interval. Other job attributes such as job run time are also typically generated by statistical modeling. The distribution of job run time is usually obtained by the empirical counting of the run time of all jobs in a period [8].

However, relying solely on statistical modeling frequently falls short of accurate workload generation. The primary focus of statistical modeling is on the distribution of workloads, neglecting potential correlations between various workload sequences. In this paper, we combine statistical modeling and autoregressive analysis for automatic HPC workload generation. At first, the job arrivals are obtained by modeling and simulation. Then other job attributes are further generated for the generated job arrival sequence.

In our workload generation approach, a statistical model combining Poisson regression and Poisson-Gamma regression is used to generate job arrivals. The commonly used Poisson regression works well in clusters with large amounts of job arrivals [9]. But in clusters with sparse workloads, the Poisson distribution assumption of job arrival distribution is difficult to be satisfied. Overdispersion of job arrivals occurs in these cases [32], which we deal with by introducing a Poisson-Gamma distribution. The hybrid model of Poisson regression and Poisson-Gamma regression can handle more job arrival distributions than Poisson regression alone.

Other job attributes, such as job requested resource and job run time, are generated by the autoregressive model in our approach. These job attributes are organized as time series in the historical job arrival sequences. In our approach, we perform attentive sequential generation for these job attributes based on autoregressive analysis of serial autocorrelation. Compared with statistical modeling, our model can better reproduce the data correlation in the series [4]. Based on the seasonal scheduling behavior derived from historical workloads, coupled with simulated job sequences provided by the job arrival model, job attributes of each job can be reasonably generated in our approach.

We evaluate the proposed workload generation approach on real HPC workload traces from four different real-world supercomputing centers. Our results show that our model can reveal the characteristics of real workload accurately. The synthetic workload generated by the proposed model reproduces scheduling information in the real workloads. In the evaluation, we use multiple metrics to measure the effectiveness of existing methods as baselines and show the superiority of our model.

The main contributions of this study are summarized as follows:

  • We simulate job arrivals based on statistical modeling combining Poisson regression and Poisson-Gamma regression, with which more diverse arrival situations can be accurately represented.

  • We conduct an autoregressive analysis to better model serialized workload information so that we can perform attentive sequential generation for job attributes realistically. So that the job attributes can be realistically generated for the simulated job sequence.

  • We evaluate the proposed method on real workload traces from four supercomputing centers. All evaluations demonstrate the superiority of using our model to generate synthetic workloads.

Table 1. Specifications of workload traces from four platforms.

2 Preliminary

We conduct our experiments on real traces from four different supercomputing centers. They are Supercomputing Center of University of Science and Technology of China [3], Center for High Performance Computing in Shanghai Jiao Tong University [1], Shanghai Supercomputing Center [2], and Gansu Supercomputing Center respectively, and are subsequently referred to as platforms A, B, C, and D. The workload trace data from platform A is a typical dataset to conduct our experiments, and the inspiration for building our workload generation model is mainly based on our exploration of it. The workload trace data from the other three platforms are used as supplementary datasets in our evaluation experiments to verify the generalization of our approach. The specifications of traces from these four platforms are shown in Table 1. Among them, the traces of the first three platforms have similar specifications, while the data size of platform D is smaller.

These workload traces are recorded in the Standard Workload Format (SWF) proposed by David Talby and refined through discussions by Dror Feitelson, James Patton Jones, and others [6]. For each job, we mainly pay attention to its user ID, submit time, run time, and requested resources. A complete job sequence plus these job attributes can describe a complete scheduling process. For evaluation of workload generation, the last 10 weeks of each trace will be used as a test window, while all the data will be used to train the workload generation model.

The above platforms used for evaluation experiments mainly carry scientific computing workloads for researchers in various fields [1,2,3]. They are dominated by these tightly coupled MPI jobs. Common scientific computing software, such as VASP (Vienna Ab initio Simulation Package) [18] and Gaussian [10], often submits a large-scale parallel computing job to the cluster. For reasons such as adaptation or price, users’ main choice for computing resources is the CPU rather than other acceleration devices such as GPU. The computing power of the computing nodes of the above platforms is indeed mainly provided by the CPU. So in this paper, job requested resources refer to the number of requested computing nodes, most of which are computing nodes with multiple CPU cores. Moreover, as the computing model expands, the run time of scientific computing tasks will increase beyond linear corresponding increases [19]. Therefore, the run times of jobs in these workloads vary greatly, with the longest scientific computing jobs running for weeks and the shortest test jobs running for just seconds.

Fig. 1.
figure 1

Overview of modeling framework.

3 Modeling Methodology

The workload generation model proposed in this paper consists of two major components, the job arrival model and the job attribute model. The overview of the modeling framework is shown in Fig. 1.

First, the arrival of jobs within each hour will be described by stochastic processes. The job arrival model employs two regression tasks to model job arrivals. Then, for each generated job with unknown attributes, the job attribute model autoregressively generates job attributes that conform to user habits at the user granularity, including requested resources, run time, job queue, etc.

3.1 Job Arrival Model

The job arrival model employs two regression tasks, Poisson regression and Poisson-Gamma regression to model the stochastic process of job arrivals. We treat job arrivals at one-hour intervals within the trace range as random arrivals. The feature vector of each interval is determined by the following three features:

  • HoD: One-hot code, to indicate this interval belongs to which hour of the day. Its dimension is 24.

  • DoW: One-hot code, to indicate this interval belongs to which day of the week. Its dimension is 7.

  • WoY: One-hot code, to indicate this interval belongs to which week of the year. Its dimension is 52.

HoD and DoW determine the feature vector of this interval in the regression task. They describe the temporal characteristics of job arrivals on day and week scales, respectively. WoY does not participate in the regression task. We use it to calculate a trend correction term for job arrivals that changes with the No. of weeks to make the regression model more generalizable. WoY is also used as the basis for us to assign user IDs to jobs within a week.

Poisson Model. Poisson regression is widely used in the simulation of job arrivals [4, 13, 16]. The probability distribution of a single Poisson distribution is: \(f(y=k;\lambda _i) = \frac{{\lambda _i}^k}{k!} e^{-\lambda _i},k=0,1,\dots \), and Poisson process assumes that in interval i, it contains \(y_i\) events arrive randomly. For interval i, we use feature vector \(\boldsymbol{x_i}\) to describe its temporal features. HoD and DoW determine a \(168\times 168\) one-hot code feature matrix \(\boldsymbol{X}\). For any feature vector, it can be matched to a row in the feature matrix. Then, the distribution parameters can be expressed as \(\boldsymbol{\lambda }= e^{ \boldsymbol{X} \boldsymbol{\beta }}\), where \(\boldsymbol{X}\) is the feature matrix composed of independent feature vectors, and \(\boldsymbol{\beta }\) is the parameter matrix that needs to be regressed. The log-likelihood function of Poisson regression can be calculated by:

$$\begin{aligned} \begin{aligned} l(\boldsymbol{\beta }) = \sum _{i=1}^n (y_i \boldsymbol{x_i} \boldsymbol{\beta }- e^{\boldsymbol{x_i} \boldsymbol{\beta }} ), \end{aligned} \end{aligned}$$
(1)

where n represents that there are n observations in the training set, and \(y_i\) is the number of event arrivals when the corresponding feature vector is \(\boldsymbol{x_i}\). The partial derivative of the negative log-likelihood function in the direction of \(\boldsymbol{\beta }\) is used as Loss for training, and the \(\boldsymbol{\beta }\) parameter can be optimized by minimizing the negative log-likelihood estimation.

Poisson-Gamma Model. In real data, due to problems such as overdispersion and zero-inflation [32], the naive Poisson process is difficult to fully describe the arrivals in the real world. For such data, Its probability distribution can be better represented by a negative binomial distribution, or a Poisson-Gamma mixture distribution: \(f(y=k;r,p) = \frac{\varGamma (k+r)}{\varGamma (k+1)\varGamma (r)}p^r(1-p)^k ,k=0,1,\dots \). For the overdispersed part, negative binomial regression can be chosen instead of naive Poisson regression, by replacing the parameters as follows: \(\alpha _i=\frac{1}{r}\), \(\mu _i = \frac{1-p}{\alpha p}\). Then the log-likelihood function of negative binomial regression can be calculated by:

$$\begin{aligned} \begin{aligned} l(\boldsymbol{\alpha },\boldsymbol{\mu }) = \sum _{i=1}^{n}\{y_i ln\frac{\boldsymbol{\alpha }\boldsymbol{x_i} \boldsymbol{\mu }\boldsymbol{x_i}}{1+\boldsymbol{\alpha }\boldsymbol{x_i} \boldsymbol{\mu }\boldsymbol{x_i}} +ln \varGamma (y_i +\frac{1}{\boldsymbol{\alpha }\boldsymbol{x_i}}) \\ -\frac{1}{\boldsymbol{\alpha }\boldsymbol{x_i}}ln(1+\boldsymbol{\alpha }\boldsymbol{x_i} \boldsymbol{\mu }\boldsymbol{x_i}) -ln\varGamma (y_i+1)-ln\varGamma (\frac{1}{\boldsymbol{\alpha }\boldsymbol{x_i}})\}. \\ \end{aligned} \end{aligned}$$
(2)

Hybrid Model. The above two regression models will be trained to convergence taking into account zero inflation of the data. At this point, we have two different models describing job arrivals during each one-hour interval. The Poisson model is more suitable for describing data with sufficient statistics and less interference. The Poisson-Gamma model can describe overdispersed data more accurately. We fuse these two models into a unified hybrid model based on Bayesian generalized additive models [21].

This hybrid model serves as our final job arrival model to simulate job arrivals within each one-hour interval. These job arrivals will be modified based on the correction of the weekly job arrival trend according to the interval’s WoY. Then a sequence of jobs with just arrival time is generated, \( \boldsymbol{J} = \{j_1,j_2,j_3,\cdots \}\). Also based on WoY, we count the proportion of job arrivals for each user every week and randomly assign user IDs to all jobs within a week according to the generalized Bernoulli processes, which is based on the multinomial distribution of job arrivals over the user set. Let the user set be \( \boldsymbol{U} = \{u_1,u_2,u_3,\cdots \}\), and the entire job sequence \( \boldsymbol{J} \) can be divided into multiple subsequences based on user IDs: \( \boldsymbol{J}^{u_1} = \{j^{u_1}_1,j^{u_1}_2,j^{u_1}_3,\cdots \}, \boldsymbol{J}^{u_2} = \{j^{u_2}_1,j^{u_2}_2,j^{u_2}_3,\cdots \}, \cdots \).

3.2 Job Attribute Model

The function of the job attribute model is to generate reasonable job attributes for each job in the simulated job arrival sequence \(\{j_1,j_2,j_3,\cdots \}\). Then we describe the attribute embedding, the autoregressive analysis process, and the attentive sequential workload generation in our job attribute model.

Attribute Embedding. The job attributes recorded in HPC cluster trace include many items, such as requested resources, run time, job queue, job status, etc. The job attributes that are the targets of our modeling generation can be specific to the following two items: (1) Job Requested Resource, indicates the number of computing resources requested by this job, for example, 5 Computing node cores; (2) Job Run Time, indicates the time that this job occupies computing resources while running; Based on these two items, the job scheduling process within an HPC cluster can be clearly constructed. Other miscellaneous items that have a minor impact on the scheduling process are not our modeling generation targets due to the inconsistency in trace record attribute types across platforms. But all items will contribute features to our attribute embedding, let the number of items be k. In order to facilitate attribute generation, we quantize the job attributes, in which the values of these items will be mapped to k discrete sets by clustering. In this way, the job attributes can be uniquely represented by one-hot codes \(\in \mathbb {R}^{k \times D}\), where D is the dimension of discrete sets. We encode these items together into an embedding vector to represent the attributes of the job:

$$\begin{aligned} \begin{aligned} A(j) = Embed[R(j),T(j),O(j)], \end{aligned} \end{aligned}$$
(3)

where R(j), T(j), and O(j) respectively represent the requested resources, run time, and other miscellaneous items of job j. \(A(j) \in \mathbb {R}^d\) is an embedding vector representing the attributes of job j. From this, job attributes can be generated based on a multinomial distribution over k discrete sets. Our job attribute model is built based on autoregressive analysis.

Autoregressive Analysis. For a user’s sequence of jobs sorted by arrival time, job attributes are autocorrelated over time scales [4]. This temporal correlation contains the user’s habit of submitting jobs in the cluster. We autoregressively analyze this temporal correlation between jobs to obtain the possible job attribute distribution. Assume a specific user u, we need to generate job attributes \(\{A(j^{u}_1),A(j^{u}_2),\cdots \}\) for each job in u’s simulated job sequence \(\{j^{u}_1,j^{u}_2,\cdots \}\). The basis of this generation process is an autoregressive analysis on u’s historical job attribute sequence, \( \{A(j^u_{h1}),A(j^u_{h2}),A(j^u_{h3}),\cdots \}\).

Generally, for the attribute generation of user u’s i-th job, we have to extract information from the previous part of the historical sequence to obtain its attribute distribution:

$$\begin{aligned} \begin{aligned} M(j^u_i) = softmax[extrac\{A(j^u_{h(i-n)}),\cdots ,A(j^u_{h(i-2)}),A(j^u_{h(i-1)})\}], \end{aligned} \end{aligned}$$
(4)

where \(M(j^u_i)\) represents the multinomial distribution of attribute values of user u’s i-th job. In the following section, we will introduce how the approach proposed in this paper autoregressively extracts historical sequence information to perform workload generation.

Attentive Sequential Generation. The proposed approach mainly uses the attention layer to extract historical sequence information. Compared with RNNs (Recurrent neural networks) and statistical time series models such as ARIMA (Autoregressive Integrated Moving Average Model), Transformer networks based on the attention mechanism are superior in extracting long-term dependencies in sequence information [31]. The attention mechanism can be defined as:

$$\begin{aligned} \begin{aligned} Attention(\boldsymbol{Q},\boldsymbol{K},\boldsymbol{V}) = softmax(\frac{\boldsymbol{Q} \boldsymbol{K}^T}{\sqrt{d}})\boldsymbol{V}, \end{aligned} \end{aligned}$$
(5)

where \(\boldsymbol{Q}\) represents the queries, \(\boldsymbol{K}\) the keys and \(\boldsymbol{V}\) the values. The attention block calculates a weighted sum of all values, where the weight between query i and value j relates to the interaction between query i and key j. The scale factor \(\sqrt{d}\) is to avoid overly large values of the inner product. In our method, the attention layer is used to extract historical sequence information:

$$\begin{aligned} \begin{aligned} \boldsymbol{A}^u_i &= \{A(j^u_{h(i-n)}),\cdots ,A(j^u_{h(i-2)}),A(j^u_{h(i-1)})\},\\ \boldsymbol{E}^u_i &= Attention(\boldsymbol{A}^u_i\boldsymbol{W}^Q,\boldsymbol{A}^u_i\boldsymbol{W}^K,\boldsymbol{A}^u_i\boldsymbol{W}^V), \end{aligned} \end{aligned}$$
(6)

where \(\boldsymbol{W}^Q,\boldsymbol{W}^K,\boldsymbol{W}^V \in \mathbb {R}^{d\times d}\) is the linear projection matrices. The intermediate representation \(\boldsymbol{E}^u_i\) is then transformed through the feed-forward layer and softmax layer, and finally the multinomial distribution \(M(j^u_i)\) of the attributes of job \(j^u_i\) is obtained. \(M(j^u_i) \in \mathbb {R}^{k \times D}\) is the probability distribution of job attributes based on the user’s long-term habits obtained by autoregressively analyzing the user u’s historical job sequence. \(A(j^u_{h(i)}) \in \mathbb {R}^{k \times D}\) before embedding is the one-hot code of the job attributes actually submitted by the user in this order, which reflects the short-term considerations of user u. Generally, we randomly generate job attributes for all jobs in the simulated job sequence according to the probability distribution \([M(j^u_i)+A(j^u_{h(i)})]/2\). At this point, the workload generation that mimics the cluster scheduling behavior but is not restricted to historical workload patterns is completed.

4 Evaluation

In this section, we evaluate whether synthetic workloads can mimic the scheduling behavior exhibited by real workloads from multiple perspectives. We first evaluate the generated job arrivals and then evaluate the overall workload generation results. Also, we evaluate our workload generation model on a use case.

Fig. 2.
figure 2

Evaluation of job arrival models with metric Mean.

4.1 Evaluation of Job Arrival Generation

We evaluate our job arrival generation model by comparing its outputs with the outputs of Poisson regression model as the baseline. We use the following evaluation metrics:

  • Mean, is the average number of job arrivals in each feature interval.

  • Devi, is the regression deviance that is twice the difference between the maximum achievable log-likelihood and the log-likelihood of the fitted model: \(Devi = 2[l(\boldsymbol{y}) - l(\boldsymbol{\alpha }, \boldsymbol{\mu })]\).

As described in Sect. 3.1, the feature vector of each interval in the job arrival model, which is a 168-dimensional one-hot code (7 days a week, multiplied by 24 h a day), represents the one-hour interval in one hour of the day of the week. We evaluate job arrival generation results over time intervals represented by all 168 different one-hot feature vectors. The evaluation will compare whether the distribution of simulated job arrivals at each feature interval can approach the distribution of real job arrivals from the above metrics.

For the metric, Mean, it is intended to judge whether the arrival model can generate a reasonable number of job arrivals. The evaluation results for this metric are shown in Fig. 2. It can be seen that our model and the Poisson regression model both can generate a reasonable number of job arrivals in each feature interval. In each feature interval, the average number of generated job arrivals is very close to the statistical results in the real trace.

Fig. 3.
figure 3

Evaluation of job arrival models with metric Devi.

The metric, Devi, is intended to judge whether the regressions in arrival models converge well or not. The distribution obtained by Poisson regression will be approximated for calculation of deviance as a negative binomial distribution of \(\boldsymbol{\mu }= \boldsymbol{\lambda }\) and \(\alpha = 10^{-5}\). The evaluation results for this metric are shown in Fig. 3. Taking platform A as an example, it can be seen that naive Poisson regression is difficult to converge on the trace of platform A, while our model converges perfectly. This result shows that although the Poisson regression model can generate a reasonable number of job arrivals, it cannot restore the overdispersion in the original data distribution. Notably, the periodic variation of deviance in Poisson regression is strongly correlated with the periodic variation of the mean of job arrivals in all features. The periodic variation in the mean of job arrivals is due to the change in the amount of user job submissions caused by the alternation of day and night. Similar periodic variation in regression deviance of job arrivals indicates that there is a strong positive correlation between the degree of overdispersion of job arrivals and the number of job arrivals in the evaluated HPC cluster.

4.2 Evaluation of Workload Generation

We evaluate our overall workload generation results by comparing them with the results of existing methods as baselines:

  • Multinomial: This type of method directly generates workloads based on the multinomial distribution of user job attributes based on statistics on all historical traces. Many classic HPC workload modeling methods are of this type [8, 25, 35].

  • RepeatFlav: This method generates duplicate job attributes for all user jobs in a period based on the most frequently occurring user job attributes in this period. It was used as the main baseline in the work of Bergsma et al. [4].

  • Bergsma: The work of Bergsma et al. serves as our main baseline [4]. Their method can represent the state-of-the-art deep learning-based workload generation methods. They used LSTM combined with survival prediction as the backbone network to conduct autoregressive analysis of historical sequence information and gradually job attribute distribution step by step.

We use the following evaluation metrics:

  • Coverage Accuracy, indicates how accurately the distribution used for job attribute generation covers the real job attributes. For a job, if the probability of generating real job attribute values in its generation distribution exceeds 80%, then this step of generation is counted as an accurate generation.

  • Cosine Similarity, represents the cosine similarity between synthetic workload sequences and real workload sequences, Cosine Similarity\((\boldsymbol{W} ,\hat{\boldsymbol{W}}) = \boldsymbol{W} \cdot \hat{\boldsymbol{W}}/(||\boldsymbol{W} ||\ ||\hat{\boldsymbol{W}}||)\)

These two metrics measure the correlation between model output and real workloads from different perspectives.

Table 2. Evaluation of job requested resource generation.

Table 2 shows the metrics comparison on all platform traces of job requested resource sequence generated by our model and baselines. It can be seen that our model has higher coverage accuracy and can better reflect the job requested resource correlation in real workloads, compared to all baselines. In all platforms on the dataset, the variation of job requested resources is not large, so even Multinomial model can effectively generate the job requested resource for all jobs.

Table 3. Evaluation of job run time generation.

Table 3 is the metrics comparison on all platform traces of job run time generated by our model and baselines. Unlike the evaluation of the job requested resource model, the performance of Multinomial model and RepeatFlav model are poor, due to the variation of run time being too large in all platform traces. In contrast, Bergsma model and our model achieve much better results on the generation of job run time. Compared with the Bergsma model, our job running time generation method also has advantages in both metrics.

4.3 Use Case of Workload Generation

We evaluate our workload generation model on a real-world use case: cluster planning [4]. Cluster planning is when the workload in the cluster is overloaded/underloaded, we need to plan the appropriate number of nodes to add/reduce. For this problem, synthetic workloads are necessary to more fully simulate and verify whether the cluster is load balanced after adding or reducing the corresponding number of nodes. Figure 4 shows our evaluation results on the trace of platform A of whether the synthetic workloads reflect cluster load levels. In this experiment, we assume that the cluster has no limit on the number of computing resources, and then count the number of computing node cores occupied by the workloads in each time period to represent the cluster load level. It can be seen that our synthetic workloads can well reflect the real load level, where we treat the attribute distribution of each job as a discretization of the Gaussian distribution to calculate the 90% confidence interval of the synthetic load level. Our synthetic workloads can be generated in large numbers based on random patterns while simulating real load levels, making cluster planning simulations more sufficient and credible.

Fig. 4.
figure 4

Evaluation of whether the synthetic workloads reflect cluster load levels.

5 Related Work

5.1 Workload Modeling in HPC

Workload modeling has long been an active topic in the HPC community. Many existing works [8, 25, 35] were devoted to using statistical multinomial distribution to fit the workload modeling of supercomputing systems, which is a classic HPC modeling method. Rodrigo et al. further characterized the workloads in their HPC systems based on the system life cycle and evolution trends [29] and proposed a modeling method for heterogeneous workloads based on this portrayal [28]. Our approach is inspired by these existing state-of-the-art methods and further proposes a more comprehensive workload modeling for workload generation.

5.2 Workload Modeling in Cloud Computing

For cloud computing, although the main workload service types are different, their workload scheduling forms are similar to those in classic supercomputing systems. The mainstream method of workload modeling in cloud computing is to describe workloads’ changing stochastic processes. Calheiros et al. introduced a workload prediction model based on ARIMA [5]. Gao et al. conducted a detailed study on the prediction of cloud resource allocation [15]. Cortez et al. constructed a time series model of the complete workload trend from the historical characteristics of virtual resource workload sequences [9]. Bergsma et al. established workload modeling also based on historical sequence characteristics, which can effectively generate complex and realistic workloads [4]. Our approach also draws from stochastic considerations and generation techniques in modeling workloads in cloud computing.

5.3 Resource Management in HPC

Workload modeling is beneficial for managing resources to improve utilization of HPC clusters. Many resource management services are based on workload profiling. The foundation of performance modeling and optimization of existing scheduling systems such as Slurm is workload modeling [26, 34]. Cluster planning also relies on workload models for more adequate simulation [4]. Workload modeling and generation is more widely used in the evaluation of scheduling algorithms [24, 30], ranging from classic schedulers [17], heuristic-based schedulers [7, 20, 22], prediction-based schedulers [14], deep reinforcement learning-based schedulers [12, 13, 27, 33, 36, 37]. Motivated by these existing works, we plan to design a novel cluster scheduler based on our workload modeling in future work.

6 Conclusion

In this paper, we proposed a method to model the workload in HPC clusters and generate synthetic workloads for HPC clusters. We combined statistical modeling and autoregressive analysis in our model to better characterize the workload scheduling behavior exhibited by real workloads. Using the proposed model, we can generate a synthetic workload at any time involved in modeling, which can reflect the most consistent workload distribution based on accurate historical characteristics extraction of real workloads. The proposed method greatly improved the quality of automatically generated workloads and made the workload generation model more reliable for resource management services, such as cluster planning and scheduling.

In the future, we plan to further evaluate how the synthetic workload reflects real scheduling behavior and determine whether it can play a role in data augmentation. On this basis, we will design a more efficient data-driven learning-based cluster scheduler that can be further enhanced with synthetic workloads for training.