Next Article in Journal
Analysis of Phonetic Segments of Oesophageal Speech in People Following Total Laryngectomy
Next Article in Special Issue
Multilayer Measurement Methodology with Open-Source Tools for Evaluating IEEE 802.11ad Communication Quality
Previous Article in Journal
Genetic Diversity and Physicochemical Characteristics of Different Wheat Species (Triticum aestivum L., Triticum monococcum L., Triticum spelta L.) Cultivated in Romania
Previous Article in Special Issue
Wi-NN: Human Gesture Recognition System Based on Weighted KNN
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Time-Series Model for Varying Worker Ability in Heterogeneous Distributed Computing Systems

1
Samsung Electronics, Suwon 16677, Republic of Korea
2
Department of Statistics, Sungshin Women’s University, Seoul 02844, Republic of Korea
3
Data Science Center, Sungshin Women’s University, Seoul 02844, Republic of Korea
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(8), 4993; https://doi.org/10.3390/app13084993
Submission received: 7 March 2023 / Revised: 6 April 2023 / Accepted: 14 April 2023 / Published: 16 April 2023
(This article belongs to the Special Issue New Insights into Pervasive and Mobile Computing)

Abstract

:
In this paper, we consider the problem of estimating the time-dependent ability of workers participating in distributed matrix-vector multiplication over heterogeneous clusters. Specifically, we model the workers’ ability as a latent variable and introduce a log-normally distributed working rate as a function of the latent variable with parameters so that the working rate increases as the latent ability of workers increases, and takes positive values only. This modeling is motivated by the need to reflect the impact of time-dependent external factors on the workers’ performance. We estimate the latent variable and parameters using the expectation-maximization (EM) algorithm combined with the particle method. The proposed estimation and inference on the working rates are used to allocate tasks to the workers to reduce expected latency. From simulations, we observe that our estimation and inference on the working rates are effective in reducing expected latency.

1. Introduction

In order to meet the ever-growing demand for massive-scale computations using a large amount of data, distributed computing systems have become a central part of many machine learning algorithms [1]. However, distributed computing systems often face the problem of waiting for slow workers, called stragglers [2]. To overcome this issue and provide straggler tolerance, Lee et al. introduced redundancy to distributed matrix multiplication [3]. Specifically, the authors show that computational latency can be reduced using the ( n , k ) maximum distance separable (MDS) code, since collecting any k ( < n ) out of n workers’ results would complete the assigned task.
Including the idea proposed by Lee et al. [3], there has been a growing body of work using coding to mitigate the straggler problem in homogeneous clusters. This includes coding for high-dimensional matrix multiplication [4,5], matrix multiplication that considers the architecture of practical systems [6], gradient descent [7,8,9], convolution [10], and non-linear computation using a learning-based approach [11]. Mitigating the effects of stragglers is not limited to environments that consider computing clusters but can also be used in many types of distributed computing architectures that may struggle with stragglers or failures. Examples include fog computing, computation with a deadline, mobile edge computing, layered architecture of edge-assisted Internet of Things, and federated learning [10,12,13].
The system designs under homogeneous assumptions in potentially heterogeneous clusters can significantly reduce the overall performance [14]. For example, suppose that we assign the same amount of tasks to two workers with different computing capabilities. Consider a time for two workers to finish the tasks. A worker with relatively large computing capabilities takes less time to complete the given task than a worker with relatively small computing capabilities, on average. Thus, the expected time to complete the entire task is highly dependent on the time for a worker with relatively small computing capabilities to finish the task. In other words, the uniform load allocation to workers without considering the heterogeneous characteristics of the workers’ computing capabilities can lead to the performance degradation of computing systems. Reflecting on these points, research has been conducted under the assumption of heterogeneous clusters [15,16]. In [15], the authors present an asymptotically optimal load allocation to minimize the expected task completion delay for fully heterogeneous clusters. Similarly, in [16], an optimal load allocation is proposed to minimize the expected execution time under the assumption of group heterogeneity, which means that workers in different groups have different latency statistics.
In this work, we consider a scenario where the ability of each heterogeneous worker changes over time. While the work of [15,16] covers distributed computation in a heterogeneous environment, the authors assume that each worker’s ability is a fixed known constant. However, in reality, even workers designed to maintain a constant level of ability can have time-dependent abilities due to external factors such as changes in temperature and humidity over time [17]. If we do not estimate workers’ abilities, the basic load allocation method would be the uniform load allocation, which assigns the same amount of tasks to each worker. However, as we mentioned earlier, this load allocation method can potentially lead to a loss of system performance. Thus, the estimation of workers’ abilities is essential to improve system latency.
We model the aggregated impact of such factors on the worker’s ability as a normal distribution and set the worker’s ability as a latent variable. As demonstrated in several prior works [3,4,6,15,16], it is widely accepted that the runtime of a worker is often modeled by an exponential distribution. We define the working rate of a worker as the rate of the exponentially distributed runtime and assume that the working rate is log-normally distributed with two parameters: the mean and standard deviation of the logarithm of the working rate, respectively. The log-normal distribution is a good fit for the working rate since it ensures that the working rate takes only positive values and increases as the latent ability of a worker increases. We aim to estimate the latent variable and parameters associated with the ability/working rate of the worker so that one can properly allocate the workload on the workers in distributed computing. Many pieces of research have been devoted to the analysis of the time-series data with latent variables from different points of view [18,19,20,21]. In this paper, we employ particle methods to deal with both the estimation of parameters and the inference of latent variables.
Contribution: The key contributions of this work are summarized as follows.
  • To the best of our knowledge, we first model the workers’ ability that changes over time with latent variables.
  • We present an estimation algorithm for the parameters and latent variables of the workers’ ability using the expectation maximization (EM) algorithm combined with the particle method.
  • We verify the validity of the presented algorithm with Monte Carlo simulations.
  • We confirm the validity of our inference by verifying that the load allocation based on the estimated workers’ ability achieves the lower bound of the expected execution time.

2. Preliminaries

In this section, we describe our system model and model assumptions. Then an optimal load allocation is introduced according to the system model and the model assumptions.

2.1. System Model

We focus on the distributed matrix-vector multiplication over a master–worker setup in time-dependent heterogeneous clusters (described in Figure 1). At time t, we assign a task to compute the multiplication of A t and b t to the N distributed workers with different runtimes for a given matrix A t R k × d and the input vector b t R d × 1 . We assume that N workers are divided into G groups. Each group has a different number of workers and a different working rate. The workers in each group have the same working rate. We denote the number of workers in group i as N i , which implies that
i [ G ] N i = N .
The tasks are successively assigned to the workers at time points t = t 0 , t 1 , t T , t T + 1 , . To demonstrate our computation model, we present the uncoded and coded computations for a given task computing the multiplication of A t and b t assigned at time t as follows.
Figure 1. Illustration of the time-dependent heterogeneous cluster (task assigned at time t). The master sends the input vector b t to the N workers, each of which stores A i , j , t and A ˜ i , j , t in uncoded and coded computations, respectively, for i [ G ] and j [ N i ] ; (a) worker j in group i computes the multiplication of A i , j , t and b t with the working rate μ i , t , and sends back the computation results to the master; (b) worker j in group i computes the multiplication of A ˜ i , j , t and b t with the working rate μ i , t , and sends back the results to the master.
Figure 1. Illustration of the time-dependent heterogeneous cluster (task assigned at time t). The master sends the input vector b t to the N workers, each of which stores A i , j , t and A ˜ i , j , t in uncoded and coded computations, respectively, for i [ G ] and j [ N i ] ; (a) worker j in group i computes the multiplication of A i , j , t and b t with the working rate μ i , t , and sends back the computation results to the master; (b) worker j in group i computes the multiplication of A ˜ i , j , t and b t with the working rate μ i , t , and sends back the results to the master.
Applsci 13 04993 g001

2.1.1. Uncoded Computation

The rows of A t are divided into G disjoint submatrices as
A t = [ A 1 , t ; A 2 , t ; ; A G , t ]
where
A i , t = [ A i , 1 , t ; A i , 2 , t ; ; A i , N i , t ] R N i l i , t × d
and
i [ G ] N i l i , t = k .
Here, A i , j , t R l i , t × d is a submatrix of A i , t allocated to worker j in group i for i [ G ] and j [ N i ] . Worker j in group i is assigned a subtask to compute the multiplication of A i , j , t and b t after receiving the input vector b t from the master. Then, worker j in group i computes the multiplication of A i , j , t and b t and sends back the result to the master. After the master collects the computation results from all the workers, the desired computation result A t b t can be obtained.

2.1.2. Coded Computation

We apply an ( n , k ) MDS code to the rows of A t to obtain the coded matrix A t ˜ R n × d . Afterward, the rows of A ˜ t are grouped into N submatrices as
A ˜ t = [ A ˜ 1 , t ; A ˜ 2 , t ; ; A ˜ G , t ] ,
where
A ˜ i , t = [ A ˜ i , 1 , t , A ˜ i , 2 , t , , A ˜ i , N i , t ] R N i l i , t × d
is the coded data matrix allocated to the workers in group i and
i [ G ] N i l i , t = n .
Here, worker j in group i is assigned a subtask to compute the multiplication of A ˜ i , j , t R l i , t × d and b t for i [ G ] and j [ N i ] . Worker j in group i sends the computation result A ˜ i , j , t b t to the master after finishing the subtask multiplying A ˜ i , j , t and b t . The master can retrieve the desired computation result A t b t by combining the inner products of any k ( < n ) coded rows with b t from the MDS property.
For efficient load allocation to the distributed workers, we need to estimate the workers’ ability, which is assumed to change over time. To do so, for the first T time points t = t 0 , t 1 , , t T , we assign the matrix-vector multiplication tasks using the uncoded computation framework to measure the runtime of all the workers. Here, every worker is assigned the same amount of the subtask, i.e., we set l i , t = k / N for i [ G ] , j [ N i ] , and t = t 0 , t 1 , , t T . Based on the observations on the runtime, we estimate the parameters and latent variables of the working rates. Afterward ( t > t T ), the coded computation is applied using the load allocation l i , t based on the estimated working rates. The latent variable is obtained continuously for each time step to properly allocate l i , t to the workers. The detailed process of estimating the working rates will be shown in Section 3.

2.2. Model Assumptions

For the task assigned at time t, we assume the time taken for calculating an inner product of a row of A i , t ; the input vector b t follows the shifted exponential distribution with rate μ i , t , which is defined as the working rate of workers in group i. It is assumed that workers having the same working rate form a group. In other words, workers in a group have the same computing capabilities.
We model the working rate μ i , t as
μ i , t : = e X i , t
using a latent variable X i , t . Here, X i , t denotes the ability of workers in group i for a subtask assigned at time t.
For workers in group i, it is assumed that the zero state X i , 0 follows the normal distribution N ( m , σ 2 ) , where m and σ are model parameters. Let us denote θ = ( m , σ ) . We write f 0 as the probability density function of x i , 0 given parameters,
f 0 ( x i , 0 | θ = ( m , σ ) ) = 1 2 π σ e 1 2 σ 2 x i , 0 m 2 .
The ability of workers may vary over time, which is formulated by
X i , t : = X i , t 1 + ϵ i , t ,
where ϵ i , t is a white noise that is independent across group i and time t and follows the normal distribution with zero mean and variance σ ϵ 2 . Here, ϵ i , t represents the change in the worker ability as the time elapses from t 1 to t. In this work, we assume σ ϵ is given. For a precise formulation, we denote the distribution of X i , t given X i , t 1 = x i , t 1 by
f ( x i , t | x i , t 1 ) = 1 2 π σ ϵ e 1 2 σ ϵ 2 x i , t x i , t 1 2 ,
which is the probability density function of the normal distribution with mean x i , t 1 and variance σ ϵ 2 .
We assume that t j + 1 t j is a sufficient time interval for the workers to carry out their subtasks allocated at time t j . We further assume that a task is assigned at successive time points spaced at uniform intervals, i.e., t j + 1 t j = t l + 1 t l for all j l . Then, t 0 , t 1 , , t T can be rewritten as 0 , 1 , , T for notational convenience.
Let η i , j , t be the observed variable that indicates the runtime of worker j [ N i ] in group i [ G ] for a given data matrix
A i , j , t R l i , t × d
with the following probability density function:
p ( η i , j , t | X i , t , l i , t ) = μ i , t l i , t e μ i , t l i , t ( η i , j , t l i , t ) .
Since η i , j , t ’s are independent, we obtain the probability density function g of the observation η i , t given by
g ( η i , t | x i , t , l i , t ) = t [ T ] 0 j [ N i ] g ( η i , j , t | x i , t , l i , t ) .
for i [ G ] , j [ N i ] , and t [ T ] 0 . It follows from (5) that we have the probability density function of η i , 0 : T (observations for workers in group i through the time points t [ T ] 0 ) as follows:
p ( η i , 0 : T | x i , 0 : T , l i , 0 : T ) = t [ T ] 0 g ( η i , t | x i , t , l i , t ) ,
where l i , 0 : T and x i , 0 : T denote the number of rows of the data matrix assigned to workers in group i and the ability of workers in group i at time t [ T ] 0 , respectively. Then, for all of the workers,
p ( η 1 : G , 0 : T | x 1 : G , 0 : T , l 1 : G , 0 : T ) = i [ G ] t [ T ] 0 g ( η i , t | x i , t , l i , t ) = i [ G ] p ( η i , 0 : T | x i , 0 : T , l i , 0 : T ) ,
where η 1 : G , 0 : T , x 1 : G , 0 : T , and l 1 : G , 0 : T represent the overall observations, the workers’ ability, and the number of rows of the assigned matrices for workers in group i [ G ] at time t [ T ] 0 , respectively.

2.3. Optimal Load Allocation

In this subsection, we provide an analysis of load allocation given μ ^ i , t , for i [ G ] at time t. The previous work [16] presents the optimal load allocation to minimize the expected task runtime. We reproduce the results in [16] with some definitions according to our system model.
Let T i denote the task runtime taken to calculate the inner product of l i , t rows of A ˜ i , t with b t at a worker in group i at time t. We assume that T i s are independent random variables, with the shifted exponential distribution as follows
F i [ η ] = Pr [ T i η ] = 1 e μ i , t l i , t η l i , t
for i [ G ] and η l i , t . Here, it is assumed that the cumulative distribution function of the task runtime for a worker in group i to calculate one inner product is
1 e μ i , t ( η 1 )
for η 1 .
Let T r i , t : N i l i , t denote the r i , t -th order statistics of N i random variables following the distribution in (6). We define
T = max i [ G ] T r i , t : N i l i , t
as the task runtime for the master to finish the given task.
Theorem 1
(Theorem 2 in [16]). The optimal load allocation ( l 1 , t * , l 2 , t * , , l G , t * ) to achieve the minimum of
max i [ G ] E T r i , t : N i l i , t ,
denoted by T , is determined as follows:
l i , t * = k r i , t * + i ( i ) [ G ] r i , t * ξ ( r i , t * , N i ) ξ r i , t * , N i
for i [ G ] , where
r i , t * = N i 1 + 1 W 1 e μ i , t + 1 ,
ξ ( r i , t * , N i ) = 1 + 1 μ i , t log W 1 e ( μ i , t + 1 ) ,
and
r i , t * ξ r i , t * , N i = μ i , t N i W 1 ( e ( μ i , t + 1 ) ) .
Here, W 1 ( x ) is the lower branch of the Lambert W function. ( W 1 ( x ) denotes the branch satisfying W ( x ) e W ( x ) = x and W ( x ) 1 ). Then the minimum expected execution time, T , is represented as
k j [ G ] μ i , t N i W 1 e μ i , t + 1 .
Then, we have
E [ T ] max i [ G ] E T r i , t : N i l i , t T .
The first inequality follows from the definition of T in (7) and the last inequality comes from Theorem 1.
Next, we introduce the asymptotic behavior of the expected task runtime for the master to finish the given task as follows.
Theorem 2
(Theorem 3 in [16]). For the given optimal load allocation ( l 1 , t * , l 2 , t * , , l G , t * ) , E [ T ] is asymptotically equivalent to T for a sufficiently large N.
As we mentioned in Theorem 1, the optimal load allocation achieves the minimum of
max i [ G ] E T r i , t : N i l i , t ,
denoted by T . For the given ( μ 1 , t , μ 2 , t , , μ G , t ) and ( N 1 , N 2 , , N G ) , the constant T is a theoretical limit of E [ T ] . From Theorem 2, we can conclude that the load allocation given by (8) is also optimal for E [ T ] from an asymptotical perspective for sufficiently large N.
Remark 1.
For a fixed time t, the working rate of workers μ i , t in group i is the rate parameter of the shifted exponential distribution. This implies that the expected execution time for a worker in group i to finish the given task is a linear function with respect to the reciprocal of the working rate. A large μ i , t indicates that workers in group i have a relatively good ability, and the expected execution time for workers in group i tends to be relatively small. As given by Equation (1), a worker’s ability X i , t in group i can be represented by log μ i , t , where the log-normal distribution is widely used to describe the distribution of positive random variables.

3. Estimation of Latent Variable and Parameters

For the tasks allocated at time t [ T ] 0 , the master collects the computation result A i , t b t , as well as the observed data η 1 : G , t which means the runtime of the worker in group i [ G ] for the task assigned at t. After obtaining η 1 : G , 0 : T at t = T , we estimate the parameter θ and the latent variable X i , t using the EM algorithm combined with filtering and smoothing. For t > T , we obtain the estimated X i , t based on the particle filtering algorithm and the estimated parameter θ ^ .

3.1. EM Algorithm

We define the complete-data likelihood functions as follows:
L ( θ | η 1 : G , 0 : T , x 1 : G , 0 : T , l 1 : G , 0 : T ) = p ( η 1 : G , 0 : T , x 1 : G , 0 : T | θ , l 1 : G , 0 : T ) = p ( x 1 : G , 0 : T | θ ) p ( η 1 : G , 0 : T | x 1 : G , 0 : T , l 1 : G , 0 : T ) = i [ G ] p ( x i , 0 : T | θ ) p ( η i , 0 : T | x i , 0 : T , l i , 0 : T ) = i [ G ] f 0 ( x i , 0 | θ ) t [ T ] f ( x i , t | x i , t 1 ) t [ T ] 0 g ( η i , t | x i , t , l i , t )
Let
l ( θ | η 1 : G , 0 : T , x 1 : G , 0 : T , l 1 : G , 0 : T ) = log L ( θ | η 1 : G , 0 : T , x 1 : G , 0 : T , l 1 : G , 0 : T )
be the complete-data log-likelihood function where log represents the natural logarithm.
The EM algorithm is used to estimate the latent variable X i , t and parameter θ simultaneously, which consists of the following two steps: The first step is called the E-step, which starts by calculating the expected value of the complete data log-likelihood function Q ( θ | θ ( v ) ) , defined as
Q ( θ | θ ( v ) ) = E l ( θ | η 1 : G , 0 : T , X 1 : G , 0 : T , l 1 : G , 0 : T ) | η 1 : G , 0 : T , l 1 : G , 0 : T , θ ( v ) .
where the expectation is over X 1 : G , 0 : T , more precisely, p ( x 1 : G , 0 : T | η 1 : G , 0 : T , l 1 : G , 0 : T , θ ( v ) ) . Here, θ ( v ) is the v-th parameter estimate in the procedure of the EM algorithm. Then we have
Q ( θ | θ ( v ) ) = l ( θ | η 1 : G , 0 : T , x 1 : G , 0 : T , l 1 : G , 0 : T ) p ( x 1 : G , 0 : T | η 1 : G , 0 : T , l 1 : G , 0 : T , θ ( v ) ) d x 1 : G , 0 : T .
Hence, (10) implies that the distribution function for x 1 : G , 0 : T given η i , 0 : T , l 1 : G , 0 : T , and θ ( v ) are required to evaluate the Q function value at each iteration. The methodology for obtaining this distribution will be given in the next subsection. M-step finds the parameter of the next iteration using
θ ( v + 1 ) = arg max θ Q ( θ | θ ( v ) ) .
The aforementioned EM algorithm is summarized in Algorithm 1.
Algorithm 1 EM Algorithm.
(1)
Initialize θ ( 0 ) = ( α ( 0 ) , β ( 0 ) ) .
(2)
[E-Step] For the ( v + 1 ) -th iteration, compute Q ( θ | θ ( v ) ) .
(3)
[M-Step] Find
θ ( v + 1 ) = arg max θ Q ( θ | θ ( v ) ) .
(4)
Repeat the E-Step and M-Step until θ ( v ) converges.

3.2. Filtering and Smoothing

We resort to the sampling method to approximate the function Q ( θ | θ ( v ) ) since the integral (10) is analytically intractable. Our goal is to find an approximation of
p ( x 1 : G , 0 : T | η 1 : G , 0 : T , l 1 : G , 0 : T , θ ( v ) )
Since the distribution is independent across groups, we only need to infer
p ( x i , 0 : T | η i , 0 : T , l i , 0 : T , θ ( v ) )
for each group i, which is the so-called a smoothed estimate. To obtain the smoothed estimate, we use the particle method [22] since it is efficient in high-dimensional sampling and more suitable for filtering [23,24,25] compared to the other competing schemes. Specifically, we choose sequential Monte Carlo (SMC) for the filtering algorithm [26]. The algorithm is based on sequential importance sampling with resampling, which has advantages in controlling the variance of estimates [27]. Note that the SMC algorithm is based on importance sampling, and the sampled object is called the particle (hence, its name).
For workers in group i and the task assigned at t, the number of particle sequences is denoted by L and the l ( [ L ] ) -th particle is denoted by x i , t ( l ) . We need to obtain the proposal (importance) distributions q 0 ( x i , 0 | η i , 0 ) and q ( x i , t | η i , t , x i , t 1 ) . The optimal choices of the proposal distribution and proposal function are just the distributions of themselves, i.e.,
q 0 ( x i , 0 | η i , 0 ) = p ( x i , 0 | η i , 0 )
and
q ( x i , t | η i , t , x i , t 1 ) = p ( x i , t | η i , t , x i , t 1 ) .
However, evaluating these distributions is intractable in our model. Although there are several possible methods to obtain a proposal distribution, we choose the following, which are known to work well in practice:
q 0 ( x i , 0 | η i , 0 ) = f 0 ( x i , 0 | θ ( v ) ) g ( η i , 0 | x i , 0 , l i , 0 ) g ( η i , 0 | x i n i t , l i , 0 )
and
q ( x i , t | η i , t , x i , t 1 ) = f ( x i , t | x i , t 1 ) g ( η i , t | x i , t , l i , t ) g ( η i , t | x i , t 1 , l i , t )
for t [ T ] . Here, x i n i t is an initial guess of x 0 . These choices ease the computation of importance weights, which are given as follows for the l-th particle sequence:
w i , 0 ( x i , 0 ( l ) ) = f 0 ( x i , 0 | θ ( v ) ) g ( η i , 0 | x i , 0 ( l ) ) q ( x i , 0 ( l ) | η i , 0 ) = g ( η i , 0 | x i n i t , l i , 0 )
and
w i , t ( x i , t 1 : t ( l ) ) = g ( η i , t | x i , t ( l ) , l i , t ) f ( x i , t ( l ) | x i , t 1 ( l ) ) q ( x i , t ( l ) | η i , t , x i , t 1 ( l ) ) = g ( η i , t | x i , t 1 ( l ) , l i , t ) .
Note that we can skip the resampling step for t = 0 since the importance weights w i , 0 ( x i , 0 ( l ) ) for t = 0 are independent of the particles l [ L ] , which means all particles have the same weights. Therefore, x i n i t has no effects on the algorithm practically.
The algorithm for SMC filtering on workers in group i is summarized in Algorithm 2. One additional shorthand notation
x i , ( 0 : s ) | s ( l ) = ( x i , 0 | s ( l ) , x i , 1 | s ( l ) , , x i , s | s ( l ) )
is used in describing Algorithm 2.
Algorithm 2 The SMC for filtering on workers in group i.
Input: η i , 0 : s , l i , 0 : s , θ ( v ) .
-
At time t = 0 ,
  • Sample x ¯ i , 0 ( l ) from q 0 ( x i , 0 | η i , 0 ) , l = 1 , 2 , , L .
-
At time t = 1 , , s ,
  • Sample x i , t ( l ) from q ( x i , t | η i , t , x ¯ i , t 1 ( l ) ) , l = 1 , 2 , , L .
  • Set x i , 0 : t ( l ) x ¯ i , 0 : t 1 ( l ) , x i , t ( l ) .
  • Compute unnormalized importance weights
    w i , t ( x i , t 1 : t ( l ) ) = g ( η i , t | x i , t 1 ( l ) ) .
  • Compute normalized weights
    W i , t ( l ) = w i , t ( x i , t 1 : t ( l ) ) l = 1 L w i , 0 ( w i , t ( x i , t 1 : t ( l ) ) ) .
  • Resample from { W i , t ( l ) , x i , 0 : t ( l ) } l = 1 , , L to obtain L equally weighted resampled particles { 1 / L , x ¯ i , 0 : t ( l ) } l = 1 , , L .
Output: p ^ ( x i , 0 : s | η i , 0 : s , l i , 0 : s , θ ( v ) ) .
We can effectively sample from the proposal distribution q ( x i , t | η i , t , x i , t 1 ) by the adaptive rejection sampling (ARS) [28] since the proposal distribution is log-concave as shown in Proposition 1.
Proposition 1.
The proposal distributions q 0 ( x i , 0 | η i , 0 ) and q ( x i , t | η i , t , x i , t 1 ) are log-concave functions with respect to x i , 0 and x i , t , respectively.
Proof. 
By (11), we have
log q 0 ( x i , 0 | η i , 0 ) = log f 0 ( x i , 0 | θ ( v ) ) + log g ( η i , 0 | x i , 0 , l i , 0 ) log g ( η i , 0 | x i n i t , l i , 0 ) .
The second-order partial derivative of the above function with respect to x i , 0 evaluates to
1 σ 0 ( v ) 2 j [ N i ] η i , j , 0 l i , 0 e x i , 0 ,
which is negative for all real values x i , 0 . Since the logarithm of the function is concave, the proposal distribution q 0 ( x i , 0 | η i , 0 ) is log-concave in x i , 0 .
Similarly, (12) implies
log q ( x i , t | η i , t , x i , t 1 ) = log f ( x i , t | x i , t 1 ) + log g ( η i , t | x i , t , l i , t ) log g ( η i , t | x i , t 1 , l i , t ) .
The second-order partial derivative with respect to x i , t gives the following negative value for all real x i , t :
1 σ ϵ 2 j [ N i ] η i , j , t l i , t e x i , t .
Therefore, the proposal distribution q ( x i , t | η i , t , x i , t 1 ) is log-concave in x i , t .    □
We resample at every time step so that the particle sequences have equally weighted samples. In this paper, we use the multinomial resampling [22], which involves sampling from the multinomial distribution. The sampling probabilities are given by the weights of the corresponding original weighted samples.
Then, we can calculate
p ^ ( x i , 0 : T | η i , 0 : T , l i , 0 : T , θ ( v ) )
the approximation of the desired distribution by employing Algorithm 2. However, note that we cannot directly use the output
p ^ ( x i , 0 : T | η i , 0 : T , l i , 0 : T , θ ( v ) )
to the EM algorithm because the particle filtering method often suffers from the sample impoverishment, called the degeneracy problem. When the number of distinct particles
{ x i , t ( l ) } l = 1 , , L
is too small for t T , it can lead to
p ^ ( x i , t | η i , 0 : T , l i , 0 : T , θ ( v ) )
is not reliable for t T . Therefore, from Algorithm 2, we can take the filtered marginal distribution
p ^ ( x i , s | η i , 0 : s , l i , 0 : s , θ ( v ) )
of the distribution
p ^ ( x i , 0 : s | η i , 0 : s , l i , 0 : s , θ ( v ) )
at time s = 0 , 1 , , T , when we are interested in the online estimation of state variables. We have
p ^ ( x i , t | η i , 0 : t , l i , 0 : t , θ ( v ) ) = 1 L l = 1 L δ x i , t | t ( l ) ( x i , t )
for t [ T ] 0 , where δ x 0 ( x ) is the Dirac delta mass located at x 0 . Note that the weights are the same as 1 / L due to resampling at the last step.
To obtain the smoothed estimate
p ^ ( x i , t | η i , 0 : T , l i , 0 : T , θ ( v ) )
for the EM algorithm, we rely on the forward-filtering backward-sampling (FFBSa) algorithm using the filtered particles x i , t | t ( l ) , l [ L ] , in (13). The detailed process is given in Algorithm 3.
Algorithm 3 FFBSa on workers in group i.
Input: η i , 0 : T , l i , 0 : T , θ ( v ) .
  • Run Algorithm 2 to obtain the filtered distribution
    { p ^ ( x i , t | η i , 0 : t , l i , 0 : t , θ ( v ) ) } t = 0 , 1 , , T .
  • Sample x i , T ( l ) from
    p ^ ( x i , T | η i , 0 : T , l i , 0 : T , θ ( v ) ) ,
    l = 1 , 2 , , L .
  • For t = T 1 , T 2 , , 1 , 0 , sample x i , t | T ( l ) from
    p ^ ( x i , t | η i , 0 : t , x i , t + 1 | T ( l ) , l i , 0 : t , θ ( v ) ) ,
    l = 1 , 2 , , L , where
    p ^ ( x i , t | η i , 0 : t , x i , t + 1 | T ( l ) , l i , 0 : t , θ ( v ) ) = l = 1 L f ( x i , t + 1 | x i , t | T ( l ) ) δ x i , t | t ( l ) ( x i , t ) l = 1 L f ( x i , t + 1 | x i , t | T ( l ) ) .
Output: p ^ ( x i , t | η i , 0 : T , l i , 0 : T , θ ( v ) ) , t = 0 , 1 , , T .
By Algorithm 3, an estimate of the smoothed marginal distribution of x i t is represented as
p ^ ( x i , t | η i , 0 : T , l i , 0 : T , θ ( v ) ) = 1 L l = 1 L δ x i , t | T ( l ) ( x i , t )
for t = 0 , 1 , , T . Now we have the particles from the distribution
p ( x i , t | η i , 0 : T , l i , 0 : T , θ ( v ) ) .
Since all particle sequences have the same weights due to resampling, we can approximate the function Q ( θ | θ ( v ) ) as follows:
Q ( θ | θ ( v ) ) 1 L l = 1 L l ( θ | η 1 : G , 0 : T , x 1 : G , t | T ( l ) , l 1 : G , 0 : T ) .
We can run the EM algorithm given in Algorithm 1 by using the particles obtained from Algorithm 3. We begin with the initial guess θ ( 0 ) = ( α ( 0 ) , β ( 0 ) ) . In the first iteration, we compute Q ( θ | θ ( 0 ) ) and find
θ ( 1 ) = arg max θ Q ( θ | θ ( 0 ) ) .
We then run iterations until θ ( v ) converges and obtain the converged parameter estimate θ ^ .

3.3. Inference of X i , t

We have the estimated parameter θ ^ from the EM algorithm. Then the parameter θ ^ and the given values l 1 : G , 0 : T are omitted for brevity. Once a point estimate x ^ 1 : G , 0 : T of the ability x 1 : G , 0 : T is given, we can directly infer the working rates via (1). We will look at several point estimates in the following subsections based on particle filtering and smoothing algorithms.

3.3.1. Offline Inference

We consider estimating x i , t when t T , where we have observations until time t = T . Then the smoothed estimates can be exploited for this purpose. For example, offline estimation is used in the estimation of the model parameters in Algorithm 1 to avoid the particle degeneracy problem [23]. Given the observations η i , 0 : s up to time s, we denote the mean of X i , t as
x ^ i , t | s = E [ X i , t | η i , 0 : s ] .
For a point estimate on x i , t , we take the smoother mean x ^ i , t | T . To calculate x ^ i , t | T the smoothed estimate
p ^ ( x i , t | η i , 0 : T )
is obtained from Algorithm 3 considering the observed data η i , 0 : T . Then, x ^ i , t | T is represented as
m i , t | T = E [ X i , t | η i , 0 : T ] 1 L l = 1 L x i , t | T ( l ) ,
where x i , t | T ( l ) , l = 1 , 2 , , L are the smoothed samples from Algorithm 3.

3.3.2. Online Inference

This subsection considers the case of t > T ; we need a point estimate on x i , t for the load allocation at time t > T . The filtered mean x ^ i , t | t can be obtained from Algorithm 2 as soon as we have observations at time t > T . Using the observed data η i , 0 : t , x ^ i , t | t is expressed as the filtered mean
x ^ i , t | t = E [ X i , t | η i , 0 : t ] 1 L l = 1 L x i , t | t ( l ) ,
where x i , t | t ( l ) , l = 1 , 2 , , L are the filtered samples from Algorithm 2. Unlike offline inference, online inference is free from the particle degeneracy problem. We can employ the online inference procedure to obtain the ability estimates simultaneously with the entry of observations.

4. Performance Evaluation

We perform a simulation study in order to validate the suggested inference methods. We fix the true parameters m = 1 and σ = 1 so that the initial ability of groups has an average of 1 and a standard deviation of 1. We allow a minor degree of ability changes over time by setting σ ϵ = 0.01 . We consider T = 20 timepoints after the initial time and G = 10 groups of workers. The number of particles is set to L = 200 . For simplicity, we assume that all workers are assigned unit loads over time so that l i , t = 1 for i [ G ] , j [ N i ] , and t [ T ] 0 , and each group has the same number of workers N w so that N 1 = = N G = N w . We vary the number of workers, N w = 10 , 30 , 50 , to check that the algorithm works for various situations and to see how the number of workers affects the performance of the proposed algorithms.
We generate the ability x 1 : G , 0 : T of workers using (2) and (3) and use them to generate the observation data η 1 : G , 1 : N w , 0 : T that follows the distribution given in (4). This process is repeated 30 times for each case N w = 10 , 30 , 50 . Then we apply the inference algorithm to estimate the model parameters and the ability of workers.
Table 1 presents the parameter estimates of 30 repeated datasets for the cases N w = 10 , 30 , 50 . The algorithm estimates the parameters close to the true parameters m = 1 and σ = 1 in all of the cases of N w , suggesting that the algorithm works well in estimating the true parameter. Moreover, we can see that the standard deviations of all the estimates over 30 repetitions tend to decrease as N w increases, implying the larger number of workers in each group tends to yield more accurate estimates.
We now study the accuracy of the ability estimate x ^ 1 : G , 0 : T using the smoothed estimates given in (14). We use three criteria, the Pearson correlation coefficient (COR), the mean squared error (MSE), and the mean absolute error (MAE) to evaluate the performance. The measure COR is the correlation coefficient between the estimated and the true worker abilities over all the groups i [ G ] and timepoints t [ T ] 0 , given by
C O R = i [ G ] t [ T ] 0 ( x ^ i , t x ^ ¯ ) ( x i , t x ¯ ) i [ G ] t [ T ] 0 ( x ^ i , t x ^ ¯ ) 2 i [ G ] t [ T ] 0 ( x i , t x ¯ ) 2 ,
where x ^ ¯ = 1 G ( T + 1 ) i [ G ] t [ T ] 0 x ^ i , t and x ¯ = 1 G ( T + 1 ) i [ G ] t [ T ] 0 x i , t are the means of the estimated and true abilities, respectively. On the other hand, the other measures of MSE and MAE are based on the differences between the true and estimated item qualities over all the groups i [ G ] and timepoints t [ T ] 0 , given by
M S E = 1 G ( T + 1 ) i [ G ] t [ T ] 0 ( x ^ i , t x i , t ) 2 , M A E = 1 G ( T + 1 ) i [ G ] t [ T ] 0 | x ^ i , t x i , t | .
Table 2 shows the COR, MSE, and MAE results across 30 datasets. The average COR values are almost equal to 1, which suggests that the proposed algorithm properly estimates the true ability of worker groups. The average MSE and MAE values turn out to be close to 0, which suggests that the difference between the true and estimated abilities is almost the same. It is reasonable to have fewer discrepancies between the true and estimated values in the case of large N w since we have more observations.
We also depict the three measures in Figure 2, where the three measures are computed at each time t [ T ] 0 . The C O R ( t ) at each time t [ T ] 0 is given by
C O R ( t ) = i [ G ] ( x ^ i , t x ^ ¯ t ) ( x i , t x ¯ t ) i [ G ] ( x ^ i , t x ^ ¯ t ) 2 i [ G ] ( x i , t x ¯ t ) 2 ,
where x ^ ¯ t = 1 G i [ G ] x ^ i , t and x ¯ t = 1 G i [ G ] x i , t are the means of the estimated and true abilities at each time, respectively. The other two measures are represented by
M S E ( t ) = 1 G i [ G ] ( x ^ i , t x i , t ) 2 , M A E ( t ) = 1 G i [ G ] | x ^ i , t x i , t | ,
at each time t [ T ] 0 . We can see the errors of estimates tend to be uniform over time, implying that the ability estimates can be reliable at any time point.
Figure 3 presents the true and estimated abilities for each group i [ G ] for the first dataset of the case N w = 50 . The true abilities tend to lie within the ± 2 standard deviation of the L = 200 particle sequences, implying that the presented algorithm estimates the true abilities very well. The results for groups 1 and 6 indicate that the estimates can capture increasing or decreasing patterns as well. Figure 3 on the right-hand side shows the analogous plot for the working rate given in (1), suggesting that the ability estimates are useful for estimating the working rates. Note that the estimated working rates are quite accurate for a broad range of working rates. In conclusion, the simulation result demonstrates that the estimation/inference works well for both model parameters and workers’ abilities.
In Figure 4, we compare the expected execution time of the uncoded scheme, the uniform load allocation with code rate 1 2 , the load allocation in (8) with the true working rates, and the load allocation in (8) with estimated working rates. For performance comparisons, the following simulations are conducted with the fixed parameters m = 1 and σ = 1 , σ ϵ = 0.01 . Moreover, we set G = 2 , ( N 1 , N 2 ) = ( 20 , 30 ) , T = 20 , and the true working rate ( μ 1 , μ 2 ) = ( 1 , 5 ) at time t = 0 . At time T, the true working rate varies to ( μ 1 , T , μ 2 , T ) = ( 1.0109 , 4.9999 ) . We obtain the estimated working rates of the groups ( μ ^ 1 , T , μ ^ 2 , T ) = ( 0.9892 , 4.8364 ) using the proposed algorithms. This simulation shows that the load allocation in (8) with the estimated working rates can achieve the same performance as the optimal load allocation in (8) with the true working rates. It is also observed that the load allocation in (8) with the estimated working rates shows 54% and 14% reductions in the expected execution times compared to the uncoded scheme and the uniform load allocation with code rate 1 2 , respectively. This result demonstrates that our estimation and the inference of the latent variable are valid in terms of the expected execution time.

5. Concluding Remarks

In this paper, we model the time-varying ability of workers in heterogeneous distributed computing as latent variables. Since we allow for the ability to change over time, we employ a particle method to infer the latent variables. We present a method for estimating the parameters of the working rate with the latent variable using the EM algorithm combined with sequential Monte Carlo (SMC) for filtering and FFBSa. Monte Carlo simulations verify that the proposed algorithm works reasonably well in estimating the workers’ ability as well as the model parameters. Particularly, the estimation of workers’ ability shows great performance in terms of two measures: COR and MSE between the true and estimated ability.
Exploiting the proposed estimation of the workers’ ability, one can devise the optimal load allocation for minimizing the latency in the heterogeneous distributed computation with workers having time-varying abilities. Numerical simulations show that the load allocation with the estimated working rates achieves the theoretical limit of the expected execution time and reduces the expected execution time by up to 54% compared to existing schemes.

Author Contributions

Conceptualization, D.K. and H.J.; methodology, D.K. and H.J.; software, D.K., S.L. and H.J.; validation, D.K. and H.J.; formal analysis, D.K. and H.J.; writing—original draft preparation, D.K., S.L. and H.J.; writing—review and editing, D.K. and H.J.; supervision, D.K. and H.J.; project administration, D.K. and H.J.; funding acquisition, H.J. All authors have read and agreed to the published version of the manuscript.

Funding

This research is supported in part by a National Research Foundation of Korea (NRF) grant funded by the Korean government (no. 2021R1G1A109410312).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Notations

SymbolDescription
NThe number of workers participating in the matrix multiplication task
N i The number of workers in group i
GThe number of groups
A t ( A ˜ t )(coded) Data matrix in R k × d at time t
A i , j , t ( A ˜ i , j , t )(Coded) data matrix in R l i , t × d at time t allocated to worker j in group i
A i , t ( A ˜ i , t )(Coded) data matrix in R N i l i , t × d at time t allocated to the workers in group i,
i.e., A i , t = [ A i , 1 , t ; A i , 2 , t ; , A i , N i , t ] ( A ˜ i , t = [ A ˜ i , 1 , t ; A ˜ i , 2 , t ; , A ˜ i , N i , t ] )
b t Input vector in R d × 1 at time t
l i , t The number of rows of the matrix allocated to workers in group i at time t
X i , t The ability of workers in group i for a subtask assigned at time t
μ i , t Working rate of workers in group i at time t
η i , j , t Observed variable that indicates the runtime of worker j in group i at time t
N ( m , σ 2 ) Normal distribution with the mean m and variance σ 2
θ Model parameters, i.e., θ = ( m , σ )
f 0 ( x i , 0 | θ = ( m , σ ) ) Probability density function (pdf) of the initial worker’s ability X i , 0 in group i,
where X i , 0 follows a normal distribution N ( m , σ 2 )
f ( x i , t | x i , t 1 ) pdf of worker’s ability X i , t in group i at time t provided the worker’s ability
X i , t 1 at the previous time
g ( η i , j , t | x i , t , l i , t ) pdf of the runtime η i , j , t of worker j in group i at time t, given the workers’
ability X i , t and load allocation l i , t
ϵ i , t                            White noise that is independent across group i and time t and follows normal
distribution N ( 0 , σ ϵ 2 )
T i Task runtime taken to calculate the multiplication of A ˜ i , t and b t at a worker
in group i at time t
T Task runtime for the master to finish the given task

References

  1. Dean, J.; Corrado, G.; Monga, R.; Chen, K.; Devin, M.; Mao, M.; Senior, A.; Tucker, P.; Yang, K.; Le, Q.V. Large scale distributed deep networks. Proc. Adv. Neural Inform. Process. Syst. (NIPS) 2012, 1, 1223–1231. [Google Scholar]
  2. Dean, J.; Barroso, L.A. The tail at scale. Commun. ACM 2013, 56, 74–80. [Google Scholar] [CrossRef]
  3. Lee, K.; Lam, M.; Pedarsani, R.; Papailiopoulos, D.; Ramchandran, K. Speeding up distributed machine learning using codes. IEEE Trans. Inf. Theory 2018, 64, 1514–1529. [Google Scholar] [CrossRef]
  4. Lee, K.; Suh, C.; Ramchandran, K. High-dimensional coded matrix multiplication. In Proceedings of the 2017 IEEE International Symposium on Information Theory (ISIT), Aachen, Germany, 25–30 June 2017; pp. 2418–2422. [Google Scholar]
  5. Yu, Q.; Maddah-Ali, M.; Avestimehr, S. Polynomial codes: An optimal design for high-dimensional coded matrix multiplication. In Proceedings of the 31st Conference on Neural Information Processing Systems (NIPS 2017), Los Angeles, CA, USA, 4–9 December 2017; pp. 4403–4413. [Google Scholar]
  6. Park, H.; Lee, K.; Sohn, J.-Y.; Suh, C.; Moon, J. Hierarchical coding for distributed computing. In Proceedings of the 2018 IEEE International Symposium on Information Theory (ISIT), Vail, CO, USA, 17–22 June 2018; pp. 1630–1634. [Google Scholar]
  7. Tandon, R.; Lei, Q.; Dimakis, A.G.; Karampatziakis, N. Gradient coding: Avoiding stragglers in distributed learning. In Proceedings of the International Conference Machine Learning, Sydney, Australia, 6–11 August 2017; pp. 3368–3376. [Google Scholar]
  8. Raviv, N.; Tandon, R.; Dimakis, A.; Tamo, I. Gradient coding from cyclic MDS codes and expander graphs. In Proceedings of the International Conference Machine Learning, Stockholm, Sweden, 10–15 July 2018; pp. 4305–4313. [Google Scholar]
  9. Ozfaturay, E.; Gündüz, D.; Ulukus, S. Speeding up distributed gradient descent by utilizing non-persistent stragglers. In Proceedings of the 2019 IEEE International Symposium on Information Theory (ISIT), Paris, France, 7–12 July 2019. [Google Scholar]
  10. Dutta, S.; Cadambe, V.; Grover, P. Coded convolution for parallel and distributed computing within a deadline. In Proceedings of the 2017 IEEE International Symposium on Information Theory (ISIT), Aachen, Germany, 25–30 June 2017; pp. 2403–2407. [Google Scholar]
  11. Kosaian, J.; Rashmi, K.V.; Venkataraman, S. Learning-Based Coded Computation. IEEE J. Sel. Areas Commun. 2020, 1, 227–236. [Google Scholar] [CrossRef]
  12. Li, S.; Maddah-Ali, M.A.; Avestimehr, A.S. Coding for distributed fog computing. IEEE Commun. Mag. 2017, 55, 34–40. [Google Scholar] [CrossRef]
  13. Fu, X.; Wang, Y.; Yang, Y.; Postolache, O. Analysis on cascading reliability of edge-assisted Internet of Things. Reliab. Eng. Syst. Saf. 2022, 223, 108463. [Google Scholar] [CrossRef]
  14. Zaharia, M.; Konwinski, A.; Joseph, A.D.; Katz, R.H.; Stoica, I. Improving mapreduce performance in heterogeneous environments. In Proceedings of the USENIX Symposium on Operating Systems Design Implement (OSDI), San Diego, CA, USA, 8–10 December 2008; pp. 29–42. [Google Scholar]
  15. Reisizadeh, A.; Prakash, S.; Pedarsani, R.; Avestimehr, S. Coded computation over heterogeneous clusters. IEEE Trans. Inf. Theory. 2019, 65, 4227–4242. [Google Scholar] [CrossRef]
  16. Kim, D.; Park, H.; Choi, J.K. Optimal load allocation for coded distributed computation in heterogeneous clusters. IEEE Trans. Commun. 2021, 69, 44–58. [Google Scholar] [CrossRef]
  17. Gao, J. Machine Learning Applications for Data Center Optimization. Google White Pap. Available online: https://research.google/pubs/pub42542/ (accessed on 6 April 2023).
  18. Tang, B.; Matteson, D.S. Probabilistic transformer for time series analysis. Adv. Neural Inf. Process. Syst. 2021, 34, 23592–23608. [Google Scholar]
  19. Lin, Y.; Koprinska, I.; Rana, M. SSDNet: State space decomposition neural network for time series forecasting. In Proceedings of the 2021 IEEE International Conference on Data Mining (ICDM), Auckland, New Zealand, 7–10 December 2021; pp. 370–378. [Google Scholar]
  20. Hu, Y.; Jia, X.; Tomizuka, M.; Zhan, W. Causal-based time series domain generalization for vehicle intention prediction. In Proceedings of the 2022 International Conference on Robotics and Automation (ICRA), Philadelphia, PA, USA, 23–27 May 2022; pp. 7806–7813. [Google Scholar]
  21. Jung, H.; Lee, J.G.; Kim, S.H. On the analysis of fitness change: Fitness-popularity dynamic network model with varying fitness. J. Stat. Mech. Theory Exp. 2020, 4, 043407. [Google Scholar] [CrossRef]
  22. Kitagawa, G. Monte Carlo filter and smoother for non-Gaussian nonlinear state space models. J. Comput. Graph. Stat. 1996, 5, 1–25. [Google Scholar]
  23. Kantas, N.; Doucet, A.; Singh, S.S.; Maciejowski, J.; Chopin, N. On particle methods for parameter estimation in state-space models. Stat. Sci. 2015, 30, 328–351. [Google Scholar] [CrossRef]
  24. Pitt, M.K.; Shephard, N. Filtering via simulation: Auxiliary particle filters. J. Am. Stat. Assoc. 1999, 94, 590–599. [Google Scholar] [CrossRef]
  25. Carpenter, J.; Clifford, P.; Fearnhead, P. Improved particle filter for nonlinear problems. IEE Proc. Radar Sonar Navig. 1999, 146, 2–7. [Google Scholar] [CrossRef]
  26. Doucet, A.; De Freitas, N.; Gordon, N. An introduction to sequential Monte Carlo methods. Seq. Monte Carlo Methods Pract. 2001, 3–14. [Google Scholar] [CrossRef]
  27. Doucet, A.; Johansen, A.M. A tutorial on particle filtering and smoothing: Fifteen years later. Handb. Nonlinear Filter 2009, 12, 656–704. [Google Scholar]
  28. Gilks, W.R.; Wild, P. Adaptive rejection sampling for Gibbs sampling. Appl. Statist. 1992, 41, 337–348. [Google Scholar] [CrossRef]
Figure 2. C O R ( t ) , M S E ( t ) , and M A E ( t ) results for the datasets with the cases N w = 10 , 30 , 50 for each time t [ T ] 0 .
Figure 2. C O R ( t ) , M S E ( t ) , and M A E ( t ) results for the datasets with the cases N w = 10 , 30 , 50 for each time t [ T ] 0 .
Applsci 13 04993 g002
Figure 3. The ability (left) and working rate (right) estimates for all of the groups i [ G ] of the first dataset in the case of N w = 50 . The x-axis represents time. The true ability x i , 0 : T and working rate μ i , 0 : T for each group i [ G ] are represented as black solid lines with markers. The estimated ability x ^ i , 0 : T and working rate μ ^ i , 0 : T values are represented as blue solid lines. The dashed line represents the ± 2 standard deviation of L = 200 particle sequences for each group.
Figure 3. The ability (left) and working rate (right) estimates for all of the groups i [ G ] of the first dataset in the case of N w = 50 . The x-axis represents time. The true ability x i , 0 : T and working rate μ i , 0 : T for each group i [ G ] are represented as black solid lines with markers. The estimated ability x ^ i , 0 : T and working rate μ ^ i , 0 : T values are represented as blue solid lines. The dashed line represents the ± 2 standard deviation of L = 200 particle sequences for each group.
Applsci 13 04993 g003
Figure 4. Performance comparison between the uncoded scheme, uniform load allocation with code rate 1 2 , load allocation in (8) with the estimated working rates, load allocation in (8) with the true working rates, and the theoretical limit of the expected execution time.
Figure 4. Performance comparison between the uncoded scheme, uniform load allocation with code rate 1 2 , load allocation in (8) with the estimated working rates, load allocation in (8) with the true working rates, and the theoretical limit of the expected execution time.
Applsci 13 04993 g004
Table 1. Parameter estimates for the datasets with the cases N w = 10 , 30 , 50 . The true parameter is set as m = 1 , σ = 1 . We present the mean and standard deviation of 30 repetitions for each case.
Table 1. Parameter estimates for the datasets with the cases N w = 10 , 30 , 50 . The true parameter is set as m = 1 , σ = 1 . We present the mean and standard deviation of 30 repetitions for each case.
N w m σ
MeanSt. Dev.MeanSt. Dev.
101.01490.31020.97200.2512
301.01330.25380.95680.2086
501.03820.26071.03950.1977
Table 2. COR, MSE, and MAE results for the datasets with the cases N w = 10 , 30 , 50 . We present the mean and standard deviations of 30 repetitions for each case.
Table 2. COR, MSE, and MAE results for the datasets with the cases N w = 10 , 30 , 50 . We present the mean and standard deviations of 30 repetitions for each case.
N w CORMSEMSE
MeanSt. Dev.MeanSt. Dev.MeanSt. Dev.
100.99670.00200.00650.00340.06150.0125
300.99900.00060.00200.00070.03570.0069
500.99930.00040.00160.00050.03190.0054
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kim, D.; Lee, S.; Jung, H. A Time-Series Model for Varying Worker Ability in Heterogeneous Distributed Computing Systems. Appl. Sci. 2023, 13, 4993. https://doi.org/10.3390/app13084993

AMA Style

Kim D, Lee S, Jung H. A Time-Series Model for Varying Worker Ability in Heterogeneous Distributed Computing Systems. Applied Sciences. 2023; 13(8):4993. https://doi.org/10.3390/app13084993

Chicago/Turabian Style

Kim, Daejin, Suji Lee, and Hohyun Jung. 2023. "A Time-Series Model for Varying Worker Ability in Heterogeneous Distributed Computing Systems" Applied Sciences 13, no. 8: 4993. https://doi.org/10.3390/app13084993

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop