Next Article in Journal
Matter-Aggregating Systems at a Classical vs. Quantum Interface
Previous Article in Journal
Network Geometry of Borsa Istanbul: Analyzing Sectoral Dynamics with Forman–Ricci Curvature
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Workload of Queueing Systems with Autocorrelated Service Times

by
Andrzej Chydzinski
Department of Computer Networks and Systems, Silesian University of Technology, Akademicka 16, 44-100 Gliwice, Poland
Entropy 2025, 27(3), 272; https://doi.org/10.3390/e27030272
Submission received: 11 February 2025 / Accepted: 3 March 2025 / Published: 5 March 2025
(This article belongs to the Section Information Theory, Probability and Statistics)

Abstract

:
The queuing model with autocorrelated service times is studied with respect to workload, i.e., the time needed to serve all the customers in the queue. Specifically, new formulas for the probability density of workload, its tail, the average value, and entropy are derived and illustrated using numerical examples. Both time-dependent and steady-state cases are covered. It is also demonstrated that the average workload may reach surprisingly large values, exceeding several times the product of the average queue size and the average service time.

1. Introduction

In every queueing theory textbook, the study of the performance of a system begins with the queue size—its distribution and average value are usually derived first. The reason for this is perhaps that, in most real-life queueing systems, the queue size is the easiest aspect to observe and interpret.
Another important characteristic of queueing systems is workload, understood as the total time required to serve all the customers in the queue, i.e., the time needed to empty the system completely through the service process, assuming new arrivals are suspended.
In simple queueing systems, without autocorrelation in arrival or service processes, and with small variances of interarrival and service times, workload is roughly proportional to the queue size. Specifically, if the queue size is X and the average service time is F ¯ , the workload is approximately X F ¯ . In the M/M/1 queueing system, workload is exactly equal to X F ¯ .
However, when the service times are autocorrelated, such intuitive calculations of workload can be completely inaccurate. As will be seen, workload can also be equal to 3 X F ¯ rather than X F ¯ . In other words, knowledge of the queue size provides no information about the system workload when the service times are autocorrelated.
Because of that, in this paper, we will derive several formulas characterizing workload in the queueing model with autocorrelated service times. Specifically, formulas will be derived for the tail of the workload distribution, the probability density of workload, its entropy, and the average value. These formulas will encompass both time-dependent and steady-state workloads. The considered queueing model is universal in the meaning that the interarrival time distribution is general, whereas service times are modeled by the Markovian service process. This feature allows for an alignment of the autocorrelation function of the model with an empirical autocorrelation function. Finally, in addition to presenting new theorems and formulas, numerical examples will be provided to illustrate various workload-related characteristics and compare them with the average queue size.
There are several convincing reasons to study models with autocorrelated service times. For example, human-operated queueing systems may exhibit autocorrelation of service times caused by the operator’s tiredness, resulting in longer service times, or due to his/her learning process, which shortens the service times. Another reason for service autocorrelation in such systems could be the similarity of customers arriving at the queue. For instance, customers calling a call center about the same problem may require similar attendance times. In telecommunications, consecutive packet sizes can be correlated, leading to autocorrelated transmission times. In computing, autocorrelated task execution can result from resource sharing. Specifically, when a service is under heavy load caused by external tasks, it can decelerate the service in the queue of interest, leading to long, autocorrelated service times.
In short, the principal contributions of this study are as follows:
a theorem on the tail of the distribution of workload at arbitrary time;
a theorem on the probability density of workload at arbitrary time;
a theorem on the average workload at arbitrary time;
numeric examples of these characteristics, including a demonstration of counterintuitive values of workload when paired with the queue size.
The following organization applies to the rest of the article. In Section 2, the related work is reviewed, whereas in Section 3, the queueing model in question is formally defined. Section 4 is devoted to the derivation of the workload distribution. Its main results are Theorems 1 and 2, concerning the tail and the probability density of workload, respectively. In addition to that, the entropy of workload is obtained, basing on Theorems 1 and 2. In Section 5, the average workload is derived, with the final result of this section presented in Theorem 3. In Section 6, numeric examples are provided. These include sample calculations of the probability density, distribution tail, entropy, and average workload. Section 7 concludes the study with a few final remarks.

2. Related Work

To the author’s recognition, all the main results of this study, i.e., Theorems 1–3 and numerical examples, are new.
The vast majority of previous work on systems with Markovian service processes has focused solely on steady-state calculations.
Specifically, in [1], the queue size distribution in a model with a finite buffer is derived. In [2], the decay rates of large queue sizes in infinite-buffer models are investigated. In [3], the queue size and response time distributions are obtained for a model with a finite buffer and group service process. In [4], the performance of systems with finite and infinite buffers is evaluated using the queue size distribution along with the response time distribution. In [5], multiple servers with Markovian service processes are introduced. For such a model, the queue size distribution is obtained and accompanied by the idle and partly-idle period calculations, i.e., periods when every server is idle or when some servers are idle, respectively. In [6], the limit of the customer blocking probability, as the buffer size approaches infinity, is studied.
Study [7] is devoted to a differently defined service process. Namely, after an idle period, the service process returns to a fixed phase. Thus, in this model, the autocorrelation of service times isolated by an idle period is broken. Typically, in Markovian service processes, it is assumed that the correlation of service times is preserved through the idle period, i.e., the service phase remains the same as it was just before the idle period. However, both approaches may be useful in practice.
In [8], the queue with group arrivals and a finite buffer is investigated under two possible acceptance rules. The first assumes that, when the buffer is nearly full, part of the arriving group of customers can be accepted, filling the buffer completely, while the remainder is blocked. The second rule blocks the entire group in such cases. Several performance characteristics are derived for both variants, i.e., the queue size and waiting time distributions, and the blocking probability for the first, the last, and an arbitrary customer in an arriving group.
In [9], the model with group services and an infinite buffer is investigated with respect to the queue size distribution at arrival and arbitrary times. Additionally, the response time is calculated. In [10], roots of the characteristic equation are used to derive the queue size distribution for a model with single services and an infinite buffer. In [11], the response time and the autocorrelation of inter-departure times are studied in the infinite buffer model.
Two articles [12,13] focus on models with group arrivals and infinite buffers. One [12] examines the queue size along with idle and busy periods, while the other [13] investigates the distribution of waiting times.
The model with group arrivals is studied again in [14], but under a finite-buffer assumption. For this model, the average waiting time of a random customer, as well as the first and last customers in a group, is obtained. Additionally, the probability of blocking an arbitrary customer and the probability of blocking k or more consecutive customers during a busy period are derived. In [15], the model with group services and a finite buffer is revisited, aiming to optimize a profit function that combines all key performance characteristics, e.g., average queue size, waiting time, blocking probability, and others.
Article [16] compares two models with Markovian service processes regarding their queue sizes. In the first model, the service phase remains constant during idle periods, whereas in the second, it changes according to a Markov chain. In both models, group service is assumed.
The only published study involving the time-dependent analysis of a model with a Markovian service process [17] is devoted solely to the queue size. As will be seen, knowledge of the queue size provides no insight into workload when service times are autocorrelated.
Lastly, it is worth noting that the references discussed above focus on models with general interarrival distributions, as is the case here. Other studies with Markovian service processes are based either on specific interarrival distributions (e.g., Poisson, MAP, BMAP) or on discrete queueing systems with discrete-time Markovian service, e.g., [18,19,20,21,22,23,24,25,26,27].

3. Model

The model of interest is the model of a queueing system with a single server. Specifically, the next customer arrives to the queue after time is distributed according to G ( t ) , where G ( t ) is some distribution function on ( 0 , ) , not further specified, with the average value: G ¯ ,
G ¯ = 0 [ 1 G ( t ) ] d t < ,
and the Laplace–Stieltjes transform,
g ( s ) = 0 e s x d G ( x ) .
All interarrival intervals are mutually independent.
The queue of customers is formed following the arrival order and served from the head.
Service times of customers are mutually dependent and follow the Markovian service process [4]. Namely, there is an underlying random process S ( t ) , called the service phase at time t. S ( t ) may assume M values, 1 , , M . To fully parameterize S ( t ) , we need two matrices of size M × M , L 0 , and L 1 , such that L 0 is negative on the diagonal and nonnegative off the diagonal, and L 1 is nonnegative, while L = L 0 + L 1 is a transition-rate matrix, with every row summing to 0.
The evolution of S ( t ) produces various service times of different customers, in the interdependent manner, as follows. When the number of customers in the system at some t is non-zero and S ( t ) = i , then, immediately after t, at t + Δ , the service phase may jump to j with probability ( L 0 ) i j Δ + o ( Δ ) , and the ongoing service of a customer is continued, or the phase may jump to j with probability ( L 1 ) i j Δ + o ( Δ ) , and the present service is completed, so the customer is departed from the system.
If the number of customers in the system becomes zero after a departure, then the evolution of the service phase is terminated for the whole idle period. Upon a new customer arrival after that period, S ( t ) starts evolving again according to the rule given in the previous paragraph, beginning with the same phase it had just before the idle period. (Note that from this definition it follows that S ( t ) is not a Markov process).
The system volume is finite and equal to B customers, including the one being served. If, at the time of a new arrival, there are B customers in the system, the new customer is blocked. Specifically, he/she leaves unserved the system and never arrives again.
It is easy to check that the average service time is
F ¯ = 1 π L 1 e ,
where e denotes the column vector of 1’s, whereas π denotes the stationary vector of transition-rate matrix L .
The correlation of two arbitrary service times can be computed in the same way as in the case of Markovian arrival processes (see, e.g., [28,29]). Specifically, we have
C o r r ( n ) = C o r r ( s k , s k + n ) = α L 0 1 K ( K n 1 e α ) L 0 1 K e 2 π L 1 e π L 0 1 e + ( π L 1 e ) 2 ,
where s k is the k-th service time, s k + n is the service time which follows after other n 1 completed services, α is the stationary vector for transition-rate matrix K I , I is the identity matrix, and
K = L 0 1 ( L L 0 ) .
In the derivations of the next section, a vital role will be undertaken by two characteristics of the service process, p j , k , l ( x ) and q j , k , l ( x ) . The former, p j , k , l ( x ) , is defined as the probability that l customers are served in period ( 0 , x ) , and the service phase is k at the end of the l-th service time, assuming X ( 0 ) > l and S ( 0 ) = j . The latter, q j , k , l ( x ) , is defined as the probability that l customers are served in ( 0 , x ) , and the service phase is k at time x, assuming X ( 0 ) > l and S ( 0 ) = j . It will subsequently be shown how these two characteristics can be calculated.
By X ( t ) , we will represent the queue size at t, including the customer under service, if applicable. We also adopt the convention that every time the phrase “queue size” is used, it means “including the customer under service, if any”.
Lastly, it is presumed that the first arrival time has distribution G ( t ) , i.e., the arrival process is not shifted at the time origin.

4. Workload Distribution and Entropy

Let ω ( t ) denote workload at time t, i.e., the time needed to empty the system completely by the service process, assuming there were no new arrivals after t. Obviously, ω ( t ) is a random variate, so it has to be characterized by the probability distribution. For technical reasons, it is easier to start with calculations of the tail of the distribution of ω ( t ) , namely,
W m , j ( t , u ) = P { ω ( t ) > u | X ( 0 ) = m , S ( 0 ) = j } .
Having the tail, we can procure the distribution function of the workload:
W ¯ m , j ( t , u ) = 1 W m , j ( t , u ) .
We start with constructing integral equations for W m , j ( t , u ) . Firstly, for an initially empty system, X ( 0 ) = 0 , and any j { 1 , , M } , we have
W 0 , j ( t , u ) = 0 t W 1 , j ( t x , u ) d G ( x ) .
This system is obtained by conditioning upon the time of arrival of the first customer, x. Indeed, when no customer is present, the service phase remains the same, i.e., j. When the first customers arrives, the service begins with this phase and a queue size of 1. Thus, the new tail of the workload distribution at time t x is W 1 , j .
Secondly, for an initially full system, X ( 0 ) = B , and any j { 1 , , M } , we get
W B , j ( t , u ) = k = 1 M 0 t q j , k , 0 ( x ) W B , k ( t x , u ) d G ( x ) + l = 1 B 1 k = 1 M 0 t q j , k , l ( x ) W B l + 1 , k ( t x , u ) d G ( x ) + k = 1 M 0 t p j , k , B ( x ) W 1 , k ( t x , u ) d G ( x ) + ( 1 G ( t ) ) l = 0 B 1 k = 1 M q j , k , l ( t ) a = 0 B l 1 i = 1 M q k , i , a ( u ) .
The first integral of (9) handles the condition of zero services completed in ( 0 , x ) . Under such a condition, the tail of the workload distribution at time t x is W B , k because the customer arriving at x is blocked. The second integral of (9) handles the condition of l services completed in ( 0 , x ) , where 0 < l < B . Under such a condition, the size of the queue is reduced by time x, but not to zero. The new tail of the workload distribution at t x is W B l + 1 , k because the customer arriving at x is allowed to the queue. The third integral of (9) handles the condition of B services completed in ( 0 , x ) , which makes the system empty by x. Thus, counting the new customer arriving at x, the new tail of the workload distribution at t x is W l , k . The last component of (9) handles the condition of no new arrivals by t. Specifically, to have a non-zero tail of the workload distribution at t under this condition, the system must not become empty by t, which happens with the following probability:
l = 0 B 1 k = 1 M q j , k , l ( t ) .
Furthermore, under such condition, we have
P { ω ( t ) > u } = a = 0 B l 1 i = 1 M q k , i , a ( u ) ,
which completes the explanation of (9).
For a partially occupied system, 0 < X ( 0 ) = m < B , and any j { 1 , , M } , we obtain
W m , j ( t , u ) = l = 0 m 1 k = 1 M 0 t q j , k , l ( x ) W m l + 1 , k ( t x , u ) d G ( x ) + k = 1 M 0 t p j , k , m ( x ) W 1 , k ( t x , u ) d G ( x ) + ( 1 G ( t ) ) l = 0 m 1 k = 1 M q j , k , l ( t ) a = 0 m l 1 i = 1 M q k , i , a ( u ) .
Equation (12) can be explained in the same way as (9), with the exception that (12) has no analog of the first part of (9). This is because, in the case of X ( 0 ) < B , no customer can be blocked by the first arrival time, x.
Denoting
w m , j ( s , u ) = 0 e s x W m , j ( x , u ) d x ,
x j , k , l ( s ) = 0 e s x p j , k , l ( x ) d G ( x ) ,
y j , k , l ( s ) = 0 e s x q j , k , l ( x ) d G ( x ) ,
c j , k , l ( s ) = 0 e s x q j , k , l ( x ) 1 G ( x ) d x ,
we can restate Equation (8) as
w 0 , j ( s , u ) = g ( s ) w 1 , j ( s , u ) ,
while (9) can be restated as
w B , j ( s , u ) = k = 1 M y j , k , 0 ( s ) w B , k ( s , u ) + l = 1 B 1 k = 1 M y j , k , l ( s ) w B l + 1 , k ( s , u ) + k = 1 M x j , k , B ( s ) w 1 , k ( s , u ) + l = 0 B 1 a = 0 B l 1 k = 1 M i = 1 M c j , k , l ( s ) q k , i , a ( u ) ,
and (12) as
w m , j ( s , u ) = l = 0 m 1 k = 1 M y j , k , l ( s ) w m l + 1 , k ( s , u ) + k = 1 M x j , k , m ( s ) w 1 , k ( s , u ) + l = 0 m 1 a = 0 m l 1 i = 1 M k = 1 M c j , k , l ( s ) q k , i , a ( u ) , 0 < m < B .
Employing the vector
w m ( s , u ) = [ w m , 1 ( s , u ) , , w m , M ( s , u ) ] T ,
and matrices
X l ( s ) = [ x j , k , l ( s ) ] j = 1 , , M ; k = 1 , , M ,
Y l ( s ) = [ y j , k , l ( s ) ] j = 1 , , M ; k = 1 , , M ,
to (17), (18), and (19) yields
w 0 ( s , u ) = g ( s ) w 1 ( s , u ) ,
w B ( s , u ) = Y 0 ( s ) w B ( s , u ) + l = 1 B 1 Y l ( s ) w B l + 1 ( s , u ) + X B ( s ) w 1 ( s , u ) + b B ( s , u )
w m ( s , u ) = l = 0 m 1 Y l ( s ) w m l + 1 ( s , u ) + X m ( s ) w 1 ( s , u ) + b m ( s , u ) , 0 < m < B ,
respectively, with
b m ( s , u ) = l = 0 m 1 C l ( s ) a = 0 m l 1 Q a ( u ) e ,
Q l ( u ) = [ q j , k , l ( u ) ] j = 1 , , M ; k = 1 , , M ,
C l ( s ) = [ c j , k , l ( s ) ] j = 1 , , M ; k = 1 , , M .
Before presenting the solution of system (23)–(25), it ought to be underlined that all the coefficient matrices in (23)–(25) can be calculated numerically. Specifically, Y l ( s ) , C l ( s ) , and Q l ( u ) can be computed using the uniformization algorithm [30,31], whereas X l ( s ) can be obtained using the formulas proven in [17], namely,
X l ( s ) = C l 1 ( s ) L 1 s H l ( s ) , l 1 ,
with
H 1 ( s ) = C 0 ( s ) L 0 1 L 1 g ¯ ( s ) L 0 1 L 1 , g ¯ ( s ) = 0 e s x ( 1 G ( x ) ) d x ,
H l + 1 ( s ) = C l ( s ) L 0 1 L 1 H l ( s ) L 0 1 L 1 , l 1 .
We can now move to finding the solutions of (23)–(25). Specifically, applying the lemma of [32], p. 200, we realize that the solution of (25) is
w m ( s , u ) = R m ( s ) c ( s , u ) + l = 1 m R m l ( s ) a l ( s , u ) ,
where vector c ( s , u ) is unknown, but independent of m, whereas
a m ( s , u ) = Y m ( s ) w 1 ( s , u ) X m ( s ) w 1 ( s , u ) b m ( s , u ) ,
R 0 ( s ) = 0 , R 1 ( s ) = Y 0 1 ( s ) ,
R m + 1 ( s ) = Y 0 1 ( s ) ( R m ( s ) l = 0 m Y l + 1 ( s ) R m l ( s ) ) , m > 0 ,
and 0 represents the square zero matrix. Inserting m = 1 into (32) yields
c ( s , u ) = Y 0 ( s ) w 1 ( s , u ) .
Then, using (32) together with (33) and (36), we have
w m ( s , u ) = E m ( s ) w 1 ( s , u ) l = 1 m R m l ( s ) b l ( s , u ) , m > 0 ,
with
E m ( s ) = R m ( s ) Y 0 ( s ) + l = 1 m R m l ( s ) Y l ( s ) l = 1 m R m l ( s ) X l ( s ) .
What is left to find is the unknown w 1 ( s , u ) occurring in (37). After that, (37) will be fully solved. Fortunately, w 1 ( s , u ) can be found by utilizing (37) with m = B and leveraging (24) in addition to that. The final results can be gathered as follows.
Theorem 1.
The transform of the tail of the workload distribution in the model with a Markovian service process is
w 0 ( s , u ) = g ( s ) A ( s ) l = 1 B 1 Y l ( s ) k = 1 B l + 1 R B l + 1 k ( s ) b k ( s , u ) + ( Y 0 ( s ) I ) k = 1 B R B k ( s ) b k ( s , u ) b B ( s , u ) ,
w 1 ( s , u ) = w 0 ( s , u ) / g ( s ) ,
w m ( s , u ) = E m ( s ) w 0 ( s , u ) / g ( s ) k = 1 m R m k ( s ) b k ( s , u ) , 1 < m B ,
where
A ( s ) = l = 1 B 1 Y l ( s ) E B l + 1 ( s ) E B ( s ) + Y 0 ( s ) E B ( s ) + X B ( s ) 1 ,
while Y l ( s ) , R l ( s ) , b k ( s , u ) , E m ( s ) , and X B ( s ) are given in (15), (35), (26), (38), and (29), respectively.
Now we can proceed to calculations of the density of workload f m , j ( t , u ) . By definition, we have
f m , j ( t , u ) = d W ¯ m , j ( t , u ) d u , u > 0 .
(As we can see, this definition is restricted to u > 0 . The distribution of workload for u = 0 will be discussed below).
Denote
f m , j * ( s , u ) = 0 e s x f m , j ( x , u ) d x ,
and
f m ( s , u ) = [ f m , 1 * ( s , u ) , , f m , M * ( s , u ) ] T .
We have obviously
f m ( s , u ) = d w m ( s , u ) d u .
Therefore, we have to compute the derivative of (39)–(41) with respect to u. In fact, variable u occurs in (39)–(41) only in w k ( s , u ) and b k ( s , u ) , defined in (26). In order to differentiate (26), we can use the well-known formulas
d Q l ( u ) d u = Q l ( u ) L 0 + Q l 1 ( u ) L 1 , l 1 ,
Q 0 ( u ) = e L 0 u , d Q 0 ( u ) d u = L 0 e L 0 u ,
which give
b m ( s , u ) = d b m ( s , u ) d u = l = 0 m 1 C l ( s ) a = 1 m l 1 Q a ( u ) L 0 + a = 1 m l 1 Q a 1 ( u ) L 1 + L 0 e L 0 u e .
The latter formula, accompanied with (39)–(41), delivers the final result.
Theorem 2.
The transform of the probability density of the workload in the model with a Markovian service process is
f 0 ( s , u ) = g ( s ) A ( s ) l = 1 B 1 Y l ( s ) k = 1 B l + 1 R B l + 1 k ( s ) b k ( s , u ) + ( Y 0 ( s ) I ) k = 1 B R B k ( s ) b k ( s , u ) b B ( s , u ) ,
f 1 ( s , u ) = f 0 ( s , u ) / g ( s ) ,
f m ( s , u ) = E m ( s ) f 0 ( s , u ) / g ( s ) + k = 1 m R m k ( s ) b k ( s , u ) , 1 < m B ,
where b m ( s , u ) is given in (49).
Now we can proceed to the derivation of the entropy of the workload. To accomplish that, we have to observe first that density f m , j ( t , u ) is defective; i.e., in general, it does not integrate to 1 with respect to u. This comes from the fact that if the queue is empty, which has a non-zero probability, then the workload is zero. Thus, the distribution of workload has an atom at u = 0 . The size of this atom, P m , j ( t ) , is the following:
P m , j ( t ) = P { ω ( t ) = 0 | X ( 0 ) = m , S ( 0 ) = j } = 1 W m , j ( t , 0 ) .
In this way, the distribution of workload is in fact a mixture of a discrete distribution, which assumes only a value of 0, and a continues distribution, having probability density f m , j ( t , u ) / 1 P m , j ( t ) . The two-point mixing distribution is the following: P m , j ( t ) , 1 P m , j ( t ) .
Denote by h m , j ( t ) the entropy of workload at time t given X ( 0 ) = m , S ( 0 ) = j , while h m , j d ( t ) denotes the entropy of its discrete part, h m , j c ( t ) denotes the entropy of its continues part, and h m , j * ( t ) denotes the entropy of the mixing distribution. Adopting this notation yields
h m , j ( t ) = h m , j * ( t ) + P m , j ( t ) h m , j d ( t ) + 1 P m , j ( t ) h m , j c ( t ) ,
(see, e.g., Formula (7) in [33]). Furthermore, we have
h m , j * ( t ) = P m , j ( t ) log P m , j ( t ) 1 P m , j ( t ) log 1 P m , j ( t ) ,
h m , j d ( t ) = 0 ,
and
h m , j c ( t ) = 0 f m , j ( t , u ) 1 P m , j ( t ) log f m , j ( t , u ) 1 P m , j ( t ) d u = log 1 P m , j ( t ) 1 1 P m , j ( t ) 0 f m , j ( t , u ) log f m , j ( t , u ) d u .
Lastly, combining (54) with (55)–(57) and (53), we arrive at the final formula for the entropy of the workload:
h m , j ( t ) = 1 W m , j ( t , 0 ) log 1 W m , j ( t , 0 ) 0 f m , j ( t , u ) log f m , j ( t , u ) d u ,
where W m , j ( t , 0 ) can be obtained from Theorem 1, while f m , j ( t , u ) can be obtained from Theorem 2.
Indeed, Theorems 1 and 2 can be used to calculate numerically W m , j ( t , u ) and f m , j ( t , u ) for a specific t, as well as in the steady state, for t = . To calculate the steady-state values, W m , j ( , u ) and f m , j ( , u ) , we have to calculate numerically the limits s 0 + of s w 1 ( s , u ) and s f 1 ( s , u ) , respectively, directly from Theorems 1 and 2. This follows from the well-known FVT theorem. To compute these characteristics at a specific t, we have to employ a numeric inversion formula for the transform [34].

5. Average Value

Define
V m , j ( t ) = E { ω ( t ) | X ( 0 ) = m , S ( 0 ) = j } .
For an initially empty system, X ( 0 ) = 0 , and any j { 1 , , M } , we have
V 0 , j ( t ) = 0 t V 1 , j ( t x ) d G ( x ) ,
which can be rationalized in the same manner as (8). For an initially full system, X ( 0 ) = B , and any j { 1 , , M } , we obtain
V B , j ( t ) = k = 1 M 0 t q j , k , 0 ( x ) V B , k ( t x ) d G ( x ) + l = 1 B 1 k = 1 M 0 t q j , k , l ( x ) V B l + 1 , k ( t x ) d G ( x ) + k = 1 M 0 t p j , k , B ( x ) V 1 , k ( t x ) d G ( x ) + ( 1 G ( t ) ) l = 0 B 1 k = 1 M q j , k , l ( t ) a = 0 B l 1 i = 1 M 0 q k , i , a ( u ) d u .
This equation differs from (9) only in the last term, which handles the condition of no new arrivals by t, which happens with probability l = 0 B 1 k = 1 M q j , k , l ( t ) . Under such a condition, the average workload at t equals a = 0 B l 1 i = 1 M 0 q k , i , a ( u ) d u . Analogously, for a partially occupied system, we obtain
V m , j ( t ) = l = 0 m 1 k = 1 M 0 t q j , k , l ( x ) V m l + 1 , k ( t x ) d G ( x ) + k = 1 M 0 t p j , k , m ( x ) V 1 , k ( t x ) d G ( x ) + ( 1 G ( t ) ) l = 0 m 1 k = 1 M q j , k , l ( t ) a = 0 m l 1 i = 1 M 0 q k , i , a ( u ) d u , 0 < m < B .
System (60)–(62) has the same layout as (8), (9), and (12), except for free terms. Due to that, it can be solved correspondingly, with obvious modifications. Using the notation
v m , j ( s ) = 0 e s x V m , j ( x ) d x ,
v m ( s ) = [ v m , 1 ( s ) , , v m , M ( s ) ] T ,
the final solution has the following form.
Theorem 3.
The transform of the average workload in the model with a Markovian service process is
v 0 ( s ) = g ( s ) A ( s ) · l = 1 B 1 Y l ( s ) k = 1 B l + 1 R B l + 1 k ( s ) d k ( s ) + ( Y 0 ( s ) I ) k = 1 B R B k ( s ) d k ( s ) d B ( s ) ,
v 1 ( s ) = v 0 ( s ) / g ( s ) ,
v m ( s ) = E m ( s ) v 0 ( s ) / g ( s ) k = 1 m R m k ( s ) d k ( s ) , 1 < m B ,
where
d m ( s ) = l = 0 m 1 C l ( s ) a = 0 m l 1 P a e .
P a = 0 q j , k , a ( u ) d u j = 1 , , M ; k = 1 , , M .

6. Examples

In the numeric examples, the following parameter matrices of the Markovian service process are utilized:
L 0 = 0.362348 0.011973 0.009551 0.006509 2.739780 0.016561 0.013568 0.010249 0.063200 ,
L 1 = 0.339476 0.000113 0.001235 0.000405 2.716119 0.000186 0.001879 0.000124 0.037380 .
For these matrices, the average service time is F ¯ = 1 , while the 1-lag autocorrelation is C o r r ( 1 ) = 0.31 .
The interearrival time has a hyperexponential distribution, with parameters (2, 0.4) and (0.75, 0.25), which gives the average interarrival time G ¯ = 1 . Therefore, the system is fully loaded, with ρ = F ¯ / G ¯ = 1 . Finally, the system capacity is 20 customers.
In Figure 1 and Figure 2, the evolution of the average workload and queue size over time is depicted. The average workloads were obtained from Theorem 3 proven here, while the average queue sizes were obtained from Theorem 2 of [17]. Figure 1 differs from Figure 2 by the initial system occupancy, which is 0 in Figure 1 and 20 in Figure 2. In both figures, the steady-state average queue size (i.e., the limit on the right-hand side) is 9.121, while the steady-state average workload is 28.834.
These numbers, and Figure 1 and Figure 2 in general, are quite surprising. Given that the average service time is 1, one might expect that the average queue sizes and average workloads should be close to each other, rather than differ by more than three times. Such a difference emphasizes the need for calculating the workload separately, as the queue sizes computed using the formulas of [17] provide no knowledge of workload.
The following explanation for such a large discrepancy between the average queue size and average workload can be proposed: In the Markovian service process, some service phases slow down the service process below the average, while others speed it up above the average. Moreover, when the autocorrelation is positive, as here, once the service is in a slow phase, it has a tendency to stay in this slow phase for a period of time, increasing the chances of a few long service times in a row. In addition to that, an arbitrarily chosen time t is more likely to be in a long service interval than in a short service interval. To put it differently, it is more likely to be in a slow service phase than in a fast service phase.
Therefore, an arbitrarily chosen t is not only more likely to be in a slow service phase, i.e., within a longer-than-average current service time, but also it is likely that a few subsequent services will also be long. This makes the tail of the workload quite heavy, much heavier than the tail of the queue size, which explains the discrepancy between the queue size and workload in Figure 1 and Figure 2.
To illustrate this further, in Figure 3 and Figure 4, both tails are depicted, i.e., the tail of queue size distribution defined as
T m , j ( t , u ) = P { X ( t ) > u | X ( 0 ) = m , S ( 0 ) = j } ,
and the tail of workload distribution defined in (6). As seen, the tail of workload in Figure 4 decreases much more slowly with u than the tail of the queue size, and this is true for every t.
Finally, the difference between the queue size and workload can also be observed from the information theory perspective. Namely, the entropy of the queue size distribution in the steady state equals to 2.11 nats. (This was calculated from Theorem 1 of [17]). On the other hand, the entropy of the workload distribution in the steady state is 3.63 nats, which was computed by means of Formula (58), with the help of Theorems 1 and 2. As seen, the entropy of the workload is much higher than the entropy of the queue size.
In Figure 5, the probability density of workload over time is depicted for an initially empty system. Five vertical slices of Figure 5, taken at different moments in time, are depicted in Figure 6. As seen in these figures, the probability mass is initially concentrated near u = 0 , but it quickly spreads out as time passes. For a low t of just 20, there is already a significant portion of the probability mass beyond u = 100 .
Note that the steady state has not yet been reached in Figure 5. As indicated in Figure 1, the steady state is reached at around t = 250 .
As discussed in Section 5, the densities of workload presented in Figure 5 and Figure 6 are defective; i.e., they do not integrate to 1. The missing probability is P 0 , 1 ( t ) , which is the probability of zero workload at time t. This probability is depicted in Figure 7. Its steady-state value is 0.312. It is interesting that the evolution of P 0 , 1 ( t ) is not monotonic—it clearly has a minimum shortly after the system is activated.
In Figure 8, the probability density of workload is depicted over time, now for an initially full system. Five vertical slices of Figure 8, taken at different moments in time, are shown in Figure 9. As seen in these figures, the density function can assume a complicated shape, with multiple extrema. For a small t, there is no probability mass near u = 0 , as the system occupancy is non-zero with a probability close to 1. However, this changes quickly with time.
Finally, in Figure 10, the probability that the workload is zero, which is the missing probability mass in the defective density of workload, is depicted over time. Now the evolution of P 0 , 1 ( t ) is monotonic. It obviously converges to the same steady-state value as in Figure 7, i.e., 0.312.

7. Conclusions

We studied the queueing model with autocorrelated service times with respect to workload. The study was motivated by real queueing systems, in which such autocorrelation is quite common. The model we studied was universal in the meaning that the interarrival time distribution was general, while the service times were modeled by the Markovian service process, which allowed aligning the autocorrelation function of the model with an empirical autocorrelation function.
New formulas for the probability density of workload, its tail, entropy, and the average value were derived and illustrated through numeric examples, covering both time-dependent and steady-state workload of the system.
Among other findings, it was demonstrated that, in the model in question, workload can assume surprisingly large values, exceeding several times the product of the queue size and the average service time. This observation highlights the need to calculate workload as a separate characteristic. The queue size alone provides no insight into workload, as one might naively assume.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Data is contained within the article.

Conflicts of Interest

The author declares no conflict of interest.

References

  1. Bocharov, P.P. Stationary distribution of a finite queue with recurrent flow and Markovian service. Autom. Remote Control. 1996, 9, 66–78. (In Russian) [Google Scholar]
  2. Alfa, A.S.; Xue, J.; Ye, Q. Perturbation theory for the asymptotic decay rates in the queues with Markovian arrival process and/or Markovian service process. Queueing Syst. 2000, 36, 287–301. [Google Scholar] [CrossRef]
  3. Chaplygin, V.V. The mass-service G/BMSP/1/r. Inf. Process. 2003, 3, 97–108. (In Russian) [Google Scholar]
  4. Bocharov, P.P.; D’Apice, C.; Pechinkin, A.V.; Salerno, S. The stationary characteristics of the G/MSP/1/r queueing system. Autom. Remote Control. 2003, 64, 288–301. [Google Scholar] [CrossRef]
  5. Albores-Velasco, F.J.; Tajonar-Sanabria, F.S. Analysis of the GI/MSP/c/r queueing system. Inf. Process. 2004, 4, 46–57. [Google Scholar]
  6. Kim, J.; Kim, B. Asymptotic analysis for loss probability of queues with finite GI/M/1 type structure. Queueing Syst. 2007, 57, 47–55. [Google Scholar] [CrossRef]
  7. Gupta, U.C.; Banik, A.D. Complete analysis of finite and infinite buffer GI/MSP/1 queue—A computational approach. Oper. Res. Lett. 2007, 35, 273–280. [Google Scholar] [CrossRef]
  8. Banik, A.D.; Gupta, U.C. Analyzing the finite buffer batch arrival queue under Markovian service process: GIX/MSP/1/N. Top 2007, 15, 146–160. [Google Scholar] [CrossRef]
  9. Banik, A.D. Analyzing state-dependent arrival in GI/BMSP/1/ queues. Math. Comput. Model. 2011, 53, 1229–1246. [Google Scholar] [CrossRef]
  10. Chaudhry, M.L.; Samanta, S.K.; Pacheco, A. Analytically explicit results for the GI/C-MSP/1/ queueing system using roots. Probab. Eng. Informational Sci. 2012, 26, 221–244. [Google Scholar] [CrossRef]
  11. Samanta, S.K. Sojourn-time distribution of the GI/MSP/1 queueing system. Opsearch 2015, 52, 756–770. [Google Scholar] [CrossRef]
  12. Chaudhry, M.L.; Banik, A.D.; Pacheco, A. A simple analysis of the batch arrival queue with infinite-buffer and Markovian service process using roots method: GIX/C-MSP/1/∞. Ann. Oper. Res. 2017, 252, 135–173. [Google Scholar] [CrossRef]
  13. Banik, A.D.; Chaudhry, M.L.; Kim, J.J. A Note on the Waiting-Time Distribution in an Infinite-Buffer GI[X]/C-MSP/1 Queueing System. J. Probab. Stat. 2018, 2018, 7462439. [Google Scholar] [CrossRef]
  14. Banik, A.D.; Ghosh, S.; Chaudhry, M.L. On the consecutive customer loss probabilities in a finite-buffer renewal batch input queue with different batch acceptance/rejection strategies under non-renewal service. In Soft Computing for Problem Solving: SocProS 2017; Springer: Singapore, 2019; Volume 1, pp. 45–62. [Google Scholar]
  15. Banik, A.D.; Ghosh, S.; Chaudhry, M.L. On the optimal control of loss probability and profit in a GI/C-BMSP/1/N queueing system. Opsearch 2020, 57, 144–162. [Google Scholar] [CrossRef]
  16. Samanta, S.K.; Bank, B. Modelling and Analysis of GI/BMSP/1 Queueing System. Bull. Malays. Math. Sci. Soc. 2021, 44, 3777–3807. [Google Scholar] [CrossRef]
  17. Chydzinski, A. Transient GI/MSP/1/N Queue. Entropy 2024, 26, 807. [Google Scholar] [CrossRef]
  18. Ozawa, T. Analysis of Queues with Markovian Service Processes. Stoch. Model. 2004, 20, 391–413. [Google Scholar] [CrossRef]
  19. Zhang, Q.; Heindl, A.; Smirni, E. Characterizing the BMAP/MAP/1 departure process via the ETAQA truncation. Stoch. Models 2005, 21, 821–846. [Google Scholar] [CrossRef]
  20. Horváth, A.; Horváth, G.; Telek, M.A. Joint moments based analysis of networks of MAP/MAP/1 queues. Perform. Eval. 2010, 67, 759–778. [Google Scholar] [CrossRef]
  21. Wang, Y.C.; Chou, J.H.; Wang, S.Y. Loss pattern of D-BMAP/D-MSP/1/K queue and its application in wireless local communications. Appl. Math. Modell. 2011, 35, 1782–1797. [Google Scholar] [CrossRef]
  22. Sandhya, R.; Sundar, V.; Rama, G.; Ramshankar, R.; Ramanarayanan, R. BMAP/BMSP/1 queue with randomly varying environment. ISOR J. Eng. 2015, 5, 1–12. [Google Scholar]
  23. Samanta, S.K.; Chaudhry, M.L.; Pacheco, A. Analysis of BMAP/MSP/1 queue. Methodol. Comput. Appl. Probab. 2016, 18, 419–440. [Google Scholar] [CrossRef]
  24. Ghosh, S.; Banik, A.D. An algorithmic analysis of the BMAP/MSP/1 generalized processor-sharing queue. Comput. Oper. Res. 2017, 79, 1–11. [Google Scholar] [CrossRef]
  25. Ghosh, S.; Banik, A.D. Computing conditional sojourn time of a randomly chosen tagged customer in a BMAP/MSP/1 queue under random order service discipline. Ann. Oper. Res. 2018, 261, 185–206. [Google Scholar] [CrossRef]
  26. Nandi, R.; Samanta, S.K. Stationary analysis of an infinite-buffer D-MAP/D-MSP/1 queueing system. Am. J. Math. Manag. Sci. 2020, 39, 362–382. [Google Scholar] [CrossRef]
  27. Samanta, S.K.; Das, K. Detailed analytical and computational studies of D-BMAP/D-BMSP/1 queueing system. Methodol. Comput. Appl. Probab. 2023, 25, 12. [Google Scholar] [CrossRef]
  28. Chydzinski, A.; Adamczyk, B. Transient and stationary losses in a finite-buffer queue with batch arrivals. Math. Probl. Eng. 2012, 2012, 326830. [Google Scholar] [CrossRef]
  29. Mrozowski, P.; Chydzinski, A. Queues with Dropping Functions and Autocorrelated Arrivals. Methodol. Comput. Appl. Probab. 2018, 20, 97–115. [Google Scholar] [CrossRef]
  30. Latouche, G.; Ramaswami, V. Introduction to Matrix Analytic Methods in Stochastic Modeling; SIAM: Philadelphia, PA, USA, 1999. [Google Scholar]
  31. Dudin, A.N.; Klimenok, V.I.; Vishnevsky, V.M. The Theory of Queuing Systems with Correlated Flows; Springer: Cham, Swizterland, 2020. [Google Scholar]
  32. Chydzinski, A. Time to reach buffer capacity in a BMAP queue. Stoch. Model. 2007, 23, 195–209. [Google Scholar] [CrossRef]
  33. Politis, D.N. Maximum Entropy Modelling of Mixture Distributions. Kybernetes 1994, 23, 49–54. [Google Scholar] [CrossRef]
  34. Cohen, A.M. Numerical Methods for Laplace Transform Inversion; Springer: Berlin/Heidelberg, Germany, 2007. [Google Scholar]
Figure 1. Average workload and queue size in time. X ( 0 ) = 0 , S ( 0 ) = 1 .
Figure 1. Average workload and queue size in time. X ( 0 ) = 0 , S ( 0 ) = 1 .
Entropy 27 00272 g001
Figure 2. Average workload and queue size in time. X ( 0 ) = 20 , S ( 0 ) = 1 .
Figure 2. Average workload and queue size in time. X ( 0 ) = 20 , S ( 0 ) = 1 .
Entropy 27 00272 g002
Figure 3. Tail of queue size distribution in time. X ( 0 ) = 0 , S ( 0 ) = 1 .
Figure 3. Tail of queue size distribution in time. X ( 0 ) = 0 , S ( 0 ) = 1 .
Entropy 27 00272 g003
Figure 4. Tail of workload distribution in time. X ( 0 ) = 0 , S ( 0 ) = 1 .
Figure 4. Tail of workload distribution in time. X ( 0 ) = 0 , S ( 0 ) = 1 .
Entropy 27 00272 g004
Figure 5. Probability density of workload in time. X ( 0 ) = 0 , S ( 0 ) = 1 .
Figure 5. Probability density of workload in time. X ( 0 ) = 0 , S ( 0 ) = 1 .
Entropy 27 00272 g005
Figure 6. Probability density of workload in selected moments. X ( 0 ) = 0 , S ( 0 ) = 1 .
Figure 6. Probability density of workload in selected moments. X ( 0 ) = 0 , S ( 0 ) = 1 .
Entropy 27 00272 g006
Figure 7. Probability that workload is zero versus time. X ( 0 ) = 0 , S ( 0 ) = 1 .
Figure 7. Probability that workload is zero versus time. X ( 0 ) = 0 , S ( 0 ) = 1 .
Entropy 27 00272 g007
Figure 8. Probability density of workload in time. X ( 0 ) = 20 , S ( 0 ) = 1 .
Figure 8. Probability density of workload in time. X ( 0 ) = 20 , S ( 0 ) = 1 .
Entropy 27 00272 g008
Figure 9. Probability density of workload in selected moments. X ( 0 ) = 20 , S ( 0 ) = 1 .
Figure 9. Probability density of workload in selected moments. X ( 0 ) = 20 , S ( 0 ) = 1 .
Entropy 27 00272 g009
Figure 10. Probability that workload is zero versus time. X ( 0 ) = 20 , S ( 0 ) = 1 .
Figure 10. Probability that workload is zero versus time. X ( 0 ) = 20 , S ( 0 ) = 1 .
Entropy 27 00272 g010
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chydzinski, A. Workload of Queueing Systems with Autocorrelated Service Times. Entropy 2025, 27, 272. https://doi.org/10.3390/e27030272

AMA Style

Chydzinski A. Workload of Queueing Systems with Autocorrelated Service Times. Entropy. 2025; 27(3):272. https://doi.org/10.3390/e27030272

Chicago/Turabian Style

Chydzinski, Andrzej. 2025. "Workload of Queueing Systems with Autocorrelated Service Times" Entropy 27, no. 3: 272. https://doi.org/10.3390/e27030272

APA Style

Chydzinski, A. (2025). Workload of Queueing Systems with Autocorrelated Service Times. Entropy, 27(3), 272. https://doi.org/10.3390/e27030272

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop