Next Article in Journal
Numerical Investigation of Effect of Nozzle Upper Divergent Angle on Asymmetric Rectangular Section Ejector
Next Article in Special Issue
Clustered Distributed Data Storage Repairing Multiple Failures
Previous Article in Journal
Broadcast Channel Cooperative Gain: An Operational Interpretation of Partial Information Decomposition
Previous Article in Special Issue
A Post-Processing Method for Quantum Random Number Generator Based on Zero-Phase Component Analysis Whitening
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Coded Distributed Computing Under Combination Networks

1
Key Lab of Education Blockchain and Intelligent Technology, Ministry of Education, Guangxi Normal University, Guilin 541004, China
2
School of Science, Guilin University of Aerospace Technology, Guilin 541004, China
3
State Key Laboratory for Chemistry and Molecular Engineering of Medicinal Resources, School of Chemistry and Pharmaceutical Science, Guangxi Normal University, Guilin 541004, China
*
Author to whom correspondence should be addressed.
Entropy 2025, 27(3), 311; https://doi.org/10.3390/e27030311
Submission received: 3 March 2025 / Revised: 11 March 2025 / Accepted: 14 March 2025 / Published: 16 March 2025
(This article belongs to the Special Issue Network Information Theory and Its Applications)

Abstract

:
Coded distributed computing (CDC) is a powerful approach to reduce the communication overhead in distributed computing frameworks by utilizing coding techniques. In this paper, we focus on the CDC problem in ( H , L ) -combination networks, where H APs act as intermediate pivots and K = H L workers are connected to different subsets of L APs. Each worker processes a subset of the input file and computes intermediate values (IVs) locally, which are then exchanged via uplink and downlink transmissions through the AP station to ensure that all workers compute their assigned output functions. In this paper, we first novelly characterize the transmission scheme for the shuffle phase from the view point of the coefficient matrix and then obtain the scheme by using the Combined Placement Delivery Array (CPDA). Compared with the baseline scheme, our scheme significantly improves the uplink and downlink communication loads while maintaining the robustness and efficiency of the combined multi-AP network.

1. Introduction

The rapid growth of computationally intensive applications has driven considerable research into efficient distributed computing frameworks. MapReduce [1,2] and its evolution Apache Spark [3] have significant advantages in computational tasks dealing with massive datasets, where the amount of data can usually reach tens of terabytes. The MapReduce framework distributes computation tasks to multiple nodes, each storing a subset of the dataset. The computation process is decomposed into a set of “map” functions and “reduce” functions. Each map function processes a batch of data to generate intermediate values (IVs), which serve as inputs to the reduce functions. The overall process comprises three main phases: map, shuffle, and reduce phases. The nodes compute the map functions on their locally stored data, generate output IVs during the map phase, and exchange their computed IVs to ensure that each reduce function accesses to all necessary inputs during the map phase. In the reduce phase, all the nodes compute their assigned reduce functions using the gathered IVs.
Due to massive data exchange and limited communication bandwidth, the MapReduce-type system (MapReduce in short) often suffers from communication bottlenecks, i.e., the data shuffle phase consumes a significant portion of the overall task execution time. In order to reduce the communication overhead during the shuffling phase, Li et al. in [4] proposed a CDC scheme which achieves an optimal tradeoff between the computation load and communication load under a given output function assignment. The CDC scheme has also been widely studied in the other distributed computing scenarios. For example, CDC has been combined with maximum distance separable (MDS) codes to handle matrix–vector multiplication tasks and mitigate the impact of stragglers [5]. There are also many workers focusing on the stragglers in general functions [6,7], optimal resource allocation strategies [8,9], iterative computing and shuffling procedures [10,11], randomized file allocations [12], and cases with random network connectivity [13].
The CDC technique has been extended to wireless networks [14,15], where devices exchange information via wireless links. In such scenarios, owing to the decentralized nature of wireless networks, a central access point (AP) is typically required to facilitate data exchange, giving rise to uplink and downlink communication phases [16,17,18]. In real-world scenarios, we usually consider multiple APs. For example, a large place like a university campus needs to deploy multiple APs to realize campus network coverage. Moreover, a multi-AP system is able to distribute the load to multiple repeaters, thus reducing the risk of a single point of failure and improving fault tolerance.
This paper considers the CDC problem in a wireless ( H , L ) -combination network [19] which consists of H numbers of APs and K = H L computing workers. Each worker is connected to a distinct subset of L numbers of APs through wireless links. The APs act as intermediate hubs, facilitating the exchange of IVs among workers over wireless channels. Each worker is assigned a subset of the input files, processes them locally to compute IVs, and then transmits some IVs to its connected APs. The APs aggregate and broadcast the received information to other workers, ensuring that all workers obtain the necessary IVs to compute their respective output functions. This multi-AP structure provides enhanced reliability and robustness in the face of AP or node failures. We are interested in the tradeoff between the communication load and computation load (defined as the average number of nodes processing each input file), and aim to improve the overall efficiency of the distributed computing system by exploiting computation resources at workers and creating additional opportunities for parallelism and multicast transmission.
In this paper, we first obtain a baseline scheme by using CPDA. Then, by novelly characterizing the scheme from the view point of linear algebra, the problem of designing a CDC scheme for combination network is equivalent to constructing a special matrix. Finally, using CPDA, we obtain an improved CDC scheme which has a lower communication load than that of the baseline scheme, for which the computation load does not increase.
Notations: We denote the set of positive integers by N + . For n N + , the set { 1 , , n } is denoted as [ n ] , and | . | is used to represent the cardinality of a set or the length of a vector. Denote [ H ] L = { A | A [ H ] , | A | = L } , i.e., [ H ] L is the collection of subsets of [ H ] of size L.

2. System Model

We consider a ( H , L , M , N ) distributed coded computing system for a combination network, where there are N input files W = { w n | n [ N ] } , each of which has B bits, H APs H = { 1 , 2 , , H } without storage, K workers, each of which can store at most M files, and K output functions Q = { ϕ q | q [ K ] } from N files, each of which is arranged to one worker. Each worker is connected to a unique L-subset of APs over the wireless channel. So K = H L and the user set can be written as K = [ H ] L . Each worker is connected to the other workers via these wireless channels through common APs. For instance, Figure 1 shows a combination network with H = 4 and L = 2 . For each q [ K ] , the output function ϕ q maps the N input files into a streams of U bits, i.e., we have ϕ q : F 2 B N F 2 U . The map function is defined as g q : F 2 B F 2 T . For each n [ N ] , the map function maps the input file w n into intermediate value (IV) v q , n = g q ( w n ) F 2 T of T bits. Similarly, for every q [ K ] , we define the reduce function as h q : F 2 T N F 2 U , which maps the N IVs into a stream of U bits. Consequently, the output function ϕ q ( W ) is given by
ϕ q ( W ) = h q ( v q , 1 , , v q , N ) = h q ( g q ( w 1 ) , , g q ( w N ) )
as described in [4,18,20]. Thus, a ( H , L , M , N ) distributed coded computing scheme for a combination network consists of the following three phases:
Map Phase: Each worker k K stores M files, denoted by Z k . We assume that each file is stored the same number of times. Based on its local stored files, worker k K can compute all the IVs { v q , n | w n Z k , q [ K ] } .
Shuffle Phase: Each worker k K is assigned to compute an output function ϕ q , where q [ K ] . In order to obtain all the IVs { v q , n | n [ N ] } , the worker needs to request its required IVs from the other workers through their common APs. So, the communication process consists of two phases: uplink and downlink steps. Before the transmission begins, each worker encodes the IVs v q , n F 2 T as v ˜ q , n C using a randomized Gaussian encoding scheme. We assume that the transmission strategy is one-shot linear, i.e., each transmitted coded IV in a transmission could be decoded by any required worker and its local IVs. At each time slot s, in the uplink step, we denote the signal sent from the worker k K by X k ( s ) which satisfies the power constraint E [ | X i ( s ) | 2 ] P . Here, the worker can be silent, i.e., not send any data. Then, each AP h H receives the following signal:
y h ( s ) = k K c k , h X k ( s ) + z h ( s ) ,
where c k , h and z h ( s ) denote the channel coefficients from worker k to AP h and the additive white Gaussian noise, respectively. We assume that c k , h is chosen from C if worker k is connected to AP h; otherwise, c k , h = 0 . Additionally, z h ( s ) follows a normal distribution N ( 0 , 1 ) if worker k is connected to AP h; otherwise, z h ( s ) = 0 .
In the downlink step, each AP h H sends the received coded signal y h ( s ) from its connected workers. The goal of the communication is to obtain all the IVs v q , k required by each worker k K , who is arranged to compute the output function ϕ q .
Specifically, each worker k K receives the following signal from its connected APs
Y k ( s ) = h H c k , h y h ( s ) + z k ( s ) ,
where z k ( s ) is also the additive white Gaussian noise, and if h is connected to k, then z k ( s ) follows a normal distribution with N ( 0 , 1 ) ; otherwise, z k ( s ) = 0 . We can omit the additive noise terms z h ( s ) and z k ( s ) . This assumption is valid under the condition that the transmit power is sufficiently large to ensure a high signal-to-noise ratio (SNR).
Reduce Phase: After receiving all the signals sent down by the AP, each worker k K can decode all its required IVs and compute the arranged output function ϕ q ( W ) .
Definition 1. 
The following two types of loads are used as criteria for evaluating the merits of a ( H , L , M , N ) distributed coded computing scheme for combination network:
  • Computation load is defined as the total number of mapped files across the K workers, normalized by N, i.e., r k = 1 K | Z k | N ;
  • Communication load consists of the uplink load L u h = 1 H l h u K N T , where l h u is the length of the message received by the AP h in the uplink phase, i.e., the total normalized number of bits received by the H APs in the uplink process, and the downlink load L d h = 1 H l h d K N T where l h d is the length of the message sent by AP h in the downlink phase, i.e., the total normalized number of bits sent by the H APs in the downlink process.
The authors in [21] introduced the CPDA to realize a coded caching scheme for a combination network. In this paper, we will use it to construct our schemes.
Definition 2 
(CPDA). For any positive integers H, N, Z and L with L H and Z N , let K = H L , an N × K array P = ( p n , L ) n [ N ] , L [ H ] L over [ S ] { } is called a ( K , N , Z , S ) CPDA if it satisfies the following conditions:
C1. 
The symbol "∗" appears Z times in each column;
C2. 
Each s [ S ] occurs at least once in the array;
C3. 
For any two distinct entries p n 1 , L 1 and p n 2 , L 2 satisfying p n 1 , L 1 = p n 2 , L 2 = s [ S ] , then p n 1 , L 2 = p n 2 , L 1 = ;
C4. 
For any s [ S ] , the labels of all columns containing symbol s have a nonempty intersection.
  • If each integer occurs exactly g times in P , we call it g-regular CPDA.
Let us take the following example to further illustrate the concept of CPDA.
Example 1. 
When H = N = 4 and Z = L = 2, let us consider the following array.
P = ( 12 13 14 23 24 34 1 3 4 2 3 4 2 1 4 2 1 3 ) .
Clearly each column has exactly Z = 2 stars, and each integer occurs at least once in P . So the conditions C1–2 hold. Let us consider the conditions C3–4. When s = 1, we can obtain the following subarray:
P ( 1 ) = ( 12 23 24 1 1 1 ) .
Clearly, the entries
p 1 , { 2 , 3 } = p 1 , { 2 , 4 } = p 3 , { 1 , 2 } = p 3 , { 2 , 4 } = p 4 , { 1 , 2 } = p 4 , { 2 , 3 } = , p 1 , { 1 , 2 } = p 3 , { 2 , 3 } = p 4 , { 2 , 4 } = 1 .
Therefore, the above two symbols 1 satisfy condition C3. The column labels corresponding to each integer 1 are { 1 , 2 } , { 2 , 3 } , and { 2 , 4 } . Clearly, { 1 , 2 } { 2 , 3 } { 2 , 4 } = { 2 } , which satisfy condition C4. Similarly, we can obtain the subarrays which only contain integers 2, 3, and 4, respectively and show that the conditions C3–4 hold. So, P is a ( 6 , 4 , 2 , 4 ) CPDA.
There are many constructions of CPDAs in [21,22,23,24]. Here, we list some results which we will use in this paper as follows:
Lemma 1 
([21]). For any positive integers H, L, and t satisfying L | H and t < K = H 1 L 1 , there exists a ( H L , L K t , L K 1 t 1 , H K t + 1 ) CPDA.
Lemma 2 
([22]). For any positive integers H , L , b , λ satisfying 0 < L , b < H , λ < L , λ b and L + b 2 λ < H , there exists a ( H L , H b , H b L λ H L b λ , H L + b 2 λ · min { H L b + 2 λ λ , L + b 2 λ L λ } ) CPDA.

3. Main Results

In this section, the baseline scheme is introduced. Then, we characterize the CDC scheme from the view point of a coefficient matrix, thereby obtaining an optimized CDC scheme for the combination network. Finally, we use CPDA to derive the required coefficient matrix, which significantly simplifies the system design and enhances its performance. In addition, we demonstrate that the proposed scheme achieves a lower transmission load than the baseline scheme.
When each output function is computed by exactly one worker, the authors in [25] showed that the CDC problem is equivalent to the coded caching problem for device-to-device (D2D) networks. In addition, the authors in [26] proposed that a coded caching scheme for a shared link can be used to generate a coded caching scheme for a D2D network. Using the same method in [26], we can generate a CDC scheme for a combination network based on the CPDA and obtain the following result.
Theorem 1 
(Baseline scheme). Given a g- ( K , N , Z , S ) CPDA P with g > 1 and K = H L , there exists a ( H , L , M , N ) baseline coded distributed computing scheme for a ( H , L ) -combination network, with a computation load of r = K Z N and uplink and downlink communication loads of
L u = L d = S g K N ( g 1 ) .
Proof. 
We use the star positions in P to assign the files to each worker as follows. When p n , k = , worker k stores file w n . So worker k stores the following files:
Z k = { w n | n [ N ] , p n , k = } .
In the map phase, the set of IVs computed locally by worker k is
V k = { v q , n | q [ K ] , n [ N ] , p n , k = } .
According to Definition 1, the computation load is r = k = 1 K | Z k | N = K Z N .
Recall that each worker k K is assigned to compute the output function ϕ q k . From (5), worker k contains the required IVs in { v q k , n | p n , k = , n [ N ] } and needs the other required IVs in { v q k , n | p n , k , n [ N ] } , each of which should be transmitted by the other workers.
Now, we will use the integers in P to design the transmission strategy in the shuffle phase. For each integer s [ S ] , we assume that s occurs exactly g times in P . That is,
p n 1 , k 1 = p n 2 , k 2 = = p n g , k g = s .
Denote the set of workers associated with s as U s = { k 1 , k 2 , , k g } and their respective connected AP sets as
L = { L k 1 , L k 2 , , L k g } .
By the condition C4 of Definition 2, there exists at least one integer, say h s , satisfying
h s L k 1 L k 2 L k g .
Recall that in the map phase, each worker can encode the IV v q , n V k into v ˜ q , n using a random Gaussian coding scheme, where v ˜ q , n has a size of t bits. To facilitate efficient transmission, each IV v ˜ q , n in { v ˜ q k 1 , n 1 , v ˜ q k 2 , n 2 , , v ˜ q k g , n g } is divided into g 1 equal-sized packets. That is,
v ˜ q k , n = ( v ˜ q k , n ( k i ) ) k i U s , k i k ,
where v ˜ q k , n ( k i ) denotes the packet of v ˜ q k , n intended for worker k i .
In the shuffle phase, each worker k U s transmits the following encoded information of length t g 1 bits to the AP h s
X k ( s k ) = p n i , k i = s , k i k v ˜ q k i , n i ( k ) ,
where s k represents the sub-slot within time slot s allocated to worker k U s for transmitting its encoded data. Upon receiving the message y h s ( s k ) = c k , h s X k ( s k ) where the coefficient c k , h s is the channel coefficient between worker k and AP h s and is chosen from C under the independent and identically distributed condition, the AP h s broadcasts it to all workers in U s { k } . Consequently, each worker k i U s { k } receives the message
Y k i ( s k ) = c k i , h s y h s ( s k ) = c k i , h s c k , h s X k ( s k ) .
In the reduce phase, each worker k K needs to compute the output function ϕ q k . By Condition C3 of Definition 2 and Equation (5), each worker k i U s { k } already possesses all IVs in X k ( s k ) except v ˜ q k i , n i ( k ) . Therefore, each worker k i U s can decode its desired IV v ˜ q k i , n i ( k ) from Y k i ( s k ) . After time slot s, each worker k i U s obtains the IV v ˜ q k i , n i . After all S time slots, each worker k K obtains all the missing IVs { v q k , n p n , k , n [ N ] } , which are required to compute the output function ϕ q k . With all the necessary IVs, each worker k K proceeds to compute ϕ q k and completes the reduce phase.
Now, let us consider the communication load. By Definition 1, the uplink and downlink loads are
L u = L d = 1 K N t s = 1 S g t ( g 1 ) = S K N g ( g 1 ) .
Then, the proof is completed. □

3.1. Main Idea

In this subsection, we characterize the CDC scheme under the combination network from a matrix perspective. Our task is to compute Q = K output functions, where each worker k K owns a portion of the input file denoted as Z k .
Map Phase: For each mapped file w n Z k , worker k computes the IV v q , n for all K output functions. Consequently, the set of IVs available at worker k K is given by
V k = { v q , n q [ K ] , w n Z k } .
The computation load in the map phase is given by r = k = 1 K | Z k | N .
Shuffle Phase: The shuffle phase consists of two processes: (1) workers upload IVs to the AP via wireless channels, and (2) the AP broadcasts the received IVs to the connected workers. Before transmitting the information, each worker encodes the IV v q , n into v ˜ q , n using a random Gaussian coding scheme. During time slot s, a set of workers U s = { k 1 , k 2 , , k g } uploads IVs to AP h s . Assuming that the signal sent by worker k i U s at time slot s is denoted as X k i ( s ) , the signal received by AP h s can be expressed as
y h s ( s ) = C s X s = c k 1 , h s c k 2 , h s c k g , h s X k 1 ( s ) X k 2 ( s ) X k g ( s ) ,
where C s is the channel coefficient matrix for time slot s, and X s is the transmit signal vector for time slot s. To design X k i ( s ) , we define α k i = ( a k i , n 1 , a k i , n 2 , , a k i , n g ) as the encoding coefficients for worker k i . Then,
X s = X k 1 ( s ) X k 2 ( s ) X k g ( s ) = α k 1 α k 2 α k g v ˜ q k 1 , n 1 v ˜ q k 2 , n 2 v ˜ q k g , n g = a k 1 , n 1 a k 1 , n 2 a k 1 , n g a k 2 , n 1 a k 2 , n 2 a k 2 , n g a k g , n 1 a k g , n 2 a k g , n g v ˜ q k 1 , n 1 v ˜ q k 2 , n 2 v ˜ q k g , n g = A s V ˜ s ,
where A s is the coefficient matrix for time slot s, and V ˜ s is the vector of encoded IVs. Thus, the message received by AP h s is
y h s ( s ) = C s A s V ˜ s .
By designing the coefficient matrix A s , we can control the IVs uploaded by workers to AP h s , completing the uplink process.
Next, we describe the downlink process. In time slot s, AP h s broadcasts the received message y h s ( s ) to its connected workers. Each worker k i U s then receives
Y k i ( s ) = c k i , h s y h s ( s ) .
Using the IVs it already possesses, worker k i can decode the required information from Y k i ( s ) .
After S time slots, all K workers acquire the missing IVs required for the reduce function. Specifically, each worker k K obtains
{ v q , n q = q k , n [ N ] , w n Z k } .
By combining their own IVs V k , each worker now possesses all the necessary information to compute the output function.
Since the messages in both the uplink and downlink processes are y h s ( s ) , and assuming each IV has a size of t bits, the communication load per time slot is t bits. According to Definition 1, the total uplink and downlink communication loads are L u = L d = S t K N t = S K N .
Reduce Phase: After the shuffle phase, each worker k K computes the output function ϕ q k using all the IVs they possess.
Theorem 2. 
Given an ( K , N , Z , S ) CPDA with K = H L , there exists a ( H , L , M , N ) coded distributed computing scheme under a combination network, with a computation load of r = K Z N and its uplink and downlink communication loads given by
L u = L d = S K N .
Proof. 
We use a ( K , N , Z , S ) CPDA to design both the input file assignment and the coefficient matrix for the shuffle phase.
Before the computation task begins, the input file assigned to each worker k K is
Z k = { w n n [ N ] , p n , k = } .
In the map phase, each worker maps the input file it owns into K IVs v q , n . Thus, the IVs at worker k K are
V k = { v q , n q [ K ] , n [ N ] , p n , k = } .
Since each worker is allocated Z input files, the computation load according to Definition 1 is r = k = 1 K | Z k | N = K Z N .
Before the shuffle phase, each worker encodes the IV v q , n into v ˜ q , n using a random Gaussian coding scheme, where v ˜ q , n has a size of t bits. In the shuffle phase, we consider an integer s [ S ] in the CPDA P , and suppose s occurs g times in P . Let p n 1 , k 1 = p n 2 , k 2 = = p n g , k g = s . The set of workers involved in time slot s is U s = { k 1 , k 2 , , k g } , and the APs connected to all workers in U s are denoted as I s = { h h H , h is connected to all k i U s } . An arbitrary AP h s I s is selected to handle the uplink and downlink communication in time slot s.
In time slot s, we design the message received by AP h s as
y h s ( s ) = i = 1 g v ˜ q k i , n i = 1 1 1 v ˜ q k 1 , n 1 v ˜ q k 2 , n 2 v ˜ q k g , n g .
Let
V ˜ s = v ˜ q k 1 , n 1 v ˜ q k 2 , n 2 v ˜ q k g , n g .
From (13) and (18), we need to design the coefficient matrix A s such that
c k 1 , h s c k 2 , h s c k g , h s A s = 1 1 1 .
Since p k i , n i , by (17), worker k i does not possess the IV v ˜ q k i , n i . Thus, a k i , n i = 0 , and (19) can be written as
c k 1 , h s c k 2 , h s c k g , h s 0 a k 1 , n 2 a k 1 , n g a k 2 , n 1 0 a k 2 , n g a k g , n 1 a k g , n 2 0 = 1 1 1 .
By linear algebra, i.e., given the row vector ( c k 1 , h s , c k 2 , h s , , c k g , h s ) , we can always obtain a non-zero column vector a i of A s such that the ith coordinate is zero. So, we can always obtain the coefficient matrix A s in (20).
In the uplink phase, each worker k i U s transmits the encoded information X k i ( s ) = α k i V ˜ s to AP h s . In the downlink phase, AP h s broadcasts y h s ( s ) to all workers in U s . From Condition C 3 of Definition 2 and Equation (17), each worker k i U s already possesses all IVs in y h s ( s ) except v ˜ q k i , n i . Thus, each worker k i U s can decode its desired IV v ˜ q k i , n i from y h s ( s ) . After S time slots, all K workers obtain the missing IVs necessary for computing the reduce function.
Since the size of each IV v ˜ q , n is t bits, the communication load per time slot is t bits. The total uplink and downlink communication loads are
L u = L d = S t K N t = S K N .
In the reduce phase, each worker k K computes the output function ϕ q k ( w 1 , w 2 , , w N ) . Since ϕ q k ( w 1 , w 2 , , w N ) = h k ( v q k , 1 , v q k , 2 , , v q k , N ) , worker k utilizes the IVs of all N files for the q k -th output function. After the shuffle phase, each worker k K is given the IVs needed to compute ϕ q k . Therefore, all workers can compute their assigned output functions. □

3.2. Performance Evaluation

In this subsection, let us consider the performance of the proposed schemes by applying the CPDA in Lemmas 1 and 2 to Theorems 1 and 2, respectively. The resulting schemes are summarized in Table 1 and Table 2 respectively. In the following, let us consider the computation load and communication load of these schemes.
Figure 2 illustrates the relationship between the computation load and the communication load for the schemes in Table 1 and Table 2. Let us consider the distributed computing system with K = 15 computing nodes and Q = K = 15 output functions under the ( 6 , 2 ) -combination network. As shown in Figure 2, when the same class of CPDA is used, the main scheme of Theorem 2 achieves a lower communication load than the base scheme of Theorem 1 for the same computation load. This improvement results from the main scheme’s more efficient utilization of the multicast gain, which reduces the communication load.

3.3. Example for Achievable Scheme

We consider a combinatorial network with H = 3 , L = 2 , where K = 3 and N = 3 . The tasks to be computed by the 3 workers are ϕ 1 , ϕ 2 , and ϕ 3 , respectively. The workers are denoted as K = { 12 , 13 , 23 } , and the input files are denoted as W = { w 1 , w 2 , w 3 } . Before starting the task, the input files are assigned as follows: Z 12 = { w 1 } to worker 12, Z 13 = { w 2 } to worker 13, and Z 23 = { w 3 } to worker 23.

3.3.1. Map Phase

In the map phase, each worker maps each assigned file into 3 IVs. Specifically:
  • Worker 12 owns V 12 = { v 1 , 1 , v 2 , 1 , v 3 , 1 } but lacks { v 1 , 2 , v 1 , 3 } .
  • Worker 13 owns V 13 = { v 1 , 2 , v 2 , 2 , v 3 , 2 } but lacks { v 2 , 1 , v 2 , 3 } .
  • Worker 23 owns V 23 = { v 1 , 3 , v 2 , 3 , v 3 , 3 } but lacks { v 3 , 1 , v 3 , 2 } .

3.3.2. Shuffle Phase

After the map phase, the shuffle phase begins, where workers exchange IVs via the APs. Each worker encodes the IVs v q , n into v ˜ q , n using a randomized Gaussian coding scheme to ensure reliable transmission over the wireless channel.
  • Time Slot 1: Workers 12 and 13 send X 12 ( 1 ) = 1 c 12 , 1 v ˜ 2 , 1 and X 13 ( 1 ) = 1 c 13 , 1 v ˜ 1 , 2 to AP 1 simultaneously. AP 1 receives the message y 1 ( 1 ) = v ˜ 1 , 2 + v ˜ 2 , 1 , completing the uplink process. AP 1 then broadcasts y 1 ( 1 ) to workers 12 and 13, who receive Y 12 ( 1 ) = c 12 , 1 ( v ˜ 1 , 2 + v ˜ 2 , 1 ) and Y 13 ( 1 ) = c 13 , 1 ( v ˜ 1 , 2 + v ˜ 2 , 1 ) , respectively.
  • Time Slot 2: Workers 12 and 23 send X 12 ( 2 ) = 1 c 12 , 2 v ˜ 3 , 1 and X 23 ( 2 ) = 1 c 23 , 2 v ˜ 1 , 3 to AP 2 simultaneously. After the uplink and downlink phases, workers 12 and 23 receive Y 12 ( 2 ) = c 12 , 2 ( v ˜ 1 , 3 + v ˜ 3 , 1 ) and Y 23 ( 2 ) = c 23 , 2 ( v ˜ 1 , 3 + v ˜ 3 , 1 ) , respectively.
  • Time Slot 3: Workers 13 and 23 send X 13 ( 3 ) = 1 c 13 , 3 v ˜ 3 , 2 and X 23 ( 3 ) = 1 c 23 , 3 v ˜ 2 , 3 to AP 3 simultaneously. After the uplink and downlink phases, workers 13 and 23 receive Y 13 ( 3 ) = c 13 , 3 ( v ˜ 2 , 3 + v ˜ 3 , 2 ) and Y 23 ( 3 ) = c 23 , 3 ( v ˜ 2 , 3 + v ˜ 3 , 2 ) , respectively.

3.3.3. Reduce Phase

In the reduce phase, each worker decodes the missing IVs:
  • Worker 12 solves v ˜ 1 , 2 and v ˜ 1 , 3 from Y 12 ( 1 ) and Y 12 ( 2 ) using its own v ˜ 2 , 1 and v ˜ 3 , 1 .
  • Worker 13 solves v ˜ 2 , 1 and v ˜ 2 , 3 from Y 13 ( 1 ) and Y 13 ( 3 ) using its own v ˜ 1 , 2 and v ˜ 3 , 2 .
  • Worker 23 solves v ˜ 3 , 1 and v ˜ 3 , 2 from Y 23 ( 2 ) and Y 23 ( 3 ) using its own v ˜ 1 , 3 and v ˜ 2 , 3 .
After obtaining all the required IVs, each worker computes its assigned output function.
In this example, the computation load is r = 1 + 1 + 1 3 = 1 . Assuming each IV v ˜ q , n has a size of t bits, the uplink and downlink loads are calculated as L u = L d = 3 t 3 × 3 t = 1 3 .

4. Conclusions

In this paper, we studied the coded distributed computing for ( H , L ) -combination networks and provided a comprehensive framework for constructing CDC scheme from the view point of linear algebra. Then, by using CPDAs, we proposed two classes of CDC schemes for the combination network. Finally by comparisons, we showed that the proposed schemes have lower communication loads than the baseline schemes.

Author Contributions

Conceptualization, Y.Y. and X.Q.; Methodology, Y.H.; Software, Y.Y. and S.L.; Formal analysis, X.Q.; Resources, S.L.; Data curation, Y.H. and S.L.; Writing—original draft, Y.Y.; Writing—review & editing, X.Q. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Key Lab of Education Blockchain and Intelligent Technology, Ministry of Education (EBME24-15).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The data are available upon request from the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Dean, J.; Ghemawat, S. MapReduce: Simplified data processing on large clusters. Commun. ACM 2008, 51, 107–113. [Google Scholar] [CrossRef]
  2. Isard, M.; Budiu, M.; Yu, Y.; Birrell, A.; Fetterly, D. Dryad: Distributed data-parallel programs from sequential building blocks. SIGOPS Oper. Syst. Rev. 2007, 41, 59–72. [Google Scholar] [CrossRef]
  3. Zaharia, M.; Chowdhury, M.; Franklin, M.J.; Shenker, S.; Stoica, I. Spark: Cluster Computing with Working Sets. In Proceedings of the 2nd USENIX Workshop on Hot Topics in Cloud Computing (HotCloud 10), Boston, MA, USA, 22–25 June 2010. [Google Scholar]
  4. Li, S.; Maddah-Ali, M.A.; Yu, Q.; Avestimehr, A.S. A Fundamental Tradeoff Between Computation and Communication in Distributed Computing. IEEE Trans. Inf. Theory 2018, 64, 109–128. [Google Scholar] [CrossRef]
  5. Li, S.; Maddah-Ali, M.A.; Avestimehr, A.S. A Unified Coding Framework for Distributed Computing with Straggling Servers. In Proceedings of the 2016 IEEE Globecom Workshops (GC Wkshps), Washington, DC, USA, 4–8 December 2016; pp. 1–6. [Google Scholar] [CrossRef]
  6. Yan, Q.; Wigger, M.; Yang, S.; Tang, X. A Fundamental Storage-Communication Tradeoff for Distributed Computing with Straggling Nodes. IEEE Trans. Commun. 2020, 68, 7311–7327. [Google Scholar] [CrossRef]
  7. Yan, Q.; Wigger, M.; Yang, S.; Tang, X. A Fundamental Storage-Communication Tradeoff in Distributed Computing with Straggling Nodes. In Proceedings of the 2019 IEEE International Symposium on Information Theory (ISIT), Paris, France, 7–12 July 2019; pp. 2803–2807. [Google Scholar] [CrossRef]
  8. Yu, Q.; Li, S.; Maddah-Ali, M.A.; Avestimehr, A.S. How to optimally allocate resources for coded distributed computing? In Proceedings of the 2017 IEEE International Conference on Communications (ICC), Paris, France, 21–25 May 2017; pp. 1–7. [Google Scholar] [CrossRef]
  9. Attia, M.A.; Tandon, R. On the worst-case communication overhead for distributed data shuffling. In Proceedings of the 2016 54th Annual Allerton Conference on Communication, Control, and Computing (Allerton), Monticello, IL, USA, 27–30 September 2016; pp. 961–968. [Google Scholar] [CrossRef]
  10. Attia, M.A.; Tandon, R. Information Theoretic Limits of Data Shuffling for Distributed Learning. In Proceedings of the 2016 IEEE Global Communications Conference (GLOBECOM), Washington, DC, USA, 4–8 December 2016; pp. 1–6. [Google Scholar] [CrossRef]
  11. Elmahdy, A.; Mohajer, S. On the Fundamental Limits of Coded Data Shuffling. In Proceedings of the 2018 IEEE International Symposium on Information Theory (ISIT), Vail, CO, USA, 17–22 June 2018; pp. 716–720. [Google Scholar] [CrossRef]
  12. Song, L.; Srinivasavaradhan, S.R.; Fragouli, C. The benefit of being flexible in distributed computation. In Proceedings of the 2017 IEEE Information Theory Workshop (ITW), Kaohsiung, Taiwan, 6–10 November 2017; pp. 289–293. [Google Scholar] [CrossRef]
  13. Srinivasavaradhan, S.R.; Song, L.; Fragouli, C. Distributed Computing Trade-offs with Random Connectivity. In Proceedings of the 2018 IEEE International Symposium on Information Theory (ISIT), Vail, CO, USA, 17–22 June 2018; pp. 1281–1285. [Google Scholar] [CrossRef]
  14. Li, F.; Chen, J.; Wang, Z. Wireless MapReduce Distributed Computing. IEEE Trans. Inf. Theory 2019, 65, 6101–6114. [Google Scholar] [CrossRef]
  15. Parrinello, E.; Lampiris, E.; Elia, P. Coded Distributed Computing with Node Cooperation Substantially Increases Speedup Factors. In Proceedings of the 2018 IEEE International Symposium on Information Theory (ISIT), Vail, CO, USA, 17–22 June 2018; pp. 1291–1295. [Google Scholar] [CrossRef]
  16. Li, S.; Yu, Q.; Maddah-Ali, M.A.; Avestimehr, A.S. Edge-Facilitated Wireless Distributed Computing. In Proceedings of the 2016 IEEE Global Communications Conference (GLOBECOM), Washington, DC, USA, 4–8 December 2016; pp. 1–7. [Google Scholar] [CrossRef]
  17. Li, S.; Yu, Q.; Maddah-Ali, M.A.; Avestimehr, A.S. A Scalable Framework for Wireless Distributed Computing. IEEE/ACM Trans. Netw. 2017, 25, 2643–2654. [Google Scholar] [CrossRef]
  18. Yan, Q.; Tang, X.; Tao, M.; Huang, Q. A Fundamental Tradeoff Among Storage, Computation, and Communication for Distributed Computing Over Star Network. IEEE Trans. Commun. 2023, 71, 5686–5697. [Google Scholar] [CrossRef]
  19. Ji, M.; Wong, M.F.; Tulino, A.M.; Llorca, J.; Caire, G.; Effros, M.; Langberg, M. On the fundamental limits of caching in combination networks. In Proceedings of the 2015 IEEE 16th International Workshop on Signal Processing Advances in Wireless Communications (SPAWC), Stockholm, Sweden, 28 June–1 July 2015; pp. 695–699. [Google Scholar] [CrossRef]
  20. Peter, E.; Krishnan Namboodiri, K.K.; Rajan, B.S. Wireless MapReduce Arrays for Coded Distributed Computing. In Proceedings of the 2024 IEEE Information Theory Workshop (ITW), Shenzhen, China, 24–28 November 2024; pp. 163–168. [Google Scholar] [CrossRef]
  21. Yan, Q.; Wigger, M.; Yang, S. Placement Delivery Array Design for Combination Networks with Edge Caching. In Proceedings of the 2018 IEEE International Symposium on Information Theory (ISIT), Vail, CO, USA, 17–22 June 2018; pp. 1555–1559. [Google Scholar] [CrossRef]
  22. Cheng, M.; Li, Y.; Zhong, X.; Wei, R. Improved Constructions of Coded Caching Schemes for Combination Networks. IEEE Trans. Commun. 2020, 68, 5965–5975. [Google Scholar] [CrossRef]
  23. Huang, L.; Wang, J.; Cheng, M.; Deng, Q.; Zhong, B. Coded Caching for Combination Networks with Multiaccess. Information 2022, 13, 191. [Google Scholar] [CrossRef]
  24. Zhang, M.; Cheng, M.; Wu, Y.; Li, X. Coded Caching for Dense-User Combination Network in Binary Field. IEEE Trans. Commun. 2024, 72, 2716–2730. [Google Scholar] [CrossRef]
  25. Ji, M.; Caire, G.; Molisch, A.F. Fundamental Limits of Caching in Wireless D2D Networks. IEEE Trans. Inf. Theory 2016, 62, 849–869. [Google Scholar] [CrossRef]
  26. Wang, J.; Cheng, M.; Yan, Q.; Tang, X. Placement Delivery Array Design for Coded Caching Scheme in D2D Networks. IEEE Trans. Commun. 2019, 67, 3388–3395. [Google Scholar] [CrossRef]
Figure 1. A combination network with H = 4 , L = 2 .
Figure 1. A combination network with H = 4 , L = 2 .
Entropy 27 00311 g001
Figure 2. The relationship between the computation load and the communication load for the schemes in Table 1 and Table 2 is illustrated for a distributed computing system with K = 15 computing nodes and Q = K = 15 output functions under the ( 6 , 2 ) -combination network.
Figure 2. The relationship between the computation load and the communication load for the schemes in Table 1 and Table 2 is illustrated for a distributed computing system with K = 15 computing nodes and Q = K = 15 output functions under the ( 6 , 2 ) -combination network.
Entropy 27 00311 g002
Table 1. The baseline schemes in Theorem 1.
Table 1. The baseline schemes in Theorem 1.
Schemer L u , L d Original
base scheme 1 t H L K t t K Lemma 1
base scheme 2 H L b λ H b L λ H L + b 2 λ · H L b + 2 λ λ L + b 2 λ L λ H L H b ( max ( H L b + 2 λ λ , L + b 2 λ L λ ) 1 ) Lemma 2
Table 2. The main schemes in Theorem 2.
Table 2. The main schemes in Theorem 2.
Schemer L u , L d Original
main scheme 1 t H L K t ( t + 1 ) K Lemma 1
main scheme 2 H L b λ H b L λ H L + b 2 λ · min ( H L b + 2 λ λ , L + b 2 λ L λ ) H L H b Lemma 2
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yang, Y.; Huang, Y.; Qin, X.; Lu, S. Coded Distributed Computing Under Combination Networks. Entropy 2025, 27, 311. https://doi.org/10.3390/e27030311

AMA Style

Yang Y, Huang Y, Qin X, Lu S. Coded Distributed Computing Under Combination Networks. Entropy. 2025; 27(3):311. https://doi.org/10.3390/e27030311

Chicago/Turabian Style

Yang, Yongcheng, Yifei Huang, Xiaohuan Qin, and Shenglian Lu. 2025. "Coded Distributed Computing Under Combination Networks" Entropy 27, no. 3: 311. https://doi.org/10.3390/e27030311

APA Style

Yang, Y., Huang, Y., Qin, X., & Lu, S. (2025). Coded Distributed Computing Under Combination Networks. Entropy, 27(3), 311. https://doi.org/10.3390/e27030311

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop