Next Article in Journal
VR-Based Job Training System Using Tangible Interactions
Next Article in Special Issue
Dynamic Asynchronous Anti Poisoning Federated Deep Learning with Blockchain-Based Reputation-Aware Solutions
Previous Article in Journal
Dry Wearable Textile Electrodes for Portable Electrical Impedance Tomography
Previous Article in Special Issue
A Vision-Based Social Distancing and Critical Density Detection System for COVID-19
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Efficient Gradient Updating Strategies with Adaptive Power Allocation for Federated Learning over Wireless Backhaul

Division of Smart Robot Convergence and Application Engineering, Department of Electronic Engineering, Pukyong National University, Busan 48513, Korea
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(20), 6791; https://doi.org/10.3390/s21206791
Submission received: 23 August 2021 / Revised: 5 October 2021 / Accepted: 10 October 2021 / Published: 13 October 2021
(This article belongs to the Special Issue Machine Learning in Wireless Sensor Networks and Internet of Things)

Abstract

:
In this paper, efficient gradient updating strategies are developed for the federated learning when distributed clients are connected to the server via a wireless backhaul link. Specifically, a common convolutional neural network (CNN) module is shared for all the distributed clients and it is trained through the federated learning over wireless backhaul connected to the main server. However, during the training phase, local gradients need to be transferred from multiple clients to the server over wireless backhaul link and can be distorted due to wireless channel fading. To overcome it, an efficient gradient updating method is proposed, in which the gradients are combined such that the effective SNR is maximized at the server. In addition, when the backhaul links for all clients have small channel gain simultaneously, the server may have severely distorted gradient vectors. Accordingly, we also propose a binary gradient updating strategy based on thresholding in which the round associated with all channels having small channel gains is excluded from federated learning. Because each client has limited transmission power, it is effective to allocate more power on the channel slots carrying specific important information, rather than allocating power equally to all channel resources (equivalently, slots). Accordingly, we also propose an adaptive power allocation method, in which each client allocates its transmit power proportionally to the magnitude of the gradient information. This is because, when training a deep learning model, the gradient elements with large values imply the large change of weight to decrease the loss function.

1. Introduction

Recently, deep neural networks (DNNs) or convolutional neural networks (CNNs) have been widely applied to complicated signal processing, such as classification tasks and signal regression problems, due to their outstanding performances in nonlinear adaptability and feature extraction ([1,2,3] and references therein) and are also extended to the distributed sensing systems (e.g., the object recognition using distributed micro-Doppler radars in [4] and the data driven digital healthcare applications [5,6,7]). In the distributed sensing systems, centralized training strategies may be adopted to train their common DNN or CNN modules by sharing their sensing data. However, due to the data-size and the privacy issues of the locally collected data, the centralized training is not desirable, especially when the capacity of the backhaul link for the data exchange is limited.
The federated learning approach has been extensively investigated as an alternative distributed machine learning method [8,9] where, rather than sharing their locally collected dataset, the clients report the stochastic gradient information (minimizing the loss function with respect to their local dataset) to the main server. The main server then aggregates the stochastic gradient information and broadcast it to the clients. Accordingly, to achieve the unbiased stochastic gradient at the main server, the training data sampling methods are investigated [10,11]. Furthermore, in [12], to reduce the communication overhead of transmitting the updated gradient information (proportional to the number of weights in the DNNs and CNNs), an efficient weight aggregation protocol for federated learning is proposed and in [13], the structured updating method is proposed for the communication cost reduction. However, they assume that the stochastic gradient information is perfectly transferred from the multiple clients to the main server without any distortion.
In the federated learning process, when the clients are connected with wireless-connected clients, local gradient information needs to be transferred from the distributed clients to the server over the wireless backhaul link and can be distorted due to wireless channel fading. In [14,15,16,17], for the wireless backhaul, the federated learning strategies are proposed for the MNIST hand-writing image classification and the associated wireless resources are efficiently optimized. In [14,15], the average of the local stochastic gradient vectors is recovered at the server when the pre-processed local gradient vectors are transferred from the clients. In [16], the compressive sensing approach is proposed to estimate the local gradient vectors at the server. In [17], joint communication and federated learning model is developed, where the resource allocation and the client selection methods are proposed such that the packet error rates of the communication links between server and clients are optimized. We note that most of the previous works have focused on the estimation of local stochastic gradient vectors at the server.
In this paper, we also consider the federated learning system, where distributed clients are connected to the server via wireless backhaul link and develop efficient training strategies for the federated learning over wireless backhaul link. Differently from the previous works, where the average of the local stochastic gradient vectors (i.e., the equal-weight combining) is recovered at the server, we propose an efficient gradient updating method, in which the local gradients are combined such that the effective signal-to-noise ratio (SNR) is maximized at the server. In addition, we also propose a binary gradient updating strategy based on thresholding in which the round associated with all channel having small channel gains is excluded from federated learning. That is, when the backhaul links for all clients have channel gain smaller than a pre-defined threshold simultaneously, the server may have severely distorted gradient vectors, which can be avoided through the proposed updating with thresholding. Furthermore, because each client has limited transmission power, it is effective to allocate more power on the channel slots carrying specific important information, rather than allocating power equally to all channel resources (equivalently, slots). Accordingly, we also propose an adaptive power allocation method, in which each client allocates its transmit power proportionally to the magnitude of the gradient information. This is because, when training a deep learning model, the gradient elements with large values imply the large change of weight to decrease the loss function.
Through the extensive computer simulations, it can be found that the proposed gradient updating methods improve the federated learning performance over the wireless channel. Specifically, due to the distortion over wireless channel, the classification accuracy of the equal-weight combining decreases drastically as the rounds of the federated learning increase. In contrast, the proposed effective SNR maximizing scheme with thresholding exhibits the accuracy performance which is comparable to that for the federated learning over the error-free backhaul link. We note that, as the threshold level increases, the federated learning is performed stably, because the highly distorted gradient update vector due to small channel gain can be discarded by a large threshold level. However, the large threshold level may incur the gradient updating delay, but the adaptive power allocation strategy can improve the trade-off between the federated learning performance and the learning delay due to the threshold level.
The rest of this paper is organized as follows. In Section 2, the system model for the federated learning system with the wireless backhaul is presented in which the distributed clients have a common CNN module for the handwriting character recognition. In Section 4, gradient updating methods are proposed. In addition, the adaptive power allocation method is also developed considering the importance of the gradient information. In Section 5, we provide several simulation results and in Section 6, we give our conclusions.

2. System Model

In Figure 1, we consider the federated learning systems with wireless backhaul, where the L multi-clients have their own datasets to train each local network. Here, a common neural network model is shared for all clients and it is trained through the federated learning over wireless backhaul connected to the main server. The common neural network is designed for the classification problem, in which the label d ^ is induced from the network output for the lth client’s measured data with the label d, S d ( l ) . That is,
d ^ = arg max d = 1 , , D x o u t [ d ] , s . t . x o u t = f ( S d ( l ) ; θ ) R D × 1
where f ( ; θ ) denotes the non-linear neural network function with the model parameter θ ( R P × 1 ) that gives the estimate of the categorical label probability vector as its output vector. Here, P denotes the number of weights in the common neural network model and x o u t [ d ] is the dth element of the vector x o u t . We note that the size of the model parameter (P) is determined by the structure of the neural network model. Specifically, in the case of a convolutional layer with K K f 1 × K f 2 filters, the number of weights is given as K f 1 × K f 2 × K + K that accounts for the kernel size ( K f 1 × K f 2 ), the number of kernels (K) and the number of biases (K). In the case of a single fully-connected layer, the number of weights is calculated as N i n × N n r + N n r , where N i n and N n r denote the input size and the number of neurons, respectively. See also Section 2.1. We note that, because the collected data at each client are generally of a large dimension with private security issues, it is not desirable to report the collected data to the server. Furthermore, the large dimension of the data may cause the significant burden on the typical backhaul link to transmit a number of training datasets. Instead, the neural network model f ( ; θ ) will be shared over all clients and θ can be locally trained with the data obtained from each client. By denoting θ ( l ) as the model parameter trained at the lth client, θ ( l ) is reported to the server through the wireless uplink backhaul for the federated learning. The associated federated learning strategies and power allocation over the wireless backhaul will be discussed in more detail in Section 4.

2.1. CNN Architecture for Handwriting Character Recognition

Throughout the paper, multiple clients have a common neural network for the handwriting character recognition. Specifically, a typical CNN module is considered for the character image classification as in Figure 2, but the proposed federated learning strategy can be applied to other CNN models. The non-linear neural network function f ( S i ( l ) ; θ ) in (1) is composed of an input layer, convolutional layers, activation layers, max pooling layer, a fully-connected layer, and an output layer. See Section 4 for the specific values of the hyperparameters of CNN module.
– Convolution Layer: The handwriting image matrix, S i ( l ) R N w i d t h × N h e i g h t is exploited as the input of the convolution layers. In addition, each element of their output is computed through the convolution operation with a K f i 1 × K f i 2 filter (equivalently, kernel) for ith layer. Specifically, the output of the ith convolution layer can be given as:
X ( i ) [ m , n , k ] = p = 0 K f i 1 1 q = 0 K f i 2 1 k = 0 K i 1 f a X ( i 1 ) [ m + p , n + q , k ] W ( i ) [ p , q , k ] + b ( i ) [ k ] ,
where X ( i 1 ) [ m , n , k ] is the ( m , n , k ) th element of X ( i 1 ) R m i 1 × n i 1 × k i 1 , the input of the ith layer and f a · is an activation function. In addition, W ( i ) [ p , q , k ] is the ( p , q , k ) th element of the filter matrix W ( i ) at the ith layer and b ( i ) [ k ] is the kth element of a bias vector b ( i ) . Throughout the paper, rectified linear unit (ReLU) function is used as the activation function, which is given as
f a ( X ( i ) ) = m a x ( 0 , X ( i ) ) .
– Max pooling layer: In the pooling layer, to reduce the dimension of the input data without losing useful information, the elements of the input are down-sampled [18]. In the Max pooling layer, after dividing the input matrix into multiple blocks, the maximum value in each block is sampled and forwarded to the dimension-reduced output matrix.
– Flatten, Fully-Connected (FC) layer: The flatten layer is used for changing the shape of output of convolution layer into the vector which is used as the input of FC layer. We note that, in the case of a single fully-connected layer with N i n input elements and N n r neurons, the number of weights is given as N i n × N n r + N n r . In the FC layer, the output of convolution layer is associated with a proper loss function such that the label is correctly identified after the training.
Throughout the paper, the cross entropy (CE) is used as the loss function which is given as
L C E ( x o u t , L ¯ d ; θ ) = i = 1 D L ¯ d [ i ] log ( x o u t [ i ] ) ,
where x o u t R D × 1 is the output of FC and L ¯ d is a label one-hot encoded vector of size D that has zeros in all elements except the dth element, which is assigned a value of 1. Then, by using the local training datasets ( Φ t r ( l ) = { S d , t r ( l ) , L ¯ d , t r } t = 1 N t r ) at the lth client, the network function parameter can be updated as:
θ t ( l ) θ t 1 ( l ) + g t 1 ( l ) ,
where g t 1 ( l ) ( R P × 1 ) denotes the gradient such that the loss function is minimized for the local training datasets Φ t r ( l ) and is given as g t 1 ( l ) η θ L C E ( x o u t , L ¯ d , t r ; θ ) θ = θ t 1 ( l ) with a learning rate, η .

2.2. Signal Model for Wireless Backhaul

As in Figure 1, the clients are connected to the server through the wireless backhaul link. For the federated learning, the model parameters aggregated at the server are broadcast at each iteration of training phase through the wireless downlink channel, while the model parameters trained at the lth client are reported to the server through the wireless uplink backhaul link. Throughout the paper, we focus only on the uplink phase of multiple access channel and assume that the broadcast channel for the downlink phase is error-free, as done in [15,16,19].
Assuming that the clients and the server have a single antenna for the backhaul link, when total B channel resources with narrowband signal bandwidth are available (Here, we note that the channel resources may be given in the frequency axis or may be given in the time axis.), the received signal at server for the tth round of the gradient update can be given as
y t [ b ] = l = 1 L h l , t [ b ] x l , t [ b ] + n t [ b ] ,
for b = 1 , , B , where x l , t [ b ] is the precoded transmit signal of the lth client at the bth channel resource with E [ | x l , t [ b ] | 2 ] = 1 for the tth round. Here, h l , t [ b ] and n t [ b ] denote the aggregated Rayleigh fading channel and the zero-mean additive white Gaussian noise (AWGN) at the bth channel resource, respectively. That is, h l , t [ b ] follows a Gaussian distribution with a zero-mean and a variance σ h l 2 (that is, h l , t [ b ] N 0 , σ h l 2 ). Likewise, n t [ b ] N 0 , σ n 2 . In addition, the wireless channel is constant over each round of federated learning process, but changes independently from round to round. By concatenating y t [ b ] in (4), the received signal at server can be vectorized as:
y t = y t [ 1 ] y t [ B ] = l = 1 L H l , t x l , t + n t ,
where H l , t = d i a g { h l , t [ 1 ] , , h l , t [ B ] } and
x l , t = x l , t [ 1 ] x l , t [ B ] , n t = n t [ 1 ] n t [ B ] .
Here, d i a g a 1 , , a B denotes a B × B diagonal matrix having its diagonal elements as a 1 , , a B .

3. Federated Learning for Handwriting Character Recognition

Note that, as in (3), the CNN parameter θ ( l ) can be trained with the local training datasets at each client, which limits the adaptability of the CNN due to the lack of the globally measured data. Accordingly, to train their parameters globally, federated learning strategy is exploited, known as an efficient learning strategy suitable to the multi-clients environment such as our system model shown in Figure 1.
Specifically, during the tth round of the training phase, each client receives the gradient of the model parameter g t 1 from the server via the backhaul link. Then, by exploiting g t 1 instead of g t 1 ( l ) in (3) the network function parameter can be updated as:
θ t ( l ) θ t 1 ( l ) + g t 1 .
We note that g t 1 is the globally aggregated gradient computed at the server, which tends to minimize the loss function with respect to the data collected at all clients. Then, each client can compute its next local gradient g t ( l ) such that the local loss function is minimized for the locally collected datasets Φ t r ( l ) . Then, the locally updated gradient vector is reported to the server via the backhaul link. The server can then aggregate the local gradient vector to get g t as:
g t = f g ( g t ( l ) , l = 1 , , L ) ,
where the function f g ( ) represents the gradient aggregation function. In [20], the FederatedAveraging technique (i.e., equal weight combining) is proposed which is given as:
g t = 1 L l = 1 L g t ( l ) .
The aggregated gradient g t is again broadcast to the multi-clients and exploited to update the neural network model at each client. The above described steps are repeated for a given number of rounds, T.
At the beginning of the training phase, the server needs to initialize the global model parameters and, throughout the paper, the parameters are initialized based on He normal weight initialization method [21], which is advantageous when used with ReLU activation function. Based on the above description, generalized federated learning process is summarized in Algorithm 1.
Algorithm 1. Generalized federated learning train process.
  1:
Initialize θ 0 based on He normal weight initialization method
  2:
g 0 0
  3:
for t 1 to T do
  4:
      (Clients) θ t ( l ) θ t 1 ( l ) + g t 1
  5:
      (Clients) Update g t ( l ) from the datasets Φ t r ( l )
  6:
      (Clients) Report g t ( l ) to the server via the backhaul link
  7:
      (Server) g t f g ( g t ( l ) , l = 1 , , L ) as in (7)
  8:
      (Server) Broadcast g t to multi-clients
  9:
end for
Differently from the centralized learning, the datasets collected by each client are not necessarily reported to the main server in Algorithm 1. We note that, in many cases, data sharing is not free from security, regulatory and privacy issues [8]. We also note that the communication cost for the centralized learning depends on the number/size of the collected data [22,23]. In contrast, the communication cost for the federated learning is independent with the data size, but depends on the CNN architecture (specifically, the number of weights in the CNN).

4. Gradient Updating and Adaptive Power Allocation Strategies for the Federated Learning over Wireless Backhaul

In line 6 of Algorithm 1, multi-clients should report their local gradient vectors g t ( l ) through the backhaul link with B channel resources at each round. Specifically, each client should design the transmit signal x l , t to transmit g t ( l ) in (5). In addition, the server should estimate g ^ t ( l ) from the received signal y t in (5).

4.1. Linear Gradient Estimation for Federated Learning over Wireless Backhaul

To avoid the inter-channel interference over the wireless backhaul link, conventional orthogonal multiple access method with linear precoding is considered in which the wireless resource blocks are orthogonally allocated to each client. Specifically, by letting B ¯ = B L , which is assumed to be an integer, x l can be given as
x l , t = 0 B ¯ ( l 1 ) × B ¯ I B ¯ 0 ( B B ¯ l ) × B ¯ Ψ B ¯ × P ¯ g ¯ t ( l ) ,
where Ψ B ¯ × P ¯ is a predefined pseudo-random matrix satisfying the restricted isometry property (RIP) condition [24] and unitary such as:
Ψ B ¯ × P ¯ H Ψ B ¯ × P ¯ = 1 g ¯ t ( l ) 2 I P ¯ .
Note that g t ( l ) ( R P × 1 ) in (3) is split into multiple P ¯ dimensional vectors, g ¯ t ( l ) and each split vector is transmitted through B ¯ wireless resources.
Then, (5) can be rewritten as:
y ¯ l , t = H ¯ l , t Ψ B ¯ × P ¯ g ¯ t ( l ) + n ¯ t ,
where H ¯ l , t = d i a g { h l , t [ B ( l 1 ) + 1 ] , h l , t [ B ( l 1 ) + 2 ] , , h l , t [ B l ] } and
y ¯ l , t = y t [ B ¯ ( l 1 ) + 1 ] y t [ l B ¯ ] , n ¯ t = n t [ B ¯ ( l 1 ) + 1 ] n t [ l B ¯ ] .
When B ¯ P ¯ , g t ( l ) can be estimated from (11) by exploiting the linear estimation methods such as zero-forcing or MMSE estimation. That is, ZF estimate of g t ( l ) can be given as:
g ¯ ^ t ( l ) = ( G ¯ l , t H G ¯ l , t ) 1 G ¯ l , t H y ¯ l , t ,
where G ¯ l , t = H ¯ l , t Ψ B ¯ × P ¯ . When B ¯ < P ¯ and g t ( l ) is sparse, compressive sensing approach such as basis pursuit or orthogonal matching pursuit algorithms [25,26] can be applied to estimate g t ( l ) .

4.2. Proposed Gradient Updating Method Using Maximal Ratio Combining and Thresholding

From (12), the server can estimate the gradient reported from the lth client, g ¯ ^ t ( l ) . Note that, because the channel gain of the wireless backhaul link is varying over the round during the federated learning process. The ill-conditioned channel with small channel gain may increase the estimation error and distort the gradient information associated with the lth client. Accordingly, in what follows, we propose two gradient update methods based on the channel gain, H ¯ l , t .

4.2.1. Gradient Update by Maximum Ratio Combining

Note that the estimate of g t ( l ) is more reliable for larger channel gain. To see this, by considering a simple case with B ¯ = P ¯ , we can rewrite (12) as:
g ¯ ^ t ( l ) = Ψ B ¯ × P ¯ 1 ( H ¯ l , t H H ¯ l , t ) 1 H ¯ l , t H y ¯ l , t = g ¯ t ( l ) + Ψ B ¯ × P ¯ 1 ( H ¯ l , t H H ¯ l , t ) 1 H ¯ l , t H n ¯ t .
  Accordingly, the mean squared estimation error is proportional to σ n 2 H ¯ l , t F 2 . Equivalently, the effective SNR can be given as H ¯ l , t g ¯ t ( l ) F 2 σ n 2 . Therefore, when updating the aggregated gradient at the server from g t ( l ) , l = 1 , , L , instead of (8), we can exploit the weighted sum of g t ( l ) as
g t = l = 1 L w t ( l ) g ¯ ^ t ( l ) ,
where the weight w t ( l ) that maximizes the effective output SNR can be derived as:
w t ( l ) = H ¯ l , t F 2 l = 1 L H ¯ l , t F 2 ,
which is denoted as the maximum ratio combining (MRC) weights and allows the gradient vector that has undergone a better channel to contribute more to the aggregated gradient at the server. This is because it is more reliable and less-distorted through the wireless backhaul link, as observed from (13). To the best of our knowledge, the gradient update strategy by channel-based MRC in federated learning system with wireless backhaul has not been considered before.

4.2.2. Binary Gradient Update by Thresholding

When the backhaul links for all clients have small channel gain simultaneously, the server may receive severely distorted gradient vectors even though it exploits the MRC strategy, such as (14). Accordingly, we propose a method in which the round associated with all channel having small channel gains is excluded from federated learning. Specifically, if l = 1 L H ¯ l , t F 2 < ϵ , the associated gradient is not updated at the server, where ϵ is a pre-defined constant. Based on the above description, the proposed federated learning process is summarized in Algorithm 2.
Algorithm 2. Proposed federated learning train process.
  1:
Initialize θ 0 based on He normal weight initialization method
  2:
g 0 0
  3:
for t 1 to T do
  4:
      (Clients) θ t ( l ) θ t 1 ( l ) + g t 1
  5:
      (Clients) Update g t ( l ) from the datasets Φ t r ( l )
  6:
      (Clients) Report g t ( l ) to the server via the backhaul link
  7:
      if  l = 1 L H ¯ l , t F 2 < ϵ  then
  8:
         (Server) g t l = 1 L w t ( l ) g ¯ ^ t ( l ) as in (14)
  9:
      else
  10:
         (Server) g t g t 1
  11:
      end if
  12:
      (Server) Broadcast g t to multi-clients
  13:
end for

4.3. Adaptive Power Allocation Strategy Based on the Gradient Information

When the transmission power of each client is limited, rather than allocating power equally to all channel resources (equivalently, slots), it is effective to allocate more power on the channel slots carrying specific important information. Note that, when training a deep learning model, the gradient elements with large values imply the large change of weight to decrease the loss function. Accordingly, because g t ( l ) ( R P × 1 ) in (3) is split into multiple P ¯ dimensional vectors, g ¯ t ( l ) in (9), each client allocate its transmit power proportionally to the magnitude of g ¯ ( l ) in our proposed power allocation strategy. Assuming that N ¯ = P / P ¯ is an integer and then, the number of multiple split vectors is given as N ¯ . The adaptive power allocation strategy can be accomplished by setting:
Ψ B ¯ × P ¯ H Ψ B ¯ × P ¯ = N ¯ g t ( l ) 2 I P ¯ .
We note that the constraint of (10) allows the equal power to be used when transmitting the split vector g ¯ t ( l ) , while the constraint of (16) allows the power to be used in proportion to the magnitude of g ¯ t ( l ) at each transmission, exhibiting the same total transmit power as in (10). In addition, the power allocation as (16) has not been considered in the conventional federated learning methods over wireless channels.

5. Experiment Results

To see the validation of the proposed federated learning train strategy discussed in Section 4, we develop the CNN module for handwriting character recognition having the architecture in Figure 2. Specifically, the CNN module has three two-dimensional convolutional layers and the values for the hyperparameters exploited in the computer simulations are summarized in Table 1. Then, the number of elements in the gradient vector g ( l ) is given as 5.26 × 10 4 . The CNN module is shared by three clients connected to the server over the wireless channel. Throughout the simulations, we exploit the handwriting MNIST dataset where N w i d t h = N h e i g h t = 28 . In addition, three clients are considered and the received SNR at the server is defined as:
S N R r e c = l = 1 L σ h l 2 L σ n 2 ,
where σ n 2 is the variance of the AWGN. In addition, we split the gradient vector into multiple vectors having 128 elements (i.e., P ¯ = 128 in (9)).
In Figure 3 (respectively, Figure 4), we evaluate the classification accuracy and CE loss of the conventional gradient updating method based on the equal-weight combining and the proposed updating method based on MRC, discussed in Section 4.2 for high SNR ( S N R r e c = 15 dB) (respectively, low SNR ( S N R r e c = 10 dB)). For comparison purposes, the performance of the federated learning with error-free backhaul link is also evaluated. Here, the channel gain of each client is set as σ h l 2 = { 0.3 , 1.0 , 3.0 } and the threshold level in given as ϵ = 1.0 , and this value was experimentally determined. For the local training of the commonly shared CNN module, ADAM optimizer is adopted [27] at each client with a fixed learning rate, η = 0.001 .
From Figure 3, when the backhaul link is perfect and noise free, the classification accuracy increases in proportion to the rounds and the accuracy up to 0.97 can be achieved. In contrast, due to the channel fading and noise in the wireless backhaul link, training does not proceed stably when the conventional equal-weight combining is exploited. In Round 120, there is a sharp increase at the loss curve from 0.28 to 2.75 , resulting in the decrease in the accuracy from 0.92 to 0.11 . In contrast, the performance of the proposed updating method based on MRC in Section 4.2 exhibits a similar performance to that with the perfect backhaul link. In Figure 4, it can be found that, for low SNR, the classification accuracy of the equal-weight combining is not improved as the rounds increases and is below 0.15. In addition, the associated CE loss goes to infinity. At low SNR, it is difficult to recover the distortion caused over the wireless backhaul link when transmitting the gradient for model update. Especially, when there is channel distortion, the equal-weight combining does not reflect the received SNR in the gradient update and fails to train the distributed CNN modules. Interestingly, the updating method based on MRC and thresholding shows unstable peak in the CE loss, but it can avoid the CE loss divergence and improve the classification accuracy as the learning round increases.
In Figure 5, we evaluate the classification accuracy for various threshold levels ϵ with (a) S N R r e c = 15 dB and (b) S N R r e c = 10 dB when the updating method with MRC and thresholding in Section 4.2.2 is exploited. From Figure 5a, at high SNR, the federated learning can be well operated through the gradient updating method with MRC and thresholding, regardless of the threshold levels. However, for ϵ = 10.0 , the accuracy does not effectively increase as the learning round increases. That is, for a larger threshold level, more local gradient vectors transferred through the wireless channel can be discarded. In Figure 5b, it can be found that the classification performance is more sensitive to the threshold level at low SNR compared to the high SNR case. Specifically, as ϵ is larger, the federated learning is performed stably. This is also because the gradient update vector containing the amplified noise due to small channel gain can be discarded for large ϵ . We note that the large ϵ may incur the gradient updating delay, which leads the trade-off between the federated learning performance and the learning delay.
In Figure 6, to validate the adaptive power allocation strategy in Section 4.3, we evaluate the classification accuracy of various gradient updating methods with/without the adaptive power allocation strategy when the received SNR is low with different threshold levels (i.e., (a) ϵ = 1.0 and (b) ϵ = 0.1 ). It can be found that the accuracy of the MRC based gradient updating method with ϵ = 1.0 in Figure 6a is more stable compared to that with ϵ = 0.1 in Figure 6b, which coincides with the observation in Figure 5. Interestingly, by exploiting the adaptive power allocation strategy jointly with the MRC based gradient updating method in Figure 6a, the accuracy can be improved by 96.7% and it is comparable to the performance with error-free backhaul link. In addition, from Figure 6b, the adaptive power allocation strategy drastically stabilizes the federated learning performance during the learning process over wireless channel even for small ϵ = 0.1 . Accordingly, the adaptive power allocation strategy improves the trade-off between the federated learning performance and the learning delay due to the threshold level discussed in Figure 5.
In Table 2 and Table 3, the confusion matrices for the test dataset are evaluated after the federated learning is completed, where the proposed gradient updating method (Table 2) and the conventional updating method (Table 3) are, respectively, exploited. From Table 2, the proposed gradient updating method shows the classification accuracy of 0.9 or more for all labels. However, from Table 3, the CNN module trained through the conventional gradient updating method over wireless channel misclassifies most test data with specific labels.

6. Conclusions

In this paper, efficient gradient updating strategies are developed for federated learning when distributed clients are connected to the server via a wireless backhaul link. That is, a common CNN module is shared for all the distributed clients and it is trained through the federated learning over wireless backhaul connected to the main server. During the training phase, local gradients need to be transferred from the distributed clients to the server over a wireless noisy backhaul link. To overcome the distortion due to wireless channel fading, an effective SNR maximizing gradient updating method is proposed, in which the gradients are combined such that the effective SNR is maximized at the server. In addition, when the backhaul links for all clients have small channel gain simultaneously, the server may have severely distorted gradient vectors. Accordingly, we propose a binary gradient updating strategy based on thresholding in which the round associated with all channels having small channel gains is excluded from federated learning, which results in the trade-off between the federated learning performance and the learning delay. Due to the channel fading and noise in the wireless backhaul link, training does not proceed stably with the conventional equal-weight combining especially at low SNR. In contrast, the updating method based on MRC and thresholding improves the classification accuracy as the learning round increases by avoiding the CE loss divergence. Finally, we also propose an adaptive power allocation method, in which each client allocates its transmit power proportionally to the magnitude of the gradient information. Note that the gradient elements with large values imply the large change of weight to decrease the loss function. Through the computer simulations, it is confirmed that the adaptive power allocation strategy can improve the trade-off between the federated learning performance and the learning delay due to the threshold level.

Author Contributions

Conceptualization, Y.Y. and J.P.; methodology, Y.Y. and J.P.; software, Y.Y., Y.H. and J.P.; validation, J.P.; formal analysis, J.P.; investigation, Y.Y. and Y.H.; writing—original draft preparation, Y.Y.; writing—review and editing, J.P.; supervision, J.P.; project administration, J.P. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the Basic Science Research Program through the National Research Foundation of Korea, funded by the Ministry of Education (2018R1D1A1B07043786).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhang, K.; Zuo, W.; Gu, S.; Zhang, L. Learning Deep CNN Denoiser Prior for Image Restoration. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  2. Hadhrami, E.A.; Mufti, M.A.; Taha, B.; Werghi, N. Transfer learning with convolutional neural networks for moving target classification with micro-Doppler radar spectrograms. In Proceedings of the 2018 International Conference on Artificial Intelligence and Big Data (ICAIBD), Chengdu, China, 26–28 May 2018; pp. 148–154. [Google Scholar] [CrossRef]
  3. Tsagkatakis, G.; Aidini, A.; Fotiadou, K.; Giannopoulos, M.; Pentari, A.; Tsakalides, P. Survey of Deep-Learning Approaches for Remote Sensing Observation Enhancement. Sensors 2019, 19, 3929. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Kim, Y.; Moon, T. Human Detection and Activity Classification Based on Micro-Doppler Signatures Using Deep Convolutional Neural Networks. IEEE Geosci. Remote Sens. Lett. 2016, 13, 8–12. [Google Scholar] [CrossRef]
  5. Rieke, N.; Hancox, J.; Li, W.; Milletari, F.; Roth, H.R.; Albarqouni, S.; Bakas, S.; Galtier, M.N.; Landman, B.A.; Maier-Hein, K.; et al. The future of digital health with federated learning. NPJ Digit. Med. 2020, 3, 1–7. [Google Scholar] [CrossRef] [PubMed]
  6. Chaabene, S.; Bouaziz, B.; Boudaya, A.; Hökelmann, A.; Ammar, A.; Chaari, L. Convolutional Neural Network for Drowsiness Detection Using EEG Signals. Sensors 2021, 21, 1734. [Google Scholar] [CrossRef] [PubMed]
  7. Nafea, O.; Abdul, W.; Muhammad, G.; Alsulaiman, M. Sensor-Based Human Activity Recognition with Spatio-Temporal Deep Learning. Sensors 2021, 21, 2141. [Google Scholar] [CrossRef] [PubMed]
  8. Yang, Q.; Liu, Y.; Chen, T.; Tong, Y. Federated machine learning: Concept and applications. ACM Trans. Intell. Syst. Technol. (TIST) 2019, 10, 1–19. [Google Scholar] [CrossRef]
  9. Konečnỳ, J.; McMahan, B.; Ramage, D. Federated optimization: Distributed optimization beyond the datacenter. arXiv 2015, arXiv:1511.03575. [Google Scholar]
  10. Bottou, L. Large-scale machine learning with stochastic gradient descent. In Proceedings of the COMPSTAT’2010, Paris, France, 22–27 August 2010; Springer: Berlin/Heidelberg, Germany, 2010; pp. 177–186. [Google Scholar]
  11. Zhao, Y.; Li, M.; Lai, L.; Suda, N.; Civin, D.; Chandra, V. Federated learning with non-iid data. arXiv 2018, arXiv:1806.00582. [Google Scholar]
  12. Bonawitz, K.; Ivanov, V.; Kreuter, B.; Marcedone, A.; McMahan, H.B.; Patel, S.; Ramage, D.; Segal, A.; Seth, K. Practical secure aggregation for privacy-preserving machine learning. In Proceedings of the 2017 ACM SIGSAC Conference on Computer and Communications Security, Dallas, TX, USA, 30 October–3 November 2017; pp. 1175–1191. [Google Scholar]
  13. Konečnỳ, J.; McMahan, H.B.; Yu, F.X.; Richtárik, P.; Suresh, A.T.; Bacon, D. Federated learning: Strategies for improving communication efficiency. arXiv 2016, arXiv:1610.05492. [Google Scholar]
  14. Amiri, M.M.; Gündüz, D. Federated Learning Over Wireless Fading Channels. IEEE Trans. Wirel. Commun. 2020, 19, 3546–3557. [Google Scholar] [CrossRef] [Green Version]
  15. Amiri, M.M.; Gündüz, D. Machine Learning at the Wireless Edge: Distributed Stochastic Gradient Descent Over-the-Air. In Proceedings of the 2019 IEEE International Symposium on Information Theory (ISIT), Paris, France, 7–12 July 2019; pp. 1432–1436. [Google Scholar] [CrossRef] [Green Version]
  16. Jeon, Y.S.; Amiri, M.M.; Li, J.; Poor, H.V. A Compressive Sensing Approach for Federated Learning Over Massive MIMO Communication Systems. IEEE Trans. Wirel. Commun. 2021, 20, 1990–2004. [Google Scholar] [CrossRef]
  17. Chen, M.; Yang, Z.; Saad, W.; Yin, C.; Poor, H.V.; Cui, S. A Joint Learning and Communications Framework for Federated Learning Over Wireless Networks. IEEE Trans. Wirel. Commun. 2021, 20, 269–283. [Google Scholar] [CrossRef]
  18. Lecun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef] [Green Version]
  19. Zhu, G.; Wang, Y.; Huang, K. Broadband Analog Aggregation for Low-Latency Federated Edge Learning. IEEE Trans. Wirel. Commun. 2020, 19, 491–506. [Google Scholar] [CrossRef]
  20. McMahan, B.; Moore, E.; Ramage, D.; Hampson, S.; y Arcas, B.A. Communication-efficient learning of deep networks from decentralized data. In Proceedings of the 20th International Conference on Artificial Intelligence and Statistics, Fort Lauderdale, FL, USA, 20–22 April 2017; pp. 1273–1282. [Google Scholar]
  21. He, K.; Zhang, X.; Ren, S.; Sun, J. Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1026–1034. [Google Scholar]
  22. Xiong, J.; Bi, R.; Zhao, M.; Guo, J.; Yang, Q. Edge-Assisted Privacy-Preserving Raw Data Sharing Framework for Connected Autonomous Vehicles. IEEE Wirel. Commun. 2020, 27, 24–30. [Google Scholar] [CrossRef]
  23. Kang, D.; Ahn, C.W. Communication Cost Reduction with Partial Structure in Federated Learning. Electronics 2021, 10, 2081. [Google Scholar] [CrossRef]
  24. Candès, E.J. The restricted isometry property and its implications for compressed sensing. C. R. Math. 2008, 346, 589–592. [Google Scholar] [CrossRef]
  25. Kim, S.; Yun, U.; Jang, J.; Seo, G.; Kang, J.; Lee, H.; Lee, M. Reduced Computational Complexity Orthogonal Matching Pursuit Using a Novel Partitioned Inversion Technique for Compressive Sensing. Electronics 2018, 7, 206. [Google Scholar] [CrossRef] [Green Version]
  26. Chen, S.; Donoho, D. Basis pursuit. In Proceedings of the 1994 28th Asilomar Conference on Signals, Systems and Computers, Pacific Grove, CA, USA, 31 October–2 November 1994; Volume 1, pp. 41–44. [Google Scholar]
  27. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
Figure 1. Federated Learning over wireless communication system for target classification.
Figure 1. Federated Learning over wireless communication system for target classification.
Sensors 21 06791 g001
Figure 2. CNN module for handwriting character recognition.
Figure 2. CNN module for handwriting character recognition.
Sensors 21 06791 g002
Figure 3. (a) Classification accuracy and (b) CE loss curves at S N R r e c = 15 dB.
Figure 3. (a) Classification accuracy and (b) CE loss curves at S N R r e c = 15 dB.
Sensors 21 06791 g003
Figure 4. (a) Classification accuracy and (b) CE loss curves at S N R r e c = 10 dB.
Figure 4. (a) Classification accuracy and (b) CE loss curves at S N R r e c = 10 dB.
Sensors 21 06791 g004
Figure 5. Classification accuracy according to various threshold levels when the updating method with MRC and thresholding in Section 4.2.2 are exploited for (a) S N R r e c = 15 dB and (b) S N R r e c = 10 dB.
Figure 5. Classification accuracy according to various threshold levels when the updating method with MRC and thresholding in Section 4.2.2 are exploited for (a) S N R r e c = 15 dB and (b) S N R r e c = 10 dB.
Sensors 21 06791 g005
Figure 6. Comparison of classification accuracy with (a) ϵ = 1.0 and (b) ϵ = 0.1 for S N R r e c = 10 dB.
Figure 6. Comparison of classification accuracy with (a) ϵ = 1.0 and (b) ϵ = 0.1 for S N R r e c = 10 dB.
Sensors 21 06791 g006
Table 1. The values for the hyperparameters of the CNN module for handwriting character recognition.
Table 1. The values for the hyperparameters of the CNN module for handwriting character recognition.
Values
The number of layers3
The number of filters at each layer3
Filter size (The 1st layer), K f 11 × K f 12 ( 1 × 16 )
Filter size (The 2nd layer), K f 21 × K f 22 ( 16 × 32 )
Filter size (The 3rd layer), K f 31 × K f 32 ( 32 × 64 )
OptimizerADAM optimizer [27]
Learning rate, η 0.001
Table 2. Confusion matrix for hand writing character recognition of the proposed gradient updating method.
Table 2. Confusion matrix for hand writing character recognition of the proposed gradient updating method.
PredictedTrue Label
Label0123456789
00.9840.0010.0200.0040.0050.0060.0080.0030.0220.007
100.9790.00200.0050.0010.0020.0020.0010.001
200.0040.9430.0050000.0150.0040
300.0010.0060.96300.006000.0070.007
400.0010.00100.92900.0020.0010.0050.004
500.00100.0190.0010.9550.0050.0010.0030
60.0100.0050.00300.0160.0150.98000.0070.001
70.0010.0010.0160.0070.0010.00300.9490.0070.016
800.0070.0090.0020.0010.0110.0010.0030.9150.006
90.00500.00100.0420.0030.0010.0260.0290.961
Table 3. Confusion matrix for hand writing character recognition of equal-weight combining based gradient updating method.
Table 3. Confusion matrix for hand writing character recognition of equal-weight combining based gradient updating method.
PredictedTrue Label
Label0123456789
00.0100.0040.0120.0040.0110.0260.0100.0020.0050.001
10000000000
20.9770.8750.9610.8770.9320.7770.9710.8420.9190.943
30000000000
40.0090.0850.0260.1130.0390.1720.0170.1260.0590.036
50000000000
60000000000
700.0250.0010.0020.016000.02200.020
80.0040.01100.0040.0020.0260.0020.0070.0170.001
900.00100000000
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Yang, Y.; Hong, Y.; Park, J. Efficient Gradient Updating Strategies with Adaptive Power Allocation for Federated Learning over Wireless Backhaul. Sensors 2021, 21, 6791. https://doi.org/10.3390/s21206791

AMA Style

Yang Y, Hong Y, Park J. Efficient Gradient Updating Strategies with Adaptive Power Allocation for Federated Learning over Wireless Backhaul. Sensors. 2021; 21(20):6791. https://doi.org/10.3390/s21206791

Chicago/Turabian Style

Yang, Yunji, Yonggi Hong, and Jaehyun Park. 2021. "Efficient Gradient Updating Strategies with Adaptive Power Allocation for Federated Learning over Wireless Backhaul" Sensors 21, no. 20: 6791. https://doi.org/10.3390/s21206791

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop