Next Article in Journal
A Retrieval Augmentation Self-Distillation Method for Math Word Problem Solving
Previous Article in Journal
Signal Decomposition-Based MEG Analysis for Motor and Cognitive Imagery Classification
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Electrical Power Optimization of Cloud Data Centers Using Federated Learning Server Workload Allocation

Mechanical, Automotive and Materials Engineering Department, University of Windsor, Windsor, ON N9B 3P4, Canada
*
Author to whom correspondence should be addressed.
Electronics 2025, 14(17), 3423; https://doi.org/10.3390/electronics14173423
Submission received: 26 June 2025 / Revised: 18 August 2025 / Accepted: 26 August 2025 / Published: 27 August 2025

Abstract

Cloud Data Centers (CDCs) are the foundation of the digital economy, enabling data storage, processing, and connectivity for different academia/industry/commerce activities and digital services worldwide. As a result, their consistent power supply and reliable performance are critical factors; however, few works have considered power consumption optimization based on intelligent workload allocation. To this end, the proposed paper presents a Federated Learning (FL)-based server workload allocation model for optimal power optimization. In this strategy, the servers are modeled based on their Central Processing Unit (CPU), memory, storage, and network usage. A global server is considered as the global model responsible for final workload allocation decisions. Each server acts as a client in the federated learning framework, sharing its derived parameters with the global model securely and federatedly. Finally, after ten epochs of the system running, the model could optimize the system, decrease the overall power consumption, and reduce the workload pressure in each server by distributing it to other servers. The model is evaluated using different Key Performance Indicators (KPIs), and an appendix is provided, including the full performance results, workload shifting logs, and server resource status. Overall, the suggested FL allocator model shows promise in significantly lowering power consumption and alleviating server workload efficiently.

1. Introduction

Cloud Data Centers (CDCs) are considered fundamental infrastructure globally, with reliable performance being the top priority, relying on their role in the digital economy [1,2,3,4,5,6,7], as well as the Artificial Intelligence (AI) applications [8,9,10,11,12,13,14]. Based on a report by Goldman Sachs, a banking investment corporation, about $ 1 trillion will be invested in AI, chips, and a CDC power supply [15]. Given their significance, CDCs consume considerable electricity, highlighting the necessity for more efficient, clean, and sustainable power sources, making them ideal candidates for power-efficient designs [16]. To this end, several works have been conducted to pave the way for developing efficient and clean CDC power optimization [17]. Chidolue et al. [18] introduced the concept of green CDCs. These centers concentrate on power-efficient hardware, as well as renewable power utilization, advanced cooling, and resource optimization with sustainability. Despite their higher upfront costs and technical complexity, these technologies promise lower operational costs and less environmental impact. The rise of server workload analytics can significantly affect CDC power consumption and resource utilization. The dynamics of server operations, including task scheduling and resource allocation, directly impact different parts, such as cooling systems [19]. Optimizing workloads through intelligent management strategies, such as load balancing, spatio-temporal flexibility [20,21,22,23], and virtualization, can reduce power demands and improve sustainability in CDC operations. For instance, GreenGLB, an online algorithm, is presented in [24] utilizing a greedy approach for distributing interactive, indivisible workloads across geo-distributed CDCs. Indivisible workloads, sequential tasks assigned to a single CDC, are allocated based on current electricity prices, renewable energy availability, and predefined constraints. Following this concept, a game theory-based workload management framework was designed, in [25], to minimize cloud operating costs through intelligent scheduling. The framework accounts for data transfer costs, CDC queueing delays, and factors such as heterogeneous compute capabilities, cooling power, task co-location interference, time-of-use electricity pricing, renewable energy, net metering, peak demand pricing, and network pricing. Additionally, the work [26] presented a hierarchical workload management approach for geographically distributed CDCs, considering two algorithms for workload assignment and migration. These algorithms calculate a cost function to determine the optimal site for running workloads across CDCs, requiring minimal state information exchange while preserving CDC autonomy and enabling integrated management of heterogeneous platforms. Moreover, a review of optimization-based workload distribution strategies is conducted by [27], for geographically distributed CDCs, focusing on recent research. It examined optimization factors, including workload type, load balancer, renewable energy availability, energy storage, and server specifications.
The study [28] proposed an energy and workload management framework for sustainable CDCs, using stochastic optimization to adapt to variable renewable energy availability, cooling efficiency, Information Technology (IT) workload shifts, and energy price fluctuations while meeting long-term quality-of-service requirements.
Additionally, The review conducted in [29] highlighted the importance of energy efficiency in CDCs in reducing operational costs, minimizing carbon footprints, and promoting sustainable growth. Over 25 energy efficiency metrics and 250 literature references were considered, analyzing various energy-consuming components and configurations in CDCs. Through real-world case studies, the review evaluated how corporations implement these metrics, while also considering their challenges and limitations.
In addition to the significance of power consumption, the model designed to optimize power is also important. Accordingly, a FL-based workload allocation model is presented in this paper. Besides other previously published models, the proposed strategy models the CDC servers by their Central Processing Unit (CPU), network, memory, and storage devices. Next, it considers a global server as the global model and the other servers as the clients. Each client is analyzed using the local 3rd-degree polynomial regressor model, and the results are then federatively transferred to the global server to make the final evaluation and decision. Then, the proposed model is evaluated against several Key Performance Indicators (KPIs), and shifting results are derived. Moreover, the contributions of the proposed model are summarized as follows:
  • a FL-based intelligent model is developed to allocate the workload quickly, efficiently, and securely with the least errors;
  • the distribution process is conducted in just 10 intervals, tending Mean Square Error (MSE) to zero;
  • an appendix is provided, including the workload shift logs, full global/local KPIs, and server status, demonstrating the low error rates in the presented KPIs.
The paper is structured as follows: the proposed system and the FL strategy are presented in Section 2. The results and conclusion are drawn in Section 3 and Section 4, respectively. Finally, the details of log shifts, server status, and KPIs are available in Appendix A.

2. Proposed Workload Shifting Strategy

2.1. Preliminary Statement

The workload shift, in this paper, is conducted on CDC servers. In each server (Figure 1), the power consumption is mainly from the CPU, memory, storage, and network devices.
In a CDC server, the CPU is among the main components responsible for executing instructions and performing calculations, directly impacting the server’s processing power and power consumption [30]. The memory, including Random Accessible Memory (RAM), stores data that is actively used by the CPU, with higher memory usage leading [31] to increased power consumption due to the need for faster data retrieval/processing. Storage devices, typically consisting of Hard-Disk Drive (HDD) or Solid-State Drive (SSD), store large volumes of data and increase power consumption, particularly during read/write operations [32]. Finally, network devices, such as the Network Interface Card (NIC) [33] and switches [34], handle data communication between servers, often influencing power use as network traffic increases, particularly in high-throughput or low-latency environments.

2.2. Intelligent Shifting Strategy

Federated Learning (FL), firstly introduced in [35], is a decentralized approach to Machine Learning (ML), where multiple devices or data sources work together to train a global model, all while keeping their data private and secure [36]. In a typical FL setup, data resides on distributed devices, such as smartphones, Internet-of-Things (IoT) sensors, or edge servers, where each device locally processes its data to generate model updates. Instead of transmitting raw data, these updates are sent to a central server, where they are aggregated to improve the global model. This process is repeated, with the central server sending the updated model back to the devices for further training. It also solves issues related to data access by allowing data from different devices or locations to contribute to training without needing to store it all in one place. Additionally, FL reduces communication overhead compared to traditional distributed learning, as only model parameters or gradients are transferred, not entire datasets of each server. To this end, the overall structure of the developed FL model for the server’s workload allocation is shown in Figure 2.
The proposed FL-based model optimizes power consumption across a distributed system of 100 servers, each acting as a client. Each server trains a local 3 rd -degree polynomial regression model to predict power consumption based on workload characteristics, specifically CPU, memory, storage, and network usage.
These models are aggregated into a global model to guide dynamic workload redistribution, minimizing total power consumption while considering server capacity constraints and a power threshold. The proposed strategy is conceptualized in Figure 3 and presented in Algorithm 1 and Figure 4.
Algorithm 1 Proposed strategy.
  1:
Initialize: Set N 100 servers. For each server S i , define:
  2:
     Capacity vector C i [ C i CPU , C i Mem , C i Stor , C i Net ] R 0 4 .
  3:
     Dynamic state vector at time t, X i ( t ) [ X i CPU ( t ) , X i Mem ( t ) , X i Stor ( t ) , X i Net ( t ) ] R 0 4 .
  4:
Function UpdateUsage( S i , t ):
  5:
     Sample metrics from uniform distributions:
  6:
         X i CPU ( t ) U ( 10 , 90 ) , X i Mem ( t ) U ( 0.1 , C i Mem ) ,
  7:
         X i Stor ( t ) U ( 0.1 , C i Stor ) , X i Net ( t ) U ( 0.1 , C i Net ) .
  8:
     Compute power consumption:
  9:
         P i ( t ) 50 + 200 X i CPU ( t ) 100 1.5 + 0.5 X i Mem ( t ) + 3 ( X i Net ( t ) ) 1.2 + 0.1 X i Stor ( t ) + N ( 0 , 0.3 ) .
10:
Function TrainLocalModel( D i ):
11:
     Let D i { ( x i , j , y i , j ) } j = 1 m i , where x i , j R 4 (CPU, Mem, Net, Stor usage), y i , j R (power).
12:
     Standardize features: x i , j x i , j μ i σ i + 10 10 , where μ i , σ i are mean and std of D i .
13:
     Scale outputs: y i , j y i , j max ( y i , 1 , , y i , m i ) .
14:
     Split data: D i train 80 % , D i val 20 % .
15:
     Train 3rd-degree polynomial regression: y ^ i , j f ( x i , j ; θ i ) , minimizing:
16:
         J ( θ i ) 1 | D i train | j D i train ( y i , j y ^ i , j ) 2 .
17:
     Evaluate on D i val : Compute KPIs {MSE, RMSE, MAE, MAPE, MaxErr} using EvaluateModel.
18:
     Return model M i , KPIs, μ i , σ i , y max , i .
19:
Function AggregateModels( { M i } i = 1 N , { D i } i = 1 N ):
20:
     Pool data: D pool i = 1 N D i , where D pool { ( x j , y j ) } j = 1 m .
21:
     Standardize features: x j x j μ pool σ pool + 10 10 , where μ pool , σ pool are mean and std of D pool .
22:
     Scale outputs: y j y j max ( y 1 , , y m ) .
23:
     Train global 3rd-degree polynomial regression: y ^ j f ( x j ; θ g ) , minimizing:
24:
         J ( θ g ) 1 m j = 1 m ( y j y ^ j ) 2 .
25:
     Return global model M g , μ pool , σ pool , y max .
26:
Function EvaluateModel( y , y ^ ):
27:
     For true values y [ y 1 , , y m ] , predictions y ^ [ y ^ 1 , , y ^ m ] , compute:
28:
         MSE 1 m j = 1 m ( y j y ^ j ) 2 ,
29:
         RMSE MSE ,
30:
         MAE 1 m j = 1 m | y j y ^ j | ,
31:
         MAPE 100 m j = 1 m y j y ^ j max ( y j , 10 10 ) ,
32:
         MaxErr max j | y j y ^ j | .
33:
     Return {MSE, RMSE, MAE, MAPE, MaxErr}.
34:
Function ShiftWorkload( { S i } i = 1 N , M g , μ pool , σ pool , y max ):
35:
     Identify high-power servers: H { S i P i ( t ) > 150 } .
36:
     For each S i H :
37:
         Define workload to shift: Δ X i [ X i CPU ( t ) / 2 , X i Mem ( t ) / 2 , X i Net ( t ) / 2 , X i Stor ( t ) / 2 ] .
38:
         For each candidate server S k S i :
39:
             If S k satisfies: X k CPU ( t ) + Δ X i CPU 90 , X k Mem ( t ) + Δ X i Mem 0.8 C k Mem , X k Net ( t ) + Δ X i Net 0.8 C k Net , X k Stor ( t ) + Δ X i Stor 0.8 C k Stor :
40:
                   Predict power: P i old M g ( X i ( t ) ; μ pool , σ pool , y max ) , P k old M g ( X k ( t ) ; μ pool , σ pool , y max ) .
41:
                 Predict new power: P i new M g ( X i ( t ) Δ X i ; μ pool , σ pool , y max ) , P k new M g ( X k ( t ) + Δ X i ; μ pool , σ pool , y max ) .
42:
                 Compute savings: Δ P ( P i new + P k new ) ( P i old + P k old ) .
43:
             Select S k with maximum savings Δ P > 0 .
44:
         If valid S k found, update: X i ( t ) X i ( t ) Δ X i , X k ( t ) X k ( t ) + Δ X i , recompute P i ( t ) , P k ( t ) .
45:
         Log shift.
The process, as outlined in Figure 4 and Algorithm 1, manages a system of 100 servers to optimize resource usage and minimize power consumption through a FL-driven approach. Each server S i is initialized with a capacity vector ( C i ):
C i = [ C i CPU , C i Mem , C i Stor , C i Net ] R 0 4
representing its CPU ( C i CPU ), memory ( C i Mem ), storage ( C i Stor ), and network (( C i Net )) capabilities. Then, a dynamic state vector ( X i ( t ) ) is considered as
X i ( t ) = [ X i CPU ( t ) , X i Mem ( t ) , X i Stor ( t ) , X i Net ( t ) ] R 0 4
capturing real-time resource usage ( X i CPU ( t ) , X i Mem ( t ) , X i Stor ( t ) , X i Net ( t ) , for CPU, memory, storage, and the network) at time t.
The strategy proceeds by updating each server’s usage metrics through sampling from uniform distributions: X i CPU ( t ) U ( 10 , 90 ) , X i Mem ( t ) U ( 0.1 , C i Mem ) , X i Stor ( t ) U ( 0.1 , C i Stor ) , and X i Net ( t ) U ( 0.1 , C i Net ) . Power consumption is calculated based on
P i ( t ) = 50 + 200 X i CPU ( t ) 100 1.5 + 0.5 X i Mem ( t ) + 3 ( X i Net ( t ) ) 1.2 + 0.1 X i Stor ( t ) + N ( 0 , 0.3 )
incorporating contributions from all resources with an additional noise, which is Gaussian.
Next, each server trains a local 3 rd -degree polynomial regression model on its dataset by
D i = { ( x i , j , y i , j ) } j = 1 m i
where x i , j R 4 represents resource usage, and y i , j R is the power consumption. Features are standardized using
x i , j x i , j μ i σ i + 10 10
In Equation (5), x i , j R 4 is the j th workload feature vector for server i, containing a CPU, memory, network, and storage usage. The symbol μ i R 4 denotes the mean of the feature vectors in the local dataset D i of server i. Additionally, σ i R 4 is considered as the standard deviation of the feature vectors in D i . A small constant 10 10 is added to σ i to prevent division by zero.
Then, the outputs are scaled: y i , j y i , j max ( y i , 1 , , y i , m i ) . The data is split into 80% training and 20% validation sets, and the model minimizes the MSE ( J ( θ i ) ):
J ( θ i ) = 1 | D i train | j D i train ( y i , j y ^ i , j ) 2
Performance is evaluated on the validation set using KPIs (MSE, Root Mean Square Error (RMSE), Mean Absolute Error (MAE), Mean Absolute Percentage Error (MAPE), Max Error).
Local models are aggregated by pooling all datasets as in
D pool = i = 1 N D i
standardizing features and scaling outputs similarly, and training a global ( 3 rd )-degree polynomial regression model to minimize global MSE ( J ( θ g ) ):
J ( θ g ) = 1 m j = 1 m ( y j y ^ j ) 2
The global model’s performance is assessed using the same KPIs.
Workload optimization identifies high-power servers ( P i ( t ) > 150 ) in a set H . For each S i H , a workload of Equation (9) is defined for transfer:
Δ X i = [ X i CPU ( t ) / 2 , X i Mem ( t ) / 2 , X i Net ( t ) / 2 , X i Stor ( t ) / 2 ]
The proposed FL-driven strategy searches for a target server S k subject to the properties defined in
F ( X k , Δ X i , t ) = 1 if X k CPU ( t ) + Δ X i CPU 90 X k Mem ( t ) + Δ X i Mem 0.8 C k Mem X k Net ( t ) + Δ X i Net 0.8 C k Net X k Stor ( t ) + Δ X i Stor 0.8 C k Stor 0 Otherwise
The feasibility function for the system at time t, denoted F ( X k , Δ X i , t ) , is defined as 1 if the presented conditions are satisfied. In this sequence, Δ X i CPU , Δ X i Mem , Δ X i Net , Δ X i Stor are the incremental resource demands, and C k Mem , C k Net , C k Stor are the capacity limits for Memory, Network, and Storage; otherwise, F ( X k , Δ X i , t ) = 0 .
Power predictions from the global model are used to compute savings as
Δ P = ( P i new + P k new ) ( P i old + P k old )
selecting the S k with maximum savings ( Δ P > 0 ). If a valid S k is found, workloads are updated, power is recomputed, and the shift is logged. If no valid target exists, the failure is logged.
Finally, the global model is finalized. Then, performance metrics and workload shift logs are logged, showing balancing resource usage and minimizing power consumption through FL-driven optimization.
As a result, each server will require fewer resources to handle the workload, leading to reduced power consumption while all dedicated workloads will be completed as intended.
Performance is evaluated using KPIs [37], such as MSE, MAE, MAPE, RMSE, and maximum error, summarized in Table 1.
The Mean Square Error measures the average of the squared differences between predicted and actual values, emphasizing larger errors due to the squaring operation [37]. On the other side, the MAE quantifies the average magnitude of errors in a set of predictions, calculated as the average of the absolute differences between predicted and actual values [38]. Following this concept, the MAPE expresses the average absolute error as a percentage of the actual values, calculated by taking the absolute difference between predicted and actual values, dividing by the actual value, and averaging these percentages [39]. Additionally, the RMSE is considered a measure of the average magnitude of prediction errors in the same units as the original data [40,41]. Finally, maximum error represents the largest single deviation between a predicted and actual value in a dataset, capturing the worst-case scenario of a model’s predictions.

3. Results and Discussion

The proposed 100 servers are modeled in this section, and the FL strategy, with simultaneous training across all servers, is applied in the G o o g l e C o l a b platform. For the evaluation, the KPIs are calculated for the global model and each local model of servers. For the final phase, a log file is considered for the workload shift, including the initial and final servers. The results are shown in Figure 5, Figure 6, Figure 7 and Figure 8.
The proposed model reduced total power consumption across intervals, defined by workload distribution cycles, as shown in Figure 5. Power distribution became more consistent, demonstrating the model’s effectiveness in optimizing power efficiency and minimizing consumption disparities. Also, the term frequency, here means the repetitive number of servers over an identical power consumption. For example, there are about five servers that their final consumption is 150 W. As shown in Figure 6, the model balanced resource metrics, including CPU, memory, network, and storage, resulting in more uniform and optimized usage profiles. This highlights the model’s effectiveness in workload allocation, reducing bottlenecks and improving system performance.
The reduction in global model loss over successive epochs, shown in Figure 7, presents the learning convergence, confirming the effectiveness of the FL approach in adapting to server conditions and optimizing workload distribution. As depicted in Figure 6 the global model MSE decreases from approximately 2.5 to 1.0 across 10 epochs, demonstrating a steady improvement in model accuracy. This trend is complemented by the aggregated resource usage data in Figure 6 which manifests the stable average utilization of the CPU, memory, storage, and network resources over time intervals. The consistent resource usage, with the CPU and memory peaking around 40 units and storage and network remaining low, presents the efficient workload distribution and minimal overhead. Consequently, aggregated resource usage metrics remained stable over time, as shown in the results.
This observed stability, achieved through dynamic workload adjustments, underscores the robustness of the model in sustaining consistent performance even when server capacities fluctuate. Following implementation, server statuses were systematically categorized into four distinct power consumption levels—Low, Moderate, High, and Very High—as illustrated in Figure 8, providing a clear framework for monitoring and comparison. Notably, higher initial consumption levels were closely linked to increased memory and network activity, reflecting the greater strain placed on resources during peak demand. However, by intelligently redistributing workloads across servers, the system effectively alleviated these imbalances, mitigating excessive resource usage on heavily burdened nodes and narrowing disparities in utilization.
The proposed FL model reduced discrepancies in resource utilization by aligning workloads, ensuring balanced computational demands, and promoting equitable resource usage.
The model achieved the defined KPIs, with the least errors, as detailed in the Appendix A, demonstrating the precision and efficiency of the approach in optimizing workload allocation and enhancing system performance.

4. Conclusions

This study presented a novel FL-based intelligent workload allocation model designed to optimize server utilization. The approach treats each server as a client and utilizes a global model for dynamic workload distribution. Servers were categorized into five clusters, enabling targeted workload management. The FL model decentralized data processing, ensuring privacy and scalability. The global model aggregated local server data to dynamically balance workloads, minimizing bottlenecks and enhancing efficiency. The model’s effectiveness was assessed using KPIs, including resource utilization, load balancing, scalability, and power efficiency. Logs and detailed KPI analytics are provided in the appendix. By addressing critical demands for power-efficient and reliable CDCs, the proposed model presents a scalable solution to workload management. The challenges in this field involve ensuring secure and efficient communication between distributed servers acting as clients, handling heterogeneous server resources with varying CPU, memory, storage, and network capabilities. Additionally, managing the complexity of synchronizing model updates in a federated environment, addressing potential latency issues, safeguarding data privacy, and defending against cyber attacks during the FL process, such as model poisoning or data inference attacks, are critical concerns. Future work could focus on enhancing the scalability of the model to accommodate larger and more diverse CDC infrastructures, integrating adaptive learning mechanisms to respond to dynamic workload patterns in real time, exploring hybrid optimization approaches combining FL with other AI techniques, and implementing robust fault tolerance to ensure continuous operation despite server failures or communication disruptions. Moreover, extending the framework to consider renewable energy sources and considering environmental impacts, such as C O 2 footprint, could further optimize sustainability in CDC operations, while developing advanced security protocols to detect and mitigate cyber threats would strengthen the resilience of the system.

Author Contributions

A.S.: Conceptualization, Software, Validation, Visualization, Original Writing; Review/Editing; Formal Analysis, Investigation. A.R.: Supervision, Project Management, Formal Analysis, Software, Validation, Visualization, Review/Editing, Funding Acquisition. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the Natural Sciences and Engineering Research Council of Canada (NSERC) Discovery Grant stream with the grant identification number RGPIN-2020-05513.

Data Availability Statement

The data of the study can be made available from the corresponding author (A.S.) upon reasonable request.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

AIArtificial Intelligence
CDCCloud Data Center
CPUCentral Processing Unit
FLFederated Learning
HDDHard-Disk Drive
IoTInternet-of-Things
ITInformation Technology
KPIKey Performance Indicator
MAEMean Absolute Error
MAPEMean Absolute Percentage Error
MLMachine Learning
MSEMean Square Error
NICNetwork Interface Card
NSERCNatural Sciences and Engineering Research Council of Canada
RAMRandom Accessible Memory
RMSERoot Mean Square Error
SSDSolid-State Drive

Appendix A

Appendix A.1. Shift Logs

The workload shift logs by the proposed FL are available in Table A1.
Table A1. Server workload shift logs with predicted savings.
Table A1. Server workload shift logs with predicted savings.
Shift LogShift Log
Workload shifted from Server-2 to Server-62 with predicted savings of 13.52 WWorkload shifted from Server-3 to Server-81 with predicted savings of 16.48 W
Workload shifted from Server-4 to Server-69 with predicted savings of 23.72 WNo suitable target to shift workload from Server-5
Workload shifted from Server-6 to Server-90 with predicted savings of 17.81 WWorkload shifted from Server-8 to Server-43 with predicted savings of 7.75 W
Workload shifted from Server-10 to Server-49 with predicted savings of 9.42 WWorkload shifted from Server-13 to Server-2 with predicted savings of 3.01 W
Workload shifted from Server-15 to Server-3 with predicted savings of 11.31 WWorkload shifted from Server-21 to Server-1 with predicted savings of 8.63 W
Workload shifted from Server-22 to Server-13 with predicted savings of 5.19 WWorkload shifted from Server-23 to Server-97 with predicted savings of 7.65 W
No suitable target to shift workload from Server-24No suitable target to shift workload from Server-25
Workload shifted from Server-26 to Server-23 with predicted savings of 2.08 WWorkload shifted from Server-27 to Server-14 with predicted savings of 14.56 W
Workload shifted from Server-29 to Server-26 with predicted savings of 2.73 WWorkload shifted from Server-30 to Server-27 with predicted savings of 3.76 W
No suitable target to shift workload from Server-33No suitable target to shift workload from Server-37
Workload shifted from Server-39 to Server-30 with predicted savings of 0.15 WWorkload shifted from Server-40 to Server-50 with predicted savings of 13.69 W
Workload shifted from Server-41 to Server-10 with predicted savings of 6.75 WWorkload shifted from Server-42 to Server-66 with predicted savings of 9.06 W
Workload shifted from Server-44 to Server-40 with predicted savings of 7.01 WWorkload shifted from Server-45 to Server-29 with predicted savings of 7.60 W
Workload shifted from Server-47 to Server-39 with predicted savings of 5.43 WWorkload shifted from Server-48 to Server-42 with predicted savings of 2.00 W
No suitable target to shift workload from Server-54No suitable target to shift workload from Server-55
Workload shifted from Server-56 to Server-21 with predicted savings of 8.54 WWorkload shifted from Server-57 to Server-6 with predicted savings of 6.06 W
No suitable target to shift workload from Server-58Workload shifted from Server-64 to Server-7 with predicted savings of 14.04 W
No suitable target to shift workload from Server-65Workload shifted from Server-67 to Server-63 with predicted savings of 4.79 W
No suitable target to shift workload from Server-68No suitable target to shift workload from Server-70
Workload shifted from Server-72 to Server-31 with predicted savings of 7.33 WNo suitable target to shift workload from Server-76
No suitable target to shift workload from Server-77No suitable target to shift workload from Server-79
No suitable target to shift workload from Server-83No suitable target to shift workload from Server-84
No suitable target to shift workload from Server-85No suitable target to shift workload from Server-87
Workload shifted from Server-93 to Server-17 with predicted savings of 12.65 WNo suitable target to shift workload from Server-100

Appendix A.2. Servers Status

The initial and final servers status are provided in Table A2 and Table A3, respectively.
Table A2. Initial status of the servers (before FL).
Table A2. Initial status of the servers (before FL).
Server IDCPU
Usage
(%)
Memory
Usage
(GB)
Storage
Usage
(TB)
Network
Usage
(Gbps)
Power
Consumption
(W)
Server-136.1660.330.550.32124.35
Server-261.8158.541.888.93217.77
Server-332.8016.181.789.48140.56
Server-475.6328.592.666.89226.76
Server-551.881.723.796.05151.91
Server-651.1647.223.305.05168.22
Server-711.0460.220.874.66106.88
Server-879.6330.150.644.92227.60
Server-980.3519.413.017.88240.05
Server-1016.0715.652.592.0878.41
Server-1187.144.843.752.13223.20
Server-1250.8553.681.837.21181.94
Server-1350.5438.912.794.40159.08
Server-1439.0651.212.495.21146.63
Server-1571.4259.423.010.63202.16
Server-1623.7112.520.266.64108.89
Server-1768.1046.272.164.64204.19
Server-1820.1512.320.383.5888.20
Server-1945.4842.763.414.93153.87
Server-2074.5633.341.430.21196.07
Server-2113.536.222.066.7593.23
Server-2275.3434.940.934.39215.84
Server-2330.949.701.948.12127.13
Server-2439.8945.862.430.49125.02
Server-2556.3342.743.175.82181.27
Server-2650.8035.853.014.53159.07
Server-2773.5040.961.210.95199.23
Server-2812.5745.813.400.6384.29
Server-2966.1650.292.781.56187.78
Server-3076.7142.640.862.08213.18
Server-3183.2629.120.765.75240.77
Server-3220.7123.981.163.8296.38
Server-3346.9518.503.498.95164.99
Server-3478.3514.573.646.00222.24
Server-3566.3352.062.368.92225.91
Server-3687.4550.452.715.20260.92
Server-3768.6431.292.834.03195.60
Server-3845.6016.111.751.59125.24
Server-3930.029.392.006.15114.23
Server-4053.5340.343.939.66194.38
Server-4174.5657.103.869.69253.12
Server-4276.8550.340.977.22241.72
Server-4331.7718.241.451.3798.89
Server-4479.9160.751.448.42262.19
Server-4536.7110.663.238.00136.61
Server-4613.4428.371.739.54119.02
Server-4740.3315.943.449.32153.13
Server-4841.4444.163.310.62127.52
Server-4928.7814.752.930.4389.40
Server-5074.3130.823.677.77229.01
Server-5181.9010.333.307.63238.39
Server-5240.793.360.407.16135.80
Server-5328.1945.581.258.96144.10
Server-5460.3357.660.244.05188.16
Server-5545.394.170.709.83159.80
Server-5612.4843.152.462.3689.24
Server-5740.8244.781.045.86149.38
Server-5864.7030.043.514.53188.22
Server-5972.8117.860.324.49201.70
Server-6036.4827.951.343.01119.07
Server-6185.1238.033.078.67267.00
Server-6242.4833.532.466.93153.17
Server-6350.9831.372.712.90149.61
Server-6452.3630.861.289.01183.37
Server-6538.0822.572.523.80123.31
Server-6625.2047.433.842.13106.74
Server-6788.2526.420.290.51230.24
Server-6825.3443.531.206.11123.96
Server-6945.192.402.721.36116.24
Server-7044.1140.430.281.42133.25
Server-7180.4061.582.248.34263.49
Server-7268.9418.840.355.71198.49
Server-7368.2612.991.642.78179.49
Server-7479.6115.732.339.75246.59
Server-7548.1645.332.783.15151.66
Server-7635.0357.252.713.86135.83
Server-7751.5932.973.740.88143.55
Server-7823.3061.230.922.77113.50
Server-7917.7825.961.809.69123.79
Server-8030.4828.542.440.88100.65
Server-8160.1462.783.964.10191.13
Server-8228.5131.271.294.26112.87
Server-8382.4635.462.752.84228.33
Server-8489.4461.592.222.97261.25
Server-8536.7643.813.226.14143.25
Server-8611.6233.552.998.54113.91
Server-8723.9936.070.809.99139.01
Server-8864.1045.100.267.25207.58
Server-8912.1236.832.022.0684.26
Server-9085.0659.242.797.81271.58
Server-9111.8515.752.992.7877.05
Server-9268.492.382.016.15191.16
Server-9340.7318.242.769.06153.95
Server-9463.9145.151.607.61209.42
Server-9528.3641.893.397.05133.05
Server-9659.0430.193.923.92171.55
Server-9734.5126.792.911.84110.47
Server-9816.6027.790.622.6987.23
Server-9958.5318.530.663.94164.27
Server-10088.9245.531.334.66259.43
Table A3. Final status of the servers (after FL).
Table A3. Final status of the servers (after FL).
Server IDCPU
Usage
(%)
Memory
Usage
(GB)
Storage
Usage
(TB)
Network
Usage
(Gbps)
Power
Consumption
(W)
Server-126.5833.130.965.59117.96
Server-259.4343.763.874.24180.19
Server-359.067.161.426.20171.09
Server-484.053.852.941.74212.23
Server-544.3745.361.875.45154.94
Server-672.978.820.876.28205.95
Server-729.8845.821.092.43114.34
Server-867.5336.513.997.60214.25
Server-921.1953.371.136.64124.79
Server-1070.1324.281.451.83186.12
Server-1148.5422.261.561.82134.86
Server-1235.3929.713.633.61121.39
Server-1360.8928.170.799.84205.77
Server-1415.9031.400.167.55113.18
Server-1584.5851.852.836.97262.86
Server-1644.6640.290.940.66131.67
Server-1718.1449.262.881.9596.72
Server-1810.2963.211.470.3789.60
Server-1934.9038.820.238.52149.54
Server-2046.1941.041.973.75148.82
Server-2177.1428.241.862.16207.56
Server-2276.5662.563.295.96240.96
Server-2359.8115.663.414.41168.54
Server-2468.7449.883.171.20193.25
Server-2539.3763.200.768.16168.78
Server-2666.1560.651.663.25200.31
Server-2768.7513.343.130.70172.09
Server-2817.8048.952.185.25112.11
Server-2973.2840.923.543.12207.86
Server-3067.3310.932.458.53205.36
Server-3132.2420.752.495.94122.97
Server-3210.4333.883.768.94115.14
Server-3352.2160.982.641.69162.25
Server-3410.9241.200.377.69112.98
Server-3544.5025.970.690.54123.62
Server-3642.4917.292.057.51148.23
Server-3740.6058.292.517.03162.51
Server-3827.3218.773.751.0491.28
Server-3968.8348.883.270.61190.88
Server-4071.511.140.922.28179.74
Server-4177.5159.911.749.16259.87
Server-4278.1730.262.011.07206.40
Server-4327.5526.521.150.8695.21
Server-4484.2831.623.706.11247.00
Server-4584.6956.472.359.10276.21
Server-4616.2455.093.225.75115.34
Server-4779.0039.802.874.13227.36
Server-4871.1514.103.339.56222.31
Server-4921.7623.742.126.39110.37
Server-5017.6442.472.045.99112.06
Server-5139.526.651.972.52112.66
Server-5214.2752.873.817.50121.26
Server-5318.4962.222.675.32119.70
Server-5465.1124.112.031.62172.49
Server-5567.8933.791.252.19186.48
Server-5688.4821.603.369.69273.84
Server-5784.3763.333.259.68282.78
Server-5841.5129.593.417.68153.60
Server-5929.8959.173.847.88148.16
Server-6038.2926.843.203.68125.25
Server-6143.9833.700.712.88135.74
Server-6212.5720.131.121.4273.26
Server-6329.1918.162.753.11102.43
Server-6489.070.143.977.65252.98
Server-6542.7345.521.035.58152.56
Server-6627.1835.371.106.50124.48
Server-6771.0038.420.815.77213.50
Server-6839.5736.432.277.25150.25
Server-6915.7823.021.713.0685.50
Server-7068.1220.393.483.72187.67
Server-7118.6423.543.889.26121.65
Server-7284.1942.950.821.33230.10
Server-7337.4531.131.361.97118.20
Server-7440.8741.293.772.68132.76
Server-7557.3710.182.440.10142.37
Server-7668.9630.382.776.45207.53
Server-7749.867.292.257.92159.99
Server-7827.1541.152.492.36107.75
Server-7931.6747.930.179.32153.46
Server-8037.0752.710.560.12121.58
Server-8110.141.512.131.3261.33
Server-8237.6235.743.587.22146.94
Server-8341.4527.391.778.78157.53
Server-8456.4712.763.916.74170.64
Server-8552.7461.141.634.07172.81
Server-8630.8051.263.488.36148.76
Server-8753.9840.631.924.99170.22
Server-8826.5051.291.596.36130.89
Server-8922.939.783.718.63117.12
Server-9014.7442.302.674.55101.11
Server-9135.1736.471.636.95140.87
Server-9222.1260.961.271.09104.70
Server-9365.032.730.166.88186.57
Server-9439.419.182.556.96135.22
Server-9531.2461.050.745.74139.45
Server-9622.7660.690.452.94113.05
Server-9720.6242.361.361.0593.16
Server-9824.2344.982.939.11139.30
Server-9948.883.333.243.13132.04
Server-10062.6840.310.465.34191.88

Appendix A.3. Key Performance Indicators (KPIs) Results

The KPIs results, derived by the proposed FL strategy, are manifested in Table A4.
Table A4. Results of KPIs by the FL model.
Table A4. Results of KPIs by the FL model.
Server IDMSERMSEMAEMAPE (%)Max Error
Server-10.2880198860.5366748420.4074144470.3397921981.622196721
Server-20.1436790790.3790502320.2805065480.2161752820.898467968
Server-30.216654260.4654613410.366524280.2636681320.971990076
Server-40.3731738570.6108795760.4871889920.2754123021.631013131
Server-50.2087316810.4568716240.3578522930.2283208061.185042199
Server-60.1471081560.3835468110.3129869890.2243623180.747794889
Server-70.1892993560.4350854580.3360355810.2047158731.034696044
Server-80.4086051230.639222280.4572376270.2707300561.859744114
Server-90.2115870860.4599859630.3749245320.2245493420.971419173
Server-100.18251310.427215520.3395933580.239357280.85592076
Server-110.2536350510.5036219330.3799723110.2481275471.150486454
Server-120.2256699730.4750473380.4066113210.31029830.960371416
Server-130.0943604310.307181430.2524094740.2008149660.640344365
Server-140.3026612040.550146530.426207230.2530162511.221272865
Server-150.2859907360.5347810170.4092733660.269826751.205406524
Server-160.2503970890.5003969310.4003376790.3172406871.07236305
Server-170.2780198450.5272758720.3790425670.2793373471.636041854
Server-181.0205219791.0102088790.499680580.2686648014.190005521
Server-190.1496853990.3868919740.2799743470.1804483221.240260652
Server-200.1367951860.3698583320.3142813430.2012875920.742281187
Server-210.1123511250.3351881930.2763230840.186743720.824673614
Server-220.0839809360.2897946450.2322390240.1334182980.816562594
Server-230.198269160.4452742520.3574296720.2644069761.094375506
Server-240.1533070120.3915443930.3251469510.1948597830.843868315
Server-250.1580744160.3975857350.3326718080.227320260.740955843
Server-260.349611450.5912795030.4696021610.3935956191.343870624
Server-270.3831067820.6189562040.5154461550.3021643961.481931778
Server-280.2397501880.4896429190.4018466540.275367710.943554317
Server-290.1498617990.3871198760.3153674580.2259951050.831474733
Server-300.2436643580.4936237010.4426940990.3244497730.942202404
Server-310.2139655390.4625640920.3839683320.2656244381.097994193
Server-320.2831320780.5321015670.4146200360.2625002031.306626006
Server-330.240753240.4906661180.3049295080.1900573311.714911534
Server-340.241293620.491216470.4145197320.3054101911.091901808
Server-350.3102058290.5569612450.3652076490.2789797971.734774363
Server-360.1301701720.3607910360.273596110.174114010.97542151
Server-370.1753439930.4187409620.3329469710.1946986660.961919336
Server-380.1511657670.388800420.2875467950.1810598170.92740031
Server-390.1065702550.3264509990.2566486450.1471397560.891935238
Server-400.2822110620.5312354110.4419839820.2971243481.06575691
Server-410.3205067160.5661331260.4631578720.2698696831.074913764
Server-420.1722038110.414974470.3133725630.2264839971.047679801
Server-430.2485878790.4985858790.3834837120.2570470751.194580549
Server-440.2375003430.4873400690.3813575340.3078541651.390176242
Server-450.28733510.5360364720.3763446580.2188627921.542055562
Server-460.1640986360.4050908980.311282040.2176649120.866150269
Server-470.2230439290.4722752680.3893290350.2450893650.985277919
Server-480.1242919220.3525505950.2799706040.1740457530.921008532
Server-490.0741738690.2723488010.2238341750.1538608580.506896217
Server-500.1618066190.4022519350.3234806860.1683742130.818635545
Server-510.286957860.5356844780.4127627580.3057125471.136709953
Server-520.1698549740.4121346560.3311539870.2143165240.813005245
Server-530.1309654920.3618915470.2741705460.2021881080.873418091
Server-540.1462285030.3823983570.3272445190.2398733780.802019592
Server-550.3530410490.5941725750.4682466560.3422577181.595488547
Server-560.1576859050.3970968460.3557544480.2773161310.823957055
Server-570.2704279490.5200268730.4159856640.2820639821.035317278
Server-580.1106851320.3326937510.2853196330.1823750830.655095498
Server-590.3687984630.6072877920.497401750.3013836881.62412723
Server-600.1528453120.3909543610.3094571740.2032839080.787128093
Server-610.2352015820.4849758580.3803217960.2421498491.040096973
Server-620.1005510390.3170978380.2292017970.1348794040.955221638
Server-630.1338498520.365854960.2995415410.2365624030.709002268
Server-640.0687237170.2621520880.2132134770.1373889570.558875403
Server-650.5671169990.7530717090.584635210.3958679141.956126235
Server-660.1043149340.3229782250.2751173140.1669614360.604527755
Server-670.2904945960.5389755060.416879610.2775967461.431545898
Server-680.3350486330.5788338560.4187531930.314763631.453659872
Server-690.1485296110.3853953960.2975887150.2157975240.791886707
Server-700.2160420430.4648032310.3496670380.2647185251.427185467
Server-710.1521603050.3900773060.2958990580.1896006370.947513679
Server-720.1546880690.3933040410.3315600810.2071395060.863362419
Server-730.2319046180.481564760.4099817170.2598381971.131348952
Server-740.1367932130.3698556650.2921975780.1950018181.104089215
Server-750.1809655990.4254005160.3660647910.2761720570.860999199
Server-760.406144780.6372948930.5142514690.3410108141.505812213
Server-770.1751703330.4185335510.3257891480.2242553490.893631677
Server-780.1252502860.3539071710.2959595050.2204438560.648130923
Server-790.3080503470.5550228350.4344926820.2850795791.219725973
Server-800.2076923540.4557327660.3812427110.2395160910.995601431
Server-810.0923387520.3038729210.2260982910.1510000590.718063343
Server-820.2256198950.4749946260.4026163410.260164550.81739475
Server-830.1708479180.4133375360.334339230.2403236920.828825791
Server-840.2148980620.463570990.3544966290.2056285311.114321498
Server-850.285573230.5343905220.4280736370.3313297211.253368573
Server-860.1407314290.3751418780.3175418730.1997777360.981867625
Server-870.1684539610.4104314330.3401017250.2300232980.856947381
Server-880.0778542810.2790237990.231238050.1429731890.545854272
Server-890.3554869780.5962272870.4586965060.2795698431.383205455
Server-900.1496502480.3868465430.3367457850.2284024210.620414751
Server-910.3490095180.5907702750.4694418880.2828269991.225862792
Server-920.258273720.5082063750.397524750.2511081761.498208969
Server-930.243101750.4930534960.388240140.2710746491.249693837
Server-940.1471464750.3835967610.2947952320.1736351790.836771993
Server-950.1743183630.4175145070.3459452380.1844094760.768998776
Server-960.5427946640.7367459970.4921125860.3335497442.361707797
Server-970.2788937340.5281039050.4078132030.2943109271.411883016
Server-980.2395461020.4894344720.3601214150.2160577851.492884951
Server-990.1956863250.442364470.363136130.2611687070.962433125
Server-1000.157068840.3963191130.293158250.2307873650.953071812

References

  1. Xu, D.; Qu, M. Energy, environmental, and economic evaluation of a CCHP system for a data center based on operational data. Energy Build. 2013, 67, 176–186. [Google Scholar] [CrossRef]
  2. Cho, K.; Chang, H.; Jung, Y.; Yoon, Y. Economic analysis of data center cooling strategies. Sustain. Cities Soc. 2017, 31, 234–243. [Google Scholar] [CrossRef]
  3. Xie, Y.; Cui, Y.; Wu, D.; Zeng, Y.; Sun, L. Economic analysis of hydrogen-powered data center. Int. J. Hydrogen Energy 2021, 46, 27841–27850. [Google Scholar] [CrossRef]
  4. Li, C.; He, W.; Cao, E. Impact of green data center pilots on the digital economy development: An empirical study based on dual machine learning methods. Comput. Ind. Eng. 2025, 201, 110914. [Google Scholar] [CrossRef]
  5. Shehabi, A.; Smith, S.J.; Masanet, E.; Koomey, J. Data center growth in the United States: Decoupling the demand for services from electricity use. Environ. Res. Lett. 2018, 13, 124030. [Google Scholar] [CrossRef]
  6. Ebrahimi, K.; Jones, G.F.; Fleischer, A.S. Thermo-economic analysis of steady state waste heat recovery in data centers using absorption refrigeration. Appl. Energy 2015, 139, 384–397. [Google Scholar] [CrossRef]
  7. Lykou, G.; Mentzelioti, D.; Gritzalis, D. A new methodology toward effectively assessing data center sustainability. Comput. Secur. 2018, 76, 327–340. [Google Scholar] [CrossRef]
  8. Sermet, Y.; Demir, I. A semantic web framework for automated smart assistants: A case study for public health. Big Data Cogn. Comput. 2021, 5, 57. [Google Scholar] [CrossRef]
  9. Yang, J.; Xiao, W.; Jiang, C.; Hossain, M.S.; Muhammad, G.; Amin, S.U. Ai-powered green cloud and data center. IEEE Access 2018, 7, 4195–4203. [Google Scholar] [CrossRef]
  10. Richins, D.; Doshi, D.; Blackmore, M.; Nair, A.T.; Pathapati, N.; Patel, A.; Daguman, B.; Dobrijalowski, D.; Illikkal, R.; Long, K.; et al. Ai tax: The hidden cost of ai data center applications. ACM Trans. Comput. Syst. (TOCS) 2021, 37, 1–32. [Google Scholar] [CrossRef]
  11. Lu, W.; Liang, L.; Kong, B.; Li, B.; Zhu, Z. AI-assisted knowledge-defined network orchestration for energy-efficient data center networks. IEEE Commun. Mag. 2020, 58, 86–92. [Google Scholar] [CrossRef]
  12. Liu, Y.; Du, H.; Niyato, D.; Kang, J.; Xiong, Z.; Wen, Y.; Kim, D.I. Generative AI in data center networking: Fundamentals, perspectives, and case study. IEEE Netw. 2025. [Google Scholar] [CrossRef]
  13. Marahatta, A.; Xin, Q.; Chi, C.; Zhang, F.; Liu, Z. PEFS: AI-driven prediction based energy-aware fault-tolerant scheduling scheme for cloud data center. IEEE Trans. Sustain. Comput. 2020, 6, 655–666. [Google Scholar] [CrossRef]
  14. Davenport, C.; Singer, B.; Mehta, N.; Lee, B.; Mackay, J.; Modak, A.; Corbett, B.; Miller, J.; Hari, T.; Ritchie, J.; et al. AI, Data Centers and the Coming US Power Demand Surge. Goldman Sachs. 2024. Available online: https://www.spirepointpc.com/documents/FG/spirepoint/resource-center/629373_Generational_Growth__AI_data_centers_and_the_coming_US_power_demand_surge.pdf (accessed on 28 April 2024).
  15. Moss, S.; Trueman, C. Goldman Sachs: $1tn to Be Spent on AI Data Centers, Chips, and Utility Upgrades, with “Little to Show for It So Far”; Data Center Dynamics: London, UK, 2024. [Google Scholar]
  16. Van Geet, O.; Sickinger, D. Best Practices Guide for Energy-Efficient Data Center Design; Technical Report; National Renewable Energy Laboratory (NREL): Golden, CO, USA, 2024.
  17. del Rio, A.; Conti, G.; Castano-Solis, S.; Serrano, J.; Jimenez, D.; Fraile-Ardanuy, J. A guide to data collection for computation and monitoring of node energy consumption. Big Data Cogn. Comput. 2023, 7, 130. [Google Scholar] [CrossRef]
  18. Chidolue, O.; Ohenhen, P.E.; Umoh, A.A.; Ngozichukwu, B.; Fafure, A.V.; Ibekwe, K.I. Green data centers: Sustainable practices for energy-efficient it infrastructure. Eng. Sci. Technol. J. 2024, 5, 99–114. [Google Scholar] [CrossRef]
  19. Rostami, S.; Down, D.G.; Karakostas, G. Linearized Data Center Workload and Cooling Management. IEEE Trans. Autom. Sci. Eng. 2024, 22, 3502–3514. [Google Scholar] [CrossRef]
  20. Ali, A.; Özkasap, Ö. Spatial and thermal aware methods for efficient workload management in distributed data centers. Future Gener. Comput. Syst. 2024, 153, 360–374. [Google Scholar] [CrossRef]
  21. Cao, Y.; Cao, F.; Wang, Y.; Wang, J.; Wu, L.; Ding, Z. Managing data center cluster as non-wire alternative: A case in balancing market. Appl. Energy 2024, 360, 122769. [Google Scholar] [CrossRef]
  22. Chen, D.; Ma, Y.; Wang, L.; Yao, M. Spatio-temporal management of renewable energy consumption, carbon emissions, and cost in data centers. Sustain. Comput. Inform. Syst. 2024, 41, 100950. [Google Scholar] [CrossRef]
  23. Su, C.; Wang, L.; Sui, Q.; Wu, H. Optimal scheduling of a cascade hydro-thermal-wind power system integrating data centers and considering the spatiotemporal asynchronous transfer of energy resources. Appl. Energy 2025, 377, 124360. [Google Scholar] [CrossRef]
  24. Khalil, M.I.K.; Ahmad, I.; Almazroi, A.A. Energy Efficient Indivisible Workload Distribution in Geographically Distributed Data Centers. IEEE Access 2019, 7, 82672–82680. [Google Scholar] [CrossRef]
  25. Hogade, N.; Pasricha, S.; Siegel, H.J. Energy and Network Aware Workload Management for Geographically Distributed Data Centers. IEEE Trans. Sustain. Comput. 2022, 7, 400–413. [Google Scholar] [CrossRef]
  26. Forestiero, A.; Mastroianni, C.; Meo, M.; Papuzzo, G.; Sheikhalishahi, M. Hierarchical Approach for Efficient Workload Management in Geo-Distributed Data Centers. IEEE Trans. Green Commun. Netw. 2017, 1, 97–111. [Google Scholar] [CrossRef]
  27. Ahmad, I.; Khalil, M.I.K.; Shah, S.A.A. Optimization-based workload distribution in geographically distributed data centers: A survey. Int. J. Commun. Syst. 2020, 33, e4453. [Google Scholar] [CrossRef]
  28. Chen, T.; Wang, X.; Giannakis, G.B. Cooling-Aware Energy and Workload Management in Data Centers via Stochastic Optimization. IEEE J. Sel. Top. Signal Process. 2016, 10, 402–415. [Google Scholar] [CrossRef]
  29. Safari, A.; Sorouri, H.; Rahimi, A.; Oshnoei, A. A Systematic Review of Energy Efficiency Metrics for Optimizing Cloud Data Center Operations and Management. Electronics 2025, 14, 2214. [Google Scholar] [CrossRef]
  30. Möbius, C.; Dargie, W.; Schill, A. Power consumption estimation models for processors, virtual machines, and servers. IEEE Trans. Parallel Distrib. Syst. 2013, 25, 1600–1614. [Google Scholar] [CrossRef]
  31. Salmanian, Z.; Izadkhah, H.; Isazadeh, A. Optimizing web server RAM performance using birth–death process queuing system: Scalable memory issue. J. Supercomput. 2017, 73, 5221–5238. [Google Scholar] [CrossRef]
  32. Wang, S.; Lu, Z.; Cao, Q.; Jiang, H.; Yao, J.; Dong, Y.; Yang, P. BCW: Buffer-Controlled Writes to HDDs for SSD-HDD Hybrid Storage Server. In Proceedings of the USENIX Conference on File and Storage Technologies, Santa Clara, CA, USA, 24–27 February 2020. [Google Scholar]
  33. Zhan, J.; Jiang, W.; Li, Y.; Wu, J.; Zhu, J.; Yu, J. NIC-QF: A design of FPGA based Network Interface Card with Query Filter for big data systems. Future Gener. Comput. Syst. 2022, 136, 153–169. [Google Scholar] [CrossRef]
  34. Kim, D.; Liu, Z.; Zhu, Y.; Kim, C.; Lee, J.; Sekar, V.; Seshan, S. TEA: Enabling State-Intensive Network Functions on Programmable Switches. In Proceedings of the Annual Conference of the ACM Special Interest Group on Data Communication on the Applications, Technologies, Architectures, and Protocols for Computer Communication, Virtual Event USA, 10–14 August 2020. [Google Scholar]
  35. McMahan, B.; Moore, E.; Ramage, D.; Hampson, S.; y Arcas, B.A. Communication-efficient learning of deep networks from decentralized data. In Proceedings of the Artificial Intelligence and Statistics, PMLR, Fort Lauderdale, FL, USA, 20–22 April 2017; pp. 1273–1282. [Google Scholar]
  36. Zhang, C.; Xie, Y.; Bai, H.; Yu, B.; Li, W.; Gao, Y. A survey on federated learning. Knowl.-Based Syst. 2021, 216, 106775. [Google Scholar] [CrossRef]
  37. Safari, A.; Daneshvar, M.; Anvari-Moghaddam, A. Energy Intelligence: A Systematic Review of Artificial Intelligence for Energy Management. Appl. Sci. 2024, 14, 11112. [Google Scholar] [CrossRef]
  38. Dasi, H.; Ying, Z.; Ashab, M.F.B. Proposing hybrid prediction approaches with the integration of machine learning models and metaheuristic algorithms to forecast the cooling and heating load of buildings. Energy 2024, 291, 130297. [Google Scholar] [CrossRef]
  39. Indira, G.; Bhavani, M.; Brinda, R.; Zahira, R. Electricity load demand prediction for microgrid energy management system using hybrid adaptive barnacle-mating optimizer with artificial neural network algorithm. Energy Technol. 2024, 12, 2301091. [Google Scholar] [CrossRef]
  40. Hodson, T.O. Root mean square error (RMSE) or mean absolute error (MAE): When to use them or not. Geosci. Model Dev. Discuss. 2022, 2022, 5481–5487. [Google Scholar] [CrossRef]
  41. Tayalati, F.; Azmani, A.; Azmani, M. Application of supervised machine learning methods in injection molding process for initial parameters setting: Prediction of the cooling time parameter. Prog. Artif. Intell. 2024, 1–17, Erratum in Prog. Artif. Intell. 2024. [Google Scholar] [CrossRef] [PubMed]
Figure 1. A server rack schematic, including servers themselves, storage, and networking units.
Figure 1. A server rack schematic, including servers themselves, storage, and networking units.
Electronics 14 03423 g001
Figure 2. Overall workload shifting process using FL, considering a global server and local servers.
Figure 2. Overall workload shifting process using FL, considering a global server and local servers.
Electronics 14 03423 g002
Figure 3. Visualization of workload shifting for servers.
Figure 3. Visualization of workload shifting for servers.
Electronics 14 03423 g003
Figure 4. Flowchart of the overall process to distribute workload by the proposed FL strategy.
Figure 4. Flowchart of the overall process to distribute workload by the proposed FL strategy.
Electronics 14 03423 g004
Figure 5. Final results: Power consumption, and Power distribution in the servers.
Figure 5. Final results: Power consumption, and Power distribution in the servers.
Electronics 14 03423 g005
Figure 6. Usage Results: the CPU, Storage, Memory, and Network in the servers.
Figure 6. Usage Results: the CPU, Storage, Memory, and Network in the servers.
Electronics 14 03423 g006
Figure 7. Results of global model loss and aggregated resource usage (the CPU, memory, storage, and network).
Figure 7. Results of global model loss and aggregated resource usage (the CPU, memory, storage, and network).
Electronics 14 03423 g007
Figure 8. Workload distribution results, by clustering the servers.
Figure 8. Workload distribution results, by clustering the servers.
Electronics 14 03423 g008
Table 1. Considered KPIs in this work.
Table 1. Considered KPIs in this work.
Metric NameFormulationBest/Worst
MSE 1 n i = 1 n ( y i y ^ i ) 2 0/ +
MAE 1 n i = 1 n | y i y ^ i | 0/ +
MAPE 1 n i = 1 n y i y ^ i y i × 100 0%/100%
RMSE 1 n i = 1 n ( y i y ^ i ) 2 0/ +
Max. Error max i y i y ^ i 0/ +
where n, y i , and y ^ i are the total number of analyzed, real, and predicted values, respectively.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Safari, A.; Rahimi, A. Electrical Power Optimization of Cloud Data Centers Using Federated Learning Server Workload Allocation. Electronics 2025, 14, 3423. https://doi.org/10.3390/electronics14173423

AMA Style

Safari A, Rahimi A. Electrical Power Optimization of Cloud Data Centers Using Federated Learning Server Workload Allocation. Electronics. 2025; 14(17):3423. https://doi.org/10.3390/electronics14173423

Chicago/Turabian Style

Safari, Ashkan, and Afshin Rahimi. 2025. "Electrical Power Optimization of Cloud Data Centers Using Federated Learning Server Workload Allocation" Electronics 14, no. 17: 3423. https://doi.org/10.3390/electronics14173423

APA Style

Safari, A., & Rahimi, A. (2025). Electrical Power Optimization of Cloud Data Centers Using Federated Learning Server Workload Allocation. Electronics, 14(17), 3423. https://doi.org/10.3390/electronics14173423

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop