1. Introduction
Cloud computing has fundamentally transformed the computing landscape, enabling enterprises, organizations, and academic institutions to offload the management of costly IT infrastructure to cloud data centers [
1,
2]. These data centers offer elastic, reliable, and cost-effective services, hosting millions of physical servers, storage devices, and cooling facilities [
3]. This infrastructure supports a wide array of delay-sensitive applications, including e-commerce, scientific computing, high-performance simulations, and big data analytics [
4]. The scalability and flexibility of cloud data centers have led to their widespread adoption across various sectors.
However, the rapid proliferation of cloud-based applications has significantly increased the energy consumption of cloud data centers. In 2023, data centers in the United States consumed approximately 176 terawatt-hours (TWh) of electricity, accounting for about 4.4% of the nation’s total electricity consumption. Projections indicate that by 2028, U.S. data center energy consumption could reach between 325 and 580 TWh, representing up to 12% of the country’s electricity demand [
5]. Globally, data centers consumed around 340 TWh in 2023, constituting approximately 1.3% of worldwide electricity usage. This figure is expected to rise, driven by the increasing computational demands of artificial intelligence (AI) and other emerging technologies [
6].
The rising energy demands of cloud data centers (CDCs) pose significant environmental challenges, as much of their electricity still comes from fossil fuels, contributing to greenhouse gas emissions and climate change. To mitigate this impact, cloud service providers are increasingly adopting green data centers (GDCs), integrating renewable energy sources such as solar, wind, hydro-, and geothermal power to reduce their carbon footprints. For instance, as of 2024, Apple Inc. has invested heavily in renewable energy, supporting over 18 gigawatts of clean electricity across its global operations and supply chain [
7].
As cloud services expand globally, data centers are evolving into geographically distributed infrastructures, leading to the emergence of green distributed data centers (GDDCs) [
8]. These infrastructures span multiple locations, leveraging regional renewable energy sources to enhance sustainability and optimize energy use. However, renewable energy availability fluctuates across regions and deregulated electricity markets introduce dynamic pricing variations. Additionally, factors such as server availability, task service rates, and queue capacities vary spatially and temporally, making resource management increasingly complex. Furthermore, GDDCs depend on multiple Internet Service Providers (ISPs) for global data transfers, adding another layer of complexity due to variable bandwidth availability and pricing across regions. Moreover, GDDCs must efficiently schedule diverse workloads while meeting strict delay-bound constraints, ensuring cost-effective execution of heterogeneous applications [
9].
With their decentralized architecture, efficient task scheduling in GDDCs has become a critical challenge. Unlike traditional centralized data centers, GDDCs must dynamically adapt to variations in renewable energy supply, electricity costs, and network conditions. This variability necessitates intelligent workload distribution to achieve both energy efficiency and high performance and requires advanced resource allocation strategies. Scheduling tasks in GDDCs involve balancing energy consumption and Quality of Service (QoS) in highly dynamic environments [
10]. While minimizing energy usage reduces operational costs, it may compromise system performance, affecting response times and reliability, especially for delay-sensitive applications. Additionally, network bandwidth constraints and heterogeneous resource capacities across distributed data centers further complicate scheduling decisions. To address these challenges, adaptive scheduling strategies are required to dynamically allocate resources, ensuring energy-efficient execution while maintaining QoS guarantees [
11].
Extensive research in both academia and industry has focused on minimizing energy costs in CDCs [
12,
13]. While numerous task-scheduling strategies have been proposed to optimize energy consumption, many fail to account for the spatial variability of energy availability, electricity pricing, and network conditions in distributed infrastructures. Moreover, existing approaches often treat QoS as a fixed constraint rather than an optimization objective, limiting their ability to dynamically balance energy efficiency and service performance [
14]. To bridge this gap, this work addresses the joint optimization of energy cost and Quality of Service in distributed green infrastructures through a dynamic, adaptive metaheuristic framework. Specifically, the proposed work addresses three key challenges: (i) how to schedule delay-sensitive applications across geo-distributed GDDCs under highly variable energy, pricing, and network conditions; (ii) how to minimize the Average Task Loss Probability (ATLP)—a delay-bound violation metric—alongside operational cost; and (iii) how to efficiently solve the resulting high-dimensional, constrained multi-objective optimization problem in real time.
We propose a multi-objective optimization model that simultaneously minimizes the total operational cost and ATLP. Our solution is an Adaptive Firefly-Based Bi-Objective Optimization (AFBO) algorithm that introduces adaptive mechanisms to the baseline Firefly Algorithm (FA) [
15,
16] for constraint handling, randomized exploration, and convergence control. These enhancements improve global search mobility, refine the exploration–exploitation trade-off, and accelerate convergence speed, making the proposed approach highly suitable for real-time GDDC task scheduling. A representative knee solution is extracted from each generated Pareto front using the minimum Manhattan distance method, which determines the optimal service rates and task splits across GDDCs.
This work contributes to the field in several important ways. First, it formulates a realistic, delay-aware cost-QoS optimization framework tailored for dynamic GDDC environments. Second, it introduces a structured set of design extensions to the baseline Firefly Algorithm (FA), resulting in a novel adaptive firefly-based metaheuristic capable of robust, bi-objective optimization under complex constraints. Third, it combines penalty-based reformulation with Pareto-based knee selection to support efficient exploration of bi-objective trade-offs. Finally, it validates the proposed approach through extensive simulations based on real-world cloud traces, demonstrating significant improvements in service reliability, convergence speed, and energy efficiency.
This paper is structured as follows. The next section reviews related work, highlighting the novel contributions of this study.
Section 3 presents the GDDC architecture and formulates the multi-objective optimization problem for a GDDC provider.
Section 4 details the problem-solving methodology.
Section 5 discusses the proposed AFBO approach, while
Section 6 evaluates its performance through real-life trace-driven experiments and comparisons with state-of-the-art methods. Finally,
Section 7 concludes the paper with key insights and future research directions.
3. Problem Formulation
This section formulates the bi-objective optimization problem considered in this work. It presents the mathematical modeling of the system—encompassing key assumptions, operational constraints, and objective functions—that collectively define the scheduling problem in green distributed data centers (GDDCs).
Section 3.1 introduces the system architecture, including task sources, routing, and energy provisioning.
Section 3.2,
Section 3.3 and
Section 3.4 then describe the queuing behavior, energy-cost dynamics, and QoS constraints, leading to the formal problem formulation.
3.1. System Architecture and Assumptions
Green data centers (GDCs) cater to a diverse range of global users by providing various cloud services. A typical cloud provider operates multiple GDCs distributed across different geographical locations, as illustrated in the DGDC system architecture in
Figure 1. Each GDC consists of thousands of servers dedicated to processing user tasks, with power supplied from multiple energy sources, including wind, solar, and non-renewable grids. Users submit tasks to GDCs via different devices such as smartphones, tablets, and computers.
These users may include individuals engaged in latency-sensitive activities such as cloud gaming, video conferencing, and remote desktop services, as well as enterprise systems performing real-time analytics or smart factory operations. Such scenarios generate geographically distributed, delay-critical workloads that require energy-aware task scheduling to ensure both low-latency performance and sustainable operations.
Following the approach in [
10], we assume that applications are replicated and distributed across all GDCs, allowing each task to be executed independently at any available GDC. Tasks are transmitted to the cloud data centers through multiple ISPs. Based on the application type, tasks are first classified and then managed using First-In-First-Out (FIFO) queues specific to each application category [
40]. Key operational data—including queue statuses, renewable energy availability, grid electricity costs, and ISP pricing at each location—are periodically collected and transmitted to a task scheduler. This scheduler implements the proposed Adaptive Firefly-Based Bi-Objective Optimization (AFBO) algorithm to determine the Pareto-optimal task service rates, ISP-based task distribution, and optimal configuration of GDC servers. By dynamically optimizing these parameters at each time slot, the framework minimizes the total operational cost for the GDDC provider while ensuring strict Quality of Service (QoS) compliance.
3.2. Delay-Constrained Task Model
At time slot
,
represents the task arrival rate, while
denotes the server service rate for application
n in GDC
c. Following the approach in [
10], servers in GDC
c for application
n are modeled using an M/M/1/
/∞ queuing system, where
denotes the queue capacity for application
n in GDC
c. Based on this model, the average response time for application
n tasks in GDC
c is computed as follows:
where
Let
represent the maximum allowable response time for tasks of application
n. The average time response of application
n tasks in GDC
c must not exceed this constraint:
The service rate of tasks for application
n corresponds to the rate at which they are processed and removed from their respective FIFO queue. This rate is determined by incorporating the task loss probability for application
n at time slot
in GDC
c. Under the adopted M/M/1/
/∞ queuing model, the task’s loss possibility
at time slot
for application
n in GDC
c, is expressed as follows [
41]:
where
is the capacity of each server queue in GDC
c for application
n tasks and
is defined by Equation (
4). Let
define application
n’s tasks arriving rate at time slot
delivered through ISP
k to DGDC
c. The stability of an application
n’s task queue in GDC
c is guaranteed as long as the task arrival rate
does not exceed the service rate
at time slot
. This condition imposes the following constraints:
Moreover, at each time slot
, the total task arrival rate
for application
n in GDC
c must equal the sum of the task arrival rates
distributed across all
k ISPs. Additionally, the overall task arrival rate
for application
n is the sum of arrival rates across all GDCs, as expressed in the following equation:
Let
denote the average task size for application
n, and let
represent the bandwidth capacity limit of ISP
k. At each time slot
, the total allocated bandwidth through ISP
k must not exceed
, ensuring the following constraint:
3.3. Energy Consumption Model for Task Execution
Based on the homogeneous server model in [
42], each powered-on server in GDC
c processes tasks of application
n at a fixed rate of
tasks per minute. To ensure sufficient resource allocation while maintaining efficiency, the number of active servers,
, required at time slot
is given as
must not exceed
, the maximum number of servers available for application
n in GDC
c. Thus, the following constraint holds:
The number of tasks processed by each active server for application
n in GDC
c at time slot
, denoted as
, is given by
where
L is the duration of time slot
. Accordingly, let
represent the busy period (in minutes) of each powered-on server assigned to application
n in GDC
c at time slot
, defined as:
We define
as the CPU utilization of an active server handling application
n in GDC
c during time slot
, expressed as
Each powered-on server in GDC
c processing application
n’s tasks consumes energy through a combination of a base load and a dynamically varying component [
43]. The base load corresponds to the power required when the server is idle, while the proportional load scales with task execution. We define
and
as the base and load-dependent power consumption of an active server running application
n in GDC
c, respectively, computed as follows:
The parameter represents the average idle power of a server allocated to application n in GDC c, while corresponds to its average peak power. Additionally, defines the power usage effectiveness of GDC c, representing the ratio of total energy consumption (including infrastructure overhead) to the energy consumed by servers. Its typical value ranges from 1.2 to 2.0.
Therefore, the total power
of GDC
c consumed by servers in time slot
is obtained as
Thus, the total power consumed by all switched-on servers in GDC
c at time slot
is computed as follows:
Using (12) and (16), the total energy can be expressed by:
The total energy
used for processing application tasks in GDC
c at time slot
must be constrained within
, the GDC’s maximum allowable energy. The following inequality expresses this condition:
3.4. Solar and Wind Renewable Energy Model
The solar energy produced in GDC
c at time slot
, denoted as
, is determined based on the model in [
44] as follows:
where
is the active irradiation area of solar panels in GDC
c,
is the rate of converting solar energy to electricity, and
is the solar irradiance per
at time slot
.
Denoting with
the wind energy produced in GDC
c at time slot
, we compute its value using the wind energy model in [
45] as
where
is the rate to convert wind to electricity,
is the rotor area of the wind turbines,
is the on-site air density, and
is the wind speed in GDC
c.
3.5. Constrained Optimization Problem
At time slot
, the total cost of a GDDC provider, denoted as
, includes the ISP bandwidth cost
and the DGDCs energy cost
incurred for processing the scheduled tasks.
The ISP bandwidth cost at time slot
is formulated as follows:
Here, K, N, and C represent the number of available ISPs, application types, and GDCs, respectively. The ISP bandwidth price at time slot is denoted by , while defines the arrival rate of application n’s tasks delivered via ISP k to GDC c at time slot . Additionally, represents the average task size for application n and L denotes the duration of each time slot (5 min).
Let
denote the thermal power electricity price at time slot
in GDC
c; the energy cost at time slot
of the GDDC is given as
The first objective aims to minimize the operational cost of GDDCs and is formulated as follows:
The Average Task Loss Probability (ATLP), denoted as
, represents the average probability of task loss across all GDCs at time slot
. It is computed as
The second objective, which aims to minimize the ATLP,
, is expressed as
The ATLP is selected as the QoS objective because it directly captures the probability of task rejection due to service delays exceeding application-specific deadlines. This makes it highly suitable for evaluating performance in delay-sensitive GDDC environments. Compared to other QoS metrics such as latency, availability, or throughput, ATLP accounts for queue dynamics, task arrival intensity, and server capacity, offering a more integrated and service-aware performance measure. Moreover, it is mathematically aligned with the adopted M/M/1/ queuing model, allowing for tractable optimization and direct incorporation into the scheduling strategy.
Consequently, the problem is formulated as a bi-objective optimization model (
):
Solving requires jointly optimizing two interdependent decisions: (a) allocating incoming tasks among available ISPs and (b) configuring the set of active servers in each GDC at every time slot. The goal is to simultaneously minimize operational costs while optimizing Quality of Service (ATLP), ensuring that delay constraints and energy availability limits are met. An optimal solution to must effectively balance these competing objectives while adhering to the system constraints. The proposed approach for solving is detailed next, under the assumption that all system parameters, including (electricity price), (ISP bandwidth price), (wind energy availability), and (solar irradiance), are known at the beginning of each time slot .
4. Problem-Solving Methodology
This section presents the methodological foundations underlying our proposed solution to the bi-objective task-scheduling problem formulated in
Section 3. As the problem involves nonlinear, high-dimensional objectives and complex system constraints, classical optimization techniques are insufficient. We first reformulate the constrained problem using a penalty-based approach to enable tractable search in an unconstrained setting. Next, we review the limitations of conventional scalarization-based methods and discuss the advantages of direct multi-objective optimization. We then provide an overview of prominent evolutionary algorithms and motivate the selection of the Firefly Algorithm (FA) for this task. Finally, we highlight the limitations of the classical FA and introduce the rationale for our adaptive enhancements. These elements collectively establish the foundation for the Adaptive Firefly-Based Bi-Objective Optimization (AFBO) algorithm, which is formally introduced in the next section.
4.1. Reformulating the Optimization Problem
It is important to note that
and
in P1 are nonlinear with respect to
and
. As a result, P1 is a constrained nonlinear optimization problem with continuous decision variables. Constraints (5), (7)–(11), (13), and (22) are inherently nonlinear and complex. To enable metaheuristic optimization in the presence of complex constraints, we reformulate the problem by embedding constraint violations into the objective function through a penalty mechanism. Following this approach, we adopt a penalty function method [
46], eliminating explicit equality and inequality constraints by introducing a penalization function that grows to infinity when constraints are violated. Consequently, P1 is reformulated into an equivalent unconstrained nonlinear problem P2, as follows:
where
and
denote the new unconstrained objective functions,
is a large positive constant, and
represents the total constraint violation cost, defined as follows:
Here, represents the vector of decision variables and l and are positive constants.
The first summation captures penalties from violated inequality constraints , while the second term penalizes deviations from equality constraints . When all constraints are satisfied, the penalty value is zero, restoring equivalence with the original constrained problem. This reformulation enables the use of metaheuristic algorithms in a simpler unconstrained search space.
In our implementation, we set and , applying quadratic penalties for both inequality and equality constraint violations. This choice balances enforcement strength and numerical stability—penalizing larger violations more aggressively while preserving the optimizer’s ability to explore near-feasible regions. These values were selected based on commonly adopted practice in penalty-based metaheuristics and were validated empirically for convergence consistency across all scenarios tested.
4.2. Why Direct Multi-Objective Optimization?
The penalty-based reformulation in
Section 4.1 results in two conflicting objectives: minimizing operational cost and task loss probability. Selecting an appropriate multi-objective optimization strategy is thus essential. Classical scalarization techniques—such as the weighted sum or weighted metric methods—convert multi-objective problems into single-objective formulations. However, these approaches face several limitations. For instance, the weighted sum method requires predefined objective weight vectors, making it challenging to ensure a well-distributed set of Pareto-optimal solutions. Moreover, it struggles to capture certain Pareto-optimal solutions, particularly in nonconvex objective spaces, where a simple linear combination fails to explore the entire Pareto front. Additionally, this method produces only a single candidate solution per run, necessitating multiple executions with varying weights to approximate the full set of optimal trade-offs. In contrast, direct multi-objective optimization methods inherently generate a diverse set of Pareto-optimal solutions in a single run, allowing decision-makers to explore a broader range of trade-offs without repeated optimization. This improves computational efficiency and provides a more comprehensive representation of the Pareto front, making it better suited for real-world decision-making scenarios. We therefore adopt a direct multi-objective approach and review suitable evolutionary algorithms in the next subsection.
4.3. Evolutionary Approaches for Multi-Objective Optimization
Several multi-objective evolutionary optimization algorithms have been developed to simultaneously optimize multiple objectives and generate a set of Pareto-optimal solutions. Among these, multi-objective particle swarm optimization (MOPSO) [
47] is a well-known approach. However, a major drawback of MOPSO is its rapid convergence, often leading to a premature solution. As a result, MOPSO may fail to capture the true Pareto-optimal front when applied to complex multi-objective optimization problems. Another widely used method is Multi-objective Differential Evolution (MODE) [
48], a simple yet effective population-based direct search algorithm designed to handle complex multi-objective problems with constraints in continuous spaces. MODE has been validated in numerous real-world applications [
49]. However, like MOPSO, MODE is prone to premature convergence and can easily get trapped in local optima. To overcome these limitations, the authors of [
10] introduced a simulated-annealing-based bi-objective differential evolution (SBDE) algorithm to enhance solution diversity and improve the approximation of the Pareto front. In SBDE, evolutionary operations such as adaptive mutation, simulated annealing (SA)-based crossover, and SA-based selection are incorporated to improve global convergence speed and overall optimization performance.
4.4. Firefly Algorithm and Adaptation Rationale
A promising optimization technique, the Firefly Algorithm (FA) [
50], is inspired by the natural attractiveness behavior of fireflies, where individuals are drawn to others based on their flashing light signals. The basic FA was introduced in [
15], highlighting its multi-modal characteristics and superior ability to explore the search space compared to several other nature-inspired meta-heuristics. The studies in [
33,
51] further confirm the FA’s potential for efficiently solving nonlinear, non-polynomial, multidimensional, and NP-hard real-world optimization problems. Given its effectiveness in single-objective optimization, the FA has evolved into the Multi-Objective Firefly Algorithm (MOFA) to tackle multi-objective problems with complex constraints in continuous spaces [
52,
53]. MOFA leverages randomization and local search trade-offs to efficiently explore solution sets and achieve fast convergence. The balance between these trade-offs is controlled by specific algorithm parameters, which are typically tuned before execution and remain fixed during iterations [
54]. However, several modified FA variants have introduced adaptive control mechanisms that dynamically adjust these parameters during execution [
55], improving solution quality and convergence.
Building on previous advancements in FA modifications, MOFA enhancements must be tailored to the specific problem. To effectively solve our large-scale nonlinear optimization problem P2, we introduce an adaptive MOFA that dynamically optimizes the balance between exploration and exploitation. The proposed solution introduces new learning rules for parameter control, adaptation, and selection strategies to enhance its global search capability, ensuring robust and efficient global optimization of P2. The following sections first review the main steps of the basic FA, then introduce multiple novel modifications aimed at increasing FA’s global search mobility, achieving higher convergence speed and enhanced optimization efficiency. Finally, we introduce the Adaptive Firefly-based Bi-Objective Optimization (AFBO) framework and outline the methodology for selecting the final representative solution, known as the knee, from the Pareto-optimal set.
The next section formally presents the AFBO algorithm, its adaptive components, and its implementation strategy for constrained, real-time scheduling in GDDCs.
5. Adaptive Firefly-Based Bi-Objective Optimization (AFBO)
This section presents the complete Adaptive Firefly-Based Bi-Objective Optimization (AFBO) algorithm developed to solve the unconstrained problem P2. AFBO integrates new learning rules for parameter control, adaptive movement, and selection strategies to improve global search efficiency and solution quality. It extends the standard Firefly Algorithm with mechanisms for constraint handling, dominance-based evaluation, and knee selection tailored to dynamic GDDC environments. We first review the core steps of the baseline FA, followed by our adaptive enhancements, and conclude with the full AFBO framework and knee solution methodology.
5.1. Baseline Firefly Algorithm (FA): Core Steps
The pseudocode for the basic FA is outlined in [
56]. The fundamental steps of the FA are as follows:
1- Establish firefly brightness and attractiveness: At any given location
x, the brightness intensity
is determined based on the objective functions of the problem. An aggregation strategy or dominance-based mechanism is applied to define brightness for multiple objectives. The intensity also diminishes with distance from its source
, following the inverse square law, i.e.,
As light propagates through a medium with an absorption coefficient
, its attractiveness
decreases accordingly. Given
, the intensity
I at a distance
r is expressed as:
2- Calculate the distance between any two fireflies
i and
j:
where
is the dimension of one firefly vector.
3- Initialize a population of fireflies and update the position of firefly
i by moving it toward a more attractive firefly
j at each iteration
t.
The second term in (37) accounts for attractiveness, while the final term represents randomization, where
is a vector of random numbers drawn from a Gaussian distribution, and
is the randomization parameter. To control
, a geometric annealing schedule is applied, starting from its initial value
, i.e.,
where
is a damping reduction constant.
4- Update the fireflies’ light intensities, rank them accordingly, and identify the current best solution.
5.2. Adaptive Firefly Algorithm (AFA): Design Extensions
To enhance the efficiency and robustness of the Firefly Algorithm (FA), this section introduces the proposed Adaptive Firefly Algorithm (AFA). AFA integrates multiple modifications aimed at improving global search capability, convergence speed, and overall optimization efficiency. These enhancements adaptively adjust key algorithmic parameters to maintain an optimal balance between exploration and exploitation, as detailed in the following subsections.
1- Controlling the Attractiveness Parameter: In the standard FA, the initial attractiveness value is typically set to a constant
. However, we propose an adaptive rule to update
dynamically during the algorithm’s execution. Specifically, when firefly
i moves toward a more attractive firefly
j (i.e.,
), the initial brightness is adjusted as follows:
where
,
, and
is a small positive constant to prevent division by zero when
.
Moreover, in the standard FA, the absorption factor is typically set to a fixed value
, where
. To improve adaptability, we introduce a boundary-handling mechanism inspired by [
57]. Let
and
denote the upper and lower bounds, respectively. These boundaries are used to adjust the attractiveness by defining a new initial value for
as
where
D represents the maximum distance between
and
, given by
where
is the dimensionality of the firefly vector.
Consequently, the attractiveness
is computed as
2- Adaptive Randomization Parameter: We introduce an adaptive randomization parameter inspired by [
34] to enhance the balance between exploration and exploitation. A new damping scheme is proposed for the randomization factor
. While
follows a similar formulation to the standard FA, given by
the damping factor is now iteration-adaptive, ensuring a more controlled decay. In the standard FA, the randomization parameter
decreases exponentially with a fixed damping factor, which can lead to inefficient search behavior—excessive randomness in early iterations slows convergence, while insufficient randomness in later iterations risks trapping fireflies in local optima. To address this, we introduce an iteration-dependent damping factor, allowing randomness to decrease gradually in the early stages and more aggressively in later stages, striking a better balance between global exploration and local exploitation.
In the modified FA,
is defined as
where
,
t is the current iteration index, and
represents the maximum number of iterations.
3- Adaptive Scheme for the Randomization Factor: This modification introduces a novel adaptive scheme for the randomization factor by incorporating a new parameter,
, which adjusts randomization based on the fireflies’ adaptation step. The parameter is defined as:
where
is a random value drawn from a standard normal distribution and
m is a fixed positive constant. In our case,
. This ensures that randomness decreases progressively as iterations advance while still allowing slight variations due to
, maintaining diversity in the search process.
Consequently, the movement of firefly
toward a more attractive firefly
is updated as
where
dynamically scales the randomization term, ensuring a more adaptive balance between exploration and exploitation throughout the optimization process. Unlike the global damping factor
, which regulates randomness uniformly across all fireflies,
introduces a localized adaptive mechanism that varies per firefly. This refinement complements the previous modification by preventing premature stagnation, maintaining search diversity, and ensuring a more flexible adaptation of the fireflies’ movement strategy. Together, these enhancements improve both the convergence speed and the robustness of the solution.
4- Adaptive Exploration Mechanism: This modification introduces a new adaptation rule that enhances solution diversity by allowing the best fireflies to explore new solutions through random walk, while others continue moving toward superior solutions. The updated movement rule is defined as
where
is a variable fraction in
. In this approach, the best-performing fireflies execute a random walk to explore new regions of the search space, while the remaining fireflies move toward better solutions. This balances exploration and exploitation, ensuring the algorithm maintains solution diversity without sacrificing convergence.
This modification introduces two extreme cases based on the value of
. When
, the modification has no effect and the behavior remains identical to previous AFA versions. Conversely, when
, all fireflies move entirely through random walk, eliminating attraction-based movement and making the algorithm purely stochastic. Intermediate values allow the algorithm to balance convergence and exploration by selectively applying random walks to top-performing fireflies. In our design, we adopt
, following the findings in [
29], which identified this value as a robust trade-off between computational efficiency and optimization accuracy. This was also confirmed by preliminary tuning trials, where lower values (e.g.,
) reduced search diversity, while higher values (e.g.,
) introduced excessive randomness that degraded convergence speed. As such,
was selected to ensure controlled exploration without compromising the stability and convergence of the AFBO algorithm.
5.3. AFBO Framework and Multi-Objective Integration
This subsection presents the Adaptive Firefly-Based Bi-Objective Optimization (AFBO) framework, which extends the adaptive modifications introduced in the previous section to a multi-objective optimization setting. AFBO incorporates Pareto dominance, adaptive movement strategies, and a refined knee selection methodology to improve both solution quality and convergence.
1- Adaptive Pareto Front Optimization: To handle multiple objectives, AFBO operates under the Pareto dominance principle, avoiding scalarization-based methods that convert multiple objectives into a single weighted function. A solution
is said to dominate another solution
if
with at least one inequality being strict. The non-dominated solutions from the population at each iteration define the Pareto front, which forms the basis for decision making.
2- Knee Solution Selection: While the Pareto front provides a set of optimal trade-off solutions, a single representative solution—the knee solution—is often required for practical decision making. The knee point is characterized as the most balanced trade-off between conflicting objectives. It is selected based on its Manhattan distance from the ideal point.
We use Manhattan distance in this context due to its robustness and linear treatment of deviations across multiple objectives. Unlike Euclidean distance, it avoids squaring, which can exaggerate the influence of large variations in one objective. This ensures a more balanced selection between energy cost and ATLP. Moreover, cosine similarity, which captures the angular similarity between vectors, is less appropriate here because our optimization seeks absolute deviation minimization from an ideal point, not vector alignment. The Manhattan distance thus offers both simplicity and interpretability in bi-objective scheduling trade-offs.
where:
- -
and define the ideal values.
- -
The knee solution is given by
3- Roulette Wheel Selection for Knee Selection: To enhance the selection process, we introduce Roulette Wheel Selection (RWS) to ensure a more diverse and well-balanced knee selection process. Instead of always selecting the solution with the smallest
, RWS probabilistically selects a knee solution using the distribution
where:
- -
(the closest solution to the ideal point).
- -
is the annealing temperature, updated as
Using RWS ensures that the selection process remains diverse, avoiding bias toward extreme solutions, while also enhancing the algorithm’s robustness in knee selection as it converges.
5.4. The AFBO Optimization Algorithm
Based on these enhancements, the final AFBO algorithm (Algorithm 1) is presented below:
Algorithm 1 Adaptive Firefly-Based Bi-Objective Optimization (AFBO) |
- 1:
Initialize firefly population within the search space - 2:
Set parameters: - 3:
- 4:
Evaluate initial fireflies’ objective functions: - 5:
for each iteration t do - 6:
Compute damping and randomization parameters: - 7:
for each firefly i do - 8:
Compute adaptive randomization factor: - 9:
for each firefly do - 10:
if dominates (Pareto dominance) then - 11:
Compute distance: - 12:
Compute attractiveness: - 13:
Update movement: - 14:
end if - 15:
end for - 16:
end for - 17:
Compute Pareto front and store non-dominated solutions - 18:
end for - 19:
Knee Selection: - 20:
Compute Manhattan distances from ideal point - 21:
Apply RWS to select final knee solution - 22:
Return Pareto-optimal set and final knee solution
|
Algorithm 1 Computational Complexity: Let
be the number of fireflies,
T the number of iterations, and
D the number of decision variables, where
for
C data centers,
N applications, and
K ISPs. At each iteration, every firefly evaluates two objective functions, checks dominance against others, computes constraint penalties, and updates its position based on adaptive movement strategies—each operation involving all
D dimensions. Thus, the total time complexity is
. The space complexity is
, reflecting the storage of firefly positions, objective values, and the non-dominated archive. These bounds are typical for population-based multi-objective metaheuristics and are corroborated by the runtime results in
Section 6.
6. Performance Evaluation
Real-world, trace-driven experiments were conducted to evaluate the effectiveness of AFBO against state-of-the-art multiobjective evolutionary algorithms. The evaluation is based on publicly available workload traces from the Google cluster dataset [
58], a widely used benchmark in previous studies [
40,
59,
60,
61,
62]. Specifically, tasks labeled with the **SUBMIT** event type are selected as they indicate task arrivals to the scheduler. The extracted data are then grouped by day, application, and 10-min intervals. The chosen reference day for experimentation is 30 May 2024. All algorithms were implemented in MATLAB R2019b, and experiments were conducted in a standard computing environment.
6.1. System Parameters
In our model, green distributed data centers (GDDCs) are deployed across three cities: Muscat, Oman; Cairo, Egypt; and Istanbul, Turkey. Each GDC has distinct internal configurations, including the maximum number of servers, their processing capacities, and the availability of photovoltaic panels and wind generators for renewable energy supply. The workload consists of three application types, indexed as Type 1, Type 2, and Type 3, with arrival rates sampled every 10 min from the Google Cluster dataset, as illustrated in
Figure 2a. Each task type has a fixed size of 8 Mb, 5 Mb, and 2 Mb, respectively. To handle data transmission, three Internet Service Providers (ISPs) are selected, each offering different bandwidth capacities and pricing models. The maximum bandwidth capacities for ISP 1, ISP 2, and ISP 3 are 2000 Mbps, 3000 Mbps, and 4000 Mbps, respectively, with pricing structures depicted in
Figure 2b. Additionally, each GDC’s location has different solar irradiation and wind speed. The sum of solar and wind renewable energy power ratings in each location are shown in
Figure 2c.
Table 1 summarizes the configurations of renewable and thermal power sources and the characteristics of the server infrastructure at each GDC location. For this study, it is assumed that electricity prices remain constant throughout the day. The table provides the power rating of each GDC (MWh), along with electricity costs (USD/kWh). It also shows detailed configuration parameters for both wind and solar energy sources. For wind energy, the parameters include the conversion rate, rotor area (
), and air density (
), whereas the solar parameters consist of conversion efficiency and the photovoltaic panel surface area (
). The server infrastructure parameters outline critical characteristics for processing three types of delay-sensitive tasks. It specifies the delay constraints (
), queuing capacities (
), and processing rates during active (
) and idle (
) states in tasks per second. Task processing time is generated using a uniform distribution in time slots
, as described in [
42]. Moreover, the table lists the maximum number of servers (
) allocated for each task type.
6.2. Experiment 1: Performance Evaluation of AFBO vs. MOFA
Trace-driven simulations were conducted to evaluate the performance of the proposed Adaptive Firefly-Based Optimization (AFBO) in solving problem P2. Each time slot corresponds to a 10-min interval, with a total of 144 time slots simulated. To ensure reliable convergence, 1000 iterations are performed per slot.
We begin by assessing the sensitivity of the penalty factor —which controls the trade-off between constraint satisfaction and objective optimization in problem P2. For this analysis, we selected a representative time slot () and examined the normalized convergence behavior of the cost objective for different values of ranging from to . For each case, the objective function was normalized by its initial value at iteration 0, and the evolution was tracked over 200 iterations using the AFBO algorithm.
As shown in
Figure 3, small penalty values (
,
) lead to significantly slower convergence. This is expected, as constraint violations are insufficiently penalized, which delays the convergence toward feasible solutions. In contrast, for
, the objective function converges more rapidly and stabilizes earlier. This behavior confirms that the algorithm’s performance is stable within the range
, with
offering an effective trade-off between convergence speed and constraint enforcement.
Based on these results, we retain for all subsequent experiments. The next subsection continues with the comparative evaluation between AFBO and the multi-objective firefly Algorithm (MOFA). Each algorithm independently generates its own Pareto front, from which a representative knee solution is selected using the minimum Manhattan distance method. Unlike conventional approaches where QoS constraints (e.g., average task latency penalty, ATLP) are strictly enforced, this study treats ATLP as a second objective. This multi-objective formulation provides more flexibility to decision-makers by offering diverse trade-offs. As a result, the comparison is based on three key metrics: (i) total operational cost, representing the cumulative cost of ISP bandwidth and thermal grid electricity across all GDCs; (ii) ATLP, as a QoS indicator; and (iii) computational time required to reach convergence.
Figure 4 illustrates the optimized total operational costs achieved by MOFA and AFBO in all time slots. As shown in the figure, AFBO significantly outperforms MOFA in cost reduction.
Table 2 further summarizes the comparison across all three metrics. AFBO achieves the lowest average operating cost, drastically improves ATLP, and reduces runtime by approximately 53%. Specifically, operational costs are reduced nearly 22-fold, while QoS performance also improves by two orders of magnitude. Although AFBO introduces additional mechanisms for parameter adaptation and Pareto-based dominance handling, it achieves faster convergence than MOFA due to an improved exploration–exploitation balance. The adaptive controls in AFBO guide the search process more efficiently, enabling the algorithm to reach high-quality solutions in fewer iterations. These results underscore the effectiveness of the adaptive modifications in enhancing the global search capabilities and convergence efficiency of the FA framework. As a result, the overall runtime of AFBO is shorter than that of MOFA, as reported in
Table 2, despite a slightly higher computational cost per iteration. Importantly, despite the added complexity of multi-objective optimization, the convergence time of AFBO remains acceptable compared to the time slot length, making it suitable for real-time scheduling decisions.
6.3. Experiment 2: AFBO’s Performance at Solving P2
This experiment provides an in-depth performance analysis of the proposed Adaptive Firefly-Based Optimization (AFBO) in solving problem P2 under trace-driven simulations. As in Experiment 1, 1000 iterations are performed at each 10-min time slot to ensure convergence toward the global optimum.
Figure 5 illustrates the bandwidth allocated by the three Internet Service Providers (ISPs) to incoming tasks. It is evident that the bandwidth constraints of all ISPs—2000 Mbps, 3000 Mbps, and 4000 Mbps—are strictly satisfied. In particular, ISP 3 is the least utilized due to its higher bandwidth pricing, resulting in average bandwidth usages of 39.25% and 28.5% lower than ISPs 1 and 2, respectively. Conversely, ISP 1—offering the lowest data transfer costs—carries the largest task load without exceeding its 2000 Mbps limit. This behavior confirms that AFBO effectively identifies cost-efficient task distribution strategies across ISPs while adhering to bandwidth limitations.
Figure 6 presents the thermal (non-renewable) energy consumption across the three GDCs. This represents the additional power required from the grid to process tasks when the available renewable energy is insufficient. As shown, Muscat consumes the least thermal electricity among the sites, driven by Oman’s higher grid energy prices. Furthermore, during periods of high solar generation—between 6 AM and 6 PM—all GDCs operate entirely on renewable energy, avoiding any thermal electricity purchases. This behavior aligns with AFBO’s objective of maximizing GDC profitability by intelligently adjusting service rates to utilize green energy while satisfying power and processing constraints.
Figure 7 illustrates the number of switched-on servers allocated to process incoming tasks while satisfying their respective delay constraints. As shown, for all GDCs and task types, the number of active servers always remains within the maximum allowable limit defined per task type in
Table 1. Additionally, the figure reveals noticeable differences in the number of active servers used for the same task type across different GDCs. For instance, the number of active servers for type 1 tasks (
) reaches its maximum limit only at GDC 1 (
). More generally, type 1 tasks are consistently served using nearly the full server capacity across all GDCs. This behavior can be explained by two key factors: (i) type 1 tasks have the most stringent delay requirements among all task types and (ii) their associated queuing and processing capacities (see
and
in
Table 1) are smaller than those allocated to types 2 and 3. Interestingly, despite type 2 tasks having high processing demands, the number of servers assigned to them in GDC 1 remains well below its maximum limit (
) and is, in fact, the lowest compared to the allocations in GDCs 2 and 3. This is mainly due to GDC 1 (Muscat) having the least available renewable energy and the highest electricity prices. In contrast, GDC 3, which benefits from the highest level of renewable energy production, operates close to its server limit for type 2 tasks (
), making it the preferred location for processing this task type. Finally, type 3 tasks are characterized by relatively relaxed processing delay constraints, which results in fewer switched-on servers being required compared to the other task types. As a result, these tasks are predominantly processed at GDC 3, where sufficient renewable energy is consistently available to meet demand.
These findings reinforce the effectiveness of AFBO in intelligently distributing tasks across multiple GDCs while respecting delay constraints. Simultaneously, AFBO minimizes the operational costs of each GDC and maximizes the use of locally available solar and wind energy across geographically distributed infrastructures. Furthermore, the average ATLP obtained—
, as reported in
Table 2—is remarkably low, demonstrating that AFBO achieves a high Quality of Service (QoS) level. This result highlights the strength of the proposed multi-objective formulation, which not only reduces cost but also ensures near-zero latency violations, aligning well with the overall optimization objectives defined in this work.
6.4. Experiment 3: Comparison of AFBO with SBDE and MOPSO
To further validate the effectiveness of AFBO in solving problem P2, this experiment compares its performance against two state-of-the-art multi-objective evolutionary algorithms: multi-objective particle swarm optimization (MOPSO) [
47] and the simulated-annealing-based bi-objective differential evolution (SBDE) algorithm recently proposed in [
10]. The SBDE approach has been shown to significantly outperform several established methods, including NSGA-II [
30], SPEA-II [
63], and MODE [
64].
The performance comparison is based on three main evaluation criteria: convergence speed, total operational cost of the GDCs, and overall processing time. To assess convergence behavior, the time slot 35 is selected as a representative sample. The results are depicted in
Figure 8, where the base-10 logarithm of the unconstrained operational cost objective function,
, is plotted against the number of iterations for each algorithm. In all cases, constraint violations are penalized using a large penalty multiplier
to enforce feasibility. Convergence is considered achieved when the objective function stabilizes at its minimum value. The faster this minimum is reached (i.e., with fewer iterations), the better the convergence performance of the optimization algorithm.
As illustrated in
Figure 8, the proposed AFBO consistently converges to the lowest objective function values, demonstrating its effectiveness in satisfying both equality and inequality constraints while efficiently exploring the solution space for global and local optima. In contrast, the competing algorithms—particularly MOPSO and SBDE—converge to significantly higher objective values, indicating reduced capability in consistently identifying high-quality solutions. This limitation is especially pronounced for MOPSO, whose convergence stagnates far from the optimal region reached by AFBO. In addition to superior solution quality, AFBO exhibits markedly faster convergence, requiring fewer than 100 iterations compared to the approximately 350 and 850 iterations for SBDE and MOPSO, respectively. These findings underscore the robustness and efficiency of AFBO in terms of both convergence speed and optimization performance.
Figure 9 presents the total optimized operational costs for all GDCs across all time slots for each algorithm. Since each algorithm generates its own Pareto front, a representative knee solution is selected using the minimum Manhattan distance method to ensure fair comparison. The average operational costs over the simulation period are USD 20,835 for MOPSO, USD 7430 for SBDE, and just USD 325 for AFBO—corresponding to reductions by factors of 64 and 23 compared to MOPSO and SBDE, respectively. While SBDE clearly outperforms MOPSO, AFBO achieves substantially greater cost savings thanks to its ability to efficiently schedule and allocate tasks across GDCs, minimize ISP bandwidth usage, and reduce reliance on non-renewable energy. By maximizing the utilization of local wind and solar resources, AFBO ensures a more energy-efficient and sustainable operation for geographically distributed data centers.
In addition to cost optimization, AFBO exhibits excellent performance in terms of Quality of Service (QoS), measured by the second objective, which is Average Task Loss Probability (ATLP). The achieved average ATLP for AFBO is —nearly two orders of magnitude lower than that of SBDE () and several orders better than MOPSO (). This extremely low ATLP reflects a significantly reduced likelihood of task rejection, thereby ensuring more reliable service delivery across diverse system conditions. These findings reinforce the strength of AFBO in addressing complex multi-objective scheduling challenges in green distributed data centers, effectively balancing operational cost reduction with enhanced QoS.
Following the cost and QoS performance analysis, we now examine the computational efficiency of the compared algorithms.
Figure 10 illustrates the execution time of all algorithms at each time slot. It is evident that AFBO is more computationally demanding than both SBDE and MOPSO. On average, AFBO requires 305.46 s per time slot, compared to 162.64 s for SBDE and only 95 s for MOPSO. This higher execution time reflects the added complexity introduced by AFBO’s adaptive mechanisms and multi-objective optimization strategy. Although AFBO typically converges in fewer iterations, the computation time per iteration is greater, resulting in a higher total runtime. Nonetheless, this additional cost is justified by the significantly improved solution quality and lower ATLP values achieved. Moreover, AFBO remains suitable for real-time implementation, as its convergence time consistently stays well within the duration of each time slot.
7. Conclusions
The rise of green data centers (GDCs) offers a promising path toward sustainable cloud computing by integrating renewable energy sources such as solar and wind to reduce dependence on grid electricity generated from fossil fuels. However, efficient task scheduling across geo-distributed GDCs remains highly complex due to spatial variability in energy availability, electricity tariffs, and ISP bandwidth costs. These factors, combined with stringent delay and Quality of Service (QoS) requirements, necessitate intelligent scheduling strategies that can simultaneously ensure service reliability and operational cost-efficiency.
This paper formulated the underlying optimization problem as a nonlinear, bi-objective model targeting the minimization of both total operational cost and Average Task Loss Probability (ATLP)—a key Quality of Service (QoS) metric. To address this, we proposed an innovative Adaptive Firefly-Based Bi-Objective Optimization (AFBO) algorithm that integrates adaptive search mechanisms with a multi-objective Pareto-based framework. The proposed AFBO was extensively validated using real-life trace-driven simulations based on Google cluster data, ensuring realism and practical relevance.
The results confirm that AFBO consistently outperforms benchmark algorithms, including MOPSO, SBDE, and MOFA. Specifically, AFBO achieves average operational cost reductions by factors of 64, 23, and 22 compared to MOPSO, SBDE, and MOFA, respectively. From a Quality of Service (QoS) perspective, AFBO sustains an exceptionally low Average Task Loss Probability (ATLP) of —nearly two orders of magnitude lower than SBDE and MOFA and several orders lower than MOPSO. These findings highlight AFBO’s superior capability to simultaneously minimize operational costs and ensure strict adherence to QoS requirements, effectively balancing the competing objectives in multi-objective scheduling for green data centers.
In addition to its optimization accuracy, AFBO exhibits faster convergence toward global and local optima across all time slots, despite its higher computational cost. Importantly, the algorithm successfully satisfies both equality and inequality constraints for nearly all time instances, where peer methods occasionally fail to find feasible or high-quality solutions. These findings collectively validate the robustness, effectiveness, and multi-objective strength of the proposed AFBO in optimizing task placement in distributed green data centers under real-world constraints.
Given these results, the proposed AFBO framework is particularly applicable to distributed cloud platforms that deliver latency-critical services such as video conferencing, cloud gaming, industrial control, and real-time analytics. These application domains require dynamic, delay-aware scheduling under fluctuating energy and network conditions, making them well-suited for our adaptive, cost-efficient optimization strategy.
While AFBO demonstrates promising results, this work assumes fixed queue capacities and known renewable energy and price inputs at each time slot. It does not yet address the prediction errors, temporal uncertainty, or queue dynamics that may arise in real-time deployments. Furthermore, scalability on larger-scale, multi-provider federated platforms has not been evaluated. Additionally, our model does not yet incorporate explicit fault tolerance or resilience mechanisms. In real-world GDDC systems, failures in data centers or ISPs can critically impact service continuity. While this work ensures diversity in ISP-level routing, it did not integrate reliability-aware scheduling models to address node or provider failures, such as redundancy strategies, failover protocols, or risk-aware optimization formulations.
Future work will address these limitations by incorporating renewable energy and workload forecasting models, extending the framework to dynamic queues and uncertain environments, and deploying the AFBO method on live GDDC platforms to assess its real-time performance, adaptability, and robustness of scheduling at scale. Furthermore, scalability in large-scale or cross-provider cloud infrastructures—where multiple independent GDCs are coordinated—remains to be explored. We also plan to relax the fixed task size assumption by introducing heterogeneous task models that reflect practical workload diversity, enabling more realistic scheduling policies and performance evaluation. Additionally, fault tolerance mechanisms—such as provider failover handling and redundancy-aware scheduling—will be considered to improve system reliability under provider or node failures.