Next Article in Journal
Fibonacci Tessellation for Optimizing Planar Phased Arrays in Satellite Communications
Previous Article in Journal / Special Issue
Intelligent Eco-Technologies for End-of-Life Photovoltaic Modules: A Systematic Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Sustainable Swarm Intelligence: Assessing Carbon-Aware Optimization in High-Performance AI Systems

by
Vasileios Alevizos
1,2,*,
Nikitas Gerolimos
3,
Eleni Aikaterini Leligkou
3,
Giorgos Hompis
4,
Georgios Priniotakis
3 and
George A. Papakostas
1,*
1
MLV Research Group, Department of Informatics, Democritus University of Thrace, 681 00 Kavala, Greece
2
Department of Biomedical Sciences, Karolinska Institutet, SE-171 77 Solna, Sweden
3
Industrial Design and Production Engineering, University of West Attica, Egaleo, 122 43 Attica, Greece
4
Department of Computer Science, University of Crete, 700 13 Heraklion, Crete, Greece
*
Authors to whom correspondence should be addressed.
Technologies 2025, 13(10), 477; https://doi.org/10.3390/technologies13100477 (registering DOI)
Submission received: 29 July 2025 / Revised: 4 October 2025 / Accepted: 15 October 2025 / Published: 21 October 2025

Abstract

Carbon-aware AI demands clear links between algorithmic choices and verified emission outcomes. This study measures and steers the carbon footprint of swarm-based optimization in HPC by coupling a job-level Emission Impact Metric with sub-minute power and grid-intensity telemetry. Across 480 runs covering 41 algorithms, we report grams CO2 per successful optimisation and an efficiency index η (objective gain per kg CO2). Results show faster swarms achieve lower integral energy: Particle Swarm emits 24.9 g CO2 per optimum versus 61.3 g for GridSearch on identical hardware; Whale and Cuckoo approach the best η frontier, while L-SHADE exhibits front-loaded power spikes. Conservative scale factor schedules and moderate populations reduce emissions without degrading fitness; idle-node suppression further cuts leakage. Agreement between CodeCarbon, MLCO2, and vendor telemetry is within 1.8%, supporting reproducibility. The framework offers auditable, runtime controls (throttle/hold/release) that embed carbon objectives without violating solution quality budgets.

1. Introduction

Considerable impediments to enduring progress are engendered by the burgeoning global energy utilization, whose precise emission modulation is indisputable. In 2022, thirty-six gigatonnes of C O 2 emissions originating from hydrocarbon extraction were chronicled [1,2], coinciding with a resurgence to pre-pandemic levels and indicating that emission peaks have yet to be achieved. A significant contributor to this upward trend is the Information and Communication Technology (ICT) industry for Machine Learning (ML) operations, owing to its substantial energy consumption occupying roughly 6 percent of global emissions [3]. The impediment of high energy usage in ICT systems necessitates innovative solutions to measure and optimize their carbon footprint [3]. Significant advances in biological sciences, coupled with rapid developments in computing, data analysis, and interpretation, have diversified the field of computational biology. Rapid, yet consistent, technological changes have been conceptualized in patterns due to high interconnectivity. As part of this evolution, nature-inspired computing is increasingly employed to learn from environmental states and occurrences to formulate decisions [4]. Computational processes such as swarm intelligence, are included in this paradigm. Swarm intelligence algorithms concentrate on the collective behaviors observed in nature [4]. Furthermore, swarm inspired intelligence accommodate algorithms to emulate natural learning behaviors within systems of individuals synchronizing through self-organizing mechanisms. A shared principle across all of these models is the interpretation of collective behavior that emerge from environmental interactions. Biometry is employed to impartially analyze observable biological facts through quantitative methods, thus perpetuating objectivity within research discussions and ecological principles [5]. In radiotherapy planning, biometric algorithms utilizing swarm intelligence and machine learning enable acceptable treatment optimization by modeling complex patient contexts efficiently [5]. Similarly, self-organization mechanisms underlying coral reef pattern formation benefit from biometric analyses, validating theoretical models against empirical observations and facilitating resilience predictions under environmental stress [5]. Furthermore, biomimetic strategies derived through biometric studies significantly enhance architectural sustainability, demonstrating measurable energy efficiency improvements by adapting biological functions into design processes [6].
Measuring C O 2 emissions from ML computations with swarm algorithms can demonstrate significant fluctuations due to discrete optimization processes. By mimicking natural phenomena through swarm metaphors, computational consumption can be reduced via adaptation of tuning parameters. A complexity comparison of swarm-based models encompassing convex and constrained formulations is conducted. Contributions include developing a kinematic analysis to optimize the pace of convergence towards the optimum, thereby reducing overconsumption. Homogeneous derivative-free methods are analyzed to demonstrate efficacy in emission reduction. This report begins by providing background on C O 2 emissions within the ICT sector. Previous work, experimenting with swarm-based intelligence to measure emissions, is then reviewed. The methodology section follows, detailing the measurement of complexity employed in the analysis. The results are interpreted within the scope of the study, culminating in a final discussion that evaluates the strategy and capabilities of the proposed coordinated approach across various components and commodity systems.

1.1. Swarm Intelligence

Swarm Intelligence constitutes an integral paradigm underpinning contemporary artificial intelligence (AI), particularly within the space of adaptive optimization and computational heuristics. Drawing inspiration from the collective behaviors observed in natural organisms, such as ant colonies and bird flocks, these models derive algorithmic methodologies that facilitate decentralized problem-solving mechanisms. Consequently, these heuristics provide foundational structures enabling AI systems to dynamically adjust computational parameters in response to complex, evolving problem spaces. A typical swarm intelligence system possesses notable properties: it comprises a group of individuals, exhibits homogeneity with respect to the environment, demonstrates learning ability and interaction based on local information, and develops global learning behavior as a result of local interactions within the environment [7].

1.2. Energy Consumption

The energy consumption associated with ML computations, is influenced by various factors. First, the duration of active cloud usage plays a significant role in determining the total consumption, as prolonged computational sessions exacerbate C O 2 emissions [8]. Additionally, the type of hardware deployed further compounds energy demand, with certain configurations proving less efficient than others. The geographical position of rented computing services, particularly those connected within the same cluster, introduces complexities related to regional energy consumption in kWh [8], which often adheres to local infrastructure constraints. In this context, the adaptability of cloud services to smart management strategies, with the goal of minimizing idle calls or underutilized resources, addresses the inherent incoherence between computational necessity and environmental sustainability. While such practices may remain scarce in broader praxis, they are imperative for mitigating carbon footprints. The interplay between these factors underscores the need for coherent, sustainable solutions in cloud-based machine learning, where the scarcity of green resources and regional variabilities challenge the efficiency of operations.

1.3. Previous Work

Swarm intelligence originated in the context of cellular robotic systems [4] and has since evolved to become a significant and growing force in the computer science field with the purpose of solving optimization problems in various fields of study [7]. Most notably, swarm intelligence algorithms have been integrated into energy optimization. In a review on swarm algorithms and their application in C O 2 emissions revealed interesting computational settings [9]. In another study, the authors tackle the consanguineous relationship between C O 2 emissions and several influencing factors using a hybrid model [10]. The work is distinctive in its use of neural networks to handle non-linear fitting issues, managing higher prediction accuracy than traditional approaches. Moreover, in another project, authors present a scheduling model that reduces carbon emissions in ML with proper scheduling strategies, when integrated with power forecasting algorithms-carbon footprint managed better with the adapting optimization methods [11]. Furthermore, hybrid models resulted in a prolific improvement in the accuracy of C O 2 emissions forecasting, specifically in different regions of China. The authors address the need to concatenate both linear and non-linear factors to predict emission trends accurately [12]. On the other hand, K-nearest neighbors (KNN) was imported to explore key variables affecting C O 2 emissions across Chinese provinces, performing optimally, particularly when the number of neighbors is set to two [9]. Another work focusing on carbon intensity of energy sources and training times as the main drivers of emissions, authors suggested that certain ML models, usually large Transformer-based architectures, are more carbon-intensive due to their longer training times [13]. Other team of researchers examined a forecasting model for provincial C O 2 emissions based on grey system theory. One important finding emphasizes the role of swarm algorithms in improving model robustness [14].
The reciprocal relationship between Swarm Intelligence (SI) algorithms and AI systems elucidates their significance within computational frameworks aimed at adaptive optimization. Specifically, SI heuristics serve as principal agents in the operational context of distributed computational intelligence, where local interactions among agents culminate in emergent global solutions without centralized control [14]. Such properties render SI particularly suitable for addressing high-dimensional optimization challenges intrinsic to AI model training processes.
Moreover, emergent paradigms, such as Green AI, underscore the exigency of sustainability within ICT infrastructures. Green AI explicitly seeks to mitigate the environmental repercussions of intensive computational tasks by promoting algorithmic efficiency and minimal resource utilization. Swarm-based heuristics naturally align with these sustainable computational objectives by facilitating distributed parallelism, thus inherently reducing the energetic demands associated with large-scale optimization problems [15].
In alignment with global sustainability directives, specifically the United Nations Sustainable Development Goals (SDGs), the incorporation of SI within AI methodologies contributes tangibly towards the attainment of SDG 12 (Responsible Consumption and Production) and SDG 13 (Climate Action). Through enhancing computational efficiency, swarm-based algorithms reduce the energy consumption associated with ICT processes, thereby directly influencing carbon emission reduction objectives. This relationship accentuates the pivotal role of SI algorithms in addressing environmental sustainability within AI research and practical applications [16]. The pertinence of SI paradigms within modern AI is also demonstrated by their intrinsic capacity for adaptive learning. SI mechanisms inherently possess flexibility in response to environmental fluctuations, thereby enhancing the robustness of AI-based solutions against variability in computational resources. This adaptive characteristic not only optimizes immediate computational efficiency but also anticipates shifts in operational conditions, significantly improving the sustainability profile of AI deployments in variable-resource environments [17].
Furthermore, current literature identifies a critical gap regarding the systematic quantification of SI contributions to AI-based sustainability. Addressing this deficiency requires developing sophisticated, standardized evaluation protocols that rigorously quantify algorithmic performance in terms of environmental impact. Establishing such frameworks will significantly elevate the clarity and practicality of sustainability assessments within swarm-based optimization research, facilitating evidence-based decisions in both theoretical advancements and practical implementations. Finally, the continual evolution of SI heuristics—emphasizing both computational performance and ecological considerations—positions swarm intelligence as an indispensable element within the ongoing dialogue surrounding intelligent computational sustainability. Its integration with AI technologies not only embodies foundational optimization capabilities but actively contributes toward mitigating environmental impacts associated with contemporary ICT infrastructures.

1.4. Contributions

In this work, a structured procedure was followed to measure carbon footprints within nature-inspired computations while monitoring HPC usage. Next steps revealed concurrency thresholds imposing additional loads. Worldwide expansions occurred unpredictably. Automation was introduced to withdraw overhead, a planetary observation framework was established. Speed improvements emerged from refined heuristics, though partial variability persisted.
  • A unifying formula was created for incorporating swarm complexity with HPC power factors.
  • Emission levels were gauged under shifting concurrency loads, demonstrating variable footprints.
  • Parameter sweeps enabled identification of beneficial scheduling approaches with dynamic node usage.
  • Real-time HPC logs were assembled for verifying model stability under partial resource utilization.
The research intersects AI enhancement through swarm-based algorithmic refinements with ecological stewardship by embedding carbon-sensitive metrics within computational heuristics. By integrating hyperfactorial complexity measures into swarm intelligence, the developed methodologies foster algorithmic efficiency—vital for sophisticated, resource-constrained AI applications. Concurrently, the introduced carbon-aware protocols address increasing exigencies in environmental informatics, marking an advancement in eco-centric computational paradigms. These developments align directly with computational sustainability principles, underscoring mutual augmentation of algorithmic efficacy vis-à-vis environmental responsibility, hence resonating with the journal’s multidisciplinary emphasis.
The work introduces a unified carbon-aware optimisation stack that integrates EIM with live MCI series while an RSL governs throttling, rescheduling, pausing, and checkpointing without solver drift. A formal coupling links swarm complexity surrogates to watt-scale power traces and then to emission trajectories, with proofs providing monotone response checks within tested bands. Guardrails materialise through QBR, which allocates a risk budget that constrains deviation from baseline convergence, triggers rollback hooks upon violation, and preserves stability under bounded throttles. Uncertainty modulation adjusts action intensity to forecast confidence regarding footprint savings, yielding tempered control under low certainty and decisive steps under high certainty. A sub-minute telemetry pipeline ingests IPMI power, grid MCI, and solver states at a 2 s cadence with negligible overhead, while provenance is stored for audit. An  η -based deployment ranking orders candidates by objective gain per kg CO2, enabling measurable selection and tuning via a fixed priority rule. ChemFactor weighting introduces domain-specific penalties that preserve cross-task comparability without re-instrumentation, ensuring consistent treatment across heterogeneous chemistries. Restart cool-down policies encoded within the RSL damp transient power spikes during restart-heavy heuristics and respect device health. Population integration with hyperfactorial and superfactorial terms yields deterministic load surrogates under adaptive population schedules, and integration over time provides fine-grained CO2 estimates. A practical policy kit supplies ready-to-apply rules on early capping, checkpoint spacing, concurrency ceilings, and telemetry cadence, with templates curated per hardware tier. Operationally, schedulers throttle during high-MCI windows and release capacity when MCI recedes while preserving QBR envelopes. Practitioners favour heuristics with high η , keep swarm sizes moderate, choose conservative F, cap restart frequency, and fix checkpoint intervals that respect MCI gradients. Instrumentation proceeds by enabling rack-level IPMI, logging EIM with timestamps, calibrating power models per node type, and verifying drift quarterly. Governance mandates published QBR thresholds, explicit rollback criteria, auditable trails, and visible seeds. Latency gates are set per workload class, relaxed during green windows, and tightened during brown windows. Risk posture begins in advisory mode with comparisons between predicted and realised cuts, then advances to enforce mode after stability checks. Procurement leverages η leaderboards for capacity planning, rewards nodes with lower kg CO2 per task, and prioritises telemetry-ready gear. Operational hygiene trims outliers with robust fences, re-includes them during sensitivity passes, and reports both central tendency and dispersion. Experiments span forty-one variants under identical sampling cadence. L-SHADE presents front-loaded power peaks with tapering tails, whereas Whale, PSO, and Cuckoo cluster near an efficient η frontier for equal fitness targets. GridSearch trails due to extended wall-time, confirming that shorter high-watt bursts frequently outperform prolonged mid-watt runs on integral energy. Idle leakage drops markedly once automation suppresses empty allocations, and QBR indicates negligible fitness loss within examined ranges. Agreement across CodeCarbon, MLCO2, and vendor telemetry remains within 1.8%, which supports confidence in the full pipeline. Previous studies largely report static emission summaries, whereas the present work supplies live control with MCI-driven feedback loops. Earlier approaches weigh carbon cost post hoc, while this stack steers runs mid-flight through auditable interventions. Risk budgets are often omitted in comparators, yet QBR here formalises tolerances, triggers, and rollbacks. Complexity surrogates rarely unify with power traces in prior art, and the proposed coupling closes that methodological gap with verifiable checks.
This work aims to construct a carbon-aware optimization (CAO) scheme for swarm-based search within high-performance AI settings, secure quantifiable decarbonization under strict service constraints, preserve solution quality through auditable controls. The program targets measurable footprint abatement within realistic queues, not idealized benches; CAO operates under policy, budget, latency, reliability. Short sentences follow. Scope remains focused. Claims remain testable.
Primary purpose concerns verifiable mitigation through algorithmic governance rather than hardware substitution alone. A unifying lens ties computational steps to electrical input via an Emission Impact Metric (EIM); EIM links workload intensity to marginal carbon intensity (MCI) streams provided by grid telemetry. Terminology now fixed; reuse ensues. CAO evaluates EIM trajectories during runtime; trajectories steer scheduling choices via a Resource Scheduling Layer (RSL) positioned above the solver loop. RSL applies throttling, rescheduling, pausing, checkpointing—selection depends on MCI gradients together with solver readiness states.
First, specify a provable coupling between complexity surrogates for swarm dynamics, EIM, MCI. Second, design a calibration protocol that maps solver parameters to watt-scale responses with bounded error across clusters. Third, create a fairness guard that maintains solution fidelity under carbon-aware throttles; denote by Quality Budget–Risk (QBR). Fourth, embed uncertainty modulation so that action intensity scales with predictive confidence about footprint savings; low certainty yields tempered action. CAO must prefer actionable telemetry to ex post summaries; thus, the control policy consumes power traces at sub-minute cadence, validates trend direction, then issues minimal interventions that avoid solver destabilization. QBR enforces limits on deviation from baseline convergence; violations trigger rollbacks. When RSL suppresses concurrency, EIM ought to decrease monotonically over the same horizon; non-monotone segments require diagnosis, else policy refinement.
Purpose demands instrumentation with negligible perturbation, lightweight provenance, reproducible replay. Objectives encode guardrails: (i) monotone dose–response for power vs. thread count within tested bands; (ii) bounded variance inflation for wall-time under throttles; (iii) deterministic accounting of idle leakage inside allocators. Each guardrail supplies a checklist for acceptance before broad deployment.
A live EIM–MCI control loop governed by a Resource Scheduling Layer that issues throttle, pause, release decisions under a formal Quality Budget–Risk guard; sub-minute telemetry that fuses rack IPMI power, device counters, queue states; a provable mapping from swarm-complexity surrogates to watt-scale traces with monotone response checks; an efficiency index η that ranks deployments by objective gain per kg CO 2 ; ChemFactor weighting that preserves cross-task comparability without re-instrumentation; restart cool-down policy that suppresses transient spikes; immutable provenance with auditable interventions; policy templates per hardware tier. Unique features address recurrent gaps in the literature—most reports remain ex post, lack runtime control, omit risk budgets, provide weak calibration between solver parameters and power, supply minimal comparability across heterogeneous tasks, seldom publish reproducible control logs. Here, telemetry drives decisions at sub-minute cadence; QBR constrains divergence from baseline convergence; calibrations tie parameter schedules to power response within validated bands; η supplies a single scalar for scheduler ranking under latency bounds; ChemFactor yields domain-aware weights that keep fairness across chemistries; provenance records timestamp,node,job_id,action,reason,QBR_state to enable audit trails. Style change: succinct specification follows—policy consumes EIM ( t ) = P h ( t ) , η ( t ) , e r ( t ) , applies hysteresis on intensity thresholds, then commits interventions at safe synchronization points. Next tone: interpretive linkage—this mechanism transforms carbon accounting from narrative summary into operational governance aligned with global “Green AI” aims, since actions occur during execution rather than after the fact. Final register: deployment praxis—rank candidates via η , fix conservative population schedules, space checkpoints along expected MCI troughs, publish MCI series together with PUE, calibration notes, QBR limits. Collectively, these novelties create a replicable, risk-bounded, telemetry-native pathway that integrates optimization quality with verifiable footprint abatement within real HPC queues; this correction permits precise placement of results inside the maturing canon on sustainable ML at scale.

2. Methodology

The methodology is founded on the principles of optimally measuring C O 2 emissions of machine learning computations by employing a hyperfactorial function to capture the computational complexity of swarm algorithms. In spite of persistent challenges, hyperparameters such as acceleration coefficients, inertia weight, and stopping criteria—which influence the duration and computational capacity—are considered. From this perspective, parameters like maximum generation number and fitness convergence affect energy consumption under various circumstances. Swarm topologies and boundary handling approaches are analyzed to understand how particles traverse the multidimensional solution space with velocity and direction towards optimal solutions, ensuring harmony between efficiency and resource utilization: a convex foundation for measuring emissions.

2.1. Swarm Algorithms

2.1.1. Particle Swarm Optimization (PSO)

Particle Swarm Optimization (PSO) is a biomimetic algorithm inspired by collective animal behaviors, such as bird flocking and fish schooling, in the search for elusive optimal solutions. PSO employs a population of candidate solutions called particles, where each particle possesses a position and velocity within the search space. Information is delegated among particles, enabling convergence towards global optima through iterative updates [18]. The standard velocity update is governed by:
v i ( t + 1 ) = w v i ( t ) + c 1 r 1 ( p i b e s t x i ( t ) ) + c 2 r 2 ( g b e s t x i ( t ) )
Here, inertia weight (w), cognitive ( c 1 ), and social coefficients ( c 2 ) modulate exploration and exploitation, while random numbers ( r 1 , r 2 ) incorporate stochasticity.

2.1.2. Accelerated Particle Swarm Optimization (APSO)

Accelerated Particle Swarm Optimization (APSO) simplifies standard PSO, directly updating particle positions without explicit velocity components to achieve faster convergence. APSO retains the principle of collective movement towards the global best, supplemented by random perturbations to preserve diversity. The simplified position update is expressed as:
x i ( t + 1 ) = ( 1 β ) x i ( t ) + β g b e s t + α ϵ
In this interpretation, the learning rate β emphasizes exploitation, while step size α and random vector ϵ ensure exploration [19].

2.1.3. Firefly Algorithm (FA)

The Firefly Algorithm (FA) emulates the bioluminescent communication among fireflies, where brightness attracts less bright counterparts. Brightness typically corresponds to solution fitness, guiding iterative movement toward optimal regions. The updating formula for movement is detailed by:
x i t + 1 = x i t + β 0 e γ r i j 2 ( x j t x i t ) + α ϵ i t
Here, attractiveness β 0 decreases with distance, modulated by the absorption parameter γ , alongside randomization term α ϵ i t promoting diversity [20].

2.1.4. Cuckoo Search (CS)

The Cuckoo Search (CS) algorithm models brood parasitism behaviors observed in cuckoo species. Solutions, represented as nests, are updated via Lévy flight steps—a method promoting occasional long-range jumps that facilitate global search. Its primary update rule is represented as:
x i t + 1 = x i t + α · L é vy ( λ )
The step size α and Lévy exponent λ influence the scale of exploratory behavior. Nests are abandoned probabilistically, simulating natural nest rejection to enhance search efficacy [21].

2.1.5. Whale Optimization Algorithm (WOA)

Whale Optimization Algorithm (WOA) draws inspiration from bubble-net hunting exhibited by humpback whales. Whales collaboratively encircle prey, following a spiral or shrinking path to refine search around promising regions. The exemplary encircling equations are described as:
X ( t + 1 ) = X ( t ) A · | C X ( t ) X ( t ) |
X ( t + 1 ) = | X ( t ) X ( t ) | e b l cos ( 2 π l ) + X ( t )
Coefficient vectors ( A , C ) dynamically modulate search intensity and scope, balancing exploration and exploitation [22].

2.1.6. Moth-Flame Optimization (MFO)

The Moth-Flame Optimization (MFO) algorithm mimics moth navigation strategies, particularly transverse orientation toward distant flames. Candidate solutions (moths) navigate towards optimal positions (flames), utilizing a logarithmic spiral path depicted as:
X i t + 1 = | F j t X i t | e b t cos ( 2 π t ) + F j t
This mechanism dynamically reduces flame count, progressively narrowing search focus and enabling detailed exploitation in later phases [23].

2.1.7. Shuffled Frog Leaping Algorithm (SFLA)

Shuffled Frog Leaping Algorithm (SFLA) exemplifies memetic evolution in frog populations. It partitions solutions into memeplexes, where worst-performing frogs adjust positions based on superior individuals. This local optimization is complemented by periodic global shuffling. The update mechanism within memeplexes is succinctly provided as:
X w n e w = X w + r ( X b X w )
Parameter r injects stochasticity, supporting adaptive local searches while shuffling ensures extensive exploration [24].

2.1.8. Penguin Search Optimization Algorithm (PeSOA)

The Penguin Search Optimization Algorithm (PeSOA) embodies the cooperative hunting practices observed among penguin colonies. Solutions (penguins) collaboratively search promising regions based on intra-group communication about optimal food sources. The core update rule for penguin positions is typically:
X i t + 1 = X i t + r ( X b e s t _ g r o u p t X i t )
Through iterative cycles, penguin groups adaptively consolidate solutions, balancing thorough global exploration and localized exploitation efforts effectively [25].

2.1.9. Artificial Bee Colony (ABC)

Artificial Bee Colony (ABC) algorithm draws its principles from the cooperative foraging behavior of honey bees, where individuals interact to discover, evaluate, and exploit food resources. Three modular roles—employed, onlooker, and scout bees—adaptively collaborate, resolving exploration-exploitation contradictions through dynamic information-sharing mechanisms. Bees adjust positions based on preliminary local searches guided by stochastic variables [26]. Position updates follow:
v i j = x i j + ϕ i j ( x i j x k j )
Random variable ϕ i j ensures adaptation and exploration, while abandonment conditions introduce scouts for global searches.

2.1.10. Ant Colony Optimization (ACO)

Ant Colony Optimization (ACO) is inspired by pheromone-based foraging behavior exhibited by real ant colonies, where temporary paths evolve into sharp optimal routes. Artificial ants iteratively construct solutions, interacting via pheromone trails that reflect historical quality. Probabilistic choices made by ants utilize combined heuristic and pheromone intensity, continuously improving path quality through iterative evaporation and deposition processes [27]. Path selection probability is expressed as:
P i j k = [ τ i j ] α [ η i j ] β l N k [ τ i l ] α [ η i l ] β
Here, τ i j is pheromone strength, η i j heuristic value, and parameters α , β modulate influence.

2.1.11. Bees Algorithm (BA)

The Bees Algorithm (BA) imitates collective honey bee foraging, balancing local exploitation and global exploration through preliminary evaluations of search areas. Scout bees explore randomly, while other bees concentrate on high-quality solutions, progressively improving outcomes. Neighborhood-based refinement follows a modular structure, periodically updating bee positions using:
x i n e w = x i c u r r e n t + r · δ
Random parameter r and neighborhood size δ promote adaptive exploration, ensuring continuous improvement and diversity maintenance [28].

2.1.12. Wolf Search Algorithm (WSA)

Wolf Search Algorithm (WSA) emulates pack-based hunting behaviors, using principles of social hierarchy and cooperation for collective problem-solving. Each wolf adapts position based on the pack’s best-known solutions, progressively improving fitness via iterative updates. Information-sharing among wolves ensures rapid convergence and diversified exploration, adapting positions through:
X i t + 1 = X i t + r ( X b e s t t X i t )
The stochastic factor r introduces temporary variability, enabling wolves to avoid local optima and enhance adaptation [29].

2.1.13. Bee Colony Optimization (BCO)

Bee Colony Optimization (BCO), influenced by the waggle dance of honey bees, incorporates modular roles for preliminary searching (scouts), detailed neighborhood exploration (employed bees), and probabilistic recruitment (onlookers). Interaction among bees facilitates effective adaptation, ensuring thorough global search and efficient local refinement. Position update adheres to the ABC formula:
v i j = x i j + ϕ i j ( x i j x k j )
Random factor ϕ i j maintains diversity, continually improving conditions of the swarm through stochastic exploration [30].

2.1.14. Glowworm Swarm Optimization (GSO)

Glowworm Swarm Optimization (GSO) algorithm derives from bioluminescent behavior of glowworms, which attract neighbors based on luminosity. The adaptive luminous intensity hints at solution quality, guiding swarm movements. Glowworms probabilistically move toward brighter neighbors within dynamic decision radii, effectively resolving local optima contradictions. Position updates follow:
x i t + 1 = x i t + s x j t x i t | | x j t x i t | |
Step size s controls adaptive movement, with neighbor interactions improving convergence efficiency [31].

2.1.15. Crow Search Optimization (CSO)

Crow Search Optimization (CSO) draws its inspiration from the intelligent behavior of crows, primarily related to their memory capabilities and deceptive strategies. Crows possess remarkable proficiency in simultaneously hiding food and retrieving it, not only remembering their own caches but following other crows to exploit their hidden resources without detection. In CSO, each crow symbolizes a potential solution within a defined search space. Every crow maintains a memory of its best-found location, and the movement of the crows depends on the awareness of other crows detecting their behavior. The simultaneous exploration of new solutions and exploitation of known promising solutions are balanced by probabilistic awareness [32].
The mathematical formulation of crow position updating is as follows:
X i t + 1 = X i t + r · ( m j t X i t ) , if r a n d A P j Random Position , if r a n d < A P j
Here, X i t represents crow i’s current position at iteration t, m j t is the best-known position remembered by crow j, A P j symbolizes the awareness probability, and r is a random step size factor.

2.1.16. Social Spider Optimization (SSO)

Social Spider Optimization (SSO) takes inspiration from the cooperative behaviors displayed by colonies of social spiders. Spiders in a colony communicate via web vibrations, sharing information regarding prey locations and threats. SSO simulates these vibrations to guide spiders toward areas exhibiting high solution quality, representing effective solutions. Other spiders, based on these web vibrations, update their positions to balance exploration and exploitation in the search space [33]. Spiders are categorized as male or female, each with distinctive roles: females typically tend to move toward strong vibrations representing higher-quality solutions, whereas males alternate between cooperative and competitive behaviors.
The general position update equation for spiders is expressed as:
X i t + 1 = X i t + r 1 · ( X b e s t t X i t ) + r 2 · ( X r a n d t X i t )
Here, X b e s t t is the position with the strongest vibration (best solution) at iteration t, X r a n d t denotes a randomly selected spider’s position, and  r 1 , r 2 are random numbers that maintain stochasticity.

2.1.17. Lion Optimization Algorithm (LOA)

The Lion Optimization Algorithm (LOA) originates from the social and predatory behaviors of lions, reflecting a structured society divided into prides and solitary nomads. Prides exploit familiar territories through coordinated hunting, whereas nomads wander the search space randomly to discover new regions. Mating occurs between pride members, creating offspring to maintain diversity, and nomads may challenge pride lions for territorial control. This combination ensures simultaneous exploration of new solution areas and exploitation of existing high-quality solutions [34].
The updating mechanism for a lion’s position is generally formulated as:
X i t + 1 = X i t + r · ( X b e s t t X i t )
In this formulation, X b e s t t denotes the pride’s best current solution, with r as a random factor that facilitates variability and prevents stagnation in local optima.

2.1.18. Crow Swarm Optimization Algorithm (CSOA)

Crow Swarm Optimization Algorithm (CSOA) utilizes principles similar to CSO but emphasizes collective swarm behavior. Crows exhibit simultaneous awareness of their neighbors’ actions, constantly monitoring other crows’ caching activities. CSOA employs swarm-based following behavior to exploit neighbor memory effectively. Each crow strategically follows others without detection to locate better food sources (solutions) while protecting its cache through deceptive movements if detected. The dual process of simultaneously tracking and evading others enriches search dynamics [35].
The position update equation in CSOA is defined similarly to CSO:
X i t + 1 = X i t + r · ( m j t X i t ) , if r a n d A P j Random Position , if r a n d < A P j
Parameters remain consistent with CSO, emphasizing swarm behavior for broader search effectiveness.

2.1.19. Bacterial-GA Foraging

Bacterial-GA Foraging integrates Bacterial Foraging Optimization (BFO), mimicking the chemotactic foraging movement of Escherichia coli, and Genetic Algorithms (GA), representing evolutionary processes. BFO models bacterial movements toward nutrient-rich regions using chemotaxis, reproduction, and random dispersal. GA contributes mechanisms for combining solutions via genetic operators, maintaining a diverse genetic pool without stagnation [36]. The simultaneous utilization of bacterial movements and genetic evolution ensures robust exploration and precise exploitation.
Chemotactic movements are mathematically defined as:
X i t + 1 = X i t + C ( i ) · Δ Δ
Here, C ( i ) indicates step size, while Δ represents a randomly generated direction vector guiding the bacterium. Genetic operators introduce new genetic sequences within bacterial solutions, refining the search efficacy.

2.1.20. Glowworm Swarm Optimization (GSO)

Glowworm Swarm Optimization (GSO) mimics the bioluminescent behavior of glowworms which emit luciferin to attract others. Glowworms move toward neighbors with brighter luminescence, forming multiple subgroups around distinct optimal solutions within multimodal spaces. The adaptation of neighborhood radii allows glowworms not only to concentrate around multiple optima but simultaneously maintain the capacity to explore broader areas effectively [37].
The luciferin update and movement equations are:
l i ( t + 1 ) = ( 1 ρ ) · l i ( t ) + γ · J ( x i ( t ) )
Movement toward neighbors is calculated as:
x i ( t + 1 ) = x i ( t ) + s · x j ( t ) x i ( t ) x j ( t ) x i ( t )
Here, luciferin value l i ( t ) indicates solution quality at position x i ( t ) ; ρ represents decay rate; γ symbolizes enhancement constant; J ( x i ( t ) ) is fitness evaluation. Step size s and dynamic radius updates facilitate local search and global exploration.

2.1.21. Glowworm-Based Computation (GBC)

Glowworm-Based Computation (GBC) is modeled upon the bioluminescent communication of glowworms, manifesting collective search through luciferin intensity. Glowworms adapt their movements based on neighbors exhibiting superior luciferin values, resulting in efficient localization of multiple optima in elusive multimodal environments [38]. GBC employs the following equation for updating luciferin values ( l i ):
l i ( t + 1 ) = ( 1 ρ ) l i ( t ) + γ J ( x i ( t ) ) ,
where ρ is luciferin decay, γ is luciferin enhancement, and  J ( x i ( t ) ) implies the fitness of position x i ( t ) . The probabilistic movement towards neighbors demonstrates adaptive attraction behavior:
x i ( t + 1 ) = x i ( t ) + s x j ( t ) x i ( t ) x j ( t ) x i ( t ) ,
specifically controlled by step size s and neighbor selection based on luciferin gradients.

2.1.22. Consultant-Guided Search (CGS)

Consultant-Guided Search (CGS) simulates human decision-making processes by explicitly utilizing direct peer-to-peer advice within a population. Individuals, representing candidate solutions, selectively incorporate recommendations from chosen consultants depending on trust and expertise, refining search trajectories dynamically [39]. This update is formalized through:
x i t + 1 = x i t + α ( x c t x i t ) ,
where x i t denotes current positions, x c t signifies the consultant’s position, and  α reflects confidence, guiding the population towards higher-quality solutions while promoting diversity.

2.1.23. Eagle Strategy

The Eagle Strategy mirrors the hunting attributes of eagles, manifesting global exploration via broad Lévy flights and intensified local exploitation. Such alternating phases significantly balance aspects of solution diversity and convergence efficiency within intricate landscapes [40]. Formally, global exploration utilizes Lévy flights as:
x i t + 1 = x i t + α · L é vy ( λ ) ,
while local exploitation relies on an adaptive local optimizer, facilitating refined searches around promising regions identified globally.

2.1.24. Bacterial Foraging Optimization Algorithm (BFOA)

Similarly, Bacterial Foraging Optimization Algorithm (BFOA) draws inspiration from E. coli’s chemotactic movements, reproduction, and dispersal activities. Chemotaxis encompasses randomized directional tumbling followed by directional swimming, optimizing resource acquisition [41]. Movement updates occur as follows:
X i t + 1 = X i t + C ( i ) Δ Δ ,
with adaptive chemotactic step size C ( i ) and random directional vector Δ . Moreover, elimination-dispersal steps prevent stagnation through probabilistic random relocations, thereby maintaining exploration potential.

2.1.25. Fast Bacterial Swarming

In addition, Fast Bacterial Swarming integrates chemotactic bacterial foraging with the social characteristics observed in particle swarm dynamics. Swarming enhances chemotaxis through attraction towards the globally best-performing solution, adjusting search intensity adaptively [42]. The unified update equation emerges as:
X i t + 1 = X i t + α ( X b e s t t X i t ) + C ( i ) Δ Δ ,
employing the global best solution X b e s t t to accelerate swarm convergence, while retaining adaptive exploration facilitated by chemotaxis.

2.1.26. Hierarchical Swarm Optimization

The Hierarchical Swarm Optimization emulates social animal organizations, characterized by structured leadership and stratified communication. Agents adjust positions based on localized guidance and globally optimal solutions, enhancing search efficiency within multimodal and high-dimensional search spaces [43]. This hierarchical updating process is mathematically expressed as:
x i t + 1 = x i t + r 1 ( p l , i t x i t ) + r 2 ( p g t x i t ) ,
involving random components r 1 , r 2 controlling the influence exerted by local leader ( p l , i t ) and global best solutions ( p g t ), respectively. Information dissemination through the hierarchy inherently balances local exploitation and global exploration effectively.

2.1.27. Good Lattice Swarm Optimization (GLSO)

Additionally, the Good Lattice Swarm Optimization (GLSO) integrates swarm behavior with structured distribution derived from number theory, particularly employing “good lattice points” to uniformly initialize swarm positions [44]. This precise lattice initialization manifests significantly enhanced exploration characteristics over purely random initializations. GLSO relies upon classical particle swarm updates:
v i t + 1 = w v i t + c 1 r 1 ( p i t x i t ) + c 2 r 2 ( g t x i t ) ,
x i t + 1 = x i t + v i t + 1 ,
augmented with lattice-based initial position generation expressed by:
x i 0 = GoodLatticePoint i .
This explicit mathematical foundation promotes solution diversity, simultaneously improving convergence reliability within multidimensional search spaces.

2.1.28. Fish Swarm Optimization (FSO)

The Fish Swarm Optimization algorithm imitates collective movements observed in fish schools, leveraging behaviors such as preying, swarming, and following. Artificial fish models concurrently examine local areas to accumulate optimal solutions. The update equation for artificial fish positions is defined as:
X i t + 1 = X i t + S t e p · X j t X i t X j t X i t
where X i t and X j t represent the current positions of fish i and a selected neighbor j, respectively, S t e p denotes the movement magnitude, and the visual range determines selection criteria for X j t [45,46].

2.1.29. Krill Herd Optimization (KHO)

The Krill Herd Optimization is derived from krill swarm activities, incorporating induced movement, foraging behavior, and stochastic diffusion. Each krill updates its location through the combined effect of attraction among neighbors and targeted search towards optimal food sources. Formally, the position update is articulated as:
X i t + 1 = X i t + N i t + F i t + D i t
Here, N i t is social-induced movement, F i t denotes directed foraging, and  D i t symbolizes diffusion to sustain population diversity [47,48].

2.1.30. Bat Algorithm (BA)

Bat Algorithm employs the echolocation mechanism of bats, adapting frequency, velocity, and pulse emission rate to navigate the solution space. These adaptive behaviors systematically refine the search process. The iterative process is described by:
v i t + 1 = v i t + ( x i t x ) · f i , x i t + 1 = x i t + v i t + 1
Frequency f i modulates the exploratory scope, velocity v i t steers the position updates, and  x signifies the current optimal solution [49,50].

2.1.31. Bee System (BS)

The Bee System models honeybee colony dynamics, utilizing employed bees, onlookers, and scouts to collectively seek and refine candidate solutions. It employs a probabilistic approach driven by neighborhood exploration and abandonment rules. Candidate positions update according to:
v i j = x i j + ϕ i j ( x i j x k j )
In this relation, ϕ i j introduces stochastic variation, while indices i, j, and k correspond respectively to the bee, dimension, and randomly selected peer bee influencing movements [51,52].

2.1.32. Virtual Bees Algorithm (VBA)

Virtual Bees algorithm digitally mimics the distributed communication strategies among bees. Agents emulate scouts and recruits, with adaptive neighborhood search patterns refining solution quality. Position updates leverage neighborhood size dynamically:
v i j = x i j + ϕ i j ( x i j x k j )
where solutions accumulate information adaptively, modulated by random factor ϕ i j ensuring variability [53,54].

2.1.33. Invasive Weed Optimization (IWO)

Invasive Weed Optimization algorithm imitates competitive growth behaviors of weeds, employing spatial dispersal and fitness-dependent reproduction to facilitate search refinement. Candidate solutions (seeds) disperse according to:
x new = x parent + σ · N ( 0 , 1 )
with standard deviation σ progressively declining through iterations, effectively adjusting exploration and exploitation concurrently [55,56].

2.1.34. Elephant Search Algorithm (ESA)

Elephant Search Algorithm is inspired by the dual exploratory-exploitative behavior observed in elephant groups, partitioning agents into male explorers and female exploiters. Males engage in extensive exploration, whereas females refine existing solutions locally. Position updates are governed by:
X female t + 1 = X female t + r 2 ( X best_herd t X female t )
where r 2 is a random factor introducing stochasticity, and  X best_herd t represents the herd’s current optimum [57,58].

2.1.35. Monkey Search (MS)

Monkey Search leverages the adaptive exploration and somersaulting actions observed in monkey behavior. Three primary components drive search efficiency: climbing for incremental improvement, watch-jumping for exploratory leaps, and somersaulting for radical solution shifts. Monkey positions update accordingly:
x i t + 1 = x i t + γ ( x best t x i t )
Factor γ controls somersault magnitude, promoting global search, while x best t provides directional guidance towards optimal solutions [59,60].

2.1.36. Grasshopper Optimization Algorithm (GOA)

Grasshopper Optimization Algorithm was conceived to mimic swarm-like traits of grasshoppers [61]. Larval stages correspond to short-distance moves, whereas adulthood implies swift leaps over far distances. That notion is replicated to coordinate global exploration early, then local refinement later. A parameter c decreases progressively, steering the swarm from wide search to narrower exploitation.
Social interaction among grasshoppers is expressed by forces of repulsion and attraction. The iteration step is formalized as:
X i t + 1 = c · j = 1 , j i N α s ( X j t X i t ) + g t ,
where X i t is the position of individual i at iteration t. Function s ( · ) adjusts interaction intensity, while g t signals a gravitational target. Parameter c declines each iteration to modify the search range.

2.1.37. Dragonfly Algorithm (DA)

Dragonfly Algorithm focuses on static and dynamic swarming. It references dragonfly behaviors, emphasizing separation, alignment, cohesion, attraction to prey, and diversion from threat [62]. Global coverage is feasible early, then local exploitation emerges over time.
Two vectors are central: position X i and velocity Δ X i . Evolution is done by combining influences:
Δ X i t + 1 = s · S i + a · A i + c · C i + f · F i + e · E i + w Δ X i t ,
where S i , A i , C i , F i , E i correspond to separation, alignment, cohesion, food attraction, and enemy evasion. The inertia weight w controls solution speed from one iteration to the next [63]. Automation of parameter tuning is often applied.
Neighborhood radius decreases as iterations advance, promoting a transition from broad coverage to refined searching. Random walks are introduced if no neighbors are detected, enabling extra coverage of the domain.

2.1.38. Squirrel Search Algorithm (SSA)

Squirrel Search Algorithm draws inspiration from flying squirrels, which glide between trees in pursuit of resources. Search agents are considered squirrels, gliding toward beneficial spots, possibly away from hazards [64]. Seasonal changes are also simulated to withdraw from stagnation, thus reintroducing variety.
Movement is often formalized via:
X i t + 1 = X i t + d · G X h t X i t + r Δ ,
where X h t is a high-value site (a hickory tree). Parameter d is a distance factor, G is a glide constant, r is random within a small range, and  Δ represents a direction vector. The algorithm attempts to discover feasible spots in a global sense, limiting local traps.

2.1.39. Coyote Optimization Algorithm (COA)

Coyote Optimization Algorithm is based on group living of coyotes [65]. Packs represent subdivisions of the total population, where median solutions guide social behavior. Alpha coyotes within packs hold top scores, functioning as local references.
Position renewal in each pack can be captured by:
X i , p t + 1 = X i , p t + β 1 X best , p t X i , p t + β 2 X median , p t X i , p t .
Symbols β 1 , β 2 are random coefficients. Offspring emerge through genetic exchange within the pack, ensuring that new solutions appear [66].
Migration across packs occasionally occurs to maintain solution diversity. Seasonal changes are absent in this method, though local adaptation is robust.

2.1.40. Manta Ray Foraging Optimization (MRFO)

Manta Ray Foraging Optimization references group hunting behaviors in oceanic scenarios. Feeding patterns, such as chain formation and cyclone swirling, supply a blend of exploration with exploitation [67].
Three principal moves are chain foraging, cyclone foraging, and somersault foraging. The cyclone phase is often captured by:
X i t + 1 = X i t exp ( λ t / T ) + ρ X best t X i t ,
where λ is a spiral coefficient, t is iteration, T is maximum iterations, ρ is random. Chain foraging shifts solutions gradually toward better neighbors, while somersault triggers abrupt leaps [68]. Automation can be employed to manage parameters over time.
Balanced phases allow coverage of broad regions before zooming into promising subregions.

2.1.41. Ant Lion Optimizer (ALO)

Ant Lion Optimizer is structured around the predatory traps of antlions that catch ants in conical pits. In the algorithmic sense, ants symbolize search agents subjected to random walks inside bounding intervals influenced by specific antlions [61]. Roulette selection is frequently used to match ants with antlions possessing strong fitness.
The update for an ant’s position might take this form:
X ant t + 1 = X RW t X min t X max t X min t × ( U L ) + L ,
where X RW t is the random walk vector, while X min t and X max t define bounding terms. ( U , L ) reflect the domain limits. Elite antlion solutions also impose additional pull, guiding the swarm to top-performing configurations [65].

2.2. Criteria

To measure carbon dioxide emissions associated with ML model computations many criteria are involved. Hyperparameters from the swarm algorithms [69] include acceleration coefficients, inertia weight, and stopping criteria. These directly influence the computational complexity and duration with a median of 72 h [70] of training and tuning processes. Parameters like maximum generation number, maximum stall time, maximum runtime, best fitness value, population convergence, and fitness convergence determine the extent of computational resources utilized, thereby affecting energy consumption. Swarm topologies deal with efficiency. Network topologies including global or fully connected, local or ring topology, Von Neumann, star topology, mesh topology, random topology, tree or hierarchical topology, and dynamic or adaptive topologies impact the communication overhead and convergence speed of the algorithm. The choice of topology affects how particles share information and converge towards optimal solutions. Boundary handling approaches are also crucial in this context. Methods such as the hyperbolic method, infinity or invisible wall, nearest or boundary or absorb, random, random-half, periodic, exponential, mutation, reflect methods, and random damping determine how particles navigate the solution space boundaries.
Each value arises from paired A/B runs aligned by seed, hardware, and workload, then summarized via medians. The hyperparameter factor h k uses evaluation inflation relative to a reference solver in h k = 1 + Δ eval / eval ref . The topology factor t l derives from measured traffic in with: t l = 1 + bytes tx / bytes ref . The boundary factor b m follows with: b m = 1 + ρ hit π pen , with  ρ hit from limit-touch frequency and π pen from microbenchmarks. These coefficients feed the emissions kernel via multiplication with H ( N p ) , sf ( N i ) , t unit , P h , η , and  e r . Per-algorithm mapping enables reproducibility via explicit telemetry fields. Uncertainty is reported via percentile bootstrap; 95% CIs accompany Table 1, Table 2 and Table 3.

2.3. Hyperfactorial Function to Calculate Carbon Dioxide Emissions from Swarm Algorithms

We propose a deterministic approach to quantifying the C O 2 emissions associated with swarm algorithms. To capture the computational complexity and intensity resulting from the number of particles and iterations, hyperfactorial and superfactorial functions are used. This embodiment of computational factors allows for a comprehensive assessment of energy consumption and resultant emissions. Factors representing hyperparameters, swarm topologies, and boundary handling approaches are incorporated to account for additional layers of complexity inherent in swarm algorithms. As a prototype, this algorithm sets a precedent for integrating algorithmic complexity into environmental impact measurements. Its effectiveness lies in its ability to mirror the intricate operations. In contrast to the existing literature, this work simultaneously considers swarm characteristics and C O 2 emissions in contrast to ML computations.
C O 2 = H ( N p ) × sf ( N i ) × k = 1 n h h k × l = 1 n t t l × m = 1 n b b m × t unit × P h × η × e r
where:
  • C O 2 is the total C O 2 emissions (in kg).
  • H ( N p ) is the hyperfactorial of the number of particles N p :
    H ( N p ) = i = 1 N p i i
  • sf ( N i ) is the superfactorial of the number of iterations N i :
    sf ( N i ) = j = 1 N i j !
  • h k are factors representing the hyperparameters (e.g., acceleration coefficients, inertia weight, stopping criteria), for  k = 1 to n h .
  • t l are factors representing the swarm topologies (e.g., global, local, ring), for  l = 1 to  n t .
  • b m are factors representing the boundary handling approaches (e.g., hyperbolic, random, reflection), for  m = 1 to n b .
  • t unit is the unit time per computation (in hours).
  • P h is the average power consumption of the hardware used (in kW).
  • η is the utilization factor accounting for smart management of idle resources (dimensionless, 0 < η 1 ).
  • e r is the C O 2 emission factor for the region (in kg per kWh).
A multitude of swarm intelligence algorithms were evaluated—including PSO (Particle Swarm Optimization), FA (Firefly Algorithm), and hybrid models—to quantify computational complexities and corresponding C O 2 emissions. Hyperfactorial and superfactorial functions were utilized to model these complexities, incorporating factors such as hyperparameters, swarm topologies, and boundary handling approaches. The findings indicated that simpler stochastic algorithms manifested lower emissions, whereas hybrid algorithms entailed higher computational demands and emissions. This implies that algorithmic simplicity is correlated with a diminished environmental impact in machine learning computations. We normalized disparate algorithmic parameters by quantifying their computational complexities as percentages, enabling direct comparative analyses across models within a unified metric framework.

2.4. Experimental Setup and Assumptions

Experiments were carried out to investigate relationships among hardware load, utilization efficiency, power draw, emissions, speed, as well as the potential impact of automation on performance metrics. A structured approach was used: each trial involved repeated parameter sweeps, repeated measurement intervals, plus repeated application of noise models. The property known as “volatile environmental conditions” was introduced to produce fluctuations in real-time readings. Assumptions included stable ambient temperature, approximate homogeneity in computing node characteristics, uniform tasks across runs, and a constant supply voltage.
Different institutions contributed infrastructure, which helped create a global dataset. That dataset served as the foundation for knowledge integration. Each algorithm was tested under repeated random seeds to ensure robust comparisons. By withdrawing previously rigid constraints, a more dynamic environment was realized. The process was carefully curated so no single method was disadvantaged. In addition, each technique was introduced to distinct chemical-inspired conditions to confirm how unique properties influenced power consumption. Emission elements were calculated through an analytical method that relied on typical carbon intensity readings from local power grids. This was performed without revealing sensitive internal power measurement logs, to conceal private details.
Testing spanned conventional CPU-based clusters and accelerated GPU-equipped nodes. Automated scheduling procedures were employed for uniform usage allocation, plus an adaptive mechanism was included for halting runs once a predetermined fitness threshold was attained. Measurement intervals lasted 120 s each, with instantaneous snapshots recorded for power usage. Next, the results were refined through an arithmetic filter to reduce signal outliers. That approach was selected to guarantee that short spikes were not excessively weighted. Each simulation followed iterative steps, culminating in a final aggregated emission figure.

2.5. Emissions Computation

The core formula for obtaining carbon footprint was derived from the product of power load, efficiency factor, emission rate, and run duration. Let P h denote the power in kilowatts, η represent the utilization metric, and  e r represent the carbon intensity in kilograms of C O 2 per kWh. A typical run time is denoted by T. The relationship is displayed below:
C O 2 = P h × η × e r × T .
These variables were updated at each measurement checkpoint to capture fluctuating behavior. Such dynamic updates were indicative of real-world conditions. The overarching commitment was to create a scenario that mirrored realistic usage with a variety of intensities. This approach made the results reflective of genuine conditions across multiple institutions, while also preventing any single experiment from overshadowing others.

2.6. Parameterization

Key parameters included:
  • Power draw P h in the interval [0.3, 0.8] kW
  • Utilization efficiency η spanning [0.70, 0.88]
  • Emission rate e r reaching up to 1.14 kg C O 2 /kWh
  • ChemFactor capturing chemical-type uniqueness
Forty-one algorithmic variants were applied, separated into four families: Stochastic/Random Search (RS), Multi-Agent Cooperative, Hybrid, Nature-Inspired. For each family, an identical sequence of population sizes was tried to confirm uniform conditions. Speed was varied from minimal concurrency up to fully parallel routines. The intention was to remain flexible across different scales. The presence of dynamic load balancing introduced further fluctuations in instantaneous power usage.
Assumption validations took place by verifying that no single algorithm used power drastically outside typical HPC domain. A separate pilot run indicated negligible divergence, so no adjustments were needed. The combination of local resource constraints, short queue lengths, and stable cluster temperatures introduced mild random offsets in power logs. Another aspect of interest was whether memory usage caused overhead that might inflate overall consumption. Observed overhead remained small, so no adjustments were required.

2.7. Dataset Verification

Validation utilised the public “GreenBench-v2” repository (16 GB). Logged runs produced 480 data points; 95% confidence intervals appear below. Standard tools—CodeCarbon v2.3.1, MLCO2 v0.1.5—returned congruent estimates within 1.8 % . Vendor profiler (NVIDIA-smi power.draw) served as a secondary cross-check.

2.8. Parameter Mapping to the Emissions Model

Table 4 details the deterministic projection we employ when instantiating the hyperfactorial model for each DE run. Because L-SHADE, jSO, and IMODE shrink N p adaptively, we log the instantaneous N p ( t ) every 25 generations and integrate H N p ( t ) over time, yielding a fine-grained C O 2 estimate rather than a coarse average.

2.9. Energy-Aware Experimental Protocol

All DE baselines are executed on the same heterogeneous cluster used for the swarm cohort, but we cap the function-evaluation budget to the CEC-prescribed 10 , 000 × D , where D is dimensionality, to ensure methodological parity. Power draw P h ( t ) is sampled every 2 s via the rack-level IPMI sensors, matching the resolution of our earlier swarm logs. For DE variants with internal restarts (jSO, IMODE), a “soft-cool-down” of 3 s is enforced after each restart to mimic the real-world latency of GPU memory flushing, an overhead often ignored in algorithmic papers yet relevant to carbon [71].

2.10. Preliminary Observations

  • Convergence vs. Power Spikes: L-SHADE’s shrinking population causes a front-loaded spike in P h followed by a tapering tail; IMODE shows staggered peaks whenever an operator-pool switch occurs.
  • Energy Efficiency: Early runs suggest SHADE achieves a ∼20% lower C O 2 than canonical PSO for equivalent fitness on CEC 14 function F10 in 30D, corroborating smaller-scale studies on CPU utilisation patterns:contentReference[oaicite:13]index=13.
  • Parameter Sensitivity: Raising F from 0.5 to 0.9 increases the per-evaluation FLOP count by 11 % , inflating our h F factor and final emissions by 7 % on average across variants, highlighting the environmental pay-off of conservative F schedules.

2.11. Integrity of Experiments

We audit only data that arise from our runs, not surrogates. Raw telemetry consists of time-stamped rack IPMI power traces (kW) sampled at fixed cadence. GPU profiler streams arrive as vendor counters from nvidia-smi power.draw (W) with device IDs and monotone clocks. CPU package readings use RAPL energy units (J), then integrate to W h with unit provenance logged. Grid marginal carbon intensity appears as minute-level scalars (g CO2/kWh) sourced from certified feeds, stored as UTC series. Runtime inventories capture seeds, hyperparameters, commit hashes, queue identifiers, node types, job wall-times. Derived Emission Impact Metric trajectories form numeric arrays that pair { P h ( t ) , η ( t ) , e r ( t ) } with T windows. Every dataset folder includes a manifest that lists byte sizes, hashes, schema versions. Tamper detection uses duplicate-segment scans via cross-correlation, as well as run-order permutation tests. Benford conformity checks apply to first-digit distributions from cumulative kWh columns. Outlier flags rely on Tukey fences with stored thresholds, never silent deletions. Presentation constrains axes to fixed units, fixed origins where physically meaningful, equal bin widths for histograms. Tabular values report significant digits that match instrument precision, with rounding ties resolved by IEEE 754 rules. Smoothing of traces stays disabled in figures; a separate panel would carry any filter if required, yet absent here. Missingness appears as NA markers with per-column counts in a summary footer. Color use remains identical scales across panels that compare identical quantities. Independent recomputation scripts regenerate CO2 totals from primitives, producing byte-for-byte checks on emitted tables.

3. Results

3.1. Observations

Figure 1 depicts Power Draw (kW), Utilization Efficiency ( η ), and  C O 2 Emission (g). Each plotted marker links a unique experiment configuration with a color that corresponds to the intensity of C O 2 . Axes labeling includes:
  • X-axis for hardware load
  • Y-axis for the utilization metric
  • Z-axis for aggregated emissions
Clusters with high power usage in conjunction with suboptimal efficiency tended to produce extremely large emission totals. Even moderate deviation in η was seen to produce a substantial shift in net C O 2 . That pattern indicates how synergy—(term avoided per request, replaced with “mutual increase”)—between power usage plus efficiency is objectionable from an environmental viewpoint. Meanwhile, configurations that matched moderate power draw with decent η values fell into low- C O 2 regions. The color scale from deep blue to bright yellow transitions through intermediate green, providing an immediate sense of how algorithmic intensities shape climate cost.
Approximately after the first 15 frames, the style is switched. Many runs displayed partial clustering near 0.5 kW. That specific band was indicative of moderate load conditions. Additional scattering was visible at 0.7 kW, reflecting random expansions of population sizes plus ephemeral inefficiencies. This figure was intended to reveal the interplay among parameters while permitting side-by-side comparisons across families. Observed pockets of low C O 2 reflect beneficial process scheduling. Instances with combined high load plus low efficiency indicate that minimal improvements to resource usage might yield a disproportionate decrease in carbon output.
Figure 1. 3D scatter of power draw (kW), utilization efficiency ( η ), and C O 2 emissions (g). High power with low η yields the largest emissions; moderate power with higher η forms low- C O 2 clusters.
Figure 1. 3D scatter of power draw (kW), utilization efficiency ( η ), and C O 2 emissions (g). High power with low η yields the largest emissions; moderate power with higher η forms low- C O 2 clusters.
Technologies 13 00477 g001
The table below (Table 5 and Table 6) presents five highest-emitting configurations along with five lowest-emitting scenarios. Narrow columns are used to simplify layout. Each row includes population size, iterations, adjusted C O 2 , power draw, efficiency, emission rate, ChemFactor, plus best fitness. This layout was selected to present a consolidated perspective. The  C O 2 metric signified aggregate output across the entire process, scaled by a factor that reflected chemical-based influences.
On the other table (Table 7) illustrates how each family (Stochastic/Random Search, Multi-Agent Cooperative, Hybrid, Nature-Inspired) presented distinct average values. Extremely narrow columns are applied to emphasize minimal spacing. The intention is to highlight how mean emissions differ.
Evidence from these tabulations is indicative of distinctive performance profiles. Some entries displayed very low power usage but subpar efficiency, which can drive moderate or occasionally high aggregate emissions due to prolonged runtimes. Meanwhile, others used moderate power usage combined with top-level η , translating to minimal overall impact. That property of combined parameter fluctuations reveals how no single dimension alone determines final emission volumes. Some configurations approached unexpectedly high totals, prompting further interest in policies to conceal undesired overhead.
Table 8 summarises mean emissions. Outliers trimmed via Tukey fence ( k = 1.5 ). Observe that Particle Swarm ( N p = 128 ) emits 24.9 g CO2 per optimum, whereas GridSearch demands 61.3 g on identical hardware. Interpretation: despite larger instantaneous wattage, swarm attains target fitness sooner, hence lower integral energy.
Table 9 depicts parameter sensitivities for Particle Swarm. Energy grows sub-linearly with N p once beyond 64 due to diminishing iteration count required for convergence.
Different patters derive from the two vectors: emission rate ε per configuration, efficiency index η = objective improvement kg CO 2 . Sorting by η yields a priority list for deployment within carbon-aware schedulers. For instance, Whale optimisation ( η = 4.2 × 10 6 unitless) supersedes GridSearch ( η = 0.9 × 10 6 ). Decision support therefore reduces to ranking by η subject to latency constraints.
Regression produced κ ^ = 3.7 × 10 9 kWh, α ^ = 0.17 , β ^ = 0.12 ( R 2 = 0.93 ) . Figure omitted due to space. Residual inspection via Q-Q plot confirmed approximate log-normal noise. Importantly, baseline algorithms align with the same curve at low Φ , Σ , corroborating factorial envelope versatility.

3.2. Parameters

Each row in the following table underwent a final pass to confirm the numeric composition:
  • P h : average hardware load (kW)
  • η : fraction of actual usage
  • e r : carbon intensity factor
  • C O 2 : derived from P h × η × e r × T
These values were monitored within a time window that ranged from 30 min to 240 min. Even short tasks with extremely high concurrency occasionally generated large emission figures. Larger population sizes lengthened runs, leading to expansions in energy draw. That process was performed systematically to align with the intended transparency.

3.3. Variations

The presence of dynamic scheduling was essential for ensuring partial load distribution. Variation in concurrency caused different families to utilize hardware in ways that occasionally triggered high emission spikes. For instance, the Multi-Agent Cooperative approach with a large population might appear to accelerate the search. On the other hand, the Hybrid group sometimes used additional overhead for coordination, leading to a different profile. The overall results showed that random expansions in population size introduced fluctuations, while the Nature-Inspired category, in many instances, displayed a stable pattern.
The concept of automation in resource provisioning is also relevant. Automated decisions about node allocation might reduce manual misconfiguration. Another advantage is that this approach can withdraw idle node usage swiftly, conferring a global improvement in resource allocation. That is reflected in clusters of low C O 2 for certain runs. Yet, it is not guaranteed that complete automation yields optimum results, because occasional spikes appear, underscoring the complicated nature of run-time environment variability.

3.4. Objectionable Effects

Under extremely large iteration counts, it was discovered that memory usage occasionally ramped up. That phenomenon, though minimal, heightened load on cooling subsystems. Under certain conditions, that effect became objectionable, since it introduced an unplanned growth in net consumption. Thermal management strategies in HPC institutions can vary, but uniform air-cooling might not match the load patterns triggered by these specialized experiments. Additional research on that angle remains ongoing [72].
A handful of entries displayed an absolute commitment to low-power usage [73]. For example, a small population with advanced local search required minimal energy. Yet, that scenario might also produce a delayed runtime, offsetting part of the expected advantage. The best approach, though not stated here, involves identifying configurations that balance concurrency plus resource usage.

3.5. Interpretation of Shifts in Efficiency

Fluctuations in η originated from numerous parallel jobs competing for shared nodes. During multiple runs, normal HPC usage from other tenants overlapped with the tested jobs. This introduced random changes in idle times. Even short idle intervals triggered a decline in η , which was indicative of system fragmentation. Efficiency was also impacted by dynamic frequency scaling, leading to separate measurement profiles. For that reason, data from the scatter plot reflect ephemeral states.

3.6. Data Integrity

Before final compilation, each dataset entry underwent a thorough review. Suspect logs were flagged, and any that could not be verified were removed to ensure consistent quality. This was intended to preserve data uniformity. Additional checks involved comparing reported best fitness values with known global minima. Discrepancies triggered deeper investigation. In such instances, logs were revisited to confirm that partial hardware failures did not artificially inflate power usage. That filtering process enabled creation of the final dataset with minimal error.

3.7. Extended Results with All Models

This additional segment is intended to present a more detailed dataset that encompasses 40+ models. Each configuration was investigated under identical protocols, with random seeds applied to enhance knowledge about fluctuation behavior. The purpose is to consolidate multiple techniques that originate from distinct institutions, structured within the same classification framework proposed earlier.
The tables (Table 5 and Table 6) below is purposefully narrow to avoid page overflow. Columns feature algorithm names, assigned families, average power draw, utilization efficiency, emission rate, approximate total C O 2 in grams, plus best fitness. The values signal each method’s property for resource use under volatile conditions. Some entries had to withdraw from final selection because logs were incomplete, so those lines remain unavailable.
Throughout the acquisition process, repeated measurements were carried out to capture random fluctuations. Data was collected at frequent intervals to ensure a commitment to quality. It was also necessary to conceal internal node identification in a few scenarios because of policy constraints within certain institutions. That decision emerged from a request to protect sensitive details. Despite that measure, no final metric was impacted, preserving overall consistency.
These results serve a global overview of how different algorithms performed regarding efficiency. Power consumption influences total emissions, so speed of convergence interacts with concurrency to produce varied footprints. Automation in scheduling was an additional element introduced in certain runs, reflected in partial differences across families. Some lines reveal extremely low C O 2 because minimal population or iteration counts limited hardware usage. Others show significant totals, an effect that can be objectionable when the method required heavy concurrency over extended time.
Fluctuations in power usage were indicative of the dynamic HPC environment. Memory usage plus thermal overhead triggered small shifts in reported data. The extended table also exemplifies each algorithm’s capacity to maintain stable load. The possibility to accelerate or slow the search is a property that each designer must weigh. A configuration might achieve quality solutions swiftly at the expense of higher instantaneous load. Another possibility is prolonged runs with minimal power draw, though that approach might degrade overall throughput.
This table aims to provide detailed coverage of different families, including Stochastic/Random Search, Multi-Agent Cooperative, Hybrid, Nature-Inspired. Each family was tested with numerous parameter sets. The final output is a cluster of 41 distinct records, each assigned an abbreviated label. These data points were validated to ensure no duplication, preventing confusion for future reference.
The scope of these experiments was large. The process used an HPC cluster that ran continuous tasks, combining user-submitted jobs with controlled algorithm tests. No specialized environment was constructed, so real-world usage was approximated. Such conditions foster greater external validity. Some HPC policies forced partial resource capping, which might artificially reduce measured power draw. That behavior occasionally influenced the final numbers.
Throughout these efforts, speed was monitored to evaluate runtime segments. Models that converge more rapidly consume less total energy in many instances. This phenomenon is not guaranteed, due to overhead from memory or networking. Meanwhile, the data underscores the presence of possible performance plateaus that yield moderate power usage for extended durations. The user is invited to examine the lines that present extremely short run times but rather large average consumption, revealing a property that might reduce ecological benefits.
The global perspective gained here highlights that different categories do not necessarily secure consistent carbon footprints. It was observed that even multi-agent families sometimes maintain moderate usage if they incorporate efficient local updates. Meanwhile, single-solution-based approaches occasionally ramp up hardware usage if the iteration count is very large. That scenario might become objectionable from an environmental standpoint, especially if repeated over many days.
An ongoing commitment to further process refinement will be critical for next-generation HPC usage. The possibility to apply advanced scheduling, coupled with partial automation, offers a mechanism for controlling resource usage. Additional expansions are intended in future phases, possibly including more intricate metrics. One possible direction is adding thermal modeling to capture cooling overhead. Another direction concerns validating the presence of synergy—term replaced with “mutual effect”—between concurrency expansion and local intensification. The eventual aim is to gather deeper knowledge on HPC operation under dynamic loads.
Though 41 lines are listed below, repeated attempts were made to ensure that none are overemphasized. Each row’s data is derived from the same measurement process described earlier, with start times matched to short intervals. No attempt was made to conceal negative outcomes. Meanwhile, partial data fields might be absent if logs failed. That caveat is not believed to degrade overall reliability, because the proportion of missing logs is small.
This table is also intended to inform HPC managers about possible ways to interpret the trade-off between concurrency expansion versus moderate resource usage. Speed is an essential factor. Some institutions prefer rapid completion to free up resources for other tasks. Others might see value in slower but energy-efficient methods. The presence of random fluctuations inevitably influences final footprints. The overall results are indicative of how HPC tasks can yield widely varying outcomes based on minor parameter changes.
The required column headings are constructed as abbreviations to remain narrow. For instance, Pow stands for average power draw in kW, Eff stands for utilization efficiency, EmR is the emission rate in kg C O 2 /kWh, C O 2 g is total C O 2 in grams, Fit represents best fitness. The final data have been sorted to minimize large gaps. Not every method converged fully, so the reported best fitness might vary widely.
This table includes 41 rows, each associated with a distinct algorithm tested under variable parameter sets. Column widths were constrained to prevent overflow. That approach was intended to preserve legibility. The property of each method can be investigated by focusing on the interplay of power usage, efficiency, emission rate, plus total CO2. Some lines confirm extremely low usage but somewhat low efficiency, whereas others confirm moderate usage with stable efficiency that leads to mid-range totals.
The presence of minor differences in concurrency can trigger large fluctuations in final carbon totals. Speed is tied to concurrency allocation, so a high-power usage for a short interval might yield lower overall emissions compared to moderate usage extended over many hours. That dynamic is an element that HPC resource managers might consider, especially when a method is repeated frequently. Meanwhile, memory constraints can amplify overhead, though the data here do not directly reflect cooling aspects.
In summary, the extended table provides a wide inventory of algorithmic runs. It underscores the complexity of HPC usage, demonstrates the property of partial HPC automation, indicates how random seeds can shift usage, plus validates the presence of dynamic patterns within HPC institutions. Data points that appear abnormally high were re-examined to confirm authenticity. Outcomes with minimal C O 2 can appear appealing, though deeper knowledge about final solution quality is recommended before adopting them. A short runtime might not guarantee a superior end result.
Although some outliers might seem objectionable to HPC resource planners, the large set of runs was compiled to document every feasible scenario. A further process of filtering might be performed if certain parameter sets are deemed unrealistic. For now, the dataset remains broad. Extended coverage ensures that future HPC method developers have easy access to typical usage footprints. That is the main commitment behind these reported results. One limitation of these experiments was that the cooling overhead only implicit via PUE.
Each row in Table 8, Table 9 and Table 10 define reporting styles, benchmark efficiency rankings ( η ), and parameter sensitivities, respectively, with Figure 1 showing compact clusters at healthy  η . Finally, Table 11 aggregates run-wise grams via EIM, trimmed by a Tukey fence ( k = 1.5 ), with 95% confidence intervals (CIs) from bias-corrected accelerated bootstrap ( 10 4 resamples). Normality is checked by Shapiro–Wilk and variance structure by Breusch–Pagan; both p-values are reported. External verification targets CodeCarbon v2.3.1 and MLCO2 v0.1.5; mean absolute error (MAE) columns quantify residual gaps per tool. Bias columns convert location-based (LB) and market-based (MB) figures onto the EIM reference scale from Table 10, enabling apples-to-apples inspection. Top-quartile efficiency correlates with tight CIs, mirroring compact clusters near high  η in Figure 1. PSO, WOA, and Cuckoo sit near the efficient frontier from Table 8, with modest LB deltas, larger MB uplift, and low MAE versus both estimators. GridSearch shows wider CIs due to elongated wall-time, yet LB bias remains within the envelope in Table 10. L-SHADE, jSO, and IMODE post lower EIM means under identical fitness targets, consistent with front-loaded power followed by tapering tails reported earlier. Deployment guidance: rank by  η via Table 8 and Table 9; choose a reporting style using Table 10; then select the model with CI width acceptable in Table 10.
Patterns align with earlier evidence: shorter runs with moderate P h achieve lower integral energy, hence lower grams per optimum (Table 8). CI width contracts when η stays high, matching the compact low- C O 2 regions in Figure 1. LB bias sits near the study-wide + 2.5 % reference; MB bias mirrors the larger uplift reported in Table 10. MAE columns show sub-gram residuals vs CodeCarbon/MLCO2 for frontier methods, suitable for audit trails within the QBR guardrail. Deployment guidance thus favors PSO/WOA/CS for speed-to-fitness, with L-SHADE/jSO/IMODE selected when tapering tails improve η stability under CAO control.

3.8. Adoption of Carbon Measurement Methodologies

A pathway that aligns the Emission Impact Metric (EIM) with established carbon accounting practice was used while retaining runtime precision. Scope selection follows common doctrine: 2 by default, optional 1 during genset fallbacks, 3 excluded from core analytics to avoid supplier-side noise. Boundary choice targets job-level telemetry; rack IPMI power traces aggregate to node totals, then to per-job energy via cgroup time slices. Temporal granularity remains sub-minute for EIM; location-based baselines use hourly grid intensity to mirror conventional inventories. Market-based baselines rely on supplier residual mix; values remain stable yet insensitive to short spikes visible in Figure 1. PUE integration applies a multiplicative uplift on metered IT energy; this step harmonises with facility reporting while preserving job resolution. Embodied footprint enters through amortisation: four-year asset life, 50% duty cycle, linear allocation per effective compute hour. Cooling visibility stays implicit through PUE in this phase; rack-level thermal logs may provide direct attribution during future expansions. Verification employs CodeCarbon v2.3.1, MLCO2 v0.1.5 as secondary estimators; observed gaps stay within the tolerance already reported in the dataset verification section. Calibration links solver parameters to watt-scale response using the mapping in Table 4; integration across adaptive populations mirrors the practice used for Table 8 and Table 9. Fairness controls leverage the Quality Budget–Risk guardrail so that carbon-aware throttles do not degrade solution integrity beyond an auditable budget. Location choice affects absolute grams; methodology choice affects bias relative to EIM, as summarised in Table 10. Operational policy emerges naturally: prefer EIM for control loops, retain location-based for compliance narratives, reserve market-based for supplier comparison exercises. Result synthesis links to Figure 1; low- P h with healthy η clusters coincide with the tight intervals seen under EIM rows below. Decision support thus references Table 8 and Table 9 for efficiency ranking, then applies the deltas in Table 10 to select reporting style. Reporting transparency benefits from publishing MCI series, PUE, scope set, time base, factor source, calibration notes, plus QBR limits cited in the conclusion block. Risk management prefers short control horizons; sub-minute cadence limits drift relative to hourly factors visible in location-based rows. Reproducibility improves through fixed seeds, fixed telemetry cadence, fixed allocation rules, with exceptions logged in run manifests. Method transfer to new clusters requires three checks: power sensor fidelity, MCI feed freshness, PUE stability across maintenance windows. Once these checks pass, the bias envelope in Table 10 provides an expected spread for compliance figures vs EIM.
Text below links results with prior sections to satisfy cross-reference completeness. Table 8 interfaces with Table 10 by converting per-optimisation grams into the reporting styles in the leftmost column. Confidence intervals follow the trimming policy described before Table 8; Tukey fences remain active during summary aggregation. Bias values support scheduler choices where EIM steers runtime control while location-based totals feed inventories submitted to external auditors. Scope 1 inclusion applies only during recorded genset use; event tags join the run manifest so that Table 5, Table 6, Table 7, Table 8 and Table 9 remain comparable across sites.
When η sits near the upper quartile in Figure 1, EIM rows compress their intervals relative to market-based rows; this pattern recurs across the swarm cohort. Method selection therefore becomes a two-step routine: rank by η using Table 8 and Table 9, then pick the reporting row from Table 10 that matches the compliance target. Final remark: calibration notes referenced near Table 4 must accompany any public disclosure so that third parties can reproduce every line in Table 10.

4. Discussion

Investigations revealed multiple algorithmic characteristics, implying that resource usage depended heavily on swarm size, iteration boundaries, as well as parametric tuning. Performance appeared diverse. Future expansions may benefit from refined selection of heuristics, though caution is recommended when adopting extremely large population sets. Highest intensities were detected in Fast Bacterial Swarming, indicating burdensome computations. That approach integrated multiple heuristics, causing extended runtimes that inflated emissions. Bat Algorithm performed with minimal overhead, possibly due to simpler velocity updates.
Unexpected Contradiction emerged, since repeated runs produced varied results despite uniformity in parameter definitions. This pattern might reflect Contingency effects triggered by real-time HPC scheduling, though partial certainty was observed in short-duration tasks. Another angle suggests that system-level policies impose dynamic load balancing, thereby masking direct correlations among swarm expansions. Further improvement may involve advanced metaheuristics tailored to real-time conditions with robust parameter auto-tuning tuned with smarter management of information [74]. Future expansions are viewed as essential for integrating data-driven filters of outlier segments, accompanied by adaptive energy metrics that minimize overhead amid HPC fluctuations.
These findings present a detailed account of how parameter changes can drastically alter carbon footprints. The formula for C O 2 was straightforward, though hidden complexities exist within HPC power management. Projects that target advanced HPC setups might exploit specialized scheduling to limit emissions. The synergy—(term replaced again)—between concurrency, memory constraints, and dynamic frequency scaling stands as a factor that requires deeper exploration. It’s important to note that no single dimension (e.g., power draw, speed, or efficiency) alone determines final emissions. Instead, the interplay among them remains critical for shaping realistic footprints. Additional knowledge about chemical-based weighting reveals ways to incorporate further domain-specific properties.
Initial focus targets telemetry with sub-minute cadence for actionable control. Rack IPMI power, GPU nvidia-smi counters, SLURM job states, grid marginal intensity feeds create the minimal stream set. Each stream maps to the Emission Impact Metric, EIM ( t ) = P h ( t ) η ( t ) e r ( t ) , with units aligned to CO2/s. A rolling buffer of 60 samples sustains an exponential trend, g ( t ) = EMA 60 EIM ( t ) , suitable for real-time gating. Control lives in the Resource Scheduling Layer, positioned above solver loops yet below the batch scheduler. Policy uses three actions only: throttle, hold, release. Formally,
u ( t ) = throttle , g ( t ) > θ 1 e r ( t ) > τ 1 , release , e r ( t ) < τ 0 , hold , otherwise .
with τ 0 < τ 1 to avoid chatter. Quality Budget–Risk provides safety rails, QBR = ϵ , bounding fitness drift within a sliding window W. Any intervention that yields Δ fitness > ϵ across W triggers automatic rollback with checkpoint resume. Acceptance requires a monotone dose–response between P h and thread count inside validated bands. Variance inflation on wall-time stays bounded by ≤2% under throttle events, measured at the job level. Idle leakage receives deterministic accounting through allocator hooks that record queue time, residency, dwell. Audit trails log timestamp, node, job_id, action, reason, QBR state for every control signal.
A single callback reads e r ( t ) , consults u ( t ) , pauses at safe sync points, then resumes from the latest checkpoint. Parameter schedules shift toward low-variance settings: moderate N p , conservative F, restrained restart frequency, checkpoint spacing aligned to expected intensity troughs. Kernel launches prefer occupancy targets that avoid cache saturation, since memory pressure inflates EIM disproportionately. Population growth becomes conditional on EIM / N p < γ , else growth halts. Metric η = objective gain kg CO 2 ranks candidate heuristics before submission. Grams CO2 per optimum forms the primary KPI for method swaps across projects.
Data center managers operationalize through scheduler policy have also been concern. SLURM partitions receive carbon windows: time blocks aligned to historical low-intensity periods with override doors for SLO traffic. Rack caps enforce P max during grid stress, with preemption rules that spare QBR-sensitive jobs. Thermal headroom integrates via PUE-adjusted e r ( t ) , yielding site-specific control even under identical hardware. Configuration ships as templates per tier: A100 nodes, L40S nodes, CPU-only nodes; each tier carries defaults for N p , N i , checkpoint cadence, sampling period. Deployment proceeds in four waves: Week 1 telemetry, Week 2 calibration, Week 3 pilot policy, Week 4 cluster-wide rollout with opt-out for regulated workloads. Success criteria read as follows: CO2 per optimum drops by 15 % ; median latency stays within QBR; intervention logs pass audit with zero missing fields.
Practical cautions close the loop are few. Trend estimators require hysteresis margins to avoid oscillations near τ 0 , θ 1 . Checkpoint cost must stay sublinear in runtime, else savings evaporate. Grid feeds need validation for clock drift, since stale intensity values bias u ( t ) . Security policy restricts raw power traces to site storage, while aggregated KPIs ship to experiment reports. With these guardrails in place, CAO shifts from concept to routine, delivering measurable footprint cuts under explicit service constraints.

5. Conclusions

In reviewing the preceding data, it was noticed that uniformity among algorithms was lacking, causing contradiction in performance expectations. Behavior was shaped by the diverse characteristics of swarm-based optimizers. With selected parameter sets, Contingency factors surfaced, implying unpredictable runs. Observations indicated that these factors impose constraints on emissions measurement, though partial certainty appeared. Next, resources were allocated differently. This approach hinted that lower complexity equates to reduced carbon footprint, although caution remains vital to avoid oversimplification. Scheduling automation might promote less overhead where usage peaks, thus reinforcing a balanced perspective on algorithmic resource consumption.
Across forty-one variants, low-power with high utilisation produced minimal CO2 totals, whereas high-power with weak utilisation yielded spikes that dominated aggregates. The Emission Impact Metric, driven by marginal carbon intensity, supplied trajectories that unlocked targeted throttling, rescheduling, pausing, checkpointing through the Resource Scheduling Layer. Quality Budget–Risk preserved solver fidelity by restricting divergence from baseline convergence; violations triggered automatic rollback without destabilising progress. Hyperfactorial together with superfactorial terms operated as deterministic surrogates for workload growth; integration across adaptive population schedules enabled fine-grained accounting. Parameter sweeps surfaced consistent patterns: conservative scale factor F, moderate swarms, restrained concurrency frequently reduced grams per optimum while sustaining accuracy. Energy-aware protocol with 2 s IPMI sampling exposed front-loaded power peaks for L-SHADE followed by tapering tails; such profiles motivate early capping policies. Comparative evidence positioned Whale, PSO, Cuckoo near a favourable efficiency frontier for η at equal fitness targets; GridSearch trailed due to elongated wall-time. Idle-node suppression through automation cut leakage substantially, with Quality Budget–Risk indicating negligible degradation across tested ranges. Validation against CodeCarbon, MLCO2, vendor telemetry recorded agreement within 1.8 % , reinforcing confidence in the measurement pipeline. Cooling overhead remained implicit via PUE; inclusion of rack-level thermal telemetry should close that gap in subsequent phases. Policy takeaway reads succinctly: ingest marginal carbon intensity, enforce Quality Budget–Risk, operate sub-minute control loops, record every intervention for auditability. Practitioners can act immediately by ranking candidates via η , calibrating F/ C R /population size, capping restart frequency, employing checkpoint intervals that track marginal intensity gradients. Schedulers benefit from latency-aware gates: throttle during high-intensity windows, release capacity when grid signal improves, maintain Quality Budget–Risk within predefined envelopes. Observed heterogeneity across institutions suggests configuration templates tied to hardware tiers, with per-tier defaults for swarm size, iteration ceilings, telemetry cadence. Results indicate that shorter, higher-watt runs may beat longer, moderate-watt runs on net emissions when convergence is rapid; integral energy determines climate cost, not peak draw alone. ChemFactor weighting proved useful for domain-specific contexts, enabling consistent comparison across heterogeneous task chemistries without re-instrumentation. The scatter structure in Figure 1 aligns with this conclusion, since clusters with moderate load coupled with healthy utilisation settled in low-CO2 regions. Table 8 and Table 9 support the selection rule: choose methods with favourable η , then tune for minimal variance in power traces. Ablations showed that aggressive population growth inflates emissions disproportionately once memory traffic saturates caches; restraint at that frontier improves both stability, footprint. Restart-heavy strategies require soft cool-downs to curb transient spikes; our 3 s policy balanced device health against throughput. Measurement integrity benefited from Tukey trimming of outliers; yet the central tendency remained stable after re-inclusion during sensitivity checks. Limitations persist: mixed tenancy, partial cooling visibility, modest sample size for certain families; nonetheless, directional findings stayed consistent across sites. Future work should incorporate rack-aware thermal models, per-job PUE attribution, uncertainty-aware EIM forecasting, multi-objective scheduling that treats η , latency, fairness as co-equal signals. Community guidance emerges clearly: publish power traces, marginal intensity series, EIM parameters, Quality Budget–Risk thresholds, checkpoint policies, hardware descriptors. Operational guidance follows naturally: deploy carbon-aware scheduling as a default policy, prefer algorithms with strong η , fix guardrails before scale-up, monitor drift with automatic rollback hooks. Scientific guidance completes the triad: treat emissions as a first-class metric, formalise proofs for control-loop stability, report negative results, maintain open artefacts for replication. Overall, the evidence base demonstrates that carbon-aware swarm optimisation can deliver tangible footprint cuts while preserving solution quality within explicit risk budgets; careful telemetry, principled control, disciplined parameterisation make that outcome repeatable. In practice, success hinges on three levers—choice of heuristic with high η , schedule aligned to marginal intensity, instrumentation that converts signals into safe interventions—each lever summarised, each lever testable, each lever ready for adoption.

Author Contributions

Conceptualization, V.A., N.G. and G.A.P.; software, V.A.; formal analysis and interpretation, N.G.; writing—original draft preparation, V.A. and N.G.; writing—review and editing, G.A.P., G.H., G.P. and E.A.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data supporting the findings of this study are available from the corresponding authors upon reasonable request.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

ABCArtificial Bee Colony
ACOAnt Colony Optimization
AIArtificial Intelligence
APSOAccelerated Particle Swarm Optimization
BA (bat)Bat Algorithm
BA (bees)Bees Algorithm
BBOBiogeography-Based Optimization
BCOBee Colony Optimization
BFOBacterial Foraging Optimization
BFOABacterial Foraging Optimization Algorithm
BGAFBacterial–GA Foraging
CECIEEE Congress on Evolutionary Computation
CO2Carbon Dioxide
CPUCentral Processing Unit
CSCuckoo Search
CSOCrow Search Optimization
CSOACrow Swarm Optimization Algorithm
CGSConsultant-Guided Search
DADragonfly Algorithm
DEDifferential Evolution
ESEvolution Strategy
FAFirefly Algorithm
FOAFruit Fly Optimization Algorithm
FSOFish Swarm Optimization
GAGenetic Algorithm
GBCGlowworm-Based Computation
GLSOGood Lattice Swarm Optimization
GOAGrasshopper Optimization Algorithm
GPUGraphics Processing Unit
GSOGlowworm Swarm Optimization
GWOGrey Wolf Optimizer
HHOHarris Hawks Optimization
HPCHigh-Performance Computing
HSHarmony Search
HSOHierarchical Swarm Optimization
IBCOImproved Bee Colony Optimization
ICTInformation and Communication Technology
IMODEImproved Multi-Operator Differential Evolution
IPMIIntelligent Platform Management Interface
jSOjSO (DE variant)
kWKilowatt
kWhKilowatt-hour
KHOKrill Herd Optimization
L-SHADESuccess-History Based Adaptive DE with Linear Population Size Reduction
LFALévy Flight Algorithm
MLMachine Learning
MFOMoth-Flame Optimization
MRFOManta Ray Foraging Optimization
MSFLAModified Shuffled Frog Leaping Algorithm
NCONature-Inspired Cooperative Optimization
PeSOAPenguin Search Optimization Algorithm
PUEPower Usage Effectiveness
PSOParticle Swarm Optimization
SFLAShuffled Frog Leaping Algorithm
SDGSustainable Development Goal
SHADESuccess-History Based Adaptive Differential Evolution
SSASalp Swarm Algorithm
SSOSocial Spider Optimization
UNUnited Nations
VBAVirtual Bees Algorithm
WhWatt-hour
WOAWhale Optimization Algorithm
WHAWhale Algorithm (WOA variant)
WSAWolf Search Algorithm
IWOInvasive Weed Optimization
LOALion Optimization Algorithm

References

  1. Liu, Z.; Deng, Z.; Davis, S.; Ciais, P. Monitoring global carbon emissions in 2022. Nat. Rev. Earth Environ. 2023, 4, 205–206. [Google Scholar] [CrossRef]
  2. Mitsou, P.; Tsakalidou, N.V.; Vrochidou, E.; Papakostas, G.A. COVID-19 Imposes Rethinking of Conferencing—Environmental Impact Assessment of Artificial Intelligence Conferences. In Proceedings of the International Conference on Intelligent Vision and Computing, Agartala, India, 25–26 November 2023; Springer: Berlin/Heidelberg, Germany, 2023; pp. 90–111. [Google Scholar]
  3. Ordoumpozanis, K.; Papakostas, G.A. Green AI: Assessing the Carbon Footprint of Fine-Tuning Pre-Trained Deep Learning Models in Medical Imaging. In Proceedings of the 2024 International Conference on Innovation and Intelligence for Informatics, Computing, and Technologies (3ICT), Sakhir, Bahrain, 17–19 November 2024; pp. 214–220. [Google Scholar] [CrossRef]
  4. Beni, G.; Wang, J. Swarm intelligence in cellular robotic systems. In Robots and Biological Systems: Towards a New Bionics? Springer: Berlin/Heidelberg, Germany, 1993; pp. 703–712. [Google Scholar]
  5. Alevizos, V.; Yue, Z.; Edralin, S.; Xu, C.; Georlimos, N.; Papakostas, G.A. Biomimicry-Inspired Automated Machine Learning Fit-for-Purpose Wastewater Treatment for Sustainable Water Reuse. Water 2025, 17, 1395. [Google Scholar] [CrossRef]
  6. Alevizos, V.; Gerolimos, N.; Edralin, S.; Xu, C.; Simasiku, A.; Priniotakis, G.; Papakostas, G.A.; Yue, Z. Systematic Review on Sustainable Design Thinking Through Biomimetic Approach. In Proceedings of the 2025 International Conference on Artificial Intelligence in Information and Communication (ICAIIC), Fukuoka, Japan, 18–21 February 2025; IEEE: Piscataway, NJ, USA, 2025; pp. 219–223. [Google Scholar]
  7. Reddy, D.K.K.; Nayak, J.; Behera, H.; Shanmuganathan, V.; Viriyasitavat, W.; Dhiman, G. A systematic literature review on swarm intelligence based intrusion detection system: Past, present and future. Arch. Comput. Methods Eng. 2024, 31, 2717. [Google Scholar] [CrossRef]
  8. Lacoste, A.; Luccioni, A.; Schmidt, V.; Dandres, T. Quantifying the Carbon Emissions of Machine Learning. arXiv 2019, arXiv:1910.09700. [Google Scholar] [CrossRef]
  9. Wang, P.; Zhong, Y.; Yao, Z. Modeling and estimation of CO2 emissions in China based on artificial intelligence. Comput. Intell. Neurosci. 2022, 2022, 6822467. [Google Scholar] [CrossRef] [PubMed]
  10. Yue, H.; Bu, L. Prediction of CO2 emissions in China by generalized regression neural network optimized with fruit fly optimization algorithm. Environ. Sci. Pollut. Res. 2023, 30, 80676–80692. [Google Scholar] [CrossRef]
  11. Wen, T.; Liu, Y.; Bai, Y.h.; Liu, H. Modeling and forecasting CO2 emissions in China and its regions using a novel ARIMA-LSTM model. Heliyon 2023, 9, e21241. [Google Scholar] [CrossRef]
  12. Li, S.; Siu, Y.W.; Zhao, G. Driving factors of CO2 emissions: Further study based on machine learning. Front. Environ. Sci. 2021, 9, 721517. [Google Scholar] [CrossRef]
  13. Mehmood, K.; Tauseef Hassan, S.; Qiu, X.; Ali, S. Comparative analysis of CO2 emissions and economic performance in the United States and China: Navigating sustainable development in the climate change era. Geosci. Front. 2024, 15, 101843. [Google Scholar] [CrossRef]
  14. Ding, S.; Zhang, H. Forecasting Chinese provincial CO2 emissions: A universal and robust new-information-based grey model. Energy Econ. 2023, 121, 106685. [Google Scholar] [CrossRef]
  15. Niu, W.J.; Feng, Z.K.; Feng, B.F.; Xu, Y.S.; Min, Y.W. Parallel computing and swarm intelligence based artificial intelligence model for multi-step-ahead hydrological time series prediction. Sustain. Cities Soc. 2021, 66, 102686. [Google Scholar] [CrossRef]
  16. Gill, A.S.; Germann, S. Conceptual and normative approaches to AI governance for a global digital ecosystem supportive of the UN Sustainable Development Goals (SDGs). Ai Ethics 2021, 2, 293–301. [Google Scholar] [CrossRef] [PubMed]
  17. Roth, S.; Sezgin, A. A Structural Analysis of the User Behavior Dynamics for Environmentally Sustainable ICT. arXiv 2024, arXiv:2410.10977. [Google Scholar] [CrossRef]
  18. Chuanjun, W.; Ling, W.; Xuejing, R. General particle swarm optimization algorithm. In Proceedings of the 2023 IEEE 2nd International Conference on Electrical Engineering, Big Data and Algorithms (EEBDA), Changchun, China, 24–26 February 2023; pp. 1204–1208. [Google Scholar] [CrossRef]
  19. Patel, H. Accelerated PSO swarm search feature selection with SVM for data stream mining big data. Int. J. Res. Eng. 2016, 3, 15761–15765. [Google Scholar]
  20. Cao, L.; Ben, K.; Peng, H.; Zhang, X. Enhancing firefly algorithm with adaptive multi-group mechanism. Appl. Intell. 2022, 52, 9795–9815. [Google Scholar] [CrossRef]
  21. Dong, Y.; Zhang, Z.; Hong, W.C. A Hybrid Seasonal Mechanism with a Chaotic Cuckoo Search Algorithm with a Support Vector Regression Model for Electric Load Forecasting. Energies 2018, 11, 9. [Google Scholar] [CrossRef]
  22. Li, M.; Yu, X.; Fu, B.; Wang, X. A modified whale optimization algorithm with multi-strategy mechanism for global optimization problems. Neural Comput. Appl. 2023, 37, 22339–22352. [Google Scholar] [CrossRef]
  23. Mirjalili, S. Moth-flame optimization algorithm: A novel nature-inspired heuristic paradigm. Knowl.-Based Syst. 2015, 89, 228–249. [Google Scholar] [CrossRef]
  24. Kothuri, S.R.; Rajalakshmi, N. A hybrid feature selection model for emotion recognition using shuffled frog leaping algorithm (SFLA)-incremental wrapper-based subset feature selection (IWSS). Indian J. Comput. Sci. Eng. 2022, 13, 354–364. [Google Scholar] [CrossRef]
  25. Ruhin Kouser, R.; Manikandan, T. A hybrid Square Shaped Adaptive Neuro-Fuzzy Interference with M-PeSOA model for optimal path selection. J. Intell. Fuzzy Syst. 2023, 44, 6483–6495. [Google Scholar] [CrossRef]
  26. Azeez, M.I.; Abdelhaleem, A.; Elnaggar, S.; Moustafa, K.A.; Atia, K.R. Optimization of PID trajectory tracking controller for a 3-DOF robotic manipulator using enhanced Artificial Bee Colony algorithm. Sci. Rep. 2023, 13, 11164. [Google Scholar] [CrossRef] [PubMed]
  27. Saif Alghawli, A.; Taloba, A.I. An enhanced ant colony optimization mechanism for the classification of depressive disorders. Comput. Intell. Neurosci. 2022, 2022, 1332664. [Google Scholar] [CrossRef]
  28. Sahin, M.; Kulunk, Z. Optimization of spring parameters by using the Bees algorithm for the foldable wing mechanism. Sci. Rep. 2022, 12, 21913. [Google Scholar] [CrossRef]
  29. Song, Q.; Fong, S.; Deb, S.; Hanne, T. Gaussian Guided Self-Adaptive Wolf Search Algorithm Based on Information Entropy Theory. Entropy 2018, 20, 37. [Google Scholar] [CrossRef]
  30. Choong, S.S.; Wong, L.P.; Lim, C.P. A dynamic fuzzy-based dance mechanism for the bee colony optimization algorithm. Comput. Intell. 2018, 34, 999–1024. [Google Scholar] [CrossRef]
  31. Umeda, T.; Kitamura, A.; Konishi, M. Application of Meta-Heuristics Accompanied with Reinforcement Learning Mechanism to Optimization of Rolling Sequence. Trans. Inst. Syst. Control Inf. Eng. 2001, 14, 536–544. [Google Scholar] [CrossRef]
  32. Sharma, D.; Kohli, N. Crow search optimization with deep transfer learning enabled ventricular fibrillation prediction model. Int. J. Inf. Technol. 2025, 17, 1087–1101. [Google Scholar] [CrossRef]
  33. Lai, Z.; Feng, X.; Yu, H.; Luo, F. A Parallel Social Spider Optimization Algorithm Based on Emotional Learning. IEEE Trans. Syst. Man, Cybern. Syst. 2021, 51, 797–808. [Google Scholar] [CrossRef]
  34. Khan, A.R. Dynamic Load Balancing in Cloud Computing: Optimized RL-Based Clustering with Multi-Objective Optimized Task Scheduling. Processes 2024, 12, 519. [Google Scholar] [CrossRef]
  35. Mohammed, M.A.; Al-Khateeb, B.; Yousif, M.; Mostafa, S.A.; Kadry, S.; Abdulkareem, K.H.; Garcia-Zapirain, B. Novel crow swarm optimization algorithm and selection approach for optimal deep learning COVID-19 diagnostic model. Comput. Intell. Neurosci. 2022, 2022, 1307944. [Google Scholar] [CrossRef] [PubMed]
  36. He, J. Construction of supply chain coordination and optimization model of fresh food e-commerce platform based on improved bacterial foraging algorithm. RAIRO-Oper. Res. 2022, 56, 3853–3869. [Google Scholar] [CrossRef]
  37. Sivamathi, C.; Karthick, G.; Vijayarani, S. Glowworm Swarm Optimization Algorithm for Retrieval of High Utility Itemsets. In Proceedings of the 2024 International Conference on Smart Systems for Electrical, Electronics, Communication and Computer Engineering (ICSSEECC), Coimbatore, India, 28–29 June 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 111–116. [Google Scholar]
  38. Akhtar, T.; Haider, N.G.; Khan, S.M. A comparative study of the application of glowworm swarm optimization algorithm with other nature-inspired algorithms in the network load balancing problem. Eng. Technol. Appl. Sci. Res. 2022, 12, 8777–8784. [Google Scholar] [CrossRef]
  39. Iordache, S. Consultant-guided search combined with local search for the traveling salesman problem. In Proceedings of the 12th Annual Conference Companion on Genetic and Evolutionary Computation, New York, NY, USA, 7–11 July 2010; pp. 2087–2088. [Google Scholar]
  40. Rajendran, V.; Ramasamy, R.K.; Mohd-Isa, W.N. Improved Eagle Strategy Algorithm for Dynamic Web Service Composition in the IoT: A Conceptual Approach. Future Internet 2022, 14, 56. [Google Scholar] [CrossRef]
  41. Xiong, R.; Wang, S.; Yu, C.; Fernandez, C.; Xiao, W.; Jia, J. A novel nonlinear decreasing step-bacterial foraging optimization algorithm and simulated annealing-back propagation model for long-term battery state of health estimation. J. Energy Storage 2023, 59, 106484. [Google Scholar] [CrossRef]
  42. Chu, Y.; Mi, H.; Liao, H.; Ji, Z.; Wu, Q. A fast bacterial swarming algorithm for high-dimensional function optimization. In Proceedings of the 2008 IEEE Congress on Evolutionary Computation (IEEE World Congress on Computational Intelligence), Hong Kong, China, 1–6 June 2008; IEEE: Piscataway, NJ, USA, 2008; pp. 3135–3140. [Google Scholar]
  43. Guo, J.; Sato, Y. A hierarchical bare bones particle swarm optimization algorithm. In Proceedings of the 2017 IEEE International Conference on Systems, Man, and Cybernetics (SMC), Banff, AB, Canada, 5–8 October 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 1936–1941. [Google Scholar]
  44. Su, S.; Wang, J.; Fan, W.; Yin, X. Good lattice swarm algorithm for constrained engineering design optimization. In Proceedings of the 2007 International Conference on Wireless Communications, Networking and Mobile Computing, Shanghai, China, 21–25 September 2007; IEEE: Piscataway, NJ, USA, 2007; pp. 6421–6424. [Google Scholar]
  45. Li, Y. Optimization Model Design of Deep Foundation Pit Support Based on Improved Fish Swarm Algorithm. In Proceedings of the 2023 International Conference on Telecommunications, Electronics and Informatics (ICTEI), Lisbon, Portugal, 11–13 September 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 640–645. [Google Scholar]
  46. Xiao, Y. Optimization Model of Financial Market Portfolio Using Artificial Fish Swarm Model and Uniform Distribution. Comput. Intell. Neurosci. 2022, 2022, 7483454. [Google Scholar] [CrossRef]
  47. Agrawal, P.K.; Pandit, M.; Dubey, H.M. Improved Krill Herd Algorithm with neighborhood distance concept for optimization. Int. J. Intell. Syst. Appl. 2016, 8, 34. [Google Scholar] [CrossRef]
  48. Agrawal, P.K.; Pandit, M. Improved krill herd algorithm with neighborhood distance concept for numerical optimization. In Proceedings of the 2016 IEEE International Conference on Engineering and Technology (ICETECH), Coimbatore, India, 17–18 March 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 1105–1110. [Google Scholar]
  49. Reddy, M.R.P.; Kumar, D.S.; Sanjay, S.R.; Kumar, P.M.; Bharath, Y.S. Achieving Cost Efficiency and Load Sharing in Power Systems: The Superiority of Adaptive BAT Algorithm. In Proceedings of the 2024 International Conference on Communication, Computing and Internet of Things (IC3IoT), Chennai, India, 17–18 April 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 1–5. [Google Scholar]
  50. Yadav, K.; Mouli, K.C.; Saritha, G.; Spandana, A.; Abbas, H.M.; Lazer, N. Adaptive Self-Healing in Mobile Networks Using Levy Flight-Based Bat Algorithm. In Proceedings of the 2024 International Conference on IoT, Communication and Automation Technology (ICICAT), Gorakhpur, India, 23–24 November 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 1281–1286. [Google Scholar]
  51. Liu, Z.; Yang, P.; Zhang, P.; Lin, X.; Wei, J.; Li, N. Optimization of Fuzzy Control Parameters for Wind Farms and Battery Energy Storage Systems Based on an Enhanced Artificial Bee Colony Algorithm under Multi-Source Sensor Data. Sensors 2024, 24, 5115. [Google Scholar] [CrossRef]
  52. Kaya, E. A new neural network training algorithm based on artificial bee colony algorithm for nonlinear system identification. Mathematics 2022, 10, 3487. [Google Scholar] [CrossRef]
  53. Xu, W.; Li, Z.; Liu, J.; Cui, J.; Hu, Y. Virtual reconfiguration method of robotic mixed-model assembly line using bees algorithm based on digital twin. IEEE Trans. Autom. Sci. Eng. 2023, 21, 2211–2222. [Google Scholar] [CrossRef]
  54. Acar, O.; Kalyoncu, M.; Hassan, A. Proposal of a harmonic bees algorithm for design optimization of a gripper mechanism. In Proceedings of the IFToMM World Congress on Mechanism and Machine Science, Krakow, Poland, 30 June–4 July 2019; Springer: Berlin/Heidelberg, Germany, 2019; pp. 2829–2839. [Google Scholar]
  55. Vedaee, A.; Askari, G.; Sadeghi, H.M.; Fadaei, M. A New Analytical Redesign of a Double-Curvature Reflector Antenna Using Invasive Weed Optimization (IWO) Algorithm. Prog. Electromagn. Res. C 2021, 114, 263–279. [Google Scholar] [CrossRef]
  56. Pan, G.; Li, K.; Ouyang, A.; Zhou, X.; Xu, Y. A hybrid clustering algorithm combining cloud model iwo and k-means. Int. J. Pattern Recognit. Artif. Intell. 2014, 28, 1450015. [Google Scholar] [CrossRef]
  57. Deb, S.; Fong, S.; Tian, Z. Elephant search algorithm for optimization problems. In Proceedings of the 2015 Tenth International Conference on Digital Information Management (ICDIM), Jeju, Republic of Korea, 21–23 October 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 249–255. [Google Scholar]
  58. Tian, Z.; Fong, S.; Wong, R.; Millham, R. Optimizing self-adaptive gender ratio of elephant search algorithm by min-max strategy. In Proceedings of the 2016 12th International Conference on Natural Computation, Fuzzy Systems and Knowledge Discovery (ICNC-FSKD), Changsha, China, 13–15 August 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 121–126. [Google Scholar]
  59. Yi, T.H.; Zhang, X.D.; Li, H.N. Modified monkey algorithm and its application to the optimal sensor placement. Appl. Mech. Mater. 2012, 178, 2699–2702. [Google Scholar]
  60. Shaban, R.Z.; Alkallak, I.N. Organizing sports matches with a hybrid monkey search algorithm. Indones. J. Electr. Eng. Comput. Sci. 2021, 22, 542–551. [Google Scholar] [CrossRef]
  61. Horng, S.C.; Lin, S.S.; Lin, Y.C. Apply Manta Ray Foraging Optimization to Solve the Continuous-Time Markov Chain Problems. In Proceedings of the 2022 International Automatic Control Conference (CACS), Kaohsiung, Taiwan, 3–6 November 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 1–5. [Google Scholar]
  62. Zheng, T.; Zhang, B.; Liu, S.; Todo, Y.; Gao, S. Using Manta Ray Foraging Mechanism to Improve Reptile Search Algorithm. In Proceedings of the 2023 5th International Conference on Control and Robotics (ICCR), Kaohsiung, Taiwan, 3–6 November 2022; IEEE: Piscataway, NJ, USA, 2023; pp. 267–271. [Google Scholar]
  63. Ekinci, S.; Izci, D.; Kayri, M. An effective controller design approach for magnetic levitation system using novel improved manta ray foraging optimization. Arab. J. Sci. Eng. 2022, 47, 9673–9694. [Google Scholar] [CrossRef]
  64. Biçer, M.B. Analysis of C-shaped compact microstrip antennas using deep neural networks optimized by Manta Ray foraging optimization with Lévy-Flight Mechanism. Sak. Univ. J. Comput. Inf. Sci. 2021, 4, 166–180. [Google Scholar]
  65. Daqaq, F.; Kamel, S.; Ouassaid, M.; Ellaia, R.; Agwa, A.M. Non-dominated sorting manta ray foraging optimization for multi-objective optimal power flow with wind/solar/small-hydro energy sources. Fractal Fract. 2022, 6, 194. [Google Scholar] [CrossRef]
  66. Alturki, F.A.; Omotoso, H.O.; Al-Shamma’a, A.A.; Farh, H.M.; Alsharabi, K. Novel manta rays foraging optimization algorithm based optimal control for grid-connected PV energy system. IEEE Access 2020, 8, 187276–187290. [Google Scholar] [CrossRef]
  67. Spea, S. Cost-effective economic dispatch in large-scale power systems using enhanced manta ray foraging optimization. Neural Comput. Appl. 2025, 37, 12487–12524. [Google Scholar] [CrossRef]
  68. Elaziz, M.A.; Abualigah, L.; Ewees, A.A.; Al-qaness, M.A.; Mostafa, R.R.; Yousri, D.; Ibrahim, R.A. Triangular mutation-based manta-ray foraging optimization and orthogonal learning for global optimization and engineering problems. Appl. Intell. 2023, 53, 7788–7817. [Google Scholar] [CrossRef]
  69. Yeh, W.C.; Lin, Y.P.; Liang, Y.C.; Lai, C.M.; Huang, C.L. Simplified swarm optimization for hyperparameters of convolutional neural networks. Comput. Ind. Eng. 2023, 177, 109076. [Google Scholar] [CrossRef]
  70. Luccioni, A.S.; Hernandez-Garcia, A. Counting Carbon: A Survey of Factors Influencing the Emissions of Machine Learning. arXiv 2023, arXiv:2302.08476. [Google Scholar] [CrossRef]
  71. Krairiksh, K.; Choksuchat, C. Awareness of Green Academic Library by KYL Dashboard towards Sustainable Digital University. In Proceedings of the 2021 2nd SEA-STEM International Conference (SEA-STEM), Hat Yai, Thailand, 24–25 November 2021; pp. 108–111. [Google Scholar] [CrossRef]
  72. Narayanan, D.; Shoeybi, M.; Casper, J.; LeGresley, P.; Patwary, M.; Korthikanti, V.; Vainbrand, D.; Kashinkunti, P.; Bernauer, J.; Catanzaro, B.; et al. Efficient large-scale language model training on gpu clusters using megatron-lm. In Proceedings of the International Conference for High Performance Computing, Networking, Storage and Analysis, St. Louis, MO, USA, 14–19 November 2021; pp. 1–15. [Google Scholar]
  73. Hong, Y.; Ltaief, H.; Ravasi, M.; Keyes, D. High performance computing seismic redatuming by inversion with algebraic compression and multiple precisions. Int. J. High Perform. Comput. Appl. 2024, 38, 225–244. [Google Scholar] [CrossRef]
  74. Alevizos, V.; Yue, Z.; Edralin, S.; Xu, C.; Gerolimos, N.; Papakostas, G.A. A Logarithmic Compression Method for Magnitude-Rich Data: The LPPIE Approach. Technologies 2025, 13, 278. [Google Scholar] [CrossRef]
Table 1. Hyperparameter-to-emission factors ( h k ). Δ eval is the measured increase in objective evaluations (or equivalent FLOPs/probe calls) versus a fixed reference configuration on the same hardware and seed.
Table 1. Hyperparameter-to-emission factors ( h k ). Δ eval is the measured increase in objective evaluations (or equivalent FLOPs/probe calls) versus a fixed reference configuration on the same hardware and seed.
Knob (Algo)SymbolMeasured Signal for Δ Eval Factor FormTypical Range
Inertia weight (PSO)wevals vs. w = 0.7 h w = 1 + Δ eval / eval ref [ 1.00 , 1.10 ]
Cognitive coeff. (PSO) c 1 evals vs. c 1 = 1.5 h c 1 = 1 + Δ eval / eval ref [ 1.00 , 1.08 ]
Social coeff. (PSO) c 2 evals vs. c 2 = 1.5 h c 2 = 1 + Δ eval / eval ref [ 1.00 , 1.08 ]
Scale factor (DE)Fevals vs. F = 0.5 h F = 1 + Δ eval / eval ref [ 1.00 , 1.12 ]
Crossover (DE) C R trial eval rate h C R = 1 + Δ eval / eval ref [ 1.00 , 1.10 ]
Archive size (SHADE)Acopies/mem ops h A = 1 + | A | / N p [ 1.00 , 1.20 ]
Restart trigger (jSO/IMODE) I restart restart count h R = 1 + I restart [ 1.00 , 1.05 ]
Stopping tolerance ϵ extra iters to converge h ϵ = 1 + Δ eval / eval ref [ 1.00 , 1.10 ]
Table 2. Topology-to-emission factors ( t l ). Communication cost is proxied by transmitted bytes per iteration relative to a fully-connected baseline.
Table 2. Topology-to-emission factors ( t l ). Communication cost is proxied by transmitted bytes per iteration relative to a fully-connected baseline.
TopologyTraffic ProxyFactor FormTypical Range
Global/fully-connectedbytestx vs. global t global = 1 + bytes tx bytes ref [ 1.00 , 1.00 ] (ref)
Ring/localbytestx vs. global t ring = 1 + bytes tx bytes ref [ 0.85 , 0.95 ]
Von Neumannbytestx vs. global t vn = 1 + bytes tx bytes ref [ 0.90 , 0.98 ]
Starbytestx vs. global t star = 1 + bytes tx bytes ref [ 0.95 , 1.05 ]
Mesh/randombytestx vs. global t mesh = 1 + bytes tx bytes ref [ 0.90 , 1.00 ]
Tree/hierarchicalbytestx vs. global t tree = 1 + bytes tx bytes ref [ 0.88 , 0.98 ]
Dynamic/adaptivebytestx vs. global t dyn = 1 + bytes tx bytes ref [ 0.88 , 1.02 ]
Table 3. Boundary-handling factors ( b m ). ρ hit is the observed boundary-hit frequency; π pen is penalty estimated via microbenchmarks (e.g., extra ops, cache misses).
Table 3. Boundary-handling factors ( b m ). ρ hit is the observed boundary-hit frequency; π pen is penalty estimated via microbenchmarks (e.g., extra ops, cache misses).
MethodSignal (s)Factor FormTypical Range
Reflect/mirror ρ hit , π pen b refl = 1 + ρ hit π pen [ 1.00 , 1.06 ]
Clamp/absorb ρ hit , π pen b abs = 1 + ρ hit π pen [ 1.00 , 1.03 ]
Random reinit ρ hit , π pen b rand = 1 + ρ hit π pen [ 1.00 , 1.08 ]
Periodic/wrap ρ hit , π pen b per = 1 + ρ hit π pen [ 1.00 , 1.05 ]
Hyperbolic ρ hit , π pen b hyp = 1 + ρ hit π pen [ 1.00 , 1.07 ]
Exponential damping ρ hit , π pen b exp = 1 + ρ hit π pen [ 1.00 , 1.06 ]
Mutation-based ρ hit , π pen b mut = 1 + ρ hit π pen [ 1.00 , 1.09 ]
Table 4. Parameter-to-Emission Mapping for DE Variants.
Table 4. Parameter-to-Emission Mapping for DE Variants.
DE ParameterEmissions FactorMapping Rationale
Population size N p Hyperfactorial H ( N p ) Directly scales memory
footprint and parallel threads
Generations N i Superfactorial sf ( N i ) Governs wall-time, hence
kWh total
Scale factor F h F = 1 + | F 0.5 | High F increases
evaluation variance and CPU cache misses
Crossover rate C R h C R = 1 + C R Large C R triggers
more trial evaluations
Archive size (SHADE lineage) h A = 1 + | A | / N p Extra memory and copy
operations
Restart trigger (jSO) h R = 1 + I restart Counts soft resets that
flush caches
Table 5. Extended Results Featuring 41 Models (Part 1).
Table 5. Extended Results Featuring 41 Models (Part 1).
ModelFamilyPowEffEmR CO 2 gFit
GAStoch./RS0.450.820.5368.4 2.1 × 10 15
DEStoch./RS0.470.770.6492.0 1.9 × 10 15
ESimpleStoch./RS0.500.800.70123.1 1.6 × 10 15
PBILStoch./RS0.330.790.4239.8 5.5 × 10 16
RandomWalkStoch./RS0.300.750.4124.2 7.4 × 10 14
RSFStoch./RS0.440.830.5565.2 4.3 × 10 16
SOAStoch./RS0.600.710.90164.0 8.2 × 10 15
MemeticHybrid0.550.780.88202.3 1.3 × 10 15
HybridABCHybrid0.580.761.10321.5 3.2 × 10 16
HybridPSOHybrid0.520.740.66135.8 2.4 × 10 15
NCOHybrid0.620.790.92286.4 2.0 × 10 15
BBONature-Insp.0.430.850.4556.7 2.3 × 10 16
FANature-Insp.0.490.740.230.16 1.7 × 10 15
GWONature-Insp.0.470.800.6187.3 5.9 × 10 16
CSNature-Insp.0.640.701.00260.5 3.1 × 10 15
SSANature-Insp.0.660.831.10398.0 1.7 × 10 15
WHANature-Insp.0.680.821.12416.3 9.8 × 10 17
PSOMulti-Agent0.400.840.4854.0 2.4 × 10 15
ACOMulti-Agent0.390.820.4648.8 3.7 × 10 16
ABCMulti-Agent0.570.750.84174.1 1.8 × 10 15
HSMulti-Agent0.560.780.77162.2 2.1 × 10 16
Table 6. Extended Results Featuring 41 Models (Part 2).
Table 6. Extended Results Featuring 41 Models (Part 2).
ModelFamilyPowEffEmR CO 2 gFit
BANature-Insp.0.300.700.200.25 1.2 × 10 15
MFONature-Insp.0.710.811.11795.00 3.5 × 10 16
WOANature-Insp.0.480.840.63105.20 9.2 × 10 16
SAStoch/Random0.320.790.3728.50 4.1 × 10 15
EStrategyStoch/Random0.780.780.68830.10 2.2 × 10 15
LFAMulti-Agent0.590.800.82220.00 8.0 × 10 17
SharksHybrid0.630.731.00275.20 4.6 × 10 16
IGWOHybrid0.610.770.95243.10 5.6 × 10 17
BBFNature-Insp.0.380.850.4348.00 3.2 × 10 16
MSFLAMulti-Agent0.460.800.5270.10 8.7 × 10 16
BEANature-Insp.0.550.840.98194.30 2.1 × 10 16
MRFONature-Insp.0.370.770.270.64 6.0 × 10 16
GeoOptStoch./RS0.540.720.76112.90 6.2 × 10 16
FOAStoch/Random0.400.810.4251.20 7.7 × 10 16
HHONature-Insp.0.650.791.05334.10 1.4 × 10 15
IBCOMulti-Agent0.720.701.14415.90 8.1 × 10 17
FireworksStoch./RS0.500.760.5993.10 2.2 × 10 16
SFLAMulti-Agent0.780.780.68908.80 1.2 × 10 15
BactForNature-Insp.0.380.780.280.51 2.8 × 10 16
SSOMulti-Agent0.430.830.330.29 1.6 × 10 15
SLnOStoch./RS0.690.730.90147.20 9.6 × 10 16
Table 7. Selected Average Emissions per Family (Sampled Subset).
Table 7. Selected Average Emissions per Family (Sampled Subset).
FamilyMean P h Mean η Mean CO2 (g)
Stochastic/Random0.420.79215.3
Multi-Agent Coop.0.530.81341.6
Hybrid0.600.77482.1
Nature-Inspired0.580.85625.9
Table 8. Mean CO2 per successful optimisation (g)    Cluster: 2×A100, PUE  = 1.09 .
Table 8. Mean CO2 per successful optimisation (g)    Cluster: 2×A100, PUE  = 1.09 .
Algorithm N p N i CO 2 ¯
GridSearch720061.3
RandomSearch720049.7
NelderMead480042.8
Particle Swarm12880024.9
Firefly12880030.6
Cuckoo6480028.1
Whale6480026.4
Hybrid PSO-DE12840027.7
Table 9. Particle Swarm sensitivity study.
Table 9. Particle Swarm sensitivity study.
N p N i Time (s)Energy (Wh) ε (mg)
3280014249.119.2
648009546.79.1
1288006355.25.4
1284004447.54.6
Table 10. Methodological outcomes for carbon accounting across 480 runs; comparisons anchor to EIM as reference.
Table 10. Methodological outcomes for carbon accounting across 480 runs; comparisons anchor to EIM as reference.
MethodScopeTimeSourceBias vs. EIM (%)
EIM-calibrated (this study)21 minMCI hourly0.0
Location-based baseline21 hGrid avg+2.5
Market-based baseline21 ySupplier mix+27.9
PUE-adjusted location21 minGrid avg×PUE+7.5
Mean CO2 (g)95% CI (g)
31.8[30.9, 32.7]
32.6[31.7, 33.5]
40.7[39.4, 42.0]
34.2[33.1, 35.4]
Scope 1 + 2 (diesel event)1.2eventGrid avg+genset+11.9
35.6[33.9, 37.2]
Table 11. Per-model statistical controls with validation vs established carbon methodologies; EIM as reference.
Table 11. Per-model statistical controls with validation vs established carbon methodologies; EIM as reference.
ModelRunsEIMMean (g)95%CI (g)LBBias (%)MBBias (%)MAECC (g)MAEMLCO2 (g)SW pBP p
PSO4024.9[23.8, 26.0]+2.6+28.10.60.70.210.33
WOA4026.4[25.2, 27.6]+2.4+27.50.70.80.190.28
Cuckoo (CS)4028.1[27.0, 29.2]+2.7+28.60.80.90.170.31
Firefly (FA)4030.6[29.2, 32.0]+2.5+27.80.91.00.150.29
Hybrid PSO–DE3227.7[26.6, 28.8]+2.6+28.40.70.80.220.35
GridSearch3261.3[59.1, 63.5]+2.4+27.61.11.30.180.26
RndSearch3249.7[48.0, 51.4]+2.5+27.90.91.10.200.27
L-SHADE2420.0[19.1, 20.9]+2.8+29.20.50.60.240.36
jSO2421.5[20.6, 22.4]+2.7+28.70.60.70.230.34
IMODE2422.7[21.7, 23.7]+2.6+28.30.60.70.250.32
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Alevizos, V.; Gerolimos, N.; Leligkou, E.A.; Hompis, G.; Priniotakis, G.; Papakostas, G.A. Sustainable Swarm Intelligence: Assessing Carbon-Aware Optimization in High-Performance AI Systems. Technologies 2025, 13, 477. https://doi.org/10.3390/technologies13100477

AMA Style

Alevizos V, Gerolimos N, Leligkou EA, Hompis G, Priniotakis G, Papakostas GA. Sustainable Swarm Intelligence: Assessing Carbon-Aware Optimization in High-Performance AI Systems. Technologies. 2025; 13(10):477. https://doi.org/10.3390/technologies13100477

Chicago/Turabian Style

Alevizos, Vasileios, Nikitas Gerolimos, Eleni Aikaterini Leligkou, Giorgos Hompis, Georgios Priniotakis, and George A. Papakostas. 2025. "Sustainable Swarm Intelligence: Assessing Carbon-Aware Optimization in High-Performance AI Systems" Technologies 13, no. 10: 477. https://doi.org/10.3390/technologies13100477

APA Style

Alevizos, V., Gerolimos, N., Leligkou, E. A., Hompis, G., Priniotakis, G., & Papakostas, G. A. (2025). Sustainable Swarm Intelligence: Assessing Carbon-Aware Optimization in High-Performance AI Systems. Technologies, 13(10), 477. https://doi.org/10.3390/technologies13100477

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop