Next Article in Journal
Pinching Results on Totally Real Submanifolds of a Locally Conformal Kähler Manifolds
Previous Article in Journal
The Fourth Hybrid Power Mean Involving the Character Sums and Exponential Sums
Previous Article in Special Issue
High-Performance Real-Time Human Activity Recognition Using Machine Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

High-Performance Deployment Operational Data Analytics of Pre-Trained Multi-Label Classification Architectures with Differential-Evolution-Based Hyperparameter Optimization (AutoDEHypO)

by
Teo Prica
1,2 and
Aleš Zamuda
2,*
1
Institute of Information Science, Prešernova ulica 17, 2000 Maribor, Slovenia
2
Faculty of Electrical Engineering and Computer Science, University of Maribor, Koroška cesta 46, 2000 Maribor, Slovenia
*
Author to whom correspondence should be addressed.
Mathematics 2025, 13(10), 1681; https://doi.org/10.3390/math13101681
Submission received: 8 April 2025 / Revised: 18 May 2025 / Accepted: 19 May 2025 / Published: 20 May 2025
(This article belongs to the Special Issue Innovations in High-Performance Computing)

Abstract

:
This article presents a high-performance-computing differential-evolution-based hyperparameter optimization automated workflow (AutoDEHypO), which is deployed on a petascale supercomputer and utilizes multiple GPUs to execute a specialized fitness function for machine learning (ML). The workflow is designed for operational analytics of energy efficiency. In this differential evolution (DE) optimization use case, we analyze how energy efficiently the DE algorithm performs with different DE strategies and ML models. The workflow analysis considers key factors such as DE strategies and automated use case configurations, such as an ML model architecture and dataset, while monitoring both the achieved accuracy and the utilization of computing resources, such as the elapsed time and consumed energy. While the efficiency of a chosen DE strategy is assessed based on a multi-label supervised ML accuracy, operational data about the consumption of resources of individual completed jobs obtained from a Slurm database are reported. To demonstrate the impact on energy efficiency, using our analysis workflow, we visualize the obtained operational data and aggregate them with statistical tests that compare and group the energy efficiency of the DE strategies applied in the ML models.

1. Introduction

High-performance computing (HPC) continues to advance and to provide important infrastructure and a backbone for the scientific community, industry, and beyond, making efficient utilization highly important for both environmental sustainability and cost efficiency [1,2]. Despite the increasing availability of automated ML (AutoML) frameworks, most existing tools prioritize and maximize the performance (accuracy) and neglect other important objectives such as energy efficiency (EE) and resource usage and optimization, which are key factors in large-scale HPC environments and beyond [3,4]. Therefore, the enforcement of sustainable and energy-efficient solutions, including the utilization and optimization of computational resources, now plays a key role in reducing operational financial costs and minimizing environmental impacts [1,3]. Due to the broad accessibility of computational resources for HPC, the scientific community with the myriad of application areas and diverse expertise from different fields, can deploy and run their workflows [5,6]. As their needs, demands, and complexity grow, running data-intensive and parallel workflows involves both different and heterogeneous architectures and the resources of state-of-the-art systems [6,7,8]. Such heterogeneous architectures introduce challenges in workload scheduling, as there are many unknowns and uncertainties  [2,9]. However, despite the optimization capabilities and advances, most AutoML frameworks do not incorporate schedulers such as Slurm, nor do they provide support for energy monitoring, as their main focus is on improving ML model performance (accuracy) and ignore constraints such as EE, utilization, and resource allocation [3,4,10]. To address these challenges, adjusted and tailored frameworks that support sustainable and cost-efficient HPC environments are needed. Therefore, this article presents a high-performance-computing (HPC) differential-evolution-based automatic hyperparameter optimization workflow (AutoDEHypO) for energy efficiency and operational data analytics using multiple graphics processing units (GPUs), and this is deployed on the petascale EuroHPC supercomputer Vega [11]. Our proposed method is capable of determining how energy-efficient the differential evolution (DE) algorithm and ML models perform. The proposed AutoDEHypO workflow considers both their achieved accuracy and the utilization of computing resources, such as elapsed time and consumed energy; it collects runtime data through Slurm within the HPC environment, and the job allocation may impact the obtained results. Moreover, resource consumption leads to inefficient workloads, waste of computational resources, the consumption of a project quota, longer queue times, and postponement of scientific research [12]. Since Slurm dynamically schedules jobs to available, unutilized, or idle nodes within the cluster queue, proper adjustments in job submission may lead to faster allocation and faster attainment of the required job results [13]. We aim to develop the ability to operate within the constraints that Slurm users have and determine whether statistical deviations will be significant. Thus, we prepared our environment for the deployment of the AutoDEHypO workflow, where we ensured consistent resource allocation across the submitted job scripts with a Slurm script (SBATCH). Furthermore, this setup includes supervised machine learning (ML) and multi-label classification and allows for applicable optimization for aspects of hyperparameters that could be affected [14]. We chose the recent parallel implementation of the DE algorithm [15], which parallelizes the population operations in an algorithm that was initially introduced by Storn and Price in 1995 [16]. While DE provides a favorable trade-off and balance between accuracy and energy efficiency (EE) in the context of HPC [17], it additionally offers several advantages over other optimization algorithms, such as efficient and faster convergence, suitability and adaptability in a variety of different environments and optimization problems, global optimization, parallelization, scalability, energy optimization, and the possibility for combination and complementarity with different algorithms and techniques due to available implementations and integrations [16,18,19,20,21]. Furthermore, DE is suitable for the given optimization problem, as it can be used for optimizing nonlinearities in data, such as ML-related input/output data within the workflow, and it can be adapted to other use cases and problems [19]. Data are gathered from system scheduling and historical data; resource allocation and job execution times make it difficult to provide prediction and optimization [16,19,20,21]. In the evaluation phase of our experiment, we used basic ML metrics [14]. During the deployment of individual jobs on multiple GPUs in our experiment, we examined the efficiency of a chosen ML model and DE algorithm according to the accuracy [14]. DE functions and strategies show distinct behavior, and this has not yet been extensively measured or optimized within the context of EE and system resource management in HPC environments [20,22,23]. Data on the consumption of resources of individual jobs were obtained from the Slurm database of completed jobs, with energy consumption data being reported for the whole node by Intelligent Platform Management Interface (IPMI) sensors [13,24]. Based on the results of the computations, aggregated statistics were calculated, along with corresponding post hoc procedures and visualizations, to evaluate the efficiency of the combined ML model and the applied DE strategy.

1.1. Problem Statement and Objective

This work addresses the challenge of optimizing machine learning (ML) models in high-performance computing environments by balancing ML accuracy, energy consumption, and resource utilization. To address these challenges, we propose AutoDEHypO, a differential-evolution-based workflow that is specifically designed for energy-monitored hyperparameter optimization. The specific problem stated for this study consists of considering some key limitations, such as inefficient utilization of computational resources during job allocation, extended queue times, and lack of integration with HPC environments and schedulers such as Slurm. The limitations of existing workflows and frameworks are that they mostly focus on a single objective, such as ML performance or integration within HPC environments, and do not necessarily contribute to the sustainability and cost-effectiveness of HPC environments [25,26]. Therefore, we are interested in analysis of the consumption of resources and when the consumed energy and other resources can lead to different ML performance and configurations.

1.2. Main Contributions

The main contributions of this article are as follows:
  • We propose AutoDEHypO, a high-performance-computing differential-evolution-based automatic hyperparameter workflow designed to optimize the performance of ML models for energy efficiency and operational data analytics in HPC environments.
  • We deploy the AutoDEHypO workflow on the EuroHPC Vega system, utilizing multiple GPUs and Slurm scheduling and submission to execute a specialized fitness function for ML.
  • We applied and evaluated this workflow on supervised ML and multi-label classification using the CIFAR10 and CIFAR100 datasets [27].
  • We collected runtime data through Slurm within the HPC production environment.
  • We evaluated the efficiency of a chosen ML model and DE algorithm strategies according to the ML accuracy and energy efficiency, dependent on ML model architecture, datasets, and resource consumption within the HPC environment.
  • We performed aggregated statistical analyses, along with the corresponding post hoc procedures, and validated the collected data using visualizations by evaluating efficiency of combined ML models and applied DE strategies.
  • We identified significant differences in key metrics and laid the ground for future work on sustainability and cost-effectiveness using AutoDEHypO.
The following sections of this article are organized as follows. Section 2 provides an overview of the related work and existing methods. Section 3 provides a set of proposed methods. Section 4 provides the experimental results. Section 5 presents our conclusion and future work.

2. Related Work and Existing Methods

The integration of machine learning (ML) with monitoring and operational data analytics (MODA) is a step towards enhancing operational efficiency, and DE is being utilized to run DE fitness functions and enhance model performance through hyperparameter optimization and other techniques [28]. Furthermore, automated machine learning (AutoML) automates workflows and reduces manual effort [12]; here, the effectiveness of an ML model is evaluated using fundamental ML evaluation metrics, non-parametric tests, post hoc procedures, and visualizations [14]. Despite the improvements, there is still a trade-off between accuracy and energy efficiency [14]. Moreover, checkpoint, restart, and predictive models provide additional robustness and project quotas for allocated computational resources. The following subsections, therefore, describe these topics in the order in which they were mentioned here.

2.1. Machine Learning

Machine learning (ML) is used to discover patterns, make predictions, provide automation, and generate useful knowledge from datasets [29]. Several types of such discoveries have been made, through methods such as classification, regression, segmentation, clustering, error detection, sequence analysis, and others [29]. The design of layers (how they are connected and structured) represents the architecture of a model, while the trained version (system) represents an ML model itself [29]. Furthermore, in the context of learning, there is supervised machine learning and unsupervised machine learning [29]. In ML, patterns are identified within a dataset, and thus the model wants to connect attributes and classes—for example, through classification—with a focus on similarities on the one hand; on the other hand, unsupervised machine learning focuses on identifying underlying structures without a target attribute, such as segmentation [29].
In ML, optimization algorithms are used to improve results—specifically, losses within a selected space [14,30]. With the help of an ML algorithm, hyperparameters can be adjusted and optimized even before the learning phase [14]. The hyperparameters of ML are the configuration variables of ML methods [14]. Some of the most widely used optimization methods are grid search, random search, cross-validation, and Bayesian optimization [30,31]. The execution time of a manual search can be significantly longer due to the complexity of ML models and neural networks within their limits, and as a result, such a search uses significantly more computational resources [8,32].
A convolutional neural network (CNN) is suitable for processing patterns within images, processing video content, performing face recognition, processing medical data, and more. A CNN uses a filter, converts pixels into numerical values in the first layer, and processes or teaches them to recognize and analyze data with the help of subsequent layers [33].
Recurrent neural networks (RNNs) use a mechanism for storing information from previous states and process and interpret it, making them more suitable for processing when the data are structured and in some kind of sequence. They are used in areas such as natural language processing (NLP) and speech and handwriting recognition [34].
The complexity and efficiency of the model may be defined by the architecture design and structure of the NN, the layers, and their connectivity, including the number of layers, the framework, the model size, the  optimization that is applied, the input data, and more [8,35]. The efficiency and emergence of smaller ML models, which are also referred to as language models [36], may play a key role, as they are easier to manage and provide flexibility, allowing faster implementation of changes with fewer computational resources [37,38].
Multi-label classification is a type of supervised machine learning that enables each instance to be associated with multiple labels, and it differs from conventional single-label classification, where each instance can be associated with a single label [14,39]. Fine-tuning and other optimization techniques can be applied to hyperparameters, which will have an effect and improve the performance of a model [14,30]. Evaluation of progress and performance is performed using a metric (M). There are some fundamental ML metrics, such as the accuracy, precision, recall, and F-measure; in addition, metrics can be aggregated with possible prediction outcomes, e.g., total (T), true positive ( T P ), true negative ( T N ), false positive ( F P ), and false negative ( F N ), as presented in [14].
  • Accuracy (A) [14]:
    M = T P + T N T P + T N + F P + F N
  • Precision (P) [14]:
    M = T P T P + F P
  • Recall (R) [14]:
    M = T P T P + F N
  • F-Measure ( F 1 ) [14]:
    F 1 = 2 × P × R P + R
  • Metric Aggregation [14]:
    M = A + P + F 1 + R T

2.2. Monitoring and Operational Data Analytics

Monitoring and operational data analytics (MODA) in high-performance computing can be applied to broad standards and practices for the integration, collection, storage, processing, monitoring, visualization, and analysis of system data [28]. MODA is used with the goal of helping end users make informed decisions about information collected on their system through hardware power monitoring and management tools such as IPMI, RAPL, and others [40]. These tools provide information such as sensor data, usage of generic resources, job status, memory, wall time, job and node utilization, power consumption per job, nodes, blades, racks, clusters, data centers, cluster occupancy, billing, and more to be leveraged in many ways [13,28]. Furthermore, MODA can be applied and extended to either an active or a passive setup, which depends on the requirements, potential benefits, and contributions that it may offer [40]. An active setup actively influences the runtime operations, while on the other hand, a passive setup passively monitors, gathers, and collects data without affecting the runtime operations [40]. To enhance the impact of ML and gather insights into data, MODA may be used to monitor energy consumption, improve job efficiency, and improve the ML design for EE by providing real-time feedback mechanisms for adaptive and dynamic changes within ML operations, ML optimization suggestions such as predictions in ML models, or even operational insights to resolve issues by proactively predicting potential trends and patterns within ML operations [9,40,41]. Such an analysis could identify trade-offs and bring the right balance, resulting in potential reductions in the consumption of resources and bringing sustainability and EE into view [28]. Furthermore, MODA offers multiple opportunities and can potentially contribute to EE due to the possibility of monitoring energy and detecting significant differences, preventing potential issues and anomalies, finding potential root causes, detecting security vulnerabilities and malware, and providing improvements in performance, reliability, throughput, and beyond [42]. Using HPC Vega [11], the topics of MODA, including energy monitoring, operational efficiency, and optimization of HPC workloads, were investigated and recently addressed in [43].

2.3. Differential Evolution

The differential evolution (DE) algorithm is a population algorithm and was introduced by Storn and Price in 1995 [44]. DE has been used successfully in the field of optimization of numerical functions and in applications to real problems within different domains [16]. The original DE has a main evolutionary loop through which, with the help of the key and basic operators presented in Figure 1, such as initialization, mutation, crossover, and selection, the final result can be gradually improved soon after each generation by repeating these operators until the end [45]. From its introduction to the present, interest in DE has increased year on year [16,44]. Thus, DE has been developed, improved, and applied to a variety of real-world problems within multidisciplinary domain areas, such as electrical and energy systems, artificial neural networks, manufacturing and pperation research, robotics and advanced systems, pattern recognition, image processing, bioinformatics and biomedicine, electrical engineering, and others, as presented in several articles [16,46,47,48,49]. Those mechanisms can be used to make the selection process easier [50]. DE provides a favorable trade-off and balance between accuracy and energy efficiency within HPC environments and beyond [17]. Additionally, DE offers several advantages over other optimization algorithms, such as efficient and faster convergence and suitability and adaptiveness in a variety of different environments and optimization problems; often, it provides better results than other algorithms [18,51,52,53]. Furthermore, DE can be used for nonlinear, nondifferentiable, and multi-objective problems [16]. It provides robustness in complex and multidimensional spaces, global optimization in complex spaces, parallelization, scalability in high-dimensional problems, and energy optimization and distribution without compromising throughput [16,18]. It can be integrated with penalty functions and other methods to handle constraints, and it can be combined and used to complement different algorithms and techniques due to its easy implementation and integration [16,21]. Last but not least, it is being extensively researched, and constant improvements to the algorithm are being made [16,18,21].

2.3.1. Differential Evolution Operators

In DE [20], the following operators are used as highlighted in Figure 1.
First, the search space is initialized, and all search variables are randomly populated in the search space for a population of NP vectors defined with a dimension; then, the parameter vectors may change over different iterations [16]. The sequential number of the current generation [16] is denoted by t, the optimization search dimension size is denoted with d, and  i { 1 , 2 , , NP } denotes a population vector index [16].
x i ( t ) = ( x i , 1 ( t ) , x i , 2 ( t ) , , x i , d ( t ) ) .
Within this initialization stage the vector must be composed of the following components within interval limits for minimum and maximum [16]:
x i , j ( 0 ) = x min , j + rand [ 0 , 1 ] ; i , j ( x max ; j x min ; j ) .
During mutation, after the first step is completed, the algorithm selects donor and mutated vectors for each population individual [55]. The following strategies are used most often (interval from 1 to NP ), where x i ( t ) denotes the target vector, v i ( t ) denotes the mutant vector, x best ( t ) denotes the best individual in the population, R j i denotes random indexes that are mutually different and different from i, F denotes the mutation scaling factor, and t denotes the current generation [16,55,56]
  • DE/rand/1: A random vector is chosen as the basis, and a one-sided weighted difference (vector) composed of two other random vectors is added to it.
    v i ( t ) = x R 1 i ( t ) + F ( x R 2 i ( t ) x R 3 i ( t ) ) ,
  • DE/best/1: The current best random vector is used as the basis, and an additional random difference (vector) is added to it.
    v i ( t ) = x best ( t ) + F ( x R 1 i ( t ) x R 2 i ( t ) ) ,
  • DE/current-to-best/1: For an individual vector to mutate with the best vector, one random difference (vector) is added.
    v i ( t ) = x i ( t ) + F ( x best ( t ) x i ( t ) ) + F ( x R 1 i ( t ) x R 2 i ( t ) ) ,
  • DE/rand/2: A random vector is chosen as the basis, and two independent differences (vectors) of four random vectors are added to it.
    v i ( t ) = x R 1 i ( t ) + F ( x R 2 i ( t ) x R 3 i ( t ) ) + F ( x R 4 i ( t ) x R 5 i ( t ) ) ,
  • DE/best/2: The current best random vector is used as the basis, and  two random differences (vectors) of four random vectors are added to it.
    v i ( t ) = x best ( t ) + F ( x R 1 i ( t ) x R 2 i ( t ) ) + F ( x R 3 i ( t ) x R 4 i ( t ) ) .
In the crossover phase, the donor vector crosses with the targeted vector [16]. The two most commonly used methods of crossing are exponential and binomial [16], and there are a few other combinations of mutation and crossover strategies in the original DE algorithm [20,55], namely, the  DE/rand/1/exp, DE/current-to-best/1/exp, DE/best/1/exp, DE/current-to-best/1/bin, DE/rand/2/exp, and DE/best/2/exp strategies. Exponential crossover works by choosing a random number within the size of the dimension {1, 2, …, d}. The target vector is then crossed with the donor vector [16]. The binomial crossover generates a number from interval [ 0 , 1 ] for each dimension randomly and compares it to crossover value [16], in order to exchange components, or also exchange the component at random index j = K .
u i , j ( t ) = v i , j ( t ) if j = K or rand [ 0 , 1 ] ; i , j C r , x i , j ( t ) otherwise .
In selection, survival to the next generation for both vectors, i.e., the target and the trial vector, is determined [16]. If the condition is met, the target vector is simply replaced according to the strategy chosen for the next generation; otherwise, it remains and is further evolved within the next generation [16].
x i ( t + 1 ) = u ( t ) if f ( u ( t ) ) ( x i ( t ) ) x i ( t ) otherwise .

2.3.2. Improvements to the Differential Evolution Algorithm and Energy Efficiency

Since the introduction and throughout development beyond the original DE algorithm [20], its main foundational concepts were addressed and subsequently extended [57]. Especially from the efficiency perspective, extended versions and improvements have been proposed over the years [16,21]. Numerous improved algorithms have been developed and presented annually at conferences, such as at the Institute of Electrical and Electronics Engineers (IEEE) Congress on Evolutionary Computation (CEC), the Genetic and Evolutionary Computation Conference (GECCO), Parallel Problem Solving From Nature (PPSN), and competitions [58]. DE competitions are organized mainly as part of the CEC [58] and GECCO. After each competition, a review of the algorithms is conducted, and the results are included in a technical report [58]. Therefore, from such efficiency perspective defined at these competitions or other benchmarks, there are newer DE methods with benchmarks, such as L-SHADE [59] and DISH [60] with parameter control benchmarking [61]. Several related works have used DE for energy efficiency (EE), demonstrating applicability within the context of high-performance computing (HPC) and beyond [23,62], including that DE has been utilized in testing benchmarking EE with large dimensions for some CEC functions in [63]. Notably, DE can be used as a tuning platform to utilize and optimize energy-efficient and energy-aware workloads [64].
Other authors have presented a self-adaptive DE method with mutation and crossover operators and its design for NN optimization [65]. Agarwal et al. presented a differential-evolution-based approach to compress and accelerate a convolution neural network model, where the compression of various ML models improved their accuracy on different datasets, including CIFAR10 and CIFAR100 [66]. The compact differential evolution (cDE) algorithm was presented by Mininno et al. for constrained edge environments and provides support in robotics where there are limited computational resources [62]. A memetic DE (MDE) algorithm for EE with a job scheduling mechanism and possibilities for parallelization and scaling was also presented by Xueqi et al. [67]. A differential-evolution-based EE system for minimizing the consumption of resources in unmanned aerial vehicles (UAVs) and IoT devices was recently presented by Abdel-Basset et al. [68]. DE was also applied to autonomous underwater vehicles (AUVs) for underwater glider path planning (UGPP) with the objective of collecting research data within unexplored areas, which presents additional challenges; this provides robustness for such missions by addressing energy optimization for underwater glider vehicles [69,70]. Jannsen et al. also recently presented a comparative study reviewing the potential of the DE algorithm on GPUs due to its advantage of parallelism and more [71].

2.4. Comparison of Hyperparameter Optimization Methods

Differential evolution (DE) can be compared with other hyperparameter optimization methods, such as random search (RS), grid search (GS), and Bayesian optimization (BO) [25,26]. Under constraints of limited time and computational resources, neither GS nor RS proves to be ideal [25,26] and both methods tend to be inefficient, especially in the context of complex ML models [25]. Due to the exhaustive nature of GS, it suffers when dealing with a larger number of dimensions as they increase [25]. As dimensionality increases, GS evaluates a large number of configurations, which leads to the consumption of many computational resources [25]. RS, while simpler and more robust, typically requires a larger set of evaluations to achieve competitive results, regardless of resource constraints [25]. RS selects configurations at random, potentially overlooking promising regions of the search space and resigned to optimal results [25]. BO uses surrogate models, which are very efficient in low-dimensional spaces and black-box problems [25]. Unlike RS and GS, BO does not try every possibility within a given space, insted it uses predictions of outcomes where the acquired knowledge is collected and updated [25,26]. BO is a very useful method for the desire for scalability and with increasing dimensions BO loses performance rapidly with increasing dimensionality [25,26]. Additionally, in connection with DE and other optimization algorithms, a novel deep-learning expansion called Deep-BIAS was introduced and presented by van Stein et al. [72] Furthermore, Explainable AI (XAI) for evolutionary computation (EC) and the current state-of-the-art [73] introducing techniques for global sensitivity analysis in evolutionary optimization with focus on effectiveness is presented [74]. Such comparison strengthens the basis of DE as the optimizer user (within this article), especially in scenarios where the evaluation of ML models is computationally demanding, where parallelism and scalability within robust complex search and multidimensional spaces are required, as well as its implementation in real problems in various fields, including ML as presented in Section 2.3.2.

2.5. Automated Machine Learning

Automated machine learning (AutoML) is the automation of workflows within an ML process and can be applied from the beginning to the end (end-to-end) [12]. The main focus is to avoid unnecessary repeatable steps [75]. The AutoML process itself is multi-step, so we select the data source, select the data, clean and preprocess the data, execute machine learning, algorithm selection, and parameter optimization, and interpret and discover knowledge [29]. This is also reflected in the growth in the number of published works in the last 10 years, e.g.,  [3,4,26,75,76]. Due to the significant and continuous progress in the field, ML represents a popular branch of computer science and beyond [75]. It has been successfully applied within a myriad of scientific domains [39], including to exploit the potential of infrastructures such as HPC, rids, cloud and edge computing, and beyond [34,77]. Moreover, the complexity and rapid growth in the field have also created a large gap in efforts to transition to more sustainable approaches and strategies [78,79]. Improvements in optimization through Red AI and a proof of concept based on the Scikit-learn library were presented by Castellanos-Nieves et al. [80], with another work presenting an evaluation of various applied and potentially suitable algorithms with a focus on Green AI [81]. Furthermore, Red AI is mostly focused on conventional approaches, such utilizing many computational resources with the relearning of ML models, maximizing the acquired ML accuracy, and neglecting EE [4,10]. However, in recent years, as presented in recent work, advances and evolutions in the fields of ML and AutoML show the necessary contribution of the strategies and concepts of Green AI when considering EE, as well as what the trade-offs of this approach are on the path toward a more sustainable landscape and Green AI [3,4]. Geissler et al. introduced a novel energy-aware hyperparameter optimization approach (Spend More to Save More (SM2)) based on the early rejection of inefficient hyperparameter configurations to save ML training time and resource consumption [78]. While recent studies addressed EE in connection with AutoML, they were not applied to HPC or the cloud, where the utilization of resources is crucial [78]. Evolutionary algorithms have also been applied within or as optimizers as an addition to AutoML, e.g., an improved optimization framework using evolution algorithms such as DE [51], self-adaptive control parameter randomization with DE [82], the application of DE to an energy-aware auto-tuning platform [64], and an auto-selection mechanism with the application of DE [50]. Vakhnin et al. presented a novel multi-objective hybrid evolution-based tuning framework for accurately forecasting power consumption [83], and the application of DE provided better results than those of other algorithms [51,52,53]. Benchmark studies were presented by Gomes et al. [53], in addition to those for other state-of-the-art algorithms [75,84]. These were mostly tailored to specific application domains, and they lacked integration and successful deployment with an HPC or cloud environment, as well as the inclusion of scheduling systems such as Slurm [3]. Therefore, there has been significant advancement in the field of sustainability in multidisciplinary application domains within HPC environments. Researchers and developers have the option of choosing between areas related to their expertise and the task at hand according to their preferences. A collection of AutoML tools and toolkits is available, leveraging a myriad of technologies [85]. Common AutoML tools include AutoKeras [86], Auto-sklearn [87], H20 [88], TPOT [89], TensorFlow [90], Auto-PyTorch [91], and others [12,85]. Furthermore, profiling and benchmark tools are available to evaluate and optimize algorithms in both CPU and GPU architectures, e.g., COCO (COmparing Continuous Optimizers) [92,93], IOH [94,95], jMetal [84,96], irace [97], Optuna [98], BIAS [72], and, last but not least, Ray Tune [99]. Despite their optimization capabilities, none of the listed AutoML frameworks provide native support for energy efficiency [3]. Therefore, external tools such as PyJoules [100], Carbontracker [101], and others can be used for measuring, logging, and  tracking energy consumption and C O 2 emissions during execution [102]. However, Slurm incorporates MODA to gather energy data from hardware interfaces such as RAPL, IPMI, NVML, and others [13]. This data could be used in optimization processes, as part of objective function constraints, or to provide feedback for early stopping when limits are exceeded [9,29]. Taking this into account, AutoML could optimize ML model performance and provide a balance between ML accuracy and energy efficiency, as a step toward sustainability [1].

2.6. Image Datasets

The Canadian Institute for Advanced Research (CIFAR) dataset collection includes publicly available (https://www.cs.toronto.edu/∼kriz/cifar.html, accessed on 19 November 2024) datasets such as CIFAR10 and CIFAR100 [27]. The collection contains 60,000 colored images (32 × 32 × 3 pixels), of which 50,000 are for training and 10,000 are for testing; CIFAR10 contains 10 classes, while CIFAR100 has 100 [27].
The Modified National Institute of Standards and Technology (MNIST) database is a large publicly available (https://yann.lecun.com/exdb/mnist/, accessed on 19 November 2024) database that includes grayscale centered images of handwritten digits [103]. The database contains images of 28 × 28 pixels, comprising 60,000 training images and 10,000 testing images [103]. Furthermore, additional versions have been published, such as Extended MNIST (EMNIST) and Fashion MNISH, which contains 70,000 images (28 × 28 pixels); other customized collections are also available [103].
ImageNet is an image database collection that is publicly available online at address https://www.image-net.org/, accessed on 19 November 2024, and it includes colored images (64 × 64 × 3 pixels) in 1000 object classes, comprising 1,281,167 training images, 50,000 validation images, and 100,000 testing images [104]. In addition, versions with more or fewer images have also been published subsequently.

2.7. Checkpoint and Restart

Checkpoint and Restart (CR) is a mechanism that allows the running workload to be saved [105]. The saved checkpoint may be later restarted, resulting in restoration, a potentially reduced elapsed time, less consumed energy, and the possibility of debugging issues that may appear during the execution of a workload [105,106]. This mechanism provides fault tolerance and resilience to the interruption of hardware, networks, and software, as well as other potential issues [105,107]. CR may be achieved through various technologies on CPU and GPU architectures, including containerized environments [108,109]. Taking EE into account and the additional storage that is needed to store captured checkpoints, if these are properly applied, users may restore their work in the event of a crash, failure, or wall time limitation, saving potential lost time due to redundant computation and power consumption [106,108]. Additionally, users can migrate their workloads to less consuming nodes (additional partitions or constraints), save states before nodes go into sleep or idle mode at higher energy peaks, and restore when necessary, thus saving the state and throttling the CPU/GPU frequencies using dynamic voltage and frequency scaling (DVFS), restoring with lower frequencies, performing cloud bursting, and more [106,108]. While this mechanism is often sufficient, this is not the case for randomized environments and long-running jobs if everything is not captured correctly [105]. Furthermore, this additionally brings increased complexity and potential challenges, resulting in incomplete, inconsistent, and non-deterministic captured states; scaling may introduce overhead (GPUs), divergence from an original state, violations, and misled reproducibility. External systems may behave differently upon restart, in addition to other issues [105,110,111]. We can apply CR inside the basic DE loop presented in Algorithm 1, where we can continue from the last successfully calculated generation [105]. To ensure complete reproducibility, it is necessary to save after a certain number of generation cycles (10) [105]. The checkpoint must contain the current number of generations itself; we also save the current population number, fitness values, and configurations, such as seeds [105].
Algorithm 1 Differential Evolution for Machine Learning Hyperparameter Optimization
Require: ML hyperparameter optimization problem fitness function f MLHypO MODA , minimum x min and maximum x max of the search space of ML hyperparameters S for function f MLHypO MODA , DE parameters: population size NP , mutation differential weight F, crossover rate C R , number of generations G.
1:
Initialize population P 0 = { x i ( 0 ) } i = 1 NP using
x i , j ( 0 ) = x min , j + rand [ 0 , 1 ] ; i , j · ( x max , j x min , j )
2:
for  t = 1 to G do
3:
    for each individual x i ( t 1 ) P t 1  in parallel do
4:
        Mutation: Generate the donor vector v i ( t ) using one of the following DE strategies:
  • DE/rand/1, see Equation (8),
  • DE/best/1, see Equation (9),
  • DE/current-to-best/1, see Equation (10),
  • DE/rand/2, see Equation (11),
  • DE/best/2, see Equation (12).
5:
        Crossover: Generate the trial vector u i ( t ) using a binomial or exponential scheme:
u i , j ( t ) = v i , j ( t ) if j = K or rand i , j [ 0 , 1 ] C R x i , j ( t 1 ) otherwise
6:
        Selection: Evaluate f ( u i ( t ) ) and update:
x i ( t ) = u i ( t ) if f MLHypO MODA ( u i ( t ) ) f MLHypO MODA ( x i ( t 1 ) ) x i ( t 1 ) otherwise
7:
    end for
8:
end for
9:
return Best solution x best found over generations

2.8. Predictive Modeling

Resource allocation in HPC, such as the allocation of time, memory, central processing unit (CPU) cores, GPUs, and other parameters, is submitted through Slurm [13]. Limited knowledge in a research field, optimization, and beyond results in the underestimation of the required resources, which leads to job failure related to wall time or a lack of memory (OOM), resulting in a waste of computational resources, the consumption of the project quota for the allocation of resources, and the postponement of scientific research [2,11,13]. Furthermore, an ML model can be used for the prediction of consumed resources, such as time and energy, based on historical data that can be gathered from the Slurm accounting database through sacct command [2,13]. A prediction model is a step towards energy efficiency, sustainability, and the prevention of job failures; predictions can be based on historical data and fair shares [13]. The job size, age, and failure of some jobs and the consumed resources may cause a fair share score of a user to decrease [13]. Therefore, users with a higher fair share score receive a higher priority in the queue [13]. This results in reduced resources consumption with the lower queue time due to fair share policies [13]. Chu et al. presented the emerging challenges in generic and ML workloads within HPC environments and their correlations with job failures, energy consumption, and other analytics [2,13]. Such solutions based on CPU and GPU workflows have already been presented in other articles [112,113,114].

3. Proposed Methodology: AutoDEHypO

This section presents the proposed methodology and deployment based on AutoDEHypO, a workflow specifically designed for energy efficiency and operational data analytics in HPC environments.

3.1. Experimental Environment

Computing nodes on an HPC Vega partition with graphics accelerators (GPUs) were used [11]. The partition has 60 nodes, each node has 4× NVIDIA Ampere A100, 2× AMD Rome 7H12, 512 GB RAM, 2× HDR dual-port mezzanines, and 1× 1.92TB M.2 SSD [11]. Red Hat Enterprise Linux 8.10 OS, Slurm 24.05.5 Workload Manager, SingularityPRO version 4.1.6, NVIDIA driver 565.57.01, and CUDA 12.7 were installed on the computing nodes [11]. Additionally, a containerized environment based on Pytorch version 2.1.2 and the required libraries [91]. Training and evaluation were conducted on the publicly available CIFAR10 and CIFAR100 datasets [27]. Storage and dataset access were managed through large-capacity storage based on Ceph [11]. Due to the constraints within HPC environments, the execution time of a single run (i.e., 300 times calling the function f MODA AutoDEHypO ) is such that all times together (the total time of all observations) do not exceed the allocated allocation. So far, we have not limited energy consumption but only monitored it. Additional details of the experimental environment can be found in Table 1.

3.2. AutoDEHypO

We prepared our environment for the deployment of the AutoDEHypO workflow using supervised machine learning through pre-trained ML models (ResNet18, VGG11, ConvNeXtSmall, and DenseNet121) that are already available within the PyTorch framework; this included building a custom Singularity container for the PyTorch [91] framework with the required libraries. The composed code in the Python programming language takes care of loading a dataset and setting up a Distributed Data Parallel (DDP) that facilitates model parallelization and distribution, while the NVIDIA Collective Communications Library (NCCL) is used in the training phase for faster and more efficient inter-node back-end communication among multiple GPUs, as this ensures and enables efficient scaling among multiple GPUs. A set of methods was used to prepare, develop, and execute AutoDEHypO, including classification, hyperparameter optimization, metric evaluation, resource monitoring, and aggregated statistical analysis of the experimental results.

3.3. Differential-Evolution-Based Hyperparameter Optimization

A basic framework of the DE Algorithm 1 was used from a recent implementation that supports parallelization [15] for the optimization of hyperparameters, as it establishes a basis while minimizing complexity and allows future modifications and improvements as the experiment unfolds [21].

3.4. Job Scheduling, Training, Evaluation, and Visualization

The jobs are submitted through Slurm workload manager, as seen in Figure 2 [11,13]. The input SBATCH contains the required resources within the SBATCH script, the PyTorch container is invoked without modifying the underlying PyTorch code, and  the code written in Python is executed. An example of the SBATCH script that can be wrapped in a conditional loop ( 1 N ) in the command line interface (CLI) or script can be seen in Figure 3.
Furthermore, the workflow is used for training, evaluation, metrics, the storage of data in the appropriate data frame, and graph plotting based on newly acquired data, as presented in Figure 4. Finally, the dataset is loaded, and the basic steps of DE presented in Figure 1 are performed to evaluate and return the most suitable parameters for the training and evaluation phase. When the job is completed, we evaluate the performance metrics, which are saved in a data frame along with visualizations. Thus, we wanted to check the adequacy of the optimization of the hyperparameter space; i.e., we examined the number of epochs and iterations, weights, learning rate (LR), batch size, and optimizers. Given that it has been implemented and deployed, the AutoDEHypO workflow leverages the potential of utilizing multiple GPUs to run DE fitness functions for ML, as presented in Figure 4. In the evaluation phase, basic ML metrics were used (Section 2.1). To obtain, measure, and compare the utilization of computer resources of individual jobs, such as the elapsed time and energy consumption, data from Slurm were used [13].

3.5. Checkpoint and Restart, Collected Logs, and Fault Tolerance

Logging of the standard output and standard error output is enabled and generated in the event of an error. Email notifications are also set up to provide information on if a job is queued, started, completed, or failed [13]. This is important for certain cases where anomalies occur during initialization, such as when jobs need to be handled separately. The output files can be taken into account within the AutoDEHypO runtime, which can detect failures in settings, such as issues with NCCL or undetected GPUs within the initialization phase. One can look at these detections, from setting failures to the fact that CR uses feedback mechanisms to check which runs may need to be restarted (Figure 3 line 27). Furthermore, at the beginning of this experiment, we did not know how many resources were needed, and with the CR mechanism, we could accordingly adjust the input Slurm parameters and the need for the computational resource requirements, such as the job state, consumed energy, memory allocation, number of nodes, number of cores, number of cores per node, number of cores per CPU, number of tasks, number of GPUs, number of GPUs per node, and others. The Slurm command requeue allows the resubmission of a failed job when we can avoid and exclude any problematic nodes or, in the opposite case, when we do not want the submission, and a manual check is required after a certain number of—for example, 2—unsuccessful restarts; then, we can use the opposite command: no-requeue. This allows us to detect such errors in certain cases and perform an automatic restart of the job. Manual intervention is still required for cross-checking, and, if necessary, the job is resubmitted to the cluster queue, as shown in Figure 4. Due to the limitations of the experiment, the elapsed time and utilization of computer resources also needed to be taken into account.

4. Experimental Results

The experiments are divided into two phases of benchmarking. Within the first phase, we examine the efficiency of ML models and their parameters. The latter is then used in the second phase as a continuation of the experiments, to which we apply an analysis of the DE mutation strategies presented in Equations (8)–(12). The experiment had to be limited due to resource constraints such as time and the quota of computation resources granted within the development project on the largest Slovenian supercomputer, EuroHPC Vega. The ML models ResNet18, VGG11, ConvNeXtSmall, and DenseNet121 were used with public datasets. For the datasets, we chose CIFAR10 and CIFAR100, as they contain smaller images, thus providing a smaller size than other datasets, allowing them to be included, preprocessed, and trained [27]. The energy consumption is presented in Mega Joules (MJ) [13]. In order to obtain an assessment of the performance of the ML models and algorithms, aggregated statistics were calculated for the obtained results.
Taking into account that resource allocation may impact the obtained results, we ensured that all submitted job scripts had the same resource allocation [2]. Although it would have been possible to restrict our runs to a specific node using an additional Slurm option, this is not the intended use of HPC and was, hence, kept as a constraint. Therefore, jobs and runs were freely distributed across different nodes. During the initiation and execution of the experiment, we faced a few challenges on the software level that we could not avoid encountering, such as the wall time, which was set within the startup script, being too short, as well as a few others. The experiment was also run on the NVIDIA drivers and the Compute Unified Device Architecture (CUDA) Toolkit, which were updated at the beginning of the experiment, as we encountered a set of GPU nodes with incorrect configurations, which were resolved immediately, as well as hardware failures. This necessitated the replacement of DIMM modules and network cards, as well as the cleaning and resetting of network cards, GPUs, and other devices. Hardware problems in which GPUs are not available or are undetected lead to a failure of a large set of submitted jobs in less than 30 s on critical nodes due to failed internal communication through the NCCL back-end. In the event of such an error, a standard error output was generated, and an email notification was successfully received. Furthermore, the deployed CR was included as a common checkpoint for the ML model, therefore demonstrating how we saved important time with AutoDEHypO, with which we enabled resubmission. We received data on failed jobs, such as the time stamp, job state, job name, and job ID, that needed to be resubmitted. An example of the standard error output in the event of a runtime error is given in the following.
  • RuntimeError:
  • ProcessGroupNCCL is only supported with GPUs, no GPUs found!
The results obtained from the ResNet18, VGG11, ConvNeXtSmall, and DenseNet121 ML models on the CIFAR10 dataset are listed in Table 2, Table 3, Table 4 and Table 5, respectively. These tables present the batch size used and the results obtained, including the maximum achieved accuracy on the test batch, the best learning rate found, the best accuracy achieved, the CPU time consumed, the elapsed time, and the energy consumed in a Slurm job. Figure 5 provides an example of a subset for the ML performance of the ResNet18, VGG11, ConvNeXtSmall, and DenseNet121 ML models on the CIFAR10 dataset, and Figure 6 does so for the CIFAR100 dataset.
The results obtained from the ML models on the CIFAR100 dataset in the first initial phase are presented in Table 6 for ResNet18, for VGG11 in Table 7, for ConvNeXtSmall in Table 8, and for DenseNet121 in Table 9, respectively.

4.1. Obtained Results

Based on the reported data provided and using our methodology, we observed results in Table 2, Table 3, Table 4 and Table 5 for CIFAR10 and in Table 6, Table 7, Table 8 and Table 9. These tables show a minimum achieved accuracy on test trial and their corresponding impact, highlighting their significance, where the lowest observed accuracy is 62.56%, which corresponds to the calculated ML accuracy of 0.1724. The highest achieved accuracy is 99.17%, with the calculated ML accuracy of 0.19916. Furthermore, for consumed energy metric, we obtained a minimum reported value of 4.43 M (ResNet18) and a maximum value of 59.36 M (DenseNet121). We also observed the elapsed time metric, ranging from the fastest completed job of 03:38:37, to the longest job with wall time 2 d 00:00:17 that exceeded the maximum allowed job execution time within the submitted partition. The results obtained in the first phase of the experiment possibly indicated that using a batch size of 256 in a set of ML models produced the most suitable and efficient results.
We observed that the DenseNet121 model possibly consumed more energy on both datasets than the less complex ML models, such as ResNet18, as presented in Figure 5 and Figure 6. Individual ML models in combination with certain DE strategies possibly performed better and more consistently, while some possibly consumed more energy, e.g., with the exponential DE strategies applied to DenseNet121 or ResNet18, and vice versa, with binomial models possibly consuming more resources in ConvNeXSmall and VGG11. The second phase of the experiment proceeded with a performance comparison of the binomial and exponential DE strategies with each ML model. In this phase, 10 independent runs were executed, as this could more generally determine whether there were significant differences DE-strategy-wise or ML-model-wise, and confirm if there was an impact on key metrics. A mutation was applied randomly with a factor (F) presented in Table 1. The population size (P) and the maximum possible number of generations (G) were appropriately determined due to the project allocation and resource constraints, otherwise we would exceed the current allowable resource consumption within the project allocation. The convergences of metric of accuracy is plotted unified through runtimes in Figure 7 for those runs that obtained median accuracy among each of the 10 independent runs of a DE strategy for a ML model. As observed, the effects of the DE strategies DE/rand/1/bin, DE/best/1/bin, DE/current-to-best/1/bin, DE/rand/2/bin, DE/best/2/bin, as well as the exponential strategies, such as DE/rand/1/exp, DE/best/1/exp, DE/current-to-best/1/exp, DE/rand/2/exp, and DE/best/2/exp [55], possibly varies across different ML model architectures, such as ResNet18, VGG11, ConvNeXtSmall, and DenseNet121, and the CIFAR10 and CIFAR100 datasets [27]. As the results of the tables indicate, H 0 is rejected because at least one DE strategy shows a significant difference, and specific DE strategies may even perform better. Therefore, to analyse which and for how much overall when aggregated, we discuss in the next subsection.
Figure 8 shows the statistical measure of the mean for each metric in the initial phase, and it is divided into three subplots (a) elapsed time, (b) consumed energy, and (c) accuracy for the DE strategies; these values were measured in 10 runs on the CIFAR10 dataset and are grouped by ML model. Figure 9 shows the same means of the selected DE strategies measured on the CIFAR100 dataset. Moreover, if we have a well-tuned ML model with appropriate weights, and the hyperparameters are poorly tuned, we possibly achieve worse ML performance over key metrics. Furthermore, the results obtained from the ResNet18, VGG11, ConvNeXtSmall, and DenseNet121 ML models with a batch size of 256 on both datasets show that, possibly, more efficient DE strategies consumed fewer resources, as the longer execution time possibly resulted in an increase the consumption of resources, and vice versa. A shorter execution time possibly resulted in lower resource consumption. The results on the CIFAR10 dataset are presented in Table A1, Table A2, Table A3 and Table A4, while the results on the CIFAR100 dataset are presented in Table A5, Table A6, Table A7 and Table A8.

4.2. Discussion of the Aggregated Statistics

Aggregate statistical analysis using non-parametric Friedman tests and corresponding procedures for the computational results of the DE strategies was performed on the ResNet18, VGG11, ConvNeXtSmall, and DenseNet121 ML models and on the CIFAR10 and CIFAR100 datasets [27,115]. Control tests were performed on the metrics of elapsed time, consumed energy, and accuracy using custom extraction scripts and publicly available code for statistics [116]. The p-value threshold was set to 0.05, and significant differences were detected and marked (†). This analysis was conducted to determine whether the DE strategies, ML models, and their architectures or a combination thereof had a statistically significant impact and if there were significant differences in key metrics for the computational efficiency. The results of the statistical analysis are presented in Table 10, Table 11, Table 12, Table 13, Table 14, Table 15, Table 16, Table 17 and Table 18.
Table 10 presents a statistical analysis of the elapsed time in the DE strategies using a non-parametric Friedman test, where the maximum rank highlights the advantage across the DE strategies, along with the corresponding post hoc procedures. The elapsed time in the DE strategies is significantly better than that of rand2exp on CIFAR10 when using AutoDEHypO according to the post hoc procedures of Holm, Hochberg, Hommel, Holland, and Finner. The Rom procedure shows a significant difference across DE strategies in comparison with rand2exp, with the exception of rand1exp. The Li procedure shows a significant difference across DE strategies in comparison with rand2exp, with the exception of rand2bin. Therefore, AutoDEHypO suggests that on CIFAR10, with the metric of elapsed time, the DE strategies best1bin, best1exp, currenttobest1bin, currenttobest1exp, best2exp, best2bin, rand1bin, rand1exp, and rand2bin are more suitable than rand2exp.
Table 11 presents a statistical analysis of the consumed energy in DE strategies using the non-parametric Friedman test, where the minimum rank highlights the advantage across DE strategies and corresponding post hoc procedures. The consumed energy in the DE strategies rand2bin, rand2exp, best2bin, rand1bin, best2exp, rand1exp, and currenttobest1exp is significantly better than in best1bin on CIFAR10 with AutoDEHypO according to the post hoc procedures of Holm, Hochberg, Hommel, Holland, and Rom. The Finner procedure shows that the DE strategies rand2bin, rand2exp, and best2bin are significantly better than rand2bin. Furthermore, rand1exp is significantly better than rand2bin according to the Holm, Hochberg, Hommel, and  Holland procedures. The Li procedure shows significant difference across DE strategies in comparison with best1bin, with the exception of best1exp. Therefore, AutoDEHypO suggests that on CIFAR10, for the metric of consumed energy, the DE strategies rand2bin, rand2exp, best2bin, rand1bin, best2exp, rand1exp, and currenttobest1exp are more suitable than currenttobest1bin, best1exp, and best1bin.
Table 12 presents a statistical analysis of the accuracy in the DE strategies using the non-parametric Friedman test, where the highest rank highlights the advantage across DE strategies and corresponding post hoc procedures. The accuracy in the DE strategy rand1exp is significantly better than that in best1exp on CIFAR10 with AutoDEHypO according to the post hoc procedures of Holm, Holland, and Finner. Therefore, AutoDEHypO suggests that on CIFAR10, for the metric of accuracy, the DE strategy rand1exp is more suitable than best1exp, rand2bin, rand1bin, currenttobest1bin, rand2exp, best2bin, currenttobest1exp, best1bin, and best2exp.
Table 13 presents a statistical analysis of the elapsed time in the DE strategies using the non-parametric Friedman test, where the maximum rank highlights the advantage across DE strategies and corresponding post hoc procedures. The elapsed time in the DE strategy best1exp is significantly better than that in rand2bin on CIFAR100 with AutoDEHypO according to the post hoc procedures of Holm, Holland, and Finner. The Li procedure shows significant differences across DE strategies in comparison with rand2bin, with the exception of rand2exp. Therefore, AutoDEHypO suggests that on CIFAR100, for the metric of elapsed time, the DE strategy best1exp is more suitable than rand2bin, best2bin, best1bin, currenttobest1bin, currenttobest1exp, rand1bin, rand1exp, best2exp, and rand2exp.
Table 14 presents a statistical analysis of the energy consumed in the DE strategies using the non-parametric Friedman test, where the minimum rank highlights the advantage across DE strategies and corresponding post hoc procedures. The energy consumed in the DE strategy rand2bin is significantly better than that in best1bin on CIFAR100 with AutoDEHypO according to the post hoc procedures of Holm, Hommel, Holland, and Finner. The Li procedure shows significant differences across DE strategies in comparison with best1bin, with the exception of best1exp. Therefore, AutoDEHypO suggests that on CIFAR100, for the metric of consumed energy, the DE strategy rand2bin is more suitable than best1bin, rand1bin, rand2exp, best2exp, currenttobest1exp, rand1exp, currenttobest1bin, best2bin, and best1exp.
Table 15 presents a statistical analysis of the accuracy in the DE strategies using the non-parametric Friedman test, where the highest rank highlights the advantage across DE strategies and corresponding post hoc procedures. The accuracy in the DE strategy rand1bin is significantly better than that in rand1exp on CIFAR100 with AutoDEHypO according to the post hoc procedures of Holm, Hommel, Holland, Rom and Finner. The DE strategy best2bin is significantly better than rand1exp according to the procedures of Holm, Hommel, and Holland. The Li procedure shows significant differences across DE strategies in comparison with rand1exp, with the exception of best2bin. Therefore, AutoDEHypO suggests that on CIFAR100, for the metric of accuracy, the DE strategies rand1bin and rand2exp are more suitable than rand1exp, currenttobest1exp, currenttobest1bin, rand2bin, best1bin, best2exp, best1exp, and best2bin.
Table 16 presents a statistical analysis of the elapsed time in the DE strategies using the non-parametric Friedman test, where the maximum rank highlights the advantage across DE strategies and corresponding post hoc procedures. The elapsed time in the DE strategies best1exp and best1bin is significantly better than that in rand2bin on CIFAR10 and CIFAR100 with AutoDEHypO according to the post hoc procedures of Holm, Hochberg, Holland, Rom, and Finner. Additionally, the DE strategies currenttobest1bin, best2bin, and currenttobest1exp are significantly better than rand2bin according to the procedures of Holm, Hochberg, Hommel, Holland, and Rom. best2exp is significantly better than rand2bin according to the procedures of Holm, Hochberg, Hommel, and Holland. The Li procedure shows significant differences across DE strategies in comparison with rand2bin, with the exception of rand2exp. Therefore, AutoDEHypO suggests that on CIFAR10 and CIFAR100, for the metric of elapsed time, the DE strategies best1exp, best1bin, currenttobest1bin, best2bin, currenttobest1exp, and best2exp are more suitable than rand1bin, rand1exp, and rand2exp.
Table 17 presents a statistical analysis of the energy consumed in the DE strategies using the non-parametric Friedman test, where the maximum rank highlights the advantage across DE strategies and corresponding post hoc procedures. The  energy consumed in the DE strategies best1bin, best1exp, currenttobest1bin, currenttobest1exp, best2bin, rand1exp, and best2exp is significantly better than that in rand2bin on CIFAR10 and CIFAR100 with AutoDEHypO according to the post hoc procedures of Holm, Hochberg, Hommel, Holland, and Rom. According to Finner’s post hoc procedure, the DE strategies best1bin, best1exp, and currenttobest1bin are significantly better than rand2bin. The Li procedure shows significant differences across DE strategies in comparison with rand2bin, with the exception of rand2exp. Therefore, AutoDEHypO suggests that on CIFAR10 and CIFAR100, for the metric of consumed energy, the DE strategies best1bin, best1exp, currenttobest1bin, currenttobest1exp, best2bin, rand1exp, and best2exp are more suitable than rand2bin, rand1bin, and rand2exp.
Table 18 presents a statistical analysis of the accuracy in the DE strategies using the non-parametric Friedman test, where the highest rank highlights the advantage across DE strategies and corresponding post hoc procedures. The accuracy in the DE strategy rand1bin is significantly better than that in best1exp on CIFAR10 and CIFAR100 with AutoDEHypO according to the post hoc procedures of Holm, Hommel, Holland, and Finner. The Li procedure shows significant differences across DE strategies in comparison with best1exp, with the exception of best2exp. Therefore, AutoDEHypO suggests that on CIFAR10 and CIFAR100, for the metric of accuracy, the DE strategy rand1bin is more suitable than best1exp, currenttobest1bin, rand2exp, rand2bin, currenttobest1exp, rand1exp, best1bin, best2bin, and best2exp.
As shown in Table 10, Table 11 and Table 12, according to Holm, Hochberg, Hommel, Holland, and Finner, on CIFAR10, several DE strategies perform significantly better than rand2exp, with additional confirmation from the Rom and Li procedures, except for rand2bin. The DE strategy best1exp runs significantly better than rand2bin, with confirmation from Li (with the exception of rand2exp). Furthermore, the energy consumption in the DE strategies rand2bin, rand2exp, best2bin, rand1bin, best2exp, rand1exp, and currenttobest1exp is significantly better than that in best1bin, with confirmation from Li, except for rand2exp. Table 16, Table 17 and Table 18 present results on both CIFAR10 and CIFAR100, where best1bin, best1exp, currenttobest1bin, currenttobest1exp, best2bin, rand1exp, and best2exp perform significantly better than rand2bin in terms of metrics according to the Li procedure, with exception of rand2exp. In terms of accuracy, best1exp outperforms rand1exp on CIFAR10. As shown in Table 13, Table 14 and Table 15, on CIFAR100, rand1bin achieves significantly better accuracy than that of rand1exp, with confirmation from Li, with the exception of best2bin. Across both datasets, rand1bin also demonstrates significantly better accuracy than that of best1exp, except for best2exp, according to the Li procedure.
Table 19 presents an overview of the outcomes of the statistical analysis and the aggregated number (counts) of confirmed out of 450, 184 (40.8%) significant differences in the DE strategies counting each post hoc procedure across the tree key metrics. These outcomes show and confirm significant differences in elapsed time, energy efficiency, and accuracy across the DE strategies. In addition to the elapsed time and accuracy, our operational data analytics also confirm the impact on energy efficiency when selecting DE strategies.
As ML models depend on the input data characteristics and on the computational complexity of ML architecture designs, it is expected that some ML models are more suitable. To check this, we tested the suitability of some ML models by detecting differences in elapsed time, consumed energy, and accuracy. The aggregated test outcomes are seen in Table 20, where, for all three key metrics (elapsed time, consumed energy, and accuracy), it is evident from the Friedman rankings that the ResNet18 and ConvNextSmall ML models are ranked higher than the other two ML models, DenseNet121 and VGG11. These ranks are significantly different for the accuracy metric in the case of VGG11. For the metrics of elapsed time and consumed energy, in the case of VGG11, this is also significant. Moreover, significant differences from ML model DenseNet121 are also detected, i.e., these two (DenseNet121 and VGG11) require significantly more energy and time than ResNet18 and ConvNextSmall.
Furthermore, significant differences between four ML models for 10 DE strategies (denoted as k) evaluations per each run result in 20 combinations (denoted as N), with critical value q α of 2.394 from the the two-tailed Bonferroni-Dunn test for k is studentized range statistic at the significance level α threshold of 0.05 on their Friedman average ranks. To compare their ranks we calculated critical difference (CD) [117] of approximately 0.9773 . Additionally, we calculated confidence intervals (CI) based on CD, which detected significant difference in 11 (61.11%) out of 18 pairwise comparisons. The results are presented in Table 21.

5. Conclusions and Future Work

In this article, we presented the deployment of a high-performance differential-evolution-based hyperparameter optimization workflow (AutoDEHypO) for energy efficiency and operational data analytics on multiple GPUs, where a DE algorithm with different DE strategies is demonstrated and applied in hyperparameter optimization while considering key factors such as the ML model, dataset, and DE strategy. The challenge of the optimization of ML models in HPC environments was addressed by balancing the ML accuracy, energy consumption, and resource utilization. Practical limitations such as project time constraints and the computational resources allocated to this project on a part of the national share of HPC Vega may have influenced the experiment through aspects such as node availability, high cluster utilization, shared node scheduling, elapsed time and partition (wall time) limitations, network congestion, job failures, undetected GPUs, and limited granularity in energy measurement at the node level. The utilization of computing resources had to be carefully considered, and the evaluation metrics and their scope must be taken into account. AutoDEHypO detected significant differences in the utilization of HPC resources in terms of elapsed time, energy efficiency, and ML accuracy across DE strategies. Furthermore, AutoDEHypO overcomes several limitations of existing workflows, which mostly focus on ML performance, inefficient utilization of computational resources during job allocation, and extended queue times; therefore, there is a lack of integration with HPC environments and schedulers such as Slurm and limited support for sustainability and cost-effectiveness. In addition to tracking the elapsed time and accuracy, our AutoDEHypO uses operational data analytics to determine how energy efficiently DE algorithms and DE strategies perform and confirms the impact on energy efficiency when selecting DE strategies. The analytics of DE strategies influenced the consumption of resources and indicated when the consumption of energy and other resources led to different ML performance and configurations. Furthermore, the statistical analysis of the key metrics of elapsed time, consumed energy, and accuracy demonstrated significant differences in DE strategies and ML models by using the non-parametric Friedman test and corresponding post hoc procedures for CIFAR10 and CIFAR100 datasets.
Specifically, in 10 independent runs, the DE binomial mutation strategies were applied and out of 450 comparisons, 184 (40.8%) detected significant differences between strategies. The effect of used DE strategies DE/rand/1/bin, DE/best/1/bin, DE/current-to-best/1/bin, DE/rand/2/bin, DE/best/2/bin, and exponential such as DE/rand/1/exp, DE/best/1/exp, DE/current-to-best/1/exp, DE/rand/2/exp, and DE/best/2/exp varied across differentent ML model architectures such as ResNet18, VGG11, ConvNeXtSmall and DenseNet121, on datasets CIFAR10 and CIFAR100.
Additionally, as an important outcome, the ML model comparisons show that there are 41 (91.1%) out of 45 rankings in the post hoc procedures, where AutoDEHypO successfully detected significant differences in all three key metrics in the case of the VGG11 ML model; for the metrics of elapsed time and consumed energy, there were further significant differences detected in the DenseNet121 and VGG11 models when compared with the better ML models of ConvNextSmall and ResNet18. Furthermore, the calculated confidence intervals based on critical difference of approximately 0.9773 , detected significant differences in 11 (61.11%) out of 18 pairwise comparisons.
Despite the insights and contributions provided by this article, there are a few limitations to acknowledge. This workflow is designed and optimized for deployment on a single cluster. Generalizing this workflow to another HPC environment may be challenging due to differences in architecture, scheduling policies, and energy monitoring solutions. Furthermore, within the proposed workflow, we used a basic DE algorithm implementation and limited this research to a few ML models, two datasets (CIFAR10 and CIFAR100), a single DE algorithm, 10 DE strategies, and limited runs of each epoch. While this setup was sufficient for initial exploration and experimentation, further research can be conducted. Moreover, this workflow can be adopted for other ML models, such as neural networks with adjustments in their architectures, different DE algorithms and DE strategies, population sizes, and different datasets, such as ImageNet, MNIST, and others; last but not least, more computational resources can be used. Possibilities have emerged for the evaluation and validation of other algorithms and AutoML frameworks to provide a baseline comparison with state-of-the-art AutoML frameworks and beyond. The necessary validation and comparison for a different environment could take place within an additional experiment; this could include a comparison of the final results. System malfunctions may also be resolved through fault-tolerance mechanisms such as Checkpoint and Restart (CR) without affecting the obtained results, allowing one to continue after unexpected interruptions. Additionally, we can apply CR inside the DE loop, where we can continue from the last successfully saved and calculated generation. Furthermore, a real-time feedback mechanism for adaptive and dynamic changes within ML operations has not yet been implemented and may contribute to the workflow; this may be researched in the future. The prediction model and decision making based on historical data can contribute to ML performance, optimization, and energy efficiency. These improvements will not only strengthen the practical applicability of this workflow but also contribute to sustainability, and a reduction in the environmental footprint.

Author Contributions

Conceptualization, T.P. and A.Z.; methodology, T.P. and A.Z.; software, T.P. and A.Z.; validation, T.P. and A.Z.; formal analysis, T.P. and A.Z.; investigation, T.P. and A.Z.; resources, T.P. and A.Z.; data curation, T.P. and A.Z.; writing—original draft preparation, T.P. and A.Z.; writing—review and editing, T.P. and A.Z.; visualization, T.P. and A.Z.; supervision, T.P. and A.Z.; project administration, T.P. and A.Z.; funding acquisition, T.P. and A.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded and conducted within the Individual Research Work 3 Unit of the doctoral program for Computer Science and Informatics at the University of Maribor. The fee for study enrollment was financed by IZUM—Institute of Information Science (17-2141-2023/01-ab and 17-2375-2024/01-ab).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original data are included within this paper. Further inquiries may be addressed to the corresponding author.

Acknowledgments

The authors acknowledge the EuroHPC JU, HPC RIVR and SLING consortium for allocating computing resources on the national share of HPC Vega within the Development Project (S24R08-01) hosted at the Institute of Information Science (IZUM). Authors also acknowledge the project DAPHNE (Integrated Data Analysis Pipelines for Large-Scale Data Management, HPC, and Machine Learning) funded by the European Union’s Horizon 2020 research and innovation program under grant agreement No 957407. We also acknowledge COST (European Cooperation in Science and Technology) support from COST Actions: CA22137 “Randomised Optimisation Algorithms Research Network (ROAR-NET)”. Authors acknowledge the MDPI Institutional Open Access Program (IOAP) for University of Maribor and thank editors and reviewers of this paper.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

Table A1. Results obtained from 10 runs of 15 epochs of a hyperoptimized ResNet18 using a different DE strategy on CIFAR10.
Table A1. Results obtained from 10 runs of 15 epochs of a hyperoptimized ResNet18 using a different DE strategy on CIFAR10.
RunStrategyMax Accuracy (%)Best LRBest AccuracyCPU TimeElapsed TimeConsumed Energy
Run 1rand1bin100%0.00017169553081330660.2012807:17:5601:49:292.39 MJ
Run 2rand1bin99.976%0.00017169553081330660.2017608:02:2002:00:352.63 MJ
Run 3rand1bin100%0.00017169553081330660.2012806:35:4001:38:552.11 MJ
Run 4rand1bin100%0.00017169553081330660.2017609:58:5602:29:443.12 MJ
Run 5rand1bin99.96%0.00017169553081330660.19809:54:1202:28:333.11 MJ
Run 6rand1bin100%0.00017169553081330660.2017606:41:0401:40:162.14 MJ
Run 7rand1bin100%0.00017169553081330660.2012810:13:0402:33:163.27 MJ
Run 8rand1bin99.984%0.00017169553081330660.2017607:53:1601:58:192.52 MJ
Run 9rand1bin100%0.00017169553081330660.2017607:58:5201:59:432.62 MJ
Run 10rand1bin100%0.00017169553081330660.2017607:13:2801:48:222.32 MJ
Run 1best1bin99.968%0.000132447958668769260.19808:38:4802:09:422.82 MJ
Run 2best1bin100%0.000132447958668769260.2012809:53:0402:28:163.20 MJ
Run 3best1bin99.96%0.000240772301800391150.1989606:58:3201:44:382.37 MJ
Run 4best1bin99.992%0.000132447958668769260.2017609:38:2802:24:373.18 MJ
Run 5best1bin100%0.000240772301800391150.19805:56:0001:29:001.95 MJ
Run 6best1bin100%0.00021624245414767140.1989608:18:2002:04:352.86 MJ
Run 7best1bin99.968%0.000240772301800391150.2017607:53:1601:58:192.52 MJ
Run 8best1bin99.992%0.000240772301800391150.1989610:41:5602:40:293.30 MJ
Run 9best1bin99.704%0.000240772301800391150.2017610:49:1602:42:193.47 MJ
Run 10best1bin99.936%0.000240772301800391150.2017607:17:3201:49:232.29 MJ
Run 1currenttobest1bin99.952%0.000240772301800391150.2017611:21:4002:50:253.58 MJ
Run 2currenttobest1bin99.984%0.000240772301800391150.2017609:25:3202:21:233.05 MJ
Run 3currenttobest1bin100%0.00021624245414767140.1989609:46:3602:26:393.14 MJ
Run 4currenttobest1bin100%0.000240772301800391150.2017613:04:0003:16:004.27 MJ
Run 5currenttobest1bin99.984%0.000132447958668769260.2017608:57:3202:14:232.84 MJ
Run 6currenttobest1bin100%0.000240772301800391150.1989609:10:1602:17:343.00 MJ
Run 7currenttobest1bin100%0.000240772301800391150.2012809:13:2402:18:212.95 MJ
Run 8currenttobest1bin99.952%0.000132447958668769260.2017609:28:3602:22:093.06 MJ
Run 9currenttobest1bin99.976%0.00021624245414767140.19811:38:1202:54:333.93 MJ
Run 10currenttobest1bin99.264%0.000240772301800391150.2012811:49:4002:57:253.82 MJ
Run 1rand2bin99.624%0.000240772301800391150.2017612:54:5603:13:444.40 MJ
Run 2rand2bin100%0.000240772301800391150.2017611:32:4802:53:123.73 MJ
Run 3rand2bin99.992%0.00021624245414767140.2017610:10:0402:32:313.30 MJ
Run 4rand2bin99.928%0.000240772301800391150.2012809:53:0002:28:153.21 MJ
Run 5rand2bin99.992%0.00024862432592688210.19812:27:5603:06:594.02 MJ
Run 6rand2bin100%0.00021624245414767140.19816:39:4004:09:555.48 MJ
Run 7rand2bin99.944%0.000240772301800391150.2017612:32:4003:08:104.09 MJ
Run 8rand2bin100%0.000240772301800391150.19812:17:0403:04:163.95 MJ
Run 9rand2bin100%0.000240772301800391150.2017608:38:5602:09:442.82 MJ
Run 10rand2bin100%0.000281358581111197030.2017612:44:4403:11:113.99 MJ
Run 1best2bin99.144%0.00020215828617478350.1989614:49:4403:42:264.75 MJ
Run 2best2bin100%0.00020215828617478350.2012810:21:2002:35:203.30 MJ
Run 3best2bin100%0.000240772301800391150.2017611:26:3202:51:383.62 MJ
Run 4best2bin100%0.00020215828617478350.2017610:09:4802:32:273.29 MJ
Run 5best2bin99.896%0.00014475302657031410.2012810:46:0802:41:323.54 MJ
Run 6best2bin100%0.00020215828617478350.2017608:46:4002:11:402.79 MJ
Run 7best2bin100%0.000240772301800391150.2012807:18:0801:49:322.39 MJ
Run 8best2bin100%0.000281358581111197030.2012811:21:1202:50:183.83 MJ
Run 9best2bin99.936%0.000240772301800391150.1989608:07:2002:01:502.60 MJ
Run 10best2bin100%0.000123931800743550060.2017614:49:3203:42:234.69 MJ
Run 1rand1exp99.976%0.00017169553081330660.2017607:19:3601:49:542.39 MJ
Run 2rand1exp100%0.00017169553081330660.2012807:20:2001:50:052.37 MJ
Run 3rand1exp99.984%0.00017169553081330660.19810:16:0802:34:023.24 MJ
Run 4rand1exp99.992%0.00017169553081330660.2017608:01:5602:00:292.59 MJ
Run 5rand1exp99.992%0.00017169553081330660.2012807:12:3201:48:082.33 MJ
Run 6rand1exp100%0.00017169553081330660.2017608:33:1602:08:192.79 MJ
Run 7rand1exp99.992%0.00017169553081330660.1989607:16:0801:49:022.31 MJ
Run 8rand1exp100%0.00017169553081330660.2017608:34:5602:08:442.76 MJ
Run 9rand1exp100%0.00017169553081330660.2017608:35:5602:08:592.80 MJ
Run 10rand1exp99.984%0.00017169553081330660.19810:37:2802:39:223.45 MJ
Run 1rand2exp99.92%0.000240772301800391150.2017611:20:1202:50:033.81 MJ
Run 2rand2exp99.968%0.000240772301800391150.2017612:48:4403:12:114.04 MJ
Run 3rand2exp99.92%0.000240772301800391150.2017614:57:2003:44:204.86 MJ
Run 4rand2exp99.976%0.000240772301800391150.2017613:53:5203:28:284.52 MJ
Run 5rand2exp100%0.00021624245414767140.1989611:56:0402:59:013.76 MJ
Run 6rand2exp99.808%0.000240772301800391150.19814:34:2403:38:364.54 MJ
Run 7rand2exp99.88%0.000240772301800391150.2017614:13:3603:33:244.48 MJ
Run 8rand2exp100%0.000240772301800391150.2017610:30:4802:37:423.39 MJ
Run 9rand2exp100%0.00021624245414767140.1989609:19:5602:19:592.99 MJ
Run 10rand2exp99.96%0.00021624245414767140.2017619:39:0004:54:456.27 MJ
Run 1best1exp99.976%0.000132447958668769260.2017606:38:4001:39:402.15 MJ
Run 2best1exp99.984%0.000132447958668769260.2012812:34:4403:08:414.02 MJ
Run 3best1exp99.712%0.00021624245414767140.1989611:19:1602:49:493.67 MJ
Run 4best1exp100%0.000132447958668769260.2017611:52:3602:58:093.74 MJ
Run 5best1exp100%0.000240772301800391150.1989611:31:0002:52:453.68 MJ
Run 6best1exp99.96%0.000132447958668769260.2017610:16:4402:34:113.23 MJ
Run 7best1exp99.928%0.000132447958668769260.2012810:52:2002:43:053.45 MJ
Run 8best1exp99.952%0.000132447958668769260.19811:16:3202:49:083.66 MJ
Run 9best1exp100%0.000240772301800391150.2017609:32:0802:23:023.11 MJ
Run 10best1exp99.968%0.00021624245414767140.2012808:45:1602:11:192.76 MJ
Run 1best2exp100%0.000240772301800391150.2017612:03:2803:00:523.84 MJ
Run 2best2exp100%0.00020215828617478350.2017609:13:4002:18:253.20 MJ
Run 3best2exp100%0.00020215828617478350.2017610:47:1602:41:493.43 MJ
Run 4best2exp99.928%0.000210979504563787970.1989613:38:2403:24:364.24 MJ
Run 5best2exp99.584%0.000123931800743550060.2012812:44:4403:11:114.13 MJ
Run 6best2exp100%0.000123931800743550060.2017608:33:2002:08:202.71 MJ
Run 7best2exp99.968%0.000240772301800391150.2012807:16:2801:49:072.33 MJ
Run 8best2exp100%0.00020215828617478350.2017609:07:5602:16:592.90 MJ
Run 9best2exp100%0.00021624245414767140.2012813:31:3203:22:534.39 MJ
Run 10best2exp100%0.000240772301800391150.2017608:55:2802:13:522.87 MJ
Run 1currenttobest1exp100%0.000240772301800391150.1989610:41:2402:40:213.33 MJ
Run 2currenttobest1exp99.984%0.000132447958668769260.2017612:47:0403:11:464.01 MJ
Run 3currenttobest1exp99.872%0.000132447958668769260.1989609:16:4002:19:102.98 MJ
Run 4currenttobest1exp100%0.00021624245414767140.2017606:52:0401:43:012.24 MJ
Run 5currenttobest1exp100%0.00021624245414767140.2017606:27:2801:36:522.06 MJ
Run 6currenttobest1exp99.992%0.000132447958668769260.2017611:03:3202:45:533.73 MJ
Run 7currenttobest1exp99.968%0.000240772301800391150.2017607:45:1201:56:182.55 MJ
Run 8currenttobest1exp100%0.000132447958668769260.2017607:15:2801:48:522.36 MJ
Run 9currenttobest1exp99.976%0.000240772301800391150.19809:52:0002:28:003.08 MJ
Run 10currenttobest1exp99.968%0.000132447958668769260.2017607:17:2801:49:222.28 MJ
Table A2. Results obtained from 10 runs of 15 epochs of a hyperoptimized VGG11 using a different DE strategy on CIFAR10.
Table A2. Results obtained from 10 runs of 15 epochs of a hyperoptimized VGG11 using a different DE strategy on CIFAR10.
RunStrategyMax Accuracy (%)Best LRBest AccuracyCPU TimeElapsed TimeConsumed Energy
Run 1rand1bin91.136%0.00028504921786332150.194641 d 06:51:0807:42:4710.92 MJ
Run 2rand1bin89.84%0.000145334761903245030.19961 d 15:46:0409:56:3113.54 MJ
Run 3rand1bin88.904%0.000162124068907360.200082 d 08:05:2014:01:2019.72 MJ
Run 4rand1bin88.096%0.000174477153877060860.196482 d 14:03:4415:30:5621.79 MJ
Run 5rand1bin78.848%0.000132776061637921520.200081 d 17:55:4810:28:5714.62 MJ
Run 6rand1bin84.84%0.000402172631010210240.195521 d 16:39:3610:09:5414.22 MJ
Run 7rand1bin86.832%0.000137725209534367780.19961 d 13:27:4409:21:5614.32 MJ
Run 8rand1bin91.024%0.000241130783042947580.199841 d 20:35:0011:08:4515.61 MJ
Run 9rand1bin77.488%0.000139592689798095450.198482 d 07:07:0413:46:4619.14 MJ
Run 10rand1bin90.2%0.000183768490814301860.1962 d 02:35:0412:38:4618.53 MJ
Run 1best1bin88.888%0.000170846510541198340.195761 d 05:08:3207:17:0810.27 MJ
Run 2best1bin82.096%0.000149869991507051220.198641 d 07:07:0807:46:4710.74 MJ
Run 3best1bin54.912%0.00032823638086059630.197121 d 16:01:3610:00:2414.08 MJ
Run 4best1bin87.728%0.00025388759996056110.190961 d 06:27:0007:36:4510.75 MJ
Run 5best1bin86.24%0.000132447958668769260.199281 d 14:55:5209:43:5813.72 MJ
Run 6best1bin87.232%0.000187725990784069880.195441 d 10:27:2008:36:5011.88 MJ
Run 7best1bin91.024%0.00019799833108324090.197281 d 11:50:1608:57:3412.46 MJ
Run 8best1bin87.048%0.000204520793570471350.199121 d 22:41:0011:40:1516.45 MJ
Run 9best1bin86.744%0.000116978557920244460.199761 d 16:49:2410:12:2114.36 MJ
Run 10best1bin87.848%0.000118781444167737950.199282 d 01:29:0812:22:1717.48 MJ
Run 1currenttobest1bin94.408%0.00017771013478759730.198481 d 12:49:1209:12:1812.76 MJ
Run 2currenttobest1bin91.664%0.00016767751239408380.19681 d 21:12:3611:18:0915.98 MJ
Run 3currenttobest1bin87.704%0.000249274270394843360.199681 d 17:18:0810:19:3214.53 MJ
Run 4currenttobest1bin93.104%0.000125251673167103550.198641 d 10:23:4408:35:5612.22 MJ
Run 5currenttobest1bin76.968%0.000112719265752493690.20081 d 15:54:4809:58:4213.94 MJ
Run 6currenttobest1bin92.656%0.000240772301800391150.198641 d 15:43:3609:55:5414.74 MJ
Run 7currenttobest1bin83.928%0.000180056271732636770.200082 d 04:49:0013:12:1518.52 MJ
Run 8currenttobest1bin89.832%0.000180056271732636770.197281 d 14:02:0009:30:3013.37 MJ
Run 9currenttobest1bin79.304%0.000141609709395538360.200241 d 18:51:1610:42:4914.97 MJ
Run 10currenttobest1bin92.944%0.00017099711310533570.199521 d 08:37:1208:09:1811.34 MJ
Run 1rand2bin87.672%0.000140437197525316070.198642 d 09:31:2814:22:5220.78 MJ
Run 2rand2bin89.04%0.000112257164111145730.198882 d 10:33:2814:38:2220.83 MJ
Run 3rand2bin94.512%0.00021335012431247070.196242 d 02:37:0812:39:1717.67 MJ
Run 4rand2bin83.176%0.000146539418890181440.19642 d 09:09:0814:17:1719.81 MJ
Run 5rand2bin87.024%0.000119533641653748430.197282 d 05:57:2413:29:2118.39 MJ
Run 6rand2bin84.768%0.000235532893505987780.199682 d 04:24:0413:06:0117.81 MJ
Run 7rand2bin87.648%0.00022375922465134610.199361 d 17:36:3610:24:0914.43 MJ
Run 8rand2bin88.624%0.000284500768355017930.197842 d 14:21:2815:35:2221.70 MJ
Run 9rand2bin87.976%0.00013814089102834190.200082 d 01:54:1212:28:3317.34 MJ
Run 10rand2bin88.936%0.00037455290836552160.196642 d 02:32:4012:38:1017.96 MJ
Run 1best2bin87.808%0.000198024254596845550.197921 d 14:59:4809:44:5713.70 MJ
Run 2best2bin92.296%0.00012513961710271520.199441 d 12:31:3609:07:5412.92 MJ
Run 3best2bin90.4%0.00019042703378366230.19641 d 17:50:4410:27:4114.68 MJ
Run 4best2bin81.952%0.00010994762254373910.198882 d 01:55:3612:28:5417.65 MJ
Run 5best2bin75.464%0.000202918619823005050.196721 d 20:22:4811:05:4215.79 MJ
Run 6best2bin86.48%0.000100609821970013250.20042 d 10:13:3214:33:2320.22 MJ
Run 7best2bin88.752%0.000164919716241819510.199681 d 13:26:4009:21:4012.71 MJ
Run 8best2bin93.512%0.000240772301800391150.198241 d 21:12:1211:18:0316.09 MJ
Run 9best2bin89.448%0.00017147904746484410.200081 d 19:37:0010:54:1515.19 MJ
Run 10best2bin84.432%0.000144875076772761470.199282 d 13:24:0415:21:0121.30 MJ
Run 1rand1exp81.736%0.000205728364619391120.197681 d 12:40:0009:10:0012.93 MJ
Run 2rand1exp90.736%0.00012233353653173930.200241 d 13:13:2009:18:2012.97 MJ
Run 3rand1exp81.24%0.000281358581111197030.198882 d 03:22:2812:50:3718.10 MJ
Run 4rand1exp88.952%0.00017169553081330660.19882 d 01:32:4412:23:1117.78 MJ
Run 5rand1exp84.584%0.000117389803325508950.198241 d 20:13:0811:03:1715.40 MJ
Run 6rand1exp88.16%0.00017169553081330660.196882 d 12:53:4815:13:2721.12 MJ
Run 7rand1exp79.792%0.00034027337350456030.196561 d 12:26:0809:06:3212.72 MJ
Run 8rand1exp90.84%0.00018706144980548440.199681 d 22:59:5211:44:5816.24 MJ
Run 9rand1exp85.648%0.000319838608615680450.198241 d 19:27:2410:51:5115.32 MJ
Run 10rand1exp92.024%0.000180745436216547910.199681 d 22:00:3611:30:0915.91 MJ
Run 1rand2exp77.792%0.00036115292896757650.195762 d 13:49:2015:27:2021.95 MJ
Run 2rand2exp88.48%0.000178378401766325270.19882 d 03:06:4012:46:4017.93 MJ
Run 3rand2exp89.08%0.00026538470587646220.195681 d 19:58:4010:59:4014.96 MJ
Run 4rand2exp93.68%0.000125251673167103550.195682 d 09:39:2014:24:5021.03 MJ
Run 5rand2exp79.16%0.000114071407139993420.197681 d 18:18:1610:34:3414.92 MJ
Run 6rand2exp83.536%0.000173013482800916720.198562 d 02:53:2412:43:2117.94 MJ
Run 7rand2exp88.992%0.00028428807600572670.196641 d 19:29:4010:52:2515.56 MJ
Run 8rand2exp87.232%0.00037501598771371280.199522 d 15:49:4015:57:2522.13 MJ
Run 9rand2exp83.456%0.000158128240204336730.198161 d 21:09:0811:17:1716.10 MJ
Run 10rand2exp91.888%0.000177933882522167020.196241 d 22:20:5211:35:1316.15 MJ
Run 1best1exp88.696%0.000140338368546473340.197361 d 02:15:3606:33:549.31 MJ
Run 2best1exp86.968%0.00021624245414767140.197681 d 15:12:5609:48:1414.52 MJ
Run 3best1exp92.84%0.000236380949017823160.197361 d 07:52:0807:58:0211.21 MJ
Run 4best1exp86.672%0.000104271954752884740.1941 d 10:19:2808:34:5212.09 MJ
Run 5best1exp80.672%0.000194277056381570930.198881 d 15:29:5609:52:2913.53 MJ
Run 6best1exp77.904%0.000151643268009670440.197921 d 10:28:2408:37:0612.13 MJ
Run 7best1exp88.776%0.000125251673167103550.199521 d 11:42:5208:55:4312.57 MJ
Run 8best1exp86.32%0.000106351833725319330.197921 d 05:52:4007:28:1010.53 MJ
Run 9best1exp90.912%0.000164033598096865880.198321 d 13:23:4809:20:5712.89 MJ
Run 10best1exp86.616%0.000172518782329636050.198321 d 13:11:0409:17:4613.27 MJ
Run 1best2exp84.392%0.000175484195064109670.197361 d 13:47:5609:26:5913.40 MJ
Run 2best2exp85.904%0.000214385131736420150.196482 d 02:36:5612:39:1417.28 MJ
Run 3best2exp80.952%0.000106150864909526490.198641 d 15:33:1609:53:1913.77 MJ
Run 4best2exp91.496%0.00025862261362331860.195761 d 18:41:2410:40:2114.70 MJ
Run 5best2exp80.376%0.000178804635069251980.195842 d 00:10:2012:02:3516.64 MJ
Run 6best2exp75.912%0.000107198372386767010.198882 d 10:04:1614:31:0420.64 MJ
Run 7best2exp89.456%0.000185248950708825030.199281 d 23:31:0011:52:4516.85 MJ
Run 8best2exp87.856%0.000156378470833566130.195922 d 03:50:2012:57:3519.63 MJ
Run 9best2exp83.952%0.00028949860369495420.195281 d 12:47:0009:11:4513.00 MJ
Run 10best2exp90.384%0.000212672037043803580.195281 d 23:50:3611:57:3916.27 MJ
Run 1currenttobest1exp86.328%0.000161101683252510750.199841 d 15:00:1209:45:0313.84 MJ
Run 2currenttobest1exp83.032%0.00017415581518198490.19921 d 12:21:1609:05:1912.90 MJ
Run 3currenttobest1exp74.632%0.00019006815953323860.197121 d 15:17:1609:49:1914.05 MJ
Run 4currenttobest1exp94.304%0.00017378490513045970.196321 d 07:09:2407:47:2110.74 MJ
Run 5currenttobest1exp87.336%0.000339222149472093370.193841 d 12:41:5609:10:2913.80 MJ
Run 6currenttobest1exp81.44%0.000160207314299752770.198962 d 13:42:4815:25:4221.54 MJ
Run 7currenttobest1exp89.128%0.000158185662753083560.197682 d 02:56:5212:44:1318.15 MJ
Run 8currenttobest1exp88.272%0.00021787071189150580.198241 d 13:46:4009:26:4012.99 MJ
Run 9currenttobest1exp84.672%0.000170591610982738080.19922 d 06:39:4413:39:5618.98 MJ
Run 10currenttobest1exp77.736%0.000210962963167739990.199362 d 00:41:4012:10:2516.82 MJ
Table A3. Results obtained from 10 runs of 15 epochs of a hyperoptimized ConvNeXtSmall using a different DE strategy on CIFAR10.
Table A3. Results obtained from 10 runs of 15 epochs of a hyperoptimized ConvNeXtSmall using a different DE strategy on CIFAR10.
RunStrategyMax Accuracy (%)Best LRBest AccuracyCPU TimeElapsed TimeConsumed Energy
Run 1rand1bin95.976%0.00066540300794917370.201216:28:4004:07:105.76 MJ
Run 2rand1bin98.152%0.00055037828142615850.2006413:47:4403:26:564.93 MJ
Run 3rand1bin98.12%0.00026727959265821790.1974418:00:0404:30:016.28 MJ
Run 4rand1bin97.648%0.00055037828142615850.1999217:16:4004:19:106.12 MJ
Run 5rand1bin98.104%0.00021624245414767140.1972815:15:1603:48:495.21 MJ
Run 6rand1bin98.112%0.00054736094255330860.1971219:51:4404:57:567.05 MJ
Run 7rand1bin98.832%0.00060820149663338530.1999218:50:0004:42:306.60 MJ
Run 8rand1bin97.464%0.000348714254562946970.1997620:47:3605:11:547.69 MJ
Run 9rand1bin96.952%0.00053553739028156650.1970419:18:5204:49:437.12 MJ
Run 10rand1bin97.312%0.00057124394950399690.1971215:20:4403:50:115.34 MJ
Run 1best1bin98.184%0.00079275041062573270.1973617:57:2804:29:226.28 MJ
Run 2best1bin97.912%0.00079275041062573270.2005612:20:3203:05:084.39 MJ
Run 3best1bin97.752%0.00086789286879177020.197216:14:1604:03:345.78 MJ
Run 4best1bin97.384%0.00086789286879177020.1998418:27:0404:36:466.55 MJ
Run 5best1bin97.656%0.00066540300794917370.215:18:4403:49:415.40 MJ
Run 6best1bin98.496%0.00066540300794917370.2008815:22:5203:50:435.35 MJ
Run 7best1bin95.768%0.000338795628054039460.2007217:17:5604:19:296.06 MJ
Run 8best1bin97.44%0.00086789286879177020.1996816:44:5204:11:135.80 MJ
Run 9best1bin97.592%0.00066540300794917370.2008816:20:4804:05:125.62 MJ
Run 10best1bin98.184%0.00055037828142615850.1983218:05:0004:31:156.62 MJ
Run 1currenttobest1bin98.576%0.00077347680313113630.2007215:54:3203:58:385.87 MJ
Run 2currenttobest1bin97.408%0.00057397123233140840.1993616:11:0804:02:475.49 MJ
Run 3currenttobest1bin98.264%0.00079010804734894020.2008820:10:4005:02:407.23 MJ
Run 4currenttobest1bin97.288%0.000281358581111197030.2001616:00:5204:00:135.54 MJ
Run 5currenttobest1bin97.72%0.000348714254562946970.1977615:13:0403:48:165.21 MJ
Run 6currenttobest1bin98.824%0.00055037828142615850.1975213:17:3603:19:244.75 MJ
Run 7currenttobest1bin98.544%0.00038626083410303430.201213:04:4803:16:124.70 MJ
Run 8currenttobest1bin97.24%0.00039063006568036010.1997615:54:0803:58:325.59 MJ
Run 9currenttobest1bin98.752%0.00089253266255637010.2008814:28:5203:37:135.24 MJ
Run 10currenttobest1bin97.344%0.00079010804734894020.2010414:14:0003:33:304.25 MJ
Run 1rand2bin97.64%0.000281358581111197030.2011216:13:3604:03:245.71 MJ
Run 2rand2bin97.456%0.00079010804734894020.1970419:27:4404:51:566.88 MJ
Run 3rand2bin98.112%0.00079010804734894020.1972816:18:4404:04:415.68 MJ
Run 4rand2bin98.568%0.00075254780224504780.199216:20:1204:05:035.71 MJ
Run 5rand2bin98.84%0.00090.2012813:02:1203:15:334.58 MJ
Run 6rand2bin98.032%0.00021624245414767140.2002417:50:1604:27:346.38 MJ
Run 7rand2bin97.904%0.00066540300794917370.196814:55:0403:43:465.28 MJ
Run 8rand2bin97.688%0.00021624245414767140.197617:33:1204:23:186.12 MJ
Run 9rand2bin98.6%0.00039584886765497010.201219:04:1204:46:036.78 MJ
Run 10rand2bin98.672%0.00055037828142615850.1974418:23:0004:35:456.55 MJ
Run 1best2bin97.984%0.00078869257385535670.2011214:59:1203:44:485.36 MJ
Run 2best2bin98.008%0.00086789286879177020.2011216:20:2004:05:055.76 MJ
Run 3best2bin98.576%0.00079010804734894020.1964815:18:5203:49:435.37 MJ
Run 4best2bin97.688%0.00066540300794917370.1979215:37:1203:54:185.47 MJ
Run 5best2bin98.2%0.00048945331834283030.1969614:56:4403:44:115.24 MJ
Run 6best2bin97.272%0.000348714254562946970.1980822:47:2005:41:508.26 MJ
Run 7best2bin97.664%0.00048945331834283030.2006413:16:4003:19:104.91 MJ
Run 8best2bin96.464%0.00087212691909229680.1973616:35:2004:08:506.26 MJ
Run 9best2bin97.6%0.00048945331834283030.1969616:07:0004:01:455.74 MJ
Run 10best2bin96.928%0.00079010804734894020.19814:11:0803:32:474.97 MJ
Run 1rand1exp98.672%0.000281358581111197030.2000817:44:2804:26:076.06 MJ
Run 2rand1exp95.208%0.00066540300794917370.1995220:37:2805:09:227.23 MJ
Run 3rand1exp98.544%0.00055037828142615850.1974415:33:0003:53:155.52 MJ
Run 4rand1exp98.424%0.000281358581111197030.1997617:59:1604:29:496.30 MJ
Run 5rand1exp98.6%0.000240772301800391150.1995217:27:3204:21:536.11 MJ
Run 6rand1exp97.984%0.000281358581111197030.1999212:34:0403:08:314.45 MJ
Run 7rand1exp98.4%0.00055037828142615850.1968820:49:4805:12:277.56 MJ
Run 8rand1exp98.184%0.00061756673364093830.219:21:2804:50:226.77 MJ
Run 9rand1exp98.224%0.00060820149663338530.1979215:44:0403:56:015.47 MJ
Run 10rand1exp98.288%0.00054729948019861060.1969617:46:5204:26:436.27 MJ
Run 1rand2exp98.928%0.00090.2006413:29:2803:22:224.74 MJ
Run 2rand2exp97.432%0.00079010804734894020.1999215:01:4003:45:255.18 MJ
Run 3rand2exp97.512%0.00055037828142615850.1975216:30:4404:07:415.77 MJ
Run 4rand2exp97.648%0.00078306096003706240.196814:53:4403:43:265.07 MJ
Run 5rand2exp98.512%0.00086789286879177020.2011216:14:4404:03:415.74 MJ
Run 6rand2exp97.48%0.00055037828142615850.196814:13:2403:33:214.98 MJ
Run 7rand2exp98.56%0.000240772301800391150.1974420:20:0405:05:017.29 MJ
Run 8rand2exp98.28%0.00036900363372548740.2006417:17:5204:19:286.07 MJ
Run 9rand2exp96.912%0.000281358581111197030.2012820:23:2805:05:527.83 MJ
Run 10rand2exp97.904%0.00069980894237103860.19815:21:2003:50:205.42 MJ
Run 1best1exp98.336%0.00079010804734894020.2002416:19:3604:04:546.11 MJ
Run 2best1exp98.168%0.00055037828142615850.1973617:31:1604:22:496.67 MJ
Run 3best1exp97.912%0.00048945331834283030.2011215:44:2803:56:075.51 MJ
Run 4best1exp98%0.00060820149663338530.197215:38:2803:54:375.59 MJ
Run 5best1exp96.976%0.00079010804734894020.2000814:57:4803:44:275.29 MJ
Run 6best1exp98.688%0.00048945331834283030.1976816:05:3204:01:235.80 MJ
Run 7best1exp98.184%0.00086789286879177020.1966414:58:4003:44:405.23 MJ
Run 8best1exp98.248%0.000348714254562946970.2008815:40:4403:55:115.54 MJ
Run 9best1exp98.224%0.00066540300794917370.197617:06:2404:16:366.13 MJ
Run 10best1exp98.44%0.0006078153283690010.1996815:13:4803:48:275.38 MJ
Run 1best2exp98.072%0.00054987225828290920.1971220:04:3605:01:097.13 MJ
Run 2best2exp98.168%0.00087103601246564220.1974418:09:2804:32:226.40 MJ
Run 3best2exp98.416%0.00070926199239390110.1975214:38:4403:39:415.13 MJ
Run 4best2exp98.56%0.000348714254562946970.2009617:38:5604:24:446.26 MJ
Run 5best2exp98.952%0.00086789286879177020.1979213:22:0803:20:324.77 MJ
Run 6best2exp97.632%0.00086789286879177020.1978414:26:1603:36:345.07 MJ
Run 7best2exp96.784%0.000240955028862571570.199616:36:1204:09:035.79 MJ
Run 8best2exp98.008%0.000281358581111197030.2009614:44:0003:41:005.32 MJ
Run 9best2exp98.112%0.00036900363372548740.2011218:54:0804:43:326.61 MJ
Run 10best2exp97.88%0.00055037828142615850.1999216:35:0404:08:465.82 MJ
Run 1currenttobest1exp97.44%0.00053553739028156650.1970417:23:0004:20:456.06 MJ
Run 2currenttobest1exp98.736%0.00054814612519747250.197613:52:3203:28:084.79 MJ
Run 3currenttobest1exp98.104%0.0003645209082058920.200419:24:4404:51:116.78 MJ
Run 4currenttobest1exp97.152%0.00048945331834283030.1967216:59:3204:14:535.96 MJ
Run 5currenttobest1exp98.104%0.000348099763122382060.1998417:24:1204:21:036.12 MJ
Run 6currenttobest1exp97.104%0.00048945331834283030.2001616:53:2404:13:215.91 MJ
Run 7currenttobest1exp98.248%0.00067575243454840060.1999213:20:0803:20:024.73 MJ
Run 8currenttobest1exp98.032%0.00055037828142615850.197217:52:3604:28:096.45 MJ
Run 9currenttobest1exp97.144%0.00060820149663338530.1970415:00:4003:45:105.23 MJ
Run 10currenttobest1exp97.888%0.00060820149663338530.1968816:37:0804:09:175.74 MJ
Table A4. Results obtained from 10 runs of 15 epochs of a hyperoptimized DenseNet121 using a different DE strategy on CIFAR10.
Table A4. Results obtained from 10 runs of 15 epochs of a hyperoptimized DenseNet121 using a different DE strategy on CIFAR10.
RunStrategyMax Accuracy (%)Best LRBest AccuracyCPU TimeElapsed TimeConsumed Energy
Run 1rand1bin91.136%0.00028504921786332150.194641 d 06:51:0807:42:4710.92 MJ
Run 2rand1bin89.84%0.000145334761903245030.19961 d 15:46:0409:56:3113.54 MJ
Run 3rand1bin88.904%0.000162124068907360.200082 d 08:05:2014:01:2019.72 MJ
Run 4rand1bin88.096%0.000174477153877060860.196482 d 14:03:4415:30:5621.79 MJ
Run 5rand1bin78.848%0.000132776061637921520.200081 d 17:55:4810:28:5714.62 MJ
Run 6rand1bin84.84%0.000402172631010210240.195521 d 16:39:3610:09:5414.22 MJ
Run 7rand1bin86.832%0.000137725209534367780.19961 d 13:27:4409:21:5614.32 MJ
Run 8rand1bin91.024%0.000241130783042947580.199841 d 20:35:0011:08:4515.61 MJ
Run 9rand1bin77.488%0.000139592689798095450.198482 d 07:07:0413:46:4619.14 MJ
Run 10rand1bin90.2%0.000183768490814301860.1962 d 02:35:0412:38:4618.53 MJ
Run 1best1bin88.888%0.000170846510541198340.195761 d 05:08:3207:17:0810.27 MJ
Run 2best1bin82.096%0.000149869991507051220.198641 d 07:07:0807:46:4710.74 MJ
Run 3best1bin54.912%0.00032823638086059630.197121 d 16:01:3610:00:2414.08 MJ
Run 4best1bin87.728%0.00025388759996056110.190961 d 06:27:0007:36:4510.75 MJ
Run 5best1bin86.24%0.000132447958668769260.199281 d 14:55:5209:43:5813.72 MJ
Run 6best1bin87.232%0.000187725990784069880.195441 d 10:27:2008:36:5011.88 MJ
Run 7best1bin91.024%0.00019799833108324090.197281 d 11:50:1608:57:3412.46 MJ
Run 8best1bin87.048%0.000204520793570471350.199121 d 22:41:0011:40:1516.45 MJ
Run 9best1bin86.744%0.000116978557920244460.199761 d 16:49:2410:12:2114.36 MJ
Run 10best1bin87.848%0.000118781444167737950.199282 d 01:29:0812:22:1717.48 MJ
Run 1currenttobest1bin94.408%0.00017771013478759730.198481 d 12:49:1209:12:1812.76 MJ
Run 2currenttobest1bin91.664%0.00016767751239408380.19681 d 21:12:3611:18:0915.98 MJ
Run 3currenttobest1bin87.704%0.000249274270394843360.199681 d 17:18:0810:19:3214.53 MJ
Run 4currenttobest1bin93.104%0.000125251673167103550.198641 d 10:23:4408:35:5612.22 MJ
Run 5currenttobest1bin76.968%0.000112719265752493690.20081 d 15:54:4809:58:4213.94 MJ
Run 6currenttobest1bin92.656%0.000240772301800391150.198641 d 15:43:3609:55:5414.74 MJ
Run 7currenttobest1bin83.928%0.000180056271732636770.200082 d 04:49:0013:12:1518.52 MJ
Run 8currenttobest1bin89.832%0.000180056271732636770.197281 d 14:02:0009:30:3013.37 MJ
Run 9currenttobest1bin79.304%0.000141609709395538360.200241 d 18:51:1610:42:4914.97 MJ
Run 10currenttobest1bin92.944%0.00017099711310533570.199521 d 08:37:1208:09:1811.34 MJ
Run 1rand2bin87.672%0.000140437197525316070.198642 d 09:31:2814:22:5220.78 MJ
Run 2rand2bin89.04%0.000112257164111145730.198882 d 10:33:2814:38:2220.83 MJ
Run 3rand2bin94.512%0.00021335012431247070.196242 d 02:37:0812:39:1717.67 MJ
Run 4rand2bin83.176%0.000146539418890181440.19642 d 09:09:0814:17:1719.81 MJ
Run 5rand2bin87.024%0.000119533641653748430.197282 d 05:57:2413:29:2118.39 MJ
Run 6rand2bin84.768%0.000235532893505987780.199682 d 04:24:0413:06:0117.81 MJ
Run 7rand2bin87.648%0.00022375922465134610.199361 d 17:36:3610:24:0914.43 MJ
Run 8rand2bin88.624%0.000284500768355017930.197842 d 14:21:2815:35:2221.70 MJ
Run 9rand2bin87.976%0.00013814089102834190.200082 d 01:54:1212:28:3317.34 MJ
Run 10rand2bin88.936%0.00037455290836552160.196642 d 02:32:4012:38:1017.96 MJ
Run 1best2bin87.808%0.000198024254596845550.197921 d 14:59:4809:44:5713.70 MJ
Run 2best2bin92.296%0.00012513961710271520.199441 d 12:31:3609:07:5412.92 MJ
Run 3best2bin90.4%0.00019042703378366230.19641 d 17:50:4410:27:4114.68 MJ
Run 4best2bin81.952%0.00010994762254373910.198882 d 01:55:3612:28:5417.65 MJ
Run 5best2bin75.464%0.000202918619823005050.196721 d 20:22:4811:05:4215.79 MJ
Run 6best2bin86.48%0.000100609821970013250.20042 d 10:13:3214:33:2320.22 MJ
Run 7best2bin88.752%0.000164919716241819510.199681 d 13:26:4009:21:4012.71 MJ
Run 8best2bin93.512%0.000240772301800391150.198241 d 21:12:1211:18:0316.09 MJ
Run 9best2bin89.448%0.00017147904746484410.200081 d 19:37:0010:54:1515.19 MJ
Run 10best2bin84.432%0.000144875076772761470.199282 d 13:24:0415:21:0121.30 MJ
Run 1rand1exp81.736%0.000205728364619391120.197681 d 12:40:0009:10:0012.93 MJ
Run 2rand1exp90.736%0.00012233353653173930.200241 d 13:13:2009:18:2012.97 MJ
Run 3rand1exp81.24%0.000281358581111197030.198882 d 03:22:2812:50:3718.10 MJ
Run 4rand1exp88.952%0.00017169553081330660.19882 d 01:32:4412:23:1117.78 MJ
Run 5rand1exp84.584%0.000117389803325508950.198241 d 20:13:0811:03:1715.40 MJ
Run 6rand1exp88.16%0.00017169553081330660.196882 d 12:53:4815:13:2721.12 MJ
Run 7rand1exp79.792%0.00034027337350456030.196561 d 12:26:0809:06:3212.72 MJ
Run 8rand1exp90.84%0.00018706144980548440.199681 d 22:59:5211:44:5816.24 MJ
Run 9rand1exp85.648%0.000319838608615680450.198241 d 19:27:2410:51:5115.32 MJ
Run 10rand1exp92.024%0.000180745436216547910.199681 d 22:00:3611:30:0915.91 MJ
Run 1rand2exp77.792%0.00036115292896757650.195762 d 13:49:2015:27:2021.95 MJ
Run 2rand2exp88.48%0.000178378401766325270.19882 d 03:06:4012:46:4017.93 MJ
Run 3rand2exp89.08%0.00026538470587646220.195681 d 19:58:4010:59:4014.96 MJ
Run 4rand2exp93.68%0.000125251673167103550.195682 d 09:39:2014:24:5021.03 MJ
Run 5rand2exp79.16%0.000114071407139993420.197681 d 18:18:1610:34:3414.92 MJ
Run 6rand2exp83.536%0.000173013482800916720.198562 d 02:53:2412:43:2117.94 MJ
Run 7rand2exp88.992%0.00028428807600572670.196641 d 19:29:4010:52:2515.56 MJ
Run 8rand2exp87.232%0.00037501598771371280.199522 d 15:49:4015:57:2522.13 MJ
Run 9rand2exp83.456%0.000158128240204336730.198161 d 21:09:0811:17:1716.10 MJ
Run 10rand2exp91.888%0.000177933882522167020.196241 d 22:20:5211:35:1316.15 MJ
Run 1best1exp88.696%0.000140338368546473340.197361 d 02:15:3606:33:549.31 MJ
Run 2best1exp86.968%0.00021624245414767140.197681 d 15:12:5609:48:1414.52 MJ
Run 3best1exp92.84%0.000236380949017823160.197361 d 07:52:0807:58:0211.21 MJ
Run 4best1exp86.672%0.000104271954752884740.1941 d 10:19:2808:34:5212.09 MJ
Run 5best1exp80.672%0.000194277056381570930.198881 d 15:29:5609:52:2913.53 MJ
Run 6best1exp77.904%0.000151643268009670440.197921 d 10:28:2408:37:0612.13 MJ
Run 7best1exp88.776%0.000125251673167103550.199521 d 11:42:5208:55:4312.57 MJ
Run 8best1exp86.32%0.000106351833725319330.197921 d 05:52:4007:28:1010.53 MJ
Run 9best1exp90.912%0.000164033598096865880.198321 d 13:23:4809:20:5712.89 MJ
Run 10best1exp86.616%0.000172518782329636050.198321 d 13:11:0409:17:4613.27 MJ
Run 1best2exp84.392%0.000175484195064109670.197361 d 13:47:5609:26:5913.40 MJ
Run 2best2exp85.904%0.000214385131736420150.196482 d 02:36:5612:39:1417.28 MJ
Run 3best2exp80.952%0.000106150864909526490.198641 d 15:33:1609:53:1913.77 MJ
Run 4best2exp91.496%0.00025862261362331860.195761 d 18:41:2410:40:2114.70 MJ
Run 5best2exp80.376%0.000178804635069251980.195842 d 00:10:2012:02:3516.64 MJ
Run 6best2exp75.912%0.000107198372386767010.198882 d 10:04:1614:31:0420.64 MJ
Run 7best2exp89.456%0.000185248950708825030.199281 d 23:31:0011:52:4516.85 MJ
Run 8best2exp87.856%0.000156378470833566130.195922 d 03:50:2012:57:3519.63 MJ
Run 9best2exp83.952%0.00028949860369495420.195281 d 12:47:0009:11:4513.00 MJ
Run 10best2exp90.384%0.000212672037043803580.195281 d 23:50:3611:57:3916.27 MJ
Run 1currenttobest1exp86.328%0.000161101683252510750.199841 d 15:00:1209:45:0313.84 MJ
Run 2currenttobest1exp83.032%0.00017415581518198490.19921 d 12:21:1609:05:1912.90 MJ
Run 3currenttobest1exp74.632%0.00019006815953323860.197121 d 15:17:1609:49:1914.05 MJ
Run 4currenttobest1exp94.304%0.00017378490513045970.196321 d 07:09:2407:47:2110.74 MJ
Run 5currenttobest1exp87.336%0.000339222149472093370.193841 d 12:41:5609:10:2913.80 MJ
Run 6currenttobest1exp81.44%0.000160207314299752770.198962 d 13:42:4815:25:4221.54 MJ
Run 7currenttobest1exp89.128%0.000158185662753083560.197682 d 02:56:5212:44:1318.15 MJ
Run 8currenttobest1exp88.272%0.00021787071189150580.198241 d 13:46:4009:26:4012.99 MJ
Run 9currenttobest1exp84.672%0.000170591610982738080.19922 d 06:39:4413:39:5618.98 MJ
Run 10currenttobest1exp77.736%0.000210962963167739990.199362 d 00:41:4012:10:2516.82 MJ
Table A5. Results obtained from 10 runs of 15 epochs of a hyperoptimized ResNet18 using a different DE strategy on CIFAR100.
Table A5. Results obtained from 10 runs of 15 epochs of a hyperoptimized ResNet18 using a different DE strategy on CIFAR100.
RunStrategyMax Accuracy (%)Best LRBest AccuracyCPU TimeElapsed TimeConsumed Energy
Run 1rand1bin86.208%0.00068984920938774130.02081 d 18:49:5610:42:2914.01 MJ
Run 2rand1bin86.912%0.00062881623854450660.02081 d 19:05:1610:46:1914.15 MJ
Run 3rand1bin88.992%0.00064835023608667560.020722 d 02:11:4812:32:5716.05 MJ
Run 4rand1bin89.6%0.00069594787236569610.020641 d 21:23:4811:20:5714.19 MJ
Run 5rand1bin84.64%0.0003269047665347310.020882 d 07:04:1613:46:0417.39 MJ
Run 6rand1bin92.368%0.00056300591480615780.019522 d 02:48:0012:42:0016.57 MJ
Run 7rand1bin86.184%0.00055037828142615850.020721 d 17:27:2410:21:5112.85 MJ
Run 8rand1bin92.056%0.000281358581111197030.020722 d 04:50:2813:12:3716.78 MJ
Run 9rand1bin85.8%0.00066486934509922390.01962 d 05:45:1213:26:1816.98 MJ
Run 10rand1bin86.008%0.000348714254562946970.020882 d 06:00:0813:30:0217.20 MJ
Run 1best1bin85.728%0.00051630825576052520.019121 d 16:27:0810:06:4713.08 MJ
Run 2best1bin92.064%0.00046707488335297150.019361 d 23:05:0411:46:1615.44 MJ
Run 3best1bin87.264%0.00062215036995484390.019121 d 08:42:5208:10:4310.45 MJ
Run 4best1bin92.472%0.00055623072116259860.019282 d 01:43:0812:25:4716.01 MJ
Run 5best1bin87.312%0.000430210192436925540.01961 d 08:56:5208:14:1310.45 MJ
Run 6best1bin91.048%0.00086201711638765460.01962 d 08:03:2414:00:5117.56 MJ
Run 7best1bin89.144%0.00057715459547612160.020641 d 07:14:3207:48:389.91 MJ
Run 8best1bin88.952%0.00063512752531727710.020721 d 16:52:0010:13:0013.30 MJ
Run 9best1bin89.512%0.00050908104365961960.019281 d 15:08:3209:47:0813.28 MJ
Run 10best1bin88.976%0.000561500864128150.020881 d 17:22:0810:20:3213.43 MJ
Run 1currenttobest1bin87.6%0.000279847067709224030.02072¸1-11:50:0408:57:3111.53 MJ
Run 2currenttobest1bin92.32%0.00072725978208878750.020721 d 16:04:4010:01:1012.72 MJ
Run 3currenttobest1bin89.832%0.00058568945543531990.019441 d 18:41:5210:40:2813.40 MJ
Run 4currenttobest1bin91.152%0.000357274261495859670.020721 d 06:32:0407:38:019.78 MJ
Run 5currenttobest1bin87.176%0.00056656143641747470.020561 d 06:41:2407:40:219.99 MJ
Run 6currenttobest1bin84.704%0.00041770380142071810.020641 d 23:06:4411:46:4115.39 MJ
Run 7currenttobest1bin84.368%0.00052308410397135560.020881 d 20:33:4811:08:2714.57 MJ
Run 8currenttobest1bin90.44%0.000337510710015804140.01961 d 15:52:2809:58:0713.47 MJ
Run 9currenttobest1bin90.448%0.00065331106809237150.01922 d 05:18:2413:19:3617.36 MJ
Run 10currenttobest1bin83.344%0.00059970525547956540.020641 d 15:49:3209:57:2312.94 MJ
Run 1rand2bin91.8%0.000459618357711377460.020641 d 21:38:0011:24:3014.84 MJ
Run 2rand2bin85.976%0.00086789286879177020.01921 d 21:09:5211:17:2814.38 MJ
Run 3rand2bin89.632%0.00051484778073373150.020722 d 00:54:2412:13:3615.97 MJ
Run 4rand2bin90.064%0.00056500986735631390.019121 d 14:39:5609:39:5912.22 MJ
Run 5rand2bin93.344%0.00042677915067381040.020562 d 05:05:2413:16:2117.06 MJ
Run 6rand2bin87.56%0.00054778425326925080.020721 d 08:54:0008:13:3010.46 MJ
Run 7rand2bin91.312%0.00069282866076394060.020881 d 15:54:2009:58:3513.12 MJ
Run 8rand2bin90.288%0.0005454219815028210.02082 d 00:31:4812:07:5715.45 MJ
Run 9rand2bin86.84%0.00055029495472942590.020482 d 00:12:4812:03:1215.34 MJ
Run 10rand2bin88.464%0.00074105918425435070.019441 d 08:31:1208:07:4810.15 MJ
Run 1best2bin84.672%0.00047618164342913910.019122 d 01:09:2812:17:2215.34 MJ
Run 2best2bin87.872%0.00066540300794917370.02081 d 21:00:5211:15:1314.19 MJ
Run 3best2bin90.928%0.00079282416520035030.020641 d 12:51:4409:12:5612.57 MJ
Run 4best2bin92.216%0.00086789286879177020.02041 d 15:06:5209:46:4312.68 MJ
Run 5best2bin87.256%0.00089073588939520290.019441 d 10:50:3208:42:3810.86 MJ
Run 6best2bin91.592%0.000481866532636078860.019682 d 03:55:2412:58:5116.20 MJ
Run 7best2bin90.336%0.00055037828142615850.018881 d 08:30:1208:07:3310.23 MJ
Run 8best2bin89.12%0.000296935324494636360.019441 d 10:44:0808:41:0210.95 MJ
Run 9best2bin90.456%0.00048945331834283030.019121 d 10:40:4408:40:1111.33 MJ
Run 10best2bin91.448%0.000425518154201565250.02082 d 03:48:2412:57:0616.85 MJ
Run 1rand1exp86.608%0.00054736094255330860.019121 d 19:23:3210:50:5313.95 MJ
Run 2rand1exp89.832%0.0007862965598365720.020722 d 01:47:1212:26:4815.72 MJ
Run 3rand1exp85.256%0.000348714254562946970.019441 d 02:00:5206:30:138.21 MJ
Run 4rand1exp88.92%0.00060820149663338530.020641 d 20:51:3211:12:5314.37 MJ
Run 5rand1exp88.168%0.000478116039122540330.020721 d 21:33:2811:23:2214.69 MJ
Run 6rand1exp84.736%0.00088935190385597660.020722 d 00:20:4412:05:1115.44 MJ
Run 7rand1exp92.112%0.00079010804734894020.019121 d 09:10:3608:17:3910.51 MJ
Run 8rand1exp91.304%0.00054381699902039260.019121 d 11:56:5608:59:1411.46 MJ
Run 9rand1exp90.808%0.00076790847924023330.020882 d 06:16:5613:34:1417.70 MJ
Run 10rand1exp87.128%0.00060026955498432180.01962 d 06:08:5613:32:1417.20 MJ
Run 1rand2exp85.656%0.00031276001793953110.020641 d 17:26:0010:21:3012.91 MJ
Run 2rand2exp90.152%0.000481866532636078860.020721 d 22:14:1211:33:3314.81 MJ
Run 3rand2exp88.856%0.00053231768520899230.01961 d 11:39:4808:54:5711.54 MJ
Run 4rand2exp88.592%0.00034054993035967780.01921 d 14:34:5209:38:4311.98 MJ
Run 5rand2exp92.376%0.00058626254858704740.018962 d 01:45:2812:26:2217.13 MJ
Run 6rand2exp90.496%0.00064518151931689590.02081 d 13:49:2009:27:2012.45 MJ
Run 7rand2exp91.048%0.000481088784996103230.020882 d 04:12:2413:03:0616.81 MJ
Run 8rand2exp83.296%0.00048945331834283030.020721 d 05:34:5607:23:449.28 MJ
Run 9rand2exp85.08%0.000281358581111197030.020722 d 04:23:3613:05:5417.13 MJ
Run 10rand2exp84.264%0.0003654904671636250.01921 d 17:33:2010:23:2013.18 MJ
Run 1best1exp93.064%0.00054337899609328560.019121 d 13:59:4409:29:5612.15 MJ
Run 2best1exp87.8%0.00054725449679144710.020641 d 21:07:4811:16:5714.13 MJ
Run 3best1exp90.168%0.00054725449679144710.020641 d 18:50:2010:42:3513.72 MJ
Run 4best1exp89.512%0.00049063267867901810.01961 d 20:49:4411:12:2614.60 MJ
Run 5best1exp89.624%0.0006999429601028720.019121 d 10:07:2408:31:5110.92 MJ
Run 6best1exp87.144%0.00059691448335751550.019681 d 19:11:5610:47:5914.06 MJ
Run 7best1exp89.528%0.0005640138513339920.02082 d 01:29:4412:22:2615.76 MJ
Run 8best1exp78.872%0.00049024433106317030.019121 d 15:10:1209:47:3313.32 MJ
Run 9best1exp82.184%0.00066540300794917370.020641 d 08:45:4008:11:2510.76 MJ
Run 10best1exp90.832%0.00066540300794917370.02081 d 19:13:0010:48:1513.44 MJ
Run 1best2exp88.608%0.000404342692977507760.019121 d 18:05:0810:31:1713.43 MJ
Run 2best2exp91.648%0.0004253692226581140.019442 d 04:18:3613:04:3917.01 MJ
Run 3best2exp93.136%0.00053553739028156650.02081 d 23:33:4811:53:2715.06 MJ
Run 4best2exp88.712%0.00055037828142615850.018881 d 15:01:1609:45:1912.12 MJ
Run 5best2exp89.272%0.00079010804734894020.020721 d 14:47:0409:41:4612.18 MJ
Run 6best2exp84.12%0.00057308288774453760.01961 d 19:36:2010:54:0513.81 MJ
Run 7best2exp88.176%0.00036900363372548740.02081 d 11:44:0008:56:0011.46 MJ
Run 8best2exp88.624%0.000439746933463485140.01961 d 16:58:4410:14:4113.09 MJ
Run 9best2exp88.912%0.00083325181170246220.019361 d 07:06:1207:46:3310.04 MJ
Run 10best2exp85.536%0.000425518154201565250.020721 d 21:11:1211:17:4814.50 MJ
Run 1currenttobest1exp87.472%0.00086789286879177020.020961 d 23:35:1611:53:4914.96 MJ
Run 2currenttobest1exp85.928%0.00051049755728025080.019681 d 19:12:1210:48:0313.85 MJ
Run 3currenttobest1exp86.992%0.00069677327913349380.02081 d 22:33:0811:38:1715.21 MJ
Run 4currenttobest1exp88.152%0.00051156463703824240.020882 d 09:01:5614:15:2918.32 MJ
Run 5currenttobest1exp81.56%0.00039927131902173740.01881 d 08:30:5208:07:4310.46 MJ
Run 6currenttobest1exp91.048%0.00084320998785448760.020721 d 20:49:1611:12:1914.58 MJ
Run 7currenttobest1exp86.016%0.00048945331834283030.020881 d 06:42:0407:40:319.97 MJ
Run 8currenttobest1exp89.528%0.00063901829972860930.020961 d 20:04:1611:01:0413.92 MJ
Run 9currenttobest1exp93.392%0.00067797662069228890.020642 d 07:44:1613:56:0417.74 MJ
Run 10currenttobest1exp86.768%0.000281358581111197030.02082 d 06:11:2013:32:5017.20 MJ
Table A6. Results obtained from 10 runs of 15 epochs of a hyperoptimized VGG11 using a different DE strategy on CIFAR100.
Table A6. Results obtained from 10 runs of 15 epochs of a hyperoptimized VGG11 using a different DE strategy on CIFAR100.
RunStrategyMax Accuracy (%)Best LRBest AccuracyCPU TimeElapsed TimeConsumed Energy
Run 1rand1bin12.28%0.00047584408423186240.011843 d 16:49:0422:12:1631.68 MJ
Run 2rand1bin3.136%0.000281981779020857960.008563 d 18:30:2022:37:3531.84 MJ
Run 3rand1bin7.8%0.0004236808876326440.007523 d 17:03:1622:15:4930.87 MJ
Run 4rand1bin8.232%0.00021356001028830450.007443 d 18:26:0822:36:3231.36 MJ
Run 5rand1bin4.248%0.00026120968493207280.007523 d 18:13:2822:33:2231.14 MJ
Run 6rand1bin8.496%0.00019484233836430410.010163 d 20:05:3623:01:2432.09 MJ
Run 7rand1bin12.176%0.000270934942469088250.012883 d 19:56:5222:59:1332.37 MJ
Run 8rand1bin9.056%0.00028046682174055390.01283 d 19:09:1622:47:1932.17 MJ
Run 9rand1bin11.696%0.000106344293670925760.008963 d 16:50:2422:12:3631.03 MJ
Run 10rand1bin11.744%0.00021670745464969620.007923 d 20:46:4023:11:4032.23 MJ
Run 1best1bin2.48%0.00031723223263937950.00763 d 18:38:4022:39:4031.27 MJ
Run 2best1bin3.616%0.00058073742617806280.008083 d 16:38:1222:09:3331.24 MJ
Run 3best1bin4.576%0.000263010632545849130.012083 d 18:53:4822:43:2732.50 MJ
Run 4best1bin8.296%0.000285398354095132850.008083 d 22:39:4023:39:5533.27 MJ
Run 5best1bin6.304%0.000154810752832934680.012723 d 21:24:3223:21:0832.89 MJ
Run 6best1bin11.888%0.000334715063466958760.008323 d 16:45:3622:11:2430.62 MJ
Run 7best1bin10.416%0.00054994020339401770.012643 d 18:37:1222:39:1831.22 MJ
Run 8best1bin12.752%0.00042439008491612080.012563 d 19:31:2822:52:5231.19 MJ
Run 9best1bin6.832%0.000257819046275936860.012643 d 19:30:5622:52:4432.49 MJ
Run 10best1bin4.792%0.00019467380685527680.010563 d 20:33:2423:08:2132.96 MJ
Run 1currenttobest1bin11.456%0.00056314873089432990.012323 d 21:13:2823:18:2232.47 MJ
Run 2currenttobest1bin10.064%0.000150893341637552920.009283 d 20:25:5623:06:2932.70 MJ
Run 3currenttobest1bin14.8%0.00028605259560629030.008083 d 20:16:4023:04:1032.45 MJ
Run 4currenttobest1bin8.448%0.000125175101793034770.009283 d 21:55:4023:28:5532.90 MJ
Run 5currenttobest1bin8.352%0.000186906802300600930.008323 d 21:26:4823:21:4233.46 MJ
Run 6currenttobest1bin10.168%0.000188579038331683560.012643 d 18:19:3622:34:5431.69 MJ
Run 7currenttobest1bin12.568%0.000496200708902690.0123 d 14:53:1221:43:1830.73 MJ
Run 8currenttobest1bin8.728%0.000220511219859552880.013283 d 17:43:0822:25:4731.10 MJ
Run 9currenttobest1bin5.608%0.000326457930426110230.007923 d 18:37:1222:39:1831.60 MJ
Run 10currenttobest1bin6.488%0.000258156328093057060.0083 d 19:09:5222:47:2831.44 MJ
Run 1rand2bin1.08%0.00060030627844375550.012243 d 16:14:2422:03:3630.24 MJ
Run 2rand2bin9.376%0.000117765020325617310.00843 d 18:53:4822:43:2733.35 MJ
Run 3rand2bin9.952%0.000169149952991629840.008643 d 21:59:2423:29:5132.14 MJ
Run 4rand2bin11.336%0.0007201894074829240.0123 d 15:40:3621:55:0930.06 MJ
Run 5rand2bin4.76%0.000159135423440856680.013363 d 20:21:1623:05:1932.07 MJ
Run 6rand2bin12.36%0.000173886158605046570.008163 d 18:23:3222:35:5331.91 MJ
Run 7rand2bin3.416%0.00018763856715490550.00843 d 18:58:5222:44:4332.16 MJ
Run 8rand2bin3.752%0.00060567780451118880.008083 d 17:12:4022:18:1031.03 MJ
Run 9rand2bin1.104%0.000299030400595174530.007923 d 19:36:5622:54:1432.24 MJ
Run 10rand2bin9.472%0.000552888650251980.008323 d 16:42:2822:10:3732.48 MJ
Run 1best2bin3.536%0.000315161163824600750.008243 d 17:27:0422:21:4631.42 MJ
Run 2best2bin1.104%0.000479453586031871270.008163 d 17:03:2022:15:5031.38 MJ
Run 3best2bin7.808%0.000470698209856876960.0123 d 18:12:2422:33:0631.44 MJ
Run 4best2bin10.816%0.000480349523965992370.011923 d 20:37:1223:09:1831.99 MJ
Run 5best2bin8.384%0.00020735302535465910.013043 d 19:33:4422:53:2633.49 MJ
Run 6best2bin9.128%0.000222108475238624160.013123 d 15:46:0821:56:3230.58 MJ
Run 7best2bin10.336%0.000238120489657507710.008243 d 18:12:3222:33:0833.00 MJ
Run 8best2bin9.632%0.000443761460339041660.012563 d 17:34:1222:23:3331.00 MJ
Run 9best2bin7.96%0.000119557229071073480.007683 d 18:27:5222:36:5831.57 MJ
Run 10best2bin9.464%0.000411281065862328170.007443 d 15:48:3221:57:0830.42 MJ
Run 1rand1exp3.744%0.000177851449535845640.00843 d 17:36:4422:24:1131.61 MJ
Run 2rand1exp6.32%0.000231228619529638840.007683 d 18:25:4422:36:2631.52 MJ
Run 3rand1exp10.728%0.000259308954328450550.008243 d 18:17:1222:34:1831.46 MJ
Run 4rand1exp7.568%0.00033383493799703290.007683 d 19:57:2822:59:2231.56 MJ
Run 5rand1exp14.32%0.00030178952201016950.007843 d 19:44:2022:56:0531.35 MJ
Run 6rand1exp10.576%0.00044010829551402030.0083 d 16:56:3222:14:0830.49 MJ
Run 7rand1exp7.424%0.000168944076535678270.008563 d 18:07:5622:31:5931.75 MJ
Run 8rand1exp2.072%0.000285823831180450740.006883 d 18:17:0422:34:1630.64 MJ
Run 9rand1exp10.528%0.000228739831528681840.007443 d 18:55:3622:43:5432.56 MJ
Run 10rand1exp9.416%0.000223626969614144230.0123 d 19:33:3622:53:2432.41 MJ
Run 1rand2exp11.736%0.00061997894061847110.0143 d 17:02:2422:15:3631.43 MJ
Run 2rand2exp9.872%0.00042791770775277270.007683 d 17:46:0022:26:3031.44 MJ
Run 3rand2exp8.336%0.000176319741419159440.007683 d 18:45:1222:41:1831.49 MJ
Run 4rand2exp7.512%0.00056789370413517770.011683 d 23:10:4023:47:4033.71 MJ
Run 5rand2exp13.488%0.0005701324081046430.013443 d 15:35:2021:53:5030.36 MJ
Run 6rand2exp6.992%0.00051236900730920890.012963 d 18:45:5622:41:2932.21 MJ
Run 7rand2exp11.96%0.000104495370906287530.009043 d 21:32:0823:23:0233.40 MJ
Run 8rand2exp8.992%0.00073764847616296210.011923 d 19:52:5222:58:1331.66 MJ
Run 9rand2exp9.68%0.000206878981720081280.006963 d 18:52:0822:43:0231.40 MJ
Run 10rand2exp6.912%0.000175524661460829780.014723 d 19:54:0022:58:3032.53 MJ
Run 1best1exp9.488%0.000231312948537657560.007523 d 20:08:2823:02:0732.57 MJ
Run 2best1exp5.376%0.00034678248563878950.007363 d 20:37:5223:09:2832.15 MJ
Run 3best1exp4.656%0.00054725449679144710.012083 d 16:19:1622:04:4932.43 MJ
Run 4best1exp9.088%0.00021950713606257210.01323 d 19:25:1222:51:1831.09 MJ
Run 5best1exp10.32%0.00070703962611645710.011923 d 17:19:0022:19:4531.00 MJ
Run 6best1exp8.968%0.000400013644409781130.007763 d 17:12:0022:18:0031.42 MJ
Run 7best1exp9.352%0.00060856749937464260.011763 d 18:15:2022:33:5031.41 MJ
Run 8best1exp5.912%0.00023334662588551490.007363 d 19:12:2822:48:0731.99 MJ
Run 9best1exp5.496%0.00047863716746602260.008083 d 17:39:0022:24:4531.38 MJ
Run 10best1exp12.632%0.000215933424049124720.0083 d 18:00:2822:30:0731.64 MJ
Run 1best2exp13.896%0.00061671353081296440.012643 d 17:46:2822:26:3730.52 MJ
Run 2best2exp4.352%0.000176322885568682050.007443 d 17:51:5222:27:5831.16 MJ
Run 3best2exp11.44%0.00052115819866432810.013283 d 17:56:2822:29:0731.05 MJ
Run 4best2exp10.936%0.000115513385100911640.008883 d 20:16:3623:04:0931.66 MJ
Run 5best2exp11.936%0.00025795502748807250.007363 d 21:41:0423:25:1632.77 MJ
Run 6best2exp14.464%0.000203088464681672240.012163 d 21:53:4823:28:2732.60 MJ
Run 7best2exp7.72%0.0002478962130905930.012563 d 17:55:4822:28:5732.04 MJ
Run 8best2exp7.544%0.000148863802317940870.008483 d 21:18:4023:19:4032.27 MJ
Run 9best2exp9.896%0.000254055915534197650.008083 d 19:19:0822:49:4733.55 MJ
Run 10best2exp7.424%0.00055531978511636590.00843 d 15:02:4421:45:4130.28 MJ
Run 1currenttobest1exp12.896%0.000413646896573837350.012483 d 19:28:2022:52:0532.15 MJ
Run 2currenttobest1exp9.288%0.00052646793903516830.012483 d 15:57:3621:59:2430.83 MJ
Run 3currenttobest1exp7.104%0.000191177678080248580.012083 d 20:12:1623:03:0431.82 MJ
Run 4currenttobest1exp3.24%0.00015107979147780760.008243 d 19:59:2822:59:5231.85 MJ
Run 5currenttobest1exp8.952%0.000482674060411273960.012323 d 22:45:1623:41:1933.33 MJ
Run 6currenttobest1exp6.56%0.000196705301132065820.008083 d 16:54:0822:13:3230.96 MJ
Run 7currenttobest1exp8.856%0.000206565843650162560.0083 d 20:54:4423:13:4132.03 MJ
Run 8currenttobest1exp6.824%0.00059828593990298910.012163 d 19:25:0422:51:1632.40 MJ
Run 9currenttobest1exp11.28%0.000256747445733641550.01483 d 19:32:1622:53:0433.62 MJ
Run 10currenttobest1exp4.776%0.000216769933398123830.008243 d 19:14:4822:48:4231.50 MJ
Table A7. Results obtained from 10 runs of 15 epochs of a hyperoptimized ConvNeXtSmall using a different DE strategy on CIFAR100.
Table A7. Results obtained from 10 runs of 15 epochs of a hyperoptimized ConvNeXtSmall using a different DE strategy on CIFAR100.
RunStrategyMax Accuracy (%)Best LRBest AccuracyCPU TimeElapsed TimeConsumed Energy
Run 1rand1bin96.856%0.00079010804734894020.020961 d 08:40:2408:10:0611.65 MJ
Run 2rand1bin95.152%0.00068894206126785950.0209623:50:3605:57:398.54 MJ
Run 3rand1bin95.912%0.00048945331834283030.020881 d 03:21:3606:50:2410.06 MJ
Run 4rand1bin96.96%0.00079010804734894020.019681 d 03:52:3606:58:0910.06 MJ
Run 5rand1bin95.168%0.00077246695797591720.020961 d 13:22:3209:20:3813.23 MJ
Run 6rand1bin97.6%0.00068210912296230880.020961 d 03:10:0806:47:329.21 MJ
Run 7rand1bin96.544%0.00058431083616292040.0209622:45:3205:41:238.04 MJ
Run 8rand1bin94.272%0.00077246695797591720.019617:44:4004:26:106.23 MJ
Run 9rand1bin96.624%0.00079010804734894020.020961 d 02:34:3206:38:389.60 MJ
Run 10rand1bin94.296%0.00079010804734894020.0208822:40:0405:40:018.16 MJ
Run 1best1bin96.768%0.00066540300794917370.0209618:06:1604:31:346.50 MJ
Run 2best1bin98.4%0.00075254780224504780.020961 d 01:08:0406:17:018.86 MJ
Run 3best1bin97.84%0.00072946703521472210.019681 d 02:39:4006:39:559.31 MJ
Run 4best1bin97.184%0.00072284005489201620.0196819:41:5604:55:296.83 MJ
Run 5best1bin98.24%0.00060820149663338530.0209619:04:2404:46:066.84 MJ
Run 6best1bin95.816%0.00055037828142615850.0209618:34:0404:38:316.92 MJ
Run 7best1bin94.456%0.00087212691909229680.020961 d 01:10:5206:17:438.88 MJ
Run 8best1bin97.152%0.00069594787236569610.019622:53:3205:43:237.94 MJ
Run 9best1bin97.592%0.00069769342229193240.0192816:39:0804:09:475.83 MJ
Run 10best1bin96.296%0.00060011434697409240.0208823:14:3605:48:398.32 MJ
Run 1currenttobest1bin97.448%0.00079010804734894020.0196820:32:3205:08:087.84 MJ
Run 2currenttobest1bin96.792%0.00060820149663338530.020961 d 02:38:3206:39:389.94 MJ
Run 3currenttobest1bin98.48%0.00062603156445300290.019621:25:0405:21:167.69 MJ
Run 4currenttobest1bin85.752%0.00052850312379018420.020961 d 01:10:4806:17:429.12 MJ
Run 5currenttobest1bin98.36%0.00066540300794917370.0196818:50:5604:42:446.71 MJ
Run 6currenttobest1bin98.208%0.00075385180672504920.0209622:32:0805:38:028.64 MJ
Run 7currenttobest1bin98.312%0.00057397123233140840.020961 d 02:36:4806:39:129.61 MJ
Run 8currenttobest1bin97.296%0.00066540300794917370.019681 d 01:29:2406:22:219.14 MJ
Run 9currenttobest1bin89.992%0.00060534389698465670.0209619:02:4804:45:426.71 MJ
Run 10currenttobest1bin98.448%0.00086789286879177020.019681 d 05:51:1207:27:4810.39 MJ
Run 1rand2bin98.624%0.00060255197947524050.019681 d 07:56:2807:59:0711.09 MJ
Run 2rand2bin98.032%0.00060820149663338530.020961 d 00:00:2006:00:058.28 MJ
Run 3rand2bin95.968%0.00060820149663338530.0196821:12:4805:18:127.49 MJ
Run 4rand2bin94.856%0.00058672949207391940.019681 d 07:18:0407:49:3110.94 MJ
Run 5rand2bin98.888%0.0005187833862561420.020961 d 06:55:5607:43:5911.38 MJ
Run 6rand2bin95.056%0.00066540300794917370.019623:30:0805:52:328.33 MJ
Run 7rand2bin93.936%0.0007164175242331650.019681 d 08:31:4008:07:5511.33 MJ
Run 8rand2bin96.896%0.00070741162664776210.0196822:51:0805:42:478.10 MJ
Run 9rand2bin97.248%0.00050189758331921160.020961 d 14:38:4009:39:4013.91 MJ
Run 10rand2bin96.256%0.00071071412129867450.0209623:21:1605:50:198.32 MJ
Run 1best2bin94.432%0.00079010804734894020.0209622:57:4405:44:268.12 MJ
Run 2best2bin93.88%0.00060820149663338530.01961 d 10:21:5608:35:2912.41 MJ
Run 3best2bin96.88%0.00077308462297344560.019681 d 06:32:5207:38:1310.71 MJ
Run 4best2bin99.024%0.00084320998785448760.0209622:23:5205:35:587.84 MJ
Run 5best2bin92.024%0.00048945331834283030.020961 d 10:07:4808:31:5712.03 MJ
Run 6best2bin95.936%0.00069809722320173390.0209621:42:1605:25:347.68 MJ
Run 7best2bin97.44%0.00066540300794917370.020961 d 01:37:4806:24:279.08 MJ
Run 8best2bin97.008%0.00087103601246564220.020961 d 00:09:1606:02:198.28 MJ
Run 9best2bin96.92%0.00066540300794917370.0192823:56:0805:59:028.24 MJ
Run 10best2bin97.76%0.00069758809851055720.020961 d 00:20:3606:05:098.74 MJ
Run 1rand1exp97.584%0.00062041543968458620.020961 d 12:53:2009:13:2013.00 MJ
Run 2rand1exp97.904%0.00079010804734894020.0209623:29:4005:52:258.39 MJ
Run 3rand1exp96.616%0.00079010804734894020.01961 d 03:11:2406:47:519.43 MJ
Run 4rand1exp96.984%0.00071864277865022960.020961 d 03:09:4406:47:269.52 MJ
Run 5rand1exp93.584%0.00066540300794917370.019681 d 04:23:1207:05:4810.24 MJ
Run 6rand1exp97.376%0.00058431083616292040.0196822:00:0405:30:017.74 MJ
Run 7rand1exp97.616%0.00059522302225026350.019681 d 01:26:5606:21:448.82 MJ
Run 8rand1exp96.904%0.00055037828142615850.01961 d 08:47:1608:11:4911.58 MJ
Run 9rand1exp97.16%0.00060820149663338530.019681 d 05:34:4807:23:4210.62 MJ
Run 10rand1exp95.544%0.00066540300794917370.0196823:23:3605:50:548.18 MJ
Run 1rand2exp90.96%0.00071071412129867450.01961 d 03:47:1606:56:499.98 MJ
Run 2rand2exp95.184%0.00079010804734894020.019621:52:4405:28:118.15 MJ
Run 3rand2exp94.6%0.00086789286879177020.0209623:46:2005:56:358.43 MJ
Run 4rand2exp96.08%0.00048945331834283030.020881 d 02:51:3606:42:549.64 MJ
Run 5rand2exp94.264%0.00079010804734894020.01961 d 01:25:0806:21:178.80 MJ
Run 6rand2exp97.808%0.00069810236024178490.0196822:51:0805:42:478.02 MJ
Run 7rand2exp95.88%0.00063964232534174370.020961 d 12:57:4409:14:2612.77 MJ
Run 8rand2exp94.184%0.00066540300794917370.020961 d 06:18:2807:34:3710.81 MJ
Run 9rand2exp92.704%0.00060159375090086420.020961 d 00:18:1206:04:339.11 MJ
Run 10rand2exp95.224%0.00075254780224504780.020961 d 09:31:1208:22:4811.82 MJ
Run 1best1exp97.856%0.00068546282079364920.0209620:46:1605:11:347.74 MJ
Run 2best1exp96.96%0.00079275041062573270.020961 d 01:59:3606:29:549.21 MJ
Run 3best1exp97.592%0.00071898224634934250.020961 d 00:07:3606:01:548.46 MJ
Run 4best1exp97.6%0.00088644469653704340.0192820:34:4405:08:417.31 MJ
Run 5best1exp98.088%0.00055037828142615850.0209622:47:1205:41:488.20 MJ
Run 6best1exp95.664%0.00066540300794917370.0209623:48:1605:57:048.81 MJ
Run 7best1exp96.312%0.00071898224634934250.0209621:01:4805:15:277.23 MJ
Run 8best1exp96.536%0.00045929087419095750.0209620:25:4005:06:257.11 MJ
Run 9best1exp97.88%0.00086789286879177020.0209619:37:2404:54:216.93 MJ
Run 10best1exp97.952%0.00058175251026142340.0196822:56:4805:44:128.34 MJ
Run 1best2exp79.408%0.00079010804734894020.020961 d 10:02:4408:30:4111.73 MJ
Run 2best2exp96.832%0.00069993468456865930.01961 d 16:31:2410:07:5114.08 MJ
Run 3best2exp93.768%0.00079010804734894020.0209619:07:1604:46:496.67 MJ
Run 4best2exp97.744%0.00057200375201822160.01961 d 08:50:2808:12:3711.38 MJ
Run 5best2exp92.44%0.00074105918425435070.019681 d 01:40:0806:25:028.87 MJ
Run 6best2exp98.048%0.00079010804734894020.0196823:16:5205:49:138.29 MJ
Run 7best2exp95.672%0.00086789286879177020.020961 d 02:20:3206:35:089.34 MJ
Run 8best2exp95.016%0.00079010804734894020.0209618:33:2804:38:226.65 MJ
Run 9best2exp98.256%0.0006094503161481170.019681 d 05:41:4807:25:2710.45 MJ
Run 10best2exp94.192%0.00057480752182023150.0196823:10:2405:47:368.57 MJ
Run 1currenttobest1exp98.488%0.00079010804734894020.019681 d 00:31:1606:07:498.57 MJ
Run 2currenttobest1exp86.744%0.00065191915890972660.019681 d 01:54:4406:28:419.06 MJ
Run 3currenttobest1exp98.408%0.0007475747864235650.0209620:26:0005:06:307.22 MJ
Run 4currenttobest1exp97.512%0.00066540300794917370.01961 d 04:42:4407:10:4110.71 MJ
Run 5currenttobest1exp97.88%0.00060820149663338530.0209622:45:5205:41:288.12 MJ
Run 6currenttobest1exp98.288%0.00056310576067857510.019620:57:4405:14:267.52 MJ
Run 7currenttobest1exp95.728%0.00064119742490651240.01961 d 06:36:2007:39:0510.78 MJ
Run 8currenttobest1exp97.032%0.00079010804734894020.020961 d 01:05:2406:16:218.77 MJ
Run 9currenttobest1exp98.24%0.00060820149663338530.0209623:59:0005:59:458.62 MJ
Run 10currenttobest1exp96.096%0.00063797250033348350.020961 d 00:45:4006:11:258.66 MJ
Table A8. Results obtained from 10 runs of 15 epochs of a hyperoptimized DenseNet121 using a different DE strategy on CIFAR100.
Table A8. Results obtained from 10 runs of 15 epochs of a hyperoptimized DenseNet121 using a different DE strategy on CIFAR100.
RunStrategyMax Accuracy (%)Best LRBest AccuracyCPU TimeElapsed TimeConsumed Energy
Run 1rand1bin82.92%0.00066540300794917370.020563 d 15:07:3621:46:5429.99 MJ
Run 2rand1bin82.888%0.00062020476681257930.020644 d 22:05:441 d 05:31:2640.69 MJ
Run 3rand1bin76.512%0.000485645055180374640.020644 d 10:00:441 d 02:30:1136.86 MJ
Run 4rand1bin78.048%0.00080538568249625360.020484 d 02:18:401 d 00:34:4033.63 MJ
Run 5rand1bin79.576%0.0006759756887715680.020163 d 17:07:4822:16:5731.05 MJ
Run 6rand1bin78.408%0.00087673645408905010.02044 d 15:57:361 d 03:59:2441.03 MJ
Run 7rand1bin76.92%0.000605402864479460.020483 d 17:13:4022:18:2529.91 MJ
Run 8rand1bin82.168%0.00087212691909229680.020484 d 14:04:201 d 03:31:0537.22 MJ
Run 9rand1bin81.704%0.0008297371849520230.020483 d 03:12:2418:48:0626.43 MJ
Run 10rand1bin84.112%0.00054763798695194230.020323 d 05:33:0419:23:1626.64 MJ
Run 1best1bin76.288%0.00074235307420013290.02043 d 18:22:1222:35:3332.99 MJ
Run 2best1bin76.696%0.00053269483800795420.020483 d 22:52:1223:43:0334.44 MJ
Run 3best1bin78.616%0.00057976792282186480.020083 d 15:27:4021:51:5529.34 MJ
Run 4best1bin77.696%0.00062008233534556830.020482 d 22:40:4817:40:1224.30 MJ
Run 5best1bin75.056%0.00075642070942000430.020644 d 06:46:441 d 01:41:4135.42 MJ
Run 6best1bin77.44%0.00039974204276031870.01884 d 14:41:481 d 03:40:2738.67 MJ
Run 7best1bin79.648%0.00048945331834283030.020243 d 02:49:0018:42:1526.21 MJ
Run 8best1bin78.52%0.00061206821525814840.018964 d 16:57:121 d 04:14:1839.48 MJ
Run 9best1bin77.112%0.00079010804734894020.018643 d 07:10:3219:47:3827.21 MJ
Run 10best1bin78.352%0.00055495159332900770.020564 d 10:42:161 d 02:40:3436.58 MJ
Run 1currenttobest1bin76.92%0.00071322992487661750.020083 d 22:39:5223:39:5831.86 MJ
Run 2currenttobest1bin78.784%0.00063627832765638350.02043 d 21:18:4823:19:4232.47 MJ
Run 3currenttobest1bin78.712%0.0006542132084908520.018723 d 20:55:2823:13:5231.95 MJ
Run 4currenttobest1bin76.224%0.0007872889918670340.018963 d 17:49:1222:27:1830.72 MJ
Run 5currenttobest1bin81.304%0.00079808968823810980.020484 d 03:24:001 d 00:51:0034.48 MJ
Run 6currenttobest1bin80.384%0.00057534029050436970.018483 d 17:36:2822:24:0732.39 MJ
Run 7currenttobest1bin81%0.00050479137239431860.020484 d 05:25:161 d 01:21:1935.28 MJ
Run 8currenttobest1bin77.168%0.00059596352650442360.020644 d 15:42:161 d 03:55:3438.33 MJ
Run 9currenttobest1bin81.336%0.000314255571715287030.018483 d 18:30:4022:37:4030.91 MJ
Run 10currenttobest1bin75.976%0.0005035028039211230.020484 d 13:53:361 d 03:28:2437.46 MJ
Run 1rand2bin73.424%0.00079010804734894020.020163 d 23:45:5623:56:2933.48 MJ
Run 2rand2bin78.408%0.00080303311650768080.020644 d 01:54:081 d 00:28:3233.22 MJ
Run 3rand2bin83.08%0.00053457893710137220.018084 d 07:17:321 d 01:49:2335.99 MJ
Run 4rand2bin78.44%0.000484188930376752340.020484 d 09:39:201 d 02:24:5036.11 MJ
Run 5rand2bin86.144%0.00056558211822817340.020484 d 07:49:481 d 01:57:2734.62 MJ
Run 6rand2bin77.312%0.00081162892042236940.02044 d 00:45:561 d 00:11:2932.78 MJ
Run 7rand2bin78.096%0.00047788878501502830.018564 d 19:37:041 d 04:54:1638.72 MJ
Run 8rand2bin76.072%0.00075191319683944620.018322 d 22:25:2017:36:2023.72 MJ
Run 9rand2bin80.48%0.00048945331834283030.020483 d 18:42:2022:40:3532.70 MJ
Run 10rand2bin80.96%0.00076457437800185190.020644 d 18:18:561 d 04:34:4439.98 MJ
Run 1best2bin77.208%0.00048945331834283030.017843 d 02:05:1218:31:1826.87 MJ
Run 2best2bin78.64%0.00083156360585558970.020243 d 15:54:2421:58:3631.75 MJ
Run 3best2bin80.008%0.00054273905441604210.020163 d 20:26:3223:06:3831.05 MJ
Run 4best2bin75.048%0.00086789286879177020.018723 d 19:25:4422:51:2631.88 MJ
Run 5best2bin80.344%0.00075813944822731970.02084 d 11:40:161 d 02:55:0436.48 MJ
Run 6best2bin82.904%0.00084379116463186870.020564 d 06:43:521 d 01:40:5835.18 MJ
Run 7best2bin73.2%0.00075796778313877390.01843 d 12:22:5621:05:4429.16 MJ
Run 8best2bin80.864%0.00077614786174076930.018323 d 05:22:5619:20:4427.09 MJ
Run 9best2bin74.16%0.00064623408788525190.018243 d 02:33:2018:38:2026.26 MJ
Run 10best2bin74.952%0.000455012615093883240.018964 d 16:01:081 d 04:00:1738.56 MJ
Run 1rand1exp74.304%0.000524744839551060.018163 d 09:37:4420:24:2628.64 MJ
Run 2rand1exp79.888%0.00072426963283277670.018323 d 08:40:5620:10:1427.85 MJ
Run 3rand1exp77.248%0.00076315363521346150.018324 d 04:16:121 d 01:04:0333.67 MJ
Run 4rand1exp78.12%0.000434771456989763770.017843 d 16:44:1222:11:0330.92 MJ
Run 5rand1exp79.208%0.00029076980693108320.018484 d 03:13:561 d 00:48:2933.39 MJ
Run 6rand1exp80.296%0.00049941667664874480.020322 d 21:39:1217:24:4824.08 MJ
Run 7rand1exp74.568%0.000474675422146660.020644 d 09:19:481 d 02:19:5736.24 MJ
Run 8rand1exp78.84%0.00079010804734894020.020484 d 15:27:241 d 03:51:5138.04 MJ
Run 9rand1exp80.848%0.00042807830876629820.020564 d 04:37:521 d 01:09:2834.79 MJ
Run 10rand1exp80.992%0.00026101881302939070.01843 d 18:43:0422:40:4632.89 MJ
Run 1rand2exp80.432%0.00064105073296388930.020563 d 21:45:2423:26:2132.88 MJ
Run 2rand2exp77.736%0.00086789286879177020.02043 d 15:47:5221:56:5829.60 MJ
Run 3rand2exp80.592%0.0005688410618395670.018482 d 04:15:1613:03:4917.99 MJ
Run 4rand2exp82.136%0.00055461371586765770.020484 d 04:48:241 d 01:12:0634.47 MJ
Run 5rand2exp80.24%0.00076228682588160350.020645 d 00:22:081 d 06:05:3241.39 MJ
Run 6rand2exp81.224%0.00067957956620932530.020564 d 04:52:241 d 01:13:0634.98 MJ
Run 7rand2exp76.784%0.00087745377747368140.020724 d 14:45:201 d 03:41:2038.24 MJ
Run 8rand2exp76.92%0.00064458845291245660.020323 d 22:29:4023:37:2532.78 MJ
Run 9rand2exp77.456%0.00079386743873228410.020564 d 00:46:081 d 00:11:3233.52 MJ
Run 10rand2exp78.968%0.00040112334087392890.018884 d 21:11:281 d 05:17:5239.54 MJ
Run 1best1exp80.536%0.00054326689180962890.020323 d 21:49:1223:27:1832.91 MJ
Run 2best1exp85.944%0.00055665078968390940.020323 d 19:58:2822:59:3732.15 MJ
Run 3best1exp78.376%0.00073946983232652380.020563 d 19:14:4822:48:4231.12 MJ
Run 4best1exp83.952%0.00072037542401543680.02043 d 20:47:0823:11:4732.29 MJ
Run 5best1exp82.152%0.00049357761470809550.020242 d 16:46:4016:11:4022.34 MJ
Run 6best1exp77.032%0.00049464618541069930.018644 d 20:30:041 d 05:07:3139.24 MJ
Run 7best1exp78.256%0.00086789286879177020.018643 d 18:09:0822:32:1731.15 MJ
Run 8best1exp76.568%0.00050386764270450840.020643 d 21:53:4023:28:2531.41 MJ
Run 9best1exp81.928%0.00076204438746037590.01884 d 21:55:481 d 05:28:5740.68 MJ
Run 10best1exp78.616%0.00084172985033400270.020484 d 10:01:001 d 02:30:1536.33 MJ
Run 1best2exp78.04%0.00057308288774453760.018563 d 19:48:5222:57:1332.05 MJ
Run 2best2exp75.856%0.00075661472837156020.018643 d 09:51:1220:27:4827.85 MJ
Run 3best2exp82.192%0.00072625877758004640.020564 d 09:36:441 d 02:24:1136.30 MJ
Run 4best2exp81.816%0.00066540300794917370.020643 d 17:06:3622:16:3931.09 MJ
Run 5best2exp78.92%0.00058061099849956960.020724 d 07:19:441 d 01:49:5636.15 MJ
Run 6best2exp77.376%0.00071302710551429350.02084 d 22:08:161 d 05:32:0440.67 MJ
Run 7best2exp78.928%0.00077299981303713890.01884 d 06:41:321 d 01:40:2334.85 MJ
Run 8best2exp80.176%0.000400535402137857670.018563 d 19:25:3222:51:2330.82 MJ
Run 9best2exp74.688%0.00055037828142615850.020324 d 07:27:081 d 01:51:4737.24 MJ
Run 10best2exp76.312%0.00053553739028156650.020484 d 07:18:361 d 01:49:3937.86 MJ
Run 1currenttobest1exp80.632%0.00026727959265821790.018482 d 19:45:2816:56:2224.09 MJ
Run 2currenttobest1exp81.752%0.00055045657000630330.020243 d 13:43:3621:25:5429.50 MJ
Run 3currenttobest1exp79.408%0.00089644236152839140.020483 d 21:17:4023:19:2532.76 MJ
Run 4currenttobest1exp81.68%0.00049319553390841030.020563 d 13:29:2421:22:2130.02 MJ
Run 5currenttobest1exp78.368%0.00049973568168069060.018323 d 23:39:0823:54:4733.75 MJ
Run 6currenttobest1exp78.448%0.00061503469121975980.01884 d 02:30:561 d 00:37:4433.19 MJ
Run 7currenttobest1exp77.392%0.000391632351492734540.018724 d 03:53:241 d 00:58:2133.63 MJ
Run 8currenttobest1exp75.92%0.00048945331834283030.020243 d 08:05:0020:01:1528.98 MJ
Run 9currenttobest1exp76.944%0.00075138912518938410.020244 d 02:20:201 d 00:35:0533.09 MJ
Run 10currenttobest1exp77.424%0.00054452218314174150.020324 d 01:16:281 d 00:19:0733.69 MJ

References

  1. Silva, C.; Vilaça, R.; Pereira, A.; Bessa, R. A review on the decarbonization of high-performance computing centers. Renew. Sustain. Energy Rev. 2024, 189, 114019. [Google Scholar] [CrossRef]
  2. Chu, X.; Hofstätter, D.; Ilager, S.; Talluri, S.; Kampert, D.; Podareanu, D.; Duplyakin, D.; Brandic, I.; Iosup, A. Generic and ML Workloads in an HPC Datacenter: Node Energy, Job Failures, and Node-Job Analysis. arXiv 2024, arXiv:2409.08949. [Google Scholar] [CrossRef]
  3. Baratchi, M.; Wang, C.; Limmer, S.; van Rijn, J.N.; Hoos, H.; Bäck, T.; Olhofer, M. Automated machine learning: Past, present and future. Artif. Intell. Rev. 2024, 57, 122. [Google Scholar] [CrossRef]
  4. Bolón-Canedo, V.; Morán-Fernández, L.; Cancela, B.; Alonso-Betanzos, A. A review of green artificial intelligence: Towards a more sustainable future. Neurocomputing 2024, 599, 128096. [Google Scholar] [CrossRef]
  5. TOP500 Methodology. 2025. Available online: https://www.top500.org/static/media/uploads/methodology-2.0rc1.pdf (accessed on 19 January 2025).
  6. Miller, J.; Trümper, L.; Terboven, C.; Müller, M.S. A Theoretical Model for Global Optimization of Parallel Algorithms. Mathematics 2021, 9, 1685. [Google Scholar] [CrossRef]
  7. Damme, P.; Birkenbach, M.; Bitsakos, C.; Boehm, M.; Bonnet, P.; Ciorba, F.; Dokter, M.; Dowgiallo, P.; Eleliemy, A.; Faerber, C.; et al. DAPHNE: An Open and Extensible System Infrastructure for Integrated Data Analysis Pipelines. In Proceedings of the Conference on Innovative Data Systems Research, Chaminade, CA, USA, 9–12 January 2022. [Google Scholar]
  8. Alangari, N.; El Bachir Menai, M.; Mathkour, H.; Almosallam, I. Exploring Evaluation Methods for Interpretable Machine Learning: A Survey. Information 2023, 14, 469. [Google Scholar] [CrossRef]
  9. Jakobsche, T.; Lachiche, N.; Ciorba, F.M. Investigating HPC Job Resource Requests and Job Efficiency Reporting. In Proceedings of the 2023 22nd International Symposium on Parallel and Distributed Computing (ISPDC), Bucharest, Romania, 10–12 July 2023; pp. 61–68. [Google Scholar] [CrossRef]
  10. Yarally, T.; Cruz, L.; Feitosa, D.; Sallou, J.; van Deursen, A. Uncovering Energy-Efficient Practices in Deep Learning Training: Preliminary Steps Towards Green AI. In Proceedings of the 2023 IEEE/ACM 2nd International Conference on AI Engineering—Software Engineering for AI (CAIN), Melbourne, Australia, 15–16 May 2023; pp. 25–36. [Google Scholar] [CrossRef]
  11. Prica, T. Development and supporting activities on EuroHPC Vega. In Proceedings of the Austrian-Slovenian HPC Meeting 2024—ASHPC24, Grundlsee, Austria, 10–13 June 2024; p. 14. [Google Scholar]
  12. Oliveira, S.d.; Topsakal, O.; Toker, O. Benchmarking Automated Machine Learning (AutoML) Frameworks for Object Detection. Information 2024, 15, 63. [Google Scholar] [CrossRef]
  13. Yoo, A.B.; Jette, M.A.; Grondona, M. SLURM: Simple Linux Utility for Resource Management. In Proceedings of the 9th International Workshop on Job Scheduling Strategies for Parallel Processing (JSSPP), Seattle, WA, USA, 24 June 2003. [Google Scholar] [CrossRef]
  14. Han, M.; Wu, H.; Chen, Z.; Li, M.; Zhang, X. A survey of multi-label classification based on supervised and semi-supervised learning. Int. J. Mach. Learn. Cybern. 2023, 14, 697–724. [Google Scholar] [CrossRef]
  15. Virtanen, P.; Gommers, R.; Oliphant, T.E.; Haberland, M.; Reddy, T.; Cournapeau, D.; Burovski, E.; Peterson, P.; Weckesser, W.; Bright, J.; et al. SciPy 1.0: Fundamental algorithms for scientific computing in Python. Nat. Methods 2020, 17, 261–272. [Google Scholar] [CrossRef]
  16. Das, S.; Mullick, S.S.; Suganthan, P.N. Recent Advances in Differential Evolution—An Updated Survey. Swarm Evol. Comput. 2016, 27, 1–30. [Google Scholar] [CrossRef]
  17. Li, J.Y.; Zhan, Z.H.; Zhang, J. Evolutionary Computation for Expensive Optimization: A Survey. Mach. Intell. Res. 2022, 19, 3–23. [Google Scholar] [CrossRef]
  18. Qin, X.; Luo, Y.; Chen, S.; Chen, Y.; Han, Y. Investigation of Energy-Saving Strategy for Parallel Variable Frequency Pump System Based on Improved Differential Evolution Algorithm. Energies 2022, 15, 5360. [Google Scholar] [CrossRef]
  19. Dragoi, E.N.; Dafinescu, V. Parameter control and hybridization techniques in differential evolution: A survey. Artif. Intell. Rev. 2015, 45, 447–470. [Google Scholar] [CrossRef]
  20. Storn, R.; Price, K. Differential Evolution–A Simple and Efficient Heuristic for global Optimization over Continuous Spaces. J. Glob. Optim. 1997, 11, 341–359. [Google Scholar] [CrossRef]
  21. Eltaeib, T.; Mahmood, A. Differential Evolution: A Survey and Analysis. Appl. Sci. 2018, 8, 1945. [Google Scholar] [CrossRef]
  22. Nanthapodej, R.; Liu, C.H.; Nitisiri, K.; Pattanapairoj, S. Hybrid Differential Evolution Algorithm and Adaptive Large Neighborhood Search to Solve Parallel Machine Scheduling to Minimize Energy Consumption in Consideration of Machine-Load Balance Problems. Sustainability 2021, 13, 5470. [Google Scholar] [CrossRef]
  23. Chhabra, A.; Sahana, S.K.; Sani, N.S.; Mohammadzadeh, A.; Omar, H.A. Energy-Aware Bag-of-Tasks Scheduling in the Cloud Computing System Using Hybrid Oppositional Differential Evolution-Enabled Whale Optimization Algorithm. Energies 2022, 15, 4571. [Google Scholar] [CrossRef]
  24. IAM Working Group. IPMI Specification. 2006. Available online: https://openipmi.sourceforge.io/IPMI.pdf (accessed on 25 October 2024).
  25. Bischl, B.; Binder, M.; Lang, M.; Pielok, T.; Richter, J.; Coors, S.; Thomas, J.; Ullmann, T.; Becker, M.; Boulesteix, A.L.; et al. Hyperparameter optimization: Foundations, algorithms, best practices, and open challenges. WIREs Data Min. Knowl. Discov. 2023, 13, e1484. [Google Scholar] [CrossRef]
  26. Morales-Hernández, A.; Van Nieuwenhuyse, I.; Rojas Gonzalez, S. A survey on multi-objective hyperparameter optimization algorithms for machine learning. Artif. Intell. Rev. 2022, 56, 8043–8093. [Google Scholar] [CrossRef]
  27. Dataset CIFAR10 and CIFAR100. Available online: https://www.cs.toronto.edu/~kriz/cifar.html (accessed on 19 November 2024).
  28. Boito, F.; Brandt, J.; Cardellini, V.; Carns, P.; Ciorba, F.M.; Egan, H. Autonomy Loops for Monitoring, Operational Data Analytics, Feedback, and Response in HPC Operations. In Proceedings of the 2023 IEEE International Conference on Cluster Computing Workshops (CLUSTER Workshops), Santa Fe, NM, USA, 31 October 2023; pp. 37–43. [Google Scholar] [CrossRef]
  29. El Naqa, I.; Murphy, M.J. What Is Machine Learning? In Machine Learning in Radiation Oncology: Theory and Applications; Springer International Publishing: Cham, Switzerland, 2015; pp. 3–11. [Google Scholar] [CrossRef]
  30. Zhang, X.; Guo, F.; Chen, T.; Pan, L.; Beliakov, G.; Wu, J. A Brief Survey of Machine Learning and Deep Learning Techniques for E-Commerce Research. J. Theor. Appl. Electron. Commer. Res. 2023, 18, 2188–2216. [Google Scholar] [CrossRef]
  31. Bergstra, J.; Bardenet, R.; Bengio, Y.; Kégl, B. Algorithms for hyper-parameter optimization. In Proceedings of the 25th International Conference on Neural Information Processing Systems, NIPS’11, Granada, Spain, 12–15 December 2011; Curran Associates Inc.: Red Hook, NY, USA, 2011; pp. 2546–2554. Available online: https://dl.acm.org/doi/10.5555/2986459.2986743 (accessed on 7 April 2025).
  32. Zamuda, A.; Daniel Hernández Sosa, J.; Adler, L. Improving constrained glider trajectories for ocean eddy border sampling within extended mission planning time. In Proceedings of the 2016 IEEE Congress on Evolutionary Computation (CEC), Vancouver, BC, Canada, 24–29 July 2016; pp. 1727–1734. [Google Scholar] [CrossRef]
  33. Zhu, K.; Wu, J. Residual attention: A simple but effective method for multi-label recognition. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 184–193. [Google Scholar] [CrossRef]
  34. Zhou, Z.H. Machine Learning; Springer Nature: Berlin/Heidelberg, Germany, 2021. [Google Scholar]
  35. Hu, X.; Chu, L.; Pei, J.; Liu, W.; Bian, J. Model Complexity of Deep Learning: A Survey. arXiv 2021, arXiv:2103.05127. [Google Scholar] [CrossRef]
  36. Radford, A.; Narasimhan, K.; Salimans, T.; Sutskever, I. Improving Language Understanding by Generative Pre-Training; OpenAI: San Francisco, CA, USA, 2018. [Google Scholar]
  37. Menik, S.; Ramaswamy, L. Towards Modular Machine Learning Solution Development: Benefits and Trade-offs. arXiv 2023, arXiv:2301.09753. [Google Scholar] [CrossRef]
  38. Shen, Y.; Zhang, Z.; Cao, T.; Tan, S.; Chen, Z.; Gan, C. ModuleFormer: Modularity Emerges from Mixture-of-Experts. arXiv 2023, arXiv:2306.04640. [Google Scholar] [CrossRef]
  39. Barandas, M.; Famiglini, L.; Campagner, A.; Folgado, D.; Simão, R.; Cabitza, F.; Gamboa, H. Evaluation of uncertainty quantification methods in multi-label classification: A case study with automatic diagnosis of electrocardiogram. Inf. Fusion 2024, 101, 101978. [Google Scholar] [CrossRef]
  40. Maloney, S.; Suarez, E.; Eicker, N.; Guimarães, F.; Frings, W. Analyzing HPC Monitoring Data With a View Towards Efficient Resource Utilization. In Proceedings of the 2024 IEEE 36th International Symposium on Computer Architecture and High Performance Computing (SBAC-PAD), Hilo, HI, USA, 13–15 November 2024; pp. 170–181. [Google Scholar]
  41. Vontzalidis, A.; Psomadakis, S.; Bitsakos, C.; Dokter, M.; Innerebner, K.; Damme, P.; Boehm, M.; Ciorba, F.; Eleliemy, A.; Karakostas, V.; et al. DAPHNE Runtime: Harnessing Parallelism for‚ Integrated Data Analysis Pipelines. In Proceedings of the Euro-Par 2023: Parallel Processing Workshops, Limassol, Cyprus, 28 August–1 September 2023; pp. 242–246. [Google Scholar]
  42. Jakobsche, T.; Lachiche, N.; Ciorba, F.M. Challenges and Opportunities of Machine Learning for Monitoring and Operational Data Analytics in Quantitative Codesign of Supercomputers. arXiv 2022, arXiv:2209.07164. [Google Scholar] [CrossRef]
  43. Prica, T.; Zamuda, A. Monitoring Energy Consumption of Workloads on HPC Vega. In Proceedings of the 6th ISC HPC International Workshop on “Monitoring & Operational Data Analytics”, Hamburg, Germany, 9 March–13 June 2025. [Google Scholar]
  44. Chakraborty, U.K. Advances in Differential Evolution; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2008; Volume 143. [Google Scholar]
  45. Ahmad, M.F.; Isa, N.A.M.; Lim, W.H.; Ang, K.M. Differential evolution: A recent review based on state-of-the-art works. Alex. Eng. J. 2022, 61, 3831–3872. [Google Scholar] [CrossRef]
  46. Glotić, A.; Zamuda, A. Short-term combined economic and emission hydrothermal optimization by surrogate differential evolution. Appl. Energy 2015, 141, 42–56. [Google Scholar] [CrossRef]
  47. Zamuda, A.; Sosa, J.D.H.; Adler, L. Constrained differential evolution optimization for underwater glider path planning in sub-mesoscale eddy sampling. Appl. Soft Comput. 2016, 42, 93–118. [Google Scholar] [CrossRef]
  48. Lucas, C.; Hernández-Sosa, D.; Greiner, D.; Zamuda, A.; Caldeira, R. An approach to multi-objective path planning optimization for underwater gliders. Sensors 2019, 19, 5506. [Google Scholar] [CrossRef]
  49. Brest, J.; Greiner, S.; Boskovic, B.; Mernik, M.; Zumer, V. Self-Adapting Control Parameters in Differential Evolution: A Comparative Study on Numerical Benchmark Problems. IEEE Trans. Evol. Comput. 2006, 10, 646–657. [Google Scholar] [CrossRef]
  50. Fan, Q.; Yan, X.; Zhang, Y. Auto-selection mechanism of differential evolution algorithm variants and its application. Eur. J. Oper. Res. 2018, 270, 636–653. [Google Scholar] [CrossRef]
  51. Vincent, A.M.; Jidesh, P. An improved hyperparameter optimization framework for AutoML systems using evolutionary algorithms. Sci. Rep. 2023, 13, 4737. [Google Scholar] [CrossRef]
  52. Sen, A.; Gupta, V.; Tang, C. Differential Evolution Algorithm Based Hyperparameter Selection of Gated Recurrent Unit for Electrical Load Forecasting. arXiv 2023, arXiv:2309.13019. [Google Scholar] [CrossRef]
  53. Gomes, E.; Pereira, L.; Esteves, A.; Morais, H. Metaheuristic Optimization Methods in Energy Community Scheduling: A Benchmark Study. Energies 2024, 17, 2968. [Google Scholar] [CrossRef]
  54. Main Stages of the DE Algorithm. Available online: https://www.researchgate.net/figure/Main-stages-of-the-DE-algorithm_fig1_336225430 (accessed on 19 November 2024).
  55. Opara, K.; Arabas, J. Comparison of mutation strategies in Differential Evolution—A probabilistic perspective. Swarm Evol. Comput. 2018, 39, 53–69. [Google Scholar] [CrossRef]
  56. Wu, T.; Li, X.; Zhou, D.; Li, N.; Shi, J. Differential Evolution Based Layer-Wise Weight Pruning for Compressing Deep Neural Networks. Sensors 2021, 21, 880. [Google Scholar] [CrossRef]
  57. Zamuda, A. Foundational Concepts and Real-World Applications of Self-Adaptive Differential Evolution and Success History. In Swarm Intelligence—Foundational Concepts and Real-World Applications; Chibante, R., Miranda, P., Palade, V., Eds.; Artificial Intelligence; IntechOpen: London, UK, 2025; Chapter 1; pp. 1–20. Available online: https://www.intechopen.com/online-first/1222844 (accessed on 7 April 2025).
  58. Qiao, K.; Wen, X.; Ban, X.; Chen, P.; Price, K.; Suganthan, P.; Liang, J.; Wu, G.; Yue, C. Evaluation Criteria for CEC 2024 Competition and Special Session on Numerical Optimization Considering Accuarcy and Speed; Technical Report; Zhengzhou University: Zhengzhou, China; Central South University: Changsha, China; Henan Institute of Technology: Xinxiang, China; Qatar University: Doha, Qatar, 2023. [Google Scholar]
  59. Tanabe, R.; Fukunaga, A.S. Improving the search performance of SHADE using linear population size reduction. In Proceedings of the 2014 IEEE Congress on Evolutionary Computation, Beijing, China, 6–11 July 2014; pp. 1658–1665. [Google Scholar]
  60. Viktorin, A.; Senkerik, R.; Pluhacek, M.; Kadavy, T.; Zamuda, A. Distance based parameter adaptation for Success-History based Differential Evolution. Swarm Evol. Comput. 2019, 50, 100462. [Google Scholar] [CrossRef]
  61. Tanabe, R.; Fukunaga, A. Reviewing and Benchmarking Parameter Control Methods in Differential Evolution. IEEE Trans. Cybern. 2020, 50, 1170–1184. [Google Scholar] [CrossRef]
  62. Mininno, E.; Neri, F.; Cupertino, F.; Naso, D. Compact Differential Evolution. IEEE Trans. Evol. Comput. 2011, 15, 32–54. [Google Scholar] [CrossRef]
  63. Zamuda, A.; Dokter, M. Deploying DAPHNE Computational Intelligence on EuroHPC Vega for Benchmarking Randomised Optimisation Algorithms. In Proceedings of the International Conference on Broadband Communications for Next Generation Networks and Multimedia Applications (CoBCom), Graz, Austria, 9–11 July 2024; pp. 1–8. [Google Scholar]
  64. Jiang, Y.; Qi, X.; Liu, C. Energy-Aware Automatic Tuning on Many-Core Platform via Differential Evolution. In Proceedings of the 2016 45th International Conference on Parallel Processing Workshops (ICPPW), Philadelphia, PA, USA, 16–19 August 2016; pp. 258–265. [Google Scholar] [CrossRef]
  65. Baioletti, M.; Di Bari, G.; Milani, A.; Poggioni, V. Differential Evolution for Neural Networks Optimization. Mathematics 2020, 8, 69. [Google Scholar] [CrossRef]
  66. Agarwal, M.; Gupta, S.K.; Biswas, K.K. DECACNN: Differential evolution-based approach to compress and accelerate the convolution neural network model. Neural Comput. Appl. 2023, 36, 2665–2681. [Google Scholar] [CrossRef]
  67. Wu, X.; Che, A. A memetic differential evolution algorithm for energy-efficient parallel machine scheduling. Omega 2019, 82, 155–165. [Google Scholar] [CrossRef]
  68. Abdel-Basset, M.; Mohamed, R.; Alrashdi, I.; Sallam, K.M.; Hameed, I.A. Evolution-based energy-efficient data collection system for UAV-supported IoT: Differential evolution with population size optimization mechanism. Expert Syst. Appl. 2024, 245, 123082. [Google Scholar] [CrossRef]
  69. Zamuda, A.; Hernández Sosa, J.D. Differential evolution and underwater glider path planning applied to the short-term opportunistic sampling of dynamic mesoscale ocean structures. Appl. Soft Comput. 2014, 24, 95–108. [Google Scholar] [CrossRef]
  70. Zamuda, A.; Sosa, J.D.H. Success history applied to expert system for underwater glider path planning using differential evolution. Expert Syst. Appl. 2019, 119, 155–170. [Google Scholar] [CrossRef]
  71. Janssen, D.; Pullan, W.; Liew, A.W.C. GPU Based Differential Evolution: New Insights and Comparative Study. arXiv 2024, arXiv:2405.16551. [Google Scholar] [CrossRef]
  72. Van Stein, B.; Vermetten, D.; Caraffini, F.; Kononova, A.V. Deep BIAS: Detecting Structural Bias using Explainable AI. In Proceedings of the GECCO ’23 Companion: Proceedings of the Companion Conference on Genetic and Evolutionary Computation, Lisbon, Portugal, 15–19 July 2023; pp. 455–458. [Google Scholar]
  73. van Stein, N.; Kononova, A.V. (Eds.) Explainable AI for Evolutionary Computation; Springer Nature: Singapore, 2025. [Google Scholar] [CrossRef]
  74. Raponi, E.; Rodriguez, I.O.; van Stein, N. Global Sensitivity Analysis Is Not Always Beneficial for Evolutionary Computation: A Study in Engineering Design. In Explainable AI for Evolutionary Computation; Springer Nature: Singapore, 2025; pp. 13–40. [Google Scholar] [CrossRef]
  75. Barbudo, R.; Ventura, S.; Romero, J.R. Eight years of AutoML: Categorisation, review and trends. Knowl. Inf. Syst. 2023, 65, 5097–5149. [Google Scholar] [CrossRef]
  76. Salehin, I.; Islam, M.S.; Saha, P.; Noman, S.; Tuni, A.; Hasan, M.M.; Baten, M.A. AutoML: A systematic review on automated machine learning with neural architecture search. J. Inf. Intell. 2024, 2, 52–81. [Google Scholar] [CrossRef]
  77. Chatzilygeroudis, K.; Hatzilygeroudis, I.; Perikos, I. Machine Learning Basics. In Intelligent Computing for Interactive System Design: Statistics, Digital Signal Processing, and Machine Learning in Practice, 1st ed.; Association for Computing Machinery: New York, NY, USA, 2021; pp. 143–193. [Google Scholar] [CrossRef]
  78. Geissler, D.; Zhou, B.; Suh, S.; Lukowicz, P. Spend More to Save More (SM2): An Energy-Aware Implementation of Successive Halving for Sustainable Hyperparameter Optimization. arXiv 2024, arXiv:2412.08526. [Google Scholar] [CrossRef]
  79. Ferro, M.; Silva, G.D.; de Paula, F.B.; Vieira, V.; Schulze, B. Towards a sustainable artificial intelligence: A case study of energy efficiency in decision tree algorithms. Concurr. Comput. Pract. Exp. 2021, 35, e6815. [Google Scholar] [CrossRef]
  80. Castellanos-Nieves, D.; García-Forte, L. Improving Automated Machine-Learning Systems Through Green AI. Appl. Sci. 2023, 13, 11583. [Google Scholar] [CrossRef]
  81. Castellanos-Nieves, D.; García-Forte, L. Strategies of Automated Machine Learning for Energy Sustainability in Green Artificial Intelligence. Appl. Sci. 2024, 14, 6196. [Google Scholar] [CrossRef]
  82. Zamuda, A.; Brest, J. Self-adaptive control parameters’ randomization frequency and propagations in differential evolution. Swarm Evol. Comput. 2015, 25, 72–99. [Google Scholar] [CrossRef]
  83. Vakhnin, A.; Ryzhikov, I.; Niska, H.; Kolehmainen, M. A Novel Multi-Objective Hybrid Evolutionary-Based Approach for Tuning Machine Learning Models in Short-Term Power Consumption Forecasting. AI 2024, 5, 2461–2496. [Google Scholar] [CrossRef]
  84. Pătrăușanu, A.; Florea, A.; Neghină, M.; Dicoiu, A.; Chiș, R. A Systematic Review of Multi-Objective Evolutionary Algorithms Optimization Frameworks. Processes 2024, 12, 869. [Google Scholar] [CrossRef]
  85. Liuliakov, A.; Hermes, L.; Hammer, B. AutoML technologies for the identification of sparse classification and outlier detection models. Appl. Soft Comput. 2023, 133, 109942. [Google Scholar] [CrossRef]
  86. Jin, H.; Chollet, F.; Song, Q.; Hu, X. AutoKeras: An AutoML Library for Deep Learning. J. Mach. Learn. Res. 2023, 24, 1–6. [Google Scholar]
  87. Shi, M.; Shen, W. Automatic Modeling for Concrete Compressive Strength Prediction Using Auto-Sklearn. Buildings 2022, 12, 1406. [Google Scholar] [CrossRef]
  88. Omar, I.; Khan, M.; Starr, A.; Abou Rok Ba, K. Automated Prediction of Crack Propagation Using H2O AutoML. Sensors 2023, 23, 8419. [Google Scholar] [CrossRef]
  89. Olson, R.S.; Moore, J.H. TPOT: A Tree-based Pipeline Optimization Tool for Automating Machine Learning. In PMLR, Proceedings of the Workshop on Automatic Machine Learning, New York, NY, USA, 24 June 2016; Hutter, F., Kotthoff, L., Vanschoren, J., Eds.; Springer: Cham, Switzerland, 2016; Volume 64, pp. 66–74. [Google Scholar]
  90. TensorFlow. Available online: https://www.tensorflow.org/ (accessed on 28 November 2024).
  91. PyTorch. Available online: https://pytorch.org/ (accessed on 19 November 2024).
  92. Hansen, N.; Auger, A.; Ros, R.; Mersmann, O.; Tušar, T.; Brockhoff, D. COCO: A platform for comparing continuous optimizers in a black-box setting. Optim. Methods Softw. 2020, 36, 114–144. [Google Scholar] [CrossRef]
  93. Varelas, K.; El Hara, O.A.; Brockhoff, D.; Hansen, N.; Nguyen, D.M.; Tušar, T.; Auger, A. Benchmarking large-scale continuous optimizers: The bbob-largescale testbed, a COCO software guide and beyond. Appl. Soft Comput. 2020, 97, 106737. [Google Scholar] [CrossRef]
  94. Doerr, C.; Wang, H.; Ye, F.; van Rijn, S.; Bäck, T. IOHprofiler: A Benchmarking and Profiling Tool for Iterative Optimization Heuristics. arXiv 2018, arXiv:1810.05281. [Google Scholar] [CrossRef]
  95. Doerr, C.; Ye, F.; Horesh, N.; Wang, H.; Shir, O.M.; Bäck, T. Benchmarking discrete optimization heuristics with IOHprofiler. Appl. Soft Comput. 2020, 88, 106027. [Google Scholar] [CrossRef]
  96. Durillo, J.J.; Nebro, A.J. jMetal: A Java framework for multi-objective optimization. Adv. Eng. Softw. 2011, 42, 760–771. [Google Scholar] [CrossRef]
  97. López-Ibáñez, M.; Dubois-Lacoste, J.; Pérez Cáceres, L.; Birattari, M.; Stützle, T. The irace package: Iterated racing for automatic algorithm configuration. Oper. Res. Perspect. 2016, 3, 43–58. [Google Scholar] [CrossRef]
  98. Akiba, T.; Sano, S.; Yanase, T.; Ohta, T.; Koyama, M. Optuna: A Next-generation Hyperparameter Optimization Framework. In Proceedings of the KDD ’19: 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, Anchorage, AK, USA, 4–8 August 2019; Association for Computing Machinery: New York, NY, USA, 2019; pp. 2623–2631. [Google Scholar] [CrossRef]
  99. Liaw, R.; Liang, E.; Nishihara, R.; Moritz, P.; Gonzalez, J.E.; Stoica, I. Tune: A Research Platform for Distributed Model Selection and Training. arXiv 2018, arXiv:1807.05118. [Google Scholar] [CrossRef]
  100. Spirals Research Group. PyJoules: A Python Library to Capture the Energy Consumption of Code Snippets; University of Lille and Inria: Lille, France, 2021. [Google Scholar]
  101. Anthony, L.F.W.; Kanding, B.; Selvan, R. Carbontracker: Tracking and Predicting the Carbon Footprint of Training Deep Learning Models. arXiv 2020, arXiv:2007.03051. [Google Scholar] [CrossRef]
  102. Ramduny, J.; Garcia, M.; Kelly, C. Establishing a reproducible and sustainable analysis workflow. In Methods for Analyzing Large Neuroimaging Datasets; Springer: New York, NY, USA, 2024; pp. 39–60. [Google Scholar]
  103. Deng, L. The mnist database of handwritten digit images for machine learning research. IEEE Signal Process. Mag. 2012, 29, 141–142. [Google Scholar] [CrossRef]
  104. Deng, J.; Dong, W.; Socher, R.; Li, L.J.; Li, K.; Fei-Fei, L. ImageNet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
  105. Egwutuoha, I.P.; Levy, D.; Selic, B.; Chen, S. A survey of fault tolerance mechanisms and checkpoint/restart implementations for high performance computing systems. J. Supercomput. 2013, 65, 1302–1326. [Google Scholar] [CrossRef]
  106. Moran, M.; Balladini, J.; Rexachs, D.; Luque, E. Checkpoint and Restart: An Energy Consumption Characterization in Clusters. arXiv 2024, arXiv:2409.02214. [Google Scholar] [CrossRef]
  107. Kumar, M.; Gupta, S.; Patel, T.; Wilder, M.; Shi, W.; Fu, S.; Engelmann, C.; Tiwari, D. Study of interconnect errors, network congestion, and applications characteristics for throttle prediction on a large scale HPC system. J. Parallel Distrib. Comput. 2021, 153, 29–43. [Google Scholar] [CrossRef]
  108. Jiao, Y.; Lin, H.; Balaji, P.; Feng, W. Power and Performance Characterization of Computational Kernels on the GPU. In Proceedings of the 2010 IEEE/ACM Int’l Conference on Green Computing and Communications & Int’l Conference on Cyber, Physical and Social Computing, Hangzhou, China, 18–20 December 2010; pp. 221–228. [Google Scholar]
  109. Timalsina, M.; Gerhardt, L.; Tyler, N.; Blaschke, J.P.; Arndt, W. Optimizing Checkpoint-Restart Mechanisms for HPC with DMTCP in Containers at NERSC. arXiv 2024, arXiv:2407.19117. [Google Scholar] [CrossRef]
  110. Assogba, K.; Nicolae, B.; Van Dam, H.; Rafique, M.M. Asynchronous Multi-Level Checkpointing: An Enabler of Reproducibility using Checkpoint History Analytics. In Proceedings of the SC-W ’23: SC ’23 Workshops of the International Conference on High Performance Computing, Network, Storage, and Analysis, Denver, CO, USA, 12–17 November 2023; Association for Computing Machinery: New York, NY, USA, 2023; pp. 1748–1756. [Google Scholar] [CrossRef]
  111. Rojas, E.; Kahira, A.N.; Meneses, E.; Gomez, L.B.; Badia, R.M. A Study of Checkpointing in Large Scale Training of Deep Neural Networks. arXiv 2021, arXiv:2012.00825. [Google Scholar] [CrossRef]
  112. Gu, R.; Chen, Y.; Liu, S.; Dai, H.; Chen, G.; Zhang, K.; Che, Y.; Huang, Y. Liquid: Intelligent Resource Estimation and Network-Efficient Scheduling for Deep Learning Jobs on Distributed GPU Clusters. IEEE Trans. Parallel Distrib. Syst. 2022, 33, 2808–2820. [Google Scholar] [CrossRef]
  113. Daradkeh, T.; Roper, G.; Alarcon Meza, C.; Mokhov, S.A. HPC Jobs Classification and Resource Prediction to Minimize Job Failures. In Proceedings of the CompSysTech ’24: International Conference on Computer Systems and Technologies 2024, Ruse, Bulgaria, 14–15 June 2024; ACM: New York, NY, USA, 2024; pp. 95–101. [Google Scholar] [CrossRef]
  114. Tanash, M.; Yang, H.; Andresen, D.; Hsu, W. Ensemble Prediction of Job Resources to Improve System Performance for Slurm-Based HPC Systems. In Proceedings of the PEARC ’21: Practice and Experience in Advanced Research Computing, Boston, MA, USA, 18–22 July 2021; ACM: New York, NY, USA, 2021; pp. 1–8. [Google Scholar] [CrossRef]
  115. Friedman, M. The Use of Ranks to Avoid the Assumption of Normality Implicit in the Analysis of Variance. J. Am. Stat. Assoc. 1937, 32, 675–701. [Google Scholar] [CrossRef]
  116. Soft Computing and Intelligent Information Systems. 2025. Available online: https://sci2s.ugr.es/sicidm (accessed on 14 January 2025).
  117. Demšar, J. Statistical Comparisons of Classifiers over Multiple Data Sets. J. Mach. Learn. Res. 2006, 7, 1–30. [Google Scholar]
Figure 1. This figure shows the basic steps of differential evolution, including initialization, crossover, mutation, selection, and condition (loop). Until the stopping condition is met, i.e., max generations or convergence, the mutation, crossover, and selection processes are repeated over generations; then, the best solution that is found is provided [54].
Figure 1. This figure shows the basic steps of differential evolution, including initialization, crossover, mutation, selection, and condition (loop). Until the stopping condition is met, i.e., max generations or convergence, the mutation, crossover, and selection processes are repeated over generations; then, the best solution that is found is provided [54].
Mathematics 13 01681 g001
Figure 2. Basic workflow of job submission through the Slurm workload manager.
Figure 2. Basic workflow of job submission through the Slurm workload manager.
Mathematics 13 01681 g002
Figure 3. SBATCH script for job submission.
Figure 3. SBATCH script for job submission.
Mathematics 13 01681 g003
Figure 4. Schematic overview of the analytics of the potential efficiency for the AutoDEHypO workflow.
Figure 4. Schematic overview of the analytics of the potential efficiency for the AutoDEHypO workflow.
Mathematics 13 01681 g004
Figure 5. Plots of the initially obtained results of the ML metrics on the CIFAR10 dataset.
Figure 5. Plots of the initially obtained results of the ML metrics on the CIFAR10 dataset.
Mathematics 13 01681 g005
Figure 6. Plots of the initially obtained results of the ML metrics on the CIFAR100 dataset.
Figure 6. Plots of the initially obtained results of the ML metrics on the CIFAR100 dataset.
Mathematics 13 01681 g006
Figure 7. Accuracy convergences through runtime, for the median runs according to accuracy metric.
Figure 7. Accuracy convergences through runtime, for the median runs according to accuracy metric.
Mathematics 13 01681 g007
Figure 8. The statistical measures of the mean for the elapsed time (a), consumed energy (b), and accuracy (c) from the results obtained from 10 runs of 15 epochs of ML models that were hyperoptimized using the different DE strategies presented in Table A1, Table A2, Table A3 and Table A4. The x-axis represents the evaluated DE strategies, while the y-axis shows the ML models used.
Figure 8. The statistical measures of the mean for the elapsed time (a), consumed energy (b), and accuracy (c) from the results obtained from 10 runs of 15 epochs of ML models that were hyperoptimized using the different DE strategies presented in Table A1, Table A2, Table A3 and Table A4. The x-axis represents the evaluated DE strategies, while the y-axis shows the ML models used.
Mathematics 13 01681 g008aMathematics 13 01681 g008b
Figure 9. The statistical measures of the mean for the elapsed time (a), consumed energy (b), and accuracy (c) from the results obtained from 10 runs of 15 epochs of ML models that were hyperoptimized using the different DE strategies presented in Table A5, Table A6, Table A7 and Table A8. The x-axis represents the evaluated DE strategies, while the y-axis shows the ML models used.
Figure 9. The statistical measures of the mean for the elapsed time (a), consumed energy (b), and accuracy (c) from the results obtained from 10 runs of 15 epochs of ML models that were hyperoptimized using the different DE strategies presented in Table A5, Table A6, Table A7 and Table A8. The x-axis represents the evaluated DE strategies, while the y-axis shows the ML models used.
Mathematics 13 01681 g009aMathematics 13 01681 g009b
Table 1. The experimental environment outlines ML model training parameters, and essential configuration to ensure hyperparameter optimization setup and reproducibility of this setup.
Table 1. The experimental environment outlines ML model training parameters, and essential configuration to ensure hyperparameter optimization setup and reproducibility of this setup.
ParameterValueDefault
Differential Evolution Optimization
Population size ( P ) 10-
Maximum generations ( G ) 30-
Dimensions ( D ) 2-
Mutationgeneration-randomized in [0.5, 1]True
Crossover ( C r ) 0.7True
Repetitions per DE strategy100
DE Strategiesrand1bin, rand2bin, best1bin, best2bin, currenttobest1bin, currenttobest1exp, rand1exp, rand2exp, best1exp, best2expbest1bin
Learning rate[search space]-
Model and Dataset Configuration
ML Models (Architecture)ResNet18, VGG11, ConvNeXt-Small, DenseNet121-
Pretrained weightsImageNet-1kTrue
DatasetsCIFAR-10, CIFAR-100-
Input size 32 × 32 × 3 (H × W × C)True
Batch size16, 32, 64, 128, 25616
Epochs per run151
Workers per run41
Sampler shuffling (train loader)Shuffle enabled (distributed sampler)True
Sampler shuffling (test loader)Shuffle disabled (no sampler)False
Loss functionBCEWithLogitsLoss-
Evaluation metricsaccuracy, precision, recall, F1-score-
FrameworkPyTorch 2.1.2-
Python version3.10.12-
HPC Environment
Number of nodes per run11
Number of cores per run1281
Number of GPUs per run41
Memory per run0 (all)2 G
Partitiongpucpu
NVIDIA Driver, CUDA version565.57.01, 12.7True
Table 2. Results obtained in 15 epochs using the ResNet18 ML model on CIFAR10.
Table 2. Results obtained in 15 epochs using the ResNet18 ML model on CIFAR10.
Batch Size Max Accuracy %Best LRBest Accuracy Consumed CPU Time Elapsed Time Consumed Energy
1695.394%0.00060820149663338530.195921 d 02:55:3606:43:548.26 MJ
3296.008%0.00079010804734894020.195341 d 01:54:5606:28:447.94 MJ
6496.296%0.0007381244095364610.196714:50:0003:42:304.52 MJ
12896.042%0.000348714254562946970.1957415:23:3203:50:534.73 MJ
25697.988%0.00026727959265821790.1968414:07:4403:31:564.33 MJ
Table 3. Results obtained in 15 epochs using the VGG11 ML model on CIFAR10.
Table 3. Results obtained in 15 epochs using the VGG11 ML model on CIFAR10.
Batch Size Max Accuracy %Best LRBest Accuracy Consumed CPU Time Elapsed Time Consumed Energy
1697.396%0.000125251673167103550.197062 d 00:09:1612:02:1919.80 MJ
3297.182%0.0002688949516381270.197141 d 03:35:2406:53:5111.03 MJ
6497.408%0.000240772301800391150.1977419:02:3604:45:397.37 MJ
12897.686%0.000194286189592430840.1982415:15:1203:48:485.46 MJ
25697.584%0.000348714254562946970.1975818:09:4804:32:276.40 MJ
Table 4. Results obtained in 15 epochs using the ConvNeXtSmall ML model on CIFAR10.
Table 4. Results obtained in 15 epochs using the ConvNeXtSmall ML model on CIFAR10.
Batch Size Max Accuracy %Best LRBest Accuracy Consumed CPU Time Elapsed Time Consumed Energy
1698.056%0.000120348455294465370.19724 d 21:06:561 d 05:16:4437.61 MJ
3297.982%0.000140148940063677880.198063 d 08:00:0820:00:0225.60 MJ
6498.486%0.00086789286879177020.198361 d 05:25:3607:21:249.73 MJ
12898.574%0.000281358581111197030.1976823:18:4405:49:418.01 MJ
25699.17%0.00051487381018402510.1991620:23:2405:05:517.05 MJ
Table 5. Results obtained in 15 epochs using the DenseNet121 ML model on CIFAR10.
Table 5. Results obtained in 15 epochs using the DenseNet121 ML model on CIFAR10.
Batch Size Max Accuracy %Best LRBest Accuracy Consumed CPU Time Elapsed Time Consumed Energy
1695.03%0.00079126463264313010.19358 d 00:01:082 d 00:00:1757.02 MJ
3295.058%0.00079010804734894020.195463 d 12:20:1621:05:0427.31 MJ
6495.04%0.00048945331834283030.195121 d 23:24:5611:51:1415.34 MJ
12895.208%0.000348714254562946970.195821 d 01:45:2406:26:218.89 MJ
25694.8%0.00079010804734894020.1954219:09:1204:47:186.43 MJ
Table 6. Results obtained in 15 epochs using the ResNet18 ML model on CIFAR100.
Table 6. Results obtained in 15 epochs using the ResNet18 ML model on CIFAR100.
Batch Size Max Accuracy %Best LRBest Accuracy Consumed CPU Time Elapsed Time Consumed Energy
1679.978%0.000448563359543860150.18381 d 21:46:5211:26:4314.57 MJ
3286.23%0.00087174609166037770.184422:44:2805:41:077.08 MJ
6487.168%0.00053918938389076420.18719:26:2404:51:365.93 MJ
12885.288%0.00060820149663338530.189615:32:3603:53:094.77 MJ
25683.758%0.00024228749016532070.189614:34:2803:38:374.43 MJ
Table 7. Results obtained in 15 epochs using the VGG11 ML model on CIFAR100.
Table 7. Results obtained in 15 epochs using the VGG11 ML model on CIFAR100.
Batch Size Max Accuracy %Best LRBest Accuracy Consumed CPU Time Elapsed Time Consumed Energy
1683.336%0.000125251673167103550.18281 d 21:07:4811:16:5718.52 MJ
3282.176%0.00020925116218056280.18741 d 00:45:4806:11:2710.45 MJ
6481.9%0.000125251673167103550.188215:41:0003:55:155.98 MJ
12876.106%0.000132447958668769260.182221:25:2805:21:227.60 MJ
25662.56%0.000281358581111197030.172415:04:1603:46:045.22 MJ
Table 8. Results obtained in 15 epochs using the ConvNeXtSmall ML model on CIFAR100.
Table 8. Results obtained in 15 epochs using the ConvNeXtSmall ML model on CIFAR100.
Batch Size Max Accuracy %Best LRBest Accuracy Consumed CPU Time Elapsed Time Consumed Energy
1697.086%0.00037841569281883820.19885 d 23:48:481 d 11:57:1245.13 MJ
3297.494%0.000125251673167103550.19841 d 03:46:4806:56:428.86 MJ
6497.954%0.000238946523073126130.1981 d 23:54:1611:58:3415.65 MJ
12898.342%0.000161776904304399080.19921 d 07:00:0807:45:0210.65 MJ
25698.784%0.000281358581111197030.19915:59:3603:59:545.97 MJ
Table 9. Results obtained in 15 epochs using the DenseNet121 ML model on CIFAR100.
Table 9. Results obtained in 15 epochs using the DenseNet121 ML model on CIFAR100.
Batch Size Max Accuracy %Best LRBest Accuracy Consumed CPU Time Elapsed Time Consumed Energy
1678.902%0.00079126463264313010.1877 d 23:11:161 d 23:47:4959.36 MJ
3283.004%0.00056771732632273560.18543 d 17:02:5622:15:4426.45 MJ
6482.758%0.00041644293427330910.18642 d 03:18:2812:49:3715.97 MJ
12883.272%0.00066008287052118150.18141 d 09:56:5208:29:1311.57 MJ
25679.366%0.00048945331834283030.18120:16:4005:04:107.05 MJ
Table 10. Statistical analysis of elapsed time in the DE strategies using a non-parametric Friedman test with the maximum rank distribution and corresponding post hoc procedures (rejected, as marked in bold and with sign, at a statistical value below the threshold of 0.005555555555555556 for Bonferroni–Dunn, 0.05 for Holm and Hommel, 0.025 for Hochberg and Rom, 0.050000000000000044 for Holland and Finner, and 2.9216395384872545 × 10−17 for Li, respectively).
Table 10. Statistical analysis of elapsed time in the DE strategies using a non-parametric Friedman test with the maximum rank distribution and corresponding post hoc procedures (rejected, as marked in bold and with sign, at a statistical value below the threshold of 0.005555555555555556 for Bonferroni–Dunn, 0.05 for Holm and Hommel, 0.025 for Hochberg and Rom, 0.050000000000000044 for Holland and Finner, and 2.9216395384872545 × 10−17 for Li, respectively).
iDE Strategy z = ( R 0 R i ) / SE pHolm/Hochberg/HommelHolland Rom Finner Li
9best1bin5.321.05 × 10−70.00556 0.00568 0.00584 0.00568 2.92 × 10−17 
8best1exp5.321.05 × 10−70.00625 0.00639 0.00657 0.0113 2.92 × 10−17 
7currenttobest1bin3.919.07 × 10−50.00714 0.0073 0.00751 0.017 2.92 × 10−17 
6currenttobest1exp2.920.003530.00833 0.00851 0.00876 0.0225 2.92 × 10−17 
5best2exp2.620.008750.01 0.0102 0.0105 0.0281 2.92 × 10−17 
4best2bin2.510.0120.0125 0.0127 0.0131 0.0336 2.92 × 10−17 
3rand1bin2.440.01480.0167 0.017 0.0167 0.0391 2.92 × 10−17 
2rand1exp2.290.02210.025 0.0253 0.025 0.0446 2.92 × 10−17 
1rand2bin6.56 × 10−1610.05 0.05 0.05 0.05 0.05
Table 11. Statistical analysis of energy consumed in the DE strategies using the non-parametric Friedman test with the minimum rank distribution and corresponding post hoc procedures (rejected, as marked in bold and with sign, at a statistical value below the threshold at 0.005555555555555556 for Bonferroni–Dunn, 0.016666666666666666 for Holm and Hommel, 0.0125 for Hochberg, 0.016952427508441503 for Holland, 0.013109375000000001 for Rom, 0.044570249746389234 for Finner, and 0.011483473500591115 for Li, respectively).
Table 11. Statistical analysis of energy consumed in the DE strategies using the non-parametric Friedman test with the minimum rank distribution and corresponding post hoc procedures (rejected, as marked in bold and with sign, at a statistical value below the threshold at 0.005555555555555556 for Bonferroni–Dunn, 0.016666666666666666 for Holm and Hommel, 0.0125 for Hochberg, 0.016952427508441503 for Holland, 0.013109375000000001 for Rom, 0.044570249746389234 for Finner, and 0.011483473500591115 for Li, respectively).
iDE Strategy z = ( R 0 R i ) / SE pHolm/Hochberg/HommelHolland Rom Finner Li
9rand2bin5.113.15 × 10−7 0.00556 0.00568 0.00584 0.00568 0.0115 
8rand2exp4.996.19 × 10−7 0.00625 0.00639 0.00657 0.0113 0.0115 
7best2bin2.730.00628 0.00714 0.0073 0.00751 0.017 0.0115 
6rand1bin2.730.00628 0.00833 0.00851 0.00876 0.02250.0115 
5best2exp2.580.00974 0.01 0.0102 0.0105 0.02810.0115 
4rand1exp2.550.0108 0.0125 0.0127 0.0131 0.03360.0115 
3currenttobest1exp2.140.03220.0167 0.017 0.0167 0.03910.0115 
2currenttobest1bin0.7020.4830.0250.02530.0250.04460.0115 
1best1exp0.2770.7820.050.050.050.050.05
Table 12. Statistical analysis of the accuracy in the DE strategies using the non-parametric Friedman test with the minimum rank distribution and corresponding post hoc procedures (rejected, as marked in bold and with sign, at a statistical value below the threshold at 0.005555555555555556 for Bonferroni–Dunn, Holm, and Hommel, 0.005683044988048058 for Holland and Finner, and 7.7532015972022 ×10−4 for Li, respectively).
Table 12. Statistical analysis of the accuracy in the DE strategies using the non-parametric Friedman test with the minimum rank distribution and corresponding post hoc procedures (rejected, as marked in bold and with sign, at a statistical value below the threshold at 0.005555555555555556 for Bonferroni–Dunn, Holm, and Hommel, 0.005683044988048058 for Holland and Finner, and 7.7532015972022 ×10−4 for Li, respectively).
i DE Strategy z = ( R 0 R i ) / SE pHolm/Hochberg/Hommel Holland Rom Finner Li
9rand1exp1.850.06480.00556 0.00568 0.005840.00568 0.000775
8rand2bin1.50.1350.006250.006390.006570.01130.000775
7rand1bin1.440.150.007140.00730.007510.0170.000775
6currenttobest1bin1.290.1960.008330.008510.008760.02250.000775
5rand2exp0.960.3370.010.01020.01050.02810.000775
4best2bin0.5350.5920.01250.01270.01310.03360.000775
3currenttobest1exp0.480.6310.01670.0170.01670.03910.000775
2best1bin0.240.810.0250.02530.0250.04460.000775
1best2exp0.01850.9850.050.050.050.050.05
Table 13. Statistical analysis of the elapsed time in the DE strategies using the non-parametric Friedman test with the maximum rank distribution and corresponding post hoc procedures (rejected, as marked in bold and with sign, at a statistical value below the threshold of 0.005555555555555556 for Bonferroni–Dunn, Holm, and Hommel, 0.005683044988048058 for Holland and Finner, and 0.015161939437253264 for Li, respectively).
Table 13. Statistical analysis of the elapsed time in the DE strategies using the non-parametric Friedman test with the maximum rank distribution and corresponding post hoc procedures (rejected, as marked in bold and with sign, at a statistical value below the threshold of 0.005555555555555556 for Bonferroni–Dunn, Holm, and Hommel, 0.005683044988048058 for Holland and Finner, and 0.015161939437253264 for Li, respectively).
i DE Strategy z = ( R 0 R i ) / SE pHolm/Hochberg/Hommel Holland Rom Finner Li
9best1exp2.220.02670.00556 0.00568 0.005840.00568 0.0151 
8best2bin2.180.02940.006250.006390.006570.01130.0151 
7best1bin2.10.03530.007140.00730.007510.0170.0151 
6currenttobest1bin1.550.1210.008330.008510.008760.02250.0151 
5currenttobest1exp1.030.3010.010.01020.01050.02810.0151 
4rand1bin0.4430.6580.01250.01270.01310.03360.0151 
3rand1exp0.4060.6850.01670.0170.01670.03910.0151 
2best2exp0.4060.6850.0250.02530.0250.04460.0151 
1rand2exp0.3690.7120.050.050.050.050.05
Table 14. Statistical analysis of the energy consumed in the DE strategies using the non-parametric Friedman test with the minimum rank distribution and corresponding post hoc procedures (rejected, as marked in bold and with sign, at a statistical value below the threshold of 0.005555555555555556 for Bonferroni–Dunn, Holm, and Hommel, 0.005683044988048058 for Holland and Finner, and 0.014435245666076669 for Li, respectively).
Table 14. Statistical analysis of the energy consumed in the DE strategies using the non-parametric Friedman test with the minimum rank distribution and corresponding post hoc procedures (rejected, as marked in bold and with sign, at a statistical value below the threshold of 0.005555555555555556 for Bonferroni–Dunn, Holm, and Hommel, 0.005683044988048058 for Holland and Finner, and 0.014435245666076669 for Li, respectively).
i DE Strategy z = ( R 0 R i ) / SE pHolm/Hochberg/Hommel Holland Rom Finner Li
9rand2bin2.070.03860.00556 0.00568 0.005840.00568 0.0144 
8rand1bin1.980.04820.006250.006390.006570.01130.0144 
7rand2exp1.790.07330.007140.00730.007510.0170.0144 
6best2exp1.440.150.008330.008510.008760.02250.0144 
5currenttobest1exp0.9790.3280.010.01020.01050.02810.0144 
4rand1exp0.8120.4170.01250.01270.01310.03360.0144 
3currenttobest1bin0.6830.4950.01670.0170.01670.03910.0144 
2best2bin0.4250.6710.0250.02530.0250.04460.0144 
1best1exp0.3510.7260.050.050.050.050.05
Table 15. Statistical analysis of the accuracy in the DE strategies using the non-parametric Friedman test with the minimum rank distribution and corresponding post hoc procedures (rejected, as marked in bold and with sign, at a statistical value below the threshold of 0.005555555555555556 for Bonferroni–Dunn and Hochberg, 0.00625 for Holm and Hommel, 0.006391150954545011 for Holland, 0.005843911024153359 for Rom, 0.011333792975759982 for Finner, and 0.03180542233195395 for Li, respectively).
Table 15. Statistical analysis of the accuracy in the DE strategies using the non-parametric Friedman test with the minimum rank distribution and corresponding post hoc procedures (rejected, as marked in bold and with sign, at a statistical value below the threshold of 0.005555555555555556 for Bonferroni–Dunn and Hochberg, 0.00625 for Holm and Hommel, 0.006391150954545011 for Holland, 0.005843911024153359 for Rom, 0.011333792975759982 for Finner, and 0.03180542233195395 for Li, respectively).
i DE Strategy z = ( R 0 R i ) / SE pHolm/Hochberg/Hommel Holland Rom Finner Li
9rand1bin2.810.00501 0.00556 0.00568 0.00584 0.00568 0.0318 
8rand2exp2.230.02550.00625 0.00639 0.006570.01130.0318 
7currenttobest1exp2.20.0280.007140.00730.007510.0170.0318 
6currenttobest1bin1.960.05030.008330.008510.008760.02250.0318 
5rand2bin1.460.1450.010.01020.01050.02810.0318 
4best1bin1.290.1960.01250.01270.01310.03360.0318 
3best2exp1.270.2030.01670.0170.01670.03910.0318 
2best1exp1.260.2090.0250.02530.0250.04460.0318 
1best2bin0.8490.3960.050.050.050.050.05
Table 16. Statistical analysis of the elapsed time in the DE strategies using the non-parametric Friedman test with the maximum rank distribution and corresponding post hoc procedures (rejected, as marked in bold and with sign, at a statistical value below the threshold of 0.005555555555555556 for Bonferroni–Dunn, 0.0125 for Holm and Hommel, 0.01 for Hochberg, 0.012741455098566168 for Holland, 0.010515350115740741 for Rom, 0.039109465610866256 for Finner, and 0.010841964200380961 for Li, respectively).
Table 16. Statistical analysis of the elapsed time in the DE strategies using the non-parametric Friedman test with the maximum rank distribution and corresponding post hoc procedures (rejected, as marked in bold and with sign, at a statistical value below the threshold of 0.005555555555555556 for Bonferroni–Dunn, 0.0125 for Holm and Hommel, 0.01 for Hochberg, 0.012741455098566168 for Holland, 0.010515350115740741 for Rom, 0.039109465610866256 for Finner, and 0.010841964200380961 for Li, respectively).
i DE Strategy z = ( R 0 R i ) / SE pHolm/Hochberg/Hommel Holland Rom Finner Li
9best1exp5.331 × 10−70.00556 0.00568 0.00584 0.00568 0.0108 
8best1bin5.251.53 × 10−70.00625 0.00639 0.00657 0.0113 0.0108 
7currenttobest1bin3.860.0001110.00714 0.0073 0.00751 0.0170.0108 
6best2bin3.320.0009130.00833 0.00851 0.00876 0.02250.0108 
5currenttobest1exp2.790.005210.01 0.0102 0.0105 0.02810.0108 
4best2exp2.140.03230.0125 0.0127 0.01310.03360.0108 
3rand1bin2.040.04170.01670.0170.01670.03910.0108 
2rand1exp1.910.05660.0250.02530.0250.04460.0108 
1rand2exp0.2610.7940.050.050.050.050.05
Table 17. Statistical analysis of the energy consumed in the DE strategies using the non-parametric Friedman test with the maximum rank distribution and corresponding post hoc procedures (rejected, as marked in bold and with sign, at a statistical value below the threshold of 0.005555555555555556 for Bonferroni–Dunn, 0.0125 for Hochberg, 0.016666666666666666 for Holm and Hommel, 0.016952427508441503 for Holland, 0.013109375000000001 for Rom, 0.044570249746389234 for Finner, and 0.011898018581242242 for Li, respectively).
Table 17. Statistical analysis of the energy consumed in the DE strategies using the non-parametric Friedman test with the maximum rank distribution and corresponding post hoc procedures (rejected, as marked in bold and with sign, at a statistical value below the threshold of 0.005555555555555556 for Bonferroni–Dunn, 0.0125 for Hochberg, 0.016666666666666666 for Holm and Hommel, 0.016952427508441503 for Holland, 0.013109375000000001 for Rom, 0.044570249746389234 for Finner, and 0.011898018581242242 for Li, respectively).
i DE Strategy z = ( R 0 R i ) / SE pHolm/Hochberg/Hommel Holland Rom Finner Li
9best1bin5.083.8 × 10−70.00556 0.00568 0.00584 0.00568 0.01189 
8best1exp4.633.57 × 10−60.00625 0.00639 0.00657 0.0113 0.01189 
7currenttobest1bin4.14.14 × 1000.00714 0.0073 0.00751 0.017 0.01189 
6currenttobest1exp2.870.004080.00833 0.00851 0.00876 0.02250.01189 
5best2bin2.850.004420.01 0.0102 0.0105 0.02810.01189 
4rand1exp2.70.006880.0125 0.0127 0.0131 0.03360.01189 
3best2exp2.230.02560.0167 0.017 0.0167 0.03910.01189 
2rand1bin1.750.08020.0250.02530.0250.04460.01189 
1rand2exp0.2870.7740.050.050.050.050.05
Table 18. Statistical analysis of the accuracy in the DE strategies using the non-parametric Friedman test with the minimum rank distribution and corresponding post hoc procedures (rejected, as marked in bold and with sign, at a statistical value below the threshold of 0.005555555555555556 for Bonferroni–Dunn, Holm and Hommel, 0.005683044988048058 for Holland, 0.005683044988048058 for Finner, and 0.0010964059917403937 for Li, respectively).
Table 18. Statistical analysis of the accuracy in the DE strategies using the non-parametric Friedman test with the minimum rank distribution and corresponding post hoc procedures (rejected, as marked in bold and with sign, at a statistical value below the threshold of 0.005555555555555556 for Bonferroni–Dunn, Holm and Hommel, 0.005683044988048058 for Holland, 0.005683044988048058 for Finner, and 0.0010964059917403937 for Li, respectively).
i DE Strategy z = ( R 0 R i ) / SE pHolm/Hochberg/Hommel Holland Rom Finner Li
9rand1bin2.120.03440.00556 0.00568 0.005840.00568 0.00109 
8currenttobest1bin1.410.1590.006250.006390.006570.01130.00109 
7rand2exp1.370.170.007140.00730.007510.0170.00109 
6rand2bin1.20.230.008330.008510.008760.02250.00109 
5currenttobest1exp1.010.3150.010.01020.01050.02810.00109 
4rand1exp0.4180.6760.01250.01270.01310.03360.00109 
3best1bin0.1960.8450.01670.0170.01670.03910.00109 
2best2bin0.09140.9270.0250.02530.0250.04460.00109 
1best2exp0.02610.9790.050.050.050.050.05
Table 19. Overview of the outcomes of the statistical analysis for the metrics of elapsed time, energy efficiency, and accuracy in the DE strategies using the non-parametric Friedman test on both the CIFAR10 and CIFAR00 datasets.
Table 19. Overview of the outcomes of the statistical analysis for the metrics of elapsed time, energy efficiency, and accuracy in the DE strategies using the non-parametric Friedman test on both the CIFAR10 and CIFAR00 datasets.
DE StrategyHolm/Hochberg/HommelHollandRomFinnerLi
best1bin33335
best2bin44428
best1exp44345
best2exp44416
currenttobest1bin33328
currenttobest1exp44418
rand1bin44438
rand2bin33232
rand1exp44327
rand2exp22114
Table 20. Friedman rankings detected differences in elapsed time, consumed energy, and accuracy aggregated across ML models. The individual rankings demonstrate significant differences using the Holm/Hochberg/Hommel, Holland, Rom, Finner, and Li procedures, with a few exceptions for the post hoc procedures (the Rom and Li procedures for time and energy detect these differences only when comparing VGG11 and DenseNet121 against ResNet18 and ConvNextSmall, i.e., for 41 (91.1%) out of 45 rankings in the post hoc procedures, we successfully detected significant differences).
Table 20. Friedman rankings detected differences in elapsed time, consumed energy, and accuracy aggregated across ML models. The individual rankings demonstrate significant differences using the Holm/Hochberg/Hommel, Holland, Rom, Finner, and Li procedures, with a few exceptions for the post hoc procedures (the Rom and Li procedures for time and energy detect these differences only when comparing VGG11 and DenseNet121 against ResNet18 and ConvNextSmall, i.e., for 41 (91.1%) out of 45 rankings in the post hoc procedures, we successfully detected significant differences).
ML ModelElapsed TimeConsumed EnergyAccuracy
resnet181.5050000000000011.48000000000000061.394999999999994
convnextsmall1.49499999999999061.52999999999999072.6374999999999935
densenet1213.32500000000000063.29250000000000132.6725000000000008
vgg113.675000000000013.69750000000000963.295000000000002
Table 21. Ranks and confidence intervals for key metrics elapsed time, consumed energy, and accuracy across ML Models ResNet18, ConvNextSmall, DenseNet121, and VGG11 on CIFAR10 and CIFAR100 datasets.
Table 21. Ranks and confidence intervals for key metrics elapsed time, consumed energy, and accuracy across ML Models ResNet18, ConvNextSmall, DenseNet121, and VGG11 on CIFAR10 and CIFAR100 datasets.
ML Model PairMetricRank DifferenceConfidence Interval (CI)Significant
resnet18—convnextsmallelapsed time0.010000[−0.967, 0.987]No
resnet18—densenet121elapsed time−1.820000[−2.797, −0.843]Yes
resnet18—vgg11elapsed time−2.170000[−3.147, −1.193]Yes
convnextsmall—densenet121elapsed time−1.830000[−2.807, −0.853]Yes
convnextsmall—vgg11elapsed time−2.180000[−3.157, −1.203]Yes
densenet121—vgg11elapsed time−0.350000[−1.327, 0.627]No
resnet18—convnextsmallconsumed energy−0.050000[−1.027, 0.927]No
resnet18—densenet121consumed energy−1.813000[−2.79, −0.836]Yes
resnet18—vgg11consumed energy−2.218000[−3.195, −1.241]Yes
convnextsmall—densenet121consumed energy−1.763000[−2.74, −0.786]Yes
convnextsmall—vgg11consumed energy−2.168000[−3.145, −1.191]Yes
densenet121—vgg11consumed energy−0.405000[−1.382, 0.572]No
resnet18—convnextsmallaccuracy−1.242000[−2.219, −0.265]Yes
resnet18—densenet121accuracy−1.278000[−2.255, −0.301]Yes
resnet18—vgg11accuracy−1.900000[−2.877, −0.923]Yes
convnextsmall—densenet121accuracy−0.036000[−1.013, 0.941]No
convnextsmall—vgg11accuracy−0.658000[−1.635, 0.319]No
densenet121—vgg11accuracy−0.622000[−1.599, 0.355]No
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Prica, T.; Zamuda, A. High-Performance Deployment Operational Data Analytics of Pre-Trained Multi-Label Classification Architectures with Differential-Evolution-Based Hyperparameter Optimization (AutoDEHypO). Mathematics 2025, 13, 1681. https://doi.org/10.3390/math13101681

AMA Style

Prica T, Zamuda A. High-Performance Deployment Operational Data Analytics of Pre-Trained Multi-Label Classification Architectures with Differential-Evolution-Based Hyperparameter Optimization (AutoDEHypO). Mathematics. 2025; 13(10):1681. https://doi.org/10.3390/math13101681

Chicago/Turabian Style

Prica, Teo, and Aleš Zamuda. 2025. "High-Performance Deployment Operational Data Analytics of Pre-Trained Multi-Label Classification Architectures with Differential-Evolution-Based Hyperparameter Optimization (AutoDEHypO)" Mathematics 13, no. 10: 1681. https://doi.org/10.3390/math13101681

APA Style

Prica, T., & Zamuda, A. (2025). High-Performance Deployment Operational Data Analytics of Pre-Trained Multi-Label Classification Architectures with Differential-Evolution-Based Hyperparameter Optimization (AutoDEHypO). Mathematics, 13(10), 1681. https://doi.org/10.3390/math13101681

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop