You are currently viewing a new version of our website. To view the old version click .
Mathematics
  • Article
  • Open Access

20 May 2025

High-Performance Deployment Operational Data Analytics of Pre-Trained Multi-Label Classification Architectures with Differential-Evolution-Based Hyperparameter Optimization (AutoDEHypO)

and
1
Institute of Information Science, Prešernova ulica 17, 2000 Maribor, Slovenia
2
Faculty of Electrical Engineering and Computer Science, University of Maribor, Koroška cesta 46, 2000 Maribor, Slovenia
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Innovations in High-Performance Computing

Abstract

This article presents a high-performance-computing differential-evolution-based hyperparameter optimization automated workflow (AutoDEHypO), which is deployed on a petascale supercomputer and utilizes multiple GPUs to execute a specialized fitness function for machine learning (ML). The workflow is designed for operational analytics of energy efficiency. In this differential evolution (DE) optimization use case, we analyze how energy efficiently the DE algorithm performs with different DE strategies and ML models. The workflow analysis considers key factors such as DE strategies and automated use case configurations, such as an ML model architecture and dataset, while monitoring both the achieved accuracy and the utilization of computing resources, such as the elapsed time and consumed energy. While the efficiency of a chosen DE strategy is assessed based on a multi-label supervised ML accuracy, operational data about the consumption of resources of individual completed jobs obtained from a Slurm database are reported. To demonstrate the impact on energy efficiency, using our analysis workflow, we visualize the obtained operational data and aggregate them with statistical tests that compare and group the energy efficiency of the DE strategies applied in the ML models.

1. Introduction

High-performance computing (HPC) continues to advance and to provide important infrastructure and a backbone for the scientific community, industry, and beyond, making efficient utilization highly important for both environmental sustainability and cost efficiency [1,2]. Despite the increasing availability of automated ML (AutoML) frameworks, most existing tools prioritize and maximize the performance (accuracy) and neglect other important objectives such as energy efficiency (EE) and resource usage and optimization, which are key factors in large-scale HPC environments and beyond [3,4]. Therefore, the enforcement of sustainable and energy-efficient solutions, including the utilization and optimization of computational resources, now plays a key role in reducing operational financial costs and minimizing environmental impacts [1,3]. Due to the broad accessibility of computational resources for HPC, the scientific community with the myriad of application areas and diverse expertise from different fields, can deploy and run their workflows [5,6]. As their needs, demands, and complexity grow, running data-intensive and parallel workflows involves both different and heterogeneous architectures and the resources of state-of-the-art systems [6,7,8]. Such heterogeneous architectures introduce challenges in workload scheduling, as there are many unknowns and uncertainties  [2,9]. However, despite the optimization capabilities and advances, most AutoML frameworks do not incorporate schedulers such as Slurm, nor do they provide support for energy monitoring, as their main focus is on improving ML model performance (accuracy) and ignore constraints such as EE, utilization, and resource allocation [3,4,10]. To address these challenges, adjusted and tailored frameworks that support sustainable and cost-efficient HPC environments are needed. Therefore, this article presents a high-performance-computing (HPC) differential-evolution-based automatic hyperparameter optimization workflow (AutoDEHypO) for energy efficiency and operational data analytics using multiple graphics processing units (GPUs), and this is deployed on the petascale EuroHPC supercomputer Vega [11]. Our proposed method is capable of determining how energy-efficient the differential evolution (DE) algorithm and ML models perform. The proposed AutoDEHypO workflow considers both their achieved accuracy and the utilization of computing resources, such as elapsed time and consumed energy; it collects runtime data through Slurm within the HPC environment, and the job allocation may impact the obtained results. Moreover, resource consumption leads to inefficient workloads, waste of computational resources, the consumption of a project quota, longer queue times, and postponement of scientific research [12]. Since Slurm dynamically schedules jobs to available, unutilized, or idle nodes within the cluster queue, proper adjustments in job submission may lead to faster allocation and faster attainment of the required job results [13]. We aim to develop the ability to operate within the constraints that Slurm users have and determine whether statistical deviations will be significant. Thus, we prepared our environment for the deployment of the AutoDEHypO workflow, where we ensured consistent resource allocation across the submitted job scripts with a Slurm script (SBATCH). Furthermore, this setup includes supervised machine learning (ML) and multi-label classification and allows for applicable optimization for aspects of hyperparameters that could be affected [14]. We chose the recent parallel implementation of the DE algorithm [15], which parallelizes the population operations in an algorithm that was initially introduced by Storn and Price in 1995 [16]. While DE provides a favorable trade-off and balance between accuracy and energy efficiency (EE) in the context of HPC [17], it additionally offers several advantages over other optimization algorithms, such as efficient and faster convergence, suitability and adaptability in a variety of different environments and optimization problems, global optimization, parallelization, scalability, energy optimization, and the possibility for combination and complementarity with different algorithms and techniques due to available implementations and integrations [16,18,19,20,21]. Furthermore, DE is suitable for the given optimization problem, as it can be used for optimizing nonlinearities in data, such as ML-related input/output data within the workflow, and it can be adapted to other use cases and problems [19]. Data are gathered from system scheduling and historical data; resource allocation and job execution times make it difficult to provide prediction and optimization [16,19,20,21]. In the evaluation phase of our experiment, we used basic ML metrics [14]. During the deployment of individual jobs on multiple GPUs in our experiment, we examined the efficiency of a chosen ML model and DE algorithm according to the accuracy [14]. DE functions and strategies show distinct behavior, and this has not yet been extensively measured or optimized within the context of EE and system resource management in HPC environments [20,22,23]. Data on the consumption of resources of individual jobs were obtained from the Slurm database of completed jobs, with energy consumption data being reported for the whole node by Intelligent Platform Management Interface (IPMI) sensors [13,24]. Based on the results of the computations, aggregated statistics were calculated, along with corresponding post hoc procedures and visualizations, to evaluate the efficiency of the combined ML model and the applied DE strategy.

1.1. Problem Statement and Objective

This work addresses the challenge of optimizing machine learning (ML) models in high-performance computing environments by balancing ML accuracy, energy consumption, and resource utilization. To address these challenges, we propose AutoDEHypO, a differential-evolution-based workflow that is specifically designed for energy-monitored hyperparameter optimization. The specific problem stated for this study consists of considering some key limitations, such as inefficient utilization of computational resources during job allocation, extended queue times, and lack of integration with HPC environments and schedulers such as Slurm. The limitations of existing workflows and frameworks are that they mostly focus on a single objective, such as ML performance or integration within HPC environments, and do not necessarily contribute to the sustainability and cost-effectiveness of HPC environments [25,26]. Therefore, we are interested in analysis of the consumption of resources and when the consumed energy and other resources can lead to different ML performance and configurations.

1.2. Main Contributions

The main contributions of this article are as follows:
  • We propose AutoDEHypO, a high-performance-computing differential-evolution-based automatic hyperparameter workflow designed to optimize the performance of ML models for energy efficiency and operational data analytics in HPC environments.
  • We deploy the AutoDEHypO workflow on the EuroHPC Vega system, utilizing multiple GPUs and Slurm scheduling and submission to execute a specialized fitness function for ML.
  • We applied and evaluated this workflow on supervised ML and multi-label classification using the CIFAR10 and CIFAR100 datasets [27].
  • We collected runtime data through Slurm within the HPC production environment.
  • We evaluated the efficiency of a chosen ML model and DE algorithm strategies according to the ML accuracy and energy efficiency, dependent on ML model architecture, datasets, and resource consumption within the HPC environment.
  • We performed aggregated statistical analyses, along with the corresponding post hoc procedures, and validated the collected data using visualizations by evaluating efficiency of combined ML models and applied DE strategies.
  • We identified significant differences in key metrics and laid the ground for future work on sustainability and cost-effectiveness using AutoDEHypO.
The following sections of this article are organized as follows. Section 2 provides an overview of the related work and existing methods. Section 3 provides a set of proposed methods. Section 4 provides the experimental results. Section 5 presents our conclusion and future work.

3. Proposed Methodology: AutoDEHypO

This section presents the proposed methodology and deployment based on AutoDEHypO, a workflow specifically designed for energy efficiency and operational data analytics in HPC environments.

3.1. Experimental Environment

Computing nodes on an HPC Vega partition with graphics accelerators (GPUs) were used [11]. The partition has 60 nodes, each node has 4× NVIDIA Ampere A100, 2× AMD Rome 7H12, 512 GB RAM, 2× HDR dual-port mezzanines, and 1× 1.92TB M.2 SSD [11]. Red Hat Enterprise Linux 8.10 OS, Slurm 24.05.5 Workload Manager, SingularityPRO version 4.1.6, NVIDIA driver 565.57.01, and CUDA 12.7 were installed on the computing nodes [11]. Additionally, a containerized environment based on Pytorch version 2.1.2 and the required libraries [91]. Training and evaluation were conducted on the publicly available CIFAR10 and CIFAR100 datasets [27]. Storage and dataset access were managed through large-capacity storage based on Ceph [11]. Due to the constraints within HPC environments, the execution time of a single run (i.e., 300 times calling the function f MODA AutoDEHypO ) is such that all times together (the total time of all observations) do not exceed the allocated allocation. So far, we have not limited energy consumption but only monitored it. Additional details of the experimental environment can be found in Table 1.
Table 1. The experimental environment outlines ML model training parameters, and essential configuration to ensure hyperparameter optimization setup and reproducibility of this setup.

3.2. AutoDEHypO

We prepared our environment for the deployment of the AutoDEHypO workflow using supervised machine learning through pre-trained ML models (ResNet18, VGG11, ConvNeXtSmall, and DenseNet121) that are already available within the PyTorch framework; this included building a custom Singularity container for the PyTorch [91] framework with the required libraries. The composed code in the Python programming language takes care of loading a dataset and setting up a Distributed Data Parallel (DDP) that facilitates model parallelization and distribution, while the NVIDIA Collective Communications Library (NCCL) is used in the training phase for faster and more efficient inter-node back-end communication among multiple GPUs, as this ensures and enables efficient scaling among multiple GPUs. A set of methods was used to prepare, develop, and execute AutoDEHypO, including classification, hyperparameter optimization, metric evaluation, resource monitoring, and aggregated statistical analysis of the experimental results.

3.3. Differential-Evolution-Based Hyperparameter Optimization

A basic framework of the DE Algorithm 1 was used from a recent implementation that supports parallelization [15] for the optimization of hyperparameters, as it establishes a basis while minimizing complexity and allows future modifications and improvements as the experiment unfolds [21].

3.4. Job Scheduling, Training, Evaluation, and Visualization

The jobs are submitted through Slurm workload manager, as seen in Figure 2 [11,13]. The input SBATCH contains the required resources within the SBATCH script, the PyTorch container is invoked without modifying the underlying PyTorch code, and  the code written in Python is executed. An example of the SBATCH script that can be wrapped in a conditional loop ( 1 N ) in the command line interface (CLI) or script can be seen in Figure 3.
Figure 2. Basic workflow of job submission through the Slurm workload manager.
Figure 3. SBATCH script for job submission.
Furthermore, the workflow is used for training, evaluation, metrics, the storage of data in the appropriate data frame, and graph plotting based on newly acquired data, as presented in Figure 4. Finally, the dataset is loaded, and the basic steps of DE presented in Figure 1 are performed to evaluate and return the most suitable parameters for the training and evaluation phase. When the job is completed, we evaluate the performance metrics, which are saved in a data frame along with visualizations. Thus, we wanted to check the adequacy of the optimization of the hyperparameter space; i.e., we examined the number of epochs and iterations, weights, learning rate (LR), batch size, and optimizers. Given that it has been implemented and deployed, the AutoDEHypO workflow leverages the potential of utilizing multiple GPUs to run DE fitness functions for ML, as presented in Figure 4. In the evaluation phase, basic ML metrics were used (Section 2.1). To obtain, measure, and compare the utilization of computer resources of individual jobs, such as the elapsed time and energy consumption, data from Slurm were used [13].
Figure 4. Schematic overview of the analytics of the potential efficiency for the AutoDEHypO workflow.

3.5. Checkpoint and Restart, Collected Logs, and Fault Tolerance

Logging of the standard output and standard error output is enabled and generated in the event of an error. Email notifications are also set up to provide information on if a job is queued, started, completed, or failed [13]. This is important for certain cases where anomalies occur during initialization, such as when jobs need to be handled separately. The output files can be taken into account within the AutoDEHypO runtime, which can detect failures in settings, such as issues with NCCL or undetected GPUs within the initialization phase. One can look at these detections, from setting failures to the fact that CR uses feedback mechanisms to check which runs may need to be restarted (Figure 3 line 27). Furthermore, at the beginning of this experiment, we did not know how many resources were needed, and with the CR mechanism, we could accordingly adjust the input Slurm parameters and the need for the computational resource requirements, such as the job state, consumed energy, memory allocation, number of nodes, number of cores, number of cores per node, number of cores per CPU, number of tasks, number of GPUs, number of GPUs per node, and others. The Slurm command requeue allows the resubmission of a failed job when we can avoid and exclude any problematic nodes or, in the opposite case, when we do not want the submission, and a manual check is required after a certain number of—for example, 2—unsuccessful restarts; then, we can use the opposite command: no-requeue. This allows us to detect such errors in certain cases and perform an automatic restart of the job. Manual intervention is still required for cross-checking, and, if necessary, the job is resubmitted to the cluster queue, as shown in Figure 4. Due to the limitations of the experiment, the elapsed time and utilization of computer resources also needed to be taken into account.

4. Experimental Results

The experiments are divided into two phases of benchmarking. Within the first phase, we examine the efficiency of ML models and their parameters. The latter is then used in the second phase as a continuation of the experiments, to which we apply an analysis of the DE mutation strategies presented in Equations (8)–(12). The experiment had to be limited due to resource constraints such as time and the quota of computation resources granted within the development project on the largest Slovenian supercomputer, EuroHPC Vega. The ML models ResNet18, VGG11, ConvNeXtSmall, and DenseNet121 were used with public datasets. For the datasets, we chose CIFAR10 and CIFAR100, as they contain smaller images, thus providing a smaller size than other datasets, allowing them to be included, preprocessed, and trained [27]. The energy consumption is presented in Mega Joules (MJ) [13]. In order to obtain an assessment of the performance of the ML models and algorithms, aggregated statistics were calculated for the obtained results.
Taking into account that resource allocation may impact the obtained results, we ensured that all submitted job scripts had the same resource allocation [2]. Although it would have been possible to restrict our runs to a specific node using an additional Slurm option, this is not the intended use of HPC and was, hence, kept as a constraint. Therefore, jobs and runs were freely distributed across different nodes. During the initiation and execution of the experiment, we faced a few challenges on the software level that we could not avoid encountering, such as the wall time, which was set within the startup script, being too short, as well as a few others. The experiment was also run on the NVIDIA drivers and the Compute Unified Device Architecture (CUDA) Toolkit, which were updated at the beginning of the experiment, as we encountered a set of GPU nodes with incorrect configurations, which were resolved immediately, as well as hardware failures. This necessitated the replacement of DIMM modules and network cards, as well as the cleaning and resetting of network cards, GPUs, and other devices. Hardware problems in which GPUs are not available or are undetected lead to a failure of a large set of submitted jobs in less than 30 s on critical nodes due to failed internal communication through the NCCL back-end. In the event of such an error, a standard error output was generated, and an email notification was successfully received. Furthermore, the deployed CR was included as a common checkpoint for the ML model, therefore demonstrating how we saved important time with AutoDEHypO, with which we enabled resubmission. We received data on failed jobs, such as the time stamp, job state, job name, and job ID, that needed to be resubmitted. An example of the standard error output in the event of a runtime error is given in the following.
  • RuntimeError:
  • ProcessGroupNCCL is only supported with GPUs, no GPUs found!
The results obtained from the ResNet18, VGG11, ConvNeXtSmall, and DenseNet121 ML models on the CIFAR10 dataset are listed in Table 2, Table 3, Table 4 and Table 5, respectively. These tables present the batch size used and the results obtained, including the maximum achieved accuracy on the test batch, the best learning rate found, the best accuracy achieved, the CPU time consumed, the elapsed time, and the energy consumed in a Slurm job. Figure 5 provides an example of a subset for the ML performance of the ResNet18, VGG11, ConvNeXtSmall, and DenseNet121 ML models on the CIFAR10 dataset, and Figure 6 does so for the CIFAR100 dataset.
Table 2. Results obtained in 15 epochs using the ResNet18 ML model on CIFAR10.
Table 3. Results obtained in 15 epochs using the VGG11 ML model on CIFAR10.
Table 4. Results obtained in 15 epochs using the ConvNeXtSmall ML model on CIFAR10.
Table 5. Results obtained in 15 epochs using the DenseNet121 ML model on CIFAR10.
Figure 5. Plots of the initially obtained results of the ML metrics on the CIFAR10 dataset.
Figure 6. Plots of the initially obtained results of the ML metrics on the CIFAR100 dataset.
The results obtained from the ML models on the CIFAR100 dataset in the first initial phase are presented in Table 6 for ResNet18, for VGG11 in Table 7, for ConvNeXtSmall in Table 8, and for DenseNet121 in Table 9, respectively.
Table 6. Results obtained in 15 epochs using the ResNet18 ML model on CIFAR100.
Table 7. Results obtained in 15 epochs using the VGG11 ML model on CIFAR100.
Table 8. Results obtained in 15 epochs using the ConvNeXtSmall ML model on CIFAR100.
Table 9. Results obtained in 15 epochs using the DenseNet121 ML model on CIFAR100.

4.1. Obtained Results

Based on the reported data provided and using our methodology, we observed results in Table 2, Table 3, Table 4 and Table 5 for CIFAR10 and in Table 6, Table 7, Table 8 and Table 9. These tables show a minimum achieved accuracy on test trial and their corresponding impact, highlighting their significance, where the lowest observed accuracy is 62.56%, which corresponds to the calculated ML accuracy of 0.1724. The highest achieved accuracy is 99.17%, with the calculated ML accuracy of 0.19916. Furthermore, for consumed energy metric, we obtained a minimum reported value of 4.43 M (ResNet18) and a maximum value of 59.36 M (DenseNet121). We also observed the elapsed time metric, ranging from the fastest completed job of 03:38:37, to the longest job with wall time 2 d 00:00:17 that exceeded the maximum allowed job execution time within the submitted partition. The results obtained in the first phase of the experiment possibly indicated that using a batch size of 256 in a set of ML models produced the most suitable and efficient results.
We observed that the DenseNet121 model possibly consumed more energy on both datasets than the less complex ML models, such as ResNet18, as presented in Figure 5 and Figure 6. Individual ML models in combination with certain DE strategies possibly performed better and more consistently, while some possibly consumed more energy, e.g., with the exponential DE strategies applied to DenseNet121 or ResNet18, and vice versa, with binomial models possibly consuming more resources in ConvNeXSmall and VGG11. The second phase of the experiment proceeded with a performance comparison of the binomial and exponential DE strategies with each ML model. In this phase, 10 independent runs were executed, as this could more generally determine whether there were significant differences DE-strategy-wise or ML-model-wise, and confirm if there was an impact on key metrics. A mutation was applied randomly with a factor (F) presented in Table 1. The population size (P) and the maximum possible number of generations (G) were appropriately determined due to the project allocation and resource constraints, otherwise we would exceed the current allowable resource consumption within the project allocation. The convergences of metric of accuracy is plotted unified through runtimes in Figure 7 for those runs that obtained median accuracy among each of the 10 independent runs of a DE strategy for a ML model. As observed, the effects of the DE strategies DE/rand/1/bin, DE/best/1/bin, DE/current-to-best/1/bin, DE/rand/2/bin, DE/best/2/bin, as well as the exponential strategies, such as DE/rand/1/exp, DE/best/1/exp, DE/current-to-best/1/exp, DE/rand/2/exp, and DE/best/2/exp [55], possibly varies across different ML model architectures, such as ResNet18, VGG11, ConvNeXtSmall, and DenseNet121, and the CIFAR10 and CIFAR100 datasets [27]. As the results of the tables indicate, H 0 is rejected because at least one DE strategy shows a significant difference, and specific DE strategies may even perform better. Therefore, to analyse which and for how much overall when aggregated, we discuss in the next subsection.
Figure 7. Accuracy convergences through runtime, for the median runs according to accuracy metric.
Figure 8 shows the statistical measure of the mean for each metric in the initial phase, and it is divided into three subplots (a) elapsed time, (b) consumed energy, and (c) accuracy for the DE strategies; these values were measured in 10 runs on the CIFAR10 dataset and are grouped by ML model. Figure 9 shows the same means of the selected DE strategies measured on the CIFAR100 dataset. Moreover, if we have a well-tuned ML model with appropriate weights, and the hyperparameters are poorly tuned, we possibly achieve worse ML performance over key metrics. Furthermore, the results obtained from the ResNet18, VGG11, ConvNeXtSmall, and DenseNet121 ML models with a batch size of 256 on both datasets show that, possibly, more efficient DE strategies consumed fewer resources, as the longer execution time possibly resulted in an increase the consumption of resources, and vice versa. A shorter execution time possibly resulted in lower resource consumption. The results on the CIFAR10 dataset are presented in Table A1, Table A2, Table A3 and Table A4, while the results on the CIFAR100 dataset are presented in Table A5, Table A6, Table A7 and Table A8.
Figure 8. The statistical measures of the mean for the elapsed time (a), consumed energy (b), and accuracy (c) from the results obtained from 10 runs of 15 epochs of ML models that were hyperoptimized using the different DE strategies presented in Table A1, Table A2, Table A3 and Table A4. The x-axis represents the evaluated DE strategies, while the y-axis shows the ML models used.
Figure 9. The statistical measures of the mean for the elapsed time (a), consumed energy (b), and accuracy (c) from the results obtained from 10 runs of 15 epochs of ML models that were hyperoptimized using the different DE strategies presented in Table A5, Table A6, Table A7 and Table A8. The x-axis represents the evaluated DE strategies, while the y-axis shows the ML models used.

4.2. Discussion of the Aggregated Statistics

Aggregate statistical analysis using non-parametric Friedman tests and corresponding procedures for the computational results of the DE strategies was performed on the ResNet18, VGG11, ConvNeXtSmall, and DenseNet121 ML models and on the CIFAR10 and CIFAR100 datasets [27,115]. Control tests were performed on the metrics of elapsed time, consumed energy, and accuracy using custom extraction scripts and publicly available code for statistics [116]. The p-value threshold was set to 0.05, and significant differences were detected and marked (†). This analysis was conducted to determine whether the DE strategies, ML models, and their architectures or a combination thereof had a statistically significant impact and if there were significant differences in key metrics for the computational efficiency. The results of the statistical analysis are presented in Table 10, Table 11, Table 12, Table 13, Table 14, Table 15, Table 16, Table 17 and Table 18.
Table 10. Statistical analysis of elapsed time in the DE strategies using a non-parametric Friedman test with the maximum rank distribution and corresponding post hoc procedures (rejected, as marked in bold and with sign, at a statistical value below the threshold of 0.005555555555555556 for Bonferroni–Dunn, 0.05 for Holm and Hommel, 0.025 for Hochberg and Rom, 0.050000000000000044 for Holland and Finner, and 2.9216395384872545 × 10−17 for Li, respectively).
Table 11. Statistical analysis of energy consumed in the DE strategies using the non-parametric Friedman test with the minimum rank distribution and corresponding post hoc procedures (rejected, as marked in bold and with sign, at a statistical value below the threshold at 0.005555555555555556 for Bonferroni–Dunn, 0.016666666666666666 for Holm and Hommel, 0.0125 for Hochberg, 0.016952427508441503 for Holland, 0.013109375000000001 for Rom, 0.044570249746389234 for Finner, and 0.011483473500591115 for Li, respectively).
Table 12. Statistical analysis of the accuracy in the DE strategies using the non-parametric Friedman test with the minimum rank distribution and corresponding post hoc procedures (rejected, as marked in bold and with sign, at a statistical value below the threshold at 0.005555555555555556 for Bonferroni–Dunn, Holm, and Hommel, 0.005683044988048058 for Holland and Finner, and 7.7532015972022 ×10−4 for Li, respectively).
Table 13. Statistical analysis of the elapsed time in the DE strategies using the non-parametric Friedman test with the maximum rank distribution and corresponding post hoc procedures (rejected, as marked in bold and with sign, at a statistical value below the threshold of 0.005555555555555556 for Bonferroni–Dunn, Holm, and Hommel, 0.005683044988048058 for Holland and Finner, and 0.015161939437253264 for Li, respectively).
Table 14. Statistical analysis of the energy consumed in the DE strategies using the non-parametric Friedman test with the minimum rank distribution and corresponding post hoc procedures (rejected, as marked in bold and with sign, at a statistical value below the threshold of 0.005555555555555556 for Bonferroni–Dunn, Holm, and Hommel, 0.005683044988048058 for Holland and Finner, and 0.014435245666076669 for Li, respectively).
Table 15. Statistical analysis of the accuracy in the DE strategies using the non-parametric Friedman test with the minimum rank distribution and corresponding post hoc procedures (rejected, as marked in bold and with sign, at a statistical value below the threshold of 0.005555555555555556 for Bonferroni–Dunn and Hochberg, 0.00625 for Holm and Hommel, 0.006391150954545011 for Holland, 0.005843911024153359 for Rom, 0.011333792975759982 for Finner, and 0.03180542233195395 for Li, respectively).
Table 16. Statistical analysis of the elapsed time in the DE strategies using the non-parametric Friedman test with the maximum rank distribution and corresponding post hoc procedures (rejected, as marked in bold and with sign, at a statistical value below the threshold of 0.005555555555555556 for Bonferroni–Dunn, 0.0125 for Holm and Hommel, 0.01 for Hochberg, 0.012741455098566168 for Holland, 0.010515350115740741 for Rom, 0.039109465610866256 for Finner, and 0.010841964200380961 for Li, respectively).
Table 17. Statistical analysis of the energy consumed in the DE strategies using the non-parametric Friedman test with the maximum rank distribution and corresponding post hoc procedures (rejected, as marked in bold and with sign, at a statistical value below the threshold of 0.005555555555555556 for Bonferroni–Dunn, 0.0125 for Hochberg, 0.016666666666666666 for Holm and Hommel, 0.016952427508441503 for Holland, 0.013109375000000001 for Rom, 0.044570249746389234 for Finner, and 0.011898018581242242 for Li, respectively).
Table 18. Statistical analysis of the accuracy in the DE strategies using the non-parametric Friedman test with the minimum rank distribution and corresponding post hoc procedures (rejected, as marked in bold and with sign, at a statistical value below the threshold of 0.005555555555555556 for Bonferroni–Dunn, Holm and Hommel, 0.005683044988048058 for Holland, 0.005683044988048058 for Finner, and 0.0010964059917403937 for Li, respectively).
Table 10 presents a statistical analysis of the elapsed time in the DE strategies using a non-parametric Friedman test, where the maximum rank highlights the advantage across the DE strategies, along with the corresponding post hoc procedures. The elapsed time in the DE strategies is significantly better than that of rand2exp on CIFAR10 when using AutoDEHypO according to the post hoc procedures of Holm, Hochberg, Hommel, Holland, and Finner. The Rom procedure shows a significant difference across DE strategies in comparison with rand2exp, with the exception of rand1exp. The Li procedure shows a significant difference across DE strategies in comparison with rand2exp, with the exception of rand2bin. Therefore, AutoDEHypO suggests that on CIFAR10, with the metric of elapsed time, the DE strategies best1bin, best1exp, currenttobest1bin, currenttobest1exp, best2exp, best2bin, rand1bin, rand1exp, and rand2bin are more suitable than rand2exp.
Table 11 presents a statistical analysis of the consumed energy in DE strategies using the non-parametric Friedman test, where the minimum rank highlights the advantage across DE strategies and corresponding post hoc procedures. The consumed energy in the DE strategies rand2bin, rand2exp, best2bin, rand1bin, best2exp, rand1exp, and currenttobest1exp is significantly better than in best1bin on CIFAR10 with AutoDEHypO according to the post hoc procedures of Holm, Hochberg, Hommel, Holland, and Rom. The Finner procedure shows that the DE strategies rand2bin, rand2exp, and best2bin are significantly better than rand2bin. Furthermore, rand1exp is significantly better than rand2bin according to the Holm, Hochberg, Hommel, and  Holland procedures. The Li procedure shows significant difference across DE strategies in comparison with best1bin, with the exception of best1exp. Therefore, AutoDEHypO suggests that on CIFAR10, for the metric of consumed energy, the DE strategies rand2bin, rand2exp, best2bin, rand1bin, best2exp, rand1exp, and currenttobest1exp are more suitable than currenttobest1bin, best1exp, and best1bin.
Table 12 presents a statistical analysis of the accuracy in the DE strategies using the non-parametric Friedman test, where the highest rank highlights the advantage across DE strategies and corresponding post hoc procedures. The accuracy in the DE strategy rand1exp is significantly better than that in best1exp on CIFAR10 with AutoDEHypO according to the post hoc procedures of Holm, Holland, and Finner. Therefore, AutoDEHypO suggests that on CIFAR10, for the metric of accuracy, the DE strategy rand1exp is more suitable than best1exp, rand2bin, rand1bin, currenttobest1bin, rand2exp, best2bin, currenttobest1exp, best1bin, and best2exp.
Table 13 presents a statistical analysis of the elapsed time in the DE strategies using the non-parametric Friedman test, where the maximum rank highlights the advantage across DE strategies and corresponding post hoc procedures. The elapsed time in the DE strategy best1exp is significantly better than that in rand2bin on CIFAR100 with AutoDEHypO according to the post hoc procedures of Holm, Holland, and Finner. The Li procedure shows significant differences across DE strategies in comparison with rand2bin, with the exception of rand2exp. Therefore, AutoDEHypO suggests that on CIFAR100, for the metric of elapsed time, the DE strategy best1exp is more suitable than rand2bin, best2bin, best1bin, currenttobest1bin, currenttobest1exp, rand1bin, rand1exp, best2exp, and rand2exp.
Table 14 presents a statistical analysis of the energy consumed in the DE strategies using the non-parametric Friedman test, where the minimum rank highlights the advantage across DE strategies and corresponding post hoc procedures. The energy consumed in the DE strategy rand2bin is significantly better than that in best1bin on CIFAR100 with AutoDEHypO according to the post hoc procedures of Holm, Hommel, Holland, and Finner. The Li procedure shows significant differences across DE strategies in comparison with best1bin, with the exception of best1exp. Therefore, AutoDEHypO suggests that on CIFAR100, for the metric of consumed energy, the DE strategy rand2bin is more suitable than best1bin, rand1bin, rand2exp, best2exp, currenttobest1exp, rand1exp, currenttobest1bin, best2bin, and best1exp.
Table 15 presents a statistical analysis of the accuracy in the DE strategies using the non-parametric Friedman test, where the highest rank highlights the advantage across DE strategies and corresponding post hoc procedures. The accuracy in the DE strategy rand1bin is significantly better than that in rand1exp on CIFAR100 with AutoDEHypO according to the post hoc procedures of Holm, Hommel, Holland, Rom and Finner. The DE strategy best2bin is significantly better than rand1exp according to the procedures of Holm, Hommel, and Holland. The Li procedure shows significant differences across DE strategies in comparison with rand1exp, with the exception of best2bin. Therefore, AutoDEHypO suggests that on CIFAR100, for the metric of accuracy, the DE strategies rand1bin and rand2exp are more suitable than rand1exp, currenttobest1exp, currenttobest1bin, rand2bin, best1bin, best2exp, best1exp, and best2bin.
Table 16 presents a statistical analysis of the elapsed time in the DE strategies using the non-parametric Friedman test, where the maximum rank highlights the advantage across DE strategies and corresponding post hoc procedures. The elapsed time in the DE strategies best1exp and best1bin is significantly better than that in rand2bin on CIFAR10 and CIFAR100 with AutoDEHypO according to the post hoc procedures of Holm, Hochberg, Holland, Rom, and Finner. Additionally, the DE strategies currenttobest1bin, best2bin, and currenttobest1exp are significantly better than rand2bin according to the procedures of Holm, Hochberg, Hommel, Holland, and Rom. best2exp is significantly better than rand2bin according to the procedures of Holm, Hochberg, Hommel, and Holland. The Li procedure shows significant differences across DE strategies in comparison with rand2bin, with the exception of rand2exp. Therefore, AutoDEHypO suggests that on CIFAR10 and CIFAR100, for the metric of elapsed time, the DE strategies best1exp, best1bin, currenttobest1bin, best2bin, currenttobest1exp, and best2exp are more suitable than rand1bin, rand1exp, and rand2exp.
Table 17 presents a statistical analysis of the energy consumed in the DE strategies using the non-parametric Friedman test, where the maximum rank highlights the advantage across DE strategies and corresponding post hoc procedures. The  energy consumed in the DE strategies best1bin, best1exp, currenttobest1bin, currenttobest1exp, best2bin, rand1exp, and best2exp is significantly better than that in rand2bin on CIFAR10 and CIFAR100 with AutoDEHypO according to the post hoc procedures of Holm, Hochberg, Hommel, Holland, and Rom. According to Finner’s post hoc procedure, the DE strategies best1bin, best1exp, and currenttobest1bin are significantly better than rand2bin. The Li procedure shows significant differences across DE strategies in comparison with rand2bin, with the exception of rand2exp. Therefore, AutoDEHypO suggests that on CIFAR10 and CIFAR100, for the metric of consumed energy, the DE strategies best1bin, best1exp, currenttobest1bin, currenttobest1exp, best2bin, rand1exp, and best2exp are more suitable than rand2bin, rand1bin, and rand2exp.
Table 18 presents a statistical analysis of the accuracy in the DE strategies using the non-parametric Friedman test, where the highest rank highlights the advantage across DE strategies and corresponding post hoc procedures. The accuracy in the DE strategy rand1bin is significantly better than that in best1exp on CIFAR10 and CIFAR100 with AutoDEHypO according to the post hoc procedures of Holm, Hommel, Holland, and Finner. The Li procedure shows significant differences across DE strategies in comparison with best1exp, with the exception of best2exp. Therefore, AutoDEHypO suggests that on CIFAR10 and CIFAR100, for the metric of accuracy, the DE strategy rand1bin is more suitable than best1exp, currenttobest1bin, rand2exp, rand2bin, currenttobest1exp, rand1exp, best1bin, best2bin, and best2exp.
As shown in Table 10, Table 11 and Table 12, according to Holm, Hochberg, Hommel, Holland, and Finner, on CIFAR10, several DE strategies perform significantly better than rand2exp, with additional confirmation from the Rom and Li procedures, except for rand2bin. The DE strategy best1exp runs significantly better than rand2bin, with confirmation from Li (with the exception of rand2exp). Furthermore, the energy consumption in the DE strategies rand2bin, rand2exp, best2bin, rand1bin, best2exp, rand1exp, and currenttobest1exp is significantly better than that in best1bin, with confirmation from Li, except for rand2exp. Table 16, Table 17 and Table 18 present results on both CIFAR10 and CIFAR100, where best1bin, best1exp, currenttobest1bin, currenttobest1exp, best2bin, rand1exp, and best2exp perform significantly better than rand2bin in terms of metrics according to the Li procedure, with exception of rand2exp. In terms of accuracy, best1exp outperforms rand1exp on CIFAR10. As shown in Table 13, Table 14 and Table 15, on CIFAR100, rand1bin achieves significantly better accuracy than that of rand1exp, with confirmation from Li, with the exception of best2bin. Across both datasets, rand1bin also demonstrates significantly better accuracy than that of best1exp, except for best2exp, according to the Li procedure.
Table 19 presents an overview of the outcomes of the statistical analysis and the aggregated number (counts) of confirmed out of 450, 184 (40.8%) significant differences in the DE strategies counting each post hoc procedure across the tree key metrics. These outcomes show and confirm significant differences in elapsed time, energy efficiency, and accuracy across the DE strategies. In addition to the elapsed time and accuracy, our operational data analytics also confirm the impact on energy efficiency when selecting DE strategies.
Table 19. Overview of the outcomes of the statistical analysis for the metrics of elapsed time, energy efficiency, and accuracy in the DE strategies using the non-parametric Friedman test on both the CIFAR10 and CIFAR00 datasets.
As ML models depend on the input data characteristics and on the computational complexity of ML architecture designs, it is expected that some ML models are more suitable. To check this, we tested the suitability of some ML models by detecting differences in elapsed time, consumed energy, and accuracy. The aggregated test outcomes are seen in Table 20, where, for all three key metrics (elapsed time, consumed energy, and accuracy), it is evident from the Friedman rankings that the ResNet18 and ConvNextSmall ML models are ranked higher than the other two ML models, DenseNet121 and VGG11. These ranks are significantly different for the accuracy metric in the case of VGG11. For the metrics of elapsed time and consumed energy, in the case of VGG11, this is also significant. Moreover, significant differences from ML model DenseNet121 are also detected, i.e., these two (DenseNet121 and VGG11) require significantly more energy and time than ResNet18 and ConvNextSmall.
Table 20. Friedman rankings detected differences in elapsed time, consumed energy, and accuracy aggregated across ML models. The individual rankings demonstrate significant differences using the Holm/Hochberg/Hommel, Holland, Rom, Finner, and Li procedures, with a few exceptions for the post hoc procedures (the Rom and Li procedures for time and energy detect these differences only when comparing VGG11 and DenseNet121 against ResNet18 and ConvNextSmall, i.e., for 41 (91.1%) out of 45 rankings in the post hoc procedures, we successfully detected significant differences).
Furthermore, significant differences between four ML models for 10 DE strategies (denoted as k) evaluations per each run result in 20 combinations (denoted as N), with critical value q α of 2.394 from the the two-tailed Bonferroni-Dunn test for k is studentized range statistic at the significance level α threshold of 0.05 on their Friedman average ranks. To compare their ranks we calculated critical difference (CD) [117] of approximately 0.9773 . Additionally, we calculated confidence intervals (CI) based on CD, which detected significant difference in 11 (61.11%) out of 18 pairwise comparisons. The results are presented in Table 21.
Table 21. Ranks and confidence intervals for key metrics elapsed time, consumed energy, and accuracy across ML Models ResNet18, ConvNextSmall, DenseNet121, and VGG11 on CIFAR10 and CIFAR100 datasets.

5. Conclusions and Future Work

In this article, we presented the deployment of a high-performance differential-evolution-based hyperparameter optimization workflow (AutoDEHypO) for energy efficiency and operational data analytics on multiple GPUs, where a DE algorithm with different DE strategies is demonstrated and applied in hyperparameter optimization while considering key factors such as the ML model, dataset, and DE strategy. The challenge of the optimization of ML models in HPC environments was addressed by balancing the ML accuracy, energy consumption, and resource utilization. Practical limitations such as project time constraints and the computational resources allocated to this project on a part of the national share of HPC Vega may have influenced the experiment through aspects such as node availability, high cluster utilization, shared node scheduling, elapsed time and partition (wall time) limitations, network congestion, job failures, undetected GPUs, and limited granularity in energy measurement at the node level. The utilization of computing resources had to be carefully considered, and the evaluation metrics and their scope must be taken into account. AutoDEHypO detected significant differences in the utilization of HPC resources in terms of elapsed time, energy efficiency, and ML accuracy across DE strategies. Furthermore, AutoDEHypO overcomes several limitations of existing workflows, which mostly focus on ML performance, inefficient utilization of computational resources during job allocation, and extended queue times; therefore, there is a lack of integration with HPC environments and schedulers such as Slurm and limited support for sustainability and cost-effectiveness. In addition to tracking the elapsed time and accuracy, our AutoDEHypO uses operational data analytics to determine how energy efficiently DE algorithms and DE strategies perform and confirms the impact on energy efficiency when selecting DE strategies. The analytics of DE strategies influenced the consumption of resources and indicated when the consumption of energy and other resources led to different ML performance and configurations. Furthermore, the statistical analysis of the key metrics of elapsed time, consumed energy, and accuracy demonstrated significant differences in DE strategies and ML models by using the non-parametric Friedman test and corresponding post hoc procedures for CIFAR10 and CIFAR100 datasets.
Specifically, in 10 independent runs, the DE binomial mutation strategies were applied and out of 450 comparisons, 184 (40.8%) detected significant differences between strategies. The effect of used DE strategies DE/rand/1/bin, DE/best/1/bin, DE/current-to-best/1/bin, DE/rand/2/bin, DE/best/2/bin, and exponential such as DE/rand/1/exp, DE/best/1/exp, DE/current-to-best/1/exp, DE/rand/2/exp, and DE/best/2/exp varied across differentent ML model architectures such as ResNet18, VGG11, ConvNeXtSmall and DenseNet121, on datasets CIFAR10 and CIFAR100.
Additionally, as an important outcome, the ML model comparisons show that there are 41 (91.1%) out of 45 rankings in the post hoc procedures, where AutoDEHypO successfully detected significant differences in all three key metrics in the case of the VGG11 ML model; for the metrics of elapsed time and consumed energy, there were further significant differences detected in the DenseNet121 and VGG11 models when compared with the better ML models of ConvNextSmall and ResNet18. Furthermore, the calculated confidence intervals based on critical difference of approximately 0.9773 , detected significant differences in 11 (61.11%) out of 18 pairwise comparisons.
Despite the insights and contributions provided by this article, there are a few limitations to acknowledge. This workflow is designed and optimized for deployment on a single cluster. Generalizing this workflow to another HPC environment may be challenging due to differences in architecture, scheduling policies, and energy monitoring solutions. Furthermore, within the proposed workflow, we used a basic DE algorithm implementation and limited this research to a few ML models, two datasets (CIFAR10 and CIFAR100), a single DE algorithm, 10 DE strategies, and limited runs of each epoch. While this setup was sufficient for initial exploration and experimentation, further research can be conducted. Moreover, this workflow can be adopted for other ML models, such as neural networks with adjustments in their architectures, different DE algorithms and DE strategies, population sizes, and different datasets, such as ImageNet, MNIST, and others; last but not least, more computational resources can be used. Possibilities have emerged for the evaluation and validation of other algorithms and AutoML frameworks to provide a baseline comparison with state-of-the-art AutoML frameworks and beyond. The necessary validation and comparison for a different environment could take place within an additional experiment; this could include a comparison of the final results. System malfunctions may also be resolved through fault-tolerance mechanisms such as Checkpoint and Restart (CR) without affecting the obtained results, allowing one to continue after unexpected interruptions. Additionally, we can apply CR inside the DE loop, where we can continue from the last successfully saved and calculated generation. Furthermore, a real-time feedback mechanism for adaptive and dynamic changes within ML operations has not yet been implemented and may contribute to the workflow; this may be researched in the future. The prediction model and decision making based on historical data can contribute to ML performance, optimization, and energy efficiency. These improvements will not only strengthen the practical applicability of this workflow but also contribute to sustainability, and a reduction in the environmental footprint.

Author Contributions

Conceptualization, T.P. and A.Z.; methodology, T.P. and A.Z.; software, T.P. and A.Z.; validation, T.P. and A.Z.; formal analysis, T.P. and A.Z.; investigation, T.P. and A.Z.; resources, T.P. and A.Z.; data curation, T.P. and A.Z.; writing—original draft preparation, T.P. and A.Z.; writing—review and editing, T.P. and A.Z.; visualization, T.P. and A.Z.; supervision, T.P. and A.Z.; project administration, T.P. and A.Z.; funding acquisition, T.P. and A.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded and conducted within the Individual Research Work 3 Unit of the doctoral program for Computer Science and Informatics at the University of Maribor. The fee for study enrollment was financed by IZUM—Institute of Information Science (17-2141-2023/01-ab and 17-2375-2024/01-ab).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The original data are included within this paper. Further inquiries may be addressed to the corresponding author.

Acknowledgments

The authors acknowledge the EuroHPC JU, HPC RIVR and SLING consortium for allocating computing resources on the national share of HPC Vega within the Development Project (S24R08-01) hosted at the Institute of Information Science (IZUM). Authors also acknowledge the project DAPHNE (Integrated Data Analysis Pipelines for Large-Scale Data Management, HPC, and Machine Learning) funded by the European Union’s Horizon 2020 research and innovation program under grant agreement No 957407. We also acknowledge COST (European Cooperation in Science and Technology) support from COST Actions: CA22137 “Randomised Optimisation Algorithms Research Network (ROAR-NET)”. Authors acknowledge the MDPI Institutional Open Access Program (IOAP) for University of Maribor and thank editors and reviewers of this paper.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

Table A1. Results obtained from 10 runs of 15 epochs of a hyperoptimized ResNet18 using a different DE strategy on CIFAR10.
Table A1. Results obtained from 10 runs of 15 epochs of a hyperoptimized ResNet18 using a different DE strategy on CIFAR10.
RunStrategyMax Accuracy (%)Best LRBest AccuracyCPU TimeElapsed TimeConsumed Energy
Run 1rand1bin100%0.00017169553081330660.2012807:17:5601:49:292.39 MJ
Run 2rand1bin99.976%0.00017169553081330660.2017608:02:2002:00:352.63 MJ
Run 3rand1bin100%0.00017169553081330660.2012806:35:4001:38:552.11 MJ
Run 4rand1bin100%0.00017169553081330660.2017609:58:5602:29:443.12 MJ
Run 5rand1bin99.96%0.00017169553081330660.19809:54:1202:28:333.11 MJ
Run 6rand1bin100%0.00017169553081330660.2017606:41:0401:40:162.14 MJ
Run 7rand1bin100%0.00017169553081330660.2012810:13:0402:33:163.27 MJ
Run 8rand1bin99.984%0.00017169553081330660.2017607:53:1601:58:192.52 MJ
Run 9rand1bin100%0.00017169553081330660.2017607:58:5201:59:432.62 MJ
Run 10rand1bin100%0.00017169553081330660.2017607:13:2801:48:222.32 MJ
Run 1best1bin99.968%0.000132447958668769260.19808:38:4802:09:422.82 MJ
Run 2best1bin100%0.000132447958668769260.2012809:53:0402:28:163.20 MJ
Run 3best1bin99.96%0.000240772301800391150.1989606:58:3201:44:382.37 MJ
Run 4best1bin99.992%0.000132447958668769260.2017609:38:2802:24:373.18 MJ
Run 5best1bin100%0.000240772301800391150.19805:56:0001:29:001.95 MJ
Run 6best1bin100%0.00021624245414767140.1989608:18:2002:04:352.86 MJ
Run 7best1bin99.968%0.000240772301800391150.2017607:53:1601:58:192.52 MJ
Run 8best1bin99.992%0.000240772301800391150.1989610:41:5602:40:293.30 MJ
Run 9best1bin99.704%0.000240772301800391150.2017610:49:1602:42:193.47 MJ
Run 10best1bin99.936%0.000240772301800391150.2017607:17:3201:49:232.29 MJ
Run 1currenttobest1bin99.952%0.000240772301800391150.2017611:21:4002:50:253.58 MJ
Run 2currenttobest1bin99.984%0.000240772301800391150.2017609:25:3202:21:233.05 MJ
Run 3currenttobest1bin100%0.00021624245414767140.1989609:46:3602:26:393.14 MJ
Run 4currenttobest1bin100%0.000240772301800391150.2017613:04:0003:16:004.27 MJ
Run 5currenttobest1bin99.984%0.000132447958668769260.2017608:57:3202:14:232.84 MJ
Run 6currenttobest1bin100%0.000240772301800391150.1989609:10:1602:17:343.00 MJ
Run 7currenttobest1bin100%0.000240772301800391150.2012809:13:2402:18:212.95 MJ
Run 8currenttobest1bin99.952%0.000132447958668769260.2017609:28:3602:22:093.06 MJ
Run 9currenttobest1bin99.976%0.00021624245414767140.19811:38:1202:54:333.93 MJ
Run 10currenttobest1bin99.264%0.000240772301800391150.2012811:49:4002:57:253.82 MJ
Run 1rand2bin99.624%0.000240772301800391150.2017612:54:5603:13:444.40 MJ
Run 2rand2bin100%0.000240772301800391150.2017611:32:4802:53:123.73 MJ
Run 3rand2bin99.992%0.00021624245414767140.2017610:10:0402:32:313.30 MJ
Run 4rand2bin99.928%0.000240772301800391150.2012809:53:0002:28:153.21 MJ
Run 5rand2bin99.992%0.00024862432592688210.19812:27:5603:06:594.02 MJ
Run 6rand2bin100%0.00021624245414767140.19816:39:4004:09:555.48 MJ
Run 7rand2bin99.944%0.000240772301800391150.2017612:32:4003:08:104.09 MJ
Run 8rand2bin100%0.000240772301800391150.19812:17:0403:04:163.95 MJ
Run 9rand2bin100%0.000240772301800391150.2017608:38:5602:09:442.82 MJ
Run 10rand2bin100%0.000281358581111197030.2017612:44:4403:11:113.99 MJ
Run 1best2bin99.144%0.00020215828617478350.1989614:49:4403:42:264.75 MJ
Run 2best2bin100%0.00020215828617478350.2012810:21:2002:35:203.30 MJ
Run 3best2bin100%0.000240772301800391150.2017611:26:3202:51:383.62 MJ
Run 4best2bin100%0.00020215828617478350.2017610:09:4802:32:273.29 MJ
Run 5best2bin99.896%0.00014475302657031410.2012810:46:0802:41:323.54 MJ
Run 6best2bin100%0.00020215828617478350.2017608:46:4002:11:402.79 MJ
Run 7best2bin100%0.000240772301800391150.2012807:18:0801:49:322.39 MJ
Run 8best2bin100%0.000281358581111197030.2012811:21:1202:50:183.83 MJ
Run 9best2bin99.936%0.000240772301800391150.1989608:07:2002:01:502.60 MJ
Run 10best2bin100%0.000123931800743550060.2017614:49:3203:42:234.69 MJ
Run 1rand1exp99.976%0.00017169553081330660.2017607:19:3601:49:542.39 MJ
Run 2rand1exp100%0.00017169553081330660.2012807:20:2001:50:052.37 MJ
Run 3rand1exp99.984%0.00017169553081330660.19810:16:0802:34:023.24 MJ
Run 4rand1exp99.992%0.00017169553081330660.2017608:01:5602:00:292.59 MJ
Run 5rand1exp99.992%0.00017169553081330660.2012807:12:3201:48:082.33 MJ
Run 6rand1exp100%0.00017169553081330660.2017608:33:1602:08:192.79 MJ
Run 7rand1exp99.992%0.00017169553081330660.1989607:16:0801:49:022.31 MJ
Run 8rand1exp100%0.00017169553081330660.2017608:34:5602:08:442.76 MJ
Run 9rand1exp100%0.00017169553081330660.2017608:35:5602:08:592.80 MJ
Run 10rand1exp99.984%0.00017169553081330660.19810:37:2802:39:223.45 MJ
Run 1rand2exp99.92%0.000240772301800391150.2017611:20:1202:50:033.81 MJ
Run 2rand2exp99.968%0.000240772301800391150.2017612:48:4403:12:114.04 MJ
Run 3rand2exp99.92%0.000240772301800391150.2017614:57:2003:44:204.86 MJ
Run 4rand2exp99.976%0.000240772301800391150.2017613:53:5203:28:284.52 MJ
Run 5rand2exp100%0.00021624245414767140.1989611:56:0402:59:013.76 MJ
Run 6rand2exp99.808%0.000240772301800391150.19814:34:2403:38:364.54 MJ
Run 7rand2exp99.88%0.000240772301800391150.2017614:13:3603:33:244.48 MJ
Run 8rand2exp100%0.000240772301800391150.2017610:30:4802:37:423.39 MJ
Run 9rand2exp100%0.00021624245414767140.1989609:19:5602:19:592.99 MJ
Run 10rand2exp99.96%0.00021624245414767140.2017619:39:0004:54:456.27 MJ
Run 1best1exp99.976%0.000132447958668769260.2017606:38:4001:39:402.15 MJ
Run 2best1exp99.984%0.000132447958668769260.2012812:34:4403:08:414.02 MJ
Run 3best1exp99.712%0.00021624245414767140.1989611:19:1602:49:493.67 MJ
Run 4best1exp100%0.000132447958668769260.2017611:52:3602:58:093.74 MJ
Run 5best1exp100%0.000240772301800391150.1989611:31:0002:52:453.68 MJ
Run 6best1exp99.96%0.000132447958668769260.2017610:16:4402:34:113.23 MJ
Run 7best1exp99.928%0.000132447958668769260.2012810:52:2002:43:053.45 MJ
Run 8best1exp99.952%0.000132447958668769260.19811:16:3202:49:083.66 MJ
Run 9best1exp100%0.000240772301800391150.2017609:32:0802:23:023.11 MJ
Run 10best1exp99.968%0.00021624245414767140.2012808:45:1602:11:192.76 MJ
Run 1best2exp100%0.000240772301800391150.2017612:03:2803:00:523.84 MJ
Run 2best2exp100%0.00020215828617478350.2017609:13:4002:18:253.20 MJ
Run 3best2exp100%0.00020215828617478350.2017610:47:1602:41:493.43 MJ
Run 4best2exp99.928%0.000210979504563787970.1989613:38:2403:24:364.24 MJ
Run 5best2exp99.584%0.000123931800743550060.2012812:44:4403:11:114.13 MJ
Run 6best2exp100%0.000123931800743550060.2017608:33:2002:08:202.71 MJ
Run 7best2exp99.968%0.000240772301800391150.2012807:16:2801:49:072.33 MJ
Run 8best2exp100%0.00020215828617478350.2017609:07:5602:16:592.90 MJ
Run 9best2exp100%0.00021624245414767140.2012813:31:3203:22:534.39 MJ
Run 10best2exp100%0.000240772301800391150.2017608:55:2802:13:522.87 MJ
Run 1currenttobest1exp100%0.000240772301800391150.1989610:41:2402:40:213.33 MJ
Run 2currenttobest1exp99.984%0.000132447958668769260.2017612:47:0403:11:464.01 MJ
Run 3currenttobest1exp99.872%0.000132447958668769260.1989609:16:4002:19:102.98 MJ
Run 4currenttobest1exp100%0.00021624245414767140.2017606:52:0401:43:012.24 MJ
Run 5currenttobest1exp100%0.00021624245414767140.2017606:27:2801:36:522.06 MJ
Run 6currenttobest1exp99.992%0.000132447958668769260.2017611:03:3202:45:533.73 MJ
Run 7currenttobest1exp99.968%0.000240772301800391150.2017607:45:1201:56:182.55 MJ
Run 8currenttobest1exp100%0.000132447958668769260.2017607:15:2801:48:522.36 MJ
Run 9currenttobest1exp99.976%0.000240772301800391150.19809:52:0002:28:003.08 MJ
Run 10currenttobest1exp99.968%0.000132447958668769260.2017607:17:2801:49:222.28 MJ
Table A2. Results obtained from 10 runs of 15 epochs of a hyperoptimized VGG11 using a different DE strategy on CIFAR10.
Table A2. Results obtained from 10 runs of 15 epochs of a hyperoptimized VGG11 using a different DE strategy on CIFAR10.
RunStrategyMax Accuracy (%)Best LRBest AccuracyCPU TimeElapsed TimeConsumed Energy
Run 1rand1bin91.136%0.00028504921786332150.194641 d 06:51:0807:42:4710.92 MJ
Run 2rand1bin89.84%0.000145334761903245030.19961 d 15:46:0409:56:3113.54 MJ
Run 3rand1bin88.904%0.000162124068907360.200082 d 08:05:2014:01:2019.72 MJ
Run 4rand1bin88.096%0.000174477153877060860.196482 d 14:03:4415:30:5621.79 MJ
Run 5rand1bin78.848%0.000132776061637921520.200081 d 17:55:4810:28:5714.62 MJ
Run 6rand1bin84.84%0.000402172631010210240.195521 d 16:39:3610:09:5414.22 MJ
Run 7rand1bin86.832%0.000137725209534367780.19961 d 13:27:4409:21:5614.32 MJ
Run 8rand1bin91.024%0.000241130783042947580.199841 d 20:35:0011:08:4515.61 MJ
Run 9rand1bin77.488%0.000139592689798095450.198482 d 07:07:0413:46:4619.14 MJ
Run 10rand1bin90.2%0.000183768490814301860.1962 d 02:35:0412:38:4618.53 MJ
Run 1best1bin88.888%0.000170846510541198340.195761 d 05:08:3207:17:0810.27 MJ
Run 2best1bin82.096%0.000149869991507051220.198641 d 07:07:0807:46:4710.74 MJ
Run 3best1bin54.912%0.00032823638086059630.197121 d 16:01:3610:00:2414.08 MJ
Run 4best1bin87.728%0.00025388759996056110.190961 d 06:27:0007:36:4510.75 MJ
Run 5best1bin86.24%0.000132447958668769260.199281 d 14:55:5209:43:5813.72 MJ
Run 6best1bin87.232%0.000187725990784069880.195441 d 10:27:2008:36:5011.88 MJ
Run 7best1bin91.024%0.00019799833108324090.197281 d 11:50:1608:57:3412.46 MJ
Run 8best1bin87.048%0.000204520793570471350.199121 d 22:41:0011:40:1516.45 MJ
Run 9best1bin86.744%0.000116978557920244460.199761 d 16:49:2410:12:2114.36 MJ
Run 10best1bin87.848%0.000118781444167737950.199282 d 01:29:0812:22:1717.48 MJ
Run 1currenttobest1bin94.408%0.00017771013478759730.198481 d 12:49:1209:12:1812.76 MJ
Run 2currenttobest1bin91.664%0.00016767751239408380.19681 d 21:12:3611:18:0915.98 MJ
Run 3currenttobest1bin87.704%0.000249274270394843360.199681 d 17:18:0810:19:3214.53 MJ
Run 4currenttobest1bin93.104%0.000125251673167103550.198641 d 10:23:4408:35:5612.22 MJ
Run 5currenttobest1bin76.968%0.000112719265752493690.20081 d 15:54:4809:58:4213.94 MJ
Run 6currenttobest1bin92.656%0.000240772301800391150.198641 d 15:43:3609:55:5414.74 MJ
Run 7currenttobest1bin83.928%0.000180056271732636770.200082 d 04:49:0013:12:1518.52 MJ
Run 8currenttobest1bin89.832%0.000180056271732636770.197281 d 14:02:0009:30:3013.37 MJ
Run 9currenttobest1bin79.304%0.000141609709395538360.200241 d 18:51:1610:42:4914.97 MJ
Run 10currenttobest1bin92.944%0.00017099711310533570.199521 d 08:37:1208:09:1811.34 MJ
Run 1rand2bin87.672%0.000140437197525316070.198642 d 09:31:2814:22:5220.78 MJ
Run 2rand2bin89.04%0.000112257164111145730.198882 d 10:33:2814:38:2220.83 MJ
Run 3rand2bin94.512%0.00021335012431247070.196242 d 02:37:0812:39:1717.67 MJ
Run 4rand2bin83.176%0.000146539418890181440.19642 d 09:09:0814:17:1719.81 MJ
Run 5rand2bin87.024%0.000119533641653748430.197282 d 05:57:2413:29:2118.39 MJ
Run 6rand2bin84.768%0.000235532893505987780.199682 d 04:24:0413:06:0117.81 MJ
Run 7rand2bin87.648%0.00022375922465134610.199361 d 17:36:3610:24:0914.43 MJ
Run 8rand2bin88.624%0.000284500768355017930.197842 d 14:21:2815:35:2221.70 MJ
Run 9rand2bin87.976%0.00013814089102834190.200082 d 01:54:1212:28:3317.34 MJ
Run 10rand2bin88.936%0.00037455290836552160.196642 d 02:32:4012:38:1017.96 MJ
Run 1best2bin87.808%0.000198024254596845550.197921 d 14:59:4809:44:5713.70 MJ
Run 2best2bin92.296%0.00012513961710271520.199441 d 12:31:3609:07:5412.92 MJ
Run 3best2bin90.4%0.00019042703378366230.19641 d 17:50:4410:27:4114.68 MJ
Run 4best2bin81.952%0.00010994762254373910.198882 d 01:55:3612:28:5417.65 MJ
Run 5best2bin75.464%0.000202918619823005050.196721 d 20:22:4811:05:4215.79 MJ
Run 6best2bin86.48%0.000100609821970013250.20042 d 10:13:3214:33:2320.22 MJ
Run 7best2bin88.752%0.000164919716241819510.199681 d 13:26:4009:21:4012.71 MJ
Run 8best2bin93.512%0.000240772301800391150.198241 d 21:12:1211:18:0316.09 MJ
Run 9best2bin89.448%0.00017147904746484410.200081 d 19:37:0010:54:1515.19 MJ
Run 10best2bin84.432%0.000144875076772761470.199282 d 13:24:0415:21:0121.30 MJ
Run 1rand1exp81.736%0.000205728364619391120.197681 d 12:40:0009:10:0012.93 MJ
Run 2rand1exp90.736%0.00012233353653173930.200241 d 13:13:2009:18:2012.97 MJ
Run 3rand1exp81.24%0.000281358581111197030.198882 d 03:22:2812:50:3718.10 MJ
Run 4rand1exp88.952%0.00017169553081330660.19882 d 01:32:4412:23:1117.78 MJ
Run 5rand1exp84.584%0.000117389803325508950.198241 d 20:13:0811:03:1715.40 MJ
Run 6rand1exp88.16%0.00017169553081330660.196882 d 12:53:4815:13:2721.12 MJ
Run 7rand1exp79.792%0.00034027337350456030.196561 d 12:26:0809:06:3212.72 MJ
Run 8rand1exp90.84%0.00018706144980548440.199681 d 22:59:5211:44:5816.24 MJ
Run 9rand1exp85.648%0.000319838608615680450.198241 d 19:27:2410:51:5115.32 MJ
Run 10rand1exp92.024%0.000180745436216547910.199681 d 22:00:3611:30:0915.91 MJ
Run 1rand2exp77.792%0.00036115292896757650.195762 d 13:49:2015:27:2021.95 MJ
Run 2rand2exp88.48%0.000178378401766325270.19882 d 03:06:4012:46:4017.93 MJ
Run 3rand2exp89.08%0.00026538470587646220.195681 d 19:58:4010:59:4014.96 MJ
Run 4rand2exp93.68%0.000125251673167103550.195682 d 09:39:2014:24:5021.03 MJ
Run 5rand2exp79.16%0.000114071407139993420.197681 d 18:18:1610:34:3414.92 MJ
Run 6rand2exp83.536%0.000173013482800916720.198562 d 02:53:2412:43:2117.94 MJ
Run 7rand2exp88.992%0.00028428807600572670.196641 d 19:29:4010:52:2515.56 MJ
Run 8rand2exp87.232%0.00037501598771371280.199522 d 15:49:4015:57:2522.13 MJ
Run 9rand2exp83.456%0.000158128240204336730.198161 d 21:09:0811:17:1716.10 MJ
Run 10rand2exp91.888%0.000177933882522167020.196241 d 22:20:5211:35:1316.15 MJ
Run 1best1exp88.696%0.000140338368546473340.197361 d 02:15:3606:33:549.31 MJ
Run 2best1exp86.968%0.00021624245414767140.197681 d 15:12:5609:48:1414.52 MJ
Run 3best1exp92.84%0.000236380949017823160.197361 d 07:52:0807:58:0211.21 MJ
Run 4best1exp86.672%0.000104271954752884740.1941 d 10:19:2808:34:5212.09 MJ
Run 5best1exp80.672%0.000194277056381570930.198881 d 15:29:5609:52:2913.53 MJ
Run 6best1exp77.904%0.000151643268009670440.197921 d 10:28:2408:37:0612.13 MJ
Run 7best1exp88.776%0.000125251673167103550.199521 d 11:42:5208:55:4312.57 MJ
Run 8best1exp86.32%0.000106351833725319330.197921 d 05:52:4007:28:1010.53 MJ
Run 9best1exp90.912%0.000164033598096865880.198321 d 13:23:4809:20:5712.89 MJ
Run 10best1exp86.616%0.000172518782329636050.198321 d 13:11:0409:17:4613.27 MJ
Run 1best2exp84.392%0.000175484195064109670.197361 d 13:47:5609:26:5913.40 MJ
Run 2best2exp85.904%0.000214385131736420150.196482 d 02:36:5612:39:1417.28 MJ
Run 3best2exp80.952%0.000106150864909526490.198641 d 15:33:1609:53:1913.77 MJ
Run 4best2exp91.496%0.00025862261362331860.195761 d 18:41:2410:40:2114.70 MJ
Run 5best2exp80.376%0.000178804635069251980.195842 d 00:10:2012:02:3516.64 MJ
Run 6best2exp75.912%0.000107198372386767010.198882 d 10:04:1614:31:0420.64 MJ
Run 7best2exp89.456%0.000185248950708825030.199281 d 23:31:0011:52:4516.85 MJ
Run 8best2exp87.856%0.000156378470833566130.195922 d 03:50:2012:57:3519.63 MJ
Run 9best2exp83.952%0.00028949860369495420.195281 d 12:47:0009:11:4513.00 MJ
Run 10best2exp90.384%0.000212672037043803580.195281 d 23:50:3611:57:3916.27 MJ
Run 1currenttobest1exp86.328%0.000161101683252510750.199841 d 15:00:1209:45:0313.84 MJ
Run 2currenttobest1exp83.032%0.00017415581518198490.19921 d 12:21:1609:05:1912.90 MJ
Run 3currenttobest1exp74.632%0.00019006815953323860.197121 d 15:17:1609:49:1914.05 MJ
Run 4currenttobest1exp94.304%0.00017378490513045970.196321 d 07:09:2407:47:2110.74 MJ
Run 5currenttobest1exp87.336%0.000339222149472093370.193841 d 12:41:5609:10:2913.80 MJ
Run 6currenttobest1exp81.44%0.000160207314299752770.198962 d 13:42:4815:25:4221.54 MJ
Run 7currenttobest1exp89.128%0.000158185662753083560.197682 d 02:56:5212:44:1318.15 MJ
Run 8currenttobest1exp88.272%0.00021787071189150580.198241 d 13:46:4009:26:4012.99 MJ
Run 9currenttobest1exp84.672%0.000170591610982738080.19922 d 06:39:4413:39:5618.98 MJ
Run 10currenttobest1exp77.736%0.000210962963167739990.199362 d 00:41:4012:10:2516.82 MJ
Table A3. Results obtained from 10 runs of 15 epochs of a hyperoptimized ConvNeXtSmall using a different DE strategy on CIFAR10.
Table A3. Results obtained from 10 runs of 15 epochs of a hyperoptimized ConvNeXtSmall using a different DE strategy on CIFAR10.
RunStrategyMax Accuracy (%)Best LRBest AccuracyCPU TimeElapsed TimeConsumed Energy
Run 1rand1bin95.976%0.00066540300794917370.201216:28:4004:07:105.76 MJ
Run 2rand1bin98.152%0.00055037828142615850.2006413:47:4403:26:564.93 MJ
Run 3rand1bin98.12%0.00026727959265821790.1974418:00:0404:30:016.28 MJ
Run 4rand1bin97.648%0.00055037828142615850.1999217:16:4004:19:106.12 MJ
Run 5rand1bin98.104%0.00021624245414767140.1972815:15:1603:48:495.21 MJ
Run 6rand1bin98.112%0.00054736094255330860.1971219:51:4404:57:567.05 MJ
Run 7rand1bin98.832%0.00060820149663338530.1999218:50:0004:42:306.60 MJ
Run 8rand1bin97.464%0.000348714254562946970.1997620:47:3605:11:547.69 MJ
Run 9rand1bin96.952%0.00053553739028156650.1970419:18:5204:49:437.12 MJ
Run 10rand1bin97.312%0.00057124394950399690.1971215:20:4403:50:115.34 MJ
Run 1best1bin98.184%0.00079275041062573270.1973617:57:2804:29:226.28 MJ
Run 2best1bin97.912%0.00079275041062573270.2005612:20:3203:05:084.39 MJ
Run 3best1bin97.752%0.00086789286879177020.197216:14:1604:03:345.78 MJ
Run 4best1bin97.384%0.00086789286879177020.1998418:27:0404:36:466.55 MJ
Run 5best1bin97.656%0.00066540300794917370.215:18:4403:49:415.40 MJ
Run 6best1bin98.496%0.00066540300794917370.2008815:22:5203:50:435.35 MJ
Run 7best1bin95.768%0.000338795628054039460.2007217:17:5604:19:296.06 MJ
Run 8best1bin97.44%0.00086789286879177020.1996816:44:5204:11:135.80 MJ
Run 9best1bin97.592%0.00066540300794917370.2008816:20:4804:05:125.62 MJ
Run 10best1bin98.184%0.00055037828142615850.1983218:05:0004:31:156.62 MJ
Run 1currenttobest1bin98.576%0.00077347680313113630.2007215:54:3203:58:385.87 MJ
Run 2currenttobest1bin97.408%0.00057397123233140840.1993616:11:0804:02:475.49 MJ
Run 3currenttobest1bin98.264%0.00079010804734894020.2008820:10:4005:02:407.23 MJ
Run 4currenttobest1bin97.288%0.000281358581111197030.2001616:00:5204:00:135.54 MJ
Run 5currenttobest1bin97.72%0.000348714254562946970.1977615:13:0403:48:165.21 MJ
Run 6currenttobest1bin98.824%0.00055037828142615850.1975213:17:3603:19:244.75 MJ
Run 7currenttobest1bin98.544%0.00038626083410303430.201213:04:4803:16:124.70 MJ
Run 8currenttobest1bin97.24%0.00039063006568036010.1997615:54:0803:58:325.59 MJ
Run 9currenttobest1bin98.752%0.00089253266255637010.2008814:28:5203:37:135.24 MJ
Run 10currenttobest1bin97.344%0.00079010804734894020.2010414:14:0003:33:304.25 MJ
Run 1rand2bin97.64%0.000281358581111197030.2011216:13:3604:03:245.71 MJ
Run 2rand2bin97.456%0.00079010804734894020.1970419:27:4404:51:566.88 MJ
Run 3rand2bin98.112%0.00079010804734894020.1972816:18:4404:04:415.68 MJ
Run 4rand2bin98.568%0.00075254780224504780.199216:20:1204:05:035.71 MJ
Run 5rand2bin98.84%0.00090.2012813:02:1203:15:334.58 MJ
Run 6rand2bin98.032%0.00021624245414767140.2002417:50:1604:27:346.38 MJ
Run 7rand2bin97.904%0.00066540300794917370.196814:55:0403:43:465.28 MJ
Run 8rand2bin97.688%0.00021624245414767140.197617:33:1204:23:186.12 MJ
Run 9rand2bin98.6%0.00039584886765497010.201219:04:1204:46:036.78 MJ
Run 10rand2bin98.672%0.00055037828142615850.1974418:23:0004:35:456.55 MJ
Run 1best2bin97.984%0.00078869257385535670.2011214:59:1203:44:485.36 MJ
Run 2best2bin98.008%0.00086789286879177020.2011216:20:2004:05:055.76 MJ
Run 3best2bin98.576%0.00079010804734894020.1964815:18:5203:49:435.37 MJ
Run 4best2bin97.688%0.00066540300794917370.1979215:37:1203:54:185.47 MJ
Run 5best2bin98.2%0.00048945331834283030.1969614:56:4403:44:115.24 MJ
Run 6best2bin97.272%0.000348714254562946970.1980822:47:2005:41:508.26 MJ
Run 7best2bin97.664%0.00048945331834283030.2006413:16:4003:19:104.91 MJ
Run 8best2bin96.464%0.00087212691909229680.1973616:35:2004:08:506.26 MJ
Run 9best2bin97.6%0.00048945331834283030.1969616:07:0004:01:455.74 MJ
Run 10best2bin96.928%0.00079010804734894020.19814:11:0803:32:474.97 MJ
Run 1rand1exp98.672%0.000281358581111197030.2000817:44:2804:26:076.06 MJ
Run 2rand1exp95.208%0.00066540300794917370.1995220:37:2805:09:227.23 MJ
Run 3rand1exp98.544%0.00055037828142615850.1974415:33:0003:53:155.52 MJ
Run 4rand1exp98.424%0.000281358581111197030.1997617:59:1604:29:496.30 MJ
Run 5rand1exp98.6%0.000240772301800391150.1995217:27:3204:21:536.11 MJ
Run 6rand1exp97.984%0.000281358581111197030.1999212:34:0403:08:314.45 MJ
Run 7rand1exp98.4%0.00055037828142615850.1968820:49:4805:12:277.56 MJ
Run 8rand1exp98.184%0.00061756673364093830.219:21:2804:50:226.77 MJ
Run 9rand1exp98.224%0.00060820149663338530.1979215:44:0403:56:015.47 MJ
Run 10rand1exp98.288%0.00054729948019861060.1969617:46:5204:26:436.27 MJ
Run 1rand2exp98.928%0.00090.2006413:29:2803:22:224.74 MJ
Run 2rand2exp97.432%0.00079010804734894020.1999215:01:4003:45:255.18 MJ
Run 3rand2exp97.512%0.00055037828142615850.1975216:30:4404:07:415.77 MJ
Run 4rand2exp97.648%0.00078306096003706240.196814:53:4403:43:265.07 MJ
Run 5rand2exp98.512%0.00086789286879177020.2011216:14:4404:03:415.74 MJ
Run 6rand2exp97.48%0.00055037828142615850.196814:13:2403:33:214.98 MJ
Run 7rand2exp98.56%0.000240772301800391150.1974420:20:0405:05:017.29 MJ
Run 8rand2exp98.28%0.00036900363372548740.2006417:17:5204:19:286.07 MJ
Run 9rand2exp96.912%0.000281358581111197030.2012820:23:2805:05:527.83 MJ
Run 10rand2exp97.904%0.00069980894237103860.19815:21:2003:50:205.42 MJ
Run 1best1exp98.336%0.00079010804734894020.2002416:19:3604:04:546.11 MJ
Run 2best1exp98.168%0.00055037828142615850.1973617:31:1604:22:496.67 MJ
Run 3best1exp97.912%0.00048945331834283030.2011215:44:2803:56:075.51 MJ
Run 4best1exp98%0.00060820149663338530.197215:38:2803:54:375.59 MJ
Run 5best1exp96.976%0.00079010804734894020.2000814:57:4803:44:275.29 MJ
Run 6best1exp98.688%0.00048945331834283030.1976816:05:3204:01:235.80 MJ
Run 7best1exp98.184%0.00086789286879177020.1966414:58:4003:44:405.23 MJ
Run 8best1exp98.248%0.000348714254562946970.2008815:40:4403:55:115.54 MJ
Run 9best1exp98.224%0.00066540300794917370.197617:06:2404:16:366.13 MJ
Run 10best1exp98.44%0.0006078153283690010.1996815:13:4803:48:275.38 MJ
Run 1best2exp98.072%0.00054987225828290920.1971220:04:3605:01:097.13 MJ
Run 2best2exp98.168%0.00087103601246564220.1974418:09:2804:32:226.40 MJ
Run 3best2exp98.416%0.00070926199239390110.1975214:38:4403:39:415.13 MJ
Run 4best2exp98.56%0.000348714254562946970.2009617:38:5604:24:446.26 MJ
Run 5best2exp98.952%0.00086789286879177020.1979213:22:0803:20:324.77 MJ
Run 6best2exp97.632%0.00086789286879177020.1978414:26:1603:36:345.07 MJ
Run 7best2exp96.784%0.000240955028862571570.199616:36:1204:09:035.79 MJ
Run 8best2exp98.008%0.000281358581111197030.2009614:44:0003:41:005.32 MJ
Run 9best2exp98.112%0.00036900363372548740.2011218:54:0804:43:326.61 MJ
Run 10best2exp97.88%0.00055037828142615850.1999216:35:0404:08:465.82 MJ
Run 1currenttobest1exp97.44%0.00053553739028156650.1970417:23:0004:20:456.06 MJ
Run 2currenttobest1exp98.736%0.00054814612519747250.197613:52:3203:28:084.79 MJ
Run 3currenttobest1exp98.104%0.0003645209082058920.200419:24:4404:51:116.78 MJ
Run 4currenttobest1exp97.152%0.00048945331834283030.1967216:59:3204:14:535.96 MJ
Run 5currenttobest1exp98.104%0.000348099763122382060.1998417:24:1204:21:036.12 MJ
Run 6currenttobest1exp97.104%0.00048945331834283030.2001616:53:2404:13:215.91 MJ
Run 7currenttobest1exp98.248%0.00067575243454840060.1999213:20:0803:20:024.73 MJ
Run 8currenttobest1exp98.032%0.00055037828142615850.197217:52:3604:28:096.45 MJ
Run 9currenttobest1exp97.144%0.00060820149663338530.1970415:00:4003:45:105.23 MJ
Run 10currenttobest1exp97.888%0.00060820149663338530.1968816:37:0804:09:175.74 MJ
Table A4. Results obtained from 10 runs of 15 epochs of a hyperoptimized DenseNet121 using a different DE strategy on CIFAR10.
Table A4. Results obtained from 10 runs of 15 epochs of a hyperoptimized DenseNet121 using a different DE strategy on CIFAR10.
RunStrategyMax Accuracy (%)Best LRBest AccuracyCPU TimeElapsed TimeConsumed Energy
Run 1rand1bin91.136%0.00028504921786332150.194641 d 06:51:0807:42:4710.92 MJ
Run 2rand1bin89.84%0.000145334761903245030.19961 d 15:46:0409:56:3113.54 MJ
Run 3rand1bin88.904%0.000162124068907360.200082 d 08:05:2014:01:2019.72 MJ
Run 4rand1bin88.096%0.000174477153877060860.196482 d 14:03:4415:30:5621.79 MJ
Run 5rand1bin78.848%0.000132776061637921520.200081 d 17:55:4810:28:5714.62 MJ
Run 6rand1bin84.84%0.000402172631010210240.195521 d 16:39:3610:09:5414.22 MJ
Run 7rand1bin86.832%0.000137725209534367780.19961 d 13:27:4409:21:5614.32 MJ
Run 8rand1bin91.024%0.000241130783042947580.199841 d 20:35:0011:08:4515.61 MJ
Run 9rand1bin77.488%0.000139592689798095450.198482 d 07:07:0413:46:4619.14 MJ
Run 10rand1bin90.2%0.000183768490814301860.1962 d 02:35:0412:38:4618.53 MJ
Run 1best1bin88.888%0.000170846510541198340.195761 d 05:08:3207:17:0810.27 MJ
Run 2best1bin82.096%0.000149869991507051220.198641 d 07:07:0807:46:4710.74 MJ
Run 3best1bin54.912%0.00032823638086059630.197121 d 16:01:3610:00:2414.08 MJ
Run 4best1bin87.728%0.00025388759996056110.190961 d 06:27:0007:36:4510.75 MJ
Run 5best1bin86.24%0.000132447958668769260.199281 d 14:55:5209:43:5813.72 MJ
Run 6best1bin87.232%0.000187725990784069880.195441 d 10:27:2008:36:5011.88 MJ
Run 7best1bin91.024%0.00019799833108324090.197281 d 11:50:1608:57:3412.46 MJ
Run 8best1bin87.048%0.000204520793570471350.199121 d 22:41:0011:40:1516.45 MJ
Run 9best1bin86.744%0.000116978557920244460.199761 d 16:49:2410:12:2114.36 MJ
Run 10best1bin87.848%0.000118781444167737950.199282 d 01:29:0812:22:1717.48 MJ
Run 1currenttobest1bin94.408%0.00017771013478759730.198481 d 12:49:1209:12:1812.76 MJ
Run 2currenttobest1bin91.664%0.00016767751239408380.19681 d 21:12:3611:18:0915.98 MJ
Run 3currenttobest1bin87.704%0.000249274270394843360.199681 d 17:18:0810:19:3214.53 MJ
Run 4currenttobest1bin93.104%0.000125251673167103550.198641 d 10:23:4408:35:5612.22 MJ
Run 5currenttobest1bin76.968%0.000112719265752493690.20081 d 15:54:4809:58:4213.94 MJ
Run 6currenttobest1bin92.656%0.000240772301800391150.198641 d 15:43:3609:55:5414.74 MJ
Run 7currenttobest1bin83.928%0.000180056271732636770.200082 d 04:49:0013:12:1518.52 MJ
Run 8currenttobest1bin89.832%0.000180056271732636770.197281 d 14:02:0009:30:3013.37 MJ
Run 9currenttobest1bin79.304%0.000141609709395538360.200241 d 18:51:1610:42:4914.97 MJ
Run 10currenttobest1bin92.944%0.00017099711310533570.199521 d 08:37:1208:09:1811.34 MJ
Run 1rand2bin87.672%0.000140437197525316070.198642 d 09:31:2814:22:5220.78 MJ
Run 2rand2bin89.04%0.000112257164111145730.198882 d 10:33:2814:38:2220.83 MJ
Run 3rand2bin94.512%0.00021335012431247070.196242 d 02:37:0812:39:1717.67 MJ
Run 4rand2bin83.176%0.000146539418890181440.19642 d 09:09:0814:17:1719.81 MJ
Run 5rand2bin87.024%0.000119533641653748430.197282 d 05:57:2413:29:2118.39 MJ
Run 6rand2bin84.768%0.000235532893505987780.199682 d 04:24:0413:06:0117.81 MJ
Run 7rand2bin87.648%0.00022375922465134610.199361 d 17:36:3610:24:0914.43 MJ
Run 8rand2bin88.624%0.000284500768355017930.197842 d 14:21:2815:35:2221.70 MJ
Run 9rand2bin87.976%0.00013814089102834190.200082 d 01:54:1212:28:3317.34 MJ
Run 10rand2bin88.936%0.00037455290836552160.196642 d 02:32:4012:38:1017.96 MJ
Run 1best2bin87.808%0.000198024254596845550.197921 d 14:59:4809:44:5713.70 MJ
Run 2best2bin92.296%0.00012513961710271520.199441 d 12:31:3609:07:5412.92 MJ
Run 3best2bin90.4%0.00019042703378366230.19641 d 17:50:4410:27:4114.68 MJ
Run 4best2bin81.952%0.00010994762254373910.198882 d 01:55:3612:28:5417.65 MJ
Run 5best2bin75.464%0.000202918619823005050.196721 d 20:22:4811:05:4215.79 MJ
Run 6best2bin86.48%0.000100609821970013250.20042 d 10:13:3214:33:2320.22 MJ
Run 7best2bin88.752%0.000164919716241819510.199681 d 13:26:4009:21:4012.71 MJ
Run 8best2bin93.512%0.000240772301800391150.198241 d 21:12:1211:18:0316.09 MJ
Run 9best2bin89.448%0.00017147904746484410.200081 d 19:37:0010:54:1515.19 MJ
Run 10best2bin84.432%0.000144875076772761470.199282 d 13:24:0415:21:0121.30 MJ
Run 1rand1exp81.736%0.000205728364619391120.197681 d 12:40:0009:10:0012.93 MJ
Run 2rand1exp90.736%0.00012233353653173930.200241 d 13:13:2009:18:2012.97 MJ
Run 3rand1exp81.24%0.000281358581111197030.198882 d 03:22:2812:50:3718.10 MJ
Run 4rand1exp88.952%0.00017169553081330660.19882 d 01:32:4412:23:1117.78 MJ
Run 5rand1exp84.584%0.000117389803325508950.198241 d 20:13:0811:03:1715.40 MJ
Run 6rand1exp88.16%0.00017169553081330660.196882 d 12:53:4815:13:2721.12 MJ
Run 7rand1exp79.792%0.00034027337350456030.196561 d 12:26:0809:06:3212.72 MJ
Run 8rand1exp90.84%0.00018706144980548440.199681 d 22:59:5211:44:5816.24 MJ
Run 9rand1exp85.648%0.000319838608615680450.198241 d 19:27:2410:51:5115.32 MJ
Run 10rand1exp92.024%0.000180745436216547910.199681 d 22:00:3611:30:0915.91 MJ
Run 1rand2exp77.792%0.00036115292896757650.195762 d 13:49:2015:27:2021.95 MJ
Run 2rand2exp88.48%0.000178378401766325270.19882 d 03:06:4012:46:4017.93 MJ
Run 3rand2exp89.08%0.00026538470587646220.195681 d 19:58:4010:59:4014.96 MJ
Run 4rand2exp93.68%0.000125251673167103550.195682 d 09:39:2014:24:5021.03 MJ
Run 5rand2exp79.16%0.000114071407139993420.197681 d 18:18:1610:34:3414.92 MJ
Run 6rand2exp83.536%0.000173013482800916720.198562 d 02:53:2412:43:2117.94 MJ
Run 7rand2exp88.992%0.00028428807600572670.196641 d 19:29:4010:52:2515.56 MJ
Run 8rand2exp87.232%0.00037501598771371280.199522 d 15:49:4015:57:2522.13 MJ
Run 9rand2exp83.456%0.000158128240204336730.198161 d 21:09:0811:17:1716.10 MJ
Run 10rand2exp91.888%0.000177933882522167020.196241 d 22:20:5211:35:1316.15 MJ
Run 1best1exp88.696%0.000140338368546473340.197361 d 02:15:3606:33:549.31 MJ
Run 2best1exp86.968%0.00021624245414767140.197681 d 15:12:5609:48:1414.52 MJ
Run 3best1exp92.84%0.000236380949017823160.197361 d 07:52:0807:58:0211.21 MJ
Run 4best1exp86.672%0.000104271954752884740.1941 d 10:19:2808:34:5212.09 MJ
Run 5best1exp80.672%0.000194277056381570930.198881 d 15:29:5609:52:2913.53 MJ
Run 6best1exp77.904%0.000151643268009670440.197921 d 10:28:2408:37:0612.13 MJ
Run 7best1exp88.776%0.000125251673167103550.199521 d 11:42:5208:55:4312.57 MJ
Run 8best1exp86.32%0.000106351833725319330.197921 d 05:52:4007:28:1010.53 MJ
Run 9best1exp90.912%0.000164033598096865880.198321 d 13:23:4809:20:5712.89 MJ
Run 10best1exp86.616%0.000172518782329636050.198321 d 13:11:0409:17:4613.27 MJ
Run 1best2exp84.392%0.000175484195064109670.197361 d 13:47:5609:26:5913.40 MJ
Run 2best2exp85.904%0.000214385131736420150.196482 d 02:36:5612:39:1417.28 MJ
Run 3best2exp80.952%0.000106150864909526490.198641 d 15:33:1609:53:1913.77 MJ
Run 4best2exp91.496%0.00025862261362331860.195761 d 18:41:2410:40:2114.70 MJ
Run 5best2exp80.376%0.000178804635069251980.195842 d 00:10:2012:02:3516.64 MJ
Run 6best2exp75.912%0.000107198372386767010.198882 d 10:04:1614:31:0420.64 MJ
Run 7best2exp89.456%0.000185248950708825030.199281 d 23:31:0011:52:4516.85 MJ
Run 8best2exp87.856%0.000156378470833566130.195922 d 03:50:2012:57:3519.63 MJ
Run 9best2exp83.952%0.00028949860369495420.195281 d 12:47:0009:11:4513.00 MJ
Run 10best2exp90.384%0.000212672037043803580.195281 d 23:50:3611:57:3916.27 MJ
Run 1currenttobest1exp86.328%0.000161101683252510750.199841 d 15:00:1209:45:0313.84 MJ
Run 2currenttobest1exp83.032%0.00017415581518198490.19921 d 12:21:1609:05:1912.90 MJ
Run 3currenttobest1exp74.632%0.00019006815953323860.197121 d 15:17:1609:49:1914.05 MJ
Run 4currenttobest1exp94.304%0.00017378490513045970.196321 d 07:09:2407:47:2110.74 MJ
Run 5currenttobest1exp87.336%0.000339222149472093370.193841 d 12:41:5609:10:2913.80 MJ
Run 6currenttobest1exp81.44%0.000160207314299752770.198962 d 13:42:4815:25:4221.54 MJ
Run 7currenttobest1exp89.128%0.000158185662753083560.197682 d 02:56:5212:44:1318.15 MJ
Run 8currenttobest1exp88.272%0.00021787071189150580.198241 d 13:46:4009:26:4012.99 MJ
Run 9currenttobest1exp84.672%0.000170591610982738080.19922 d 06:39:4413:39:5618.98 MJ
Run 10currenttobest1exp77.736%0.000210962963167739990.199362 d 00:41:4012:10:2516.82 MJ
Table A5. Results obtained from 10 runs of 15 epochs of a hyperoptimized ResNet18 using a different DE strategy on CIFAR100.
Table A5. Results obtained from 10 runs of 15 epochs of a hyperoptimized ResNet18 using a different DE strategy on CIFAR100.
RunStrategyMax Accuracy (%)Best LRBest AccuracyCPU TimeElapsed TimeConsumed Energy
Run 1rand1bin86.208%0.00068984920938774130.02081 d 18:49:5610:42:2914.01 MJ
Run 2rand1bin86.912%0.00062881623854450660.02081 d 19:05:1610:46:1914.15 MJ
Run 3rand1bin88.992%0.00064835023608667560.020722 d 02:11:4812:32:5716.05 MJ
Run 4rand1bin89.6%0.00069594787236569610.020641 d 21:23:4811:20:5714.19 MJ
Run 5rand1bin84.64%0.0003269047665347310.020882 d 07:04:1613:46:0417.39 MJ
Run 6rand1bin92.368%0.00056300591480615780.019522 d 02:48:0012:42:0016.57 MJ
Run 7rand1bin86.184%0.00055037828142615850.020721 d 17:27:2410:21:5112.85 MJ
Run 8rand1bin92.056%0.000281358581111197030.020722 d 04:50:2813:12:3716.78 MJ
Run 9rand1bin85.8%0.00066486934509922390.01962 d 05:45:1213:26:1816.98 MJ
Run 10rand1bin86.008%0.000348714254562946970.020882 d 06:00:0813:30:0217.20 MJ
Run 1best1bin85.728%0.00051630825576052520.019121 d 16:27:0810:06:4713.08 MJ
Run 2best1bin92.064%0.00046707488335297150.019361 d 23:05:0411:46:1615.44 MJ
Run 3best1bin87.264%0.00062215036995484390.019121 d 08:42:5208:10:4310.45 MJ
Run 4best1bin92.472%0.00055623072116259860.019282 d 01:43:0812:25:4716.01 MJ
Run 5best1bin87.312%0.000430210192436925540.01961 d 08:56:5208:14:1310.45 MJ
Run 6best1bin91.048%0.00086201711638765460.01962 d 08:03:2414:00:5117.56 MJ
Run 7best1bin89.144%0.00057715459547612160.020641 d 07:14:3207:48:389.91 MJ
Run 8best1bin88.952%0.00063512752531727710.020721 d 16:52:0010:13:0013.30 MJ
Run 9best1bin89.512%0.00050908104365961960.019281 d 15:08:3209:47:0813.28 MJ
Run 10best1bin88.976%0.000561500864128150.020881 d 17:22:0810:20:3213.43 MJ
Run 1currenttobest1bin87.6%0.000279847067709224030.02072¸1-11:50:0408:57:3111.53 MJ
Run 2currenttobest1bin92.32%0.00072725978208878750.020721 d 16:04:4010:01:1012.72 MJ
Run 3currenttobest1bin89.832%0.00058568945543531990.019441 d 18:41:5210:40:2813.40 MJ
Run 4currenttobest1bin91.152%0.000357274261495859670.020721 d 06:32:0407:38:019.78 MJ
Run 5currenttobest1bin87.176%0.00056656143641747470.020561 d 06:41:2407:40:219.99 MJ
Run 6currenttobest1bin84.704%0.00041770380142071810.020641 d 23:06:4411:46:4115.39 MJ
Run 7currenttobest1bin84.368%0.00052308410397135560.020881 d 20:33:4811:08:2714.57 MJ
Run 8currenttobest1bin90.44%0.000337510710015804140.01961 d 15:52:2809:58:0713.47 MJ
Run 9currenttobest1bin90.448%0.00065331106809237150.01922 d 05:18:2413:19:3617.36 MJ
Run 10currenttobest1bin83.344%0.00059970525547956540.020641 d 15:49:3209:57:2312.94 MJ
Run 1rand2bin91.8%0.000459618357711377460.020641 d 21:38:0011:24:3014.84 MJ
Run 2rand2bin85.976%0.00086789286879177020.01921 d 21:09:5211:17:2814.38 MJ
Run 3rand2bin89.632%0.00051484778073373150.020722 d 00:54:2412:13:3615.97 MJ
Run 4rand2bin90.064%0.00056500986735631390.019121 d 14:39:5609:39:5912.22 MJ
Run 5rand2bin93.344%0.00042677915067381040.020562 d 05:05:2413:16:2117.06 MJ
Run 6rand2bin87.56%0.00054778425326925080.020721 d 08:54:0008:13:3010.46 MJ
Run 7rand2bin91.312%0.00069282866076394060.020881 d 15:54:2009:58:3513.12 MJ
Run 8rand2bin90.288%0.0005454219815028210.02082 d 00:31:4812:07:5715.45 MJ
Run 9rand2bin86.84%0.00055029495472942590.020482 d 00:12:4812:03:1215.34 MJ
Run 10rand2bin88.464%0.00074105918425435070.019441 d 08:31:1208:07:4810.15 MJ
Run 1best2bin84.672%0.00047618164342913910.019122 d 01:09:2812:17:2215.34 MJ
Run 2best2bin87.872%0.00066540300794917370.02081 d 21:00:5211:15:1314.19 MJ
Run 3best2bin90.928%0.00079282416520035030.020641 d 12:51:4409:12:5612.57 MJ
Run 4best2bin92.216%0.00086789286879177020.02041 d 15:06:5209:46:4312.68 MJ
Run 5best2bin87.256%0.00089073588939520290.019441 d 10:50:3208:42:3810.86 MJ
Run 6best2bin91.592%0.000481866532636078860.019682 d 03:55:2412:58:5116.20 MJ
Run 7best2bin90.336%0.00055037828142615850.018881 d 08:30:1208:07:3310.23 MJ
Run 8best2bin89.12%0.000296935324494636360.019441 d 10:44:0808:41:0210.95 MJ
Run 9best2bin90.456%0.00048945331834283030.019121 d 10:40:4408:40:1111.33 MJ
Run 10best2bin91.448%0.000425518154201565250.02082 d 03:48:2412:57:0616.85 MJ
Run 1rand1exp86.608%0.00054736094255330860.019121 d 19:23:3210:50:5313.95 MJ
Run 2rand1exp89.832%0.0007862965598365720.020722 d 01:47:1212:26:4815.72 MJ
Run 3rand1exp85.256%0.000348714254562946970.019441 d 02:00:5206:30:138.21 MJ
Run 4rand1exp88.92%0.00060820149663338530.020641 d 20:51:3211:12:5314.37 MJ
Run 5rand1exp88.168%0.000478116039122540330.020721 d 21:33:2811:23:2214.69 MJ
Run 6rand1exp84.736%0.00088935190385597660.020722 d 00:20:4412:05:1115.44 MJ
Run 7rand1exp92.112%0.00079010804734894020.019121 d 09:10:3608:17:3910.51 MJ
Run 8rand1exp91.304%0.00054381699902039260.019121 d 11:56:5608:59:1411.46 MJ
Run 9rand1exp90.808%0.00076790847924023330.020882 d 06:16:5613:34:1417.70 MJ
Run 10rand1exp87.128%0.00060026955498432180.01962 d 06:08:5613:32:1417.20 MJ
Run 1rand2exp85.656%0.00031276001793953110.020641 d 17:26:0010:21:3012.91 MJ
Run 2rand2exp90.152%0.000481866532636078860.020721 d 22:14:1211:33:3314.81 MJ
Run 3rand2exp88.856%0.00053231768520899230.01961 d 11:39:4808:54:5711.54 MJ
Run 4rand2exp88.592%0.00034054993035967780.01921 d 14:34:5209:38:4311.98 MJ
Run 5rand2exp92.376%0.00058626254858704740.018962 d 01:45:2812:26:2217.13 MJ
Run 6rand2exp90.496%0.00064518151931689590.02081 d 13:49:2009:27:2012.45 MJ
Run 7rand2exp91.048%0.000481088784996103230.020882 d 04:12:2413:03:0616.81 MJ
Run 8rand2exp83.296%0.00048945331834283030.020721 d 05:34:5607:23:449.28 MJ
Run 9rand2exp85.08%0.000281358581111197030.020722 d 04:23:3613:05:5417.13 MJ
Run 10rand2exp84.264%0.0003654904671636250.01921 d 17:33:2010:23:2013.18 MJ
Run 1best1exp93.064%0.00054337899609328560.019121 d 13:59:4409:29:5612.15 MJ
Run 2best1exp87.8%0.00054725449679144710.020641 d 21:07:4811:16:5714.13 MJ
Run 3best1exp90.168%0.00054725449679144710.020641 d 18:50:2010:42:3513.72 MJ
Run 4best1exp89.512%0.00049063267867901810.01961 d 20:49:4411:12:2614.60 MJ
Run 5best1exp89.624%0.0006999429601028720.019121 d 10:07:2408:31:5110.92 MJ
Run 6best1exp87.144%0.00059691448335751550.019681 d 19:11:5610:47:5914.06 MJ
Run 7best1exp89.528%0.0005640138513339920.02082 d 01:29:4412:22:2615.76 MJ
Run 8best1exp78.872%0.00049024433106317030.019121 d 15:10:1209:47:3313.32 MJ
Run 9best1exp82.184%0.00066540300794917370.020641 d 08:45:4008:11:2510.76 MJ
Run 10best1exp90.832%0.00066540300794917370.02081 d 19:13:0010:48:1513.44 MJ
Run 1best2exp88.608%0.000404342692977507760.019121 d 18:05:0810:31:1713.43 MJ
Run 2best2exp91.648%0.0004253692226581140.019442 d 04:18:3613:04:3917.01 MJ
Run 3best2exp93.136%0.00053553739028156650.02081 d 23:33:4811:53:2715.06 MJ
Run 4best2exp88.712%0.00055037828142615850.018881 d 15:01:1609:45:1912.12 MJ
Run 5best2exp89.272%0.00079010804734894020.020721 d 14:47:0409:41:4612.18 MJ
Run 6best2exp84.12%0.00057308288774453760.01961 d 19:36:2010:54:0513.81 MJ
Run 7best2exp88.176%0.00036900363372548740.02081 d 11:44:0008:56:0011.46 MJ
Run 8best2exp88.624%0.000439746933463485140.01961 d 16:58:4410:14:4113.09 MJ
Run 9best2exp88.912%0.00083325181170246220.019361 d 07:06:1207:46:3310.04 MJ
Run 10best2exp85.536%0.000425518154201565250.020721 d 21:11:1211:17:4814.50 MJ
Run 1currenttobest1exp87.472%0.00086789286879177020.020961 d 23:35:1611:53:4914.96 MJ
Run 2currenttobest1exp85.928%0.00051049755728025080.019681 d 19:12:1210:48:0313.85 MJ
Run 3currenttobest1exp86.992%0.00069677327913349380.02081 d 22:33:0811:38:1715.21 MJ
Run 4currenttobest1exp88.152%0.00051156463703824240.020882 d 09:01:5614:15:2918.32 MJ
Run 5currenttobest1exp81.56%0.00039927131902173740.01881 d 08:30:5208:07:4310.46 MJ
Run 6currenttobest1exp91.048%0.00084320998785448760.020721 d 20:49:1611:12:1914.58 MJ
Run 7currenttobest1exp86.016%0.00048945331834283030.020881 d 06:42:0407:40:319.97 MJ
Run 8currenttobest1exp89.528%0.00063901829972860930.020961 d 20:04:1611:01:0413.92 MJ
Run 9currenttobest1exp93.392%0.00067797662069228890.020642 d 07:44:1613:56:0417.74 MJ
Run 10currenttobest1exp86.768%0.000281358581111197030.02082 d 06:11:2013:32:5017.20 MJ
Table A6. Results obtained from 10 runs of 15 epochs of a hyperoptimized VGG11 using a different DE strategy on CIFAR100.
Table A6. Results obtained from 10 runs of 15 epochs of a hyperoptimized VGG11 using a different DE strategy on CIFAR100.
RunStrategyMax Accuracy (%)Best LRBest AccuracyCPU TimeElapsed TimeConsumed Energy
Run 1rand1bin12.28%0.00047584408423186240.011843 d 16:49:0422:12:1631.68 MJ
Run 2rand1bin3.136%0.000281981779020857960.008563 d 18:30:2022:37:3531.84 MJ
Run 3rand1bin7.8%0.0004236808876326440.007523 d 17:03:1622:15:4930.87 MJ
Run 4rand1bin8.232%0.00021356001028830450.007443 d 18:26:0822:36:3231.36 MJ
Run 5rand1bin4.248%0.00026120968493207280.007523 d 18:13:2822:33:2231.14 MJ
Run 6rand1bin8.496%0.00019484233836430410.010163 d 20:05:3623:01:2432.09 MJ
Run 7rand1bin12.176%0.000270934942469088250.012883 d 19:56:5222:59:1332.37 MJ
Run 8rand1bin9.056%0.00028046682174055390.01283 d 19:09:1622:47:1932.17 MJ
Run 9rand1bin11.696%0.000106344293670925760.008963 d 16:50:2422:12:3631.03 MJ
Run 10rand1bin11.744%0.00021670745464969620.007923 d 20:46:4023:11:4032.23 MJ
Run 1best1bin2.48%0.00031723223263937950.00763 d 18:38:4022:39:4031.27 MJ
Run 2best1bin3.616%0.00058073742617806280.008083 d 16:38:1222:09:3331.24 MJ
Run 3best1bin4.576%0.000263010632545849130.012083 d 18:53:4822:43:2732.50 MJ
Run 4best1bin8.296%0.000285398354095132850.008083 d 22:39:4023:39:5533.27 MJ
Run 5best1bin6.304%0.000154810752832934680.012723 d 21:24:3223:21:0832.89 MJ
Run 6best1bin11.888%0.000334715063466958760.008323 d 16:45:3622:11:2430.62 MJ
Run 7best1bin10.416%0.00054994020339401770.012643 d 18:37:1222:39:1831.22 MJ
Run 8best1bin12.752%0.00042439008491612080.012563 d 19:31:2822:52:5231.19 MJ
Run 9best1bin6.832%0.000257819046275936860.012643 d 19:30:5622:52:4432.49 MJ
Run 10best1bin4.792%0.00019467380685527680.010563 d 20:33:2423:08:2132.96 MJ
Run 1currenttobest1bin11.456%0.00056314873089432990.012323 d 21:13:2823:18:2232.47 MJ
Run 2currenttobest1bin10.064%0.000150893341637552920.009283 d 20:25:5623:06:2932.70 MJ
Run 3currenttobest1bin14.8%0.00028605259560629030.008083 d 20:16:4023:04:1032.45 MJ
Run 4currenttobest1bin8.448%0.000125175101793034770.009283 d 21:55:4023:28:5532.90 MJ
Run 5currenttobest1bin8.352%0.000186906802300600930.008323 d 21:26:4823:21:4233.46 MJ
Run 6currenttobest1bin10.168%0.000188579038331683560.012643 d 18:19:3622:34:5431.69 MJ
Run 7currenttobest1bin12.568%0.000496200708902690.0123 d 14:53:1221:43:1830.73 MJ
Run 8currenttobest1bin8.728%0.000220511219859552880.013283 d 17:43:0822:25:4731.10 MJ
Run 9currenttobest1bin5.608%0.000326457930426110230.007923 d 18:37:1222:39:1831.60 MJ
Run 10currenttobest1bin6.488%0.000258156328093057060.0083 d 19:09:5222:47:2831.44 MJ
Run 1rand2bin1.08%0.00060030627844375550.012243 d 16:14:2422:03:3630.24 MJ
Run 2rand2bin9.376%0.000117765020325617310.00843 d 18:53:4822:43:2733.35 MJ
Run 3rand2bin9.952%0.000169149952991629840.008643 d 21:59:2423:29:5132.14 MJ
Run 4rand2bin11.336%0.0007201894074829240.0123 d 15:40:3621:55:0930.06 MJ
Run 5rand2bin4.76%0.000159135423440856680.013363 d 20:21:1623:05:1932.07 MJ
Run 6rand2bin12.36%0.000173886158605046570.008163 d 18:23:3222:35:5331.91 MJ
Run 7rand2bin3.416%0.00018763856715490550.00843 d 18:58:5222:44:4332.16 MJ
Run 8rand2bin3.752%0.00060567780451118880.008083 d 17:12:4022:18:1031.03 MJ
Run 9rand2bin1.104%0.000299030400595174530.007923 d 19:36:5622:54:1432.24 MJ
Run 10rand2bin9.472%0.000552888650251980.008323 d 16:42:2822:10:3732.48 MJ
Run 1best2bin3.536%0.000315161163824600750.008243 d 17:27:0422:21:4631.42 MJ
Run 2best2bin1.104%0.000479453586031871270.008163 d 17:03:2022:15:5031.38 MJ
Run 3best2bin7.808%0.000470698209856876960.0123 d 18:12:2422:33:0631.44 MJ
Run 4best2bin10.816%0.000480349523965992370.011923 d 20:37:1223:09:1831.99 MJ
Run 5best2bin8.384%0.00020735302535465910.013043 d 19:33:4422:53:2633.49 MJ
Run 6best2bin9.128%0.000222108475238624160.013123 d 15:46:0821:56:3230.58 MJ
Run 7best2bin10.336%0.000238120489657507710.008243 d 18:12:3222:33:0833.00 MJ
Run 8best2bin9.632%0.000443761460339041660.012563 d 17:34:1222:23:3331.00 MJ
Run 9best2bin7.96%0.000119557229071073480.007683 d 18:27:5222:36:5831.57 MJ
Run 10best2bin9.464%0.000411281065862328170.007443 d 15:48:3221:57:0830.42 MJ
Run 1rand1exp3.744%0.000177851449535845640.00843 d 17:36:4422:24:1131.61 MJ
Run 2rand1exp6.32%0.000231228619529638840.007683 d 18:25:4422:36:2631.52 MJ
Run 3rand1exp10.728%0.000259308954328450550.008243 d 18:17:1222:34:1831.46 MJ
Run 4rand1exp7.568%0.00033383493799703290.007683 d 19:57:2822:59:2231.56 MJ
Run 5rand1exp14.32%0.00030178952201016950.007843 d 19:44:2022:56:0531.35 MJ
Run 6rand1exp10.576%0.00044010829551402030.0083 d 16:56:3222:14:0830.49 MJ
Run 7rand1exp7.424%0.000168944076535678270.008563 d 18:07:5622:31:5931.75 MJ
Run 8rand1exp2.072%0.000285823831180450740.006883 d 18:17:0422:34:1630.64 MJ
Run 9rand1exp10.528%0.000228739831528681840.007443 d 18:55:3622:43:5432.56 MJ
Run 10rand1exp9.416%0.000223626969614144230.0123 d 19:33:3622:53:2432.41 MJ
Run 1rand2exp11.736%0.00061997894061847110.0143 d 17:02:2422:15:3631.43 MJ
Run 2rand2exp9.872%0.00042791770775277270.007683 d 17:46:0022:26:3031.44 MJ
Run 3rand2exp8.336%0.000176319741419159440.007683 d 18:45:1222:41:1831.49 MJ
Run 4rand2exp7.512%0.00056789370413517770.011683 d 23:10:4023:47:4033.71 MJ
Run 5rand2exp13.488%0.0005701324081046430.013443 d 15:35:2021:53:5030.36 MJ
Run 6rand2exp6.992%0.00051236900730920890.012963 d 18:45:5622:41:2932.21 MJ
Run 7rand2exp11.96%0.000104495370906287530.009043 d 21:32:0823:23:0233.40 MJ
Run 8rand2exp8.992%0.00073764847616296210.011923 d 19:52:5222:58:1331.66 MJ
Run 9rand2exp9.68%0.000206878981720081280.006963 d 18:52:0822:43:0231.40 MJ
Run 10rand2exp6.912%0.000175524661460829780.014723 d 19:54:0022:58:3032.53 MJ
Run 1best1exp9.488%0.000231312948537657560.007523 d 20:08:2823:02:0732.57 MJ
Run 2best1exp5.376%0.00034678248563878950.007363 d 20:37:5223:09:2832.15 MJ
Run 3best1exp4.656%0.00054725449679144710.012083 d 16:19:1622:04:4932.43 MJ
Run 4best1exp9.088%0.00021950713606257210.01323 d 19:25:1222:51:1831.09 MJ
Run 5best1exp10.32%0.00070703962611645710.011923 d 17:19:0022:19:4531.00 MJ
Run 6best1exp8.968%0.000400013644409781130.007763 d 17:12:0022:18:0031.42 MJ
Run 7best1exp9.352%0.00060856749937464260.011763 d 18:15:2022:33:5031.41 MJ
Run 8best1exp5.912%0.00023334662588551490.007363 d 19:12:2822:48:0731.99 MJ
Run 9best1exp5.496%0.00047863716746602260.008083 d 17:39:0022:24:4531.38 MJ
Run 10best1exp12.632%0.000215933424049124720.0083 d 18:00:2822:30:0731.64 MJ
Run 1best2exp13.896%0.00061671353081296440.012643 d 17:46:2822:26:3730.52 MJ
Run 2best2exp4.352%0.000176322885568682050.007443 d 17:51:5222:27:5831.16 MJ
Run 3best2exp11.44%0.00052115819866432810.013283 d 17:56:2822:29:0731.05 MJ
Run 4best2exp10.936%0.000115513385100911640.008883 d 20:16:3623:04:0931.66 MJ
Run 5best2exp11.936%0.00025795502748807250.007363 d 21:41:0423:25:1632.77 MJ
Run 6best2exp14.464%0.000203088464681672240.012163 d 21:53:4823:28:2732.60 MJ
Run 7best2exp7.72%0.0002478962130905930.012563 d 17:55:4822:28:5732.04 MJ
Run 8best2exp7.544%0.000148863802317940870.008483 d 21:18:4023:19:4032.27 MJ
Run 9best2exp9.896%0.000254055915534197650.008083 d 19:19:0822:49:4733.55 MJ
Run 10best2exp7.424%0.00055531978511636590.00843 d 15:02:4421:45:4130.28 MJ
Run 1currenttobest1exp12.896%0.000413646896573837350.012483 d 19:28:2022:52:0532.15 MJ
Run 2currenttobest1exp9.288%0.00052646793903516830.012483 d 15:57:3621:59:2430.83 MJ
Run 3currenttobest1exp7.104%0.000191177678080248580.012083 d 20:12:1623:03:0431.82 MJ
Run 4currenttobest1exp3.24%0.00015107979147780760.008243 d 19:59:2822:59:5231.85 MJ
Run 5currenttobest1exp8.952%0.000482674060411273960.012323 d 22:45:1623:41:1933.33 MJ
Run 6currenttobest1exp6.56%0.000196705301132065820.008083 d 16:54:0822:13:3230.96 MJ
Run 7currenttobest1exp8.856%0.000206565843650162560.0083 d 20:54:4423:13:4132.03 MJ
Run 8currenttobest1exp6.824%0.00059828593990298910.012163 d 19:25:0422:51:1632.40 MJ
Run 9currenttobest1exp11.28%0.000256747445733641550.01483 d 19:32:1622:53:0433.62 MJ
Run 10currenttobest1exp4.776%0.000216769933398123830.008243 d 19:14:4822:48:4231.50 MJ
Table A7. Results obtained from 10 runs of 15 epochs of a hyperoptimized ConvNeXtSmall using a different DE strategy on CIFAR100.
Table A7. Results obtained from 10 runs of 15 epochs of a hyperoptimized ConvNeXtSmall using a different DE strategy on CIFAR100.
RunStrategyMax Accuracy (%)Best LRBest AccuracyCPU TimeElapsed TimeConsumed Energy
Run 1rand1bin96.856%0.00079010804734894020.020961 d 08:40:2408:10:0611.65 MJ
Run 2rand1bin95.152%0.00068894206126785950.0209623:50:3605:57:398.54 MJ
Run 3rand1bin95.912%0.00048945331834283030.020881 d 03:21:3606:50:2410.06 MJ
Run 4rand1bin96.96%0.00079010804734894020.019681 d 03:52:3606:58:0910.06 MJ
Run 5rand1bin95.168%0.00077246695797591720.020961 d 13:22:3209:20:3813.23 MJ
Run 6rand1bin97.6%0.00068210912296230880.020961 d 03:10:0806:47:329.21 MJ
Run 7rand1bin96.544%0.00058431083616292040.0209622:45:3205:41:238.04 MJ
Run 8rand1bin94.272%0.00077246695797591720.019617:44:4004:26:106.23 MJ
Run 9rand1bin96.624%0.00079010804734894020.020961 d 02:34:3206:38:389.60 MJ
Run 10rand1bin94.296%0.00079010804734894020.0208822:40:0405:40:018.16 MJ
Run 1best1bin96.768%0.00066540300794917370.0209618:06:1604:31:346.50 MJ
Run 2best1bin98.4%0.00075254780224504780.020961 d 01:08:0406:17:018.86 MJ
Run 3best1bin97.84%0.00072946703521472210.019681 d 02:39:4006:39:559.31 MJ
Run 4best1bin97.184%0.00072284005489201620.0196819:41:5604:55:296.83 MJ
Run 5best1bin98.24%0.00060820149663338530.0209619:04:2404:46:066.84 MJ
Run 6best1bin95.816%0.00055037828142615850.0209618:34:0404:38:316.92 MJ
Run 7best1bin94.456%0.00087212691909229680.020961 d 01:10:5206:17:438.88 MJ
Run 8best1bin97.152%0.00069594787236569610.019622:53:3205:43:237.94 MJ
Run 9best1bin97.592%0.00069769342229193240.0192816:39:0804:09:475.83 MJ
Run 10best1bin96.296%0.00060011434697409240.0208823:14:3605:48:398.32 MJ
Run 1currenttobest1bin97.448%0.00079010804734894020.0196820:32:3205:08:087.84 MJ
Run 2currenttobest1bin96.792%0.00060820149663338530.020961 d 02:38:3206:39:389.94 MJ
Run 3currenttobest1bin98.48%0.00062603156445300290.019621:25:0405:21:167.69 MJ
Run 4currenttobest1bin85.752%0.00052850312379018420.020961 d 01:10:4806:17:429.12 MJ
Run 5currenttobest1bin98.36%0.00066540300794917370.0196818:50:5604:42:446.71 MJ
Run 6currenttobest1bin98.208%0.00075385180672504920.0209622:32:0805:38:028.64 MJ
Run 7currenttobest1bin98.312%0.00057397123233140840.020961 d 02:36:4806:39:129.61 MJ
Run 8currenttobest1bin97.296%0.00066540300794917370.019681 d 01:29:2406:22:219.14 MJ
Run 9currenttobest1bin89.992%0.00060534389698465670.0209619:02:4804:45:426.71 MJ
Run 10currenttobest1bin98.448%0.00086789286879177020.019681 d 05:51:1207:27:4810.39 MJ
Run 1rand2bin98.624%0.00060255197947524050.019681 d 07:56:2807:59:0711.09 MJ
Run 2rand2bin98.032%0.00060820149663338530.020961 d 00:00:2006:00:058.28 MJ
Run 3rand2bin95.968%0.00060820149663338530.0196821:12:4805:18:127.49 MJ
Run 4rand2bin94.856%0.00058672949207391940.019681 d 07:18:0407:49:3110.94 MJ
Run 5rand2bin98.888%0.0005187833862561420.020961 d 06:55:5607:43:5911.38 MJ
Run 6rand2bin95.056%0.00066540300794917370.019623:30:0805:52:328.33 MJ
Run 7rand2bin93.936%0.0007164175242331650.019681 d 08:31:4008:07:5511.33 MJ
Run 8rand2bin96.896%0.00070741162664776210.0196822:51:0805:42:478.10 MJ
Run 9rand2bin97.248%0.00050189758331921160.020961 d 14:38:4009:39:4013.91 MJ
Run 10rand2bin96.256%0.00071071412129867450.0209623:21:1605:50:198.32 MJ
Run 1best2bin94.432%0.00079010804734894020.0209622:57:4405:44:268.12 MJ
Run 2best2bin93.88%0.00060820149663338530.01961 d 10:21:5608:35:2912.41 MJ
Run 3best2bin96.88%0.00077308462297344560.019681 d 06:32:5207:38:1310.71 MJ
Run 4best2bin99.024%0.00084320998785448760.0209622:23:5205:35:587.84 MJ
Run 5best2bin92.024%0.00048945331834283030.020961 d 10:07:4808:31:5712.03 MJ
Run 6best2bin95.936%0.00069809722320173390.0209621:42:1605:25:347.68 MJ
Run 7best2bin97.44%0.00066540300794917370.020961 d 01:37:4806:24:279.08 MJ
Run 8best2bin97.008%0.00087103601246564220.020961 d 00:09:1606:02:198.28 MJ
Run 9best2bin96.92%0.00066540300794917370.0192823:56:0805:59:028.24 MJ
Run 10best2bin97.76%0.00069758809851055720.020961 d 00:20:3606:05:098.74 MJ
Run 1rand1exp97.584%0.00062041543968458620.020961 d 12:53:2009:13:2013.00 MJ
Run 2rand1exp97.904%0.00079010804734894020.0209623:29:4005:52:258.39 MJ
Run 3rand1exp96.616%0.00079010804734894020.01961 d 03:11:2406:47:519.43 MJ
Run 4rand1exp96.984%0.00071864277865022960.020961 d 03:09:4406:47:269.52 MJ
Run 5rand1exp93.584%0.00066540300794917370.019681 d 04:23:1207:05:4810.24 MJ
Run 6rand1exp97.376%0.00058431083616292040.0196822:00:0405:30:017.74 MJ
Run 7rand1exp97.616%0.00059522302225026350.019681 d 01:26:5606:21:448.82 MJ
Run 8rand1exp96.904%0.00055037828142615850.01961 d 08:47:1608:11:4911.58 MJ
Run 9rand1exp97.16%0.00060820149663338530.019681 d 05:34:4807:23:4210.62 MJ
Run 10rand1exp95.544%0.00066540300794917370.0196823:23:3605:50:548.18 MJ
Run 1rand2exp90.96%0.00071071412129867450.01961 d 03:47:1606:56:499.98 MJ
Run 2rand2exp95.184%0.00079010804734894020.019621:52:4405:28:118.15 MJ
Run 3rand2exp94.6%0.00086789286879177020.0209623:46:2005:56:358.43 MJ
Run 4rand2exp96.08%0.00048945331834283030.020881 d 02:51:3606:42:549.64 MJ
Run 5rand2exp94.264%0.00079010804734894020.01961 d 01:25:0806:21:178.80 MJ
Run 6rand2exp97.808%0.00069810236024178490.0196822:51:0805:42:478.02 MJ
Run 7rand2exp95.88%0.00063964232534174370.020961 d 12:57:4409:14:2612.77 MJ
Run 8rand2exp94.184%0.00066540300794917370.020961 d 06:18:2807:34:3710.81 MJ
Run 9rand2exp92.704%0.00060159375090086420.020961 d 00:18:1206:04:339.11 MJ
Run 10rand2exp95.224%0.00075254780224504780.020961 d 09:31:1208:22:4811.82 MJ
Run 1best1exp97.856%0.00068546282079364920.0209620:46:1605:11:347.74 MJ
Run 2best1exp96.96%0.00079275041062573270.020961 d 01:59:3606:29:549.21 MJ
Run 3best1exp97.592%0.00071898224634934250.020961 d 00:07:3606:01:548.46 MJ
Run 4best1exp97.6%0.00088644469653704340.0192820:34:4405:08:417.31 MJ
Run 5best1exp98.088%0.00055037828142615850.0209622:47:1205:41:488.20 MJ
Run 6best1exp95.664%0.00066540300794917370.0209623:48:1605:57:048.81 MJ
Run 7best1exp96.312%0.00071898224634934250.0209621:01:4805:15:277.23 MJ
Run 8best1exp96.536%0.00045929087419095750.0209620:25:4005:06:257.11 MJ
Run 9best1exp97.88%0.00086789286879177020.0209619:37:2404:54:216.93 MJ
Run 10best1exp97.952%0.00058175251026142340.0196822:56:4805:44:128.34 MJ
Run 1best2exp79.408%0.00079010804734894020.020961 d 10:02:4408:30:4111.73 MJ
Run 2best2exp96.832%0.00069993468456865930.01961 d 16:31:2410:07:5114.08 MJ
Run 3best2exp93.768%0.00079010804734894020.0209619:07:1604:46:496.67 MJ
Run 4best2exp97.744%0.00057200375201822160.01961 d 08:50:2808:12:3711.38 MJ
Run 5best2exp92.44%0.00074105918425435070.019681 d 01:40:0806:25:028.87 MJ
Run 6best2exp98.048%0.00079010804734894020.0196823:16:5205:49:138.29 MJ
Run 7best2exp95.672%0.00086789286879177020.020961 d 02:20:3206:35:089.34 MJ
Run 8best2exp95.016%0.00079010804734894020.0209618:33:2804:38:226.65 MJ
Run 9best2exp98.256%0.0006094503161481170.019681 d 05:41:4807:25:2710.45 MJ
Run 10best2exp94.192%0.00057480752182023150.0196823:10:2405:47:368.57 MJ
Run 1currenttobest1exp98.488%0.00079010804734894020.019681 d 00:31:1606:07:498.57 MJ
Run 2currenttobest1exp86.744%0.00065191915890972660.019681 d 01:54:4406:28:419.06 MJ
Run 3currenttobest1exp98.408%0.0007475747864235650.0209620:26:0005:06:307.22 MJ
Run 4currenttobest1exp97.512%0.00066540300794917370.01961 d 04:42:4407:10:4110.71 MJ
Run 5currenttobest1exp97.88%0.00060820149663338530.0209622:45:5205:41:288.12 MJ
Run 6currenttobest1exp98.288%0.00056310576067857510.019620:57:4405:14:267.52 MJ
Run 7currenttobest1exp95.728%0.00064119742490651240.01961 d 06:36:2007:39:0510.78 MJ
Run 8currenttobest1exp97.032%0.00079010804734894020.020961 d 01:05:2406:16:218.77 MJ
Run 9currenttobest1exp98.24%0.00060820149663338530.0209623:59:0005:59:458.62 MJ
Run 10currenttobest1exp96.096%0.00063797250033348350.020961 d 00:45:4006:11:258.66 MJ
Table A8. Results obtained from 10 runs of 15 epochs of a hyperoptimized DenseNet121 using a different DE strategy on CIFAR100.
Table A8. Results obtained from 10 runs of 15 epochs of a hyperoptimized DenseNet121 using a different DE strategy on CIFAR100.
RunStrategyMax Accuracy (%)Best LRBest AccuracyCPU TimeElapsed TimeConsumed Energy
Run 1rand1bin82.92%0.00066540300794917370.020563 d 15:07:3621:46:5429.99 MJ
Run 2rand1bin82.888%0.00062020476681257930.020644 d 22:05:441 d 05:31:2640.69 MJ
Run 3rand1bin76.512%0.000485645055180374640.020644 d 10:00:441 d 02:30:1136.86 MJ
Run 4rand1bin78.048%0.00080538568249625360.020484 d 02:18:401 d 00:34:4033.63 MJ
Run 5rand1bin79.576%0.0006759756887715680.020163 d 17:07:4822:16:5731.05 MJ
Run 6rand1bin78.408%0.00087673645408905010.02044 d 15:57:361 d 03:59:2441.03 MJ
Run 7rand1bin76.92%0.000605402864479460.020483 d 17:13:4022:18:2529.91 MJ
Run 8rand1bin82.168%0.00087212691909229680.020484 d 14:04:201 d 03:31:0537.22 MJ
Run 9rand1bin81.704%0.0008297371849520230.020483 d 03:12:2418:48:0626.43 MJ
Run 10rand1bin84.112%0.00054763798695194230.020323 d 05:33:0419:23:1626.64 MJ
Run 1best1bin76.288%0.00074235307420013290.02043 d 18:22:1222:35:3332.99 MJ
Run 2best1bin76.696%0.00053269483800795420.020483 d 22:52:1223:43:0334.44 MJ
Run 3best1bin78.616%0.00057976792282186480.020083 d 15:27:4021:51:5529.34 MJ
Run 4best1bin77.696%0.00062008233534556830.020482 d 22:40:4817:40:1224.30 MJ
Run 5best1bin75.056%0.00075642070942000430.020644 d 06:46:441 d 01:41:4135.42 MJ
Run 6best1bin77.44%0.00039974204276031870.01884 d 14:41:481 d 03:40:2738.67 MJ
Run 7best1bin79.648%0.00048945331834283030.020243 d 02:49:0018:42:1526.21 MJ
Run 8best1bin78.52%0.00061206821525814840.018964 d 16:57:121 d 04:14:1839.48 MJ
Run 9best1bin77.112%0.00079010804734894020.018643 d 07:10:3219:47:3827.21 MJ
Run 10best1bin78.352%0.00055495159332900770.020564 d 10:42:161 d 02:40:3436.58 MJ
Run 1currenttobest1bin76.92%0.00071322992487661750.020083 d 22:39:5223:39:5831.86 MJ
Run 2currenttobest1bin78.784%0.00063627832765638350.02043 d 21:18:4823:19:4232.47 MJ
Run 3currenttobest1bin78.712%0.0006542132084908520.018723 d 20:55:2823:13:5231.95 MJ
Run 4currenttobest1bin76.224%0.0007872889918670340.018963 d 17:49:1222:27:1830.72 MJ
Run 5currenttobest1bin81.304%0.00079808968823810980.020484 d 03:24:001 d 00:51:0034.48 MJ
Run 6currenttobest1bin80.384%0.00057534029050436970.018483 d 17:36:2822:24:0732.39 MJ
Run 7currenttobest1bin81%0.00050479137239431860.020484 d 05:25:161 d 01:21:1935.28 MJ
Run 8currenttobest1bin77.168%0.00059596352650442360.020644 d 15:42:161 d 03:55:3438.33 MJ
Run 9currenttobest1bin81.336%0.000314255571715287030.018483 d 18:30:4022:37:4030.91 MJ
Run 10currenttobest1bin75.976%0.0005035028039211230.020484 d 13:53:361 d 03:28:2437.46 MJ
Run 1rand2bin73.424%0.00079010804734894020.020163 d 23:45:5623:56:2933.48 MJ
Run 2rand2bin78.408%0.00080303311650768080.020644 d 01:54:081 d 00:28:3233.22 MJ
Run 3rand2bin83.08%0.00053457893710137220.018084 d 07:17:321 d 01:49:2335.99 MJ
Run 4rand2bin78.44%0.000484188930376752340.020484 d 09:39:201 d 02:24:5036.11 MJ
Run 5rand2bin86.144%0.00056558211822817340.020484 d 07:49:481 d 01:57:2734.62 MJ
Run 6rand2bin77.312%0.00081162892042236940.02044 d 00:45:561 d 00:11:2932.78 MJ
Run 7rand2bin78.096%0.00047788878501502830.018564 d 19:37:041 d 04:54:1638.72 MJ
Run 8rand2bin76.072%0.00075191319683944620.018322 d 22:25:2017:36:2023.72 MJ
Run 9rand2bin80.48%0.00048945331834283030.020483 d 18:42:2022:40:3532.70 MJ
Run 10rand2bin80.96%0.00076457437800185190.020644 d 18:18:561 d 04:34:4439.98 MJ
Run 1best2bin77.208%0.00048945331834283030.017843 d 02:05:1218:31:1826.87 MJ
Run 2best2bin78.64%0.00083156360585558970.020243 d 15:54:2421:58:3631.75 MJ
Run 3best2bin80.008%0.00054273905441604210.020163 d 20:26:3223:06:3831.05 MJ
Run 4best2bin75.048%0.00086789286879177020.018723 d 19:25:4422:51:2631.88 MJ
Run 5best2bin80.344%0.00075813944822731970.02084 d 11:40:161 d 02:55:0436.48 MJ
Run 6best2bin82.904%0.00084379116463186870.020564 d 06:43:521 d 01:40:5835.18 MJ
Run 7best2bin73.2%0.00075796778313877390.01843 d 12:22:5621:05:4429.16 MJ
Run 8best2bin80.864%0.00077614786174076930.018323 d 05:22:5619:20:4427.09 MJ
Run 9best2bin74.16%0.00064623408788525190.018243 d 02:33:2018:38:2026.26 MJ
Run 10best2bin74.952%0.000455012615093883240.018964 d 16:01:081 d 04:00:1738.56 MJ
Run 1rand1exp74.304%0.000524744839551060.018163 d 09:37:4420:24:2628.64 MJ
Run 2rand1exp79.888%0.00072426963283277670.018323 d 08:40:5620:10:1427.85 MJ
Run 3rand1exp77.248%0.00076315363521346150.018324 d 04:16:121 d 01:04:0333.67 MJ
Run 4rand1exp78.12%0.000434771456989763770.017843 d 16:44:1222:11:0330.92 MJ
Run 5rand1exp79.208%0.00029076980693108320.018484 d 03:13:561 d 00:48:2933.39 MJ
Run 6rand1exp80.296%0.00049941667664874480.020322 d 21:39:1217:24:4824.08 MJ
Run 7rand1exp74.568%0.000474675422146660.020644 d 09:19:481 d 02:19:5736.24 MJ
Run 8rand1exp78.84%0.00079010804734894020.020484 d 15:27:241 d 03:51:5138.04 MJ
Run 9rand1exp80.848%0.00042807830876629820.020564 d 04:37:521 d 01:09:2834.79 MJ
Run 10rand1exp80.992%0.00026101881302939070.01843 d 18:43:0422:40:4632.89 MJ
Run 1rand2exp80.432%0.00064105073296388930.020563 d 21:45:2423:26:2132.88 MJ
Run 2rand2exp77.736%0.00086789286879177020.02043 d 15:47:5221:56:5829.60 MJ
Run 3rand2exp80.592%0.0005688410618395670.018482 d 04:15:1613:03:4917.99 MJ
Run 4rand2exp82.136%0.00055461371586765770.020484 d 04:48:241 d 01:12:0634.47 MJ
Run 5rand2exp80.24%0.00076228682588160350.020645 d 00:22:081 d 06:05:3241.39 MJ
Run 6rand2exp81.224%0.00067957956620932530.020564 d 04:52:241 d 01:13:0634.98 MJ
Run 7rand2exp76.784%0.00087745377747368140.020724 d 14:45:201 d 03:41:2038.24 MJ
Run 8rand2exp76.92%0.00064458845291245660.020323 d 22:29:4023:37:2532.78 MJ
Run 9rand2exp77.456%0.00079386743873228410.020564 d 00:46:081 d 00:11:3233.52 MJ
Run 10rand2exp78.968%0.00040112334087392890.018884 d 21:11:281 d 05:17:5239.54 MJ
Run 1best1exp80.536%0.00054326689180962890.020323 d 21:49:1223:27:1832.91 MJ
Run 2best1exp85.944%0.00055665078968390940.020323 d 19:58:2822:59:3732.15 MJ
Run 3best1exp78.376%0.00073946983232652380.020563 d 19:14:4822:48:4231.12 MJ
Run 4best1exp83.952%0.00072037542401543680.02043 d 20:47:0823:11:4732.29 MJ
Run 5best1exp82.152%0.00049357761470809550.020242 d 16:46:4016:11:4022.34 MJ
Run 6best1exp77.032%0.00049464618541069930.018644 d 20:30:041 d 05:07:3139.24 MJ
Run 7best1exp78.256%0.00086789286879177020.018643 d 18:09:0822:32:1731.15 MJ
Run 8best1exp76.568%0.00050386764270450840.020643 d 21:53:4023:28:2531.41 MJ
Run 9best1exp81.928%0.00076204438746037590.01884 d 21:55:481 d 05:28:5740.68 MJ
Run 10best1exp78.616%0.00084172985033400270.020484 d 10:01:001 d 02:30:1536.33 MJ
Run 1best2exp78.04%0.00057308288774453760.018563 d 19:48:5222:57:1332.05 MJ
Run 2best2exp75.856%0.00075661472837156020.018643 d 09:51:1220:27:4827.85 MJ
Run 3best2exp82.192%0.00072625877758004640.020564 d 09:36:441 d 02:24:1136.30 MJ
Run 4best2exp81.816%0.00066540300794917370.020643 d 17:06:3622:16:3931.09 MJ
Run 5best2exp78.92%0.00058061099849956960.020724 d 07:19:441 d 01:49:5636.15 MJ
Run 6best2exp77.376%0.00071302710551429350.02084 d 22:08:161 d 05:32:0440.67 MJ
Run 7best2exp78.928%0.00077299981303713890.01884 d 06:41:321 d 01:40:2334.85 MJ
Run 8best2exp80.176%0.000400535402137857670.018563 d 19:25:3222:51:2330.82 MJ
Run 9best2exp74.688%0.00055037828142615850.020324 d 07:27:081 d 01:51:4737.24 MJ
Run 10best2exp76.312%0.00053553739028156650.020484 d 07:18:361 d 01:49:3937.86 MJ
Run 1currenttobest1exp80.632%0.00026727959265821790.018482 d 19:45:2816:56:2224.09 MJ
Run 2currenttobest1exp81.752%0.00055045657000630330.020243 d 13:43:3621:25:5429.50 MJ
Run 3currenttobest1exp79.408%0.00089644236152839140.020483 d 21:17:4023:19:2532.76 MJ
Run 4currenttobest1exp81.68%0.00049319553390841030.020563 d 13:29:2421:22:2130.02 MJ
Run 5currenttobest1exp78.368%0.00049973568168069060.018323 d 23:39:0823:54:4733.75 MJ
Run 6currenttobest1exp78.448%0.00061503469121975980.01884 d 02:30:561 d 00:37:4433.19 MJ
Run 7currenttobest1exp77.392%0.000391632351492734540.018724 d 03:53:241 d 00:58:2133.63 MJ
Run 8currenttobest1exp75.92%0.00048945331834283030.020243 d 08:05:0020:01:1528.98 MJ
Run 9currenttobest1exp76.944%0.00075138912518938410.020244 d 02:20:201 d 00:35:0533.09 MJ
Run 10currenttobest1exp77.424%0.00054452218314174150.020324 d 01:16:281 d 00:19:0733.69 MJ

References

  1. Silva, C.; Vilaça, R.; Pereira, A.; Bessa, R. A review on the decarbonization of high-performance computing centers. Renew. Sustain. Energy Rev. 2024, 189, 114019. [Google Scholar] [CrossRef]
  2. Chu, X.; Hofstätter, D.; Ilager, S.; Talluri, S.; Kampert, D.; Podareanu, D.; Duplyakin, D.; Brandic, I.; Iosup, A. Generic and ML Workloads in an HPC Datacenter: Node Energy, Job Failures, and Node-Job Analysis. arXiv 2024, arXiv:2409.08949. [Google Scholar] [CrossRef]
  3. Baratchi, M.; Wang, C.; Limmer, S.; van Rijn, J.N.; Hoos, H.; Bäck, T.; Olhofer, M. Automated machine learning: Past, present and future. Artif. Intell. Rev. 2024, 57, 122. [Google Scholar] [CrossRef]
  4. Bolón-Canedo, V.; Morán-Fernández, L.; Cancela, B.; Alonso-Betanzos, A. A review of green artificial intelligence: Towards a more sustainable future. Neurocomputing 2024, 599, 128096. [Google Scholar] [CrossRef]
  5. TOP500 Methodology. 2025. Available online: https://www.top500.org/static/media/uploads/methodology-2.0rc1.pdf (accessed on 19 January 2025).
  6. Miller, J.; Trümper, L.; Terboven, C.; Müller, M.S. A Theoretical Model for Global Optimization of Parallel Algorithms. Mathematics 2021, 9, 1685. [Google Scholar] [CrossRef]
  7. Damme, P.; Birkenbach, M.; Bitsakos, C.; Boehm, M.; Bonnet, P.; Ciorba, F.; Dokter, M.; Dowgiallo, P.; Eleliemy, A.; Faerber, C.; et al. DAPHNE: An Open and Extensible System Infrastructure for Integrated Data Analysis Pipelines. In Proceedings of the Conference on Innovative Data Systems Research, Chaminade, CA, USA, 9–12 January 2022. [Google Scholar]
  8. Alangari, N.; El Bachir Menai, M.; Mathkour, H.; Almosallam, I. Exploring Evaluation Methods for Interpretable Machine Learning: A Survey. Information 2023, 14, 469. [Google Scholar] [CrossRef]
  9. Jakobsche, T.; Lachiche, N.; Ciorba, F.M. Investigating HPC Job Resource Requests and Job Efficiency Reporting. In Proceedings of the 2023 22nd International Symposium on Parallel and Distributed Computing (ISPDC), Bucharest, Romania, 10–12 July 2023; pp. 61–68. [Google Scholar] [CrossRef]
  10. Yarally, T.; Cruz, L.; Feitosa, D.; Sallou, J.; van Deursen, A. Uncovering Energy-Efficient Practices in Deep Learning Training: Preliminary Steps Towards Green AI. In Proceedings of the 2023 IEEE/ACM 2nd International Conference on AI Engineering—Software Engineering for AI (CAIN), Melbourne, Australia, 15–16 May 2023; pp. 25–36. [Google Scholar] [CrossRef]
  11. Prica, T. Development and supporting activities on EuroHPC Vega. In Proceedings of the Austrian-Slovenian HPC Meeting 2024—ASHPC24, Grundlsee, Austria, 10–13 June 2024; p. 14. [Google Scholar]
  12. Oliveira, S.d.; Topsakal, O.; Toker, O. Benchmarking Automated Machine Learning (AutoML) Frameworks for Object Detection. Information 2024, 15, 63. [Google Scholar] [CrossRef]
  13. Yoo, A.B.; Jette, M.A.; Grondona, M. SLURM: Simple Linux Utility for Resource Management. In Proceedings of the 9th International Workshop on Job Scheduling Strategies for Parallel Processing (JSSPP), Seattle, WA, USA, 24 June 2003. [Google Scholar] [CrossRef]
  14. Han, M.; Wu, H.; Chen, Z.; Li, M.; Zhang, X. A survey of multi-label classification based on supervised and semi-supervised learning. Int. J. Mach. Learn. Cybern. 2023, 14, 697–724. [Google Scholar] [CrossRef]
  15. Virtanen, P.; Gommers, R.; Oliphant, T.E.; Haberland, M.; Reddy, T.; Cournapeau, D.; Burovski, E.; Peterson, P.; Weckesser, W.; Bright, J.; et al. SciPy 1.0: Fundamental algorithms for scientific computing in Python. Nat. Methods 2020, 17, 261–272. [Google Scholar] [CrossRef]
  16. Das, S.; Mullick, S.S.; Suganthan, P.N. Recent Advances in Differential Evolution—An Updated Survey. Swarm Evol. Comput. 2016, 27, 1–30. [Google Scholar] [CrossRef]
  17. Li, J.Y.; Zhan, Z.H.; Zhang, J. Evolutionary Computation for Expensive Optimization: A Survey. Mach. Intell. Res. 2022, 19, 3–23. [Google Scholar] [CrossRef]
  18. Qin, X.; Luo, Y.; Chen, S.; Chen, Y.; Han, Y. Investigation of Energy-Saving Strategy for Parallel Variable Frequency Pump System Based on Improved Differential Evolution Algorithm. Energies 2022, 15, 5360. [Google Scholar] [CrossRef]
  19. Dragoi, E.N.; Dafinescu, V. Parameter control and hybridization techniques in differential evolution: A survey. Artif. Intell. Rev. 2015, 45, 447–470. [Google Scholar] [CrossRef]
  20. Storn, R.; Price, K. Differential Evolution–A Simple and Efficient Heuristic for global Optimization over Continuous Spaces. J. Glob. Optim. 1997, 11, 341–359. [Google Scholar] [CrossRef]
  21. Eltaeib, T.; Mahmood, A. Differential Evolution: A Survey and Analysis. Appl. Sci. 2018, 8, 1945. [Google Scholar] [CrossRef]
  22. Nanthapodej, R.; Liu, C.H.; Nitisiri, K.; Pattanapairoj, S. Hybrid Differential Evolution Algorithm and Adaptive Large Neighborhood Search to Solve Parallel Machine Scheduling to Minimize Energy Consumption in Consideration of Machine-Load Balance Problems. Sustainability 2021, 13, 5470. [Google Scholar] [CrossRef]
  23. Chhabra, A.; Sahana, S.K.; Sani, N.S.; Mohammadzadeh, A.; Omar, H.A. Energy-Aware Bag-of-Tasks Scheduling in the Cloud Computing System Using Hybrid Oppositional Differential Evolution-Enabled Whale Optimization Algorithm. Energies 2022, 15, 4571. [Google Scholar] [CrossRef]
  24. IAM Working Group. IPMI Specification. 2006. Available online: https://openipmi.sourceforge.io/IPMI.pdf (accessed on 25 October 2024).
  25. Bischl, B.; Binder, M.; Lang, M.; Pielok, T.; Richter, J.; Coors, S.; Thomas, J.; Ullmann, T.; Becker, M.; Boulesteix, A.L.; et al. Hyperparameter optimization: Foundations, algorithms, best practices, and open challenges. WIREs Data Min. Knowl. Discov. 2023, 13, e1484. [Google Scholar] [CrossRef]
  26. Morales-Hernández, A.; Van Nieuwenhuyse, I.; Rojas Gonzalez, S. A survey on multi-objective hyperparameter optimization algorithms for machine learning. Artif. Intell. Rev. 2022, 56, 8043–8093. [Google Scholar] [CrossRef]
  27. Dataset CIFAR10 and CIFAR100. Available online: https://www.cs.toronto.edu/~kriz/cifar.html (accessed on 19 November 2024).
  28. Boito, F.; Brandt, J.; Cardellini, V.; Carns, P.; Ciorba, F.M.; Egan, H. Autonomy Loops for Monitoring, Operational Data Analytics, Feedback, and Response in HPC Operations. In Proceedings of the 2023 IEEE International Conference on Cluster Computing Workshops (CLUSTER Workshops), Santa Fe, NM, USA, 31 October 2023; pp. 37–43. [Google Scholar] [CrossRef]
  29. El Naqa, I.; Murphy, M.J. What Is Machine Learning? In Machine Learning in Radiation Oncology: Theory and Applications; Springer International Publishing: Cham, Switzerland, 2015; pp. 3–11. [Google Scholar] [CrossRef]
  30. Zhang, X.; Guo, F.; Chen, T.; Pan, L.; Beliakov, G.; Wu, J. A Brief Survey of Machine Learning and Deep Learning Techniques for E-Commerce Research. J. Theor. Appl. Electron. Commer. Res. 2023, 18, 2188–2216. [Google Scholar] [CrossRef]
  31. Bergstra, J.; Bardenet, R.; Bengio, Y.; Kégl, B. Algorithms for hyper-parameter optimization. In Proceedings of the 25th International Conference on Neural Information Processing Systems, NIPS’11, Granada, Spain, 12–15 December 2011; Curran Associates Inc.: Red Hook, NY, USA, 2011; pp. 2546–2554. Available online: https://dl.acm.org/doi/10.5555/2986459.2986743 (accessed on 7 April 2025).
  32. Zamuda, A.; Daniel Hernández Sosa, J.; Adler, L. Improving constrained glider trajectories for ocean eddy border sampling within extended mission planning time. In Proceedings of the 2016 IEEE Congress on Evolutionary Computation (CEC), Vancouver, BC, Canada, 24–29 July 2016; pp. 1727–1734. [Google Scholar] [CrossRef]
  33. Zhu, K.; Wu, J. Residual attention: A simple but effective method for multi-label recognition. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 184–193. [Google Scholar] [CrossRef]
  34. Zhou, Z.H. Machine Learning; Springer Nature: Berlin/Heidelberg, Germany, 2021. [Google Scholar]
  35. Hu, X.; Chu, L.; Pei, J.; Liu, W.; Bian, J. Model Complexity of Deep Learning: A Survey. arXiv 2021, arXiv:2103.05127. [Google Scholar] [CrossRef]
  36. Radford, A.; Narasimhan, K.; Salimans, T.; Sutskever, I. Improving Language Understanding by Generative Pre-Training; OpenAI: San Francisco, CA, USA, 2018. [Google Scholar]
  37. Menik, S.; Ramaswamy, L. Towards Modular Machine Learning Solution Development: Benefits and Trade-offs. arXiv 2023, arXiv:2301.09753. [Google Scholar] [CrossRef]
  38. Shen, Y.; Zhang, Z.; Cao, T.; Tan, S.; Chen, Z.; Gan, C. ModuleFormer: Modularity Emerges from Mixture-of-Experts. arXiv 2023, arXiv:2306.04640. [Google Scholar] [CrossRef]
  39. Barandas, M.; Famiglini, L.; Campagner, A.; Folgado, D.; Simão, R.; Cabitza, F.; Gamboa, H. Evaluation of uncertainty quantification methods in multi-label classification: A case study with automatic diagnosis of electrocardiogram. Inf. Fusion 2024, 101, 101978. [Google Scholar] [CrossRef]
  40. Maloney, S.; Suarez, E.; Eicker, N.; Guimarães, F.; Frings, W. Analyzing HPC Monitoring Data With a View Towards Efficient Resource Utilization. In Proceedings of the 2024 IEEE 36th International Symposium on Computer Architecture and High Performance Computing (SBAC-PAD), Hilo, HI, USA, 13–15 November 2024; pp. 170–181. [Google Scholar]
  41. Vontzalidis, A.; Psomadakis, S.; Bitsakos, C.; Dokter, M.; Innerebner, K.; Damme, P.; Boehm, M.; Ciorba, F.; Eleliemy, A.; Karakostas, V.; et al. DAPHNE Runtime: Harnessing Parallelism for‚ Integrated Data Analysis Pipelines. In Proceedings of the Euro-Par 2023: Parallel Processing Workshops, Limassol, Cyprus, 28 August–1 September 2023; pp. 242–246. [Google Scholar]
  42. Jakobsche, T.; Lachiche, N.; Ciorba, F.M. Challenges and Opportunities of Machine Learning for Monitoring and Operational Data Analytics in Quantitative Codesign of Supercomputers. arXiv 2022, arXiv:2209.07164. [Google Scholar] [CrossRef]
  43. Prica, T.; Zamuda, A. Monitoring Energy Consumption of Workloads on HPC Vega. In Proceedings of the 6th ISC HPC International Workshop on “Monitoring & Operational Data Analytics”, Hamburg, Germany, 9 March–13 June 2025. [Google Scholar]
  44. Chakraborty, U.K. Advances in Differential Evolution; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2008; Volume 143. [Google Scholar]
  45. Ahmad, M.F.; Isa, N.A.M.; Lim, W.H.; Ang, K.M. Differential evolution: A recent review based on state-of-the-art works. Alex. Eng. J. 2022, 61, 3831–3872. [Google Scholar] [CrossRef]
  46. Glotić, A.; Zamuda, A. Short-term combined economic and emission hydrothermal optimization by surrogate differential evolution. Appl. Energy 2015, 141, 42–56. [Google Scholar] [CrossRef]
  47. Zamuda, A.; Sosa, J.D.H.; Adler, L. Constrained differential evolution optimization for underwater glider path planning in sub-mesoscale eddy sampling. Appl. Soft Comput. 2016, 42, 93–118. [Google Scholar] [CrossRef]
  48. Lucas, C.; Hernández-Sosa, D.; Greiner, D.; Zamuda, A.; Caldeira, R. An approach to multi-objective path planning optimization for underwater gliders. Sensors 2019, 19, 5506. [Google Scholar] [CrossRef]
  49. Brest, J.; Greiner, S.; Boskovic, B.; Mernik, M.; Zumer, V. Self-Adapting Control Parameters in Differential Evolution: A Comparative Study on Numerical Benchmark Problems. IEEE Trans. Evol. Comput. 2006, 10, 646–657. [Google Scholar] [CrossRef]
  50. Fan, Q.; Yan, X.; Zhang, Y. Auto-selection mechanism of differential evolution algorithm variants and its application. Eur. J. Oper. Res. 2018, 270, 636–653. [Google Scholar] [CrossRef]
  51. Vincent, A.M.; Jidesh, P. An improved hyperparameter optimization framework for AutoML systems using evolutionary algorithms. Sci. Rep. 2023, 13, 4737. [Google Scholar] [CrossRef]
  52. Sen, A.; Gupta, V.; Tang, C. Differential Evolution Algorithm Based Hyperparameter Selection of Gated Recurrent Unit for Electrical Load Forecasting. arXiv 2023, arXiv:2309.13019. [Google Scholar] [CrossRef]
  53. Gomes, E.; Pereira, L.; Esteves, A.; Morais, H. Metaheuristic Optimization Methods in Energy Community Scheduling: A Benchmark Study. Energies 2024, 17, 2968. [Google Scholar] [CrossRef]
  54. Main Stages of the DE Algorithm. Available online: https://www.researchgate.net/figure/Main-stages-of-the-DE-algorithm_fig1_336225430 (accessed on 19 November 2024).
  55. Opara, K.; Arabas, J. Comparison of mutation strategies in Differential Evolution—A probabilistic perspective. Swarm Evol. Comput. 2018, 39, 53–69. [Google Scholar] [CrossRef]
  56. Wu, T.; Li, X.; Zhou, D.; Li, N.; Shi, J. Differential Evolution Based Layer-Wise Weight Pruning for Compressing Deep Neural Networks. Sensors 2021, 21, 880. [Google Scholar] [CrossRef]
  57. Zamuda, A. Foundational Concepts and Real-World Applications of Self-Adaptive Differential Evolution and Success History. In Swarm Intelligence—Foundational Concepts and Real-World Applications; Chibante, R., Miranda, P., Palade, V., Eds.; Artificial Intelligence; IntechOpen: London, UK, 2025; Chapter 1; pp. 1–20. Available online: https://www.intechopen.com/online-first/1222844 (accessed on 7 April 2025).
  58. Qiao, K.; Wen, X.; Ban, X.; Chen, P.; Price, K.; Suganthan, P.; Liang, J.; Wu, G.; Yue, C. Evaluation Criteria for CEC 2024 Competition and Special Session on Numerical Optimization Considering Accuarcy and Speed; Technical Report; Zhengzhou University: Zhengzhou, China; Central South University: Changsha, China; Henan Institute of Technology: Xinxiang, China; Qatar University: Doha, Qatar, 2023. [Google Scholar]
  59. Tanabe, R.; Fukunaga, A.S. Improving the search performance of SHADE using linear population size reduction. In Proceedings of the 2014 IEEE Congress on Evolutionary Computation, Beijing, China, 6–11 July 2014; pp. 1658–1665. [Google Scholar]
  60. Viktorin, A.; Senkerik, R.; Pluhacek, M.; Kadavy, T.; Zamuda, A. Distance based parameter adaptation for Success-History based Differential Evolution. Swarm Evol. Comput. 2019, 50, 100462. [Google Scholar] [CrossRef]
  61. Tanabe, R.; Fukunaga, A. Reviewing and Benchmarking Parameter Control Methods in Differential Evolution. IEEE Trans. Cybern. 2020, 50, 1170–1184. [Google Scholar] [CrossRef]
  62. Mininno, E.; Neri, F.; Cupertino, F.; Naso, D. Compact Differential Evolution. IEEE Trans. Evol. Comput. 2011, 15, 32–54. [Google Scholar] [CrossRef]
  63. Zamuda, A.; Dokter, M. Deploying DAPHNE Computational Intelligence on EuroHPC Vega for Benchmarking Randomised Optimisation Algorithms. In Proceedings of the International Conference on Broadband Communications for Next Generation Networks and Multimedia Applications (CoBCom), Graz, Austria, 9–11 July 2024; pp. 1–8. [Google Scholar]
  64. Jiang, Y.; Qi, X.; Liu, C. Energy-Aware Automatic Tuning on Many-Core Platform via Differential Evolution. In Proceedings of the 2016 45th International Conference on Parallel Processing Workshops (ICPPW), Philadelphia, PA, USA, 16–19 August 2016; pp. 258–265. [Google Scholar] [CrossRef]
  65. Baioletti, M.; Di Bari, G.; Milani, A.; Poggioni, V. Differential Evolution for Neural Networks Optimization. Mathematics 2020, 8, 69. [Google Scholar] [CrossRef]
  66. Agarwal, M.; Gupta, S.K.; Biswas, K.K. DECACNN: Differential evolution-based approach to compress and accelerate the convolution neural network model. Neural Comput. Appl. 2023, 36, 2665–2681. [Google Scholar] [CrossRef]
  67. Wu, X.; Che, A. A memetic differential evolution algorithm for energy-efficient parallel machine scheduling. Omega 2019, 82, 155–165. [Google Scholar] [CrossRef]
  68. Abdel-Basset, M.; Mohamed, R.; Alrashdi, I.; Sallam, K.M.; Hameed, I.A. Evolution-based energy-efficient data collection system for UAV-supported IoT: Differential evolution with population size optimization mechanism. Expert Syst. Appl. 2024, 245, 123082. [Google Scholar] [CrossRef]
  69. Zamuda, A.; Hernández Sosa, J.D. Differential evolution and underwater glider path planning applied to the short-term opportunistic sampling of dynamic mesoscale ocean structures. Appl. Soft Comput. 2014, 24, 95–108. [Google Scholar] [CrossRef]
  70. Zamuda, A.; Sosa, J.D.H. Success history applied to expert system for underwater glider path planning using differential evolution. Expert Syst. Appl. 2019, 119, 155–170. [Google Scholar] [CrossRef]
  71. Janssen, D.; Pullan, W.; Liew, A.W.C. GPU Based Differential Evolution: New Insights and Comparative Study. arXiv 2024, arXiv:2405.16551. [Google Scholar] [CrossRef]
  72. Van Stein, B.; Vermetten, D.; Caraffini, F.; Kononova, A.V. Deep BIAS: Detecting Structural Bias using Explainable AI. In Proceedings of the GECCO ’23 Companion: Proceedings of the Companion Conference on Genetic and Evolutionary Computation, Lisbon, Portugal, 15–19 July 2023; pp. 455–458. [Google Scholar]
  73. van Stein, N.; Kononova, A.V. (Eds.) Explainable AI for Evolutionary Computation; Springer Nature: Singapore, 2025. [Google Scholar] [CrossRef]
  74. Raponi, E.; Rodriguez, I.O.; van Stein, N. Global Sensitivity Analysis Is Not Always Beneficial for Evolutionary Computation: A Study in Engineering Design. In Explainable AI for Evolutionary Computation; Springer Nature: Singapore, 2025; pp. 13–40. [Google Scholar] [CrossRef]
  75. Barbudo, R.; Ventura, S.; Romero, J.R. Eight years of AutoML: Categorisation, review and trends. Knowl. Inf. Syst. 2023, 65, 5097–5149. [Google Scholar] [CrossRef]
  76. Salehin, I.; Islam, M.S.; Saha, P.; Noman, S.; Tuni, A.; Hasan, M.M.; Baten, M.A. AutoML: A systematic review on automated machine learning with neural architecture search. J. Inf. Intell. 2024, 2, 52–81. [Google Scholar] [CrossRef]
  77. Chatzilygeroudis, K.; Hatzilygeroudis, I.; Perikos, I. Machine Learning Basics. In Intelligent Computing for Interactive System Design: Statistics, Digital Signal Processing, and Machine Learning in Practice, 1st ed.; Association for Computing Machinery: New York, NY, USA, 2021; pp. 143–193. [Google Scholar] [CrossRef]
  78. Geissler, D.; Zhou, B.; Suh, S.; Lukowicz, P. Spend More to Save More (SM2): An Energy-Aware Implementation of Successive Halving for Sustainable Hyperparameter Optimization. arXiv 2024, arXiv:2412.08526. [Google Scholar] [CrossRef]
  79. Ferro, M.; Silva, G.D.; de Paula, F.B.; Vieira, V.; Schulze, B. Towards a sustainable artificial intelligence: A case study of energy efficiency in decision tree algorithms. Concurr. Comput. Pract. Exp. 2021, 35, e6815. [Google Scholar] [CrossRef]
  80. Castellanos-Nieves, D.; García-Forte, L. Improving Automated Machine-Learning Systems Through Green AI. Appl. Sci. 2023, 13, 11583. [Google Scholar] [CrossRef]
  81. Castellanos-Nieves, D.; García-Forte, L. Strategies of Automated Machine Learning for Energy Sustainability in Green Artificial Intelligence. Appl. Sci. 2024, 14, 6196. [Google Scholar] [CrossRef]
  82. Zamuda, A.; Brest, J. Self-adaptive control parameters’ randomization frequency and propagations in differential evolution. Swarm Evol. Comput. 2015, 25, 72–99. [Google Scholar] [CrossRef]
  83. Vakhnin, A.; Ryzhikov, I.; Niska, H.; Kolehmainen, M. A Novel Multi-Objective Hybrid Evolutionary-Based Approach for Tuning Machine Learning Models in Short-Term Power Consumption Forecasting. AI 2024, 5, 2461–2496. [Google Scholar] [CrossRef]
  84. Pătrăușanu, A.; Florea, A.; Neghină, M.; Dicoiu, A.; Chiș, R. A Systematic Review of Multi-Objective Evolutionary Algorithms Optimization Frameworks. Processes 2024, 12, 869. [Google Scholar] [CrossRef]
  85. Liuliakov, A.; Hermes, L.; Hammer, B. AutoML technologies for the identification of sparse classification and outlier detection models. Appl. Soft Comput. 2023, 133, 109942. [Google Scholar] [CrossRef]
  86. Jin, H.; Chollet, F.; Song, Q.; Hu, X. AutoKeras: An AutoML Library for Deep Learning. J. Mach. Learn. Res. 2023, 24, 1–6. [Google Scholar]
  87. Shi, M.; Shen, W. Automatic Modeling for Concrete Compressive Strength Prediction Using Auto-Sklearn. Buildings 2022, 12, 1406. [Google Scholar] [CrossRef]
  88. Omar, I.; Khan, M.; Starr, A.; Abou Rok Ba, K. Automated Prediction of Crack Propagation Using H2O AutoML. Sensors 2023, 23, 8419. [Google Scholar] [CrossRef]
  89. Olson, R.S.; Moore, J.H. TPOT: A Tree-based Pipeline Optimization Tool for Automating Machine Learning. In PMLR, Proceedings of the Workshop on Automatic Machine Learning, New York, NY, USA, 24 June 2016; Hutter, F., Kotthoff, L., Vanschoren, J., Eds.; Springer: Cham, Switzerland, 2016; Volume 64, pp. 66–74. [Google Scholar]
  90. TensorFlow. Available online: https://www.tensorflow.org/ (accessed on 28 November 2024).
  91. PyTorch. Available online: https://pytorch.org/ (accessed on 19 November 2024).
  92. Hansen, N.; Auger, A.; Ros, R.; Mersmann, O.; Tušar, T.; Brockhoff, D. COCO: A platform for comparing continuous optimizers in a black-box setting. Optim. Methods Softw. 2020, 36, 114–144. [Google Scholar] [CrossRef]
  93. Varelas, K.; El Hara, O.A.; Brockhoff, D.; Hansen, N.; Nguyen, D.M.; Tušar, T.; Auger, A. Benchmarking large-scale continuous optimizers: The bbob-largescale testbed, a COCO software guide and beyond. Appl. Soft Comput. 2020, 97, 106737. [Google Scholar] [CrossRef]
  94. Doerr, C.; Wang, H.; Ye, F.; van Rijn, S.; Bäck, T. IOHprofiler: A Benchmarking and Profiling Tool for Iterative Optimization Heuristics. arXiv 2018, arXiv:1810.05281. [Google Scholar] [CrossRef]
  95. Doerr, C.; Ye, F.; Horesh, N.; Wang, H.; Shir, O.M.; Bäck, T. Benchmarking discrete optimization heuristics with IOHprofiler. Appl. Soft Comput. 2020, 88, 106027. [Google Scholar] [CrossRef]
  96. Durillo, J.J.; Nebro, A.J. jMetal: A Java framework for multi-objective optimization. Adv. Eng. Softw. 2011, 42, 760–771. [Google Scholar] [CrossRef]
  97. López-Ibáñez, M.; Dubois-Lacoste, J.; Pérez Cáceres, L.; Birattari, M.; Stützle, T. The irace package: Iterated racing for automatic algorithm configuration. Oper. Res. Perspect. 2016, 3, 43–58. [Google Scholar] [CrossRef]
  98. Akiba, T.; Sano, S.; Yanase, T.; Ohta, T.; Koyama, M. Optuna: A Next-generation Hyperparameter Optimization Framework. In Proceedings of the KDD ’19: 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, Anchorage, AK, USA, 4–8 August 2019; Association for Computing Machinery: New York, NY, USA, 2019; pp. 2623–2631. [Google Scholar] [CrossRef]
  99. Liaw, R.; Liang, E.; Nishihara, R.; Moritz, P.; Gonzalez, J.E.; Stoica, I. Tune: A Research Platform for Distributed Model Selection and Training. arXiv 2018, arXiv:1807.05118. [Google Scholar] [CrossRef]
  100. Spirals Research Group. PyJoules: A Python Library to Capture the Energy Consumption of Code Snippets; University of Lille and Inria: Lille, France, 2021. [Google Scholar]
  101. Anthony, L.F.W.; Kanding, B.; Selvan, R. Carbontracker: Tracking and Predicting the Carbon Footprint of Training Deep Learning Models. arXiv 2020, arXiv:2007.03051. [Google Scholar] [CrossRef]
  102. Ramduny, J.; Garcia, M.; Kelly, C. Establishing a reproducible and sustainable analysis workflow. In Methods for Analyzing Large Neuroimaging Datasets; Springer: New York, NY, USA, 2024; pp. 39–60. [Google Scholar]
  103. Deng, L. The mnist database of handwritten digit images for machine learning research. IEEE Signal Process. Mag. 2012, 29, 141–142. [Google Scholar] [CrossRef]
  104. Deng, J.; Dong, W.; Socher, R.; Li, L.J.; Li, K.; Fei-Fei, L. ImageNet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
  105. Egwutuoha, I.P.; Levy, D.; Selic, B.; Chen, S. A survey of fault tolerance mechanisms and checkpoint/restart implementations for high performance computing systems. J. Supercomput. 2013, 65, 1302–1326. [Google Scholar] [CrossRef]
  106. Moran, M.; Balladini, J.; Rexachs, D.; Luque, E. Checkpoint and Restart: An Energy Consumption Characterization in Clusters. arXiv 2024, arXiv:2409.02214. [Google Scholar] [CrossRef]
  107. Kumar, M.; Gupta, S.; Patel, T.; Wilder, M.; Shi, W.; Fu, S.; Engelmann, C.; Tiwari, D. Study of interconnect errors, network congestion, and applications characteristics for throttle prediction on a large scale HPC system. J. Parallel Distrib. Comput. 2021, 153, 29–43. [Google Scholar] [CrossRef]
  108. Jiao, Y.; Lin, H.; Balaji, P.; Feng, W. Power and Performance Characterization of Computational Kernels on the GPU. In Proceedings of the 2010 IEEE/ACM Int’l Conference on Green Computing and Communications & Int’l Conference on Cyber, Physical and Social Computing, Hangzhou, China, 18–20 December 2010; pp. 221–228. [Google Scholar]
  109. Timalsina, M.; Gerhardt, L.; Tyler, N.; Blaschke, J.P.; Arndt, W. Optimizing Checkpoint-Restart Mechanisms for HPC with DMTCP in Containers at NERSC. arXiv 2024, arXiv:2407.19117. [Google Scholar] [CrossRef]
  110. Assogba, K.; Nicolae, B.; Van Dam, H.; Rafique, M.M. Asynchronous Multi-Level Checkpointing: An Enabler of Reproducibility using Checkpoint History Analytics. In Proceedings of the SC-W ’23: SC ’23 Workshops of the International Conference on High Performance Computing, Network, Storage, and Analysis, Denver, CO, USA, 12–17 November 2023; Association for Computing Machinery: New York, NY, USA, 2023; pp. 1748–1756. [Google Scholar] [CrossRef]
  111. Rojas, E.; Kahira, A.N.; Meneses, E.; Gomez, L.B.; Badia, R.M. A Study of Checkpointing in Large Scale Training of Deep Neural Networks. arXiv 2021, arXiv:2012.00825. [Google Scholar] [CrossRef]
  112. Gu, R.; Chen, Y.; Liu, S.; Dai, H.; Chen, G.; Zhang, K.; Che, Y.; Huang, Y. Liquid: Intelligent Resource Estimation and Network-Efficient Scheduling for Deep Learning Jobs on Distributed GPU Clusters. IEEE Trans. Parallel Distrib. Syst. 2022, 33, 2808–2820. [Google Scholar] [CrossRef]
  113. Daradkeh, T.; Roper, G.; Alarcon Meza, C.; Mokhov, S.A. HPC Jobs Classification and Resource Prediction to Minimize Job Failures. In Proceedings of the CompSysTech ’24: International Conference on Computer Systems and Technologies 2024, Ruse, Bulgaria, 14–15 June 2024; ACM: New York, NY, USA, 2024; pp. 95–101. [Google Scholar] [CrossRef]
  114. Tanash, M.; Yang, H.; Andresen, D.; Hsu, W. Ensemble Prediction of Job Resources to Improve System Performance for Slurm-Based HPC Systems. In Proceedings of the PEARC ’21: Practice and Experience in Advanced Research Computing, Boston, MA, USA, 18–22 July 2021; ACM: New York, NY, USA, 2021; pp. 1–8. [Google Scholar] [CrossRef]
  115. Friedman, M. The Use of Ranks to Avoid the Assumption of Normality Implicit in the Analysis of Variance. J. Am. Stat. Assoc. 1937, 32, 675–701. [Google Scholar] [CrossRef]
  116. Soft Computing and Intelligent Information Systems. 2025. Available online: https://sci2s.ugr.es/sicidm (accessed on 14 January 2025).
  117. Demšar, J. Statistical Comparisons of Classifiers over Multiple Data Sets. J. Mach. Learn. Res. 2006, 7, 1–30. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.