Next Article in Journal
Enhanced ResNet50 for Diabetic Retinopathy Classification: External Attention and Modified Residual Branch
Previous Article in Journal
TextNeX: Text Network of eXperts for Robust Text Classification—Case Study on Machine-Generated-Text Detection
Previous Article in Special Issue
A Model for Learning-Curve Estimation in Efficient Neural Architecture Search and Its Application in Predictive Health Maintenance
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Comparative Study of Modern Differential Evolution Algorithms: Perspectives on Mechanisms and Performance

by
Janez Brest
* and
Mirjam Sepesy Maučec
Faculty of Electrical Engineering and Computer Science, University of Maribor, SI-2000 Maribor, Slovenia
*
Author to whom correspondence should be addressed.
Mathematics 2025, 13(10), 1556; https://doi.org/10.3390/math13101556
Submission received: 14 March 2025 / Revised: 5 May 2025 / Accepted: 6 May 2025 / Published: 9 May 2025

Abstract

:
Since the discovery of the Differential Evolution algorithm, new and improved versions have continuously emerged. In this paper, we review selected algorithms based on Differential Evolution that have been proposed in recent years. We examine the mechanisms integrated into them and compare the performance of algorithms. To compare their performances, statistical comparisons were used as they enable us to draw reliable conclusions about the algorithms’ performances. We use the Wilcoxon signed-rank test for pairwise comparisons and the Friedman test for multiple comparisons. Subsequently, the Mann–Whitney U-score test was added. We conducted not only a cumulative analysis of algorithms, but we also focused on their performances regarding the function family (i.e., unimodal, multimodal, hybrid, and composition functions). Experimental results of algorithms were obtained on problems defined for the CEC’24 Special Session and Competition on Single Objective Real Parameter Numerical Optimization. Problem dimensions of 10, 30, 50, and 100 were analyzed. In this paper, we highlight promising mechanisms for further development and improvements based on the study of the selected algorithms.

1. Introduction

Differential Evolution (DE) is a simple and effective evolutionary algorithm for solving global optimization problems in continuous space. It has also been adapted for use in discrete, most commonly binary, spaces. The original version was defined for single-objective optimization and was later extended to multi-objective optimization. This line of research, which goes in many directions, highlights the significance of this metaheuristic.
In this paper, we focus our research on DE for single-objective optimization in continuous space. It was introduced in 1996 by Storn and Price. Since then, numerous modifications and improvements have been proposed. This trend continues today and has been especially noticeable in recent years. In the variety of algorithm improvements, it is difficult to determine which one represents a truly significant enhancement. In the paper, we examine the improvements of the DE algorithm that have been published primarily in the past year of algorithm development.
Each year at the Congress on Evolutionary Computation (CEC), a competition for single-objective real parameter numerical optimization is held, in which DE-based algorithms hold a prominent place. In the year 2024, among the six competing algorithms, four were derived from DE. These four are of particular interest to us in this paper. Three algorithms from the recent past are also included in comparison to assess whether 2024 brings significant progress in development.
In the literature, algorithms are compared using different statistical techniques to perform single-problem and multiple-problem analysis. Non-parametric statistical tests are commonly used as they are less restrictive than parametric tests. In this paper, we adapted the same tests and added the Mann–Whitney U-score test, which was used to determine the winners in the most recent CEC 2024 competition. The computation of the U-score in our research slightly differs from that used in CEC 2024, as will be described later in the paper. The paper is organized as follows: Section 2 provides the necessary background knowledge to understand the main ideas of the paper. This section outlines the basic DE algorithm and briefly describes the statistical tests used in the research. Related work is discussed in Section 3. The algorithms used for comparison are described in Section 4, where their main components are highlighted.
The algorithms are complex, with several integrated mechanisms and applied settings. Complete descriptions of them are provided with references to the original papers in which they were defined. Section 5 outlines the methodology for the comparative analysis. Extensive experiments conducted to evaluate the performance of the comparative algorithms are presented in Section 6. This section also discusses the obtained results. Section 7 highlights the most promising mechanisms integrated into the algorithms under research. Finally, Section 8 summarizes the findings and concludes the paper, offering suggestions for future research directions.

Contributions of the Paper

This paper makes the following contribution:
It describes the recent advances in DE proposed in 2024, focusing on modifications to enhance the algorithm’s effectiveness and efficiency. The algorithms are evaluated across dimensions of 10D, 30D, 50D, and 100D. The improvements are statistically validated using the Wilcoxon signed-rank test, the Friedman test, and the Mann–Whitney U-score test.

2. Preliminary

2.1. Differential Evolution

In this subsection, we review the DE algorithm [1,2], as it serves as the core algorithm for the further advancements studied in this paper. An optimization algorithm aims to identify a vector x so as to optimize f ( x ) ; x = ( x 1 , x 2 , , x D ) . D is the dimensionality of the function f. The variables’ domains are defined by their lower and upper bounds: x j , l o w , x j , u p p ; j { 1 , , D } .
DE is a population-based algorithm where each individual in the population is represented by a vector x i , g ; i = 1 , 2 , , Np . Np denotes the population size, and g represents the generation number. During each generation, DE applies mutation, crossover, and selection operations to each vector (target vector) to generate a trial vector (offspring).
Mutation generates the mutant vector v i , g + 1 according to
v i , g + 1 = x r 1 , g + F · ( x r 2 , g x r 3 , g ) .
Indexes r 1 , r 2 , r 3 are randomly selected within the range { 1 , Np } and they are pairwise different from each other and from index i.
Crossover generates a new vector u i , g + 1 :
u j i , g + 1 = v j i , g + 1 if rand ( j ) Cr or j = rn ( i ) , x j i , g otherwise .
r a n d ( j ) [ 0 , 1 ] is the j-th evaluation of the uniform random generator number. rn ( i ) { 1 , 2 , , D } is a randomly chosen index.
Selection performs a greedy selection scheme:
x i , g + 1 = u i , g + 1 if f ( u i , g + 1 ) f ( x i , g ) , x i , g otherwise .
The initial population is selected uniformly at random within the specified lower and upper bounds defined for each variable x j . The algorithm iterates until the termination criterion is satisfied.
Subsequently, in addition to (1) and (2), other mutation and crossover strategies were proposed in the studies in [3,4,5]. Although the core DE has significant potential, modifications to its original structure were recognized as necessary to achieve substantial performance improvements.

2.2. Statistical Tests

DE belongs to the stochastic algorithms that can return a different solution in each run, as they use random generators. Therefore, we run the stochastic algorithms several times and statistically compare the obtained results. Based on this, we can then say that a given algorithm is statistically better (or worse) than another algorithm with a certain level of significance.
Statistical tests designed for statistical analyses are classified into two categories: parametric and non-parametric. Parametric tests are based on assumptions that are most probably violated when analyzing the performance of stochastic algorithms based on computational intelligence. For this reason, in this paper, we use several non-parametric tests for pairwise and multiple comparisons.
To infer information about two or more algorithms based on the given results, two hypotheses are formulated: the null hypothesis H 0 and the alternative hypothesis H 1 . The null hypothesis is a statement of no difference, whereas the alternative hypothesis represents a difference between algorithms. When using a statistical procedure to reject a hypothesis, a significance level α is set to determine the threshold at which the hypothesis can be rejected. Instead of predefining a significance level ( α ), the smallest significance level that leads to the rejection of H 0 can be calculated. This value, known as the p-value, represents the probability of obtaining a result at least as extreme as the observed one, assuming H 0 is true. The p-value provides insight into the strength of the evidence against H 0 . The smaller the p-value, the stronger the evidence against H 0 . Importantly, it achieves this without relying on a predetermined significance level.
The simple sign test for pairwise comparison assumes that if the two algorithms being compared have equivalent performance, the number of instances where each algorithm outperforms the other would be approximately equal. The Wilcoxon signed-rank test extends this idea by considering not only the number of wins but also the magnitude of differences. Specifically, it ranks the absolute differences in performance for each benchmark function and uses these ranks to determine whether the differences are statistically significant. This test is non-parametric, meaning it does not assume that the performance data follow a normal distribution, which makes it especially suitable for comparing optimization algorithms. This test takes the mean performance from multiple runs for each benchmark function of the two algorithms. Unlike parametric tests, where the null hypothesis H 0 assumes the equivalence of means, these tests define the null hypothesis ( H 0 ) as the equivalence of medians. The test adds the ranks of the positive and negative differences separately. The smaller of the two rank sums is used to compute the test statistic. If the test finds that the sum of ranks for positive and negative differences is significantly different, the null hypothesis (that the two algorithms have equivalent median performance) is rejected. This suggests that one algorithm consistently outperforms the other.
The Friedman test is a non-parametric test used to detect differences in the performance of multiple algorithms across multiple benchmark functions. It is an alternative to the repeated-measures ANOVA when the assumptions of normality are not met. This makes the Friedman test well-suited for comparing optimization algorithms, whose performance data is often non-normal. In this test, the null hypothesis H 0 states that the medians of the different algorithms are equivalent. In contrast, the alternative hypothesis H 1 suggests that at least two algorithms have significantly different medians, which means that they differ significantly in performance. To perform the test, each algorithm’s performance is ranked for every benchmark problem: the best-performing algorithm receives rank 1, the second-best receives rank 2, and so on. These ranks are assigned independently for each problem. Once all ranks are assigned, the test calculates the average rank of each algorithm across all problems. The Friedman test statistic is then computed based on these average ranks to assess whether the observed differences are greater than what would be expected by chance. If the result of the Friedman test is statistically significant, a post hoc analysis (e.g., the Nemenyi test) is conducted to determine which specific pairs of algorithms differ. In the Nemenyi test, the critical distance (CD) is calculated. It is a threshold used to determine whether the difference in performance is statistically significant. If the difference between the average ranks of two algorithms exceeds this CD, their performance is considered significantly different.
The Mann–Whitney U-score test, which is also called the Wilcoxon rank-sum test, is another non-parametric statistical test used to compare two algorithms and determine whether one tends to have higher values than the other. Unlike the Wilcoxon signed-rank test, which compares two related or paired samples (e.g., same problems run by both algorithms), the U-score test is designed for independent samples, such as comparing the results of two algorithms across different trials or problem instances without assuming any pairing between results. The test ranks all results from both algorithms together. The smallest value receives rank 1, the next smallest rank 2, and so on, regardless of which algorithm the value came from. Once all values are ranked, the ranks are then separated back into their respective groups (i.e., by algorithm), and the sum of ranks for each algorithm is calculated. The test statistic (U-score) is derived from these rank sums and represents the number of times observations in one group precede observations in the other. A lower U-score value indicates a stronger difference between the two distributions. The test then assesses whether this observed difference in rank sums will likely have occurred by chance under the null hypothesis ( H 0 ), which states that the two algorithms are drawn from the same distribution (i.e., neither consistently performs better).
Readers interested in a more detailed explanation of statistical tests are advised to consult the literature [6,7]. Our research focuses on the results at the end of the optimization process, while others also consider the convergence of their results, as it is also a desirable characteristic [8].

3. Related Works

One of the first DE surveys was published by Neri and Tirronen in 2010 [9], and another in 2011 by Das and Sugathan [10]. The first one includes some experimental results obtained with the algorithms available at that time. The second reviews the basic steps of the DE, discusses different modifications defined at that time, and also provides an overview of the engineering applications of the DE. Both surveys stated that DE will remain an active research area in the future. The time that followed confirmed this, as noted by the same authors who published the updated survey 5 years later [3]. Several years later, Opara and Arabas in [4] provided a theoretical analysis of DE and discussed the characteristics of DE genetic operators and an overview of the population diversity and dynamics models. In 2020, Bilal et al. [5] published another survey of the literature on DE, which also provided a bibliometric analysis of DE. Some papers review specific components of DE. DE is very sensitive to its parameter settings; the authors in [11,12,13,14] reviewed parameter control methods. Piotrowski et al. [15] analyzed population size settings and adaptation. Parouha and Verma reviewed hybrids with DE [16]. Ma et al. [17] made an exhaustive listing of more than 500 nature-inspired meta-heuristic algorithms. They also empirically analyzed 11 newly proposed meta-heuristics with many citations, comparing them against 4 state-of-the-art algorithms, 3 of which were DE-based. Their results show that 10 out of 11 algorithms are less efficient and robust than 4 state-of-the-art algorithms; they are also inferior to them in terms of convergence speed and global search ability. Cited papers show that, over time, DE has evolved in many different directions and remains one of the most promising optimization algorithms. The present paper analyzes the DE algorithms that were proposed very recently. The focus is on their mechanisms and performance.
The algorithms proposed in the literature are very diverse in terms of exploration and exploitation capabilities, so it is important how we evaluate and compare them. In [6], Derrac et al. give a practical tutorial on the use of statistical tests developed to perform both pairwise and multiple comparisons for multi-problem analysis. Traditionally, algorithms are evaluated using non-parametric statistical tests with the sign test, the Friedman test, the Mann–Whitney test, and both the Wilcoxon signed-rank and rank-sum tests are among the most commonly used. They are based on ordinal results, like the final fitness values of multiple trials. Convergence speed is also an important characteristic of the optimization algorithm. A trial may also terminate upon reaching a predefined target value. In such cases, the outcome is characterized by both the time taken to achieve the target value (if achieved) and the final fitness value. The authors in [8] propose a way to make existing non-parametric tests to compare both the speed and accuracy of algorithms. They demonstrate trial-based dominance using the Mann–Whitney U-score test to compute U-scores with trial data from multiple algorithms. The authors demonstrate that U-scores are much more effective than traditional ways to rank solutions of algorithms. Non-parametric statistical methods are time-consuming and primarily focus on mean performance measures. To use them, raw results of algorithms are needed for comparison. Goula and Sotiropoulo [18] propose a multi-criteria decision-making technique to evaluate the performance of algorithms and rank them. Multiple criteria included best, worst, median, mean, and standard deviation of the error values. They employed equal weighting because the true weights of the criteria are generally unknown in practice.
To perform a comparative analysis of algorithms, the selection of problems on which the algorithms are evaluated is crucial [19]. During the last 20 years, several problem sets were defined. CEC’17 seems to be the most difficult, as it contains a low percentage of unimodal functions (7%), and a high percentage (70%) of hybrid or composition functions.
If we compare our research with related works, we would highlight the following: We conduct an in-depth comparison of the latest DE-based algorithms—particularly those proposed in 2024—and the basic DE algorithm. L-SHADE and jSO were included in the comparison, as they form the core of many improved DE-based algorithms. NL-SHADE-LBC was included as it was one of the best-performing DE-based algorithms in the CEC’22 Competition. The other four are DE-based algorithms from the CEC’24 Competition. The analysis includes a wide range of dimensions, D = 10 , D = 30 , D = 50 , and D = 100 , and three statistical tests. To the best of our knowledge, such an analysis has not yet been performed.

4. Algorithms in the Comparative Study

In this section, we briefly describe the algorithms under comparison. We expose their main characteristics. Readers interested in detailed descriptions of the algorithms are encouraged to study the given references. All algorithms are based on the original DE.

4.1. L-SHADE

The L-SHADE algorithm was proposed in 2014 [20]. Its denominator is the SHADE (Success-History based Adaptive DE) algorithm [21], proposed in 2013, which introduces external archive A into DE to enrich population diversity. The archive A saves target vectors if trial vectors improve them. The SHADE algorithm also introduces a history-based adaptation scheme for adjusting F i and Cr i . Successful F i and Cr i are saved into historical memory by using the weighted Lehmer mean in each generation. The mutant vector is generated using a “DE/currect-to-pbest/1” mutation strategy using the external archive. In 2014, L-SHADE [20] added a linear population size reduction strategy to balance exploration and exploitation.

4.2. NL-SHADE-LBC

The NL-SHADE-LBC (Non-Linear population size reduction Success-History Adaptive Differential Evolution with Linear Bias Change) algorithm was proposed in 2022 [22] as an improved variant of the NL-SHADE-RSP [23] approach, which integrates several key concepts, such as non-linear reduction of population size, rank-based selective pressure in the mutation strategy “DE/currect-to-pbest/1”, automatic tuning of archive usage probability, and a set of rules for regulating the crossover rate. In contrast, NL-SHADE-LBC employs a fixed probability for archive usage and introduces a revised archive update strategy. It also applies a biased parameter adaptation approach using the generalized Lehmer mean for both the scaling factor F and the crossover rate C r . The NL-SHADE-LBC uses a modified bound constraint handling technique. Whenever a mutant vector is generated outside the boundaries, it is generated again, with new parameters. The resampling procedure is repeated up to 100 times. If the new point still lies outside the boundaries, it is then handled using the midpoint target method.

4.3. jSO

jSO was proposed in 2017 [24]. It significantly improves the performance of basic DE. Its denominator is the L-SHADE algorithm. The iL-SHADE [25] algorithm modifies its memory update mechanisms using an adjustment rule that depends on the evolution stage. Higher values for Cr and lower values for F were propagated at the early stages of evolution with the idea of enriching the population diversity. It also uses a linear increase in greediness factor p. jSO is based on iL-SHADE. It introduces a new parameter F w to the scaling factor F to the mutation strategy for tuning the exploration and exploitation ability. A smaller factor F w is applied during the early stages of the evolutionary process, while a larger factor F w is utilized in the later stages. Using a parameter F w , jSO adapts a new weighted version of the mutation strategy “DE/currect-to-pbest-w/1”.

4.4. jSOa

The jSOa (jSO with progressive Archive) algorithm was proposed in 2024 [26]. It proposes a more progressive update of the external archive A. In the SHADE algorithm, and subsequently in the jSO algorithm, when archive A reaches a predefined size, space for new individuals is created by removing random individuals. A newly proposed jSOa introduces a more systematic approach for storing outperformed parent individuals in the archive of the jSO algorithm. When the archive reaches its full capacity, its individuals are sorted based on their functional values, dividing the archive into two sections: better and worse individuals. The newly outperformed parent solution is then inserted into the “worse” section, ensuring that better solutions are preserved while the less effective individuals in the archive are refreshed. Notably, the new parent solution can still be inserted into the archive, even if it performs worse than an existing solution already in the archive. Using this approach, 50% of the better individuals in the archive are not replaced. Except for the archive, the other steps of the jSO algorithm remain unchanged in jSOa.

4.5. mLSHADE-RL

The mLSHADE-RL (multi-operator Ensemble LSHADE with Restart and Local Search) algorithm was also proposed in 2024 [27]. Its core algorithm is L-SHADE, which was extended into the LSHADE-EpSin algorithm with an ensemble approach to adapt the scaling factor using an efficient sinusoidal scheme [28]. Additionally, LSHADE-EpSin uses a local search method based on Gaussian Walks, which is used in later generations to improve exploitation. In LSHADE-EpSin, a mutation strategy DE/current-to-pbest/1 with an external archive is applied. LSHADE-cnEpSin [29] is the improved version of LSHADE-EpSin, which uses a practical selection for scaling parameter F and a covariance matrix learning with Euclidean neighborhoods to optimize the crossover operator. mLSHADE-RL further enhances the LSHADE-cnEpSin algorithm. It integrates three mutation strategies such as “DE/current-topbest-weight/1” with archive, “DE/current-to-pbest/1” without archive, and “DE/current-to-ordpbest-weight/1”. It has been shown that multi-operator DE approaches adaptively emphasize better-performing operators at various stages. mLSHADE-RL [27] also uses a restart mechanism to overcome the local optima tendency. It consists of two parts: detecting stagnating individuals and enhancing population diversity. Additionally, mLSHADE-RL applies a local search method in the later phase of the evolutionary procedure to enhance the exploitation capability.

4.6. RDE

The RDE (Reconstructured Differential Evolution) algorithm was also proposed in 2024 [30]. Just like the previously introduced algorithms, RDE can also be understood as improving the L-SHADE algorithm. Like other algorithms, starting with SHADE, it uses an external archive. The research has demonstrated that incorporating multiple mutation strategies is beneficial; however, adding all-new strategies does not necessarily guarantee the best performance. RDE uses a hybridization of two mutation strategies, the “DE/current-to-pbest/1” and “DE/current-to-order-pbest/1” strategies. The allocation of population resources between the two strategies is governed by an adaptive mechanism that considers the average fitness improvement achieved by each strategy. In RDE, parameters F and Cr are adapted similarly to how they are in jSO. In LSHADE-RSP, the use of a fitness-based rank pressure was proposed for the first time [31]. The idea is to correct the probability of different individuals being selected in DE, with rank indicators assigned to each of them. In RDE, the selection strategy from LSHADE-RSP was slightly modified, as it is based on three instead of two terms.

4.7. L-SRTDE

Like the previous three algorithms, L-SRTDE (Linear population size reduction Success RaTe-based Differential Evolution) was also proposed in 2024 [32]. It focuses on one of the most important parameters of the DE, the scaling factor. L-SRTDE adapts its value based on the success rate, which is defined as the ratio of improved solutions in each generation. The success rate also influences the parameter of the mutation strategy, which determines the number of top solutions from which the randomly chosen one is selected. Another minor modification in L-SRTDE is the usage of repaired crossover rate Cr . The idea is that instead of using the sampled crossover rate Cr for crossover, the actual Cr a value is calculated as the number of replaced components divided by the dimensionality D. All other characteristics of the L-SRTDE algorithm are taken from the L-NTADE algorithm [33], which introduces the new mutation strategy, called “r-new-to-ptop/n/t”, and significantly alters the selection step.

5. Methodology

The purpose of the analysis is to evaluate the performance of recent DE-based algorithms. To evaluate their contributions, we include two well-established baselines and promising new algorithms. To maintain depth in the analysis without overwhelming complexity, we limited the selection of algorithms to recent CEC (Congress on Evolutionary Computation) competition. In addition to the four DE-based algorithms from CEC’24, the basic DE, L-SHADE, NL-SHADE-LBC, and jSO algorithms were used in comparison. L-SHADE is a denominator of many improved versions of DE. NL-SHADE-LBC ranked second in the CEC 2022 Special Session and Competition on Single-Objective Real-Parameter Numerical Optimization. jSO is incorporated in the comparison, as it ranked in second place in the CEC 2017 Special Session and Competition on Single-Objective Real Parameter Numerical Optimization. jSO is also a highly cited algorithm, as many authors use it as a baseline algorithm for further development.
The selected algorithms were analyzed separately for each problem dimension: 10, 30, 50, and 100. They were run on all test problems a predetermined number of times. The best error value for every 10 × D evaluations was recorded for each run. The run was finished after a maximum number of function evaluations had been reached. The maximum number of evaluations was determined based on the problem’s dimensionality and is the same for all algorithms.
Having raw results for all algorithms, five statistical measures were calculated: best, worst, median, mean, and standard deviation of the error values. The algorithms were evaluated using the following statistical tests: Wilcoxon signed-ranks test, Friedman test, and Mann–Whitney U-score test. It is important to note that the computation of the U-score in this article differs from the U-score used in the CEC 2024 competition [8,34]. In our study, we applied the classic Mann–Whitney U-score test, whereas the U-score in the CEC 2024 competition incorporated the “target-to-reach” value. Convergence speed was also analyzed using convergence plots.

6. Experiments

We conducted experiments in which we empirically evaluated the progress brought by the improvements to the DE algorithm over the past year. We included the following algorithms: basic DE, L-SHADE [20], NL-SHADE-LBC [22], jSO [24], jSOa [26], mLSHADE-RL [27], RDE [30], and L-SRTDE [32]. For DE, we used the following settings: F = 0.5 , Cr = 0.9 , and Np = 100 ; these remained fixed during the optimization process. In contrast, other algorithms use a population size reduction mechanism and employ self-adaptive or adaptive mechanisms to control parameters. Each algorithm includes additional parameters specific to its design. We used their default configurations, as described in their original publications, and did not alter any of these parameter settings. Although the authors of the algorithms have published some results, we conducted the runs ourselves in the study based on the publicly available source code of the algorithms. The experiments were performed following the CEC’24 Competition’s [35] instructions: Experiments with the dimensions D = 10 , D = 30 , D = 50 , and D = 100 were performed. Each algorithm was run 25 times for each test function. With 25 runs, it becomes possible to estimate central tendencies (such as the median or mean) and dispersion (such as variance) with reasonable accuracy, while also ensuring that statistical tests (e.g., the Wilcoxon signed-rank test, Mann–Whitney U-score test, and Friedman test) have sufficient power to detect meaningful differences between algorithms. Each run stopped either when the error obtained was less than 10 8 or when the maximal number of evaluations M a x F E s = 10,000 × D was achieved.

6.1. CEC’24 Benchmark Suite

The CEC’24 Special Session and Competition on Single Objective Real Parameter Numerical Optimization [35] was based on the CEC’17 benchmark suite, which is a collection of 29 optimization problems that are based on shifted, rotated, non-separable, highly ill-conditioned, and complex functions. The benchmark problems aim to replicate the behavior of real-world problems, which often exhibit complex features that basic optimization algorithms may struggle to capture. The functions are as follows:
  • Unimodal functions: Bent Cigar function and Zakharov function. Unimodal functions are theoretically the simplest and should present only moderate challenges for well-designed algorithms.
  • Simple multimodal functions: Rosenbrock’s function, Rastrigin’s function, expanded Scaffer’s F6 function, Lunacek Bi_Rastrigin function, non-continuous Rastrigin’s function, Levy function, and Schwefel’s function. Simple multimodal functions, characterized by multiple optima, are rotated and shifted. However, their fitness landscape often exhibits a relatively regular structure, making them suitable for exploration by various types of algorithms.
  • Hybrid functions formed as the sum of different basic functions: Each function contributes a varying weighted influence to the overall problem structure across different regions of the search space. Consequently, the fitness landscape of these functions often varies in shape across different areas of the search space.
  • Composition functions formed as a weighted sum of basic functions plus a bias according to which component optimum is the global one: They are considered the most challenging for optimizers because they extensively combine the characteristics of sub-functions and incorporate hybrid functions as sub-components.
The search range is [ 100 , 100 ] D . The functions are treated as black-box problems. Table 1 presents function groups.

6.2. Results

In this subsection, we present experimental results using three metrics for a comparison of algorithms, namely the U-score, the Wilcoxon test, and the Friedman test. The experimental results were obtained on 29 benchmark functions for eight selected algorithms (DE, L-SHADE, NL-SHADE-LBC, jSO, jSOa, mLSHADE-RL, RDE, and L-SRTDE) on four dimensions (10, 30, 50, and 100). The main experimental results are summarized in the rest of this subsection, while additional detailed and collected results are placed in the Supplementary Materials:
  • Tables S1–S32 present the best, worst, median, mean, and standard deviation values after maximal number of function evaluations for eight algorithms on 29 benchmark functions. For dimension D = 10 , DE solved 9 functions; jSO and mLSHADE-RL each solved 11; NL-SHADE-LBC and L-SHADE solved 12; jSOa solved 13; while RDE and L-SRTDE solved 14 functions. At D = 30 , DE solved three functions; NL-SHADE-LBC, L-SHADE, jSO, and jSOa each solved four; mLSHADE-RL and RDE solved five; and L-SRTDE achieved the best performance with eight functions solved. For D = 50 , DE solved only two functions; NL-SHADE-LBC managed one; L-SHADE, jSO, jSOa, mLSHADE-RL, and RDE each solved four; and L-SRTDE solved five functions. At the highest dimension, D = 100 , DE did not solve any function; NL-SHADE-LBC and L-SRTDE each solved one; L-SHADE, jSO, mLSHADE-RL, and RDE each solved two; and jSOa performed the best with three functions solved.
  • Convergence speed graphs are depicted in Figures S1–S16. L-SRTDE demonstrates the best performance, exhibiting the steepest and most consistent convergence curves. RDE also performs competitively, while the remaining algorithms generally take longer to reach lower objective values. DE, on the other hand, shows minimal improvement throughout the runs, maintaining consistently high objective values.

6.2.1. U-Score

Table 2, Table 3, Table 4 and Table 5 show the U-score values for eight algorithms on 29 benchmark functions for dimensions D = 10 , D = 30 , D = 50 , and D = 100 , respectively. For each function, the rank of each algorithm is presented after ‘/’. At the bottom of the table, a sum of the U-score values and a sum of ranks, RS, are shown.
Overall U-score values for each dimension are collected in Table 6. In this table, “Total” indicates a sum of the U-scores values over all dimensions (higher value is better), while “totalRS” summarizes ranks of algorithms (lower value is better). Based on these results, we can see that L-SRTDE has the first rank, followed by RDE, jSOa, jSO, L-SHADE, mLSHADE-RL, NL-SHADE-LBC, and DE. Table 7 presents the overall ranks of the algorithms for each dimension separately, as well as the cumulative rank across all dimensions, based on the U-score. L-SRTDE holds rank one across all dimensions, while RDE holds rank 2 in all dimensions except for D = 10. jSO and jSOa alternate between ranks 3 and 4. L-SHADE consistently holds rank five across all dimensions. NL-SHADE-LBC and mLSHADE-RL alternate between ranks 6 and 7. DE consistently holds the lowest rank, which is eight, across all dimensions.
We also analyzed the performance of the algorithms across different function families (see Table 1). To achieve this, we summed the U-score values of the algorithms for each function family and identified the best-performing algorithm for each function family across different problem dimensions. The obtained results are collected in Table 8. Unimodal functions are typically easier to solve. For dimension 10, all algorithms perform equally well. In higher dimensions (D = 50, 100), the best performers are jSOa, jSO, L-SHADE, RDE, and L-SRTDE, which all perform similarly well. DE and NL-SHADE-LBC fall behind significantly as the dimension increases, especially at D = 50 and 100. The overall winner is jSOa, with the highest cumulative U-score (22,537.5), indicating strong and consistent performance across all dimensions. L-SRTDE, despite leading in many benchmarks, underperforms slightly in this group (17,170.5), mainly due to lower scores at D = 100. For multimodal functions, L-SRTDE achieves the best results, showing excellent ability to escape local optima and converge effectively. RDE (75,422) also dominates here. DE and NL-SHADE-LBC perform poorly, with much lower totals (22,424 and 39,809, respectively), confirming their difficulty with complex landscapes. jSO, jSOa, and L-SHADE are competitive but consistently fall behind L-SRTDE and RDE in all dimensions. On hybrid functions, L-SRTDE leads significantly (135,343), showing exceptional adaptability. RDE again ranks second (101,431), indicating strong hybrid problem-solving capability. DE and mLSHADE-RL perform weakly (33,028.5 and 43,945.5, respectively), suggesting challenges in handling mixed landscapes. jSO and jSOa improve in higher dimensions but still fall short of L-SRTDE and RDE. Composition functions are the most complex. L-SRTDE (117,833.5) and RDE (113,761) once again show superior performance, likely due to better exploration/exploitation balance. Moderate performers are jSO, jSOa, and L-SHADE with scores between 66,000 to 86,000. DE and NL-SHADE-LBC again show the lowest cumulative scores (41,440 and 64,510).
Based on the U-score tests, we can conclude that L-SRTDE is the most consistently dominant algorithm across all function groups and dimensions, excelling particularly in multimodal, hybrid, and composition functions, which are typically more difficult. RDE is also a strong overall performer, often second to L-SRTDE. jSO and jSOa perform well on unimodal and hybrid functions, but lag in composition problems. DE shows consistently poor performance across all categories, especially as dimensionality increases. NL-SHADE-LBC and mLSHADE-RL are mid-tier at best and struggle with higher-dimensional and complex functions.

6.2.2. U-Score-CEC’24

The obtained results of all eight algorithms were further evaluated using the U-score metric, which was used in CEC’24 [35]. The metric is referred to as U-scores-CEC’24. The sum of U-scores-CEC’24 and the sum of RS over all dimensions are collected in Table 9. We can see that the difference between the algorithms based on the U-score-CEC’24 is smaller than with the Mann–Whitney U-score (see, for example, the difference between DE and L-SRTDE). Overall ranks of algorithms based on the U-score-CEC’24 are given in Table 10. L-SRTDE and RDE once again occupy the top two positions, while DE remains last. The remaining algorithms appear in a slightly different order.

6.2.3. Wilcoxon Signed-Rank Test

The Wilcoxon signed-rank test is used to compare two algorithms, and the number of all pairwise comparisons for eight algorithms per one dimension is 8 × 5 / 2 = 20 . Altogether, we have 80 comparisons. Usually, we only compare the newly designed algorithm with other algorithms, which reduces the number of comparisons. We chose L-SRTDE to compare it with other algorithms and the summarized results are presented in Table 11. For each dimension, the comparison is performed, using the Wilcoxon signed-rank test with α = 0.05 , of the L-SRTDE algorithm with other algorithms, and the summarized results are presented as +/≈/−. Sign ‘+’ means that L-SRTDE performs significantly better than the compared algorithm, ‘−’ means that L-SRTDE performs significantly worse than the compared algorithm, while ‘≈’ means that there is no significant difference between the compared algorithms. Detailed results are shown in Tables S33–S36, while summarized results are collected in Table 11. The lines labeled ‘Total’ contain the sum of the values for each individual algorithm according to the dimension.
We can see that L-SRTDE performs well but not overwhelmingly better in low-dimensional problems. There is greater performance overlap with competitors. L-SRTDE becomes more clearly superior in D = 30 . Most differences are now statistically significant, especially against older or less adaptive methods. As the problem dimension increases, L-SRTDE pulls further ahead of its peers, with most comparisons resulting in statistically significant superiority. In a high-dimensional setting D = 100 , L-SRTDE is consistently and significantly better than all other algorithms, solidifying its position as the best performer.

6.2.4. Friedman Test

As a third test to compare the algorithms, we used the Friedman test. The obtained results for each dimension are presented in Figure 1, Figure 2, Figure 3 and Figure 4 and Table 12, Table 13, Table 14 and Table 15.
From Table 12, Table 13, Table 14 and Table 15, we can conclude that L-SRTDE consistently ranks first across all dimensions, demonstrating dominance that is not affected by problem dimensionality. RDE secures second place in every case except for D = 10, where it drops to the penultimate position. jSO and jSOa alternate between third and fourth place, while L-SHADE consistently holds fifth. mLSHADE-RL follows in sixth, with NL-SHADE-LBC in seventh. DE finishes last across all dimensions, confirming that more modern adaptive or hybrid algorithms clearly outperform it.
In Figure 1, Figure 2, Figure 3 and Figure 4, a critical distance indicates that there exists a statistical difference among the two compared algorithms if the difference between their Friedman rank values is smaller than the critical value. We plot a dotted line of the best-obtained algorithm and the blue line to show the critical distance compared to the best-performed algorithm, which is depicted in green.
Figure 1 shows the comparison for D = 10 , and we can see that all eight algorithms are below the critical distance from L-SRTDE, which is the best algorithm. This means there are no significant differences between all eight algorithms. Table 12 tabulates Friedman ranks, shows a performance order in the Rank column, and at the bottom of the table, the χ 2 and p values of the Friedman test are reported. If the p-value is lower than predefined significance level ( α ), it indicates that the algorithms have statistically significant differences in performance.
The obtained results for D = 30 , D = 50 , and D = 100 are shown in Figure 2, Figure 3 and Figure 4, respectively. One can see that L-SRTDE is statistically better than DE, L-SHADE, NL-SHADE-LBC, jSO, jSOa, and mLSHADE-RL, while there is no statistical difference between L-SRTDE and RDE. This holds for each dimension of D = 30 , D = 50 , and D = 100 .
On D = 100 , we can also see that algorithms RDE, jSOa, and jSO are close to the blue line (i.e., critical distance), and RDE is the only one that is slightly below the blue line.
In general, all three tests that were used in the comparison of the eight algorithms report very consistent comparison results for each dimension.

7. Discussion

The classic DE algorithm lacks the sophisticated mechanisms introduced in the algorithms, which were proposed later. While it is generally effective, it struggles with maintaining diversity and balancing exploration and exploitation without additional enhancements. It does not have adaptive mechanisms for F and Cr , leading to suboptimal parameter tuning. It does not have an external archive or hybrid strategies, limiting diversity control. It does not use a restart mechanism, making it prone to premature convergence.
The ranking of the compared algorithms reflects each algorithm’s ability to balance exploration and exploitation, maintain population diversity, and adapt to different stages of evolution. L-SRTDE leads due to its success rate-driven adaptation, refining both mutation and scaling. The repaired crossover rate ( Cr a ) refines the crossover mechanism, ensuring more effective variation across dimensions. RDE follows closely, leveraging hybrid strategies and adaptive resource allocation. It combine “DE/current-to-pbest/1” and “DE/current-to-order-pbest/1”. The key improvement is in how resources are allocated between these strategies—an adaptive mechanism evaluates their average fitness improvement and adjusts accordingly. Additionally, RDE incorporates rank pressure-based selection, refining the probability of selecting individuals based on fitness ranks. jSOa stands out for its archive management, ensuring stability. Instead of randomly deleting older solutions, jSOa divides the archive into “better” and “worse” sections based on functional values. This ensures that stronger solutions are preserved while allowing some weaker solutions to be refreshed. This more structured archive update leads to better diversity retention without compromising convergence speed. jSO remains strong, though jSOa refines its archive. mLSHADE-RL’s multi-operator approach seems to be powerful. Non-linear population size reduction strategy combined with linear bias change for both the scaling factor F and the crossover rate C r improve the efficiency of the NL-SHADE-LBC algorithm. The L-SHADE algorithm, with its robust adaptive mechanism and dynamic population control, serves as the basis for many enhancements to the DE algorithm.
Each of these algorithms offers unique strengths, but L-SRTDE currently represents the most advanced evolutionary strategy in this group of algorithms.

8. Conclusions

In this study, we conducted a comprehensive review of state-of-the-art algorithms based on Differential Evolution (DE) that have been proposed in recent years, including L-SHADE, jSO, jSOa, NL-SHADE-LBC, mLSHADE-RL, RDE, and L-SRTDE, along with classic DE. We examined the key mechanisms incorporated into these algorithms and systematically compared their performance. Specifically, we identified shared and unique mechanisms among the algorithms.
To ensure robust performance evaluation, we employed statistical comparisons, which provide reliable insights into algorithm efficacy. The Wilcoxon signed-rank test was utilized for pairwise comparisons, while the Friedman test was applied for multiple comparisons. Additionally, the Mann–Whitney U-score test was incorporated to enhance the statistical rigor of our analysis. We also performed a cumulative assessment of the eight algorithms and evaluated their performance across different function families, including unimodal, multimodal, hybrid, and composition functions.
The experimental evaluation was conducted on benchmark problems defined for the CEC’24 Special Session and Competition on Single Objective Real Parameter Numerical Optimization. We analyzed problem dimensions of 10, 30, 50, and 100, running all algorithms independently with parameter settings as recommended by the original authors.
The key findings from our experiments indicate that all three statistical tests consistently ranked the algorithms in the following order: L-SRTDE achieved the highest rank, followed by RDE. Next are jSOa and jSO, which alternate between third and fourth positions. LSHADE consistently follows in fifth, succeeded by mLSHADE-RL and NL-SHADE-LBC, with DE ranking last. Notably, a similar ranking was obtained by the method proposed in [8], which was also employed for algorithm ranking in CEC 2024 [34]. However, the CEC 2024 competition was performed only at dimension D = 30 [34]. We extend the analysis to a lower dimension ( D = 10 ) and higher dimensions ( D = 50 and D = 100 ). The analysis across different function families ranks the algorithms slightly differently at dimension D = 10 ; however, at D = 100 , L-SRTDE remains the best across all function families.
In conclusion, significant progress has been made since the first published Differential Evolution algorithm. The enhanced versions of DE now incorporate a variety of mechanisms, all of which collectively contribute to improved performance.

Supplementary Materials

The following supporting information can be downloaded at https://www.mdpi.com/article/10.3390/math13101556/s1.

Author Contributions

Conceptualization, M.S.M.; methodology, M.S.M. and J.B.; software, J.B.; validation, M.S.M. and J.B.; investigation, M.S.M. and J.B.; writing—original draft preparation, J.B. and M.S.M.; writing—review and editing, J.B. and M.S.M.; visualization, J.B. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Slovenian Research and Innovation Agency (research core funding No. P2-0069—Advanced Methods of Interaction in Telecommunications and P2-0041—Computer Systems, Methodologies, and Intelligent Services).

Data Availability Statement

The data are available as Supplementary Materials.

Acknowledgments

The authors would like to thank the developers of the jSOa, mLSHADE-RL, RDE, and L-SRTDE algorithms for making their source code publicly available.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Storn, R.; Price, K. Differential evolution—A simple and efficient heuristic for global optimization over continuous spaces. J. Glob. Optim. 1997, 11, 341–359. [Google Scholar] [CrossRef]
  2. Price, K. Differential Evolution: A Practical Approach to Global Optimization; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2005. [Google Scholar]
  3. Das, S.; Mullick, S.S.; Suganthan, P.N. Recent advances in differential evolution—An updated survey. Swarm Evol. Comput. 2016, 27, 1–30. [Google Scholar] [CrossRef]
  4. Opara, K.R.; Arabas, J. Differential Evolution: A survey of theoretical analyses. Swarm Evol. Comput. 2019, 44, 546–558. [Google Scholar] [CrossRef]
  5. Bilal; Pant, M.; Zaheer, H.; Garcia-Hernandez, L.; Abraham, A. Differential Evolution: A review of more than two decades of research. Eng. Appl. Artif. Intell. 2020, 90, 103479. [Google Scholar] [CrossRef]
  6. Derrac, J.; García, S.; Molina, D.; Herrera, F. A practical tutorial on the use of nonparametric statistical tests as a methodology for comparing evolutionary and swarm intelligence algorithms. Swarm Evol. Comput. 2011, 1, 3–18. [Google Scholar] [CrossRef]
  7. Carrasco, J.; García, S.; Rueda, M.; Das, S.; Herrera, F. Recent trends in the use of statistical tests for comparing swarm and evolutionary computing algorithms: Practical guidelines and a critical review. Swarm Evol. Comput. 2020, 54, 100665. [Google Scholar] [CrossRef]
  8. Price, K.V.; Kumar, A.; Suganthan, P.N. Trial-based dominance for comparing both the speed and accuracy of stochastic optimizers with standard non-parametric tests. Swarm Evol. Comput. 2023, 78, 101287. [Google Scholar] [CrossRef]
  9. Neri, F.; Tirronen, V. Recent advances in differential evolution: A survey and experimental analysis. Artif. Intell. Rev. 2010, 33, 61–106. [Google Scholar] [CrossRef]
  10. Das, S.; Suganthan, P.N. Differential evolution: A survey of the state-of-the-art. IEEE Trans. Evol. Comput. 2011, 15, 4–31. [Google Scholar] [CrossRef]
  11. Dragoi, E.N.; Dafinescu, V. Parameter control and hybridization techniques in differential evolution: A survey. Artif. Intell. Rev. 2016, 45, 447–470. [Google Scholar] [CrossRef]
  12. Tanabe, R.; Fukunaga, A. Reviewing and benchmarking parameter control methods in differential evolution. IEEE Trans. Cybern. 2019, 50, 1170–1184. [Google Scholar] [CrossRef] [PubMed]
  13. Al-Dabbagh, R.D.; Neri, F.; Idris, N.; Baba, M.S. Algorithmic design issues in adaptive differential evolution schemes: Review and taxonomy. Swarm Evol. Comput. 2018, 43, 284–311. [Google Scholar] [CrossRef]
  14. Reyes-Davila, E.; Haro, E.H.; Casas-Ordaz, A.; Oliva, D.; Avalos, O. Differential Evolution: A Survey on Their Operators and Variants. Arch. Comput. Methods Eng. 2024, 32, 83–112. [Google Scholar] [CrossRef]
  15. Piotrowski, A.P. Review of differential evolution population size. Swarm Evol. Comput. 2017, 32, 1–24. [Google Scholar] [CrossRef]
  16. Parouha, R.P.; Verma, P. State-of-the-art reviews of meta-heuristic algorithms with their novel proposal for unconstrained optimization and applications. Arch. Comput. Methods Eng. 2021, 28, 4049–4115. [Google Scholar] [CrossRef]
  17. Ma, Z.; Wu, G.; Suganthan, P.N.; Song, A.; Luo, Q. Performance assessment and exhaustive listing of 500+ nature-inspired metaheuristic algorithms. Swarm Evol. Comput. 2023, 77, 101248. [Google Scholar] [CrossRef]
  18. Goula, E.M.K.; Sotiropoulos, D.G. HRA: A Multi-Criteria Framework for Ranking Metaheuristic Optimization Algorithms. arXiv 2024, arXiv:2409.11617. [Google Scholar]
  19. Piotrowski, A.P.; Napiorkowski, J.J.; Piotrowska, A.E. Choice of benchmark optimization problems does matter. Swarm Evol. Comput. 2023, 83, 101378. [Google Scholar] [CrossRef]
  20. Tanabe, R.; Fukunaga, A.S. Improving the search performance of SHADE using linear population size reduction. In Proceedings of the 2014 IEEE Congress on Evolutionary Computation (CEC), Beijing, China, 6–11 July 2014; IEEE: Piscataway, NJ, USA, 2014; pp. 1658–1665. [Google Scholar]
  21. Tanabe, R.; Fukunaga, A. Evaluating the performance of SHADE on CEC 2013 benchmark problems. In Proceedings of the 2013 IEEE Congress on Evolutionary Computation, Cancun, Mexico, 20–23 June 2013; IEEE: Piscataway, NJ, USA, 2013; pp. 1952–1959. [Google Scholar]
  22. Stanovov, V.; Akhmedova, S.; Semenkin, E. NL-SHADE-LBC algorithm with linear parameter adaptation bias change for CEC 2022 Numerical Optimization. In Proceedings of the 2022 IEEE Congress on Evolutionary Computation (CEC), Padua, Italy, 18–23 July 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 01–08. [Google Scholar]
  23. Stanovov, V.; Akhmedova, S.; Semenkin, E. NL-SHADE-RSP algorithm with adaptive archive and selective pressure for CEC 2021 numerical optimization. In Proceedings of the 2021 IEEE Congress on Evolutionary Computation (CEC), Kraków, Poland, 28 June–1 July 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 809–816. [Google Scholar]
  24. Brest, J.; Maučec, M.S.; Bošković, B. Single objective real-parameter optimization: Algorithm jSO. In Proceedings of the 2017 IEEE Congress on Evolutionary Computation (CEC), Donostia, Spain, 5–8 June 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 1311–1318. [Google Scholar]
  25. Brest, J.; Maučec, M.S.; Bošković, B. iL-SHADE: Improved L-SHADE algorithm for single objective real-parameter optimization. In Proceedings of the 2016 IEEE Congress on Evolutionary Computation (CEC), Vancouver, BC, Canada, 24–29 July 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 1188–1195. [Google Scholar]
  26. Bujok, P. Progressive Archive in Adaptive jSO Algorithm. Mathematics 2024, 12, 2534. [Google Scholar] [CrossRef]
  27. Chauhan, D.; Trivedi, A.; Shivani. A Multi-operator Ensemble LSHADE with Restart and Local Search Mechanisms for Single-objective Optimization. arXiv 2024, arXiv:2409.15994. [Google Scholar]
  28. Awad, N.H.; Ali, M.Z.; Suganthan, P.N.; Reynolds, R.G. An ensemble sinusoidal parameter adaptation incorporated with L-SHADE for solving CEC2014 benchmark problems. In Proceedings of the 2016 IEEE Congress on Evolutionary Computation (CEC), Vancouver, BC, Canada, 24–29 July 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 2958–2965. [Google Scholar]
  29. Awad, N.H.; Ali, M.Z.; Suganthan, P.N. Ensemble sinusoidal differential covariance matrix adaptation with Euclidean neighborhood for solving CEC2017 benchmark problems. In Proceedings of the 2017 IEEE Congress on Evolutionary Computation (CEC), Donostia, Spain, 5–8 June 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 372–379. [Google Scholar]
  30. Tao, S.; Zhao, R.; Wang, K.; Gao, S. An Efficient Reconstructed Differential Evolution Variant by Some of the Current State-of-the-art Strategies for Solving Single Objective Bound Constrained Problems. arXiv 2024, arXiv:2404.16280. [Google Scholar]
  31. Stanovov, V.; Akhmedova, S.; Semenkin, E. LSHADE algorithm with rank-based selective pressure strategy for solving CEC 2017 benchmark problems. In Proceedings of the 2018 IEEE Congress on Evolutionary Computation (CEC), Rio de Janeiro, Brazil, 8–13 July 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 1–8. [Google Scholar]
  32. Stanovov, V.; Semenkin, E. Success Rate-based Adaptive Differential Evolution L-SRTDE for CEC 2024 Competition. In Proceedings of the 2024 IEEE Congress on Evolutionary Computation (CEC), Yokohama, Japan, 30 June–5 July 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 1–8. [Google Scholar]
  33. Stanovov, V.; Akhmedova, S.; Semenkin, E. Dual-Population Adaptive Differential Evolution Algorithm L-NTADE. Mathematics 2022, 10, 4666. [Google Scholar] [CrossRef]
  34. Qiao, K.; Wen, X.; Ban, X.; Chen, P.; Price, K.; Suganthan, P.; Liang, J.; Wu, G.; Yue, C. Performance Comparison of CEC 2024 Competition Entries on Numerical Optimization Considering Accuracy and Speed; Technical Report; Zhengzhou University: Zhengzhou, China; Qatar University: Doha, Qatar, 2024; Available online: https://github.com/P-N-Suganthan/2024-CEC/blob/main/CEC2024%20Comparison%20Slides.pdf (accessed on 25 April 2025).
  35. Qiao, K.; Wen, X.; Ban, X.; Chen, P.; Price, K.; Suganthan, P.; Liang, J.; Wu, G.; Yue, C. Evaluation Criteria for CEC 2024 Competition and Special Session on Numerical Optimization Considering Accuracy and Speed; Technical Report; Zhengzhou University: Zhengzhou, China; Central South University: Changsha, China; Henan Institute of Technology: Xinxiang, China; Qatar University: Doha, Qatar, 2023; Available online: https://github.com/P-N-Suganthan/2024-CEC/blob/main/Tech_Report_U_Score_Comparison.pdf (accessed on 25 April 2025).
Figure 1. Friedman ranks of algorithms with a critical distance for 10D.
Figure 1. Friedman ranks of algorithms with a critical distance for 10D.
Mathematics 13 01556 g001
Figure 2. Friedman ranks of algorithms with a critical distance for 30D.
Figure 2. Friedman ranks of algorithms with a critical distance for 30D.
Mathematics 13 01556 g002
Figure 3. Friedman ranks of algorithms with a critical distance for 50D.
Figure 3. Friedman ranks of algorithms with a critical distance for 50D.
Mathematics 13 01556 g003
Figure 4. Friedman ranks of algorithms with a critical distance for 100D.
Figure 4. Friedman ranks of algorithms with a critical distance for 100D.
Mathematics 13 01556 g004
Table 1. Function groups.
Table 1. Function groups.
Function GroupFunctions
Unimodal functionsf1, f2
Multimodal functionsf3, f4, f5, f6, f7, f8, f9
Hybrid functionsf10, f11, f12, f13, f14, f15, f16, f17, f18, f19
Composition functionsf20, f21, f22, f23, f24, f25, f26, f27, f28, f29
Table 2. U-scores of algorithms for 10D. Sum denotes the total of U-scores, and RS represents the sum of ranks across all 29 functions.
Table 2. U-scores of algorithms for 10D. Sum denotes the total of U-scores, and RS represents the sum of ranks across all 29 functions.
Func.DENL-SHADE-LBCL-SHADEjSOjSOamLSHADE-RLRDEL-SRTDE
f12187.5/4.52187.5/4.52187.5/4.52187.5/4.52187.5/4.52187.5/4.52187.5/4.52187.5/4.5
f22187.5/4.52187.5/4.52187.5/4.52187.5/4.52187.5/4.52187.5/4.52187.5/4.52187.5/4.5
f31906/71794/82300/3.52300/3.52300/3.52300/3.52300/3.52300/3.5
f40/82264.5/61155.5/72685/32425.5/52749.5/22508/43712/1
f52212.5/42212.5/42212.5/42212.5/42212.5/42212.5/42212.5/42012.5/8
f60/82619/32157/52540/42060/62917/21771/73436/1
f70/82714/31300.5/72377/42270/52785.5/22194/63859/1
f82187.5/4.52187.5/4.52187.5/4.52187.5/4.52187.5/4.52187.5/4.52187.5/4.52187.5/4.5
f90/82834/22298/42137.5/52304/32081.5/72084/63761/1
f101037.5/82412.5/2.52304/62412.5/2.52412.5/2.52200.5/72412.5/2.52308/5
f112609/41884/72082/62817/22881.5/1124.5/82734/32368/5
f121944/72581/32307.5/42599/22173/51089.5/81946.5/62859.5/1
f132134/62750/21232/72655/52750/2563/82750/22666/4
f143872/11939/52606/21770/71886/6835/82013/42579/3
f153064/21137/82715/31600/73290/12075/41627.5/61991.5/5
f16797/81447/73872/12336/41961/52618/32902/21567/6
f173820/11804/72710/22191.5/62369.5/324/82258/52323/4
f182592.5/41390.5/72242.5/52761.5/33420.5/1275.5/82813/22004/6
f191556/72676/33263/2498/83375/11730/62159/52243/4
f201452/82494/22003/62382.5/52396/42471/31684.5/72617/1
f21790/82362.5/42362.5/42362.5/42362.5/42362.5/42262/72635.5/1
f22189/83164.5/1889.5/72837.5/42528/51795.5/63159.5/22936.5/3
f231622/72969.5/11197/82207.5/51714.5/62864.5/22263/42662/3
f242882/12676/22605/32329/52517/42006/61156/81329/7
f252187.5/4.52187.5/4.52187.5/4.52187.5/4.52187.5/4.52187.5/4.52187.5/4.52187.5/4.5
f261856/6425/71971/4.51971/4.52804/34318/13803/2352/8
f272437.5/1.52235.5/42233.5/52224.5/61614.5/82068.5/72248.5/32437.5/1.5
f28411/82312/52347/41388/61382/73276/23062.5/33321.5/1
f291035/71643/62533/42946/22931/3681/83649/12082/5
sum/RS48,969/163.563,491/127.563,649/13265,291/129.569,090.5/11657,174/145.568,723/12371,112.5/107
Rank86542731
Table 3. U-scores of algorithms for 30D. Sum denotes the total of U-scores, and RS represents the sum of ranks across all 29 functions.
Table 3. U-scores of algorithms for 30D. Sum denotes the total of U-scores, and RS represents the sum of ranks across all 29 functions.
Func.DENL-SHADE-LBCL-SHADEjSOjSOamLSHADE-RLRDEL-SRTDE
f12187.5/4.52187.5/4.52187.5/4.52187.5/4.52187.5/4.52187.5/4.52187.5/4.52187.5/4.5
f20/82500/42500/42500/42500/42500/42500/42500/4
f31864.5/61241/72087.5/42087.5/42087.5/43836/13625/2671/8
f40/8811/73389/22655/42057/51178/63059/34351/1
f51577.5/81952.5/62470/32365/42575/1.52250/52575/1.51735/7
f60/8859/73564/22681/42170/51158/63000/34068/1
f70/8939/73200/32520/41969/51154/63374.5/24343.5/1
f82187.5/4.52187.5/4.52187.5/4.52187.5/4.52187.5/4.52187.5/4.52187.5/4.52187.5/4.5
f90/8625/72576/32400/62657/22408/52468/44366/1
f10384/81644/51595/63222.5/22987.5/3732/72979.5/43955.5/1
f110/81569/51383/63194/33317/2940/72745/44352/1
f12384/72152/62586/32583/42161/5366/83324/23944/1
f13573/71337/62287/52791/43179/252/83001/34280/1
f14625/71596/62243/43167/32203/50/83522/24144/1
f15363/81496/71923/62124/32006/41998/53386/24204/1
f169/81114/72794/22323/52634/32020/62426/44180/1
f17641/71569/62052/52979/43606/11/83302/33350/2
f18694/71434/62063/52861/32667/40/83523/24258/1
f192619/3351/82094/62503/42326/5985/72704/23918/1
f200/8916/73065/32695/41901/51308/63276/24339/1
f212200/42200/42200/42200/42200/42100/82200/42200/4
f220/82691/32311/52679/41706/61056/72821/24236/1
f230/83677/21711/62187/51304/72217/42350/34054/1
f241005.5/6632/8836/72566.5/52602/42706/34372/12780/2
f25335/82652/32169/62270/42249/5815/72885/24125/1
f262506/31375/6588.5/82133/41347.5/71630/53873/24047/1
f272065/72177.5/61383/82239.5/52321.5/42442.5/12435/32436/2
f28302/8534/72323/52499/42554/31577/63788/23923/1
f291926/5386/81615/72472/32830/22128/44372/11771/6
sum/RS24,448.5/19844,805/17063,383/13773,272/11668,492/116.543,932.5/16588,261/78.5100,906/63
Rank86534721
Table 4. U-scores of algorithms for 50D. Sum denotes the total of U-scores, and RS represents the sum of ranks across all 29 functions.
Table 4. U-scores of algorithms for 50D. Sum denotes the total of U-scores, and RS represents the sum of ranks across all 29 functions.
Func.DENL-SHADE-LBCL-SHADEjSOjSOamLSHADE-RLRDEL-SRTDE
f1142/8483/72850/32850/32850/32625/62850/32850/3
f20/8625/72825/32825/32825/32750/62825/32825/3
f31165/7539/81526.5/62083/52142.5/44182/12787/33075/2
f40/8629/73556/22609/41978/51351/63002/34375/1
f53151.5/31469/62366/42046/53365/20/83812.5/11290/7
f60/8758/73532/22793/41869/51573/63440/33535/1
f70/8644/73524/22491/42106/51300/63066/34369/1
f82150.5/72512.5/3.52512.5/3.52512.5/3.52512.5/3.5274.5/82512.5/3.52512.5/3.5
f90/8625/72560/32390/62481/52578/22491/44375/1
f10218/81597/51407/62770.5/42956.5/3547/73803/24201/1
f114/8693/72003/62692/42748/32006/52979/24375/1
f12434/71146/62347/53102/22797/4307/83009/34358/1
f13623/71275/62003/53001/33478/22/82867/44251/1
f14625/71568/61643/52771/43358/20/83179/34356/1
f1579/81618/61880/52067/41603/72459/33430/24364/1
f1636/81286/62671/32365/42281/51135/73574/24152/1
f1743/81545/51535/63253/23171/3683/72947/44323/1
f18600/71348/61749/52863/43514/265/82986/34375/1
f1940/81209/71628/61841/52726/32396/43285/24375/1
f200/8673/73461/22712/41735/51474/63088/34357/1
f211408/63375/1.51057/82247/41329/72841/31868/53375/1.5
f220/81751/62465/42532/32083/51086/73214/24369/1
f230/83597/22022/52334/41275/6955/72984/34333/1
f241493.5/6373/82035/52483/42975.5/3774/73657/23709/1
f2531/81619/62468/42840/32220/5647/73300/24375/1
f262286/41473/6893/72464/32281/510/83857/24236/1
f272543/4616/82255/52900/2.52900/2.5661/74350/11275/6
f28491/7193/82574/32113/52452/41668/64163/13846/2
f292173/4628/7586/81892/51528/63662/24375/12656/3
sum/RS19,736.5/20435,867.5/17963,934/131.573,842/11171,540/11840,011.5/17493,701/75.5108,868/51
Rank87534621
Table 5. U-scores of algorithms for 100D. Sum denotes the total of U-scores, and RS represents the sum of ranks across all 29 functions.
Table 5. U-scores of algorithms for 100D. Sum denotes the total of U-scores, and RS represents the sum of ranks across all 29 functions.
Func.DENL-SHADE-LBCL-SHADEjSOjSOamLSHADE-RLRDEL-SRTDE
f1327/8365/73425/2.53425/2.53425/2.51925/53425/2.51183/6
f20/8625/72287/53216/34375/13436/22311/41250/6
f31359/7821/81911.5/62150/32112.5/44268/12949/21929/5
f40/8701/73293/22635/42208/51180/63112/34371/1
f52010/51939/6798/72500/44309/10/83076/22868/3
f60/8650/73438/22824/42851/32121/51622/63994/1
f70/8677/73350/22574/42495/51215/62824.5/34364.5/1
f8652.5/71978/51855.5/62926.5/43362.5/20/83362.5/23362.5/2
f90/8626/72007/52825/32800/43048/21819/64375/1
f10628/71730/51362/63112/33391/237/82865/44375/1
f1190/8535/71601/62205/52550/33848/22427/44244/1
f12739/7401/81832/53169/32542/4839/63624/24354/1
f130/81834/51168/63151/33379/2748/72845/44375/1
f141040/81152/71303/63166/22798/31444/52262/44335/1
f150/81819/51848/41711/61636/73686/22425/34375/1
f160/81396/72455/41661/61899/52585/33129/24375/1
f170/81753/61978/52724/33208/21470/71992/44375/1
f18379/81464/5966/73063/33538/21202/62513/44375/1
f190/81011/71466/62122/52330/43374/22822/34375/1
f200/8939/73008/32670/42309/5977/63222/24375/1
f210/84332/11223/72020/52264/42599/31269/63793/2
f22744/71047/62541/52648/43301/2125/82832/34262/1
f23157/83573/21519/62476/41892/5475/73046/34362/1
f241967/6691/81464/72285/32113/52205/42506/24269/1
f25652/71891/51563/62952/32571/4113/83383/24375/1
f262222/51094/6845/72625/32594/40/84350/13770/2
f27970/7155/82295/52481/32211/62447/43525/13416/2
f2889/81095/72361/42400/32023/51514/63820/24198/1
f29464/8707/72418/32403/41802/51632/64375/13699/2
sum/RS14,489.5/21737,001/18057,581/145.576,119.5/106.578,289/106.548,513/15183,733/87.5111,774/50
Rank87543621
Table 6. Overall U-scores of algorithms. Total denotes the sum of U-scores, and totalRS represents the sum of RS across all dimensions.
Table 6. Overall U-scores of algorithms. Total denotes the sum of U-scores, and totalRS represents the sum of RS across all dimensions.
DimensionDENL-SHADE-LBCL-SHADEjSOjSOamLSHADE-RLRDEL-SRTDE
D10/RS48,969/163.563,491/127.563,649/13265,291/129.569,090.5/11657,174/145.568,723/12371,112.5/107
D30/RS24,448.5/19844,805/17063,383/13773,272/11668,492/116.543,932.5/16588,261/78.5100,906/63
D50/RS19,736.5/20435,867.5/17963,934/131.573,842/11171,540/11840,011.5/17493,701/75.5108,868/51
D100/RS14,489.5/21737,001/18057,581/145.576,119.5/106.578,289/106.548,513/15183,733/87.5111,774/50
Total/totalRS107,644/782.5181,164/656.5248,547/546288,524/463287,412/457189,631/635.5334,418/364.5392,660/271
Table 7. Overall rank of algorithms based on the U-score.
Table 7. Overall rank of algorithms based on the U-score.
DimensionDENL-SHADE-LBCL-SHADEjSOjSOamLSHADE-RLRDEL-SRTDE
D1086542731
D3086534721
D5087534621
D10087543621
Overall rank87534621
Table 8. U-scores arranged according to functions’ families.
Table 8. U-scores arranged according to functions’ families.
Function GroupDimensionDENL-SHADE-LBCL-SHADEjSOjSOamLSHADE-RLRDEL-SRTDE
Unimodal
functions
D1043754375437543754375437543754375
D302187.54687.54687.54687.54687.54687.54687.54687.5
D501421108567556755675537556755675
D100327990571266417800536157362433
Total7031.511,160.520,449.521,378.522,537.519,798.520,473.517,170.5
Multimodal
functions
D10630616,625.513,61116,439.515,759.517,233.515,25721,268
D305629.5861519,47416,89615,70314,171.520,28921,722
D5064677176.519,57716,924.516,45411,258.521,11123,531.5
D1004021.5739216,65318,434.520,13811,83218,76525,264
Total22,42439,80969,31568,694.568,054.554,495.575,42291,785.5
Hybrid
functions
D1022,388.517,608.523,03019,22824,106.59334.521,20320,601
D30590812,61819,42524,52524,099636227,93336,630
D50248411,68817,45923,95525,676905328,25638,929
D100224811,36514,61722,97223,88019,19624,03939,183
Total33,028.553,279.574,53190,68097,761.543,945.5101,431135,343
Composition
functions
D1013,41019,975.518,32620,453.520,04121,559.523,79119,943.5
D3010,339.516,324.515,136.521,24619,11416,671.529,09629,572
D5010,425.513,62516,35521,80519,043.512,30431,76832,174
D100726514,58516,22922,29020,77111,11029,10636,144
Total41,44064,51066,046.585,794.578,969.561,645113,761117,833.5
Table 9. Overall U-scores-CEC’24 [35] of algorithms. Total denotes the sum of U-scores, and totalRS represents the sum of RS across all dimensions.
Table 9. Overall U-scores-CEC’24 [35] of algorithms. Total denotes the sum of U-scores, and totalRS represents the sum of RS across all dimensions.
DimensionDENL-SHADE-LBCL-SHADEjSOjSOamLSHADE-RLRDEL-SRTDE
D10/RS47,469/16169,703/11358,466/13345,841/16142,809/17367,845/17373,864/11210,1503/63
D30/RS40,530/16659,120/14076,171/10758,745/14151,148/15356,844/13870,204/12094,738/79
D50/RS33,101/17954,938/14566,733/12565,118/13263,267/13664,095/12767,809/11992,439/81
D100/RS33,301/18249,109/15563,294/13877,414/10675,190/10461,741/13470,408/12177,043/104
Total/totalRS154,401/688232,870/553264,664/503247,118/540232,414/566250,525/527282,285/472365,723/327
Table 10. Overall rank of algorithms based on the U-score-CEC’24 [35].
Table 10. Overall rank of algorithms based on the U-score-CEC’24 [35].
DimensionDENL-SHADE-LBCL-SHADEjSOjSOamLSHADE-RLRDEL-SRTDE
D106.5356.58421
D3085267431
D5087356421
D10087631.5541.5
Overall rank86357421
Table 11. The Wilcoxon signed-rank results of L-SRTDE versus other algorithms ( α = 0.05 ). (+ means that L-SRTDE performs significantly better than a compared algorithm).
Table 11. The Wilcoxon signed-rank results of L-SRTDE versus other algorithms ( α = 0.05 ). (+ means that L-SRTDE performs significantly better than a compared algorithm).
DimensionAlgorithm+/≈/−
10DE13/10/6
NL-SHADE-LBC9/18/2
L-SHADE8/17/4
jSO7/17/5
jSOa8/15/6
mLSHADE-RL11/16/2
RDE6/19/4
Total62/112/29
30DE22/7/0
NL-SHADE-LBC22/7/0
L-SHADE22/6/1
jSO19/9/1
jSOa19/8/2
mLSHADE-RL20/7/2
RDE17/8/4
Total141/52/10
50DE27/1/1
NL-SHADE-LBC26/3/0
L-SHADE25/3/1
jSO23/5/1
jSOa23/5/1
mLSHADE-RL25/2/2
RDE20/3/6
Total169/22/12
100DE28/1/0
NL-SHADE-LBC28/0/1
L-SHADE26/1/2
jSO25/2/2
jSOa24/2/3
mLSHADE-RL26/0/3
RDE21/2/6
Total178/8/17
Table 12. Algorithms’ ranks for 10D.
Table 12. Algorithms’ ranks for 10D.
AlgorithmMeanRankRank
DE4.74137936
NL-SHADE-LBC4.50000004
L-SHADE4.65517245
jSO4.44827593
jSOa4.10344832
mLSHADE-RL4.84482768
RDE4.82758627
L-SRTDE3.87931031
  χ 2 = 5.0974 , p = 0.64808
Table 13. Algorithms’ ranks for 30D.
Table 13. Algorithms’ ranks for 30D.
AlgorithmMeanRankRank
DE6.68965528
NL-SHADE-LBC5.72413797
L-SHADE4.82758625
jSO4.17241383
jSOa4.18965524
mLSHADE-RL5.67241386
RDE2.74137932
L-SRTDE1.98275861
  χ 2 = 95.3302 , p = 9.911 × 10−18
Table 14. Algorithms’ ranks for 50D.
Table 14. Algorithms’ ranks for 50D.
AlgorithmMeanRankRank
DE7.27586218
NL-SHADE-LBC6.17241387
L-SHADE4.29310345
jSO3.77586213
jSOa3.94827594
mLSHADE-RL6.06896556
RDE2.74137932
L-SRTDE1.72413791
  χ 2 = 123.1061 , p = 1.7262 × 10−23
Table 15. Algorithms’ ranks for 100D.
Table 15. Algorithms’ ranks for 100D.
AlgorithmMeanRankRank
DE7.51724148.0
NL-SHADE-LBC6.13793107.0
L-SHADE4.98275865.0
jSO3.67241383.5
jSOa3.67241383.5
mLSHADE-RL5.06896556.0
RDE3.18965522.0
L-SRTDE1.75862071.0
  χ 2 = 111.5434 , p = 4.3921 × 10−21
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Brest, J.; Sepesy Maučec, M. Comparative Study of Modern Differential Evolution Algorithms: Perspectives on Mechanisms and Performance. Mathematics 2025, 13, 1556. https://doi.org/10.3390/math13101556

AMA Style

Brest J, Sepesy Maučec M. Comparative Study of Modern Differential Evolution Algorithms: Perspectives on Mechanisms and Performance. Mathematics. 2025; 13(10):1556. https://doi.org/10.3390/math13101556

Chicago/Turabian Style

Brest, Janez, and Mirjam Sepesy Maučec. 2025. "Comparative Study of Modern Differential Evolution Algorithms: Perspectives on Mechanisms and Performance" Mathematics 13, no. 10: 1556. https://doi.org/10.3390/math13101556

APA Style

Brest, J., & Sepesy Maučec, M. (2025). Comparative Study of Modern Differential Evolution Algorithms: Perspectives on Mechanisms and Performance. Mathematics, 13(10), 1556. https://doi.org/10.3390/math13101556

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop