For the experimental comparisons, we selected the optimization methods GA, WOA, PSO, ACO, and SaDE because they represent different categories of metaheuristic algorithms and are widely established in the literature. These methods are based on diverse principles (biological/physical mimicry and population-based techniques) and do not require gradient information, making them particularly suitable for the non-differentiable multimodal problems examined in our study. We specifically included SaDE as an adaptive DE variant to demonstrate that our proposed NEWDE + MDM outperforms even advanced versions through its simplicity. The improved methods, NSGA-II and NSGA-III, were excluded because our study focuses on single-objective optimization problems, whereas NSGA-II/III are designed for multi-objective problems with multiple targets. Furthermore, the termination rules we propose specifically address convergence in single-objective optimization and are not directly compatible with multi-objective optimization requirements.
The new termination rules (BSS, WSS, TSS, BOSS, SRS, IRS, and doublebox) were carefully selected as they collectively cover different aspects of the convergence process: BSS monitors best solution stability, WSS tracks worst solution evolution, TSS and BOSS assess population-wide behavior, SRS measures value range contraction, IRS evaluates improvement rate, and doublebox analyzes solution distribution. This comprehensive set of rules enables holistic convergence assessment beyond simply tracking the optimal solution. Potential additional termination rules for future research could include metrics like diversity difference, gradient change, and entropy level. However, these would incur additional computational costs and may not be universally applicable across all optimization problem types. Therefore, our current study concentrates on the aforementioned seven rules, which together provide balanced coverage of various convergence process aspects while maintaining computational efficiency.
Table 2.
Comparison of classic DE with new DE and new DE + MDM.
Function | DE | NEWDE | NEWDE + MDM |
---|
ACKLEY | 19,688 | 15,089 | 7249 |
BF1 | 10,683 | 8537 | 5204 |
BF2 | 10,419 | 8416 | 4775 |
BF3 | 10,013 | 7572 | 4285 |
BRANIN | 5670 | 5852 | 2762 |
CAMEL | 8116 | 9206 | 5358 |
DIFFPOWER2 | 14,547 | 11,793 | 10,734 |
DIFFPOWER5 | 31,332 | 33,168 | 25,736 |
DIFFPOWER10 | 39,317 | 38,228 | 40,297 |
EASOM | 4662 | 4589 | 3664 |
ELP10 | 7834 | 7471 | 6053 |
ELP20 | 10,648 | 10,180 | 8809 |
ELP30 | 13,003 | 12,573 | 10,957 |
EXP4 | 6986 | 7134 | 3824 |
EXP8 | 7326 | 7141 | 4329 |
GKLS250 | 7238 | 9351 | 3181 |
GKLS350 | 8107 | 8314 | 1811 (0.96) |
GOLDSTEIN | 9013 | 7751 | 5026 |
GRIEWANK2 | 12,785 | 10,309 | 4397 (0.66) |
GRIEWANK10 | 16,527 | 16,607 | 9553 |
HANSEN | 7748 | 7768 | 6432 |
HARTMAN3 | 6434 | 6450 | 3179 |
HARTMAN6 | 7181 | 6827 | 4790 |
POTENTIAL3 | 7871 | 7677 | 5994 |
POTENTIAL5 | 11,961 | 11,698 | 10,653 |
POTENTIAL6 | 15,460 (0.56) | 15,449 (0.7) | 12,775 (0.86) |
POTENTIAL10 | 22,602 | 21,695 | 20,237 |
RASTRIGIN | 10,597 | 9556 | 4639 (0.93) |
ROSENBROCK4 | 10,336 | 10,513 | 8729 |
ROSENBROCK8 | 12,909 | 12,799 | 11,259 |
ROSENBROCK16 | 16,527 | 16,920 | 15,377 |
SHEKEL5 | 7772 | 7533 | 5306 |
SHEKEL7 | 8013 | 7586 | 5052 |
SHEKEL10 | 8306 | 7527 | 5120 |
SINU4 | 8904 | 7515 | 6723 |
SINU8 | 8968 | 7411 | 6795 |
SINU16 | 12,241 | 10,472 | 8577 |
TEST2N4 | 7874 | 8107 | 4292 |
TEST2N5 | 9347 | 9109 | 3907 (0.96) |
TEST2N7 | 12,006 | 11,153 (0.9) | 4565 (0.7) |
TEST30N3 | 7296 | 7572 | 4558 |
TEST30N4 | 9103 | 8804 | 5958 |
Total | 483,370 (0.98) | 459,422 (0.99) | 332,921 (0.97) |
Table 3.
Comparison of new DE with majority dimension mechanism method versus others.
Function | GA | WOA | PSO | ACO | SaDE | NEWDE + MDM |
---|
ACKLEY | 10,436 | 33,824 | 9279 | 9091 | 11,396 | 7249 |
BF1 | 6265 | 13,475 | 6311 | 7047 (0.66) | 9280 | 5204 |
BF2 | 5995 | 13693 | 5884 | 6908 (0.76) | 8668 | 4775 |
BF3 | 5559 | 21,432 | 5509 | 6450 (0.83) | 7436 | 4285 |
BRANIN | 4536 | 8619 | 4667 | 5944 | 5424 | 2762 |
CAMEL | 5017 | 8902 | 5050 | 4757 | 6558 | 5358 |
DIFFPOWER2 | 8704 | 13,806 | 10,988 | 11,556 | 13,321 | 10,734 |
DIFFPOWER5 | 23,774 | 41,028 | 27,029 | 55,183 | 32,831 | 25,736 |
DIFFPOWER10 | 25,511 | 58,060 | 34,319 | 84,710 | 38,694 | 40,297 |
EASOM | 4080 | 5437 | 4134 | 4203 | 4611 | 3664 |
ELP10 | 5663 | 22,804 | 6588 | 4637 | 7493 | 6053 |
ELP20 | 8800 | 36,649 | 8953 | 5065 | 11,227 | 8809 |
ELP30 | 12,757 | 42,506 | 11,075 | 5416 | 15,027 | 10,957 |
EXP4 | 5163 | 8397 | 5163 | 5935 | 6612 | 3824 |
EXP8 | 5318 | 11,478 | 5440 | 6197 | 6935 | 4329 |
GKLS250 | 4575 | 7612 | 4628 | 4459 | 5963 | 3181 |
GKLS350 | 5184 | 10,088 | 4769 | 4614 | 7826 | 1811 (0.96) |
GOLDSTEIN | 5932 | 11,804 | 6051 | 7248 | 6918 | 5026 |
GRIEWANK2 | 7485 | 11,135 (0.83) | 5317 | 6430 (0.43) | 12,110 | 4397 (0.63) |
GRIEWANK10 | 9393 | 51,041 | 10,239 | 7848 | 14,334 | 9553 |
HANSEN | 6025 | 15,091 | 5033 | 5091 (0.66) | 6967 | 6432 |
HARTMAN3 | 4936 | 10,911 | 5102 | 5408 | 6098 | 3179 |
HARTMAN6 | 5419 | 21,302 | 5825 | 6154 (0.7) | 7141 | 4790 |
POTENTIAL3 | 6455 | 12,705 | 6998 | 6854 | 8026 | 5994 |
POTENTIAL5 | 9878 | 66,605 | 12,339 | 9702 | 12,204 | 10,653 |
POTENTIAL6 | 13,891 (0.76) | 10,648 (0.93) | 13,945 (0.46) | 11,166 (0.06) | 15,368 (0.76) | 12,775 (0.86) |
POTENTIAL10 | 16,834 | 197,044 | 17,948 | 12,540 (0.3) | 28,241 | 20,237 |
RASTRIGIN | 6868 | 10,530 | 5756 | 5346 (0.4) | 9198 | 4639 (0.93) |
ROSENBROCK4 | 6414 | 18,576 | 7611 | 4848 | 9668 | 8729 |
ROSENBROCK8 | 8128 | 25,777 | 10,198 | 5374 | 12,075 | 11,259 |
ROSENBROCK16 | 11,678 | 37,759 | 13,529 | 5987 | 17,178 | 15,377 |
SHEKEL5 | 5705 | 22,886 | 5915 | 6815 (0.56) | 7374 | 5306 |
SHEKEL7 | 5741 | 26,964 | 5938 | 6777 (0.63) | 7469 | 5052 |
SHEKEL10 | 5829 | 20,334 | 5915 | 6670 (0.46) | 7471 | 5120 |
SINU4 | 5334 | 13,266 | 5355 | 5687 (0.73) | 7324 | 6723 |
SINU8 | 5839 | 21,358 | 6188 | 6472 (0.9) | 8606 | 6795 |
SINU16 | 7278 | 47,713 | 6745 | 8465 (0.73) | 12,963 | 8577 |
TEST2N4 | 5813 | 16,104 | 5339 | 5752 (0.56) | 7425 | 4292 |
TEST2N5 | 6516 | 18,131 | 5644 | 5893 (0.36) | 8810 | 3907 (0.76) |
TEST2N7 | 8205 (0.96) | 23,489 (0.63) | 6057 (0.93) | 6199 (0.03) | 10,655 (0.9) | 4565 (0.7) |
TEST30N3 | 5635 | 11,307 | 5634 | 6591 | 6663 | 4558 |
TEST30N4 | 6594 | 18,229 | 6464 | 8813 | 7837 | 5958 |
Total | 335,162 (0.99) | 1,098,519 (0.98) | 350,871 (0.98) | 406,302 (0.8) | 457,425 (0.99) | 332,921 (0.97) |
3.2. Experimental Results
For the aforementioned functions, a series of tests were conducted on a computer equipped with an AMD Ryzen 5950× processor and 128 GB of RAM, running Debian Linux. Each test was repeated 30 times with new random values in each iteration, and the average results were recorded. The tool used was developed in ANSI C++ using the GLOBALOPTIMUS [
41] platform, which is open-source and available at
https://github.com/itsoulos/GLOBALOPTIMUS. The parameter settings of the method are shown in
Table 1.
In the following experimental results, the values in the cells correspond to the average number of function calls over 30 repetitions. The numbers in parentheses indicate the percentage of cases where the method successfully found the global minimum. If no parentheses are present, it means the method was 100% successful in all the tests.
Table 2 presents comparative results between three versions of the differential evolution algorithm: the classic DE, the modified NEWDE, and NEWDE + MDM. The data concern performance on a series of standard test functions, with measurements including the number of objective function evaluations and the success rate of finding the global minimum. From the analysis of the results, we observe that the modified NEWDE outperforms the classic DE in most cases, with a reduced number of objective function evaluations. For example, for the ACKLAY function, NEWDE requires 15,089 evaluations compared to 19,688 for classic DE, while, for the BF1 function, the respective evaluations are 8537 versus 10,683. This improvement becomes even more pronounced with the addition of the MDM, where the evaluation numbers decrease significantly, −7249 for ACKLAY and 5204 for BF1. The MDM appears to offer significant advantages, particularly for complex functions like GKLS350, where the numbers of evaluations drop from 8107 (classic DE) and 8314 (NEWDE) to just 1811, with a 96% success rate. A similarly impressive improvement is observed for the RASTRIGIN function, with the evaluations decreasing from 10,597 to 4639 (93% success rate). In some cases, such as the DIFFPOWER10 and ROSENBROCK16 functions, the improvements are less significant, suggesting that the algorithm’s effectiveness depends on each function’s characteristics. However, the total sum of the evaluations across all the functions shows a clear reduction from 483,370 (classic DE) to 459,422 (NEWDE) and finally to 332,921 (NEWDE + MDM), with success rates of 98%, 99%, and 97%, respectively. The results confirm that combining the modified DE with the MDM leads to significant performance improvement, with reduced objective function evaluations and high success rates in finding the global minimum. This improvement is particularly notable for functions with multiple local minima and high dimensionality, where classical methods struggle to achieve good performance.
The statistical analysis (Pairwise Wilcoxon Test [
42]) conducted using the R programming language to compare classical differential evolution (DE), modified DE (NEWDE), and modified DE with the majority dimension mechanism (NEWDE + MDM) yielded significant conclusions regarding the statistical differences between these methods, as shown in
Figure 1. The
p-values, which express the levels of statistical significance, revealed a highly significant difference (
p < 0.01) between classical DE and modified NEWDE. This indicates that the improvements introduced in NEWDE led to a statistically significant performance enhancement. The comparison between classical DE and NEWDE + MDM showed an extremely significant difference (
p < 0.0001), confirming that the addition of the majority dimension mechanism contributes very substantially to improving the algorithm’s effectiveness. Furthermore, the comparison between modified NEWDE and NEWDE + MDM also demonstrated an equally extremely significant difference (
p < 0.0001). This proves that the majority dimension mechanism provides additional significant advantages even when compared to the already improved version of the algorithm. Overall, the statistical results confirm that both enhanced versions of the algorithm (NEWDE and NEWDE + MDM) are significantly better than classical DE, with NEWDE + MDM being particularly outstanding due to the incorporation of the majority dimension mechanism. These findings highlight the importance of the proposed modifications for improving the performance of the differential evolution algorithm.
Table 3 and
Figure 2 present a comparative performance analysis of various optimization methods, including GA, WOA, PSO, Ant Colony Optimization (ACO) [
43], self-adaptive differential evolution (SaDE), and NEWDE + MDM. The NEWDE + MDM method demonstrates the best overall performance, with the lowest total number of function evaluations (332,921) and a high success rate (97%). Specifically, for functions such as ACKLAY (7249 evaluations), BF1 (5204), and GKLS350 (1811 evaluations with a 96% success rate), NEWDE + MDM clearly outperforms the other methods. Furthermore, its performance on complex functions like RASTRIGIN (4639 evaluations with a 93% success rate) and HARTMAN3 (3179 evaluations) is particularly impressive. The WOA shows the highest number of evaluations (1,098,519), indicating significantly higher computational costs compared to the other methods. However, its success rate remains high (98%), demonstrating that, despite its inefficiency, the algorithm can provide reliable results for certain problems. Both the GA and PSO show similar performance, with total evaluations of 335,162 and 350,871, respectively, and success rates of 99% and 98%. However, for specific functions like DIFFPOWER10, PSO appears to be slightly superior, while the GA performs better on functions like SHEKEL5 and SHEKEL7. The ACO algorithm shows the lowest success rate (80%) and a high evaluation count (406,302), making it less efficient compared to the other methods. Nevertheless, for certain functions like ELP10 (4637 evaluations) and GKLS250 (4459 evaluations), ACO can be competitive. The SaDE method demonstrates interesting results, with a total number of objective function evaluations (457,425) higher than NEWDE + MDM but lower than the WOA while maintaining a very high success rate (0.99), the highest among all the methods. This indicates that SaDE, while less efficient than NEWDE + MDM in terms of required function evaluations, offers exceptional reliability in finding the global minimum, particularly in complex optimization problems where convergence to local minima is a frequent pitfall. Overall, the results confirm that NEWDE + MDM is the most effective method, particularly for problems with multiple local minima and high dimensionality. While the other methods may be effective in specific cases, they cannot match the overall performance of NEWDE + MDM. This analysis highlights the importance of selecting the appropriate optimization algorithm based on the problem characteristics.
The results of the statistical analysis (Friedman test [
44]) conducted in R for comparing various optimization methods revealed significant differences in their performance (
Figure 3). Specifically, a very extremely significant difference (
p < 0.0001) was observed between the GA and WOA, indicating that these two methods differ substantially in terms of performance. However, comparisons between the GA and PSO as well as the GA and ACO showed non-significant differences (
p > 0.05), meaning their performances are statistically indifferent. The comparison between the GA and SaDE showed a significant difference (
p < 0.05), suggesting that SaDE performs differently compared to the genetic algorithm. In contrast, the comparison between the WOA and PSO showed no significant difference (
p > 0.05), while the WOA versus ACO again showed a very extremely significant difference (
p < 0.0001). Furthermore, the WOA differs significantly (
p < 0.05) from SaDE, but this difference becomes highly significant (
p < 0.01) when examined in another context. The comparison between the WOA and the modified differential evolution (NEWDE + MDM) showed a very extremely significant difference (
p < 0.0001), confirming the superiority of the latter method. The comparisons between PSO and ACO, as well as between ACO and NEWDE + MDM, showed non-significant differences (
p > 0.05). However, PSO differs significantly (
p < 0.05) from SaDE, as does ACO when compared to SaDE. Finally, the comparison between SaDE and the WOA confirms an extremely significant difference (
p < 0.001), highlighting that SaDE has statistically better performance compared to the WOA. Overall, the results demonstrate that NEWDE + MDM stands out for its superior performance compared to the other methods, while the WOA shows large statistical differences relative to most of the other techniques. These findings can help in selecting the optimal method depending on the optimization problem at hand.
Table 5 presents a detailed comparison of various termination criteria for the proposed optimization algorithm, including BSS, WSS, TSS, BOSS, SRS, IRS, doublebox, and the combined “All” criterion that incorporates all the previous rules. The analysis reveals that the IRS criterion achieves the lowest total number of function evaluations (263,582) with a 96% success rate, making it the most efficient among the individual termination rules. The “All” criterion, which combines all the rules, demonstrates similar performance, with 257,860 evaluations and a 96% success rate, confirming the superiority of this combined approach. In contrast, the BOSS criterion requires the highest number of evaluations (1,318,249) despite its excellent success rate (99%). This indicates that, while BOSS reliably finds the global optimum, it does so at a significantly increased computational cost. Similarly, the TSS and SRS criteria also require relatively high evaluation counts (465,279 and 420,140, respectively). For specific test functions like GKLS350, all the termination criteria achieve high success rates (96%), with IRS and “All” requiring the fewest evaluations (1811). For the RASTRIGIN function, IRS and “All” stand out with 3548 evaluations and a 93% success rate compared to BOSS, requiring 15,210 evaluations. Similarly, for HARTMAN3, the BSS, IRS, and All criteria need the fewest evaluations (3179, 3042, and 2651, respectively). In cases like POTENTIAL6, the doublebox criterion achieves the highest success rate (96%) but requires substantially more evaluations (35,090) compared to IRS (9616), which has a lower success rate (70%). This demonstrates the need to balance accuracy against computational cost. Overall, the results confirm that either combining multiple criteria (“All”) or using IRS provides the best balance between reduced function evaluations and high success rates. Conversely, criteria like BOSS, while reliable, significantly increase the computational burden. The optimal termination rule selection depends on the problem requirements, particularly regarding the trade-off between solution accuracy and computational efficiency.
The results of the statistical analysis conducted to compare various termination rules revealed interesting findings regarding the statistical significance of their differences (
Figure 4). Specifically, the comparison between the BSS (Best Solution Stopping) and WSS (Worst Solution Stopping) rules showed no significant difference (
p > 0.05), indicating that these two termination approaches do not differ significantly in terms of their effectiveness. Similarly, non-significant differences were observed in comparisons between BSS and TSS (Total Solution Stopping), SRS (Solution Range Stopping), IRS (Iteration Range Stopping), doublebox, and “All”. However, the comparison between BSS and BOSS (Best Overall Solution Stopping) showed an extremely significant difference (
p < 0.001), suggesting that the BOSS rule differs significantly from BSS. A similar, although less statistically significant, difference was observed between WSS and BOSS (
p < 0.05). Notable differences were observed in the other comparisons. The TSS vs. IRS comparison showed a highly significant difference (
p < 0.01), while TSS vs. “All” showed an extremely significant difference (
p < 0.001). The differences between BOSS and IRS, as well as between BOSS and "All", were very extremely significant (
p < 0.0001 in both cases). Significant differences were also observed between SRS and IRS (
p < 0.05), and between SRS and “All” (
p < 0.001). Finally, the IRS vs. “All” comparison showed a very extremely significant difference (
p < 0.0001), while the doublebox vs. “All” comparison also showed a very extremely significant difference (
p < 0.0001). These findings highlight that certain termination rules, such as BOSS and "All", show statistically significant differences compared to other rules, which may have important implications for selecting the optimal termination rule for different optimization problems.