Next Article in Journal
Investigation of Droplet Spreading and Rebound Dynamics on Superhydrophobic Surfaces Using Machine Learning
Previous Article in Journal
Improvement of Anti-Collision Performance of Concrete Columns Using Bio-Inspired Honeycomb Column Thin-Walled Structure (BHTS)
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Hybrid Multi-Strategy Differential Creative Search Optimization Algorithm and Its Applications

School of Mathematics and Computer Science, Shaanxi University of Technology, Hanzhong 723001, China
*
Author to whom correspondence should be addressed.
Biomimetics 2025, 10(6), 356; https://doi.org/10.3390/biomimetics10060356
Submission received: 27 April 2025 / Revised: 24 May 2025 / Accepted: 30 May 2025 / Published: 1 June 2025

Abstract

To address the issues of uneven initial distribution and limited search accuracy with the traditional divergent quantum-inspired differential search (DCS) algorithm, a hybrid multi-strategy variant, termed DQDCS, is proposed. This improved version overcomes these limitations by integrating the refined set strategy and clustering process for population initialization, along with the double Q-learning model to balance exploration and exploitation This enhanced version replaces the conventional pseudo-random initialization with a refined set generated through a clustering process, thereby significantly improving population diversity. A novel position update mechanism is introduced based on the original equation, enabling individuals to effectively escape from local optima during the iteration process. Additionally, the table reinforcement learning model (double Q-learning model) is integrated into the original algorithm to balance the probabilities between exploration and exploitation, thereby accelerating the convergence towards the global optimum. The effectiveness of each enhancement is validated through ablation studies, and the Wilcoxon rank-sum test is employed to assess the statistical significance of performance differences between DQDCS and other classical algorithms. Benchmark simulations are conducted using the CEC2019 and CEC2022 test functions, as well as two well-known constrained engineering design problems. The comparison includes both recent state-of-the-art algorithms and improved optimization methods. Simulation results demonstrate that the incorporation of the refined set and clustering process, along with the table reinforcement learning model (double Q-learning model) mechanism, leads to superior convergence speed and higher optimization precision.

1. Introduction

Optimization theory and its practical applications are undergoing unprecedented transformation in contemporary industrial practice, with their core value lying in the targeted enhancement of key system performance indicators. Driven by the continuous expansion of engineering frontiers and the rapid advancements in computational science, classical optimization paradigms—such as Newton’s method and gradient descent—have increasingly revealed their limitations [1,2,3]. Applied to modern engineering problems with high dimensionality and strong nonlinearity, these methods often face exponential time complexity, making exact solutions practically infeasible. This practical challenge has catalyzed the vigorous development of metaheuristic optimization techniques. These algorithms, inspired by natural phenomena or social behavior [4,5], offer promising alternatives by delivering global or high-accuracy approximate solutions under acceptable computational costs. The balance between solution quality and computational efficiency makes modern metaheuristic algorithms—such as genetic algorithms (GA) [6], particle swarm optimization (PSO) [7], and ant colony optimization (ACO) [8]—powerful tools with wide-ranging applicability. These methods have demonstrated exceptional performance in solving complex optimization problems across various interdisciplinary domains, including intelligent manufacturing, logistics scheduling, and financial modeling [9,10,11].
As a vital branch of metaheuristic algorithms, swarm intelligence optimization techniques have attracted increasing attention in the academic community in recent years. These algorithms emulate the cooperative behavior of biological populations—such as the flock migration model in particle swarm optimization (PSO), the mating behavior pattern in butterfly optimization algorithm (BOA) [12], and the deep-sea migration strategy in the salp swarm algorithm (SALP) [13]—demonstrating distinctive computational advantages. Their simple and intuitive structural design, rapid convergence, and multimodal optimization capabilities enable excellent adaptability in solving complex, nonlinear, and high-dimensional engineering optimization problems. Notably, bio-inspired optimization paradigms are undergoing continuous innovation. Modern engineering design methodologies have accelerated the iterative process through parallel search mechanisms and have adopted dynamic balance strategies to coordinate exploration and exploitation. These approaches have delivered significant application value in improving mechanical system performance, reducing operational costs, and optimizing resource allocation efficiency. For instance, Houssein et al. [14] proposed a dimension learning-enhanced equilibrium optimizer that constructs multi-level feature interaction mechanisms, significantly improving the lesion segmentation accuracy of COVID-19 lung CT images, particularly in low-contrast regions. Alkayem et al. [15] developed an adaptive pseudo-inverse stochastic fractal search algorithm that employs intelligent dimensionality reduction strategies within the solution space, enabling robust detection of subtle defects in complex steel structure damage assessments and greatly enhancing diagnostic reliability. Abdollahzadeh et al. [16] introduced the puma optimization (PO) algorithm, which simulates jaguar-inspired intelligent behavior for designing exploration–exploitation mechanisms. By integrating a hyper-heuristic phase transition strategy, the algorithm adaptively balances optimization stages according to problem characteristics, thereby substantially improving the capability to solve complex optimization problems. Liu et al. [17] integrated the Q-learning mechanism from reinforcement learning to dynamically hybridize the Aquila optimizer (AO) with the improved arithmetic optimization algorithm (IAOA). By reconstructing the mathematical acceleration function, they achieved a better balance between global search and local exploitation while optimizing the reward model to enhance decision-making efficiency. Wang et al. [18] proposed an underwater image enhancement method based on reinforcement learning. The reward was constructed by a human visual perception index, the features were extracted by residual network, and the image was gradually optimized by combining various enhancement algorithms. The visual effect could be improved without reference to the image. Li et al. [19] proposed a joint detection and tracking (IDT) method based on reinforcement learning, which guided the detector to dynamically optimize the threshold through the feedback prediction information of the tracker, and established an adaptive detection gate in the range-Doppler diagram, thus effectively improving the detection probability and tracking continuity of high-frequency ground wave radar for weak targets. Zhu et al. [20] combined the global exploration capability of the black-winged kite algorithm (KA), the local optimization strength of particle swarm optimization (PSO), and the mutation strategy of differential evolution (DE) to balance search abilities, effectively preventing premature convergence and significantly improving both the convergence speed and solution accuracy in high-dimensional optimization problems. However, with the increasing complexity of engineering optimization tasks—characterized by multi-constraint coupling and high-dimensional nonlinearity—current swarm intelligence algorithms have begun to exhibit theoretical limitations when handling non-convex solution spaces and adapting to dynamic environments. In particular, for complex system designs involving strong time-varying properties and multi-objective conflicts, challenges arise in ensuring convergence stability and maintaining well-distributed solution sets.
The differential creative search (DCS) algorithm [21] represents a cutting-edge advancement in the field of metaheuristic optimization. This algorithm introduces an innovative multi-stage co-evolution framework in which a heterogeneous knowledge transfer mechanism dynamically integrates population experience, while a dynamic solution space reconstruction strategy enhances adaptability to complex optimization problems. DCS demonstrates strong global convergence performance in tackling high-dimensional, non-convex, and multimodal engineering optimization tasks. Liu et al. [22] proposed a hybrid approach incorporating an opposition-based learning strategy along with an adaptive reset mechanism that balances fitness and distance. This design encourages low-performance individuals to migrate toward the vicinity of the optimal solution, thereby facilitating the exploration of promising regions in the search space. Cai et al. [23] improved the shortcomings of DCS algorithm through collaborative development mechanism and population evaluation strategy. However, DCS still suffers from limitations in terms of solution distribution bias and insufficient local convergence precision during iterations, which restrict its ability to effectively explore the boundary regions of the search space and identify the global optimum. To address these challenges, we propose an improved initialization strategy that integrates a refined set with a clustering process to enhance the diversity of the initial population. Furthermore, we incorporate a double Q-learning strategy using dual Q-tables, enabling a more effective balance between exploration and exploitation. This balance empowers the algorithm to explore unknown regions in complex environments more effectively, thereby reducing the risk of premature convergence to local optima.
The main contributions of the proposed algorithm in this study are summarized as follows.
(1)
A novel algorithm is developed by integrating a refined set, a clustering process, and a double Q-learning strategy into the differential creative search (DCS) framework. Ablation studies are conducted to verify that each of these strategies contributes positively to the performance enhancement of the DCS algorithm.
(2)
The proposed DQDCS algorithm is benchmarked against ten state-of-the-art algorithms on the CEC2019 and CEC2022 test suites. Extensive simulations demonstrate the superior performance of DQDCS. Its improvements are visualized through convergence curves and boxplots, and further validated by the Wilcoxon rank-sum test to confirm its overall effectiveness.
(3)
The DQDCS algorithm is applied to two real-world constrained engineering design problems: the design of hydrostatic thrust bearings and the synchronous optimal pulse width modulation (SOPWM) problem in three-level inverters. In both cases, the goal is to minimize the objective function under complex constraints. The results show that DQDCS is particularly suitable for solving practical engineering optimization problems.
The remainder of this study is organized as follows. Section 2 introduces the fundamentals of the differential creative search (DCS) algorithm. Section 3 elaborates on the three enhancement strategies incorporated into DQDCS. Section 4 reports the experimental comparisons between DQDCS and other algorithms on the CEC2019 and CEC2022 benchmarks. Section 5 presents the application of DQDCS to practical engineering problems and offers a comprehensive analysis of the results. Finally, Section 6 concludes the study.

2. Differentiated Creative Search Algorithm Thoughts and Process

The differentiated creative search (DCS) algorithm is a swarm intelligence-based optimization method whose core framework integrates differentiated knowledge acquisition (DKA) and creative realism (CR). A dual-strategy mechanism is employed to balance divergent and convergent thinking. In this framework, high-performing individuals adopt a divergent thinking strategy, utilizing existing knowledge and creative reasoning to conduct global exploration guided by the Linnik distribution. In contrast, the remaining individuals adopt a convergent thinking strategy, integrating feedback from both the team leader and peer members to perform local exploitation. This process involves local optimization informed by both elite and randomly selected individuals. The overall procedure of the DCS algorithm is summarized as follows.

2.1. Initialization

The initial population X is randomly generated, with each individual represented as a random solution:
X i = [ x i , 1 , x i , 2 , x i , 3 , , x i , D ] satisfying Equation (1):
x i , d = L + U ( 0 , 1 ) × ( U L )
where U and L represent the lower and upper bounds of the d-th dimension, respectively, and U ( 0 , 1 ) follows a uniform distribution on the interval ( 0 , 1 ) .

2.2. Differentiated Knowledge Acquisition

Differentiated knowledge acquisition (DKA) emphasizes the rate at which new knowledge is acquired, exerting differential effects on individual agents. These effects are primarily manifested through the modification of the individual’s existing knowledge attributes or dimensional components. The parameter η i , t denotes the quantified knowledge acquisition rate of the i-th individual at iteration t, as defined in Equation (2).
η i , t = 1 2 × U 0 , 1 × φ i , t + 1 , i f   U 0 , 1 φ i , t 1 2 × U 0 , 1 × φ i , t + 0 , otherwise
In Equation (2), φ i , t represents the coefficient associated with variable φ for the individual at the t-th iteration, and is calculated using Equation (3). Here, NP denotes the population size, and R i , t indicates the rank of the i-th individual at the beginning of iteration t. The influence of the DKA process on each component of X i is described in Equation (4), where j r a n d U 1 , 2 , , D denotes an integer uniformly selected from the set 1 , 2 , , D and D represents the dimensionality of the problem.
φ i , t = 0.25 + 0.55 × R i , t / N P
j r a n d U 1 , 2 , , D v i , d = v i , d ,   i f   U 0.1 η i , t   o r   d = j r a n d x i , d ,   o t h e r w i s e

2.3. Convergent Thinking

The strategy for low-performing individuals leverages the knowledge base of high performers and incorporates the stochastic contributions of two randomly selected team members into the solution proposed by the current individual. The inertia weight F is generated randomly, and this process is described by Equation (5), where F the inertia weight and X b e s t represents the best global solution in the current population.
V i = F × X b e s t F × X r 2 X i + F × X r 1 X i
X i draws upon convergent thinking by integrating the information provided by team members X r 1 and X r 2 , thereby refining the knowledge of the team leader, X b e s t , as described by Equation (6). Here, the coefficient λ t governs the extent to which peer influence shapes an individual’s social cognition within the team environment. It reflects the degree to which team dynamics affect individual perspectives. The value of λ t decreases over time, as defined in Equation (7), where NFET and NFEmax denote the number of function evaluations in the current iteration and the maximum allowable number of function evaluations, respectively.
V i = ω × X best + λ t × X r 2 X i + ω i , t × X r 1 X i
λ t = 0.1 × 0.518 × ( 1 N F E T / N F E max )
Two random individuals, x r 1 and x r 2 , are selected, and a new candidate solution is generated by incorporating the best individual with stochastic components, as defined in Equation (8).
v i , d = ω x best , d + λ t ( x r 2 , d x i , d ) + ω i , t ( x r 1 , d x i , d )

2.4. Divergent Thinking

A random individual x r 1 is selected, and a new candidate solution is generated using the Linnik distribution, as defined in Equation (9), where L K ( α , σ ) denotes a random variable drawn from the Linnik distribution with parameters α = 0.618 and σ = 0.05 .
v i , d = x r 1 + L K ( α , σ )

2.5. Team Diversification

As the team continues to evolve, it generates increasingly diverse ideas. To maintain diversity and adaptability, the DCS algorithm replaces underperforming members with newly generated individuals. The equation used to generate new individuals is provided in Equation (10).
V N P = L + U ( 0 , 1 ) × ( U L )

2.6. Offspring Population

For each individual X i , t , a trial solution V i , t is generated. The decision to retain or replace the original solution is based on a comparison of their fitness values. If the trial solution V i , t exhibits superior fitness, it replaces the original solution X i , t ; otherwise, the original solution is retained. This process is formulated in Equation (11).
X i , t + 1 = V i , t ,   i f   f ( V i , t )     f ( X i , t ) X i , t ,   otherwise
In each iteration, all individuals X i , t + 1 in the newly generated population are evaluated, and the best global solution X best , t is subsequently updated. This process is formally defined in Equation (12).
X best , t + 1 = X i , t + 1 ,   i f   f ( X i , t + 1 ) < f ( X best , t ) X best , t ,   otherwise
Algorithm 1 outlines the detailed steps of the DCS algorithm described above.
Algorithm 1: Particle swarm optimization
Step 1: Random Initialization of the Population
The initial population is generated randomly to ensure diversity in the solution space.
Step 2: Fitness Evaluation
The fitness of each individual is evaluated by computing its objective function value, which reflects the individual’s performance on the optimization problem.
Step 3: Determination of the Number of High-Performance Individuals
The number of top-performing individuals is determined based on a predefined proportion of the population.
Step 4: Initialization of Iteration Counter and Parameters
Set the iteration counter t = 1, the number of function evaluations NFE = 0, and the probability of population migration to 0.5.
Step 5: Main Optimization Loop
Continue the optimization process while NFE < NFEmax, repeating the position updating and fitness evaluation steps.

3. Hybrid Multi-Strategy Differentiated Creative Search Algorithm

Firstly, the DCS algorithm initializes the population using pseudo-random numbers, which results in limited population diversity and a lack of target-oriented search in the early stages, thereby reducing optimization efficiency. Secondly, the algorithm relies heavily on the best-performing individual during position updates, making it susceptible to premature convergence and hindering its ability to escape local optima. These issues ultimately lead to reduced optimization accuracy and a slower convergence rate.
To address these limitations, targeted strategies are introduced to enhance the algorithm’s overall performance. Previous studies have shown that the diversity of the initial population significantly influences the algorithm’s ability to converge rapidly and accurately. A higher degree of initial diversity allows the algorithm to explore a broader solution space during the search process, thereby increasing the likelihood of identifying the global optimum.
To improve population diversity, generate higher-quality initial solutions, and effectively overcome the limitations of the basic DCS algorithm, a refined set strategy and clustering strategy are incorporated during the population initialization phase. Although the DCS algorithm enhances global exploration through its “creativity” mechanism, it lacks sufficient flexibility in dynamic environments, such as path planning with moving obstacles for UAVs (unmanned aerial vehicles) or real-time demand changes in industrial scheduling.
The incorporation of double Q-learning allows the algorithm to interact continuously with the environment, facilitating real-time perception and autonomous decision-making. This enables the DCS algorithm to adapt its search strategy more precisely under dynamic conditions, thereby maintaining high operational efficiency. Furthermore, the learning capability of double Q-learning enhances the algorithm’s generalization ability, enabling robust performance in unseen scenarios and increasing the practical value of intelligent optimization techniques.

3.1. Refined Set Initialization

Compared to traditional pseudo-random initialization, the refined set strategy achieves a more uniform distribution of the population across the search space through a carefully designed sampling method. This uniformity reduces the likelihood of the algorithm becoming trapped in local optima during the early stages of optimization. By promoting a broader distribution of individuals, the algorithm gains increased opportunities to explore diverse regions and identify superior solutions.
The mathematical formulation of the refined set strategy is presented in Equations (13) and (14), where L and U represent the lower and upper bounds of the d-th dimension, respectively, and i indicates the index of the i-th individual. The parameter p denotes the smallest prime number that satisfies specific conditions, and mod d × i , p refers to the result of the modulo operation applied to d × i with respect to p. Specifically, it involves calculating the remainder of d × i divided by p. The value of p directly influences the initial distribution of the population. A larger p value enhances global exploration capability by promoting a more uniform distribution, while a smaller p value may cause the algorithm to prematurely enter the local exploitation phase. The parameter p is calculated using Equation (15), where N represents the set of natural numbers and n e x t _ p r i m e ( d × i ) denotes the smallest prime number greater than d × i   .
M = mod d × i , p p
x i , d = L + M × ( U L )
p = d × i d × i N n e x t _ p r i m e ( d × i ) d × i N

3.2. Clustering Process

After the population is generated using the refined set strategy, a clustering algorithm is applied for further enhancement. In this study, the k-means clustering method is adopted to allow the DCS algorithm to focus on salient features during the learning phase while minimizing the impact of low-quality data. This process improves the algorithm’s generalization ability on unseen data, thereby increasing its robustness and stability in practical applications. The detailed implementation steps are presented as follows.
(1)
Select the Cluster Centers
Randomly select k points as the initial cluster centers C 1 , C 2 , , C k , where C j = c j , 1 , c j , 2 , , c j , D . In this experiment, the value of k is determined by Equation (16), where NP denotes the population size. The value of k linearly varies with the population size, ensuring that each cluster contains a sufficient number of individuals. When the population is large, increasing the value of k allows for capturing more intricate details. Conversely, when the population is small, reducing the value of k helps prevent excessive subdivision.
k = N P 5
(2)
Calculate the distance and assign individuals
For each individual x i in the population, the Euclidean distance between x i and each cluster center C j is calculated. The individual x i is then assigned to the cluster corresponding to the nearest cluster center. This is specifically expressed by Equation (17).
d x i , C j = d = 1 D x i , d c j , d 2
(3)
Update the cluster centers
For each cluster, the cluster center is recalculated. Let the set of individuals in the j-th cluster be denoted S j . The new cluster center C j = 1 S j x i S j x i is computed as the mean of the individuals in S j , where S j denotes the number of individuals in the set S j .
(4)
Iterative process
Repeat steps 2 and 3 until the cluster centers no longer undergo significant changes.
(5)
Density-based uniform selection
Calculate the local density of each individual x i S j , denoted ρ i = 1 S j x k S j exp d x i , x k d c u t 2 , where d c u t is the truncation distance used in the density calculation. Individuals with moderate density are selected, and the final initialized population is given by Equation (18).
X final = U j = 1 k S e l e c t e d j

3.3. Double Q-Table Reinforcement Learning Model (Double Q-Learning Model)

The Q-learning model consists of five fundamental components: the agent, environment, action, state, and reward. Its operational procedure can be succinctly described as a cyclical interaction of state transition, action selection mechanism, and reward feedback. In traditional Q-learning, a single Q-table is employed to store the estimated values of state–action pairs. However, this approach often suffers from issues such as overestimation bias and limited exploration capacity. To mitigate these limitations, double Q-learning introduces two independent Q-tables, each of which is updated independently. This dual-table framework significantly reduces the estimation bias inherent in the single-table approach and enables a more balanced, robust evaluation of the learned policy.
Let S = s 1 , s 2 , , s m denote the set of environment states and A = a 1 , a 2 , , a n represent the set of actions that the agent can execute. In each iteration, the agent occupies a certain state s S and selects an action a A to perform. After executing the action, the environment provides a reward r t + 1 and a new state s t + 1 . The reward is computed according to Equation (19), where X new denotes the solution at the new position and X i , 1 represents the solution at the current position. Upon receiving this information, the agent evaluates the expected value Q s t , a t for each possible action.
In double Q-learning, either Q 1 or Q 2 is selected for updating with equal probability. The update equations are provided in Equations (20) and (21). In the equations, s t represents the state of the agent, a t denotes the action executed by the agent, r t + 1 is the immediate reward obtained by the agent after executing the action, and γ is the discount factor, which penalizes future rewards. When γ = 0 , double Q-learning considers only the current reward; when γ = 1 , it prioritizes long-term rewards, α is the learning rate, typically within the interval 0 , 1 . Double Q-learning maintains two Q-tables, traditional Q-learning, but with distinct update logic. Figure 1 simplified illustration example of the operational process of double Q-learning, the arrows in the figure indicate the flow direction of the data, while the dotted line represents the dividing line of time, the transition from time t to t+1. The agent alternates between updating the two Q-tables and combines information from both when selecting actions, thereby reducing estimation bias and enhancing policy robustness.
r t + 1 = f ( X new ) f ( X i , 1 )
Q 1 s t , a t = Q 1 s t , a t + α r t + 1 + γ Q 2 s t + 1 , arg max a Q 1 s t + 1 , a Q 1 s t , a t
Q 2 s t , a t = Q 2 s t , a t + α r t + 1 + γ Q 1 s t + 1 , arg max a Q 2 s t + 1 , a Q 2 s t , a t
In the DCS algorithm, the last individual is considered inefficient and updated using random initialization, which often results in low-quality solutions. To address this issue, double Q-learning introduces two independent Q-tables and employs a dual Q-table mechanism to improve solution quality. This separation makes the calculation of target Q-values more reliable, helping to reduce estimation bias.

3.4. Ablation Experiment

To evaluate the effectiveness of the proposed strategies, four representative test functions from the CEC2019 benchmark suite were selected. The performance of the refined set strategy, the clustering process strategy, their combination, and the double Q-learning strategy was systematically compared. Each algorithm was executed for 500 iterations across 30 independent runs. Specifically, D1 adopts the refined set strategy, D2 employs the clustering process strategy, D3 integrates both the refined set and clustering process strategies, and D4 incorporates the double Q-learning mechanism. The corresponding results are summarized in Table 1. Experimental results demonstrate that population initialization using the combined refined set and clustering process significantly enhances the algorithm’s capability to approach the global optimum. Moreover, the D4 variant, augmented with the double Q-learning strategy, achieves the best overall fitness across all test functions.

3.5. Hybrid Multi-Strategy DQDCS Algorithm

The DQDCS algorithm integrates both the refined-point set strategy and a clustering-based approach, and further incorporates a double Q-learning model to construct a multi-strategy hybrid optimization framework. The refined-point set is generated through mathematically guided sampling techniques to ensure a more uniform distribution of individuals across the search space, thereby replacing conventional pseudo-random initialization methods. Such a distribution promotes broader coverage of the solution space and mitigates early-stage search blind spots. A high-quality initial population enhances the algorithm’s ability to converge more reliably toward optimal solutions and reduces performance fluctuations caused by poor initial positioning.
In contrast to purely random initialization, the refined-point set strategy employs structured sampling to diminish the randomness-induced variability in the initial population, thereby improving the algorithm’s robustness. Moreover, this strategy can be tailored to the specific characteristics of the optimization problem; for example, in constrained optimization scenarios, it helps ensure that the initial population satisfies constraint conditions, thus avoiding infeasible solutions and enhancing overall algorithmic stability.
The clustering strategy divides the population into multiple subgroups, with each subgroup representing distinct regions or features of the search space. This structural partitioning aids in preserving diverse solution patterns throughout the optimization process, thereby reducing the risk of premature convergence to local optima. By maintaining diversity and promoting exploration, the algorithm is better positioned to locate the global optimum efficiently.
By integrating the probability distributions and maximum values derived from two independent Q-tables, the double Q-learning mechanism enables more balanced action selection between exploration and exploitation. Within the DQDCS algorithm, this dual-Q-table framework effectively mitigates the estimation bias commonly associated with single-Q-table implementations and facilitates more comprehensive policy evaluation, thereby enhancing the algorithm’s global search capability. Specifically, double Q-learning selects the optimal action based on the current state, which subsequently guides population updates. This approach allows for broader exploration during the early stages of optimization while gradually shifting toward refined exploitation of promising regions in the later stages. As such, the algorithm achieves accelerated convergence without compromising solution quality. A flowchart of the DQDCS algorithm is illustrated in Figure 2.
Algorithm 2 presents the pseudocode of the proposed DQDCS algorithm.
Algorithm 2: DQDCS algorithm.
Initialize the population using Equation (1);
Evaluate fitness for all individuals;
Determine the refined set via the clustering process;
Initialize Q-tables to zero;
Set key parameters: exploration threshold pc, golden ratio, η and φ values;
while the number of function evaluations (nfe) < max_nfe do
  Sort the population by fitness;
  Identify the best solution x_best;
  Compute λt using Equation (7);
  for each individual i do
    Compute ηi and φi using Equations (2) and (3);
    Determine behavior category (high-, average-, or low-performing);
    if i is low-performing and rand < pc then
      Generate a new solution randomly;
    else if i is high-performing then
      Select r1 ≠ i;
      Update selected dimensions using Equation (8);
    else//average-performing
      Select r1, r2 ≠ i;
      Compute ωi;
      Update selected dimensions using Equation (8);
    end if
    Apply reflection-based boundary handling;
    Evaluate fitness of the new solution;
    If improved, update position and fitness;
    Compute reward from fitness change;
    Update Q1 using Q2 for value estimation (Equation (20));
    Update Q2 using Q1 similarly (Equation (21));
  end for
  Update the best solution and record convergence data;
end while
Return best solution, best fitness;

3.6. Complexity Analysis of the Algorithm

The implementation of the DQDCS algorithm involves certain design challenges, yet it maintains relatively low computational complexity. The overall complexity is stage-dependent, as each phase—initialization, fitness evaluation, and solution generation—contributes differently to the total computational cost.
In general, the DQDCS algorithm comprises three fundamental procedures: population initialization, fitness evaluation, and generation of new solutions. The main loop iterates for a maximum of Max_iter iterations, and in each iteration, operations are performed for each of the N search agents.
The computational complexity of the initialization phase is O(N), owing to the refined set-based initialization method and clustering process. Additionally, the time complexity for fitness evaluation depends on the complexity of the objective function, denoted O(F). Therefore, the overall computational complexity of the algorithm can be expressed as O(Max_iter × N × F).

4. Simulation Environment and Result Analysis

In the field of optimization, particularly in the study of evolutionary algorithms and metaheuristic methods, validating the effectiveness of proposed algorithms is of paramount importance, as these approaches are expected to address complex challenges encountered in real-world applications. To assess their performance, standardized test cases or well-established benchmark problems are commonly employed. These benchmark evaluations offer a unified platform for objective comparison, enabling fair and consistent performance assessment across different algorithms and facilitating a rigorous analysis of their strengths and limitations.
This experiment was conducted using MATLAB 2023 and a Windows 11 operating system with Intel Core i5-13400 CPU and 8 GB RAM (Intel Corporation, Santa Clara, CA, USA). The CEC2019 [24] and CEC2022 [25] benchmark functions were employed to evaluate the performance of the DQDCS algorithm. These benchmark functions enable a systematic comparison between DQDCS and other state-of-the-art metaheuristic algorithms, thereby verifying the competitiveness and applicability of the proposed method in solving complex optimization problems. This evaluation framework ensures scientific rigor and provides clear directions for further algorithmic enhancements.
Considering the inherent stochastic nature of metaheuristic algorithms, relying on a single run for each benchmark function may lead to unreliable conclusions. Therefore, multiple simulations were conducted for each algorithm, including puma algorithm [16], the original differentiated creative search algorithm (DCS) [21], the multi-strategy hybrid DQDCS algorithm, a differentiated creative search algorithm with multi-strategy improvement (MSDCS) [22], Chernobyl disaster optimizer (CDO) [26], adaptive spiral flying sparrow search algorithm (ASFSSA) [27], waterwheel plant algorithm (WWAP) [28], subpopulation improved grey wolf optimizer with Gaussian mutation and Lévy flight (SPGWO) [29], dung beetle optimizer (DBO) [30] and nonlinear randomly reuse-based mutated whale optimization algorithm (NRRMWOA) [31]. The detailed experimental results are presented as follows.

4.1. The CEC2019 Benchmark Functions Are Employed for Performance Evaluation

The CEC2019 benchmark functions are specifically designed to evaluate and compare the performance of optimization algorithms. They encompass a diverse set of challenging optimization problems, including multimodal, high-dimensional, and dynamic characteristics, thereby closely simulating real-world complexities. Given the inherent stochastic nature of metaheuristic algorithms, a single run per benchmark function is insufficient to reliably demonstrate an algorithm’s effectiveness. To enhance the reliability and fairness of performance evaluation, each algorithm was independently tested 100 times on each benchmark function, with a maximum of 500 iterations per run. The population size was consistently established at 200 individuals.

4.1.1. CEC2019 Optimization Accuracy Analysis

To accurately evaluate and compare the performance on the CEC2019 benchmark functions, the experimental results are summarized in terms of the best, mean, and standard deviation (Std) values. In the result tables, the best mean values, which serve as key performance indicators, are highlighted with underlining. The detailed outcomes are presented in Table 2.
As shown in the statistical results in Table 2, the proposed DQDCS algorithm consistently achieves the best values on all functions from F1 to F10. Moreover, it obtains the best mean performance on F3–F10 compared to other competing algorithms. In addition, the DQDCS achieves the smallest standard deviation on functions F3, F5, F6, F8, and F9, demonstrating superior robustness. These findings indicate that DQDCS not only offers stable performance but also exhibits highly competitive exploration capability among the compared algorithms.

4.1.2. CEC2019 Convergence Curve Analysis

To facilitate a more intuitive comparison of the convergence behavior across different functions in the CEC2019 benchmark functions, convergence curves for the DQDCS algorithm and ten other competing algorithms were plotted. As shown in Figure 3, the horizontal axis represents the number of iterations, while the vertical axis denotes the fitness value. It can be observed that the DQDCS algorithm demonstrates superior convergence accuracy on functions F1–F7, F9, and F10. Notably, it is capable of approaching the optimal value at the early stages of the search process, particularly on functions F1 and F10, where the curves converge almost linearly to the optimum. These results indicate that the DQDCS algorithm exhibits relatively strong performance and that the incorporation of multi-strategy enhancements is both feasible and effective in improving optimization precision.

4.1.3. CEC2019 Boxplot Analysis

To compare the performance of different algorithms across multiple runs—particularly in terms of stability and robustness—boxplots are employed as an evaluation tool. As shown in Figure 4, for the DQDCS algorithm, the median values for functions F3, F4, F6, F7, and F9 are located near the lower boundaries of the boxes, indicating a right-skewed (positively skewed) distribution. This suggests that the majority of runs yielded relatively favorable results. Moreover, no outliers are observed in functions F1, F4, F5, F7, or F9 and only a few outliers appear in F2, F3, F6, F8, and F10, demonstrating that DQDCS exhibits stronger robustness compared to the other algorithms.

4.1.4. CEC2019 Wilcoxon Rank-Sum Test

The Wilcoxon rank-sum test [32], a non-parametric statistical test, was employed to evaluate the significant differences between two algorithms. The objective was to verify whether there exists a significant difference between the DQDCS algorithm and the other ten comparison algorithms, thereby assessing the optimization performance of DQDCS. A p-value below 0.05 indicates the rejection of the null hypothesis, signifying a significant difference between the two algorithms. Table 3 presents the p-values obtained from the Wilcoxon rank-sum test conducted between DQDCS and the other ten representative comparison algorithms when solving the CEC2019 benchmark. As shown in Table 3, DQDCS outperforms the other comparison algorithms in solving the CEC2019 functions, with p-values between DQDCS and each of the comparison algorithms being lower than 0.05. The results unequivocally emphasize that DQDCS exhibits significant differences compared to the other algorithms in the majority of the functions, highlighting its distinct advantage in solution performance.

4.2. The CEC2022 Benchmark Functions Are Employed for Performance Evaluation

The CEC2022 benchmark functions are a standardized set of problems used to assess the performance of optimization algorithms in research and development. These test functions simulate various aspects of real-world optimization problems, including local minima, maxima, global optima, and a range of complexities such as nonlinearity and discontinuities. By validating algorithms on these diverse test sets, it ensures that the algorithms exhibit high robustness and stability when confronted with different challenging scenarios. This approach helps avoid the issue of algorithms performing exceptionally well in specific environments while failing in others, thereby enhancing the generality and reliability of the algorithm. For a comprehensive evaluation, the complex functions described in the CEC2022 test suite are used to assess the effectiveness of the DQDCS algorithm. The number of iterations is set to 500, and 100 independent tests are conducted for each benchmark function. The population size is consistently established at 200 individuals.

4.2.1. CEC2022 Optimization Accuracy Analysis

To visually observe and compare the results of the CEC2022 benchmark functions, the following table presents the optimal values (best), mean values (mean), and standard deviations (Std) for each test function. In these tables, the mean value is used as the performance indicator, and the best mean values are underlined. These results are provided in Table 4 below.
From the statistical results in Table 4, it can be observed that for test functions F1 to F3, DQDCS achieved the best mean values and found the optimal solutions when compared to the other algorithms. Moreover, when searching for the optimal values of functions F1 to F3, DQDCS exhibited the lowest standard deviation among all algorithms.
For test functions F4 to F8, DQDCS found the optimal solutions for functions F5 to F8. It achieved the best mean values across all algorithms for functions F6, F7, and F8. Additionally, for function F4, DQDCS outperformed DCS in terms of the mean value, and DQDCS also achieved lower standard deviations than the other algorithms in functions F4, F5, and F7.
In the case of test functions F9 to F12, DQDCS found the optimal solutions for functions F11 and F12. For functions F9, F10, and F12, DQDCS achieved the best mean values across all algorithms. For function F11, DQDCS outperformed DCS in terms of the mean value. Furthermore, DQDCS exhibited the lowest standard deviation in functions F10 and F12, and its standard deviation was lower than that of DCS in functions F9 and F11.

4.2.2. CEC2022 Convergence Curve Analysis

To more intuitively compare the optimization accuracy and convergence speed of various algorithms, the convergence curves for each algorithm based on the CEC2022 benchmark functions are presented in Figure 5, which shows a comparison of the convergence curves for eleven algorithms, with the horizontal axis representing the number of iterations and the vertical axis representing fitness values. The DQDCS algorithm exhibits the highest convergence speed on functions F1, F4, F7, F8, F10, F11, and F12, achieving the highest convergence accuracy on functions F1 to F3 and F5 to F12. Furthermore, it almost finds the optimal value at the beginning, especially for functions F2, F8, F9, F10, and F12, where it converges to the optimal value in an almost linear manner. This further confirms that the DQDCS algorithm performs relatively well, and the multi-strategy improvements are effective and feasible in enhancing both the convergence speed and accuracy of the algorithm.

4.2.3. CEC2022 Boxplot Analysis

In order to compare the performance of different algorithms, a boxplot evaluation based on the CEC2022 benchmark functions was drawn to assess and compare the algorithms. As shown in Figure 6, the DQDCS algorithm exhibits no outliers across functions F1 to F12, indicating superior robustness compared to the other algorithms. Furthermore, the boxplots for functions F1 to F3 and F5 to F12 are relatively flat, suggesting that the DQDCS algorithm demonstrates stable data behavior.

4.2.4. CEC2022 Wilcoxon Rank-Sum Test

To further evaluate the effectiveness of the DQDCS algorithm, the Wilcoxon rank-sum test was utilized. This test is ideal for comparing the performance of original and improved algorithms, as it can assess significant differences between two independent samples, particularly when the data do not follow a normal distribution. A key advantage of the Wilcoxon rank-sum test is its non-parametric nature, which makes it particularly effective for comparing two independent sets of samples without assuming a specific distribution, regardless of sample size.
By conducting the Wilcoxon rank-sum test on the DQDCS algorithm and other algorithms based on the CEC2022 test set, the superiority and reliability of the DQDCS algorithm were further evaluated. The specific results are shown in Table 5. The p-value indicates the degree of significance between the two algorithms. When the p-value is less than 5%, the difference is considered significant; otherwise, it is not. The results presented in Table 5 indicate that the p-values are all less than 5%, which demonstrates a significant difference between DQDCS and the other comparison algorithms, further validating the superiority and effectiveness of the DQDCS algorithm.
In conclusion, the DQDCS algorithm significantly enhances the overall performance of the DCS algorithm. When compared with other algorithms, it exhibits outstanding performance and strong overall capability. However, there are still some limitations to the DQDCS algorithm. For instance, the increased diversity in the initialization stage sacrifices some of the algorithm’s convergence speed. Additionally, when solving certain functions, the increased computational load leads to slight performance degradation. Therefore, there remains room for improvement. These results suggest that the hybrid multi-strategy approach is effective for most of the test functions, which aligns with the “no free lunch” theorem.

5. Engineering Case Studies and Results Analysis

In the research and development of metaheuristic optimization algorithms, the construction and refinement of algorithm performance evaluation systems have always been a central focus in the academic community. Traditional evaluation paradigms are often based on benchmark test function sets. While these functions provide a standardized testing environment and clear theoretical optimal solutions, they significantly differ from the complexities of real-world engineering problems, making it difficult to fully reflect the practical applicability of algorithms in real-world scenarios. In contrast, real-world engineering problems typically exhibit highly complex characteristics. First, the optimal global solution is difficult to determine in advance using analytical methods, and its existence and uniqueness often lack rigorous mathematical proof. Second, the problem space generally contains various complex constraints, such as nonlinear constraints, inequality constraints, and boundary conditions, which are interwoven and greatly increase the difficulty of solving the problem. These features make real-world engineering problems an ideal benchmark for testing the robustness, adaptability, and engineering practicality of optimization algorithms.
From both the algorithm validation and engineering application perspectives, employing real-world problems with actual engineering constraints as test cases holds irreplaceable significance for thoroughly evaluating the practical performance of optimization algorithms. This testing approach not only more accurately simulates the algorithm’s performance in real operational environments, but also effectively addresses the limitations of benchmark function tests in reflecting the generalization capability of algorithms. Based on this, this study carefully selected the design of static pressure thrust bearings [33] and the application of synchronous optimal pulse width modulation (SOPWM) in three-level inverters [34] as typical test cases. This study aimed to conduct a systematic empirical analysis to evaluate the effectiveness and superiority of the proposed algorithm in solving complex engineering optimization problems.
The design of static pressure thrust bearings is a typical multidisciplinary optimization problem involving fields such as fluid mechanics, materials science, and thermodynamics. The optimization of its design parameters requires balancing multiple performance metrics, including load-bearing capacity, stability, and energy consumption. On the other hand, the application of synchronous optimal pulse width modulation (SOPWM) in three-level inverters focuses on the field of power electronics, aiming to find the optimal solution among several objectives, such as ensuring output voltage waveform quality, reducing switching losses, and suppressing harmonics. These two engineering cases are highly representative and complementary: on one hand, they encompass various constraints from different engineering fields, such as mechanical and power electronics, which allows for a comprehensive assessment of the algorithm’s capability in handling complex constrained optimization problems; on the other hand, through in-depth analysis of real engineering data, the performance of the algorithm in terms of solution quality, efficiency, and convergence speed can be directly evaluated. The research results not only provide essential empirical evidence for further optimization of the algorithm but also lay a solid theoretical and practical foundation for its broader application in various engineering domains.

5.1. Static Pressure Thrust Bearing

The primary objective of this design problem is to optimize the bearing power loss using four design variables, with the goal of minimizing the power loss. These design variables include the bearing radius R x 1 , groove radius R 0 x 2 , oil viscosity μ x 3 , and flow rate Q x 4 . The problem involves seven nonlinear constraints, labeled g 1 g 7 , which are defined in Equations (23)–(29). These constraints pertain to the load-carrying capacity W, the inlet oil pressure P 0 , and the oil film thickness h, as specified in Equations (30), (31), and (32), respectively.
The objective function f ( x ) primarily includes the flow rate of the lubricant, inlet oil pressure, and the power loss function resulting from friction under specific constraints. The detailed formulation is provided in Equation (22). Additionally, the power loss caused by friction is closely related to the temperature rise of the lubricant and the oil film thickness. The axial load constraint W ensures that the bearing can withstand the specified axial load, which is a fundamental requirement for its proper operation. This constraint is directly related to the bearing’s load-carrying capacity and stability. The inlet oil pressure constraint P 0 guarantees a reasonable inlet pressure, which is crucial for effective lubrication and normal operation of the bearing. Insufficient oil pressure may result in inadequate lubrication, while excessive pressure could lead to leakage or other failures. The oil film thickness constraint h ensures an appropriate film thickness, which is essential for reducing friction, minimizing power loss, and preventing surface wear. This constraint plays a critical role in maintaining the bearing’s performance and extending its service life. Figure 7 illustrates the structure of the static pressure thrust bearing, the blue background indicates regional visualization.
f ( x ) = Q P 0 0.7 + E f
In the objective function f ( x ) : Q represents the flow rate of the lubricant oil; P 0 denotes the inlet oil pressure; and E f represents the power loss caused by friction. The constraints associated with the objective function are mainly composed of the following seven inequality constraints.
g 1 ( x ) = W 10 , 1000 0
g 2 ( x ) = 5000 W π ( R 2 R 0 2 ) 0
g 3 ( x ) = 50 P 0 0
g 4 ( x ) = 0.001 0.0307 386.4 P 0 ( Q 2 π R h ) 0
g 5 ( x ) = R R 0 0
g 6 ( x ) = h 0.001 0
g 7 ( x ) = h 0.001 0
W = π P 0 2 R 2 R 0 2 ln ( R R 0 )
P 0 = 6 μ Q π h 3 ln ( R R 0 )
The temperature rise expression can be calculated using Equations (32) and (33).
Δ T = 2 ( 10 P 559.7 )
P = log 10 log 10 ( 8.122 × 10 6 μ + 0.8 ) + 3.55 10.04
The frictional power loss, E f , is given by Equation (34).
E f = 9336 Q × 0.0307 × 0.5 Δ T
The film thickness, h, is defined as shown in Equation (35).
h = ( 2 π × 750 60 ) 2 2 π μ E f ( R 4 4 R 0 4 4 )
The remaining parameters are defined as shown in Equations (36)–(39).
1 R 16
1 R 0 16
1 × 10 6 μ 16 × 10 6
1 Q 16
To compare the performance of DQDCS with several classical algorithms, the population size was set to 30, the number of iterations to 200, and each algorithm was executed 30 times. The results of the static pressure thrust bearing design problem are presented in Table 6, where the best values are underlined for clarity. It can be observed that DQDCS ranks first in terms of best solution, variance, mean, and worst-case performance. Additionally, the median value of DQDCS ties for first place with DBO. In summary, DQDCS demonstrates superior overall performance in solving this engineering problem.

5.2. Application of SOPWM (Synchronous Optimal Pulse Width Modulation) in Three-Level Inverter

Synchronous optimal pulse width modulation (SOPWM) is an advanced technique used for controlling medium-voltage (MV) drives. It significantly reduces the switching frequency without introducing additional distortion, thereby decreasing switching losses and improving the performance of inverters. Within one fundamental period, the switching angles are computed to minimize current distortion simultaneously. SOPWM can be transformed into a scalable constrained optimization problem. For inverters with different voltage levels, the application of SOPWM in three-level inverters can be described as follows. The primary objective of this problem is to minimize the current distortion f, subject to the constraints g and h, as described by Equations (40)–(43). Figure 8 illustrates the structure of the SOPWM application in a three-level inverter.
f = k k 4 i = 1 N s i cos k α i 2 k k 4
k = 5 , 7 , 11 , 13 , , 97 ,
N = f s , max f . m
s i = 1 i + 1
The constraint on the relationship between adjacent switching angles is defined by Equation (44). Specifically, it requires that the difference between two consecutive switching angles must exceed a threshold 10 5 . This is intended to ensure a certain degree of regularity and stability in the variation of switching angles, thereby preventing potential control issues caused by excessively close switching angles. The constraint condition h ( i ) involves a modulation-related parameter m, and is expressed by Equation (45). This condition states that under specific circumstances, the sum of cosine terms involving s i and α ( i ) must be equal to m . It serves to constrain the switching angles and other related parameters to ensure compliance with the operational requirements of the system. Equation (46) defines the constraint on the switching angle α i , restricting its values to a range between 0 and π 2 . This limitation is established based on the physical characteristics and operational requirements of the inverter, ensuring that the switching angle varies within a reasonable range. Such a constraint is essential for maintaining proper inverter functionality and achieving optimal performance.
The proposed improved algorithm is compared against several classical algorithms using a population size of 30 and 200 iterations, with each algorithm executed 30 times. The experimental results for the application of synchronous optimal pulse width modulation (SOPWM) in a three-level inverter are presented in Table 7, where the best values are underlined for clarity. As observed, DQDCS achieves the best rank in terms of the optimal value and standard deviation, ranks second in both mean and median values, and ranks fourth in the worst-case performance. Taken together, these results demonstrate that DQDCS exhibits the best overall performance in solving this problem.
g ( i ) = α i + 1 α i 10 5 > 0 , i = 1 , 2 , , N 1
h ( i ) = m i = 1 N s i cos α i = 0
0 < α i < π 2 , i = 1 , 2 , , N

5.3. Analysis of CPU Running Time for Each Algorithm

The CPU running time of an algorithm is a critical performance metric, directly affecting both the efficiency and responsiveness of a program. By analyzing time complexity, we can quantify the algorithm’s running time under various input sizes. Time complexity represents the number of CPU cycles required for the algorithm’s execution, assisting in predicting performance in large-data environments. The CPU running time is influenced not only by the algorithm itself but also by factors such as hardware architecture, compiler optimizations, and the operating environment. Practical testing and performance benchmarking enable the validation of theoretical analysis and provide a foundation for algorithm optimization.
Therefore, a comprehensive analysis of CPU running time serves as an essential tool for algorithm selection and system design. By comparing the CPU running times of different algorithms under identical conditions, the most efficient solution can be identified. As shown in Figure 9, the DQDCS algorithm ranks third in CPU running time when applied to the static pressure bearing problem. It significantly outperforms the original DCS algorithm, with shorter running times reducing hardware load, extending equipment lifespan, and lowering operational costs.
Moreover, the DQDCS algorithm also ranks third in CPU running time for solving the SOPWM (synchronous optimal pulse width modulation) problem in a three-level inverter, as depicted in Figure 10. The reduced CPU running time contributes to lower energy consumption, particularly in large-scale computations, resulting in savings in both electricity and hardware resources.
In summary, the DQDCS algorithm demonstrates excellent performance when addressing specific engineering problems, with its shorter CPU running time bringing significant economic and environmental benefits to real-world applications. These results further validate the importance of algorithm optimization and provide valuable insights for future algorithm design and enhancement. Through an in-depth analysis and comparison of CPU running times, we can select the most suitable algorithm for various computationally intensive tasks, thereby ensuring optimal resource utilization while maintaining computational efficiency.

6. Conclusions

A novel variant of the differential creative search (DCS) algorithm, termed DQDCS, is proposed to address the issue of uneven optimization performance in engineering applications. This improved version integrates a refined set-based clustering process and a double Q-learning mechanism. By leveraging a uniformly distributed initial population derived from the refined set and clustering process, the algorithm significantly reduces the risk of premature convergence to local optima in the early search phase, introducing a diverse set of individuals characterized by high randomness and non-determinism. The double Q-learning strategy is employed to effectively balance exploration and exploitation, enhancing the algorithm’s ability to escape local optima and substantially improving search efficiency and convergence accuracy.
Comparative simulation experiments were conducted between DQDCS, the original DCS, and several other state-of-the-art algorithms using both standard benchmark functions and constrained engineering optimization problems. The results demonstrate that DQDCS offers superior optimization speed and accuracy, maintaining its ability to avoid local optima even in later stages of the search. Specifically, DQDCS achieved 19 first-place rankings across the CEC2019 and CEC2022 benchmark functions, indicating robust and consistent performance. Furthermore, in the static thrust bearing design problem and the SOPWM (synchronous optimal pulse width modulation) application in three-level inverters, DQDCS consistently ranked first in overall performance, validating its effectiveness for solving real-world engineering optimization tasks.
The DQDCS algorithm involves substantial computational processes during each iteration, including population initialization, fitness evaluation, and position updating. These procedures can lead to extended computational times and increased demands on hardware resources, particularly when addressing large-scale or complex optimization problems. To enhance the efficiency of the algorithm, parallel computing strategies can be adopted. In this approach, individuals within the population are distributed across multiple processors or computational nodes, enabling concurrent execution and significantly accelerating the algorithm’s performance. While the DQDCS algorithm is primarily designed for continuous optimization problems, its application to discrete or combinatorial optimization tasks requires specific modifications. For discrete optimization problems, a feasible solution is to encode discrete variables into a continuous representation, execute the optimization within the continuous domain, and subsequently decode the solutions back to their original discrete forms. Furthermore, it is crucial to investigate and develop mutation, crossover, and other evolutionary operators explicitly adapted for discrete search spaces. Such enhancements are expected to improve the algorithm’s capability and efficiency in addressing discrete optimization challenges.
Future research will prioritize extending the application of the DQDCS algorithm to electro-hydraulic servo control and autonomous robotic arm control. Particular emphasis will be placed on the deep integration of reinforcement learning with DQDCS to further enhance its optimization capabilities. Additionally, future studies will explore the development of hybrid metaheuristic algorithms that incorporate multiple strategies or novel mathematical concepts. These improvements aim to enhance population diversity, maintain a balanced exploration-exploitation trade-off, and mitigate premature convergence, thereby improving optimization accuracy and accelerating convergence speed.
The DQDCS algorithm exhibits substantial potential in various engineering design and manufacturing fields. It can be effectively applied to mechanical structure optimization, enhancing strength, stiffness, and fatigue life while simultaneously reducing weight and minimizing costs. In the energy and power sectors, DQDCS can be employed to optimize power system generation scheduling and energy storage configurations, thereby improving overall efficiency and cost-effectiveness. Beyond these applications, DQDCS demonstrates promise in autonomous driving and robotics, where it can facilitate real-time path planning and trajectory optimization, ultimately enhancing safety and operational efficiency.
Despite its promising applications, DQDCS still faces certain limitations, such as computational inefficiencies in high-dimensional problems and local convergence challenges in complex constrained optimization tasks. To address these issues, future work will focus on refining the algorithm’s structure and parameter settings to enhance its solution efficiency and accuracy in high-dimensional, complex scenarios. Additionally, research will investigate more effective constraint-handling mechanisms to improve DQDCS’s global convergence ability and robustness in solving complex constrained optimization problems. These advancements aim to elevate the algorithm’s performance, making it more adaptable and resilient in real-world engineering applications.

Author Contributions

Y.Z. analyzed the data and wrote the paper. L.Y. and Y.C. revised the manuscript. J.Y. and M.Z. designed the study. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Natural Science Foundation of Shaanxi Province (2024JC-YBMS-014), the Foundation of Shaanxi University of Technology (SLGNL202409), the Foundation of Shaanxi University of Technology (SLGYCX2537).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data used to support the findings of this study are available from the corresponding author upon request.

Conflicts of Interest

The authors declare that there are no conflicts of interest regarding the publication of this paper.

References

  1. Berahas, A.S.; Bollapragada, R.; Nocedal, J. An investigation of Newton-sketch and subsampled Newton methods. Optim. Methods Softw. 2020, 35, 661–680. [Google Scholar] [CrossRef]
  2. Li, M. Some New Descent Nonlinear Conjugate Gradient Methods for Unconstrained Optimization Problems with Global Convergence. Asia-Pac. J. Oper. Res. 2024, 41, 2350020. [Google Scholar] [CrossRef]
  3. Upadhyay, B.B.; Pandey, R.K.; Liao, S. Newton’s method for interval-valued multiobjective optimization problem. J. Ind. Manag. Optim. 2024, 20, 1633–1661. [Google Scholar] [CrossRef]
  4. Altay, E.V.; Alatas, B. Intelligent optimization algorithms for the problem of mining numerical association rules. Phys. A Stat. Mech. Its Appl. 2020, 540, 123142. [Google Scholar] [CrossRef]
  5. Dai, C.; Ma, L.; Jiang, H.; Li, H. An improved whale optimization algorithm based on multiple strategies. Comput. Eng. Sci. 2024, 46, 1635. [Google Scholar]
  6. Katoch, S.; Chauhan, S.S.; Kumar, V. A review on genetic algorithm: Past, present, and future. Multimed. Tools Appl. 2021, 80, 8091–8126. [Google Scholar] [CrossRef]
  7. Shami, T.M.; El-Saleh, A.A.; Alswaitti, M.; Al-Tashi, Q.; Summakieh, M.A.; Mirjalili, S. Particle swarm optimization: A comprehensive survey. IEEE Access 2022, 10, 10031–10061. [Google Scholar] [CrossRef]
  8. Chaudhari, B.; Gulve, A. An edge detection method based on Ant Colony System for medical images. J. Integr. Sci. Technol. 2025, 13, 1073. [Google Scholar] [CrossRef]
  9. Dong, Y.; Sun, C.; Han, Y.; Liu, Q. Intelligent optimization: A novel framework to automatize multi-objective optimization of building daylighting and energy performances. J. Build. Eng. 2021, 43, 102804. [Google Scholar] [CrossRef]
  10. Li, W.; Wang, G.-G.; Gandomi, A.H. A survey of learning-based intelligent optimization algorithms. Arch. Comput. Methods Eng. 2021, 28, 3781–3799. [Google Scholar] [CrossRef]
  11. Kamal, M.; Tariq, M.; Srivastava, G.; Malina, L. Optimized security algorithms for intelligent and autonomous vehicular transportation systems. IEEE Trans. Intell. Transp. Syst. 2021, 24, 2038–2044. [Google Scholar] [CrossRef]
  12. Arora, S.; Singh, S. Butterfly optimization algorithm: A novel approach for global optimization. Soft. Comput. 2019, 23, 715–734. [Google Scholar] [CrossRef]
  13. Mirjalili, S.; Gandomi, A.H.; Mirjalili, S.Z.; Saremi, S.; Faris, H.; Mirjalili, S.M. Salp Swarm Algorithm: A bio-inspired optimizer for engineering design problems. Adv. Eng. Softw. 2017, 114, 163–191. [Google Scholar] [CrossRef]
  14. Houssein, E.H.; Emam, M.M.; Ali, A.A. Improved manta ray foraging optimization for multi-level thresholding using COVID-19 CT images. Neural Comput. Appl. 2021, 33, 16899–16919. [Google Scholar] [CrossRef]
  15. Alkayem, N.F.; Shen, L.; Mayya, A.; Asteris, P.G.; Fu, R.; Di Luzio, G.; Cao, M. Prediction of concrete and FRC properties at high temperature using machine and deep learning: A review of recent advances and future perspectives. J. Build. Eng. 2024, 83, 108369. [Google Scholar] [CrossRef]
  16. Abdollahzadeh, B.; Khodadadi, N.; Barshandeh, S.; Trojovský, P.; Gharehchopogh, F.S.; El-kenawy, E.S.M.; Abualigah, L.; Mirjalili, S. Puma optimizer (PO): A novel metaheuristic optimization algorithm and its application in machine learning. Clust. Comput. 2024, 27, 5235–5283. [Google Scholar] [CrossRef]
  17. Liu, H.; Zhang, X.; Zhang, H.; Li, C.; Chen, Z. A reinforcement learning-based hybrid Aquila Optimizer and improved Arithmetic Optimization Algorithm for global optimization. Expert Syst. Appl. 2023, 224, 119898. [Google Scholar] [CrossRef]
  18. Wang, H.; Sun, S.; Chang, L.; Li, H.; Zhang, W.; Frery, A.C.; Ren, P. INSPIRATION: A reinforcement learning-based human visual perception-driven image enhancement paradigm for underwater scenes. Eng. Appl. Artif. Intell. 2024, 133, 108411. [Google Scholar] [CrossRef]
  19. Li, X.; Sun, W.; Ji, Y.; Huang, W. A Joint Detection and Tracking Paradigm Based on Reinforcement Learning for Compact HFSWR. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2024, 18, 1995–2009. [Google Scholar] [CrossRef]
  20. Zhu, X.; Zhang, J.; Jia, C.; Liu, Y.; Fu, M. A Hybrid Black-Winged Kite Algorithm with PSO and Differential Mutation for Superior Global Optimization and Engineering Applications. Biomimetics 2024, 10, 236. [Google Scholar] [CrossRef]
  21. Duankhan, P.; Sunat, K.; Chiewchanwattana, S.; Nasa-Ngium, P. The Differentiated Creative search (DCS): Leveraging Differentiated knowledge-acquisition and Creative realism to address complex optimization problems. Expert Syst. Appl. 2024, 252, 123734. [Google Scholar] [CrossRef]
  22. Liu, J.; Lin, Y.; Zhang, X.; Yin, J.; Zhang, X.; Feng, Y.; Qian, Q. Agricultural UAV path planning based on a differentiated creative search algorithm with multi-strategy improvement. Machines 2024, 12, 591. [Google Scholar] [CrossRef]
  23. Cai, X.; Zhang, C. An Innovative Differentiated Creative Search Based on Collaborative Development and Population Evaluation. Biomimetics 2025, 10, 260. [Google Scholar] [CrossRef] [PubMed]
  24. Chakraborty, S.; Saha, A.K.; Sharma, S.; Mirjalili, S.; Chakraborty, R. A novel enhanced whale optimization algorithm for global optimization. Comput. Ind. Eng. 2021, 153, 107086. [Google Scholar] [CrossRef]
  25. Cai, X.; Wang, W.; Wang, Y. Multi-strategy enterprise development optimizer for numerical optimization and constrained problems. Sci. Rep. 2025, 15, 10538. [Google Scholar] [CrossRef]
  26. Shehadeh, H.A. Chernobyl disaster optimizer (CDO): A novel meta-heuristic method for global optimization. Neural Comput. Appl. 2023, 35, 10733–10749. [Google Scholar] [CrossRef]
  27. Ouyang, C.; Qiu, Y.; Zhu, D. Adaptive spiral flying sparrow search algorithm. Sci. Program. 2021, 2021, 6505253. [Google Scholar] [CrossRef]
  28. Abdelhamid, A.A.; Towfek, S.K.; Khodadadi, N.; Alhussan, A.A.; Khafaga, D.S.; Eid, M.M.; Ibrahim, A. Waterwheel plant algorithm: A novel metaheuristic optimization method. Processes 2023, 11, 1502. [Google Scholar] [CrossRef]
  29. Yu, X.; Duan, Y.; Cai, Z. Sub-population improved grey wolf optimizer with Gaussian mutation and Lévy flight for parameters identification of photovoltaic models. Expert Syst. Appl. 2023, 232, 120827. [Google Scholar] [CrossRef]
  30. Xue, J.; Shen, B. Dung beetle optimizer: A new meta-heuristic algorithm for global optimization. J. Supercomput. 2023, 79, 7305–7336. [Google Scholar] [CrossRef]
  31. Wu, L.; Xu, D.; Guo, Q.; Chen, E.; Xiao, W. A nonlinear randomly reuse-based mutated whale optimization algorithm and its application for solving engineering problems. Appl. Soft Comput. 2024, 167, 112271. [Google Scholar] [CrossRef]
  32. Punia, P.; Raj, A.; Kumar, P. Enhanced zebra optimization algorithm for reliability redundancy allocation and engineering optimization problems. Clust. Comput. 2025, 28, 267. [Google Scholar] [CrossRef]
  33. Guo, L.; Gu, W. PMSOMA: Optical microscope algorithm based on piecewise linear chaotic map** and sparse adaptive exploration. Sci. Rep. 2024, 14, 20849. [Google Scholar] [CrossRef] [PubMed]
  34. Liu, J.; Feng, J.; Yang, S.; Zhang, H.; Liu, S. Dynamic ε-multilevel hierarchy constraint optimization with adaptive boundary constraint handling technology. Appl. Soft Comput. 2024, 152, 111172. [Google Scholar] [CrossRef]
Figure 1. Operation process of double Q-learning.
Figure 1. Operation process of double Q-learning.
Biomimetics 10 00356 g001
Figure 2. DQDCS flowchart.
Figure 2. DQDCS flowchart.
Biomimetics 10 00356 g002
Figure 3. Convergence curve of CEC2019 benchmark tests.
Figure 3. Convergence curve of CEC2019 benchmark tests.
Biomimetics 10 00356 g003aBiomimetics 10 00356 g003b
Figure 4. Boxplots of CEC2019 benchmark functions. “+” is the abnormal data in the box plot.
Figure 4. Boxplots of CEC2019 benchmark functions. “+” is the abnormal data in the box plot.
Biomimetics 10 00356 g004aBiomimetics 10 00356 g004b
Figure 5. Convergence curve of CEC2022 benchmark tests.
Figure 5. Convergence curve of CEC2022 benchmark tests.
Biomimetics 10 00356 g005aBiomimetics 10 00356 g005b
Figure 6. Boxplots of CEC2022 benchmark functions.
Figure 6. Boxplots of CEC2022 benchmark functions.
Biomimetics 10 00356 g006aBiomimetics 10 00356 g006b
Figure 7. Static pressure thrust bearing. Arrows indicate the flow direction or movement direction.
Figure 7. Static pressure thrust bearing. Arrows indicate the flow direction or movement direction.
Biomimetics 10 00356 g007
Figure 8. SOPWM for 3-level inverters. The arrow part represents the signal flow direction in the SOPWM schematic diagram.
Figure 8. SOPWM for 3-level inverters. The arrow part represents the signal flow direction in the SOPWM schematic diagram.
Biomimetics 10 00356 g008
Figure 9. Comparison charts of CPU running times of various algorithms for static pressure thrust bearing.
Figure 9. Comparison charts of CPU running times of various algorithms for static pressure thrust bearing.
Biomimetics 10 00356 g009
Figure 10. Comparison charts of CPU running times of various algorithms for SOPWM for 3-level inverters.
Figure 10. Comparison charts of CPU running times of various algorithms for SOPWM for 3-level inverters.
Biomimetics 10 00356 g010
Table 1. Ablation experiment results for each strategy.
Table 1. Ablation experiment results for each strategy.
AlgorithmBestMeanStd
F2DCS3.432233.440148.401
D13.172510.559518.8376
D23.172315.738319.5717
D33.152315.738319.5717
D43.151732.832949.4357
F5DCS1.02971.07930.047848
D11.03571.19970.1064
D21.03651.20230.1112
D31.01981.20190.103
D41.00081.05770.034179
F8DCS3.04233.53220.56543
D12.89244.450.55146
D22.39742.85180.3392
D32.02332.87580.40339
D42.01292.83640.3775
F10DCS7.717820.06273.5227
D16.976518.31285.2696
D25.354719.65284.7076
D32.793719.5355.1848
D41.000119.99294.4706
Table 2. Experimental results on the CEC2019 benchmark functions. The underlines are used to highlight the optimal average values.
Table 2. Experimental results on the CEC2019 benchmark functions. The underlines are used to highlight the optimal average values.
FunctionAlgorithmBestMeanStd
F1DCS11.17460.066539
DQDCS11.01290.017456
MSDCS1.000154.4005138.3435
CDO11.20
Puma1103,218.6127227,810.8443
WWPA1.6834238.3049583.4071
SPGWO13656.33918551.4486
DBO1414,533.6708721,175.8267
NRRMWOA180.8782785,017.93781,311,767.1446
SABO15.774121.3505
ASFSSA110
F2DCS3.432233.440148.401
DQDCS3.151732.832949.4357
MSDCS5.11329.67974.1544
CDO550
Puma4.23284.70360.37262
WWPA5.20018.31473.6477
SPGWO63.7169259.0299142.4842
DBO4.2752384.2964194.8378
NRRMWOA11.9124721.1422965.2022
SABO4.59938.78575.9168
ASFSSA4.21894.32890.18135
F3DCS2.24242.93710.4211
DQDCS1.00042.04390.2884
MSDCS11.726912.40150.34845
CDO4.70965.86120.64681
Puma1.46562.14040.57859
WWPA8.068810.14370.86708
SPGWO2.45043.24250.74601
DBO1.40913.97661.8562
NRRMWOA1.41064.98931.953
SABO5.73876.90260.57406
ASFSSA1.01343.64141.5732
F4DCS5.18487.25521.0675
DQDCS5.015.59511.8712
MSDCS85.6623136.802921.366
CDO63.543271.43575.479
Puma5.974813.04095.7055
WWPA116.9494142.256213.0633
SPGWO4.002114.89379.5953
DBO11.096523.74728.109
NRRMWOA17.031143.54318.719
SABO35.163545.19177.9818
ASFSSA7.986541.07129.6237
F5DCS1.02971.07930.047848
DQDCS1.00081.05770.034179
MSDCS76.1748153.011342.5413
CDO53.721972.89354.7508
Puma1.02711.16040.096242
WWPA122.0759173.334521.0507
SPGWO1.17231.52550.22175
DBO1.04421.14620.06739
NRRMWOA1.27851.52290.17431
SABO1.75722.94260.93688
ASFSSA1.06151.16460.072103
F6DCS1.00041.93021.0947
DQDCS1.00291.46850.73351
MSDCS9.624914.18521.4021
CDO7.81889.44170.90669
Puma1.0041.75170.89464
WWPA11.264612.86440.87031
SPGWO1.2812.14440.92531
DBO2.05464.61421.7611
NRRMWOA4.95137.53651.5727
SABO2.84284.7480.97675
ASFSSA1.00092.64291.1644
F7DCS119.6257591.2146214.4368
DQDCS80.447347.2542160.8203
MSDCS1910.9472563.8317271.9363
CDO1174.98171493.4973180.7765
Puma126.3932605.3863286.0483
WWPA2117.42312407.1589144.384
SPGWO342.6493736.2504217.0422
DBO417.6764786.1656309.1151
NRRMWOA499.52931150.2235299.2043
SABO1214.08371760.7036189.8127
ASFSSA293.3793800.6022282.4334
F8DCS3.04233.53220.56543
DQDCS2.4473.3650.022699
MSDCS4.6545.25710.0764
CDO3.7574.20630.20053
Puma2.43073.61690.41838
WWPA4.95535.24670.10746
SPGWO1.32783.40540.52501
DBO2.90843.87750.4717
NRRMWOA3.50714.37470.35839
SABO3.88484.53320.23968
ASFSSA3.13814.07520.33229
F9DCS1.04631.12880.052384
DQDCS1.00671.04260.040301
MSDCS3.88635.33930.61068
CDO3.65484.18620.14797
Puma1.08231.16720.051977
WWPA4.40265.23770.37346
SPGWO1.0731.15430.040058
DBO1.17221.27690.08308
NRRMWOA1.18241.37980.15444
SABO1.12241.28980.066172
ASFSSA1.09161.18010.075728
F10DCS1.004420.23724.5283
DQDCS1.000119.99294.4706
MSDCS21.649421.64990.0010212
CDO21.273121.40230.061844
Puma20.961521.01330.027353
WWPA21.264121.69660.13883
SPGWO21.231521.39230.081047
DBO2121.23870.18582
NRRMWOA20.99621.05390.086024
SABO20.894721.33330.15323
Table 3. Wilcoxon rank-sum test results for CEC2019 benchmark.
Table 3. Wilcoxon rank-sum test results for CEC2019 benchmark.
DQDCS vs.DCSMSDCSCDOPumaWWPASPGWODBONRRMWOASABOASFSSA
F16.39 × 10−51.83 × 10−41.82 × 10−41.72 × 10−41.83 × 10−41.13 × 10−24.45 × 10−21.83 × 10−41.83 × 10−46.35 × 10−5
F26.39 × 10−53.30 × 10−41.83 × 10−41.83 × 10−41.83 × 10−41.40 × 10−16.40 × 10−21.83 × 10−42.46 × 10−48.75 × 10−5
F32.11 × 10−21.83 × 10−47.30 × 10−35.83 × 10−41.83 × 10−42.80 × 10−32.57 × 10−25.83 × 10−45.80 × 10−31.73 × 10−2
F41.13 × 10−21.83 × 10−41.83 × 10−42.12 × 10−21.83 × 10−42.11 × 10−24.31 × 10−12.20 × 10−31.83 × 10−41.01 × 10−3
F56.40 × 10−31.83 × 10−41.83 × 10−41.83 × 10−41.83 × 10−41.83 × 10−41.40 × 10−21.13 × 10−21.83 × 10−41.83 × 10−4
F61.83 × 10−44.40 × 10−41.83 × 10−41.83 × 10−47.69 × 10−41.83 × 10−41.73 × 10−22.20 × 10−31.83 × 10−41.83 × 10−4
F72.46 × 10−41.83 × 10−41.83 × 10−43.76 × 10−21.83 × 10−47.69 × 10−44.52 × 10−25.80 × 10−31.83 × 10−45.80 × 10−3
F81.73 × 10−41.83 × 10−43.76 × 10−24.52 × 10−21.82 × 10−41.40 × 10−24.40 × 10−42.57 × 10−21.83 × 10−47.69 × 10−4
F97.75 × 10−31.83 × 10−41.83 × 10−45.80 × 10−31.83 × 10−43.39 × 10−22.57 × 10−24.52 × 10−27.69 × 10−42.20 × 10−3
F102.43 × 10−21.83 × 10−41.83 × 10−41.71 × 10−31.83 × 10−41.83 × 10−42.83 × 10−34.59 × 10−34.40 × 10−47.76 × 10−4
Table 4. Testing results of CEC2022 benchmark functions. The underline is the average value of the optimal performance in the operation test function and is used to highlight the optimal average value.
Table 4. Testing results of CEC2022 benchmark functions. The underline is the average value of the optimal performance in the operation test function and is used to highlight the optimal average value.
FunctionAlgorithmBestMeanStd
F1DCS3003013.8519 × 10−13
DQDCS3003006.8317 × 10−14
MSDCS10,766.750310,766.75710.013964
CDO7490.196420,512.767313,607.6136
Puma300.0023301.27112.5835
WWPA8455.8816236,150.71161,491,237.9285
SPGWO304.8303797.50611028.1107
DBO300361.9946341.4715
NRRMWOA358.81573032.31312230.1134
SABO1195.78143379.00631311.4349
ASFSSA309.4622586.0842281.7238
F2DCS400405.52753.4253
DQDCS400.0035403.4253.6114
MSDCS750.17372491.5488960.6972
CDO570.9415835.285949.4592
Puma400406.39437.2844
WWPA972.56713270.42141414.9336
SPGWO400.5352415.789614.3696
DBO400.0218424.242729.16
NRRMWOA400.0781422.051827.9517
SABO404.4744448.468221.9722
ASFSSA400.0102413.440419.3205
F3DCS6006003.5703 × 10−7
DQDCS600.0001600.00053.2297 × 10−7
MSDCS643.6113682.755110.18
CDO627.5045634.87253.539
Puma6006000.00026748
WWPA662.8221687.37787.5308
SPGWO600.0541600.49670.68095
DBO600601.77392.6285
NRRMWOA601.6986621.400811.8169
SABO603.078612.02947.4192
ASFSSA600602.29586.5591
F4DCS803.5611809.18242.9573
DQDCS801.7438809.08591.0729
MSDCS867.3885896.9954.0671
CDO828.5143845.47716.2154
Puma806.9647818.3576.8807
WWPA864.4186887.32758.2043
SPGWO800.3831800.99996.6974
DBO807.9597829.886610.517
NRRMWOA809.95835.823814.4457
SABO819.7184838.02768.1862
ASFSSA811.9395829.61985.4414
F5DCS9009002.2852 × 10−14
DQDCS9009002.0549 × 10−14
MSDCS2066.30812246.889121.1599
CDO1248.37161377.543469.9206
Puma900900.38610.57066
WWPA1860.82092380.0621185.2868
SPGWO900900.004210.5999
DBO900916.412157.1606
NRRMWOA906.3261253.0823278.8837
SABO901.1913925.924116.9796
ASFSSA901.72821398.8497174.1701
F6DCS1800.03681805.69471.5327
DQDCS1800.02381801.9392.0455
MSDCS70,252,227.0095207,374,163.459333,963,697.1425
CDO15,167,928.0263157,093,751.8526240,080,753.4813
Puma1807.92352063.024734.8686
WWPA8,608,986.8261185,293,465.081573,329,210.9276
SPGWO1960.64645799.18612292.3454
DBO1895.14124731.54192375.8384
NRRMWOA1846.68693977.14992076.5281
SABO2459.448920,110.622811,920.7947
ASFSSA1930.12485432.43661903.6817
F7DCS2027.9742049.277714.7551
DQDCS2000.2292001.63852.837
MSDCS2337.86972447.373869.1571
CDO2231.40562300.05230.0378
Puma2151.46212192.527243.688
WWPA2306.81392451.107390.2488
SPGWO2034.34982057.403225.2011
DBO2032.34332090.274934.7803
NRRMWOA2101.60252198.902460.3948
SABO2105.66822178.991837.0986
ASFSSA2030.95442094.598529.2648
F8DCS2222.02122225.73485.0729
DQDCS2200.71862208.87647.0158
MSDCS2846.84523952.59941101.3132
CDO2243.6442251.9466.6606
Puma2226.32812353.5188125.6302
WWPA2551.54222917.5036244.8562
SPGWO2224.44042229.96774.8594
DBO2233.75822301.607663.7515
NRRMWOA2237.56652262.52634.2888
SABO2278.16332356.111964.2682
ASFSSA2222.18352227.17964.6573
F9DCS2529.28442529.28440
DQDCS2480.78212480.29420.027562
MSDCS3371.80153996.9709412.0453
CDO3151.46523426.8386126.0153
Puma2480.79762480.82020.017425
WWPA3370.77734492.3652979.6033
SPGWO2481.18052500.771622.002
DBO2480.91252496.643120.4099
NRRMWOA2481.14862491.646914.7268
SABO2603.38142699.623946.6498
ASFSSA2480.78132480.80640.067892
F10DCS2500.34382515.185646.6588
DQDCS2500.15422503.50418.4699
MSDCS6937.00147963.4014476.9702
CDO4832.64145873.0597506.4387
Puma2500.6382515.046845.15
WWPA7228.17727552.068210.5838
SPGWO2500.50733312.8457742.2508
DBO2500.81572930.6677678.0253
NRRMWOA2501.31944062.99871135.7437
SABO2858.035556.15441635.5186
ASFSSA2500.73842630.7374410.1658
F11DCS26002646.5043108.0761
DQDCS26002639103.2576
MSDCS5105.14665105.17690.082861
CDO3329.12243343.27815.6324
Puma2600.00012637.3103118.4636
WWPA3712.74144917.2808262.7002
SPGWO2601.18842904.431130.4169
DBO26002814.9987171.69
NRRMWOA2600.6982901.0184130.9309
SABO2832.49723233.2798104.4736
ASFSSA26002663.175113.4852
F12DCS2988.26993219.0423181.9443
DQDCS2700.61862722.03632.1318
MSDCS3624.33464070.0287301.5474
CDO3478.46913508.670525.8794
Puma2939.25422951.624213.3077
WWPA2900.00482900.0055.3612 × 10−5
SPGWO2937.21742955.925613.3377
DBO2939.84992973.734740.5269
NRRMWOA2958.05793044.933670.3033
SABO2994.78043054.333838.4792
ASFSSA2945.59172962.468113.8506
Table 5. Wilcoxon rank-sum test results for CEC2022 benchmark functions.
Table 5. Wilcoxon rank-sum test results for CEC2022 benchmark functions.
DQDCS vs.DCSMSDCSCDOPumaWWPASPGWODBONRRMWOASABOASFSSA
F14.40 × 10−41.83 × 10−41.83 × 10−41.83 × 10−41.83 × 10−47.69 × 10−41.83 × 10−44.40 × 10−41.83 × 10−41.83 × 10−4
F23.61 × 10−31.83 × 10−41.83 × 10−41.83 × 10−41.83 × 10−41.83 × 10−48.90 × 10−31.83 × 10−41.83 × 10−41.83 × 10−4
F32.83 × 10−31.83 × 10−41.83 × 10−41.83 × 10−41.83 × 10−42.20 × 10−31.83 × 10−41.83 × 10−41.83 × 10−41.83 × 10−4
F49.11 × 10−31.83 × 10−41.83 × 10−44.52 × 10−21.83 × 10−41.13 × 10−22.20 × 10−38.90 × 10−31.83 × 10−45.21 × 10−3
F52.57 × 10−21.83 × 10−41.83 × 10−41.83 × 10−41.83 × 10−41.01 × 10−31.83 × 10−41.83 × 10−41.83 × 10−41.83 × 10−4
F66.23 × 10−21.83 × 10−41.83 × 10−41.83 × 10−41.83 × 10−41.73 × 10−21.40 × 10−22.20 × 10−31.83 × 10−44.40 × 10−4
F73.12 × 10−21.83 × 10−41.83 × 10−41.83 × 10−41.83 × 10−43.34 × 10−22.83 × 10−31.83 × 10−41.83 × 10−46.40 × 10−2
F83.12 × 10−21.83 × 10−41.83 × 10−44.40 × 10−41.83 × 10−42.20 × 10−31.04 × 10−21.83 × 10−41.83 × 10−44.40 × 10−4
F91.83 × 10−41.83 × 10−41.83 × 10−41.83 × 10−41.83 × 10−41.83 × 10−42.80 × 10−31.83 × 10−41.83 × 10−41.83 × 10−4
F101.71 × 10−21.83 × 10−41.83 × 10−42.21 × 10−31.83 × 10−41.73 × 10−29.11 × 10−31.83 × 10−41.83 × 10−43.09 × 10−2
F114.52 × 10−21.83 × 10−41.83 × 10−41.83 × 10−41.83 × 10−41.40 × 10−22.20 × 10−31.83 × 10−41.83 × 10−41.83 × 10−4
F121.73 × 10−21.83 × 10−41.83 × 10−41.01 × 10−31.83 × 10−47.69 × 10−41.01 × 10−32.46 × 10−41.83 × 10−42.46 × 10−4
Table 6. Optimization results of static pressure thrust bearing. The underline is the average value of the optimal performance in the operation test function and is used to highlight the optimal average value.
Table 6. Optimization results of static pressure thrust bearing. The underline is the average value of the optimal performance in the operation test function and is used to highlight the optimal average value.
Namex1x2x3x4WorstBestStdMeanMedian
DCS1.59727828
× 10−5
1.0002377514.791535515.99424421.06923430
× 109
1.07206086 × 1096.61825327 × 1051.07014876
× 109
1.06997430 × 109
DQDCS1.60000000
× 10−5
1.0000000014.800168916.00000001.06894753
× 109
1.06954809 × 1091.34312464 × 1051.06898081
× 109
1.06894753 × 109
MSDCS1.17291046
× 10−5
1.791309425.7027400415.94131011.40136705
× 109
5.22419187 × 10141.19485993 × 10143.57212782
× 1013
1.17985612 × 1011
Puma1.60000000
× 10−5
1.0000000014.800347416.00000001.06896145
× 109
1.12825463 × 1091.37399596 × 1071.07613384
× 109
1.07044585 × 109
CDO1.60000000
× 10−5
1.0000000014.816361716.00000001.07047191
× 109
1.15997503 × 1092.15231379 × 1071.09774625
× 109
1.09665275 × 109
WWPA0.473396391.60123665
× 106
3.94298347
× 105
3.91671659
× 105
−2.22716152
× 1011
2.35796666 × 10105.37550187 × 10101.52405939
× 1010
1.15527859 × 109
SPGWO1.60000000
× 10−5
1.0000000014.80124441.600000001.06904448
× 106
1.32319148 × 1095.65555937 × 1071.08297369
× 109
1.07012621 × 109
DBO1.60000000
× 10−5
1.0000000014.800168916.00000001.06894753
× 109
1.47869424 × 1099.15951602 × 1071.08955220
× 109
1.06894753 × 109
NRRMWOA1.60000000
× 10−5
1.0000000014.800157216.00000001.06894764
× 109
1.10966892 × 1091.11408274 × 1071.07630514
× 109
1.07166285 × 109
SABO1.60000000
× 10−5
1.0000000014.814560916.00000001.07029905
× 109
1.49353781 × 1091.44018113 × 1081.18985405
× 109
1.12582885 × 109
ASFSSA1.60000000
× 10−5
1.0000000014.800244216.00000001.06895191
× 109
1.07108086 × 1096.86835160 × 1051.06948901
× 109
1.06929758 × 109
Table 7. Optimization results of SOPWM for 3-level inverters. The underline is the average value of the optimal performance in the operation test function and is used to highlight the optimal average value.
Table 7. Optimization results of SOPWM for 3-level inverters. The underline is the average value of the optimal performance in the operation test function and is used to highlight the optimal average value.
Namex1x2x3x4WorstBestStdMeanMedian
DCS0.5569908201.556919291.556322641.570690230.9957098961.002728981.67135890
× 10−3
0.9983009570.998148422
DQDCS0.5564231391.570796331.570375211.570796330.9954051320.9954404607.97475504
× 10−6
0.9954085500.995405849
MSDCS0.3988186491.417582511.491399571.549233551.61761236105.74127030.133819927.684256218.69082921
Puma0.5564146031.570796331.570378181.570796330.9954051290.9955808965.19507407
× 10−5
0.9954307610.995406675
CDO0.555834510011.570796331.570796331.570796330.9955930451.140821886.15210384
× 10−2
1.051101861.00665276
WWPA2.85068113
× 10−3
1.48395345
× 10−4
1.5715065
× 10−4
7.21160329
× 104
1.4889583578.043126418.217723911.93109684.09729542
SPGWO0.5567143111.570796331.570277341.570796330.9954183491.167559466.26457711
× 10−2
1.028497980.995582535
DBO0.5564183171.570796331.570377501.570796330.9954051271.167556793.97313485
× 10−2
1.006878100.995580896
NRRMWOA0.5564161131.570796331.570378741.570796330.9954051281.167557725.29815832
× 10−2
1.012637610.995405148
SABO0.5556606481.570796331.570796331.570796330.9955812360.9964587352.43692395
× 10−4
0.9957267610.995624926
ASFSSA0.5564183231.570796331.570377501.570796330.9954051271.206654757.59187361
× 10−2
1.041514730.995508388
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, Y.; Yong, L.; Chen, Y.; Yang, J.; Zhang, M. A Hybrid Multi-Strategy Differential Creative Search Optimization Algorithm and Its Applications. Biomimetics 2025, 10, 356. https://doi.org/10.3390/biomimetics10060356

AMA Style

Zhang Y, Yong L, Chen Y, Yang J, Zhang M. A Hybrid Multi-Strategy Differential Creative Search Optimization Algorithm and Its Applications. Biomimetics. 2025; 10(6):356. https://doi.org/10.3390/biomimetics10060356

Chicago/Turabian Style

Zhang, Yuanyuan, Longquan Yong, Yijia Chen, Jintao Yang, and Mengnan Zhang. 2025. "A Hybrid Multi-Strategy Differential Creative Search Optimization Algorithm and Its Applications" Biomimetics 10, no. 6: 356. https://doi.org/10.3390/biomimetics10060356

APA Style

Zhang, Y., Yong, L., Chen, Y., Yang, J., & Zhang, M. (2025). A Hybrid Multi-Strategy Differential Creative Search Optimization Algorithm and Its Applications. Biomimetics, 10(6), 356. https://doi.org/10.3390/biomimetics10060356

Article Metrics

Back to TopTop