You are currently viewing a new version of our website. To view the old version click .
Mathematics
  • Article
  • Open Access

11 November 2025

Reinforcement Learning Enhanced Multi-Objective Social Network Search Algorithm for Engineering Design Problems

,
,
and
1
School of Mechanical and Automotive Engineering, South China University of Technology, Guangzhou 510640, China
2
School of Aeronautical Manufacturing and Mechanical Engineering, Nanchang Hangkong University, Nanchang 330063, China
*
Author to whom correspondence should be addressed.

Abstract

To address real-world engineering design optimization problems, this study proposes a reinforcement learning enhanced multi-objective social network search algorithm (QMOSNS), which represents a novel approach for solving multi-objective optimization problems. QMOSNS utilizes Halton sequences for population initialization to enhance the diversity of the initial population. A multi-objective archive mechanism is implemented to store Pareto-optimal solutions and select parental individuals through a reassigned fitness evaluation strategy. Furthermore, Q-learning is incorporated to adaptively select mutation operators, thereby dynamically balancing the algorithm’s exploration and exploitation capabilities. QMOSNS was rigorously evaluated through 50 prominent case studies, including 22 unconstrained multi-objective benchmark problems, 18 constrained multi-objective benchmark problems, and 10 multi-objective engineering design problems, to comprehensively validate its computational capabilities and effectiveness. Moreover, statistical results obtained using consistent performance metrics were compared with those of other highly regarded algorithms to ensure a fair and objective performance assessment. The comparative results show that QMOSNS is robust and superior in handling a wide variety of multi-objective problems. This study underscores the efficacy of integrating reinforcement learning with social intelligence for tackling complex multi-objective optimization in engineering and computational domains.

1. Introduction

Multi-objective optimization (MOO) is a subfield of multi-criteria decision making that deals with mathematical problems involving the simultaneous optimization of multiple objective functions. Unlike single-objective optimization problems that seek a single optimal solution based on a single criterion, MOO requires considering and balancing two or more conflicting objectives []. MOO is of significant practical importance in real-world applications, spanning disciplines such as economics, science, and engineering [,,]. It provides a systematic framework to support decision-makers in navigating complex trade-off environments and identifying optimal solutions that satisfy multiple competing requirements. The task of MOO algorithms is to find the best trade-off solution set among multiple objectives, known as the Pareto optimal set [].
Generally, the use of stochastic optimization algorithms to address multi-objective optimization problems (MOPs) can be categorized into two main classes: priori and posterior []. A priori method primarily involves transforming MOPs into new single-objective optimization problems through a set of weighting coefficients. Single-objective optimization algorithms can then readily solve this transformed problem after objective aggregation. Typically, domain experts determine these weights based on empirical knowledge, adjusting the weighting of each objective according to their relative importance []. However, this approach has notable limitations: Firstly, weight determination often relies on the decision-maker’s subjective judgment, lacking unified objective criteria. Different decision-makers varying perceptions of objective importance may lead to inconsistent weight assignments. Furthermore, its weighted aggregation approach makes it difficult to optimize all objectives at once. Consequently, a single run cannot obtain the complete Pareto-optimal set []. In a posteriori approach, multiple objectives are simultaneously preserved and optimized, allowing for the acquisition of the complete Pareto optimal solution set in a single optimization run. The Pareto optimal set provides decision-makers with a wide range of options, enabling them to make informed choices among multiple solutions based on specific requirements []. This highlights the importance of solution diversity in MOO. Moreover, this approach can effectively handle any type of Pareto front. The algorithms based on a posteriori method are highly prevalent in the literature [,].
Moreover, posterior methods can be classified into three major categories: Multi-Objective Evolutionary Algorithms (MOEA) and Multi-Objective Swarm Algorithms (MOSA). In 1984, David Schaffer introduced the Vector Evaluated Genetic Algorithm (VEGA) by extending the original Genetic Algorithm (GA) to multi-objective problems []. This work marked the inception of evolutionary algorithms in multi-objective optimization contexts, and VEGA is widely recognized as the first Pareto-based MOEA. The core principle underlying most MOEAs is non-dominated sorting (NDS), which was adopted by the original Non-dominated Sorting Genetic Algorithm (NSGA) []. However, NSGA suffered from several limitations, including the absence of an explicit sharing parameter, lack of elitism consideration, and high computational cost of non-dominated sorting. To address these issues, its successor Non-dominated Sorting Genetic Algorithm II (NSGA-II) introduced three key innovations: a fast non-dominated sorting approach, a diversity preservation mechanism, and a crowded-comparison operator []. These improvements established NSGA-II as one of the most widely used and influential MOEAs globally. Multi-objective Evolutionary Algorithm Based on Decomposition (MOEA/D) represent another significant advancement in the field of MOEA []. This class of algorithms breaks down a multi-objective problem into separate subproblems. It then uses aggregation strategies to share information between neighboring subproblems, which optimizes computational efficiency. However, such methods may fail when addressing problems with discontinuous Pareto fronts and exhibit certain limitations in solving complex MOPs. In the domain of MOSAs, Coello Coello & Lechuga proposed the Multi-Objective Particle Swarm Optimization (MOPSO) algorithm by extending the conventional Particle Swarm Optimization (PSO) framework []. The approach incorporates mechanisms such as multi-objective function handling, mutation strategies, and a grid-based technique to enhance the distribution of solutions along the Pareto front. MOSAs represent a class of optimization techniques inspired by collective behaviors observed in human societies, animal populations, and physical phenomena []. Since the introduction of MOPSO, numerous novels and effective MOSAs have been continuously developed. Notable examples include the Multi-objective Grey Wolf Optimizer (MOGWO) [], Multi-objective Water Cycle Algorithm (MOWCA) [], Decomposition-Based Multi-Objective Symbiotic Organism Search (MOSOS/D) [], Multi-objective Artificial Vultures Optimization Algorithm (MOAVOA) [], Multi-objective Group Learning Algorithm (MOGLA) [], Multi-objective Stochastic Paint Optimizer (MOSPO) [], Multi-objective Teaching-Learning-Based Optimizer (MOTLBO) [], and Multi-objective Sine-Cosine Algorithm (MOSCA) []. These algorithms have made significant contributions to addressing complex MOPs by providing efficient and diverse solutions.
In MOO algorithms, parameter selection plays a critical role in modulating the balance between exploration and exploitation capabilities of evolutionary operators []. For instance, in the NSGA-II, the crossover and mutation probabilities directly affect convergence behavior (exploitation). Similarly, in MOEA/D, the size of the neighborhood and the aggregation function parameters determine the balance between global search and local intensification. Moreover, distinct operator types inherently possess characteristic exploration-exploitation properties []. Differential evolution operators are highly exploratory, using stochastic vectors to probe the search space. In contrast, simulated binary crossover and polynomial mutation are more exploitative, refining solutions through local search. These characteristics are fundamentally influenced by their respective mechanistic structures and interaction modes with the solution space. Therefore, it is crucial to intelligently adjust algorithm parameters or combine different operators during optimization. This is key to maintaining an effective balance between exploration and exploitation. This adaptive approach enables dynamic adjustment to problem characteristics and enhances the algorithm’s ability to escape local optima while preserving convergence efficiency []. Reinforcement Learning (RL) is a machine learning paradigm focused on training agents to make optimal sequential decisions through interaction with an unknown environment []. In this learning process, an agent tries different actions and adjusts its strategy according to the results of these actions. Recent years have seen many studies successfully harness Reinforcement Learning (RL) to control evolutionary algorithms. By enabling real-time adaptation of parameters or on-the-fly selection of operators, RL helps these algorithms dynamically balance exploration and exploitation [,,]. This hybrid approach has demonstrated significant improvements in solving complex optimization problems, as evidenced by enhanced convergence properties and solution diversity across various benchmark and real-world problems [].
Although numerous MOO algorithms have been proposed to address MOPs, challenges such as slow convergence speed and susceptibility to local optima persist []. These algorithms aim to find a set of Pareto-optimal solutions that are evenly distributed. This even distribution provides a comprehensive range of high-quality choices for decision making. According to the No-Free-Lunch (NFL) theorem, no single algorithm can solve all optimization problems []. This fundamental principle underscores the persistent need to develop generalized MOO algorithms capable of effectively addressing diverse types of MOPs.
Social Network Search algorithm (SNS) is a metaheuristic algorithm proposed in 2021 for solving global optimization problems, based on human behavior []. The core innovation of the SNS algorithm lies in its guidance of the search process by simulating four distinct behavioral modes of social network users. It features a simple structure, high hierarchical organization, no need for manual parameter tuning, and fast convergence. SNS has demonstrated remarkable effectiveness in addressing diverse engineering problems, representing a novel and promising methodology with substantial potential for further research and development [,]. Building upon this foundation, this paper proposes a Reinforcement Learning Enhanced Multi-Objective Social Network Search Algorithm (QMOSNS) for Engineering Design Problems. The primary contributions of this research work can be expressed as follows:
  • The proposed QMOSNS enhances population diversity through Halton sequence initialization and maintains a multi-objective archive for storing Pareto-optimal solutions and selecting parent individuals. Moreover, Q-learning is integrated to dynamically adjust user behavior modes within the SNS framework, improving the balance between exploration and exploitation;
  • The performance of QMOSNS was rigorously evaluated on a suite of test problems, including constrained and unconstrained benchmarks as well as engineering design problems. The algorithm was compared against state-of-the-art MOO methods using two performance metrics, with the results validated by statistical tests, robustness analysis, and Pareto front visualizations;
  • Analysis shows that QMOSNS successfully handles MOPs, achieving well-balanced Pareto fronts across diverse benchmarks and engineering applications. Comparative results confirm its highly competitive performance, outperforming other algorithms in most cases.
The paper is structured as follows: Section 2 defines definitions related to this research. Section 3 details the original SNS algorithm, followed by Section 4′s presentation of QMOSNS. Section 5 validates the proposed algorithm through extensive experiments on 50 test problems and provides a detailed analysis of the results. Lastly, Section 6 conclude with a discussion of future research directions.

2. Background

2.1. Multi-Objective Optimization

Typically, an optimization problem only requires finding a single solution for a single objective function. However, in many real-world applications, we often face situations where multiple conflicting objectives need to be addressed simultaneously []. Taking minimization problems as an example, a multi-objective optimization problem can be formulated as follows:
min F ( x ) = min [ f 1 ( x ) , f 2 ( x ) , , f m ( x ) ] T F   s . t . g j ( x ) 0 , j = 1 , , l h k ( x ) = 0 , k = 1 , , n x = ( x 1 , x 2 , , x d ) T S
where x is a d-dimensional decision vector, S is the decision space, F ( x ) consists of m objective functions, F is the objective space, f i ( x ) is the i-th objective function, g j ( x ) is the j-th inequality constraint, h k ( x ) is the k-th equality constraint, and l and n represent the number of inequality constraints and equality constraints, respectively.
In single-objective optimization, the quality of a solution can be determined simply by comparing its fitness value. However, when comparing solutions for problems with multiple objectives, traditional comparison methods fail to evaluate multiple criteria effectively. To address this issue, most multi-objective algorithms incorporate the concept of dominance in their comparison mechanisms to handle multiple objectives and identify Pareto optimal solutions. Within this context, the relevant concepts in multi-objective optimization are defined as follows.

2.1.1. Pareto Dominance

Suppose x 1 = ( x 1 1 , x 2 1 , , x n 1 ) and x 2 = ( x 1 2 , x 2 2 , , x n 2 ) are two feasible solutions in the objective space that satisfy all constraints. Solution x 1 is said to dominate solution x 2 (denoted as x 1 x 2 ) if and only if the following conditions are met:
  i 1 , 2 , , m : f i ( x 1 ) f i ( x 2 )   j 1 , 2 , , m : f j ( x 1 ) < f j ( x 2 )

2.1.2. Pareto Optimality

A solution x * is defined as a Pareto optimal solution if it satisfies all constraints within the feasible region Ω and there exists no other solution in Ω that dominates x * .
x * Ω x Ω : x x *

2.1.3. Pareto Optimal Set

The set composed of all Pareto optimal solutions is called the Pareto Optimal Set (PS), defined as follows:
P S = x * | x * Ω x Ω : x x *

2.1.4. Pareto Optimal Front

The Pareto Optimal Front (PF) represents the projection of all Pareto optimal solutions x * in the PS onto the objective space F , formally defined as:
P F = f ( x * ) F | x * P S

2.2. Multi-Objective Performance Metrics

Unlike the evaluation of single-objective optimization algorithms, multi-objective optimization has two primary goals: finding a set of Pareto optimal solutions that satisfy the constraints and ensuring the diversity of solutions within the Pareto optimal set. Consequently, a comprehensive set of performance metrics is required to adequately assess the performance of multi-objective optimization algorithms. The following subsection details the metrics employed in this study.

2.2.1. Inverted Generational Distance

The Inverted Generational Distance (IGD) is a metric used to evaluate the comprehensive performance of multi-objective optimization algorithms []. It measures algorithm performance by computing the average distance from points on the true Pareto front to the solution set generated by the algorithm. A smaller IGD value indicates that the generated solution set is closer to the true Pareto front and exhibits better distribution uniformity. The definition of IGD is as follows:
I G D ( P * , A ) = y * P * d ( y * , A ) | P * | d ( y * , A ) = min y A { i = 1 m ( y i * y i ) 2 }
where P * represents a set of representative solutions on the true Pareto front, A denotes the approximate PF obtained by the algorithm, and m indicates the number of objectives.

2.2.2. Hypervolume

The Hypervolume (HV) is another key metric for evaluating the comprehensive performance of multi-objective optimization algorithms []. Unlike IGD, which requires prior knowledge of the true Pareto front, HV does not necessitate information about the true Pareto front for computation. It quantifies solution set quality by measuring the hypervolume enclosed between the solution set and a predefined reference point. A larger HV value indicates superior overall performance of the solution set, reflecting broader coverage and more uniform distribution in the objective space. The formal definition is as follows:
H V ( S ) = V O L ( x S [ f 1 ( x ) , z 1 r ] × × [ f m ( x ) , z m r ] )
where VOL(.) denotes the Lebesgue measure, m represents the number of objectives, and z r = ( z 1 r , , z m r ) T is a user-defined reference point in the objective space.

3. Social Network Search Algorithm

With the advancement of Internet technology, social networks have emerged, significantly enhancing the efficiency of interpersonal communication. The Social Network Search (SNS) algorithm draws inspiration from user interactions within these platforms. In social networks, users frequently influence others’ opinions by sharing their perspectives, insights, and ideas—a process that can be conceptualized as an optimization mechanism, as users inherently seek to maximize their influence and popularity. The SNS algorithm generates novel solutions to optimization problems by emulating four distinct behavioral modes observed in social network users: imitation, conversation, disputation, and innovation. The mathematical modeling and interpretation of these behavioral modes are elaborated as follows [].

3.1. Imitation Mood

In this mood, social network users often try to emulate the popular views of other users to get their views noticed. This mood is mathematically formulated in Equation (4):
x i n e w = x j + r a n d ( 1 , 1 ) × R R = r a n d ( 0 , 1 ) × r r = x i x j
where x i and x j represent the i-th and j-th of the current users, respectively, and i is not equal to j. rand(−1, 1) and rand(0, 1) represent random vectors in the range [−1, 1] and [0, 1], respectively. Moreover, R represents the shock radius of the imitation space and r represents the popular radius of user j.

3.2. Conversation Mood

In this mood, users will interact with each other and discuss different issues to generate new perspectives. This mood can be mathematically formulated as follows:
x i n e w = x k + R , R = r a n d ( 0 , 1 ) × D D = s i g n ( f i f j ) × ( x j x i )
where x k and x j are two randomly selected users. D represents the difference in users’ views, where X i represents the viewpoint of the current i-th user, and i, j, and k are different. Moreover, the expression ( s i g n ( f i f j ) ) describes the direction of motion of x k . It is worth noting that the user’s viewpoints are sent to change through the conversation with user j.

3.3. Disputation Mood

In this mood, the user explains his or her view of the issue to others and defends his or her point of view. In this case, the user sees the different views of others and may be influenced by them. The new affected view in disputation is as follows:
x i n e w = x i + r a n d ( 0 , 1 ) × ( M A F × x i ) M = t N r x t N r A F = 1 + r o u n d ( r a n d )
where M is the average of the views of the users involved in the discussion, and AF is the admitting effect that represents the statement of the user’s viewpoints during the discussion and takes the value integer 1 or 2. The symbol (round) stands for taking the input value to the closest integer value and ( N r ) is the number of people involved in the discussion.

3.4. Innovation Mood

In this mood, users share their personal beliefs and experiences in approaching an issue in a creative way with others, and therefore, new perspectives are presented. The mathematical explanation of innovation is as follows:
x i n e w d = t x j d + ( 1 t ) × n n e w d n n e w d = L B d + r a n d ( 0 , 1 ) × ( U B d L B d ) t = r a n d ( 0 , 1 )
where d represents the d-th variable in the user’s viewpoint; U B d and L B d represent the upper and lower bounds of the d-th dimensional variable; n n e w d is the new thought in the d-th dimension of the viewpoint; and x j d represents the d-th dimension of the j-th user’s viewpoint where i is not equal to j. x i n e w d consists of the pre-existing thought ( n j d ) and the new thought ( n n e w d ).A change in the d-th dimension of the user’s viewpoint also means that the user’s viewpoint on the event has changed.

3.5. Network Rules

There are rules defined in social networks that users need to follow when sharing their views. The rules that restrict the user’s views are formulated as follows:
x i n e w d = min ( x i n e w d , U B d ) x i n e w d = max ( x i n e w d , L B d ) d = 1 , 2 , 3 , , v i e w s   d i m i = 1 , 2 , 3 , , u s e r s   n u m b e r
During the next iteration of the algorithm, one of the four moods will be randomly selected to update the user’s view, and the new view generated will need to adhere to the rules of the social network. The main process of SNS is shown in Figure 1.
Figure 1. The flowchart of the SNS algorithm.

4. The Proposed QMOSNS

4.1. Population Initialization Based on Halton Sequence

The population initialization method significantly impacts both the iterative process and the final results of an algorithm. A uniformly distributed initial population enhances the exploration capability of optimization algorithms and improves search efficiency []. However, the original SNS algorithm lacks mechanisms for generating uniform initial populations, relying solely on random initialization methods. This often results in poor coverage uniformity across the solution space, leading to insufficient population diversity during early optimization stages and inefficient acquisition of solution space information. Existing literature demonstrates that low-discrepancy sequences like the Halton sequence serve as effective initialization mechanisms for metaheuristic algorithms, as they tend to uniformly cover the search space to facilitate optimal solution discovery []. Therefore, this study incorporates the Halton sequence into the initialization phase. By leveraging the deterministic characteristics of the Halton sequence to generate populations instead of random initialization, the proposed method achieves more comprehensive coverage of the solution space compared to traditional random approaches. As shown in Figure 2, a population of 1000 two-dimensional individuals was generated using random sampling, the Sobol sequence, and the Halton sequence method, with each dimension constrained to the interval [0, 1]. It can be seen that the Halton sequence achieved the best performance. The use of Halton sequences for initial population generation effectively prevents clustering phenomena in initial populations, thereby improving population diversity and uniformity. Such enhancements maintain strong exploration capability during early optimization stages and establish a solid foundation for global optimization.
Figure 2. Comparison of the distribution of the initial population.

4.2. Multi-Objective Archive

The QMOSNS algorithm is developed based on the SNS algorithm described earlier. In the SNS algorithm, the quality of a solution is evaluated according to its fitness value for a single objective, thereby enabling the selection of the globally optimal solution. However, in multi-objective optimization, solutions possess fitness values across multiple objectives. Their quality cannot be determined simply by directly comparing these objective fitness values. Inspired by SPEA-II [], this study adopts the same approach to reassign a new fitness value F i t ( X ) to each multi-objective solution X, thereby evaluating the quality of solutions. The specific definition is as follows:
F i t ( x ) = y S y x | R x | + 1 d i s t ( x , x ) + 2
where R x is the set of solutions dominated by the solution x, S is the set of solutions dominating the solution x and x denotes the nearest neighbor to x. The first component of Equation (9) represents the total number of solutions dominated by x in the solution set, while the second component corresponds to the inverse of the Euclidean distance between x and its nearest neighbor x . These two components, respectively, assess convergence and diversity performance. A smaller F i t ( x ) indicates better solution quality of x. Specifically, when F i t ( x ) < 1 , it signifies that x is non-dominated.
The utilization of external archives is prevalent in multi-objective optimization algorithms []. In this algorithm, an external archive A is likewise adopted to preserve candidate solutions. The external archive A maintains a fixed population size N of solutions, which serve as parent individuals for generating offspring in the subsequent generation. During each evolutionary iteration, both offspring individuals and solutions in archive A are evaluated and assigned fitness F i t ( x ) , which is then used to update the archive. The specific archive update mechanism is implemented as follows:
  • If the number of solutions satisfying F i t ( x ) < 1 is precisely equal to N, these solutions are stored in the A;
  • If the number of solutions satisfying F i t ( x ) < 1 exceeds N, the top N solutions are selected based on F i t ( x ) and stored in the A;
  • If the number of solutions satisfying F i t ( x ) < 1 is less than N, then the k compliant solutions are stored in the A, and the remaining (N – k) solutions are selected from the top-ranked candidates based on F i t ( x ) and stored in the A.

4.3. Q-Learning-Based Mood Selection Strategy

The SNS algorithm randomly selects the moods model and the user’s optimization mechanism, which is conducive to improving the algorithm’s global exploration ability and avoiding the algorithm’s premature convergence in the process of optimization, but it also reduces the algorithm’s optimization accuracy and efficiency to a certain extent. Applying reinforcement learning to social network search algorithms helps to help algorithms rationally select sentiment models in the process of searching for excellence, in order to balance the algorithm’s global exploration and local exploitation capabilities, and thus improve the efficiency and accuracy of searching for excellence.
The Q-learning algorithm is one of the most frequently used algorithms for RL, which consists of five parts: the intelligent body, the Q-table, the reward value R, the state set st, and the action set ac []. The Q-learning algorithm first perceives the state of the environment through the intelligent body, and then makes the optimal decision based on the value of the Q-table. After the decision is applied to the environment, the state of the intelligent body will be changed and the reward value for the decision will be fed back to the intelligent body from the environment to update the Q-table. The dynamic updating mechanism of the Q-table is as follows:
Q s t t , a c t Q s t t , a c t + λ R t + 1 + γ max Q s t t + 1 , a c Q s t , a c t
where λ represents the learning efficiency, γ represents the discount factor, R t + 1 denotes the immediate reward obtained by executing the action a c t in state s t t , and Q ( s t t , a c t ) represents the Q-table value of executing the action a c t in state s t t . λ is crucial to the Q-learning algorithm, and the larger λ is, the greater the influence of the expected future reward on the computation of the future Q-value is; and the smaller λ is, the greater the influence of the current Q-value on the computation of the future Q-value is. Therefore, the learning effect of the intelligence in the search space can be improved by adaptively decreasing λ in the process of SNS iteration, and λ is calculated as follows:
λ = λ i n i t i a l + λ f i n a l 2 λ i n i t i a l + λ f i n a l 2 cos π 1 t M a x g e n
where λ i n i t i a l and λ f i n a l represent the initial and final values of learning efficiency, respectively.
In constrained multi-objective optimization methods, algorithm performance must simultaneously consider convergence and feasibility. The Hypervolume (HV) indicator serves as a comprehensive evaluation metric that assesses both the convergence and diversity of Pareto optimal solution sets by measuring the volume of the multidimensional space dominated by these solutions relative to a reference point []. The feasibility of the population is evaluated using both the constraint violation (CV) and the feasibility rate (FR). Notably, the changes in HV and CV (denoted Δ H V and Δ C V ) between two consecutive generations can effectively reflect the efficacy of the evolutionary strategy. Based on this, four states can be defined according to the changes in HV and CV:
  • State 1 ( S 1 ): Both HV and CV improve simultaneously;
  • State 2 ( S 2 ): Only the HV improves;
  • State 3 ( S 3 ): Only the CV improves;
  • State 4 ( S 4 ): There is no improvement in either HV or CV.
The immediate reward of applying an action to a state, denoted as R, is determined by the improvement of the evolution. The reward value R varies as shown in Equation (12).
R = 1         ,         Δ H V > 0   &   FR = 1 0.5   , Δ H V > 0   &   FR 1   &   Δ C V 0 1     ,       others
When using the Q-learning algorithm in SNS algorithms, a Q-table needs to be created for each user. Given that the SNS incorporates four mood models and four states have been defined previously, the Q-table is designed as a matrix of 4 rows and 4 columns, where the rows represent states s t and the columns represent actions a c . As shown in Figure 3, before the Q-table update, the agent selected the imitation mood model while in S 3 , and the corresponding Q-value for this position was updated to 1.42 according to Equation (11). After the Q-table update, the current state transitions to S 1 . Subsequently, a random number r will be generated, and if r is smaller than the set greedy threshold ε , the action corresponding to the maximum Q-value will be selected; otherwise, an action will be randomly chosen from the four actions. By following this cycle, the appropriate mood model for the current situation can be selected based on the Q value.
Figure 3. Operation of the Q-learning-based mood selection strategy.
The pseudo code of QMOSNS is presented in Algorithm 1.
Algorithm 1 QMOSNS Algorithm
Input: Popsize: Number of population, Maxgen: Maximum iterations, D: Dimension size, UB: Upper bound of a variable, LB: Lower bound of a variable.
Output: Return the archive.
Initialize the Q-table.
Initialize the population using Halton sequence.
Calculate the fitness value of each individual in the population.
Update the archive.
for t = 1: Maxgen
  if t = 1
    Randomly select a mood.
  else
    Select a mood using the Q-table.
  End if
  Generate offspring using the selected mood.
  Calculate the fitness value of each individual in the offspring.
  Update the archive and select parent solutions.
  Update the Q-table.
end for

4.4. Computational Complexity

The complexity of the proposed QMOSNS is based on the number of decision variables, the objective variables, and the population size. Imagine a multi-objective problem with D decision variables, M objective variables, and N particles as an example. At each generation, the QMOSNS mainly spends the computational cost for updating user’s viewpoint, updating archive, and implementing the Q-learning-based emotion selection strategy. The complexity of update user’s viewpoint is decided by N and D, so the computational complexity of this process is O(DN). During the archive update process, fitness assignment constitutes the most computationally complex part, with a computational complexity of O(MN3). The Q-learning-based operator selection process consists of three steps: state determination, Q-table update, and optimal action selection. Their computational complexities are O(N2), O(1), and O(1), respectively. To sum up, the total computational complexity of the proposed algorithm is O(MN3), which is highly sensitive to increases in N and is a key factor determining the actual runtime. When N increases by a factor of k, the theoretical runtime grows by k3 times. This implies that while the algorithm is feasible for small to medium-sized populations, it faces computational bottlenecks in ultra-large-scale scenarios.

5. Numerical Examples and Results

5.1. Experimental Setup

In this section, 50 well-established multi-objective benchmark problems and engineering design problems derived from authoritative studies are employed to rigorously evaluate the optimization performance of QMOSNS. These test problems incorporate objective functions with distinct characteristics and span varying dimensions of design variables, which are systematically classified into three primary categories:
  • Unconstrained multi-objective problems: ZDT1, ZDT2, ZDT3, ZDT4, ZDT6, DTLZ1, DTLZ2, DTLZ3, DTLZ4, DTLZ5, DTLZ6, DTLZ7, UF1, UF2, UF3, UF4, UF5, UF6, UF7, UF8, UF9, UF10 [,,];
  • Constrained multi-objective problems: C1_DTLZ1, C1_DTLZ2, C2_DTLZ3, C2_DTLZ4, MW1, MW2, MW3, MW4, MW5, MW6, MW7, MW8, MW9, MW10, MW11, MW12, MW13, MW14 [,];
  • Multi-objective engineering design problems: Pressure Vessel Design [], Vibrating Platform Design [], Welded Beam Design [], Disc Brake Design [], Car Side Impact Design [], Four Bar Plane Truss [], Multiple Disk Clutch Brake Design [], Spring Design [], Multi-product Batch Plant [], Crash Energy Management for High-speed Train [].
To validate the effectiveness of the proposed QMOSNS algorithm, comparative experiments were conducted against six state-of-the-art algorithms, including two widely used approaches (NSGA-II [], MOEA/D []) and four advanced methods (ARMOEA [], MOEADDAE [], CMOES [], PPS []). All experiments were performed in a computational environment equipped with an Intel(R) Core(TM) i5-12600KF CPU @3.70 GHz and 32.00 GB RAM. The operating system was Windows 11 Pro, and MATLAB R2023b served as the execution platform. The experimental studies were implemented on PlatEMO [], with all algorithmic parameters maintained at their default settings in the platform.
Each algorithm is independently executed 30 times for each multi-objective optimization problem, and the statistical results are subjected to a Wilcoxon rank-sum test with a significance level of 0.05. The symbols “+”, “−”, and “=“ denote that the algorithm performs better than, worse than, or equal to the compared algorithm, respectively. To ensure reliable and fair comparisons, the population size for all algorithms is set to 100. The maximum number of evaluations for the ZDT, DTLZ, UF, C_DTLZ, and MW test functions is set to 25,000, 100,000, 100,000, 100,000, and 60,000, respectively. For multi-objective engineering optimization problems, the maximum number of evaluations is determined according to the following criteria (where m represents the number of objectives in the problem and d denotes the problem’s dimensionality):
2 × 10 4 ,     i f ( m = 2 ) & ( d 10 ) 8 × 10 4 ,     else i f ( m = 2 ) & ( d > 10 ) 2.6250 × 10 4 ,     else i f ( m = 3 ) & ( d 10 ) 1.05 × 10 5 ,     else i f ( m = 3 ) & ( d > 10 )

5.2. Multi-Objective Benchmark Problems

5.2.1. Unconstrained Multi-Objective Problems

Table 1 presents the final statistical results of the IGD metrics obtained by QMOSNS, NSGA-II, MOEA/D, ARMOEA, MOEADDAE, CMOES and PPS. The results indicate that in terms of both the average and standard deviation values of the IGD metric, QMOSNS significantly outperforms NSGA-II, MOEA/D, ARMOEA, MOEADDAE, CMOES, and PPS on ZDT1, ZDT2, ZDT3, and ZDT6, only underperforming NSGA-II and MOEADDAE on ZDT4. This demonstrates that QMOSNS can achieve solutions closer to the true Pareto front on most test functions compared to the other algorithms. Table 2 shows the statistical results of the HV performance metrics for the DTLZ benchmark functions obtained by the QMOSNS algorithm and other multi-objective algorithms. The experimental results reveal that the QMOSNS algorithm outperforms the other algorithms on the ZDT1, ZDT2, and ZDT6 test functions. MOEADDAE achieves the best performance on ZDT3, while NSGA-II performs best on ZDT4, but QMOSNS still secures the second rank on these two test functions. Therefore, QMOSNS demonstrates superior convergence and diversity on the ZDT test problems. Figure 4 displays a graphical representation of the true Pareto front of the ZDT test suite and the Pareto fronts obtained by QMOSNS. All plots demonstrate that the Pareto optimal fronts obtained by QMOSNS exhibit uniform distribution along the true Pareto optimal fronts.
Table 1. Statistical Results of MOO Algorithms in Terms of IGD For ZDT Test Suites.
Table 2. Statistical Results of MOO Algorithms in Terms of HV For ZDT Test Suites.
Figure 4. Pareto optimal fronts generated by QMOSNS for test functions ZDT1, ZDT2, ZDT3, ZDT4, and ZDT6.
Table 3 and Table 4 present the statistical results of IGD and HV obtained by QMOSNS, NSGA-II, MOEA/D, ARMOEA, MOEADDAE, CMOES, and PPS. Overall, the IGD metric results indicate that QMOSNS outperforms all other compared algorithms across all test problems except DTLZ4, demonstrating the superior convergence performance of the proposed algorithm on these test functions. The significant differences in IGD values between QMOSNS and other algorithms further support this conclusion. Although QMOSNS shows slightly inferior performance to MOEA/D on DTLZ4 and performs comparably to ARMOEA, this does not compromise its overall effectiveness. Regarding the HV metric, QMOSNS achieves significantly better results than the other algorithms on four test functions: DTLZ2, DTLZ5, DTLZ6, and DTLZ7. MOEA/D delivers the best performance on DTLZ1, DTLZ3, and DTLZ4. However, QMOSNS remains highly competitive with MOEA/D on DTLZ1 and DTLZ3, only slightly underperforming on DTLZ3. In summary, QMOSNS still exhibits strong performance across the DTLZ test suite. Figure 5 illustrates the true Pareto front and the obtained Pareto fronts of all algorithms on the DTLZ test problems. The results in Table 3 and Table 4 are visually confirmed by the graphical representations in Figure 5, where it can be observed that the Pareto front generated by QMOSNS not only converges well to the true Pareto front but also maintains a uniform distribution.
Table 3. Statistical Results of MOO Algorithms in Terms of IGD For DTLZ Test Suites.
Table 4. Statistical Results of MOO Algorithms in Terms of HV For DTLZ Test Suites.
Figure 5. Pareto optimal fronts generated by QMOSNS for test functions DTLZ1-DTLZ7.
Similarly to Table 1, Table 2, Table 3 and Table 4, Table 5 and Table 6 present the statistical results of IGD and HV metrics obtained by QMOSNS, NSGA-II, MOEA/D, ARMOEA, MOEADDAE, CMOES, and PPS on the UF test functions, respectively. In terms of the IGD metric, QMOSNS demonstrates the best convergence performance on UF2, UF5, UF6, and UF7. PPS achieves the best convergence on UF1 and UF3, CMOES performs best on UF4 and UF10, while MOEADDAE shows superior convergence on UF8 and UF9. Regarding the HV metric, the QMOSNS algorithm outperforms the other algorithms on most functions. PPS only exhibits competitive performance on UF1, UF3, and UF7. Additionally, MOEADDAE also demonstrates better performance than other algorithms on UF8. According to the Wilcoxon rank-sum test, it can be observed from Table 5 and Table 6 that QMOSNS significantly outperforms the other methods. Figure 6 illustrates the Pareto optimal fronts obtained by QMOSNS and the true Pareto optimal fronts for each function.
Table 5. Statistical Results of MOO Algorithms in Terms of IGD For UF Test Suites.
Table 6. Statistical Results of MOO Algorithms in Terms of HV For UF Test Suites.
Figure 6. Pareto optimal fronts generated by QMOSNS for test functions UF1-UF10.
The aggregated performance metrics across Table 1, Table 2, Table 3, Table 4, Table 5 and Table 6 demonstrate that QMOSNS emerges as the superior algorithm among the seven evaluated methods for handling unconstrained multi-objective optimization problems.

5.2.2. Constrained Multi-Objective Problems

In the previous section on unconstrained multi-objective test problems, the proposed QMOSNS demonstrated superior performance. To further evaluate its capability in handling constrained multi-objective optimization problems, we conducted performance assessments on two constrained test suites.
Table 7 and Table 8 present the statistical results of IGD and HV metrics obtained by QMOSNS and comparative algorithms on the C_DTLZ test set. The results indicate that QMOSNS achieves the best performance on C1_DTLZ1, C2_DTLZ2, and C3_DTLZ4, with its IGD values demonstrating excellent convergence properties. On the C1_DTLZ3 function, CMOES yields the optimal results, followed by MOEADDAE, while the other algorithms fail to converge close to the true Pareto optimal front. These observations are further supported by the HV metric results in Table 8. Overall, QMOSNS exhibits the strongest performance across this test set. Figure 7 illustrates the Pareto optimal front obtained by QMOSNS and the corresponding true Pareto front for this test set. Figure 7 reveals that QMOSNS achieves excellent convergence to the true Pareto optimal front with proper distribution and spread for the Pareto optimal solutions, except in the case of C1_DTLZ3.
Table 7. Statistical Results of MOO Algorithms in Terms of IGD For C_DTLZ Test Suites.
Table 8. Statistical Results of MOO Algorithms in Terms of HV For C_DTLZ Test Suites.
Figure 7. Pareto optimal fronts generated by QMOSNS for test functions C1-DTLZ1, C1-DTLZ3, C2-DTLZ2, and C3-DTLZ4.
The MW test suite presents significant optimization challenges due to its notably small feasible regions and multiple nonlinear constraints. Table 9 and Table 10 display the statistical results of IGD and HV metrics obtained by QMOSNS, NSGA-II, MOEA/D, ARMOEA, MOEADDAE, CMOES, and PPS on the MW test set, respectively. Analysis of IGD metrics reveals that QMOSNS achieves optimal performance on most functions, including MW1, MW2, MW4, MW6, MW8, MW10, and MW13. This superior performance is consistently reflected in the corresponding HV metric results as well. While QMOSNS shows relatively weaker performance compared to some algorithms on the remaining functions, the Wilcoxon rank-sum test results for both IGD and HV metrics demonstrate that the proposed algorithm maintains substantial advantages over NSGA-II, MOEA/D, ARMOEA, MOEADDAE, and PPS, with only marginal superiority when compared to CMOES. Figure 8 provides a graphical representation comparing the true Pareto front of the MW test suite with the Pareto fronts obtained by QMOSNS. These figures show that some of the test problems have special Pareto optimal fronts. Figure 8 reveals that QMOSNS consistently achieves solutions with both strong convergence to the true Pareto fronts and uniform distribution across the entire optimal regions.
Table 9. Statistical Results of MOO Algorithms in Terms of IGD For MW Test Suites.
Table 10. Statistical Results of MOO Algorithms in Terms of HV For MW Test Suites.
Figure 8. Pareto optimal fronts generated by QMOSNS for test functions MW1-MW14.
The test results on constrained multi-objective optimization problems demonstrate that QMOSNS effectively handles constraints and obtains Pareto optimal fronts with both high convergence and good distribution across different Pareto optimal regions.

5.3. Multi-Objective Engineering Design Problems

While multi-objective benchmark functions are effective for algorithm testing, real-world engineering optimization problems pose greater challenges to algorithm performance. To evaluate the applicability of the proposed algorithm, this study further tests the algorithm on 10 multi-objective engineering design problems. The basic information of these 10 test problems is presented in Table 11, where m denotes the number of objectives, d represents the problem dimension, ng indicates the number of inequality constraints, and nh stands for the number of equality constraints. Since real multi-objective engineering design problems lack known true Pareto fronts, the HV metric is exclusively used to evaluate algorithm performance on these problems. Table 12 provides the statistical results of HV metrics obtained by each algorithm.
Table 11. Basic Information on Multi-objective Engineering Design Problems.
Table 12. Statistical Results of MOO Algorithms in Terms of HV For Engineering Design Problems.
Table 12 demonstrates that QMOSNS achieves the best HV metric values on 6 engineering design problems: Vibrating Platform Design, Disc Brake Design, Multiple Disk Clutch Brake Design, Spring Design, Multi-product Batch Plant, and Crash Energy Management for High-speed Train. It ranked second only to ARMOEA on problems Pressure Vessel Design, Car Side Impact Design, and Four Bar Plane Truss, while ranking third on problem Welded Beam Design, performing inferior to NSGA-II and CMOES. Furthermore, the results of the Wilcoxon rank-sum test provide additional evidence that, across these 10 real-world engineering design problems, QMOSNS significantly outperforms NSGA-II, MOEA/D, ARMOEA, MOEADDAE, CMOES, and PPS algorithms while maintaining favorable robustness. These experimental results confirm that QMOSNS exhibits stronger competitiveness in solving practical engineering design problems compared to other algorithms. Figure 9 provides a graphical representation of the Pareto fronts obtained by QMOSNS.
Figure 9. Pareto optimal fronts generated by QMOSNS for engineering design problems.

5.4. Ablation Study on Initialization Strategy

To evaluate the effectiveness of the population initialization based on Halton sequence, an ablation study was conducted on ten test functions selected from the aforementioned 50. In the ablation study, the algorithm without Halton sequence-based population initialization is referred to as QMOSNS-1, while the version employing this initialization method is designated as QMOSNS-2. The performance is measured using the IGD and HV metrics, with all other settings consistent with the previous experiments. According to the results shown in Table 13, QMOSNS-1 and QMOSNS-2 yield generally similar results, though QMOSNS-2 performs slightly better. This indicates that the Halton sequence-based initialization method contributes to some extent to improving the optimization outcomes, thereby demonstrating the effectiveness of the approach.
Table 13. Statistical Results in the Ablation Study.

6. Conclusions

This study proposes a novel multi-objective optimization algorithm called QMOSNS, which is developed based on the Social Network Search algorithm framework to solve complex multi-objective optimization problems. QMOSNS employs the Halton sequence to generate the initial population, thereby enhancing population diversity. By replacing the random selection mechanism in the traditional SNS emotion model with Q-learning, the algorithm can effectively identify optimal mutation strategies during iterations. The approach incorporates an external archive population combined with non-dominated sorting and crowding distance to handle multi-objective optimization problems. The performance of QMOSNS was evaluated against NSGA-II, MOEA/D, ARMOEA, MOEADDAE, CMOES, and PPS using three test suites: 22 unconstrained multi-objective benchmark problems, 18 constrained multi-objective benchmark problems, and 10 multi-objective engineering design problems, with both IGD and HV metrics employed for performance assessment. The experimental results across all test categories demonstrate that QMOSNS outperforms NSGA-II, MOEA/D, ARMOEA, MOEADDAE, CMOES, and PPS in solving multi-objective optimization problems. In the real world, most multi-objective engineering optimization problems often involve complex constraints and intricate Pareto front structures, posing significant challenges to the MOO algorithms. The QMOSNS method proposed in this study adaptively selects evolutionary operators through a Q-learning mechanism, effectively balancing global exploration and local exploitation processes, thereby enhancing its problem-solving capability in complex optimization scenarios. It has demonstrated strong performance across multiple test problems, proving its considerable applicability and effectiveness in addressing real-world multi-objective optimization challenges. For future research, QMOSNS will be applied to address dynamic optimization, uncertainty optimization, and practical engineering problems, particularly optimization challenges in mechanical design. Additionally, comparative studies will be conducted between QMOSNS and other state-of-the-art multi-objective optimization algorithms to further investigate its performance in solving real-world engineering problems.

Author Contributions

Conceptualization, W.P. and Z.L.; methodology, W.P.; software, Z.L.; validation, W.P., Z.L. and G.H.; formal analysis, W.P.; investigation, J.L.; resources, W.P.; data curation, Z.L.; writing—original draft preparation, W.P., Z.L., J.L. and G.H.; writing—review and editing, W.P., Z.L., J.L. and G.H.; visualization, Z.L.; supervision, J.L. and G.H.; project administration, G.H.; funding acquisition, W.P., J.L. and G.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National High Technology Research and Development Program of China (Grant No. 2014AA042001), the National Natural Science Foundation of China (Grant No. 52365012 and 52165008), the Key Project of the Natural Science Foundation of Jiangxi (Grant No. 20224ACB204015), and the Jiangxi Provincial Natural Science Foundation (Grant No. 20252BAC200332).

Data Availability Statement

The data presented in this study are openly available from the PlatEMO 4.11 platform, which provides publicly available benchmark test sets (ZDT, DTLZ, UF, C_DTLZ, MW, and RWMOP). The data can be accessed at: https://github.com/BIMK/PlatEMO(accessed on 14 March 2025).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. He, M.; Wang, Z.; Chen, H.; Cao, Y.; Ma, L. Multi-Objective Evolutionary Algorithm Based on Decomposition with Orthogonal Experimental Design. Expert Syst. 2025, 42, e13802. [Google Scholar] [CrossRef]
  2. Yin, L.; Sun, Z. Distributed Multi-Objective Grey Wolf Optimizer for Distributed Multi-Objective Economic Dispatch of Multi-Area Interconnected Power Systems. Appl. Soft Comput. 2022, 117, 108345. [Google Scholar] [CrossRef]
  3. Tang, J.; Dou, T.; Wu, F.; Hu, L.; Yu, T. Scheduling Optimization of a Vehicle Power Battery Workshop Based on an Improved Multi-Objective Particle Swarm Optimization Method. Mathematics 2025, 13, 2790. [Google Scholar] [CrossRef]
  4. Méndez, M.; Rossit, D.A.; González, B.; Frutos, M. Proposal and Comparative Study of Evolutionary Algorithms for Optimum Design of a Gear System. IEEE Access 2020, 8, 3482–3497. [Google Scholar] [CrossRef]
  5. Zhou, X.; Cai, X.; Zhang, H.; Zhang, Z.; Jin, T.; Chen, H.; Deng, W. Multi-Strategy Competitive-Cooperative Co-Evolutionary Algorithm and Its Application. Inf. Sci. 2023, 635, 328–344. [Google Scholar] [CrossRef]
  6. Marler, R.T.; Arora, J.S. The Weighted Sum Method for Multi-Objective Optimization: New Insights. Struct. Multidiscip. Optim. 2010, 41, 853–862. [Google Scholar] [CrossRef]
  7. Konak, A.; Coit, D.W.; Smith, A.E. Multi-Objective Optimization Using Genetic Algorithms: A Tutorial. Reliab. Eng. Syst. Saf. 2006, 91, 992–1007. [Google Scholar] [CrossRef]
  8. Mirjalili, S.; Gandomi, A.H.; Mirjalili, S.Z.; Saremi, S.; Faris, H.; Mirjalili, S.M. Salp Swarm Algorithm: A Bio-Inspired Optimizer for Engineering Design Problems. Adv. Eng. Softw. 2017, 114, 163–191. [Google Scholar] [CrossRef]
  9. Mirjalili, S.; Jangir, P.; Saremi, S. Multi-Objective Ant Lion Optimizer: A Multi-Objective Optimization Algorithm for Solving Engineering Problems. Appl. Intell. 2017, 46, 79–95. [Google Scholar] [CrossRef]
  10. Premkumar, M.; Jangir, P.; Sowmya, R.; Alhelou, H.H.; Heidari, A.A.; Chen, H. MOSMA: Multi-Objective Slime Mould Algorithm Based on Elitist Non-Dominated Sorting. IEEE Access 2021, 9, 3229–3248. [Google Scholar] [CrossRef]
  11. Kalita, K.; Jangir, P.; Pandya, S.B.; Cep, R.; Abualigah, L.; Migdady, H.; Daoud, M.S. Many-Objective Artificial Hummingbird Algorithm: An Effective Many-Objective Algorithm for Engineering Design Problems. J. Comput. Des. Eng. 2024, 11, 16–39. [Google Scholar] [CrossRef]
  12. Schaffer, J.D. Multiple Objective Optimization with Vector Evaluated Genetic Algorithms. In Proceedings of the First International Conference on Genetic Algorithms and Their Applications, Pittsburgh, PA, USA, 24–26 July 1985; Lawrence Erlbaum: Hillsdale, NJ, USA, 1985; pp. 93–100. [Google Scholar]
  13. Srinivas, N.; Deb, K. Multiobjective Optimization Using Nondominated Sorting in Genetic Algorithms. Evol. Comput. 1994, 2, 221–248. [Google Scholar] [CrossRef]
  14. Deb, K.; Pratap, A.; Agarwal, S.; Meyarivan, T. A Fast and Elitist Multiobjective Genetic Algorithm: NSGA-II. IEEE Trans. Evol. Comput. 2002, 6, 182–197. [Google Scholar] [CrossRef]
  15. Zhang, Q.; Li, H. MOEA/D: A Multiobjective Evolutionary Algorithm Based on Decomposition. IEEE Trans. Evol. Comput. 2007, 11, 712–731. [Google Scholar] [CrossRef]
  16. Coello, C.A.C.; Pulido, G.T.; Lechuga, M.S. Handling Multiple Objectives with Particle Swarm Optimization. IEEE Trans. Evol. Comput. 2004, 8, 256–279. [Google Scholar] [CrossRef]
  17. Fu, S.; Li, K.; Huang, H.; Ma, C.; Fan, Q.; Zhu, Y. Red-Billed Blue Magpie Optimizer: A Novel Metaheuristic Algorithm for 2D/3D UAV Path Planning and Engineering Design Problems. Artif. Intell. Rev. 2024, 57, 134. [Google Scholar] [CrossRef]
  18. Mirjalili, S.; Saremi, S.; Mirjalili, S.M.; Coelho, L.d.S. Multi-Objective Grey Wolf Optimizer: A Novel Algorithm for Multi-Criterion Optimization. Expert Syst. Appl. 2016, 47, 106–119. [Google Scholar] [CrossRef]
  19. Sadollah, A.; Eskandar, H.; Kim, J. Water Cycle Algorithm for Solving Constrained Multi-Objective Optimization Problems. Appl. Soft Comput. 2015, 27, 279–298. [Google Scholar] [CrossRef]
  20. Ganesh, N.; Shankar, R.; Kalita, K.; Jangir, P.; Oliva, D.; Perez-Cisneros, M. A Novel Decomposition-Based Multi-Objective Symbiotic Organism Search Optimization Algorithm. Mathematics 2023, 11, 1898. [Google Scholar] [CrossRef]
  21. Khodadadi, N.; Soleimanian Gharehchopogh, F.; Mirjalili, S. MOAVOA: A New Multi-Objective Artificial Vultures Optimization Algorithm. Neural Comput. Appl. 2022, 34, 20791–20829. [Google Scholar] [CrossRef]
  22. Rahman, C.M.; Mohammed, H.M.; Abdul, Z.K. Multi-Objective Group Learning Algorithm with a Multi-Objective Real-World Engineering Problem. Appl. Soft Comput. 2024, 166, 112145. [Google Scholar] [CrossRef]
  23. Khodadadi, N.; Abualigah, L.; Mirjalili, S. Multi-Objective Stochastic Paint Optimizer (MOSPO). Neural Comput. Appl. 2022, 34, 18035–18058. [Google Scholar] [CrossRef]
  24. Lin, W.; Yu, D.Y.; Zhang, C.; Liu, X.; Zhang, S.; Tian, Y.; Liu, S.; Xie, Z. A Multi-Objective Teaching−learning-Based Optimization Algorithm to Scheduling in Turning Processes for Minimizing Makespan and Carbon Footprint. J. Clean. Prod. 2015, 101, 337–347. [Google Scholar] [CrossRef]
  25. Tawhid, M.A.; Savsani, V. Multi-Objective Sine-Cosine Algorithm (MO-SCA) for Multi-Objective Engineering Design Problems. Neural Comput. Appl. 2019, 31, 915–929. [Google Scholar] [CrossRef]
  26. Yang, Z.; Hu, X.; Li, Y.; Liang, M.; Wang, K.; Wang, L.; Tang, H.; Guo, S. A Q-Learning-Based Improved Multi-Objective Genetic Algorithm for Solving Distributed Heterogeneous Assembly Flexible Job Shop Scheduling Problems with Transfers. J. Manuf. Syst. 2025, 79, 398–418. [Google Scholar] [CrossRef]
  27. Zhang, Z.; Tang, Q.; Zhang, L.; Li, Z.; Cheng, L. A Q-Learning-Based Multi-Population Algorithm for Multi-Objective Distributed Heterogeneous Assembly No-Idle Flowshop Scheduling with Batch Delivery. Expert Syst. Appl. 2025, 263, 125690. [Google Scholar] [CrossRef]
  28. Zheng, J.; Chen, S. A Q-Learning Multi-Objective Grey Wolf Optimizer for the Distributed Hybrid Flowshop Scheduling Problem. Eng. Optim. 2024, 57, 2609–2628. [Google Scholar] [CrossRef]
  29. Zhang, Z.; Shao, Z.; Shao, W.; Chen, J.; Pi, D. MRLM: A Meta-Reinforcement Learning-Based Metaheuristic for Hybrid Flow-Shop Scheduling Problem with Learning and Forgetting Effects. Swarm Evol. Comput. 2024, 85, 101479. [Google Scholar] [CrossRef]
  30. Huang, Y.; Guo, Y.; Chen, G.; Wei, H.; Zhao, X.; Yang, S.; Ge, S. Q-Learning Assisted Multi-Objective Evolutionary Optimization for Low-Carbon Scheduling of Open-Pit Mine Trucks. Swarm Evol. Comput. 2025, 92, 101778. [Google Scholar] [CrossRef]
  31. Ebrie, A.S.; Kim, Y.J. Reinforcement Learning-Based Multi-Objective Optimization for Generation Scheduling in Power Systems. Systems 2024, 12, 106. [Google Scholar] [CrossRef]
  32. Zheng, R.; Zhang, Y.; Sun, X.; Yang, L.; Song, X. A Reinforcement Learning-Assisted Multi-Objective Evolutionary Algorithm for Generating Green Change Plans of Complex Products. Appl. Soft Comput. 2025, 170, 112660. [Google Scholar] [CrossRef]
  33. Qi, R.; Li, J.; Wang, J.; Jin, H.; Han, Y. QMOEA: A Q-Learning-Based Multiobjective Evolutionary Algorithm for Solving Time-Dependent Green Vehicle Routing Problems with Time Windows. Inf. Sci. 2022, 608, 178–201. [Google Scholar] [CrossRef]
  34. Huy, T.H.B.; Nallagownden, P.; Truong, K.H.; Kannan, R.; Vo, D.N.; Ho, N. Multi-Objective Search Group Algorithm for Engineering Design Problems. Appl. Soft Comput. 2022, 126, 109287. [Google Scholar] [CrossRef]
  35. Adam, S.P.; Alexandropoulos, S.-A.N.; Pardalos, P.M.; Vrahatis, M.N. No Free Lunch Theorem: A Review. In Approximation and Optimization: Algorithms, Complexity and Applications; Demetriou, I.C., Pardalos, P.M., Eds.; Springer International Publishing: Cham, Switzerland, 2019; pp. 57–82. ISBN 978-3-030-12767-1. [Google Scholar]
  36. Talatahari, S.; Bayzidi, H.; Saraee, M. Social Network Search for Global Optimization. IEEE Access 2021, 9, 92815–92863. [Google Scholar] [CrossRef]
  37. Zamli, K.Z.; Alhadawi, H.S.; Din, F. Utilizing the Roulette Wheel Based Social Network Search Algorithm for Substitution Box Construction and Optimization. Neural Comput. Appl. 2023, 35, 4051–4071. [Google Scholar] [CrossRef]
  38. M. Shaheen, A.; El-Sehiemy, R.A.; Hasanien, H.M.; Ginidi, A. An Enhanced Optimizer of Social Network Search for Multi-Dimension Optimal Power Flow in Electrical Power Grids. Int. J. Electr. Power Energy Syst. 2024, 155, 109572. [Google Scholar] [CrossRef]
  39. Han, H.; Zhang, L.; Yinga, A.; Qiao, J. Adaptive Multiple Selection Strategy for Multi-Objective Particle Swarm Optimization. Inf. Sci. 2023, 624, 235–251. [Google Scholar] [CrossRef]
  40. Sierra, M.R.; Coello Coello, C.A. Improving PSO-Based Multi-Objective Optimization Using Crowding, Mutation and ε-Dominance. In Evolutionary Multi-Criterion Optimization; Coello Coello, C.A., Hernández Aguirre, A., Zitzler, E., Eds.; Springer: Berlin/Heidelberg, Germany, 2005; pp. 505–519. [Google Scholar]
  41. Zitzler, E.; Thiele, L. Multiobjective Evolutionary Algorithms: A Comparative Case Study and the Strength Pareto Approach. IEEE Trans. Evol. Comput. 1999, 3, 257–271. [Google Scholar] [CrossRef]
  42. Li, Q.; Liu, S.-Y.; Yang, X.-S. Influence of Initialization on the Performance of Metaheuristic Optimizers. Appl. Soft Comput. 2020, 91, 106193. [Google Scholar] [CrossRef]
  43. Tharwat, A.; Schenck, W. Population Initialization Techniques for Evolutionary Algorithms for Single-Objective Constrained Optimization Problems: Deterministic vs. Stochastic Techniques. Swarm Evol. Comput. 2021, 67, 100952. [Google Scholar] [CrossRef]
  44. Zitzler, E.; Laumanns, M.; Thiele, L. SPEA2: Improving the Strength Pareto Evolutionary Algorithm; ETH Zurich, Computer Engineering and Networks Laboratory: Zurich, Switzerland, 2001. [Google Scholar]
  45. Xia, M.; Dong, M. A Novel Two-Archive Evolutionary Algorithm for Constrained Multi-Objective Optimization with Small Feasible Regions. Knowl.-Based Syst. 2022, 237, 107693. [Google Scholar] [CrossRef]
  46. Gao, M.; Gao, K.; Ma, Z.; Tang, W. Ensemble Meta-Heuristics and Q-Learning for Solving Unmanned Surface Vessels Scheduling Problems. Swarm Evol. Comput. 2023, 82, 101358. [Google Scholar] [CrossRef]
  47. Zitzler, E.; Deb, K.; Thiele, L. Comparison of Multiobjective Evolutionary Algorithms: Empirical Results. Evol. Comput. 2000, 8, 173–195. [Google Scholar] [CrossRef] [PubMed]
  48. Deb, K.; Thiele, L.; Laumanns, M.; Zitzler, E. Scalable Test Problems for Evolutionary Multiobjective Optimization. In Evolutionary Multiobjective Optimization: Theoretical Advances and Applications; Abraham, A., Jain, L., Goldberg, R., Eds.; Springer: London, UK, 2005; pp. 105–145. ISBN 978-1-84628-137-2. [Google Scholar]
  49. Zhang, Q.; Zhou, A.; Zhao, S.; Suganthan, P.N.; Liu, W.; Tiwari, S. Multiobjective Optimization Test Instances for the CEC 2009 Special Session and Competition. In Special Session on Performance Assessment of Multi-Objective Optimization Algorithms, Technical Report; University of Essex: Colchester, UK; Nanyang Technological University: Singapore, 2008. [Google Scholar]
  50. Jain, H.; Deb, K. An Evolutionary Many-Objective Optimization Algorithm Using Reference-Point Based Nondominated Sorting Approach, Part II: Handling Constraints and Extending to an Adaptive Approach. IEEE Trans. Evol. Comput. 2014, 18, 602–622. [Google Scholar] [CrossRef]
  51. Ma, Z.; Wang, Y. Evolutionary Constrained Multiobjective Optimization: Test Suite Construction and Performance Comparisons. IEEE Trans. Evol. Comput. 2019, 23, 972–986. [Google Scholar] [CrossRef]
  52. Narayanan, S.; Azarm, S. On Improving Multiobjective Genetic Algorithms for Design Optimization. Struct. Optim. 1999, 18, 146–155. [Google Scholar] [CrossRef]
  53. Chiandussi, G.; Codegone, M.; Ferrero, S.; Varesio, F.E. Comparison of Multi-Objective Optimization Methodologies for Engineering Applications. Comput. Math. Appl. 2012, 63, 912–942. [Google Scholar] [CrossRef]
  54. Deb, K. Evolutionary Algorithms for Multi-Criterion Optimization in Engineering Design. Evol. Algorithms Eng. Comput. Sci. 1999, 2, 135–161. [Google Scholar]
  55. Kumar, A.; Wu, G.; Ali, M.; Luo, Q.; Mallipeddi, R.; Suganthan, P.; Das, S. A Benchmark-Suite of Real-World Constrained Multi-Objective Optimization Problems and Some Baseline Results. Swarm Evol. Comput. 2021, 67, 100961. [Google Scholar] [CrossRef]
  56. Cheng, F.Y.; Li, X.S. Generalized Center Method for Multiobjective Engineering Optimization. Eng. Optim. 1999, 31, 641–661. [Google Scholar] [CrossRef]
  57. Steven, G. Evolutionary Algorithms for Single and Multicriteria Design Optimization. Struct. Multidisc. Optim. 2002, 24, 88–89. [Google Scholar] [CrossRef]
  58. Dhiman, G.; Kumar, V. Multi-Objective Spotted Hyena Optimizer: A Multi-Objective Optimization Algorithm for Engineering Problems. Knowl.-Based Syst. 2018, 150, 175–197. [Google Scholar] [CrossRef]
  59. Zhang, H.; Peng, Y.; Hou, L.; Tian, G.; Li, Z. A Hybrid Multi-Objective Optimization Approach for Energy-Absorbing Structures in Train Collisions. Inf. Sci. 2019, 481, 491–506. [Google Scholar] [CrossRef]
  60. Tian, Y.; Cheng, R.; Zhang, X.; Cheng, F.; Jin, Y. An Indicator-Based Multiobjective Evolutionary Algorithm with Reference Point Adaptation for Better Versatility. IEEE Trans. Evol. Comput. 2018, 22, 609–622. [Google Scholar] [CrossRef]
  61. Zhu, Q.; Zhang, Q.; Lin, Q. A Constrained Multiobjective Evolutionary Algorithm with Detect-and-Escape Strategy. IEEE Trans. Evol. Comput. 2020, 24, 938–947. [Google Scholar] [CrossRef]
  62. Zhang, K.; Xu, Z.; Yen, G.G.; Zhang, L. Two-Stage Multiobjective Evolution Strategy for Constrained Multiobjective Optimization. IEEE Trans. Evol. Comput. 2024, 28, 17–31. [Google Scholar] [CrossRef]
  63. Fan, Z.; Li, W.; Cai, X.; Li, H.; Wei, C.; Zhang, Q.; Deb, K.; Goodman, E. Push and Pull Search for Solving Constrained Multi-Objective Optimization Problems. Swarm Evol. Comput. 2019, 44, 665–679. [Google Scholar] [CrossRef]
  64. Tian, Y.; Cheng, R.; Zhang, X.; Jin, Y. PlatEMO: A MATLAB Platform for Evolutionary Multi-Objective Optimization [Educational Forum]. IEEE Comput. Intell. Mag. 2017, 12, 73–87. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.