You are currently viewing a new version of our website. To view the old version click .
Symmetry
  • Article
  • Open Access

29 December 2025

An Improved Multi-Objective Moth Flame Optimization Algorithm Based on r-Domination and Its Application in Parameter Optimization of Multi-Spacecraft Attitude Control

,
,
and
1
School of Mathematics and Statistics, Taishan University, Tai’an 271000, China
2
Key Lab of Industrial Computer Control Engineering of Hebei Province, Yanshan University, Qinhuangdao 066004, China
3
Key Lab of Intelligent Data Information Processing and Control of Hebei Province, Tangshan University, Tangshan 063000, China
4
Key Lab of Intelligent Motion Control System of Tangshan City, Tangshan University, Tangshan 063000, China
Symmetry2026, 18(1), 62;https://doi.org/10.3390/sym18010062 
(registering DOI)
This article belongs to the Section Computer

Abstract

To solve the multi-objective optimization problem (MOP) for parameter optimization of a multiple spacecraft attitude controller, an improved multi-objective moth–flame optimization algorithm based on r-domination (rIMOMFO) is proposed. In rIMOMFO, Cauchy mutation is utilized to produce flames that can improve the global search ability. Based on the difference of fitness variance between the two generations of moths, a hybrid mutation strategy is used to perturb moths, which can improve the population diversity of moths. Meanwhile, an adaptive parameter is applied to the moth position update mechanism, which can further balance the exploitation and exploration. In addition, r-domination is used to lead the search toward the preference of the decision maker. The effectiveness of the rIMOMFO is verified using test problems (CF, UF and WFG). Finally, rIMOMFO is used to optimize the parameters of a distributed cooperative controller for multiple spacecraft attitudes. The experimental topological structure is symmetrical. The results show that the optimized controller can accelerate the convergence speed and improve the control accuracy.

1. Introduction

Cooperative control techno [logy of multiple spacecraft has a wide and important application in the development of national defense and science technology [1,2,3,4]. Distributed synthetic aperture radar, three-dimensional imaging and other earth observation tasks involving multiple spacecraft require each spacecraft to point to a specific position on the earth. The above situation must ensure that each spacecraft can track the preset attitude trajectory [5]. Especially for multiple spacecraft tasks, rapidly tracking accuracy and stability is a very challenging problem.
The distributed cooperative control system for multiple spacecraft attitudes is a multi-variable, strong coupling and nonlinear composite system. For the last several years, many researchers have studied the attitude distributed cooperative control systems of multiple spacecraft. In [6], an adaptive control method is designed to realize the coordinated attitude tracking control of a group of spacecraft without speed. In [7], according to the problem of finite time and velocity free attitude coordinated control in undirected communication graph, a continuous distributed finite time spacecraft attitude coordinated control law without angular velocity measurement is designed, which ensures a group of spacecraft can track a common time-varying reference attitude in a limited time. In [8], to settle the problem of relative position coordination control of a spacecraft formation flight under an undirected communication diagram, a new event-based coordinated control scheme is proposed that reduces the communication load. In addition, researchers have also studied the problem of multi-spacecraft preset time event-triggered attitude consensus control with time-varying state constraints, collaborative control problems, distributed finite-time attitude coordination control of spacecraft formation under multiple constraints on angular velocity and control torque, and spacecraft cluster cooperative obstacle avoidance control [9,10,11,12]. However, different controller parameters produce different control results for the cooperative control of multiple spacecraft. It is an MOP of how to find the appropriate controller parameters and optimize the performance of such multiple spacecraft when there are multiple constraints in the real system.
In recent years, scholars have conducted extensive research on traditional mathematical optimization methods and intelligent optimization methods for complex optimization problems [13,14,15]. Related research has been conducted on intelligent optimization methods. The multi-objective evolutionary algorithms (MOEAs) for complex optimization problems have been deeply studied [15,16,17]. With the development of MOEAs, many excellent algorithms were developed, such as non-dominated sorting genetic algorithm II (NSGA-II) [18], the multi-objective evolutionary algorithm based on decomposition (MOEA/D) [19], strength pareto evolutionary algorithm 2 (SPEA2) [20] and non-dominated sorting genetic algorithm III (NSGA-III) [21]. The MOEAs are composed of an evolutionary algorithm and a multi-objective mechanism. Among them, the multi-objective mechanism includes non dominated sorting methods, decomposition-based mechanisms and so on. For large-scale multi-objective problems with large decision variables, ref. [22] proposes an evolutionary strategy based on a two-stage accelerated search optimizer. Specifically, in the first stage, a convergence optimizer was designed, and in the second stage, a reverse learning differential optimization algorithm was used as the diversity optimizer. For the Constrained Multi-Objective Optimization Problem with Deception Constraints (DCMOP), ref. [23] proposes a novel evolutionary multitasking algorithm using a collaborative multi-step mutation strategy. This algorithm transforms DCMOP into a multitask optimization problem, where the main task is the original DCMOP and the auxiliary task aims to provide effective support for the main task. Specifically, the designed collaborative multi-step mutation strategy includes two contributions to addressing deception constraints. For problems with multiple objectives, strong constraints, large-scale, and mixed variables, ref. [24] proposes a multitask and multi-objective optimization framework that combines evolutionary algorithms and mathematical programming. In this framework, the main task uses an evolutionary algorithm with global search capability to solve multi-objective optimization problems. Auxiliary tasks utilize mathematical programming methods with robust linear constraint handling capabilities to solve multiple weighted single objective optimization problems. In addition, swarm-intelligence-based search algorithms for multi-objective optimization have also been developed and applied to a variety of multi-objective optimization problems [25,26,27].
Since the 1950s, inspired by the laws of biological evolution and optimization phenomena in nature, many intelligent optimization algorithms have been proposed, such as particle swarm optimization (PSO) [28] and genetic algorithm (GA) [29]. The differential evolution (DE) [30] algorithm appeared earlier, but there are many variants that can be utilized to solve different problems. In the past two decades, meta-heuristic algorithms have attracted widespread attention, such as Coyote Optimization Algorithm (COA) [31], Sand Cat Swarm Optimization (SCSO) [32], firefly algorithm (FA) [33], Equilibrium optimizer (EO) [34], Heap-based Optimizer algorithm (HBOA) [35], Cuckoo Search (CS) [36], Sine Cosine Algorithm (SCA) [37], Bacterial Foraging Optimization algorithm (BFO) [38], Barnacles Mating Optimizer (BMO) [39] and moth–flame optimization (MFO) [40].
These intelligent optimization algorithms are used to solve multi-objective optimization problems, such as PSO [41], Whale Optimization Algorithm [42], Grey Wolf Optimization (GWO) [43], CS [44] and Fish Algorithm (FA) [45]. It is worth noting that the MFO is effective in solving some optimization problems [46], but it can become stuck in bad local optima because it focuses on exploitation rather than exploration [47].
To solve the above problems, an improved moth–flame optimization (IMFO) algorithm is proposed. In IMFO, new moth position updating and flame generation mechanisms are designed. Cauchy mutation is utilized to produce flames that lead the moths to their updated positions; a hybrid mutation operator is used to perturb the positions of the moths. Among them, variations in the number of flames and moths can be dynamically adjusted by calculating the difference of fitness variance between the two generations of moths. In addition, an adaptive parameter is applied to the moth position update mechanism to further balance the exploitation and exploration. We combine IMFO with r-dominated sorting to produce rIMOMFO. Moreover, rIMOMFO is utilized to solve the MOP of parameter optimization for a multiple spacecraft attitude controller. To sum up, the main contributions of this paper are as follows:
(1)
Cauchy mutation is utilized to produce flames that lead the moths to their updated positions; a hybrid mutation operator is used to perturb the positions of the moths. Among them, variations in the number of flames and moths can be dynamically adjusted by calculating the difference of fitness variance between the two generations of moths. In addition, an adaptive parameter is applied to the moth position update mechanism to further balance exploitation and exploration.
(2)
The IMFO and r-dominated sorting are combined to produce a multi-objective evolutionary algorithm called rIMOMFO. rIMOMFO is evaluated by benchmark test problems (CF, UF and WFG). The results show that the rIMOMFO algorithm is feasible and effective.
(3)
The rIMOMFO algorithm is used to optimize the parameters of a distributed cooperative controller for multiple spacecraft attitudes. The results show that the optimized controller can improve the control accuracy and accelerate the convergence speed.
The rest of the paper is organized as follows: In Section 2, a multi-objective optimization model of multiple spacecraft attitude distributed cooperative control is established. In Section 3, the rIMOMFO is discussed in detail. In Section 4, the benchmark problems are utilized to test the performance of the rIMOMFO algorithm and how it compares with other algorithms. The rIMOMFO is used to optimize the control parameters of the attitude cooperative controller for multiple spacecraft in Section 5. In Section 6, the conclusions and future works of this study are given.

2. Problem Formulation

2.1. System Description

In this paper, the Modified Rodrigues Parameters (MRPs) [48] are utilized to describe the rigid body attitude motion of each spacecraft. The dynamic equations of the i ( i = 1 , 2 , , m ) spacecraft are as follows:
σ ˙ i = G i σ i ω i J i ω ˙ i = ω i × J i ω i + u i + ε i
where σ i = σ i 1 , σ i 2 , σ i 3 T represents the ith spacecraft attitude, ω i = ω i 1 , ω i 2 , ω i 3 T is the angular velocity of the ith spacecraft body coordinate system relative to the inertial coordinate system and ω i × denotes a skew-symmetric matrix of ω i , which is given in Equation (2). u i = u c i + u f i is the actual input torque, u c i is the expected input torque of the spacecraft (to be designed) and u f i is the actuator fault, where the fault is expressed as deviation fault and efficiency fault. u f i = f a i + Ξ i I u c i , f a i is the actuator deviation fault and f a i is bounded; Ξ i = d i a g Ξ i 1 , Ξ i 2 , Ξ i 3 is the efficiency factor of the actuator and 0 < Ξ i j 1 , j = 1 , 2 , 3 . When Ξ i j = 1 , the actuator of the ith spacecraft is in a healthy state; when Ξ i j = 0 , the actuator of the ith spacecraft fails completely; when 0 < Ξ i j 1 , there is partial failure of the actuator of the ith spacecraft. ε i R 3 is the external interference torque, G i = 1 2 1 σ i T σ i 2 I 3 + σ i × + σ i T σ i , I 3 is the unit matrix of order 3, σ i × is the skew-symmetric matrix of σ i and J i R 3 × 3 is the moment of inertia.
ω × = 0 ω 3 ω 2 ω 3 0 ω 1 ω 2 ω 1 0
According to reference [48], the Euler–Lagrange equation of multiple spacecraft kinematics is obtained as follows:
M i ( q i ) q ¨ i + C i ( q i , q ˙ i ) = u i + ε i
where M i ( q i ) = G i T ( q i ) J i G i 1 ( q i ) , C i ( q i , q ˙ i ) = G i T ( q i ) ( J i G i 1 ( q i ) q ˙ i ) × G i 1 ( q i ) G i T ( q i ) J i G i 1 ( q i ) G ˙ i ( q i ) × G i 1 ( q i ) .

2.2. Distributed Coordinative Controller Based on Super Twisting and an Integral Sliding Surface

To cause the attitude of multiple spacecraft to track the preset trajectory quickly with external disturbance and actuator failure, a distributed cooperative controller based on super twisting and an integral sliding mode surface is designed.
Defining x i , 1 = q i , x i , 2 = q ˙ i , the kinematics Equation (3) of the ith spacecraft is transformed into
x ˙ i , 1 = x i , 2 x ˙ i , 2 = f i ( x i , 1 , x i , 2 ) + g i ( x i , 1 ) u c i + d i y i = x i , 1
where f i ( x i , 1 , x i , 2 ) = M i 1 ( x i , 1 ) C i ( x i , 1 , x i , 2 ) x i , 2 ; g i ( x i , 1 ) = M i 1 ( x i , 1 ) ; d i = M i 1 ( x i , 1 ) ( ε i + u f i ) .
The consensus errors are given
χ i , 1 = j = 1 n a i j x i , 1 x j , 1 + b i x i , 1 q 0 χ i , 2 = j = 1 n a i j x i , 2 x j , 2 + b i x i , 2
Define the sliding-mode surface
S i = x i , 2 + 0 t c i , 1 s i g β 1 χ i , 1 τ + c i , 2 s i g β 2 χ i , 2 τ d τ
where S i = S i , 1 , S i , 2 , S i , 3 T , β 1 = β 2 2 β 2 , β 2 0 , 1 , c i , 1 > 0 , c i , 2 > 0 , and satisfy that polynomial λ 2 + c i , 2 λ + c i , 1 is Hurwitz, i = 1 , 2 , , m .
Based on reference [49], the distributed consensus controller is designed as
u c i = ( g i ( x i , 1 ) ) 1 u i n o m + u i s w
u i n o m = f i ( x i , 1 , x i , 2 ) + d ^ i + c i , 1 s i g β 1 χ i , 1 + c i , 2 s i g β 2 χ i , 2
u i s w = δ i , 1 ϕ i S i 0 t z ˙ i S i d τ
where ϕ i S i = μ i , 1 s i g 1 2 S i + μ i , 2 S i , z ˙ i S i = δ i , 2 μ i , 1 2 2 s i g 0 S i + μ i , 2 2 S i + 3 2 μ i , 1 μ i , 2 s i g 1 2 S i = δ i , 2 Γ i ϕ i S i , with Γ i = μ i , 1 2 d i a g S i , 1 1 2 , S i , 2 1 2 , S i , 3 1 2 + μ i , 2 I 3 , δ i , 1 > 0 , δ i , 2 > 0 , μ i , 1 > 0 and μ i , 2 > 0 are constants. d ^ i is the estimation of lumped system interference  d i . For the ith spacecraft, the disturbance observer is designed as
x ^ ˙ i , 2 = f i ( x i , 1 , x i , 2 ) + g i ( x i , 1 ) u c i + r i s i g γ i e i + d ^ i d ^ ˙ i = η i s i g n σ d i + α ^ i s i g n σ d i
where η i > 0 , γ i 0 , 1 , α ^ i is the estimate of α i , σ d i = e ˙ i + r i s i g γ i e i and e i = x i , 2 x ^ i , 2 .

2.3. The MOP Formulation for the Attitude Control of Multiple Spacecraft

To make the designed controller improve the control accuracy and accelerate the convergence speed of the system, in this paper, f a ( a = 1 , 2 , 3 ) and f b ( b = 4 , 5 , 6 ) are selected as the two kinds of objective functions, and the detailed description is as follows:
(1) f a is the adjust time for the three axes of the follower spacecraft to reach the specified accuracy. And it is used to embody the rapidity of the multiple spacecraft cooperative control system.
f a ( x ) = t s , I f t h e f o l l o w e r c a n r e a c h a c e r t a i n a c c u r a c y i n t h e s e t t i m e t c , o t h e r w i s e
where t s is adjustment time and t c is penalty time.
(2) f b is the root mean square error of the attitude of followers after stabilization. And it is used to embody the stability of the multiple spacecraft cooperative control system.
f b ( x ) = 1 N s n = 1 N s max { ( q i , 1 ( b ) q 0 ( b ) ) 2 }
where N s is the sample times after spacecraft stabilization, i ( i = 1 , 2 , , m ) .
The MOP is to search the minimum value of f a and f b . And the mathematical expression of the optimization problem in this paper is as follows:
min x Ω m F ( x ) = { f 1 ( x ) , f 2 ( x ) , f 3 ( x ) , f 4 ( x ) , f 5 ( x ) , f 6 ( x ) }
s u b j e c t t o 0 < δ i 1 2 , 0 < δ i 2 2 0 < μ i 1 2 , 0 < μ i 2 2 , i = 1 , 2 , 3 , 4 β 1 = β 2 2 β 2 , β 2 0 , 1
where x = δ 11 , δ 21 , δ 31 , δ 41 , δ 12 , δ 22 , δ 32 , δ 42 , μ 11 , μ 21 , μ 31 , μ 41 , μ 12 , μ 22 , μ 32 , μ 42 , β 2 and Ω m is the feasible region, including all controller parameters to be found.

3. Improved Multi-Objective Moth–Flame Algorithm Based on r-Domination (rIMOMFO)

3.1. Moth–Flame Optimization Algorithm and r-Dominance Relationship

3.1.1. Moth–Flame Optimization Algorithm

The MFO algorithm is a population-based swarm intelligence algorithm. Moreover, it is inspired by the horizontal positioning mechanism of the moths in flight [40]. There are two populations in MFO: moths (M) and flames (F). They and their corresponding fitness values are shown below.
M = M 1 M 2 M n = m 11 m 12 m 1 d m 21 m 22 m 2 d m n 1 m n 2 m n d
OM = O M 1 O M 2 O M n
F = F 1 F 2 F n = f 11 f 12 f 1 d f 21 f 22 f 2 d f n 1 f n 2 f n d
OF = O F 1 O F 2 O F n
where n the population number of moths/flames and d is the number of dimensions. O M i and O F i represent the fitness function value of the corresponding moth M i = m i 1 , m i 2 , , m i d and F i = f i 1 , f i 2 , , f i d , i { 1 , 2 , , n } , respectively.
The population of moths is initialized as follows [50]:
m i j = l b i + w j u b i l b i
where j = 1 , 2 , , d ; w j is a random number in [0,1]; and u b i and l b i indicate the upper and lower bounds of the i-th variables ( ub = u b 1 , u b 2 , , u b n , lb = l b 1 , l b 2 , , l b n ).
The position updating formulas of moths are as follows:
M i ( l + 1 ) = D i ( l ) · e b t · cos 2 π t + F i ( l ) , i f n o D i ( l ) · e b t · cos 2 π t + F f n o ( l ) , i > f n o
where b = 1 ; t is a random number in [r,1]; r = 1 + l ( 1 / L ) ; l and L represent the current number and the maximum number of iterations, respectively; f n o denotes the number of adaptive flames and is calculated by Equation (19); and D i indicates the distance between the ith flame and its corresponding moth. The formula for D i is as follows:
D i ( l ) = | F i ( l ) M i ( l ) | , i f n o | F f n o ( l ) M i ( l ) | , i > f n o
Meanwhile, to balance exploration and exploitation, MFO uses the adaptive flame number method to determine the number of flames in each iteration. The mathematical expression is as follows:
f n o = r o u n d ( n l ( n 1 ) L )

3.1.2. r-Dominance Relation

The r-dominance (reference-solution-based dominance) is the fusion of the Pareto dominance principle and the reference point method (RPM). Its advantage is that it keeps the sequence induced via Pareto advantage and prefers to choose solutions closer to the reference points. The definition of r-dominance relation is given as
Definition 1 
([51]). The r-Dominance Relation Assuming a population of individuals P, a reference vector g and a weight vector w, a solution x is said to r-dominate a solution y( x r y ) if one of the following statements holds true: (1) x dominates y in the Pareto sense; (2) x and y are Pareto-equivalent and D ( x , y , g ) < δ , where δ [ 0 , 1 ] and
D ( x , y , g ) = D i s t ( x , g ) D i s t ( y , g ) D i s t max D i s t min
D i s t max = M a x z P D i s t ( z , g )
D i s t min = M i n z P D i s t ( z , g )
δ is termed the non-r-dominance threshold.
The r-dominance is mainly used to find the partial order between Pareto-equivalent solutions. And it can partially distinguish non-dominant solutions based on a given expected level vector. Therefore, it not only makes r-dominance more stressful than Pareto dominance, but also incorporates the preference of decision makers in the selection process.

3.2. IMFO Based on a Dynamic Adjustment Mechanism and Hybrid Mutation Strategy

To improve the search performance of the MFO, Cauchy mutation is utilized to produce flames that can lead moths to their updated positions; a hybrid mutation operator is used to perturb the positions of moths. Among them, the variations in the number of flames and moths can be dynamically adjusted by calculating the difference of fitness variance between the two generations of moths. In addition, an adaptive parameter is applied to the moth position update mechanism to further balance the exploitation and exploration. The detailed IMFO description is as follows:
(1)
Dynamic adjustment mechanism
According to the difference in fitness variance between the two generations of the moths, the number of disturbed flames and moths is adjusted dynamically. Studying the overall change of fitness value for a moth population can track the state of the population. The fitness variance of a moth is expressed as follows:
σ 2 = i = 1 n O M i O M ¯ f
f = max 1 , max O M i O M ¯
K = 0 , l = 1 min n 1 , f l o o r l n 3 σ l 2 σ l 1 2 , l 2
where O M ¯ is the current average fitness of the moths’ population, f is a standardized calibration factor, K is the number of disturbed flames, σ l 2 is the variance of the fitness values of the lth generation moth population and f l o o r means taking the greatest integer less than or equal to a given real number.
In this paper, according to the fitness variance of moth population in two iterations, the number of flames and moths to be disturbed is dynamically adjusted. In the early stage of the search, more flame is mutated to improve the search ability. And the more moths gathered, the more are mutated. This mechanism can improve the ability to jump out of the local optimum.
(2)
Improvement of flame generation
Based on the MFO algorithm, K flames are selected for Cauchy mutation and used to lead the moths to their updated positions. This method is beneficial to jump out of the local optimum. Cauchy mutation is used in many meta-heuristic algorithms, such as DE [52] and KH [53]. In this paper, Cauchy mutation is used to avoid premature convergence and improve the global search ability of MFO. The Cauchy distribution function can be described as follows:
y ( x ) = 1 π k 0 ( x x 0 ) 2 + k 0 2
where k 0 is the scale parameter, x 0 is the position of the peak and the value here is 0. The density function of Cauchy distribution is shown as follows:
Y ( x ) = 1 π arctan ( x k 0 ) + 1 2
In the iteration process, the best flame of each iteration is selected as the base flame, and then the variant flames are generated via a Cauchy operator. And the mathematical expression of variant flame (P) is as follows:
P j = F b e s t j + F ¯ j · C ( a , b 2 )
where P j is the j-th variable of P ; the number of variation flames is K under this strategy; F b e s t j indicates the jth dimension of the best flame; and a is set to 0. To balance exploration and exploitation, the value form of b is b = 0.6 0.4 ( l / L ) . F ¯ j is given as follows:
F ¯ j = ( i = 1 n f i j ) / n
(3)
Improved moth position updating
An adaptive parameter, which is generated by the cosine function, is introduced into the original MFO position update mechanism to further balance the exploitation and exploration. And the improved position update mechanism of moths is as follows:
M i ( l + 1 ) = ( 1 c ) · D i · e b t · cos 2 π t + c · F i ( l ) , i f n o D i · e b t · cos 2 π t + F f n o ( l ) , i > f n o
c = c m i n + ( c m a x c m i n ) · cos ( 2 π l / L ) ;
where c m a x and c m i n are the maximum and minimum values of the cosine coefficient, respectively.
In IMFO, the moths are mutated by a mixed mutation strategy that is a combination of Gauss mutation and Lévy mutation. And the number of moths to mutate is determined through the dynamic adjustment mechanism. This mutate operator expression is as follows:
M i ( l + 1 ) = M i ( l + 1 ) + F ¯ · ( C ( a , b 2 , d ) + L ( β ) )
F ¯ = ( i = 1 n f i ) / n
where L ( β ) is the Lévy random number and can be described as [54]
L ( β ) φ × μ ν 1 / β
where μ and ν are the standard normal distributions, β = 3 / 2 .
φ = Γ 1 + β × sin π × β / 2 Γ 1 + β 2 × β × 2 β 1 2 1 / β
where φ is known as a standard Gamma function.
The framework of the IMFO is given in Algorithm 1, and the steps of IMFO are given as follows:
Step 1 Initialization
Initialize the variables (dimensions): number (d) of moths per flame, population numbers (n) of moths/flames, maximum number of function evaluations ( m a x _ F E s ) and the current number of function evaluations ( F E s ), l, L.
Step 2 Randomly initialize moth populations M using Equation (19).
Step 3 Continue the iterative process or stop.
Run Step 4 to Step 9 until the F E s has reached the m a x _ F E s .
Step 4 Calculate the number of flames guiding moths according to Equation (22).
Step 5 Evaluate the fitness function values of moths for the current iteration.
Step 6 Update flames
The flame population is generated by sorting the moths of two adjacent generations. And K flames are selected for Cauchy mutation and used to guide the moths to update their position.
Step 7 Judge whether the fitness function value of the optimal flame reaches the set value. If it does, output the optimal solution F b e s t and terminate the program; otherwise, continue to the next step.
Step 8 Update moths
According to the improved position updating mechanism, M i is updated using Equation (33).
Step 9 Gauss mutation and Lévy mutation
The moths are mutated by a combination of Gauss mutation and Lévy mutation according to Equation (35), and the number of moths to mutate is determined through the dynamic adjustment mechanism.
Step 10 Return the best solution F b e s t .
Algorithm 1: Pseudocode of IMFO
1: Initialize m a x _ F E s , n , d , F E s = 0 , l = 1 , L = c e i l ( m a x _ F E s / n ) ;
2: Randomly initialize M using Equation (19);
3: while  F E s m a x _ F E s
4:       Calculate f n o using Equation (22);
5:       Evaluate the fitness value of moths: OM = F i t n e s s F u n c t i o n ( M ) ; F E s = F E s + n ;
6:       If  l = = 1
7:              F = s o r t ( M ) ; OF = s o r t ( OM ) ;
            Select the optimum solution F b e s t ;
8:       else
9:              F = s o r t ( M ( l 1 ) , M ( l ) ) ;
             OF = s o r t ( OM ( l 1 ) , OM ( l ) ) ;
            Select the optimum solution F b e s t ;
10:       end
11:       Calculate the number of disturbed flames K using Equation (28)
12:       If  K n
13:        K = n 1 ;
14:       end
15:       Update the flames using (31);
16:       for  i = 1 : n
17:             for  j = 1 : d
18:                   Calculate r and t; Calculate D i using Equation (21);
19:                   Update the variables in M i using Equation (33);
20:             end
21:       end
22:       for  i = n K + i : n
23:       Update the moths? Position by Cauchy mutation and Lévy mutation using (35);
24:       end
25:        l = l + 1 ;
26: end while
27: return the best solution F b e s t .

3.3. Framework of rIMOMFO

The rIMOMFO is an MOEA based on IMFO and r-dominance. The IMFO algorithm overcomes the local optimum and accelerates the convergence speed. The main improvements are as follows: (1) Cauchy mutation is utilized to produce flames that can lead moths to their updated positions. (2) The moths are mutated by a combination of Gauss mutation and Lévy mutation. (3) An adaptive parameter is applied to the moth position update mechanism. (4) The variation of flame and moth numbers is determined by the variance of the fitness values of two generations of moths. In addition, the r-domination has the ability of preference-guided search based on the decision maker and is oriented to the Pareto optimal frontier in a reasonable number of evaluations. The procedure and flowchart for rIMOMFO are shown in Algorithm 2 and Figure 1, and the steps of rIMOMFO are given as follows:
Figure 1. The flowchart of rIMOMFO.
Step 1 Initialization
Initialize the d, n, m a x _ F E s , F E s , l, L and δ .
Step 2 Randomly initialize parent population M using Equation (19), and each objective function of every moth is evaluated.
Step 3 Use a non-r-dominated sort method to sort the M and calculate crowding distance [18] for each front.
Step 4 Continue the iterative process or stop.
Run Step 5 to Step 6 until the F E s has reached the m a x _ F E s .
Step 5 The offspring population M O is generated by the position update mechanism (Equations (33) and (35)) of the IMFO algorithm that is proposed in the previous section.
Step 6 Merge the parent population M and offspring population M O to create U( U = M M O ); sort the U and calculate the crowding distance for each front. Then, replace M for the first n members of U.
Step 7 Return the best solutions.
Algorithm 2: rIMOMFO
1: Initialize m a x _ F E s , n , d , F E s = 0 , l = 1 , L = c e i l ( m a x _ F E s / ( n + 7 ) ) , δ ;
2: Randomly initialize parent population M using Equation (19);
3: Sort the M based on the non-r-dominated sort method, F E s = F E s + n ;
4: Compute crowding distance [18] for each front;
5: while  F E s m a x _ F E s
6:    Generate the offspring population M O by the position update mechanism (Equations (33) and (35)) of IMFO;
7:    Merge parent population M and offspring population M O to create U( U = M M O );
8:    Sort the U based on the non-r-dominated sort method;
9:    Compute crowding distance for each front;
10:    Replace M for first n members of U;
11:     l = l + 1 ; F E s = F E s + n ;
12: end while
13: return the best solutions

3.4. Computational Complexity of rIMOMFO

The computational complexity of rIMOMFO in one iteration mainly consists of the following two aspects: position update of moths and flames, and non-r-dominated sorting and crowding distance. The complexity order of moth flame position updating is O ( n 2 ) . The computational complexity of non-r-dominated sorting is O ( M n 2 ) ( M is the number of objectives); the complexity of crowding distance is O ( M l o g n ) . Therefore, the complexity order of rIMOMFO is O ( M n 2 ) .

4. Performance Test and Analysis

To verify the effectiveness of rIMOMFO, this section carries out simulation analysis through benchmark test problems. Firstly, benchmark test problems and performance metrics are introduced. Secondly, the test results are given and analyzed.

4.1. Benchmark Test Problems

In this paper, the WFG [55] test problems and 20 test problems [56] in CEC 2009 are used to test the performance of rIMOMFO. WFG is a complex optimization problem, including convexity or concavity and other forms; UF is used for multi-objective unconstrained problems; CF is used for multi-objective constraint problems. In this study, the number of objectives ( M ) is set to 2, 3 (UF,CF) and 4, 7 (WFG), which tests the performance of the algorithm on MOPs and MaOPs, respectively.

4.2. Performance Metrics

In this paper, inverted generational distance (IGD) and hypervolume (HV) are utilized to evaluate UF and CF; generational distance (GD) and spacing (S) are used to evaluate WFG. Their detailed description is as follows:
(1)
Generational distance (GD)
GD is utilized to embody the convergence of a solution set generated by MOAs [57]. A smaller GD value indicates that the convergence of an algorithm is better. If the value of GD is equal to 0, it indicates that all the solutions generated by the algorithm are on the Pareto front. GD is given as
G D A , V = 1 A M j A min M i V   d M i , M j 2
where V = v 1 , v 2 , v n is a set of n targeted points. A is the points of the solution set gained by a specific algorithm. d M i , M j indicates the Euclidean distance between M i and M j .
(2)
Inverted generational distance (IGD)
IGD can simultaneously calculate the convergence and distribution of solutions [58]. The smaller the IGD value, the better the set A. GD is defined as
I G D A , V = 1 V i = 1 V min M j A d v i , M j
(3)
Hypervolume (HV)
HV can evaluate the convergence and distribution of the solution set at the same time [59]. A larger HV value indicates that the convergence and distribution of the algorithm is better. HV is given as
H V A , r = v o l u m e f A f 1 , r 1 × f m , r m
where r = r 1 , r 2 , , r m T is a reference point. HV of the A with regard to r is the volume of the region dominated by A and bounded by r [60].
(4)
Spacing (S)
The Spacing metric evaluates the uniformity of the distribution of non-dominated solutions [61]. A smaller S value indicates that the distribution of an algorithm is better. Spacing can be formulated as follows:
S = 1 n 1 i = 1 n d ¯ d i 2
d i = k O k i min m = 1 M O M m i O M m k
where d ¯ is the average of the distances d i ( i = 1 , , n ). Spacing is the standard deviation of the distance between two continuous solutions of a set O.

4.3. Test Design

To better analyze the performance of rIMOMFO, two groups of comparative tests are utilized. Firstly, based on CF and UF, the proposed rIMOMFO algorithm is compared with other algorithms (NSGAII [18], MAEA [62], MOEAD [19], MOEADDAE [63], CMOPSO [64], dMOPSO [65], MOPSO [66], NSLS [67], NSGAIISDR [68] and NSMFO [69]). Secondly, the rIMOMFO algorithm is tested based on WFG and compared with other algorithms (NSGAII, NSGAIII, MAEA, MOEADDAE, dMOPSO and MOPSO). We reference their original paper for the parameter setting of these comparison algorithms in Table 1. Among them, the selection principle of δ = 0.1 refers to the reference literature [51]. c m a x = 0.9 , c m i n = 0.5 . For the test problems of objective function to 2,3,4,7, the population size (n) is set to 100, 100, 286 and 294, respectively. The m a x F E s for all test problems is n 1000 . For WFG test problems, on the basis of [70], the number of position-related variables is k = 18 and the number of distance-related variables is l = 14 . Additionally, the number of decision variables is set as d = k + l = 32 . In the two groups of comparative tests, each algorithm is run independently 30 times for each problem.
Table 1. Parameter settings.
In this paper, five performance evaluation methods are used: Mean, STD, Wilcoxon rank-sum test, multiple-problem Wilcoxon signed-rank test and Friedman rank test. Mean is the average fitness function value of 30 independent runs; STD is the standard deviation of the fitness function value of independent runs; the Wilson rank-sum test is utilized to characterize whether the rIMOMFO is significantly better than other comparative algorithms. In this paper, the significance level of 0.05 is used. “−” means that the rIMOMFO algorithm is better than other algorithms, “+” means that the comparison method is better than rIMOMFO, “=” means that the rIMOMFO and the comparison method are at the same level. In addition, the multiple-problem Wilcoxon signed-rank test is also given for the two groups of simulations. “R+” indicates a positive rank and “R−” indicates a negative rank. “+” and “−” indicate the better than and worse than other algorithms, respectively. At the same time, they indicate that rIMOMFO is different from other algorithms. “=” indicates that the rIMOMFO algorithm is at the same level as the other algorithms and that there is no significant difference. For the Friedman rank test, “Ave” represents the average ranking of the algorithm and “Rank” represents the final ranking of the algorithm. In addition, the bold data in the table represents the results of the optimal algorithm on the corresponding test problem.

4.4. Test Results and Analysis

In this subsection, the WFG and CEC2009 problems are utilized to test the rIMOMFO algorithm, which is compared with the other algorithms. Table 2 and Table 3 show the average and standard deviation for the IGD and HV of 12 algorithms in the CEC2009 test problem, respectively. Table 4, Table 5, Table 6 and Table 7 show the results of the multiple-problem Wilcoxon signed-rank test and Friedman rank test for the IGD and HV of 12 algorithms. The results of nine algorithms on the WFG test are shown in Table 8, Table 9, Table 10, Table 11, Table 12, Table 13, Table 14, Table 15, Table 16, Table 17, Table 18 and Table 19. The tables contain the average, standard deviation, multiple-problem Wilcoxon signed-rank test and Friedman rank test of GD and S of 4 objectives and 7 objectives.
Table 2. Comparison results on IGD value with Mean and STD for CEC 2009.
Table 3. Comparison results on HV value with Mean and STD for CEC 2009.
Table 4. Results(IGD) of multiple-problem Wilcoxon signed-rank test.
Table 5. Friedman rank test results(IGD) of the 12 algorithms.
Table 6. Results(HV) of multiple-problem Wilcoxon signed-rank test.
Table 7. Friedman rank test results(HV) of the 12 algorithms.
Table 8. Results on GD value with Mean and STD for WFG with 4 objectives.
Table 9. Results(GD) of multiple-problem Wilcoxon signed-rank test.
Table 10. Friedman rank test results(GD) of the 9 algorithms.
Table 11. Results on S value with Mean and STD for WFG with 4 objectives.
Table 12. Results(S) of multiple-problem Wilcoxon signed-rank test.
Table 13. Friedman rank test results(S) of the 9 algorithms.
Table 14. Results on GD value with Mean and STD for WFG with 7 objectives.
Table 15. Results(GD) of multiple-problem Wilcoxon signed-rank test.
Table 16. Friedman rank test results(gd) of the 9 algorithms.
Table 17. Results on S value with Mean and STD for WFG with 7 objectives.
Table 18. Results(S) of multiple-problem Wilcoxon signed-rank test.
Table 19. Friedman rank test results(S) of the 9 algorithms.
In Table 2, rIMOMFO is better than the other 11 algorithms in CF5, UF7 and UF10, and it has the same level as the other algorithms in the other test problems. NSMFO and NSIMFO are inferior to rIMOMFO in 12 test problems and have the same performance in 7 test problems. IMFO has the same or better performance than MFO in most test problems, which shows that the improvement of IMFO is very effective. It can be seen intuitively in Table 5 that the ranking of rIMOMFO is 3; the average ranking is 4.11. MAEA and NSGAII are the first and second, and the ranking of the other nine algorithms is behind rIMOMFO. At the same time, it can be seen in Table 4 that although MAEA and NSGAII are better than rIMOMFO, there is no significant difference between them and rIMOMFO. In Table 5, rIMOMFO is superior to the other 11 algorithms in CF5, UF7, UF8 and UF10. The overall performance of NSGAII and MOEAD in CEC 2009 is better than that of rIMOMFO, but they have similar convergence and distribution in most test problems. In Table 7, the ranking of rIMOMFO is 3, with an average ranking of 4.16. For the first- and second-ranked MAEA and NSGAII, it can be seen in Table 6 that there is no significant difference between them and rIMOMFO.
In Table 8, rIMOMFO is superior to the other 11 algorithms in WFG3 and WFG8. And it also shows good competitiveness in other test problems. It can be seen intuitively in Table 10 that the ranking of rIMOMFO is 4; the average ranking is 4.22. MOEADDAE, MAEA and NSGAII are the first, second and third, and the other five algorithms are ranked behind rIMOMFO. At the same time, Table 9 shows that although MOEADDAE, MAEA and NSGAIII are better than rIMOMFO, there is no significant difference between them and rIMOMFO. In addition, rIMOMFO shows good performance on all problems except WFG9 in Table 11. In Table 13, the ranking of rIMOMFO is 1, with an average ranking of 1.00. From Table 12, there is a significant difference between rIMOMFO and the other algorithms. For the WFG problems with 7 objectives, Table 14 and Table 17 show that the results of rIMOMFO are also competitive. It can be seen intuitively in Table 16 that the ranking of rIMOMFO is 4; the average ranking is 4.22. NSGAIII, MAEA and MOEADDAE are the first, second and third, and the ranking of the other five algorithms is behind rIMOMFO. At the same time, it can be seen in Table 15 that although NSGAIII, MAEA and MOEADDAE are better than rIMOMFO, there is no significant difference between them and rIMOMFO. In addition, in Table 19, the ranking of rIMOMFO is 1, with an average ranking of 1.78. From Table 18, there is a significant difference between rIMOMFO and the other algorithms.
Therefore, through the above tests, it can be concluded that the proposed rIMOMFO method is competitive with the other 12 methods in terms of search performance.

4.5. Ablation Experiments

To verify the effectiveness of each module of the algorithm proposed in this article, ablation experiments were conducted. Table 20 presents the experimental results, indicating the algorithm names corresponding to each improved mechanism. Among them, ✓ indicates the adoption of corresponding strategies, while × indicates the non adoption of corresponding strategies. From the table, it can be seen that rIMOMFO has better convergence and diversity than rIMOMFO-DHI, rIMOMFO-DH and rIMOMFO-I.
Table 20. Comparative results for ablation experiments.

5. Application of rIMOMFO for Parameter Optimization of a Multiple Spacecraft Attitude Controller

In this section, the rIMOMFO algorithm is utilized to solve the MOP of the parameter optimization of a multiple spacecraft attitude controller. Firstly, spacecraft parameters are given. Then, the effectiveness of the rIMOMFO algorithm is verified. Finally, the attitude coordinative control performance of the multiple spacecraft controller is obtained based on the optimized parameters.

5.1. Simulation Conditions

The interaction between different spacecraft is symmetrical, and the simulation topology presented in this paper is also symmetrical. In this paper, a multiple spacecraft model with one virtual leader and four followers is adopted. Figure 2 [71] shows the communication topology of the multiple spacecraft model.
Figure 2. Communication topology of multiple spacecraft model.
The inertial matrices of the four followers are J 1 = d i a g 17 , 12 , 9 , J 2 = d i a g 14 , 13 , 10 , J 3 = d i a g 20 , 10 , 12 and J 4 = d i a g 15 , 9 , 16 . The virtual leader’s posture is q 0 = 0.5 , 0 , 0 T . The external interference of the system is ε i = 0.1 cos 0.2 t , sin 0.3 t , cos 0.4 t T N·m. The efficiency factor of the actuator fault is Ξ i = d i a g 0.2 , 0.4 , 1 . The deviation fault is f a i = f a i 1 , f a i 2 , f a i 3 T , f a i 1 = 0.05 0.9 + 0.1 sin t / 10 + 0.01 ϖ N·m, f a i 2 = 0.05 0.9 + 0.1 cos t / 10 + 0.015 ϖ N·m, f a i 3 = 0.05 0.9 + 0.1 sin t / 10 + 0.02 ϖ N·m; ϖ 1 , 1 is random number. The parameters of the adaptive sliding mode interference observer are r i = 0.05 , γ i = 0.5 , η i = 0.1 and i , 1 = 0.2 .
If parameter optimization is not used to determine the controller parameters of multiple spacecraft, the parameters need to be determined through the controller design process and experience. This method requires substantial knowledge of controller design and multiple spacecraft systems and takes a lot of time. In this paper, four algorithms (rIMOMFO, rNSMFO, NSIMFO and NSMFO) are utilized to solve the MOP of a multiple spacecraft system. rNSMFO, NSIMFO and NSMFO are the most similar algorithms to rIMOMFO, which can further illustrate the effectiveness of rIMOMFO in practical applications. The population size is 40 and the m a x F E s is 4000. The other parameters of these algorithms are given in the previous section.

5.2. Results and Analysis

The root mean square error between 10 s and 15 s is used to reflect the stability of attitude tracking; 10 s is a hypothetical threshold of the maximum adjust time to reach stability. Figure 3 shows that all the solutions generated by rIMOMFO are lower than the threshold. The other three methods are not be stable within 8 s. This means that it is difficult for them to solve the multiple spacecraft optimization problem in a given evaluation range. Therefore, the proposed rIMOMFO algorithm has the ability to solve the multiple spacecraft optimization problem.
Figure 3. Solutions obtained by rIMOMFO, rNSMFO, NSIMFO and NSMFO.
This section gives the running results of these four algorithms, and the selected solutions are given in Table 21 and Table 22. According to these solutions, the performance of the attitude coordinative control of multiple spacecraft is shown in Figure 4, Figure 5, Figure 6 and Figure 7.
Table 21. Solutions from rIMOMFO, rNSMFO, NSIMFO and NSMFO.
Table 22. Objectives value of the corresponding solution.
Figure 4. Multiple spacecraft attitudes and their derivative tracking curves with rIMOMFO.
Figure 5. Multiple spacecraft attitudes and their derivative tracking curves with rNSMFO.
Figure 6. Multiple spacecraft attitudes and their derivative tracking curves with NSIMFO.
Figure 7. Multiple spacecraft attitudes and their derivative tracking curves with NSMFO.
Figure 3 shows the attitude tracking curve based on rIMOMFO. q i , j and q ˙ i , j are the j-th attitude and its derivatives of the i-th spacecraft. Figure 4 shows three states of the four spacecraft attitudes and their derivatives. Accordingly, the attitude tracking curves based on rNSMFO, NSIMFO and NSMFO are shown in Figure 5, Figure 6 and Figure 7, respectively. It can be seen from the figures that the attitudes of multiple spacecraft can converge to a stable state. In addition, Table 22 shows that the tracking accuracy of the rIMOMFO-optimized controller can reach 10 6 , which is better than that for the other three algorithms. In addition, the adjustment time is longer. However, for the whole control process, the adjustment time of rIMOMFO is acceptable. Therefore, controllers optimized by rIMOMFO can obtain higher control accuracies with acceptable adjustment times.

6. Conclusions and Future Work

In this paper, the MOP of parameter optimization for a multiple spacecraft attitude controller is established, and an improved multi-objective moth–flame optimization algorithm based on r-dominant sorting (rIMOMFO) is proposed. In rIMOMFO, the Cauchy mutation is utilized to produce flames that lead moths to updated positions to improve the search performance of MFO. Secondly, a hybrid mutation operator is utilized to perturb the positions of the moths. Among them, the variation in the numbers of flames and moths can be dynamically adjusted by calculating the difference of fitness variance between the two generations of moths. In addition, an adaptive parameter is applied to the moth position update mechanism to further balance the exploitation and exploration. Finally, r-domination is used to guide the search toward the preference of the decision maker. Based on the WFG and CEC2009 test problems, rIMOMFO is compared with 12 other MOEAs, and the results show that the rIMOMFO algorithm is feasible and effective. Furthermore, rIMOMFO is used to optimize the parameters of an attitude cooperative controller for multiple spacecraft, and the results show that the optimized controller can improve the control accuracy and accelerate the convergence speed. Therefore, the rIMOMFO algorithm proposed in this paper can be utilized to solve MOPs in practical engineering.
In further research work, the improvement of MFO performance and the domination relationship with a stronger distribution can be studied. Simultaneously, the proposed rIMOMFO can be applied to practical devices, considering issues such as communication delay, actuator saturation and scalability for larger formations.

Author Contributions

Conceptualization, X.Z. and Y.F.; methodology, X.Z.; software, X.Z.; validation, X.Z., Y.F. and L.L.; formal analysis, L.L.; investigation, M.X.; resources, L.L.; data curation, L.L.; writing—original draft preparation, X.Z.; writing—review and editing, Y.F.; visualization, X.Z.; supervision, M.X.; project administration, Y.F.; funding acquisition, X.Z., Y.F., L.L. and M.X. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Foundation of China (Grant Number 62573375), Natural Science Foundation of Shandong Province (Grant Number ZR2023QF044), Natural Science Foundation of Hebei Province (Grant Number F2024203038), Science and Technology Research and Development Plan Project of Qinhuangdao City (Grant Number 202501A008). Science Research Project of Hebei Education Department (Grant Number QN2025401).

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

This work was conducted without any conflicts of interest.

References

  1. Wu, B.; Xu, C.; Zhang, Y. Decentralized adaptive control for attitude synchronization of multiple spacecraft via quantized information exchange. Acta Astronaut. 2020, 175, 57–65. [Google Scholar] [CrossRef]
  2. Ren, W.; Beard, R. A decentralized scheme for spacecraft formation flying via the virtual structure approach. In Proceedings of the 2003 American Control Conference, Denver, CO, USA, 4–6 June 2003; Volume 2, pp. 1746–1751. [Google Scholar]
  3. Du, H.; Li, S.; Qian, C. Finite-Time Attitude Tracking Control of Spacecraft With Application to Attitude Synchronization. IEEE Trans. Autom. Control 2011, 56, 2711–2717. [Google Scholar] [CrossRef]
  4. Liu, X.; Kumar, K.D. Network-Based Tracking Control of Spacecraft Formation Flying with Communication Delays. IEEE Trans. Aerosp. Electron. Syst. 2012, 48, 2302–2314. [Google Scholar] [CrossRef]
  5. Xu, M.; Fang, Y.; Li, J.; Zhao, X. Attitude coordinated output feedback control based on event triggering for multiple spacecrafts. Int. J. Control 2022, 95, 2438–2447. [Google Scholar] [CrossRef]
  6. Hu, Q.; Zhang, J.; Zhang, Y. Velocity-free attitude coordinated tracking control for spacecraft formation flying. ISA Trans. 2018, 73, 54–65. [Google Scholar] [CrossRef]
  7. Zou, A.M.; de Ruiter,, A.H.; Kumar, K.D. Distributed finite-time velocity-free attitude coordination control for spacecraft formations. Automatica 2016, 67, 46–53. [Google Scholar] [CrossRef]
  8. Hu, Q.; Shi, Y.; Wang, C. Event-Based Formation Coordinated Control for Multiple Spacecraft Under Communication Constraints. IEEE Trans. Syst. Man Cybern. Syst. 2021, 51, 3168–3179. [Google Scholar] [CrossRef]
  9. Tan, L.; Luo, J.; Liu, J.; Zhang, S. Cooperative autonomous obstacle avoidance control of spacecraft cluster via event-triggered reference governors. J. Frankl. Inst. 2024, 361, 107027. [Google Scholar] [CrossRef]
  10. Tang, Y.; Lai, B.; Li, X.; Zou, A.M.; Yang, X. Distributed finite-time attitude coordination control of spacecraft formations with multiple constraints. Int. J. Robust Nonlinear Control 2024, 34, 15. [Google Scholar] [CrossRef]
  11. Li, L.; Du, L.; Hu, Y. Cooperative control method of multiple spacecraft formation based on graphtheory. J. Comput. Methods Sci. Eng. 2024, 24, 1237–1251. [Google Scholar] [CrossRef]
  12. Yang, X.; Chen, Q.; Xie, S.; He, X. Adaptive Predefined-Time Event-Triggered Control for Attitude Consensus of Multiple Spacecraft With Time-Varying State Constraints. IEEE Trans. Syst. Man, Cybern. Syst. 2025, 55, 3756–3768. [Google Scholar] [CrossRef]
  13. Gong, Z.; Liu, C.; Teo, K.L.; Wu, Y. A gradient-based optimization algorithm to solve optimal control problems with conformable fractional-order derivatives. J. Comput. Appl. Math. 2025, 454, 116169. [Google Scholar] [CrossRef]
  14. Naifar, O. Tempered fractional gradient descent: Theory, algorithms, and robust learning applications. Neural Netw. 2026, 193, 108005. [Google Scholar]
  15. Chen, K.; Luo, W.; Zhou, Q.; Liu, Y.; Xu, P.; Shi, Y. A Novel Immune Algorithm for Multiparty Multiobjective Optimization. IEEE Trans. Emerg. Top. Comput. Intell. 2025, 9, 1238–1252. [Google Scholar] [CrossRef]
  16. Tanabe, R.; Ishibuchi, H. A Review of Evolutionary Multimodal Multiobjective Optimization. IEEE Trans. Evol. Comput. 2020, 24, 193–200. [Google Scholar] [CrossRef]
  17. Moldovan, D.; Slowik, A. Energy consumption prediction of appliances using machine learning and multi-objective binary grey wolf optimization for feature selection. Appl. Soft Comput. 2021, 111, 107745. [Google Scholar] [CrossRef]
  18. Deb, K.; Pratap, A.; Agarwal, S.; Meyarivan, T. A fast and elitist multiobjective genetic algorithm: NSGA-II. IEEE Trans. Evol. Comput. 2002, 6, 182–197. [Google Scholar] [CrossRef]
  19. Zhang, Q.; Li, H. MOEA/D: A Multiobjective Evolutionary Algorithm Based on Decomposition. IEEE Trans. Evol. Comput. 2007, 11, 712–731. [Google Scholar] [CrossRef]
  20. Kim, M.; Hiroyasu, T.; Miki, M.; Watanabe, S. SPEA2+: Improving the Performance of the Strength Pareto Evolutionary Algorithm 2. In Proceedings of the International Conference on Parallel Problem Solving from Nature, Birmingham, UK, 18–22 September 2004; pp. 742–751. [Google Scholar]
  21. Deb, K.; Jain, H. An Evolutionary Many-Objective Optimization Algorithm Using Reference-Point-Based Nondominated Sorting Approach, Part I: Solving Problems With Box Constraints. IEEE Trans. Evol. Comput. 2014, 18, 577–601. [Google Scholar] [CrossRef]
  22. Cui, Z.; Wu, Y.; Zhao, T.; Zhang, W.; Chen, J. A two-stage accelerated search strategy for large-scale multi-objective evolutionary algorithm. Inf. Sci. 2025, 686, 121347. [Google Scholar] [CrossRef]
  23. Qiao, K.; Yu, K.; Yue, C.; Qu, B.; Liu, M.; Liang, J. A Cooperative Multistep Mutation Strategy for Multiobjective Optimization Problems With Deceptive Constraints. IEEE Trans. Syst. Man, Cybern. Syst. 2024, 54, 6670–6682. [Google Scholar] [CrossRef]
  24. Ma, J.; Zhang, Y.; Wang, Y.; Gong, D.; Sun, X.; Zeng, B. A Multitask Multiobjective Operation Optimization Method for Coal Mine Integrated Energy System. IEEE Trans. Ind. Inform. 2024, 20, 11149–11160. [Google Scholar] [CrossRef]
  25. Chen, K.; Luo, W.; Lin, X.; Song, Z.; Chang, Y. Evolutionary Biparty Multiobjective UAV Path Planning: Problems and Empirical Comparisons. IEEE Trans. Emerg. Top. Comput. Intell. 2024, 8, 2433–2445. [Google Scholar] [CrossRef]
  26. Li, Z.; Luo, C.; Parr, G.; Min, G. Optimization for Control and Data Traffic in Aerial–Terrestrial AxV Networks: A Multiobjective Policy-Based Learning Approach. IEEE Internet Things J. 2025, 12, 33026–33040. [Google Scholar] [CrossRef]
  27. Pham, T.H. On efficient solutions of nonsmooth fractional multiobjective optimization problems with mixed constraints. Appl. Anal. 2025, 1–34. [Google Scholar] [CrossRef]
  28. Coello, C.; Pulido, G.; Lechuga, M. Handling multiple objectives with particle swarm optimization. IEEE Trans. Evol. Comput. 2004, 8, 256–279. [Google Scholar] [CrossRef]
  29. Konak, A.; Coit, D.W.; Smith, A.E. Multi-objective optimization using genetic algorithms: A tutorial. Reliab. Eng. Syst. Saf. 2006, 91, 992–1007. [Google Scholar] [CrossRef]
  30. Das, S.; Suganthan, P.N. Differential Evolution: A Survey of the State-of-the-Art. IEEE Trans. Evol. Comput. 2011, 15, 4–31. [Google Scholar] [CrossRef]
  31. Pierezan, J.; Dos Santos Coelho, L. Coyote Optimization Algorithm: A New Metaheuristic for Global Optimization Problems. In Proceedings of the 2018 IEEE Congress on Evolutionary Computation (CEC), Rio de Janeiro, Brazil, 8–13 July 2018; pp. 1–8. [Google Scholar]
  32. Yuan, X.; Zhu, Z.; Yang, Z.; Zhang, Y. MSCSO: A Modified Sand Cat Swarm Optimization for Global Optimization and Multilevel Thresholding Image Segmentation. Symmetry 2025, 17, 2012. [Google Scholar] [CrossRef]
  33. Yang, X.S. Firefly Algorithm. Eng. Optim. Introd. Metaheuristic Appl. 2010, 9, 221–230. [Google Scholar]
  34. Faramarzi, A.; Heidarinejad, M.; Stephens, B.; Mirjalili, S. Equilibrium optimizer: A novel optimization algorithm. Knowl.-Based Syst. 2020, 191, 105190. [Google Scholar] [CrossRef]
  35. Ginidi, A.R.; Elsayed, A.M.; Shaheen, A.M.; Elattar, E.E.; El-Sehiemy, R.A. A Novel Heap-Based Optimizer for Scheduling of Large-Scale Combined Heat and Power Economic Dispatch. IEEE Access 2021, 9, 83695–83708. [Google Scholar] [CrossRef]
  36. Gandomi, A.; Yang, X.; Alavi, A. Cuckoo search algorithm: A metaheuristic approach to solve structural optimization problems. Eng. Comput. 2013, 29, 17–35. [Google Scholar] [CrossRef]
  37. Mirjalili, S. SCA: A Sine Cosine Algorithm for solving optimization problems. Knowl.-Based Syst. 2016, 96, 120–133. [Google Scholar] [CrossRef]
  38. Sharma, V.; Pattnaik, S.S.; Garg, T. A review of bacterial foraging optimization and its applications. Procedia-Soc. Behav. Sci. 2012, 48, 1294–1303. [Google Scholar]
  39. Sulaiman, M.H.; Mustaffa, Z.; Saari, M.M.; Daniyal, H.; Daud, M.R.; Razali, S.; Mohamed, A.I. Barnacles Mating Optimizer: A Bio-Inspired Algorithm for Solving Optimization Problems. In Proceedings of the 2018 19th IEEE/ACIS International Conference on Software Engineering, Artificial Intelligence, Networking and Parallel/Distributed Computing (SNPD), Busan, Republic of Korea, 27–29 June 2018; pp. 265–270. [Google Scholar] [CrossRef]
  40. Mirjalili, S. Moth-flame optimization algorithm: A novel nature-inspired heuristic paradigm. Knowl.-Based Syst. 2015, 89, 228–249. [Google Scholar] [CrossRef]
  41. Liang, S.; Liu, Z.; You, D.; Pan, W.; Zhao, J.; Cao, Y. PSO-NRS: An online group feature selection algorithm based on PSO multi-objective optimization. Appl. Intell. 2023, 53, 15095–15111. [Google Scholar] [CrossRef]
  42. Ghasemi, M.; Bagherifard, K.; Parvin, H.; Nejatian, S.; Pho, K.H. Multi-objective whale optimization algorithm and multi-objective grey wolf optimizer for solving next release problem with developing fairness and uncertainty quality indicators. Appl. Intell. 2021, 51, 5358–5387. [Google Scholar] [CrossRef]
  43. Mirjalili, S.; Saremi, S.; Mirjalili, S.M.; dos S. Coelho, L. Multi-objective grey wolf optimizer: A novel algorithm for multi-criterion optimization. Expert Syst. Appl. 2016, 47, 106–119. [Google Scholar] [CrossRef]
  44. Balasubbareddy, M.; Sivanagaraju, S.; Suresh, C.V. Multi-objective optimization in the presence of practical constraints using non-dominated sorting hybrid cuckoo search algorithm. Eng. Sci. Technol. Int. J. 2015, 18, 603–615. [Google Scholar] [CrossRef]
  45. Zhai, Y.K.; Xu, Y.; Gan, J.Y. A novel artificial fish swarm algorithm based on multi-objective optimization. In Proceedings of the International Conference on Intelligent Computing, Huangshan, China, 25–29 July 2012; Volume 7390, pp. 67–73. [Google Scholar]
  46. Hassanien, A.E.; Gaber, T.; Mokhtar, U.; Hefny, H. An improved moth flame optimization algorithm based on rough sets for tomato diseases detection. Comput. Electron. Agric. 2017, 136, 86–96. [Google Scholar] [CrossRef]
  47. Shehab, M.; Abualigah, L.; Hamad, H.A.; Alabool, H.; Alshinwan, M.; Khasawneh, A.M. Moth-flame optimization algorithm: Variants and applications. Neural Comput. Appl. 2020, 32, 9859–9884. [Google Scholar] [CrossRef]
  48. Chung, S.J.; Ahsun, U.; Slotine, J. Application of Synchronization to Formation Flying Spacecraft: Lagrangian Approach. J. Guid. Control Dyn. 2008, 32, 512–526. [Google Scholar] [CrossRef]
  49. Xu, M.; Fang, Y.; Li, J.; Zhao, X. Finite time distributed coordinated control for attitude of multi-spacecraft based on super-twisting algorithm. Control Theory Appl. 2021, 38, 924–932. [Google Scholar]
  50. Zhao, X.; Fang, Y.; Ma, S.; Liu, Z. Multi-swarm improved moth-flame optimization algorithm with chaotic grouping and Gaussian mutation for solving engineering optimization problems. Expert Syst. Appl. 2022, 204, 117562. [Google Scholar] [CrossRef]
  51. Ben Said, L.; Bechikh, S.; Ghedira, K. The r-Dominance: A New Dominance Relation for Interactive Evolutionary Multicriteria Decision Making. IEEE Trans. Evol. Comput. 2010, 14, 801–818. [Google Scholar] [CrossRef]
  52. Yu, X.; Cai, M.; Cao, J. A novel mutation differential evolution for global optimization. J. Intell. Fuzzy Syst. 2015, 28, 1047–1060. [Google Scholar] [CrossRef]
  53. Wang, G.G.; Deb, S.; Gandomi, A.H.; Alavi, A.H. Opposition-based krill herd algorithm with Cauchy mutation and position clamping. Neurocomputing 2016, 177, 147–157. [Google Scholar] [CrossRef]
  54. Luo, J.; Chen, H.; Zhang, Q.; Xu, Y.; Huang, H.; Zhao, X. An improved grasshopper optimization algorithm with application to financial stress prediction. Appl. Math. Model. 2018, 64, 654–668. [Google Scholar] [CrossRef]
  55. Huband, S.; Hingston, P.; Barone, L.; While, L. A review of multiobjective test problems and a scalable test problem toolkit. IEEE Trans. Evol. Comput. 2006, 10, 477–506. [Google Scholar] [CrossRef]
  56. Zhang, Q.; Zhou, A.; Zhao, S.Z.; Suganthan, P.N.; Liu, W.; Tiwari, S. Multiobjective optimization Test Instances for the CEC 2009 Special Session and Competition. Mech. Eng. 2008, 264, 1–30. [Google Scholar]
  57. Hu, J.; Yu, G.; Zheng, J.; Zou, J. A preference-based multi-objective evolutionary algorithm using preference selection radius. Soft Comput. 2017, 21, 5025–5051. [Google Scholar] [CrossRef]
  58. Zitzler, E.; Thiele, L.; Laumanns, M.; Fonseca, C.; da Fonseca, V. Performance assessment of multiobjective optimizers: An analysis and review. IEEE Trans. Evol. Comput. 2003, 7, 117–132. [Google Scholar] [CrossRef]
  59. Zitzler, E.; Thiele, L. Multiobjective evolutionary algorithms: A comparative case study and the strength Pareto approach. IEEE Trans. Evol. Comput. 1999, 3, 257–271. [Google Scholar] [CrossRef]
  60. Zhang, Z.; Qin, H.; Yao, L.; Liu, Y.; Jiang, Z.; Feng, Z.; Ouyang, S. Improved Multi-objective Moth-flame Optimization Algorithm based on R-domination for cascade reservoirs operation. J. Hydrol. 2020, 581, 124431. [Google Scholar] [CrossRef]
  61. Schott, J. Fault Tolerant Design Using Single and Multicriteria Genetic Algorithm Optimization. Master’s Thesis, Massachusetts Institute of Technology, Cambridge, MA, USA, 1995. [Google Scholar]
  62. Xiong, Z.; Yang, J.; Zhao, Z.; Wang, Y.; Yang, Z. Maximum angle evolutionary selection for many-objective optimization algorithm with adaptive reference vector. J. Intell. Manuf. 2023, 34, 961–984. [Google Scholar] [CrossRef]
  63. Zhu, Q.; Zhang, Q.; Lin, Q. A Constrained Multiobjective Evolutionary Algorithm With Detect-and-Escape Strategy. IEEE Trans. Evol. Comput. 2020, 24, 938–947. [Google Scholar] [CrossRef]
  64. Zhang, X.; Zheng, X.; Cheng, R.; Qiu, J.; Jin, Y. A competitive mechanism based multi-objective particle swarm optimizer with fast convergence. Inf. Sci. 2018, 427, 63–76. [Google Scholar] [CrossRef]
  65. Zapotecas Martínez, S.; Coello Coello, C.A. A Multi-Objective Particle Swarm Optimizer Based on Decomposition. Assoc. Comput. Mach. GECCO ’11 2011, 8, 69–76. [Google Scholar]
  66. Coello Coello, C.; Lechuga, M. MOPSO: A Proposal for Multiple Objective Particle Swarm. In Proceedings of the 2002 Congress on Evolutionary Computation, CEC’02, Honolulu, HI, USA, 12–17 May 2002; Volume 2, pp. 1051–1056. [Google Scholar]
  67. Chen, B.; Zeng, W.; Lin, Y.; Zhang, D. A New Local Search-Based Multiobjective Optimization Algorithm. IEEE Trans. Evol. Comput. 2015, 19, 50–73. [Google Scholar] [CrossRef]
  68. Tian, Y.; Cheng, R.; Zhang, X.; Su, Y.; Jin, Y. A Strengthened Dominance Relation Considering Convergence and Diversity for Evolutionary Many-Objective Optimization. IEEE Trans. Evol. Comput. 2019, 23, 331–345. [Google Scholar] [CrossRef]
  69. Savsani, V.; Tawhid, M.A. Non-dominated sorting moth flame optimization (NS-MFO) for multi-objective problems. Eng. Appl. Artif. Intell. 2017, 63, 20–32. [Google Scholar] [CrossRef]
  70. Wang, R.; Purshouse, R.C.; Fleming, P.J. Preference-inspired co-evolutionary algorithms using weight vectors. Eur. J. Oper. Res. 2015, 243, 423–441. [Google Scholar] [CrossRef]
  71. Xu, M.; Fang, Y.; Li, J.; Zhao, X. Finite Time Fully Distributed Consensus Control for Multi-agent System With Input Saturation and Limited Communication Resources. Int. J. Control. Autom. Syst. 2023, 21, 3659–3672. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.