1. Introduction
The rapid advancement of modern society necessitates the continuous adoption of intelligent methods for optimization. Whether the objective is maximization or minimization, such optimization problems are essential to societal progress [
1]. Optimization algorithms—defined as stochastic or deterministic methods that optimize decision variables subject to objective functions and constraints [
2,
3,
4]—are employed to address problems involving multiple decision variables under complex nonlinear constraints [
5]. Classical optimization techniques include gradient descent [
6], Newton’s method [
7], the method of Lagrange multipliers [
8], and simulated annealing [
9]. However, real-world optimization tasks often exceed the capabilities of traditional solvers due to their large scale, high dimensionality, and inherently nonlinear nature. Such problems are frequently multimodal, causing search processes to easily become trapped in local optima and miss global solutions. Moreover, the proliferation of variables leads to an exponential expansion of the search space, resulting in the “curse of dimensionality.” The nonlinear and often non-differentiable behavior of objective functions further restricts the applicability of classical algorithms that rely on smoothness and convexity [
10,
11,
12]. Engineering problems, especially multi-criteria optimization, are core issues in engineering design and manufacturing. Most practical optimization problems inherently involve multiple conflicting objectives [
13]. Optimization algorithms can achieve material and manufacturing cost reductions while maintaining structural performance. Therefore, simple yet efficient metaheuristic algorithms are particularly suitable for solving difficult engineering optimization problems [
14].
In contrast to conventional optimization approaches, genetic algorithms introduce a novel perspective on optimization [
15]. As an evolution-inspired strategy and the first metaheuristic algorithm, they are capable of identifying high-quality solutions within computationally feasible time and resource constraints by evolving their own problem-solving ability, thereby enabling efficient handling of real-world optimization challenges [
16]. Dokeroglu et al. argue that a new generation of metaheuristics is increasingly becoming one of the most effective avenues for tackling complex optimization problems. These methods are celebrated for their potent global search capabilities, minimal assumptions about problem structure, and exceptional scalability. They are widely applied to high-dimensional, nonlinear, and multimodal optimization landscapes that defy conventional techniques, demonstrating remarkable efficiency and practicality in real-world applications [
17]. Those are also related to the challenge mentioned by Cebeci & Timur. In addition to calculating the mean ranking, the method to verify the successful development of metaheuristic algorithms also includes recording the median and the number of times the solution converges to the global optimum, since the solution is usually non-normally distributed. This successfully verifies the success of genetic algorithms in the origin of metaheuristic algorithms and the powerful engineering problem-solving capabilities they bring [
18]. However, the outcomes derived from different research questions often vary. When deciding whether to use the mean or the median for comparison, researchers need to conduct analyses based on the specific context of the problem. In general, each statistical method emphasizes different aspects, thereby reflecting the diversity in how results are presented across various issues.
The new generation of metaheuristic algorithms can be primarily categorized into four types. The first category comprises variants inspired by evolutionary behavior, which are developed based on biological evolution or genetic mechanisms. The Genetic Algorithm (GA) was the first to employ evolutionary principles, simulating chromosomal crossover and mutation [
19]. The Differential Evolution (DE) algorithm, proposed by Price et al., introduced vector differences between individuals in a population as a perturbation mechanism to guide the search direction, enhancing both solution accuracy and convergence efficiency in continuous optimization problems [
20]. The Evolutionary Mating Algorithm (EMA), developed by Sulaiman et al., made a significant contribution by incorporating the Hardy–Weinberg equilibrium principle from population genetics into evolutionary computation. This provided a quantifiable model of genetic drift for discrete optimization, establishing a search framework based on explicit genetic mechanisms [
21]. Meanwhile, Dong et al. proposed a Chaotic Enhanced Evolutionary Optimization (CEO) method, which integrates a memristive hyperchaotic system with differential evolution. This hybrid approach effectively balances global exploration and convergence speed, offering a novel dynamical optimization pathway for tackling NP-hard problems [
22].
The second category consists of algorithms inspired by physical and chemical phenomena, tracing back to the Simulated Annealing (SA) algorithm by Bertsimas and Tsitsiklis. SA, inspired by the physical process of cooling, has solved numerous optimization problems and provided foundational insights into physicochemical-inspired metaheuristics [
23]. Abdel-Basset et al. proposed the Kepler Optimization Algorithm (KOA), which is inspired by celestial mechanics. By simulating the orbital motion of planets under gravitational forces, KOA effectively guides the movement of candidate solutions, significantly improving global search capability and convergence efficiency [
24]. Su et al. developed the Rime Optimization Algorithm (RIME), emulating the formation mechanism of frost ice in nature. Through modeling the growth behavior of soft and hard rime under varying temperature conditions, the algorithm achieves an adaptive balance between exploration and exploitation during the search process [
25]. Additionally, Kashan proposed the Optics-Inspired Optimization (OIO) algorithm, which innovatively applies Fermat’s principle from optics to optimization. By mapping the path of light propagation in media to search strategies in the solution space, OIO demonstrates exceptional ability to avoid local optima in non-convex and complex function optimization [
26].
The third category, and currently the most popular, is swarm intelligence metaheuristics inspired by collective behavior. Notable recent representatives are as follows: The pioneers of metaheuristic algorithms—particle swarm optimization (PSO) based on the response of particle groups and ant colony optimization (ACO) based on the communication behavior of ants—have set a precedent for the development of group behavior in metaheuristic algorithms [
27,
28]. The Dhole optimization algorithm (DOA) by Mohammed et al. was inspired by the vocal communication and cooperative hunting strategies of dhole. By simulating sound-guided adaptive decision-making and dynamic clustering behavior, DOA effectively balances exploration and exploitation during the search process [
29]. Rüppell’s Fox Optimizer (RFO) by Braik et al., which models the diurnal and nocturnal foraging behavior of Rüppell’s fox, leverages acute sensory capabilities such as a 260° field of vision and 150° auditory perception. Its cognitive component dynamically coordinates the transition between global exploration and local exploitation [
30]. The Animated Oat Optimization algorithm (AOO) by Wang et al. was based on the propagation mechanisms of animated oat seeds—including dispersal via external forces, obstacle-avoiding propulsion, and hygroscopic twisting movement. This offers a novel nature-inspired search framework for engineering optimization, such as wireless sensor network layout [
31]. The Superb Fairy-Wren Optimization algorithm by Jia et al. draws inspiration from the development, brooding, and predator avoidance behaviors of superb fairy-wrens. By simulating these natural behaviors in distinct phases, the algorithm efficiently handles feature selection problems [
32].
The final category represents the most recent advancement: human-inspired metaheuristics based on individual and collective human behaviors. Ahmadi pioneered the Human Behavior-Based Optimization (HBO) model, abstracting search rules from multiple cognitive stages such as knowledge acquisition, consultation, and belief evolution. This establishes an optimization framework adaptable to complex constraints with cognitive interpretability [
33]. Bourechekara proposed the Most Valuable Player Algorithm (MVP), which draws on the dual mechanisms of individual excellence and collective collaboration in sports teams. It constructs a search strategy that simultaneously optimizes individual performance and group cooperation, effectively mitigating the inherent conflict between exploration and exploitation in traditional swarm intelligence algorithms [
34]. Trojovský designed the Preschool Education Optimization Algorithm (PEOA) based on the phased teaching mechanisms in early childhood cognitive development. By simulating the influence gradient of teachers and a three-stage cognitive evolution process, the algorithm significantly enhances the ability to escape local optima in complex multimodal environments [
35].
In summary, metaheuristic algorithms can be categorized as shown in
Figure 1. The pursuit of robust and efficient optimizers is fundamental in tackling the ever-growing complexity of real-world problems. Despite the proliferation of metaheuristic algorithms, significant challenges remain, particularly in achieving a delicate balance between exploratory and exploitative behaviors—a common limitation that often leads to suboptimal performance or premature convergence. While various natural and physical phenomena have inspired numerous algorithms, the sophisticated strategic interactions inherent in human economic behavior remain a relatively untapped source of inspiration. To bridge this gap, this paper proposes the Market Game Optimization Algorithm (MGOA). Its core novelty lies in a unique two-stage mechanism that simulates, with equal probability, the competitive strategies of merchants and their coordination with customers. This design fosters a dynamic and effective search process. The efficacy of MGOA is rigorously validated through comprehensive parameter analysis, behavioral and visual inspection, extensive experiments on the CEC 2017 and 2022 benchmark suites, and applications to 11 real-world engineering design problems.
The remainder of this paper is structured as follows:
Section 2 delineates the interactive behaviors between merchants and customers and elucidates the underlying mechanisms that inspire the algorithm.
Section 3 formalizes the merchant-customer game behavior model and introduces the proposed MGOA.
Section 4 presents the experimental setup, including parameter configuration, qualitative discussions, performance evaluations, and practical applications.
Section 5 concludes the study by summarizing the findings and discussing limitations and future research directions.
4. Experiment and Analysis
To demonstrate the advantages and characteristics of MGOA, this study employs behavioral and visual analysis and parameter configuration experiments to illustrate the algorithm’s optimization capability. Subsequently, a multi-dimensional performance comparison is conducted against ten well-established metaheuristic algorithms on the CEC2017 benchmark, highlighting the superior performance of MGOA. Furthermore, comparative experiments involving ten state-of-the-art enhanced algorithms on CEC2022 and 11 real-world engineering problems are carried out, providing dual validation of MGOA’s applicability and potential in both benchmark testing and practical optimization scenarios.
4.1. Computing Environment
To ensure the fairness and repeatability of the experiments, all experiments in this paper were conducted in a unified environment. The experimental configuration is detailed in
Table 2.
4.2. Behavioral and Visual Analysis
This section employs 8 classic benchmark functions to design four experiments for the behavioral and visual analysis of MGOA. First, the distribution of the optimal merchant strategies within the solution space of MGOA is analyzed to visualize the search strategy characteristics of merchants. The historical trajectories of marketing strategy searches are recorded by tracking the position of the optimal marketing strategy in each iteration. Additionally, an experiment on business strategy evolution is designed by documenting the best business strategy at every iteration. Then, to examine the overall iterative trend of the algorithm, the fitness value of the best solution after each iteration is recorded, and a convergence curve experiment is constructed. Finally, to analyze the variation trend of the fitness value over iterations, a fitness value change experiment is designed by logging the best fitness value after each update.
Figure 5 presents the results of MGOA on the four qualitative experiments described above. Specifically,
Figure 5a illustrates the 3D spatial distribution of all solutions for each benchmark function, while
Figure 5b depicts the two-dimensional distribution of the MGOA search history. It can be observed that a small number of historical best solutions are scattered across the solution space, while the majority cluster around the global optimum—particularly evident in functions F1, F3, F7, F11, and F13. This indicates that MGOA is capable of rapidly identifying near-optimal solutions within the solution space and transitioning earlier into the exploitation phase, thereby enhancing solution accuracy.
Figure 5c tracks the convergence behavior of MGOA throughout the iterative process. For most test functions, MGOA consistently improves solution quality as iterations proceed without becoming trapped in local optima. This holds true not only for unimodal functions such as F1 and F3 but also for complex multimodal and composite functions like F10 and F11. Through the synergistic effect of the attraction phase and the cooperation phase, MGOA conducts incremental search and refinement during iterations, effectively avoiding local optima.
As shown in
Figure 5d, due to the influence of the attraction phase in the early search stages, MGOA exhibits relatively large search step sizes in functions such as F1, F3, F10, F11, and F13. This characteristic facilitates escaping local optima and locating the global optimum. In later iterations, the cooperation phase plays a critical role by reducing step sizes, which contributes to refining the accuracy of the best solution.
Figure 5e records the best fitness value of MGOA after each iteration. It can be seen that the fitness values fluctuate to some extent across iterations, which stems from merchants’ active strategy adjustments during updates. This behavior confirms the effectiveness of the algorithm in dynamically exploring optimal solutions and aligns with its intended design.
In summary, MGOA not only demonstrates the ability to quickly converge to near-global optima and focus on intensive exploitation for improved precision but also actively alters search positions during updates. This enhances global exploration capability and the capacity to escape local optima, all while maintaining competitive convergence speed.
4.3. Strategy Coefficient and Dimension Selection Analysis
The CEC2017 benchmark set [
38] comprises a heterogeneous suite of test functions, including both unimodal and multimodal problems. These functions exhibit non-homogeneous properties—such as non-separability, robustness, and rotational invariance—enabling a comprehensive evaluation of an algorithm’s adaptability across diverse search environments. This benchmark has become a well-established standard for assessing the robustness and efficiency of optimization methods. In the performance evaluation experiments, the population size is set to 50, the maximum number of iterations is 5000, and 50 independent runs are conducted in parallel.
Table 3 presents all function suites within the CEC2017 test set.
In this section, we vary the value of parameter
in CEC 2017—introduced in
Section 3.2 as the decay coefficient for strategy adjustment, which significantly influences the overall precision and stability of the algorithm. Experiments are conducted with
set to 1, 3, 5, 7, and 10, respectively, and compared with the original configuration, where
, to illustrate the impact of parameter variation on the performance of MGOA.
As illustrated in
Figure 6, the value of α directly affects the behavior of the adjustment intensity
. A constant value of
indicates that the algorithm is conducting intensive exploitation around the current best solution, while a decreasing
suggests broader exploration. Therefore,
plays a critical role in directly governing the balance between exploration and exploitation, making appropriate selection essential. It can be observed that as the number of iterations increases,
gradually stabilizes when
or 10. For
or 5, although
exhibits mild fluctuations, the amplitude remains limited. Only when
does
demonstrate significant and persistent oscillations, indicating that a balance between exploration and exploitation has been achieved.
Table 4 presents the comparative results and rankings of different
parameter configurations on the CEC2017 benchmark suite under 100-dimensional settings. Among these, bold text indicates the optimal result. MGOA (with
) achieves the highest overall ranking, and its average rank is superior to those of algorithms with other parameter values. This indicates that MGOA (with
) does not exhibit inferior performance in optimization tasks.
Figure 7 similarly displays the ranking performance of various α values across different benchmark functions. Darker, nearly black shades indicate higher algorithm rankings under the corresponding parameter, whereas lighter, orange-like shades correspond to lower ranks. It can be observed that MGOA (with
) achieves the highest number of first-place rankings—prevailing in 16 out of 30 functions—significantly outperforming all other parameter configurations.
Therefore, based on comprehensive consideration of algorithmic accuracy and stability, the decay coefficient is set to 1 in this study.
After that, this study selects MGOA (with
) for a qualitative analysis of dimension selection, setting D to 1, 3, 5, 7, and 10. As can be seen in
Table 5, when the dimension selection ranges from 1 to 5, the algorithm performance improves, and when it ranges from 5 to 10, it starts to deteriorate, but it is still better than when the selection is small. In addition, MGOA (with
) has the lowest Friedman ranking, ranking first, indicating that dimension is not always better when it is larger. In addition, the methodology shows that it is related to the problem dimension. Therefore, the key to improving performance is to reasonably select a moderate number of algorithm dimensions. The number of dimensions extracted was 5 in this study.
4.4. Exploration and Development Evaluation
MGOA optimizes market strategies by balancing exploration and exploitation behaviors among merchant agents through the alternating execution of an attraction phase and a cooperation phase. To quantitatively assess the algorithm’s exploration–exploitation tendency, we track the dynamic variation in components across dimensions to quantify its search behavior. The exploration–exploitation balance is evaluated using the following formula:
Figure 8 shows that as iterations proceed, the exploration curve gradually declines, while the exploitation curve rises and eventually approaches 100%. This trend indicates that the algorithm initially emphasizes exploration and progressively shifts its focus toward exploitation, refining local solutions as the search continues. Through this dynamic balance, MGOA effectively integrates global search with local refinement, thereby converging toward an optimal market strategy.
4.5. Comparison of Mature Algorithms
Subsequently, this study compares MGOA against several established and recently proposed metaheuristic algorithms on the CEC2017 benchmark to validate its effectiveness. The compared algorithms include four highly cited classical methods: Whale Optimization Algorithm (WOA) [
39], Harris Hawks Optimization (HHO) [
40], Aquila Optimizer (AO) [
41], and Dung Beetle Optimizer (DBO) [
42]; four mature algorithms published in 2024: Black-Winged Kite Algorithm (BKA) [
43], Newton-Raphson-Based Optimizer (NRBO) [
44], Information Acquisition Optimizer (IAO) [
45], and Bermuda Triangle Optimizer (BTO) [
46]; and two latest high-performance algorithms published in 2025: Tornado Optimizer with Coriolis (TOC) [
47] and Chinese Pangolin Optimizer (CPO) [
48].
To ensure reproducibility,
Table 6 summarizes the initial parameter configurations for all algorithms involved (excluding adaptive parameters, only fixed parameters are used).
According to the results presented in
Table 7, the classical algorithms perform well on unimodal functions but exhibit significant performance divergence on complex multimodal and hybrid functions. Some algorithms, such as BTO, even suffer a drastic decline in solution quality by several orders of magnitude in certain cases. Within this context, MGOA demonstrates comprehensive and superior performance: it achieves the theoretical optimal values on functions F1, F2, F3, F4, and F9, indicating perfect convergence capability. Moreover, it consistently yields the lowest mean values or ranks among the top performers across the majority of the remaining functions, highlighting exceptional stability and robustness. These results strongly validate the powerful global exploration and local exploitation capabilities of MGOA.
According to
Table 8, where the test dimension of CEC2017 is increased to 30, MGOA continues to demonstrate comprehensive and superior performance. It not only exhibits outstanding global optimization capability, reliably locating the global optimal region, but also maintains high convergence speed and exceptional stability.
As shown in
Table 9, with the test dimension further raised to 50, MGOA consistently retains top-tier performance. Its powerful global exploration ability enables it to effectively escape local optima and approach the true global optimum, reflecting an excellent balance between broad search coverage and precise local exploitation.
Based on
Table 10, where the dimension is elevated to 100, MGOA achieves optimal or near-optimal results stably across the majority of test functions. Its remarkable convergence accuracy, robustness, and reliability fully validate the significant advantages in both global exploration and local exploitation capabilities.
Figure 9 displays selected convergence behaviors under the 100-dimensional CEC2017 benchmark. The dark blue curve represents MGOA, which demonstrates rapid convergence in the early stages on functions such as F11, F14, F18, F21, and F30, while significantly outperforming other comparative algorithms in later-phase exploitation, underscoring its competitiveness among state-of-the-art methods.
As observed in
Figure 10, MGOA converges to the optimal value in nearly all test functions except F30, where it exhibits relatively high fluctuation variance. Its consistently low mean values across the benchmark illustrate its powerful global optimization capability.
Figure 11 presents the average computation time of each algorithm. The most notable trend is that IAO requires significantly more time than the others. Algorithms such as CPO, NRBO, and MGOA show noticeably lower computational costs, with MGOA achieving the shortest average time, indicating relatively high efficiency.
Table 11 reports the outcomes of the Wilcoxon rank-sum test on the CEC2017 benchmark suite (10 dimensions), conducted at the 95% confidence level (
α = 0.05). In this table, the symbol “+” indicates that the reference algorithm MGOA performs significantly better than its counterpart, “−” denotes significantly worse performance, and “=” reflects no statistically significant difference. Inspection of the summary row reveals a striking pattern: against HHO, NRBO, and BTO, MGOA achieves superiority across all 30 test functions (30/0/0). When compared with WOA and TOC, only a single function in each case fails to reach statistical significance (29/1/0). Similarly, MGOA demonstrates a clear advantage over CPO and BKA (28/2/0 and 27/3/0, respectively). Even in comparison with AO, the results (26/4/0) remain dominated by positive outcomes. By contrast, modest instability emerges in the cases of DBO and IAO, where a handful of unfavorable instances are observed (23/6/1 and 5/24/1, respectively)—the only two columns in the table containing “−,” with IAO characterized primarily by ties. Coupled with the prevalence of exceedingly small
p-values across individual functions (e.g., 1.83 × 10
−4), these findings substantiate the conclusion that, on the 30 test functions of CEC2017, MGOA consistently exhibits statistically significant superiority over the majority of competing algorithms. The exceptions lie in its comparisons with IAO, where ties predominate, and with DBO, where occasional disadvantages arise. Results for other dimensional settings are provided in
Appendix B. Taken together, the scarcity of “−” outcomes and the predominance of “+” and “=” outcomes underscore that MGOA’s performance is, in most cases, not inferior to—and frequently superior to—that of the other algorithms.
Figure 12 displays the Friedman mean rankings of the algorithms across different dimensions (10, 30, 50, and 100), highlighting their relative performance under various settings. MGOA consistently maintains favorable rankings across all dimensions, demonstrating outstanding stability and efficiency. It achieves competitive rankings while retaining low computational time, particularly in the 100-dimensional case. In contrast, other algorithms such as IAO and BTO exhibit considerable volatility. These results distinguish MGOA as the most balanced and efficient algorithm overall, excelling in both time complexity and dimensional scalability.
4.6. Comparison of Advanced Improved Algorithms
Subsequently, this study evaluates MGOA against ten recently published improved algorithms on the CEC2022 (shown in
Table 12) benchmark to further verify its efficiency and advanced performance [
49]. The compared algorithms include: Multi-population mutative moth-flame optimization algorithm (MM_MFO) [
50], Dung Beetle Optimization Algorithm Based on Improved Multi-Strategy Fusion (MSFDBO) [
51], Enhanced gorilla troops optimizer (EGTO) [
52], Hybrid constriction coefficient-based PSO and gravitational search algorithm (CPSOGSA) [
53], Black-winged kite algorithm fused with osprey (OCBKA) [
54], Improved RIME (IRIME) [
55], Enhanced adaptive butterfly optimization algorithm (EABOA) [
56], Hybrid whale optimization algorithm (hGWOA) [
57], ESHO benchmarks for hyperthermia therapy optimization (ESHO) [
58], and Enhanced walrus optimization algorithm (EWO) [
59].
According to the results in
Table 13 and
Figure 13, MGOA demonstrates comprehensive and superior performance: it achieves better mean values than other state-of-the-art improved algorithms on functions F1, F2, F3, F6, F7, F9, and F11, while matching their convergence speed. These results reflect its well-balanced global exploration and local exploitation capabilities.
As illustrated in
Figure 14, MGOA attains the lowest mean values and standard deviations in almost every test function. This consistent advantage over advanced improved algorithms confirms MGOA’s strong capability in handling robustness and sensitivity challenges.
Table 14 reports the results of the Wilcoxon rank-sum test on the CEC2022 benchmark (20 dimensions) at the 95% significance level (α = 0.05). In the table, “+” indicates that the reference algorithm, MGOA, is significantly superior to the comparator; “−” indicates it is significantly inferior; and “=” denotes no statistically significant difference. The aggregate statistics in the final row reveal that MGOA attains a 12/0/0 record against ESHO, reflecting the most consistent and pronounced overall advantage. Against hGWOA, MM_MFO, MSFDBO, OCBKA, and EWO, the tallies are 10/2/0, 11/1/0, 10/2/0, 10/2/0, and 7/4/1, respectively—outcomes dominated by “+,” with relatively few ties. Results against CPSOGSA and IRIME are 9/3/0, with no losses; versus EGTO, 7/4/1, with only a single significant loss; and against EABOA, 5/5/2, which yields the greatest number of ties and two losses, marking it as the most challenging competitor. Examination of the per-function
p-values shows that most entries lie well below the 0.05 threshold (e.g., the many 1.83e−4 values in F3–F7 and F12), corroborating these significance claims; several functions (e.g., F1 and F8–F10) exhibit more “=” outcomes, indicating negligible differences. The few defeats arise primarily on F3 and F5 versus EABOA, on F11 versus EGTO, and on F10 versus EWO. Taken together, the scarcity of “−” outcomes and the preponderance of “+” and “=” indicate that, across the 12 CEC2022 test functions, MGOA is generally not inferior to—and often significantly better than—most state-of-the-art enhanced algorithms.
Figure 15 illustrates the computational time complexity of MGOA compared to state-of-the-art improved algorithms. It can be observed that MGOA achieves the lowest average computational time. Combined with its previously demonstrated superior performance, these results indicate that MGOA is not only conceptually concise but also a highly powerful metaheuristic algorithm.
4.7. Practical Engineering Design Issues
Engineering design optimization problems aim to achieve optimal system performance through mathematical modeling and algorithmic solutions while satisfying physical constraints and performance metrics. To validate the effectiveness of the proposed method, this subsection employs 11 engineering design problems, systematically recording statistical metrics—including best value, mean value, and standard deviation—for each algorithm.
4.7.1. Tension/Compression Spring Design Problem
The tension/compression spring design problem aims to minimize the weight of a spring under constraints of minimum deflection, surge frequency, and shear stress. As illustrated in
Figure 16, the problem involves three continuous decision variables: wire diameter (
d), mean coil diameter (
D), and number of active coils (
P). The mathematical formulation is provided in
Appendix A.1.
As shown in
Table 15, which presents the statistical results of the tension/compression spring design problem, MGOA achieves the minimum values in both the best and mean result categories, along with the smallest standard deviation. This indicates that MGOA can rapidly converge to the optimal solution for this problem, demonstrating superior performance compared to other algorithms and highlighting its effectiveness in addressing the tension/compression spring design challenge.
4.7.2. Pressure Vessel Design Problem
The objective of the pressure vessel design problem is to minimize the total cost, which includes material, forming, and welding costs. As shown in
Figure 17, the cylindrical pressure vessel is sealed at both ends with hemispherical heads. This optimization problem involves four decision variables: shell thickness (
Ts), head thickness (
Th), inner radius (
R), and the length of the cylindrical section (
L). The mathematical model is provided in
Appendix A.2.
As can be seen from
Table 16, MGOA achieves the lowest values in both the best and mean results, along with the smallest standard deviation, indicating the minimal optimized cost. These results demonstrate that MGOA exhibits higher stability in locating the optimal solution, which has positive implications for this practical application.
4.7.3. Three-Bar Truss Design Problem
The objective of the three-bar truss design problem is to minimize the volume of a truss structure by adjusting the cross-sectional areas (
x1 and
x2). The truss is subject to stress constraints (
σ) on each structural member, as illustrated in
Figure 18. This optimization problem involves one nonlinear fitness function, three nonlinear inequality constraints, and two continuous decision variables. The mathematical formulation is provided in
Appendix A.3.
As shown in
Table 17, MGOA again ranks first with particularly notable performance—its standard deviation of zero far surpasses that of other algorithms, maintaining a significant advantage. These results confirm MGOA’s exceptional stability and superior performance, consistently enabling precise convergence to the global optimum for this problem.
4.7.4. Welded Beam Design Problem
As depicted in
Figure 19, the welded beam structure consists of a beam and welded connections. The design is constrained by multiple factors, including shear stress, bending stress in the beam, buckling load on the bar, and end deflection of the beam. Solving this problem requires identifying an optimal combination of the following structural parameters: weld thickness (
h), clamped bar length (
L), bar height (
t), and bar thickness (
b). These variables are represented as a vector
x = [
x1,
x2,
x3,
x4], where
x1 to
x4 correspond to
h,
L,
t, and
b, respectively. The mathematical model is detailed in
Appendix A.4.
Table 18 demonstrates that MGOA achieves the best results in terms of both the best and mean values, while ranking third in standard deviation, behind EABOA and EWO. However, the computational time complexity of MGOA is considerably lower than that of EABOA and EWO, particularly significantly lower than EABOA. Moreover, the standard deviation of MGOA is only marginally larger than that of EABOA and EWO, indicating that MGOA is capable of rapidly converging to high-quality solutions within a short timeframe.
4.7.5. Speed Reducer Design Problem
The speed reducer design problem involves seven design variables: face width (
b), gear module (
m), number of gear teeth (
p), length of the first shaft between bearings (
l1), length of the second shaft between bearings (
l2), diameter of the first shaft (
d1), and diameter of the second shaft (
d2). The primary objective is to minimize the total weight of the speed reducer while satisfying constraints including bending stress of the gear teeth, surface contact stress, transverse deflection of the shafts, and stress in the shafts. The mathematical model is provided in
Appendix A.5.
As illustrated in
Table 19, the method achieves optimal results in both the best and mean value categories, while ranking closely behind EABOA and EWO in standard deviation. Notably, the disadvantage of MGOA in standard deviation is substantially outweighed by its significant advantage in computational time complexity compared to EABOA and EWO. These outcomes indicate that MGOA maintains its lightweight characteristic while simultaneously delivering competitive stability.
4.7.6. Gear Train Design Problem
The gear train design problem represents a practical optimization challenge in mechanical engineering with no additional constraints, making it relatively straightforward. The goal is to minimize the specific transmission cost of the gear train. The variables are the number of teeth of four gears:
Na (
x1),
Nb (
x2),
Nd (
x3), and
Nf (
x4). The mathematical formulation is described in
Appendix A.6.
As shown in
Table 20, MGOA ranks second—behind only ESHO—in terms of both the mean and the standard deviation. This is attributable to the stringent constraints and quasi-discrete nature of the gear-train design. MGOA’s exploration-oriented updates may overshoot the narrow feasible region near the optimum, occasionally yielding near-feasible solutions and thereby increasing the mean and variance. In contrast, ESHO features a stronger late-stage exploitation or repair mechanism that better preserves feasibility and continues to refine solutions once ratios close to the optimum are identified. Nevertheless, MGOA also achieves the best-known optimum, indicating strong optimization capability and confirming its effectiveness for the gear system design problem.
4.7.7. Cantilever Beam Design Problem
The cantilever beam consists of five hollow square sections, each defined by one structural parameter while maintaining a constant thickness, resulting in a total of five decision variables. The objective of the cantilever beam optimization is to minimize its total weight under the constraint of satisfying a specified vertical displacement limit. The mathematical formulation is provided in
Appendix A.7.
As shown in
Table 21, MGOA achieves a standard deviation of zero and the lowest optimized weight, reflecting ideal performance. These results demonstrate that MGOA can stably and efficiently address this engineering challenge, further confirming its excellence and precision in search capability.
4.7.8. Minimize I-Beam Vertical Deflection Problem
Minimizing the vertical deflection of an I-beam is a classical structural optimization problem. The main objective is to reduce the beam’s vertical deflection under load through optimal design of its cross-section and material properties. An I-beam, characterized by its I-shaped cross-section, typically consists of two flanges and a web. The beam is fixed at one end and subjected to concentrated or distributed loads at the other, leading to bending and vertical deflection. The goal is to design a beam that minimizes this deflection under applied loading. The mathematical model is described in
Appendix A.8.
Table 22 shows that MGOA achieved the lowest values in terms of the best, mean, and standard deviation; however, the same results were also attained by MSFDBO and EABOA. Nevertheless, MGOA exhibited the shortest average computation time, demonstrating its lightweight nature, lower computational time complexity, and overall superior performance.
4.7.9. Tubular Column Design Problem
Tubular columns are typically hollow cylindrical structures used to support other components or resist bending. Since they primarily bear axial forces, these columns are susceptible to buckling, especially under high axial compression. Thus, the main objective in tubular column design is to prevent buckling under external loads while optimizing geometric dimensions—such as outer diameter and wall thickness—so that the structure can withstand design loads with minimal material cost. The mathematical model is provided in
Appendix A.9.
As shown in
Table 23, while MGOA, EGTO, EABOA, and EWO all achieved the same minimum values in terms of the best, mean, and standard deviation, MGOA demonstrated a consistently shorter average computation time compared to the other three algorithms. This advantage underscores MGOA’s lightweight nature, enabling it to rapidly converge to optimal solutions and effectively address the tubular column design problem.
4.7.10. Piston Lever Design Problem
The piston lever design problem is a mechanical optimization task aimed at developing an efficient and stable lever system to ensure desired force transmission while minimizing system size or material cost. A piston generates linear reciprocating motion via fluid pressure, and a lever is used to modify the magnitude and direction of this linear force—commonly applied in mechanical transmission, force amplification, or torque conversion. Designing such a system often involves optimizing lever dimensions within reasonable spatial and weight constraints to achieve required performance metrics such as force transmission, effectiveness, and stability. The mathematical formulation is described in
Appendix A.10.
Table 24 indicates that MGOA secured the optimal best value, while it was only surpassed by MSFDBO and EABOA in the mean value, and by EABOA alone in standard deviation. Notably, the performance gaps in both the mean and standard deviation are minimal. Furthermore, MGOA exhibits slightly lower computational complexity than MSFDBO and significantly lower complexity compared to EABOA. Collectively, these findings affirm that MGOA—by virtue of its ability to identify optimal solutions with low time complexity—offers a meaningful contribution to addressing the piston lever design challenge.
4.7.11. Reinforced Concrete Beam Design Problem
This problem focuses on designing a reinforced concrete beam that meets strength, stiffness, and stability requirements while minimizing material consumption and cost, without compromising structural safety. A reinforced concrete beam combines two materials: concrete, which resists compressive stresses, and steel reinforcement, which bears tensile stresses. The design must ensure the beam can support design loads without failure or excessive deformation. Key variables include beam dimensions, amount and distribution of reinforcement, and concrete strength grade. The mathematical formulation is detailed in
Appendix A.11.
Table 25 demonstrates that MGOA delivers performance comparable to EGTO, EABOA, and EWO, with all four algorithms achieving identical optimal results in terms of best, mean, and standard deviation values. Moreover, MGOA required the shortest computation time among all algorithms evaluated, underscoring its inherently lightweight architecture and low computational complexity. These attributes position MGOA as a viable and efficient solution for the reinforced concrete beam design problem.
5. Conclusions
The proposed high-performance algorithm, named the Market Game Optimization Algorithm (MGOA), draws inspiration from the competitive dynamics between merchants and customers and aims to solve complex real-world optimization problems. MGOA computationally models how merchants adapt their strategies in response to market changes and engage customers through attraction and collaboration. The algorithm is structured around two complementary stages, each activated with a 50% probability. In the first stage, extensive search and strategy adjustments are performed, while the second stage focuses on refining strategies based on customer feedback and coordination. This design introduces innovative behavioral inspiration and delivers highly efficient optimization performance. Qualitative analysis shows that the algorithm consistently converges toward the global optimum. Parameter analysis further identifies the optimal adjustment intensity for reacting to market volatility. In subsequent performance evaluations, MGOA was compared against 10 widely recognized algorithms and 10 recently enhanced variants, demonstrating superior effectiveness on unimodal, multimodal, hybrid, and composite functions from CEC benchmark suites. Finally, applications to 11 real-world engineering design problems confirm that MGOA achieves the most effective results in the majority of cases among all compared algorithms.
Therefore, the contributions of this work are threefold: (1) Its design inspiration stems from the intelligent behaviors of advanced biological entities, opening a novel and forward-looking heuristic design pathway in the field of optimization algorithms. (2) By incorporating the attraction and synergy mechanisms between merchants and customers, the algorithm achieves extensive exploration in the early iterations, thereby accelerating convergence, while focusing on refined exploitation in later phases to effectively avoid local optima. (3) In practical applications, MGOA exhibits superior solving efficiency and stability, contributing to the effective resolution of numerous real-world optimization problems. In addition, compared to traditional gradient-based methods, which rely heavily on gradient information and are often limited to differentiable objective functions, MGOA, as a gradient-free heuristic approach, demonstrates significant advantages in handling non-differentiable, highly nonlinear, or noisy problems. However, it should be noted that gradient-based methods may still achieve faster convergence on smooth and convex problems where explicit gradient information is available.
In summary, MGOA demonstrates exceptional global optimization capabilities. Future research directions could encompass the introduction of chaotic mechanisms to enhance its global exploration performance, the improvement of mutation strategies to strengthen its ability to escape local optima, or its integration with other intelligent behavioral models or advanced algorithms. After that, we will prioritize rigorous, head-to-head comparisons with leading CEC algorithms—such as EA4Eig, NL-SHADE-LBC, S-LSHADE-DP, and LSHADE-cnEpSin—under standardized benchmark settings and statistical testing (subject to the availability of verified implementations) to more comprehensively assess enhanced MGOA’s competitiveness. Such developments would further advance the emergence of powerful hybrid algorithms capable of efficiently handling large-scale complex optimization problems.