Test and Validation of the Surrogate-Based, Multi-Objective GOMORS Algorithm against the NSGA-II Algorithm in Structural Shape Optimization the Algorithm

: Nowadays, product development times are constantly decreasing, while the requirements for the products themselves increased significantly in the last decade. Hence, manufacturers use Computer-Aided Design (CAD) and Finite-Element (FE) Methods to develop better products in shorter times. Shape optimization offers great potential to improve many high-fidelity, numerical problems such as the crash performance of cars. Still, the proper selection of optimization algorithms provides a great potential to increase the speed of the optimization time. This article reviews the optimization performance of two different algorithms and frameworks for the structural behavior of a b-pillar. A b-pillar is the structural component between a car ’ s front and rear door, loaded under static and crash requirements. Furthermore, the validation of the algorithm includes a feasibility constraint. Recently, an optimization routine was implemented and validated for a Non-dominated Sorting Genetic Algorithm (NSGA-II) implementation. Different multi-objective optimization algorithms are reviewed and methodically ranked in a comparative study by given criteria. In this case, the Gap Optimized Multi-Objective Optimization using Response Surfaces (GOMORS) framework is chosen and implemented into the existing Institut für Konstruktionstechnik Optimizes Shapes (IKOS) framework. Specifically, the article compares the NSGA-II and GOMORS directly for a linear, non-linear, and feasibility optimization scenario. The results show that the GOMORS outperforms the NSGA-II vastly regarding the number of function calls and Pareto-efficient results without the feasibility constraint. The problem is reformulated to an unconstrained, three-objective optimization problem to analyze the influence of the constraint. The constrained and unconstrained approaches show equal performance for the given scenarios. Accordingly, the authors provide a clear recommendation towards the surrogate-based GOMORS for costly and multi-objective evaluations. Furthermore, the algorithm can handle the feasibility constraint properly when formulated as an objective function and as a constraint. GOMORS Algorithm against


Introduction
A common way to save weight in structural design is to use optimization approaches in the product development process (PDP) [1,2], such as topology, shape, and size optimization. Certainly, engineers apply the mentioned optimization types in different stages of the PDP, for instance, the preliminary or conceptual design. Engineers mostly use shape optimization in all design-relevant stages of the PDP. Primarily, Computer-aided Design (CAD) and Non-Uniform Rational B-Splines (NURBS)-based approaches vary the geometry of parts before the geometry is meshed and calculated [3,4]. Especially in automotive of the NSGA-II to the implementation of the selected GOMORS framework in a comparative study [21]. The optimizations use the implemented b-pillar from [20] as a model. Furthermore, different load cases are analyzed for both algorithms: a torsional (linear), a side-impact (non-linear), and the side-impact load case under the geometrical constraint. Finally, the constraint is changed to an objective function to test different configurations of the GOMORS. The Discussion section analyzes and classifies the results of the study. Finally, the Conclusion and Outlook closes the article and gives an outlook on further possible areas of application of the method.

Materials and Methods
For structural optimization in the conceptual and preliminary design, mainly two different types of optimizations are used: topology and shape optimization. Only shape and size optimization are applied [3]. Accordingly, engineers often perform shape and size optimizations in preliminary and detailed design stages. Both approaches rely on optimizing possible design parameters of the structure, mostly with heuristics or metaheuristics. Shape variation uses two main techniques: Mesh-and CAD-based [3,6] processes. CAD-based processes can be realized with the combination of CAD programs for the geometry and pre-processors for the mesh-generation and solver-deck setup or with CAD programs with integrated pre-processor, such as SFE CONCEPT [4,22,23]. In particular, SFE CONCEPT offers the ability to generate structures quickly and pre-process Finite Element (FE) decks easily to follow Finite Element Analysis (FEA). Hence, the SFE CON-CEPT tool takes over the design and pre-processing to optimize structural components [18,24,25]. A vector of design variables (DVs) describes the geometry variably. The upper and lower design variable boundaries define the geometrical limits of optimized parts. The response of the calculated FE simulation leads to the response function, which can be subdivided into the objective and constraint function, such as mass, intrusion, or stiffness [15]. Ryberg and Domeij [9] formulate the general form of an optimization problem as: Determining the design variables that minimize the objective function is the target of the optimization problem. The vectors and ℎ represent the inequality and equality constraints. Each objective and constraint function value depends on the design variables . Feasible design points fulfill all constraints, while the design points which fail to comply with the constraints are infeasible. If the optimization problem has no constraints, it is called an unconstrained problem, while the contrary is called a constrained problem. If the problem has more than one objective function, it is called a multi-objective optimization. By the reformulation of the problem with objective functions, the following condition results:

Surrogate Models and Optimization Methods
From the viewpoint of the optimizer, the system of geometry change, pre-processing, calculation, and evaluation is a black box. Surrogate models can predict the dependency of the constraints and objective functions on the input variables. Furthermore, different coupling methods of surrogates and optimization algorithms exist, classified as decoupled, surrogate-driven, and surrogate-assisted methods. Decoupled methods are the most straightforward approach as combining a surrogate and an optimization algorithm requires less effort. The surrogate models approximate the target functions, and the optimization algorithm searches the optima and predicts a good solution. The framework updates the surrogate model accordingly until it depletes the budget for expensive evaluations. The surrogate model is intrinsically built into the optimization procedure in surrogate-driven methods such as Bayesian optimization techniques, which use probabilistic surrogate models with sequential updates. A rather unpopular approach is surrogate-assisted methods, where surrogate models support the algorithm in a specific domain, e.g., local search [11].
An extensive range of optimization algorithms is applicable for this kind of problem. As the behavior response functions are typically highly multimodal and not present in their mathematical closed form, one needs to use heuristic or metaheuristic approaches for the optimization [9,26]. Metaheuristics often adopt biological behavior for the optimization of the target function. Hence, a broad spectrum of different optimization approaches exists, which allow the optimization of black-box problems in varying disciplines. Paas and van Dijk [5] compare a Multi-Objective Evolutionary Algorithm Based on Decomposition (MOEA/D) to an NSGA-II and apply both algorithms to a BIW-optimization task. The NSGA-II converged faster to the Pareto-frontier and covered a more significant part of the solution space. The benchmark results of different automotive crash optimizations are summarized in [7]. The comparison of an Evolutionary Algorithm (EA), Genetic Algorithms (GAs), and Simulated Annealing (SA) for Noise, Vibration, and Harshness (NVH) in BIW optimization showed that for the GA and EA, the mutation and population size play an outstanding role, and so do the number of parents and children. Elitism improves multi-objective optimizations, while non-elitist approaches work well for single-objective optimizations. Furthermore, the EA outperformed the other algorithms under high noise and non-linearity. Due to the diversification and good exploration, EAs tend to be very successful, as they consider the whole solution space in early conceptual design. Indeed, one should consider the more exploitive algorithm should. All of these metaheuristics differ in their intensification and diversification [27]. The Bacterial Foraging Algorithm (BFO) tends to be less performant than the GA, Particle Swarm Optimization (PSO), and Ant Colony Optimization (ACO) in the scenario of an aeroelastic wing shape optimization. Therefore, modified Cooperative Bacterial Foraging Optimization (CFBO) tends to perform like the mentioned algorithm regarding the optimization performance [28]. Single algorithms have characteristics of either good exploitation (intensification) or exploration (diversification). Thus, new approaches arise that combine the benefits of intensification and diversification by combining different metaheuristics, e.g., presented in [29]. Due to its simplicity and effectiveness, recent research implemented the NSGA-II into the IKOS framework. Furthermore, it was validated using two different generic b-pillar structures under a geometrical constraint [19,30]. Due to the high computational demand of crash simulations, surrogate models should improve optimization speed [9,12].
The following section reviews frameworks and Python libraries, which use optimization algorithms combined with surrogate models. Wang and Chai combine the PSO with a BFO to form the PSO-BFO algorithm [31]. The algorithm uses a Radial Basis Function (RBF) surrogate model to optimize the behavior of a full-frontal impact and side impact of an SFE CONCEPT BIW model. While the PSO tends to have excellent explorative behavior, the BFO is better for exploitation. Hence, combining both algorithms with a surrogate improves the structural performance in both crash scenarios. The combined algorithm outperformed the PSO-GA and the simple PSO and BFO.
Besides the structural applications, different black-box optimization approaches exist. The following paragraph mentions notable exemplary implementations. Based on the MOCE+ algorithm, Haber and Beruvides propose the Simple Multi-Objective Cross-Entropy (SMOCE) algorithm [32]. The algorithm uses elitism clustering by histograms of the objective functions. One of the main benefits of this algorithm is its robustness and the low amount of algorithm parameters that makes the algorithm very consistent. The authors performed a comparative study for commonly used test problems with two objective functions. The study found that the SMOCE outperforms the MOCE+, NSGA-II, MOEA/D, MOPSO, SPEA-II, and PESA-II for hyper area exploration. The other algorithm may benefit from a smaller sampling size and generations (epochs); the SMOCE appears suitable for inexpensive function calls, e.g., analytical functions, as demonstrated in the article. Belakaria and Deshwal describe the Max-value Entropy Search for Multi-Objective Optimization (MESMO) with Bayesian Optimization for Network on Chip optimizations in [33]. The MESMO uses an output-space, entropy-based acquisition function to detect evaluation points, which leads to the efficient detection of the Pareto-frontier points. The authors compare MESMO to different Bayes approaches, which are Pareto-efficient global optimization (ParEGO), the Predictive entropy search for Multi-objective Bayesian Optimization (PESMO), S-Metric-Selection-based Efficient Global Optimization (SMSego) expected improvement in Pareto hypervolume (EHI), and the probability of improvement in stepwise uncertainty reduction (SUR). The paper compares all algorithms for the same two synthetic benchmark problems, which combine classical, single-objective benchmark problems. MESMO outperforms the other algorithms in terms of robustness and convergence.
Another approach is the GOMORS framework, a multi-objective EA that uses RBFsurrogate models [21]. The EA predicts new function points based on the surrogate model and generates possible candidates. A few points are selected based on rules, which balance exploration, exploitation, and frontier diversification. The selected candidates are then evaluated based on the costly function evaluation. The expensive black-box evaluations then update the RBF-surrogates, which serve as the foundation for further optimizations with the EA. Indeed, different embedded EAs have been evaluated, such as the NSGA-II, MOEA/D, and AMALGAM. The AMALGAM outperformed the other algorithms slightly. The GOMORS was then directly compared to the ParEGO and NSGA-II on a groundwater remediation problem and synthetic test functions. The GOMORS topped the ParEGO [34] and the NSGA-II regarding the effectivity. The authors state GO-MORS to find suitable solutions under a low amount of costly function evaluations.

IKOS: The Framework
As this article deals with implementing a new optimization algorithm into an existing optimization framework, the following paragraph briefly introduces the present implementation. The basic framework and approach are described in [18]. In addition, an extension of the method and more detailed optimizations are performed in [20]. The general framework is a Python 3-based implementation for the multi-objective structural shape and size optimization of linear and non-linear load cases for structural components. In particular, the framework currently uses SFE CONCEPT as a CAD-modeling tool and as a pre-processor. The solver OptiStruct calculates the linear load cases, and RADIOSS deals with the non-linear calculation of the structural components. Both load cases are evaluated individually or as multidisciplinary optimization. The framework uses the multi-objective NSGA-II as an optimization algorithm [19]. Again, one can formulate the optimization problem as a black-box function, wherein the calculation time for the non-linear evaluations is high and hence costly. The NSGA-II optimized the parts but showed slow convergence ratios and many function evaluations. Still, the NSGA-II is a robust algorithm for optimizing a b-pillar under the given geometrical constraint [20]. Furthermore, the framework allows for the parallelization of the calculations and server support. Hence, the user can surveil the optimization routine locally while the algorithm performs costly calculations on a server of choice, as shown in Figure 1.   The infeasible designs are penalized following the general approaches presented in the articles mentioned above, which the following rule can summarize: If the design intersects with the boundary surface, the intersecting surface area is calculated to the penalty factor. Accordingly, the infeasibility factor penalty varies between zero and one. A design is fully infeasible (penalty = 1) if the hull surface penetrates every element of the proposed design. The factor turns 0 if the check finds no penetrations. For every mth objective function, the objective function value is penalized.

Requirements towards Optimization Algorithms
The missing availability of criteria for choosing a proper optimization algorithm is one of the fundamental problems for engineers in optimization. Most publications use different test functions, and the authors validate the algorithms for computationally inexpensive test functions. Another problem is the number of validated objectives. Furthermore, the behavior of the black box (multimodality, non-linearity) and requirements from the framework itself (particular environment, e.g., Python 3) need to be considered. Papers and articles present different measures to choose a suitable optimization algorithm. Hence, this paragraph briefly reviews the requirements for the well-founded selection of an optimization algorithm. Chugh and Sindhya [35] surveyed a wide variety of algorithms; the authors found that for 45 published algorithms from 2008 to 2016, a maximum of three objective functions, mainly two, were evaluated. The amount of DVs commonly lies below 30. Mostly, no time for the training of the surrogate models is mentioned, which is founded on the argument that the expensive function evaluations are predominant. Still, training can be a significant factor for large data sets or quick assessment. In addition, the factors are mainly relative to the other algorithms analyzed in a specific paper: the following criteria can be formulated considering the efficiency of the algorithms: • The dimension of the objective and decision space; • Maximum number of expensive function evaluations; • Applicable problem characteristics (e.g., multimodality, divided Pareto-frontier).
Hence, the authors summarize that a good algorithm for optimizing black-box functions generates legitimate solutions with few evaluations even for unknown characteristics. The main questions towards the choice of algorithms are stated in [36], which are: • Which algorithm deals the best with the problem? • For which kind of problem can the algorithm be used?
Commonly, both questions can only be answered after testing the algorithm for the problem. The user's experience influences the algorithm's choice, as do the software, licensing, calculation (evaluation) cost, and the maximum time for the optimization. Because of their discontinuity, the numerical noise, and the multimodal character of crash simulations' local search strategies are not applicable for multi-objective optimization. Global search algorithms such as SAs, GAs, and EAs can usually be applied [7], while newer (adaptive) approaches focus on a balanced behavior of exploration and exploitation [12].
The following section reviews the choice, implementation, and validation of one of the mentioned algorithms into the IKOS framework.

Choice of the Optimization Algorithm
The mentioned algorithms are all potential candidates for the optimization problem. The linear load case does not play an important role, as calculation times are about 30 s. With 30 min for a simple and more than 12 h for a complex crash calculation, the algorithm and its efficiency are more important for the non-linear optimization time. From the survey, and according to the experience with the framework, the following requirements are suggested: • Multi-objective optimization algorithm; • Effective search strategy; • A limited number of costly function evaluations (e.g., use of surrogates); • Able to deal with the constraint function; • Parallelization possible.
As all these criteria are case-dependent and notably hard to evaluate for each optimization algorithm, a pre-selection with a weighted cost-benefit analysis is processed to choose an optimization algorithm.
Accordingly, the three main criteria for this specific optimization scenario are defined with the weights:
The evaluation and ranking of all the algorithms are still subjective, but it helps quantify and compare the benefits and drawbacks of all algorithms and frameworks. Appendix A shows the detailed cost-benefit analysis in tabular format. From this analysis, the GO-MORS framework is chosen and used for the following analysis due to its potentially high convergence ratio, the parallel data evaluation with up to four parallel calculations, and its global search strategy. Furthermore, GOMORS is already implemented in Python.
Accordingly, the following paragraphs briefly describe the NSGA-II and the GO-MORS to understand both methods better.

NSGA-II Algorithm
The Elitist Nondominated Sorting Genetic Algorithm (NSGA-II) is one of the most commonly applied GAs in engineering optimization. The main principle of the algorithm is based on fast, nondominated sorting, with diversity preservation, called crowding distance sorting. Deb described the NSGA-II initially in [18]. The following paragraph gives a brief overview of the articles' implementation.

•
Step 0-Define Algorithm inputs: Firstly, the user sets the maximum number of expensive function evaluations, the number of parents, and the mutation rate. • Step 1-Initial Evaluation Points Selection: The experimental design is processed using expensive simulations for a predefined sampling size. • Step 2-Iterative Improvement: To select the Pareto points of the sampling and iteratively optimize the designs, the algorithm repeats the steps of fast, nondominated sorting, diversity preservation, crossover, and mutation until it reaches the maximum number of costly evaluations. • Step 2.1-Fast, Nondominated Sorting: For each design, a domination count, which represents the number of designs dominating the current design, is set up. Furthermore, a set of designs, dominated by the current design, is assigned. By sorting the designs by their domination count, frontiers are built, for which the most dominating frontiers represent the individuals with the highest potential.

•
Step 2.2a-Density Estimation: The method calculates the crowding distance by the distance of the current solution to its two closest neighbors for each point of each frontier. It then sorts the designs by their crowding distance.

•
Step 2.2b-Crowded Comparison Operator: The method prioritizes designs if they exhibit a prior nondominated rank. Accordingly, if two designs are in the same frontier and have the same nondomination rank, it prefers the design in the less crowded area.

•
Step 2.3-Offspring Creation: The initial population based on the sampling is now sorted according to the mentioned strategies. Then, binary tournament selection, recombination, and mutation create the offspring population. Finally, elitism selects a certain percentage of parents.

•
Step 2.4-Calculate Responses: Expensive simulations return the demanded objective function values.
Finally, the procedure is repeated until the maximum number of expensive evaluations is reached.
The parameter values for the NSGA-II are set by experience. Hence, the number of parents for each generation is set to 20% of the sampling size. A total of 30% of the created offspring are mutated for more extensive exploration. Furthermore, the change of the selected value by mutation is 10%.

GOMORS Framework
Akhtar and Shoemaker [21] implemented the Gap-Optimized Multi-objective Optimization using Response Surfaces (GOMORS) framework. Indeed, GOMORS is a surrogate-assisted algorithm that uses the response surface method to estimate costly objective functions for the optimization with a GA. GOMORS updates the response surface model with the points evaluated in parallel. It identifies new, high-potential points in each iteration for the expensive data evaluation by selection rules based on the surrogate model. The following steps comprehensively describe the algorithm based on the original article.

•
Step 0-Define Algorithm inputs: Firstly, the maximum number of expensive function evaluations is selected. Furthermore, the gap parameter and the number of costly evaluations after each iteration are chosen. • Step 1-Initial Evaluation Points Selection: The experimental design selects the initial set of points and expensively evaluates the selected points via simulation. • Step 2-Iterative Improvement: The algorithm runs iteratively until the maximum number of expensive function evaluations (abort criteria) is reached. • Step 2.1-Fit/Update the Response Surface Models: The response surface models fit each objective for the expensively evaluated simulation points.

•
Step 2.2-Surrogate-Assisted Global Search: GOMORS uses an MOEA to minimize the objective functions represented by the response surface models, selecting a set of potential candidates.

•
Step 2.3a-Identify Least Crowded Solution: The least crowded design is selected by calculating the crowding distance of the expensively computed data.

•
Step 2.3b-Local Search: The framework uses the least crowded solution as input for an MOEA-based neighborhood search, based on the response surface models, called the "Gap Optimization Problem". It processes the search in the radius of the initially defined gap parameter around the least crowded solution.

•
Step 2.4-Select Points for Expensive Function Evaluation: A candidate selection is processed based on selection rules, the data sets from the surrogate-assisted global search, and the local search.

•
Step 2.5-Perform expensive function evaluations and update the non-dominated solution set: The method evaluates candidate points by costly simulations and updates the response surface model with the data of the candidate points.

•
Step 3-Return Best Approximated Front: After the final loop, when the maximum number of expensive function evaluations is reached, the best non-dominated frontier is returned.
For further details on GOMORS, the reader is referred to [21]. One problem with the current version of the GOMORS framework is that it is an implementation in Python 2.7, incompatible with Python 3.7. Hence, the GOMORS framework was ported to Python 3.7 to run in the existing IKOS environment. The article provides this modified GOMORScode via GitHub [37]. Mainly, it consists of the following changes compared to the implementation from [21]: • It uses a symmetric Latin hypercube sampling instead of a Latin hypercube sampling; • The MOEA of choice is the epsilon-NSGA-II following Kollat and Reed; • Instead of the hypervolume improvement, the epsilon-progress-based selection is used.

Results
The optimization setup was already established and surveyed in recent papers. The b-pillar has two main load cases: Nastran calculates the linear load case and RADIOSS a non-linear load case. Accordingly, for the linear and the non-linear load case, the b-pillar offers multiple objective functions for evaluation. The potential objective functions are the mass ( ), the static deflection (torsion) ( ), the intrusion in the lateral direction ( ), and the absorbed energy during the impact ( ). The optimization scenario is always multi-objective and can be a multidisciplinary optimization scenario in the case of the deflection and crash case. Figure 3 shows the two load cases. The stiffener varies by implicit parametrization between the upper and the lower shell, as shown in Figure 4. Eight sections control the shape and create many possible designs. By setting the DVs implicitly, no infeasible designs are created. Still, the effort it takes to develop such an implicit design is very high.

Linear Load Case Comparison without Infeasibility
The first scenario optimizes the b-pillar performance towards the static torsional load case and the mass. The challenge is the minimization of both objective functions. Hence, the unconstrained MOO optimization problem is formulated to: In Figure 5, the optimization results are displayed, which compares the two objective NSGA-II optimization runs with the two objective GOMORS optimization runs for the linear displacement load case. The z-displacement and the mass are normalized to give the viewer a better understanding of the improvement on the scale by the following formula for feature scaling: where ′ is the normalized data for the ith data point for the objective function value . All values are relative to the initial design . The weight change is relative to the initial part for each design, so it excludes the rest of the structure. In this case, the mass of each design only relates to the initial stiffener design in the first scenario. One can see that the GOMORS outperforms the NSGA-II. The GOMORS converges faster and reaches a Pareto-frontier earlier. Still, due to the conflicting objective functions, the algorithms strive into the minimum mass, and the structural stiffness decreases. Still, the GOMORS finds designs that reduce structural weight by 13% while maintaining equal stiffness. The shape optimization of the stiffener reduces the mass on the component level by approximately 7%, while the design achieves a minor displacement of 1%. Further structures lie in the lower left quadrant.  Figure 3 also shows the second scenario, the intrusion of the impactor in the b-pillar. In this scenario, the algorithms minimize the intrusion and the mass. Again, the min-max feature scaling method is used for the value scaling in Figure 6. Again, the GOMORS converges significantly faster than the NSGA-II. The NSGA-II slightly improves the design compared to the sampling points, as multiple iterations lie in the sampling area. The GOMORS, on the other hand, has a more explorative behavior. The designs are determined to converge towards the Pareto-frontier. Nearly no points are in the sampling area. The GOMORS can decrease the weight by 12.5% while maintaining equal intrusion. The NSGA-II can only reduce the weight by 8% under similar intrusion. Significant decreases in the intrusion up to 12% are possible, reducing weight by 5%.

Non-Linear Load Case Comparison without Infeasibility
The preceding paragraphs demonstrate the performance of the GOMORS in comparison to the NSGA-II for linear and non-linear calculations. Hence, further analysis will only assess linear calculations due to the high computational resources needed for nonlinear calculations.

Optimization Results for the Linear Load Case with Infeasibility
The third analysis compares the two algorithms for the same b-pillar with two more design variables and the infeasibility constraint. Figure 7 shows three different configurations of the b-pillar in its thinnest, the initial, and the thickest configuration. The used design and package-constraint surfaces are the same as in [19]. The stiffener can still vary between the upper and the lower shell. Furthermore, the design space, representing the external geometric constraint from package and design, is highlighted in red.  Another problem is the ratio of infeasible designs by the intersecting hull surface, which reduces the number of calculated designs [18,20]. Accordingly, the sampling size of actual calculated samples shrinks by a factor of 0.2 for the given sample. The sampling size is initially set to 1000 members to compensate for that effect. Hence, approximately 200 members are calculated during the optimization; still, the number of design variables is increased by two, which increases the size of the solution space. The number of iterative samples is set to 200. Finally, Figure 8. shows the resulting point diagram with the normalized mass and z-displacement.  The distribution of the samples differs due to the different model setups. Conversely to the previous optimization runs, the optimization algorithms perform very differently. The NSGA-II and the GOMORS act pretty similar for the shape optimization task. Both form an interval of potential candidates but do not converge into a region of ideal design. For equal mass, both algorithms find solutions, which decrease the displacement by 20%. For a similar displacement, the max decrease is about 1.5%. Notably, the NSGA-II shows a slightly better performance than the GOMORS, which the small sampling size could cause. Both algorithms mostly create feasible designs considering the infeasibility constraint. The GOMORS creates three infeasible members, and the NSGA-II creates four infeasible members. Still, both results seem to be insufficient in the improvement of performance. Hence, a different approach is needed, discussed in the following section.

Implementation for Surrogate-Based Optimization Schemes
The preceding Sections 3.1 and 3.2 demonstrate that the surrogate-based GOMORS framework mostly has a superior performance to the NSGA-II in the case of classical shape optimization. However, penalty factors may cause surrogate models to map the behavior of the actual physical model poorly. This behavior was already predicted in [20] and is now validated in Section 3.3 of this article. Hence, the following section tests a second approach to validate the feasibility of the penalization scheme for surrogate-based processes, which are represented in this case by GOMORS.
As the unsteady behavior of the penalization scheme may cause an inferior quality of the surrogate models [20], the penalty is therefore set as the third objective function. The approach benefits from the lower amount of needed mesh evaluations, as it does not manipulate the infeasible designs. Still, the number of required calculations increased. The objective function value of the infeasibility factor ( ) is zero if no infeasibility is detected, and ( ) is larger than zero if a surface penatration is detected. As GOMORS performed well on the unconstrained and provided worse results in the constrained approach, the approach is only validated for GOMORS under consideration of the following three objective functions: mass ( ), stiffness ( ) and feasibility ( ): With respect to the preceding optimizations, the sampling size is set to 200, and the number of iterations is set to 100. Figure 9 shows the Pareto-frontier, which is built during the optimization. The normalization is relative concerning the initial design, and it considers the outer sheet metal, the inner sheet metal, and the stiffener. Hence, the percentage relations are the same as in Figure 9. With the presented optimization setup, only a feasible design outperforms the initial setup in terms of Pareto-optimality. Still, the improvement is minimal, which means no increase in terms of mass and ~10% in displacement. Conversely, there are potential lighter candidates, but they exhibit a more significant static deformation. This lack of potential candidates may result from the larger solution space (10 DV) and the strong influence of the outer and inner shell movement on the weight and resulting displacement due to the feature importance.
The sampling size is increased to 1000 samples to improve the surrogate model quality, while 200 iterations allow a more detailed optimization run. Furthermore, the number of design variables is increased to 24. The 16 DVs for the outer and inner shells are activated to ensure that the influence of the DVs is evenly weighted and that more design freedom leads to candidates of higher potential. Furthermore, the feasibility approach with penalty constraint and the feasibility approach with three objective functions are directly compared for the chosen sampling and the NSGA-II and the GOMORS iterations, as shown in Table 1.  GOMORS  1000  1000  2  24  NSGA-II  1000  1000  2  24  GOMORS  1000  200  3  24 According to Table 1, the three scenarios are optimized and visualized in the following paragraphs. Again, the b-pillar is analyzed, but the number of DVs is increased to 24. The optimization problem can be formulated as: For a better understanding of the outer shell movement, four possible feasible candidates are shown in Figure 10. The stiffener can move in between the upper and lower shell, as presented in Figure 4.    The figure shows that the mass and displacement can be decreased significantly, while mostly feasible designs are created. The decrease in mass goes up to 8% for an equal static deformation, while other candidates exhibit a reduced mass by 5.5% and nearly 10% lower deformation. The sampling contains almost no feasible designs, but the algorithm produces mostly feasible designs during the optimization. Figure 12 shows the convergence of the feasibility for the three other optimization runs to demonstrate the effectiveness of the different approaches towards the feasibility. Therefore, it offers the infeasible designs per 10 samples for the sampling and optimization iterations. The sampling size, as described in Table 1, is 1000.  Figure 12 shows that infeasible designs are very high in the sampling. After the sampling, the number of infeasible designs decreases to zero in the case of the constrained two objective approach. Conversely, the three-objective approach still creates some infeasible designs. To visualize the convergence of the algorithms, Figure 13 shows how the NSGA-II and the GOMORS perform for the given optimization with 24 DVs for the two-objective constraint approach and the three-objective approach. In the two-objective approach, the infeasible designs are penalized and not part of the frontier, while in the three-objective approach, the final designs are all created feasible. For each algorithm, the frontier only represents feasible designs. Figure 13. The Pareto-frontier for the GOMORS with two and three objective functions and the NSGA-II with two objective functions. The performance is relative to the initial design. One can see that the GOMORS outperforms the NSGA-II. Both GOMORS frontiers nearly converge equally, while the NSGA-II frontier does not converge this far. For an equal displacement of 100%, the GOMORS with three objective functions reduces the mass to 92.5%, while the GOMORS with two objective functions reduces the mass to 93.5%. The NSGA-II, on the other hand, only achieves 95% of the initial mass.

Discussion
The presented analysis dealt with implementing the GOMORS optimization algorithm into an existing framework and its potential to optimize the structural performance of BIW parts under a geometrical constraint. The results of the structural optimization for a linear and a non-linear crash-load case are promising, as the GOMORS outperforms the NSGA-II in terms of convergence ratio. Figure 4; Figure 5 show that GOMORS reduces the time for calculation and the number of expensive computations. The second part of the paper dealt with whether surrogate-based algorithms can handle the feasibility constraint. This question was successfully answered with Figure 7. The surrogate represents the model behavior properly so that the optimization converges quickly if the sampling size is of sufficient size. In [20], the authors tested Random Forrest Regression as a surrogate model and found that the penalized infeasible sampling's mean squared error was very high. Concerning the optimization of the linear load case with infeasibility constraint, this hypothesis can be partly refuted, as the GOMORS converges much better than the NSGA-II, which could be reasoned by a higher performance of the RBF surrogate model.
The final test of the GOMORS framework for three-objective optimization analyzes the performance of the constraining approach by direct penalization of infeasible designs. It considers the displacement, mass, and feasibility. Firstly, Figure 11 shows the functionality of the approach as the final frontier only exhibits feasible results, which also show better structural performance than the initial design. Figure 13 compares the performance of the two-objective with the three-objective approach. While comparing the two-objective to the three-objective GOMORS approach, the three-objective approach performs slightly better in this specific scenario. Furthermore, GOMORS outperforms the NSGA-II for the 24 DV example. The results demonstrate the performance of the GOMORS framework in comparison to the NSGA-II. Still, one drawback of the GOMORS implementation is the number of parallel evaluations, which is limited to four, and for the current implementation, to one. Long calculation runs may benefit a substantial parallelization; a wider parallelization will make the GOMORS even more performant. It could help to reduce the time of optimization even further. Even if the convergence ratio of the NSGA-II is lower than the one of GOMORS, the nearly unlimited scalable parallelization makes it attractive for cheaper optimization tasks, such as linear calculations. Still, GOMORS may be of benefit here, as areas that tend to be not explored by the NSGA-II are explored.

Conclusions and Outlook
The multi-objective optimization of structural BIW parts, especially with costly crash calculation, is challenging for standard metaheuristic algorithms, such as the NSGA-II. Especially when the design is constrained, as in the demonstrated geometric constraint, the NSGA-II tends to converge slowly and may use many costly evaluations. Implementing the GOMORS optimization framework into the IKOS framework showed that surrogate-based frameworks can outperform typical metaheuristic approaches on a large scale, even concerning constrained objective functions and linear and non-linear calculations. Here, the method of infeasibility penalization was validated for the GOMORS algorithm and can now be effectively used for the presented scenarios. Furthermore, the article demonstrated the importance of choosing influential DVs as a critical factor for the success of optimizations. The study confirmed that more DVs enable larger improvements in the initial design proposal, e.g., 7.5% less weight while maintaining equal structural stiffness without any need for different tooling or manufacturing technologies. Further research will focus on analyzing the optimization parameter of the GOMORS framework, for example, the influence of the gap and additional internal parameters, which have been defaulted in this analysis. Furthermore, the optimization of more complex non-linear calculations will be of significant interest, mainly non-linear crash calculations of longitudinal structures. Problems such as bifurcation and high uncertainties occur during the non-linear calculation for these calculations. This behavior may be a problem for the radial basis function, failing to handle the significant uncertainties. Here, different surrogate approaches may be of major interest. In particular, the question should be answered if the surrogate-assisted process has major benefits compared to classic metaheuristics such as the NSGA-II [7,19].
At the same time, further applications of the feasibility factor are of significant interest. The approach could be tested in all domains, where shape optimization is used under geometric constraints, such as structural applications, e.g., automotive chassis components [38] or aeroelastic analysis for planes [39]. Primary applications tend to be aerodynamics in planes [40] or preliminary race car design, where large geometrical designs are explored under strictly regulative design spaces [41]. Funding: We acknowledge support from the Open Access Publication Funds of Technische Universität Braunschweig.

Data Availability Statement:
The implemented GOMORS for Python 3.7 can be downloaded via the following link on GitHub [37].