You are currently viewing a new version of our website. To view the old version click .
Algorithms
  • Article
  • Open Access

28 November 2025

A Multi-Objective Evolutionary Computation Approach for Improving Neural Network-Based Surrogate Models in Structural Engineering

,
and
Institute of Intelligent Systems and Numerical Applications in Engineering (SIANI), University of Las Palmas de Gran Canaria (ULPGC), 35017 Las Palmas de Gran Canaria, Spain
*
Author to whom correspondence should be addressed.
Algorithms2025, 18(12), 754;https://doi.org/10.3390/a18120754 
(registering DOI)
This article belongs to the Special Issue Multi-Objective and Multi-Level Optimization: Algorithms and Applications (2nd Edition)

Abstract

Surrogate models are widely used in science and engineering to approximate other methods that are usually computationally expensive. Here, artificial neural networks (ANNs) are employed as surrogate regression models to approximate the finite element method in the problem of structural analysis of steel frames. The focus is on a multi-objective neural architecture search (NAS) that minimizes the training time and maximizes the surrogate accuracy. To this end, several configurations of the non-dominated sorting genetic algorithm (NSGA-II) are tested versus random search. The robustness of the methodology is demonstrated by the statistical significance of the hypervolume indicator. Non-dominated solutions (consisting of the set of best designs in terms of accuracy for each training time or in terms of training time for each accuracy) reveal the importance of multi-objective hyperparameter tuning in the performance of ANNs as regression surrogates. Non-evident optimal values were attained for the number of hidden layers, the number of nodes per layer, the batch size, and alpha parameter of the Leaky ReLU transfer function. These results are useful for comparing with state-of-the-art ANN regression surrogates recently attained in the recent structural engineering literature. This approach facilitates the selection of models that achieve the optimal balance between training speed and predictive accuracy, according to the specific requirements of the application.

1. Introduction

Solving real-world problems through computational engineering often requires a high computational time, even with increasingly efficient computers. Among the methods to reduce computational time and/or the number of evaluations of the objective function for improving optimal design in engineering using metaheuristics/evolutionary algorithms, we can fundamentally list the use of parallel evolutionary algorithms, the use of surrogate models or metamodels, and the use of game theory-based evolutionary algorithms []. This manuscript focuses on the second of these methods: efficient surrogate modeling is handled, particularly in the context of structural engineering and regression surrogate modeling, through artificial neural networks.
The application of neural networks in the calculation and optimization of skeletal structures is a field that was started some thirty years ago, as can be seen in references [,,]. However, the recent surge in deep learning [] and surrogate-based evolutionary algorithms [] has recently fostered the research for enhancing single and multi-objective optimization [] in real-world applications through surrogate-based evolutionary algorithms in many engineering fields, such as biomedical engineering [], reliability engineering [], and others []. In the field of structural engineering, ref. [] proposes a single neural network architecture based on iterative tuning for the construction of a surrogate model. However, it does not provide details on the number of trials performed. In this case, the model is applied to the problem of reinforced concrete rigid node bar structures and had a six-layer hidden architecture with 128 neurons in the first two layers and 64 neurons in the last four layers. In ref. [,], surrogate models based on neural networks are proposed for an offshore wind turbine metal support bar structure. In the first case [], 28 neural network architectures are tested, with possible combinations of one to four hidden layers and a uniform number of neurons per hidden layer of 10, 25, 50, 75, 100, 125 and 150. In the second case [], only seven neural network architectures are tested consisting of 2, 3, 4, 5, 6, 7, 8 hidden layers and 100, 150, 200, 250, 300, 350, 400 neurons per layer, respectively. Whether detailing and proposing a single architecture or proposing even a limited set of architectures for the neural network-based surrogate model, there is scope to improve the performance of the surrogate if an optimization procedure is handled to improve the architecture and/or hyperparameters of the neural network.
The enhancement of the neural network architecture and parameters has been handled through the procedure known as Neural Architecture Search (NAS) [,]. The most commonly used algorithms for NAS are reinforcement learning, Bayesian optimization, and evolutionary algorithms/metaheuristics, with the latter being employed successfully in multi-objective optimization approaches. In single-objective NAS, where the minimization of error (or alternatively the maximization of accuracy) is typically sought, the complexity of the neural network (or the computational training time required) is not a factor that is taken into consideration. Conversely, within the paradigm of multi-objective optimization, a secondary objective function is incorporated, namely training time, as elucidated in this study. This enables the identification of a set of non-dominated solutions, signifying the attainment of the ANN architecture that corresponds to the optimal accuracy for a specific training time. Alternatively, this approach facilitates the identification of the ANN architecture that attains the lowest training time for a given level of accuracy. The extant literature on the use of neural networks as surrogate models/metamodels in structural engineering has recently been augmented with the incorporation of four recent reviews, which are indicative of the growing interest in this field of research [,,,]. It is evident from the research undertaken to the best of the authors’ knowledge that this work is pioneering in its proposal, implementation, and successful demonstration of a multi-objective NAS in the field of structural engineering. A comprehensive survey of the literature reveals no previous suggestion or successful application of a multi-objective NAS in this area. The conflicting objectives are commonly related to the accuracy of the surrogate model accuracy and the efficiency of the hardware [,], although the complexity of the neural network is also considered (the number of parameters is frequently taken into account). In this manuscript, the conflicting objectives are related to improving first, surrogate model accuracy and second, surrogate model calculation speed.
This work focuses on enhancing efficient neural network-based surrogate models. A multi-objective approach using evolutionary algorithms is proposed for the improvement of surrogate models in bar element structures. This approach aims to produce models that are as accurate as possible (being able to faithfully reproduce the associated numerical method) and simultaneously as fast as possible (training/calculation time of the surrogate model as short as possible). To this end, this work employs a state-of-the-art multi-objective evolutionary algorithm based on the Pareto dominance criterion (the non-dominated sorting genetic algorithm NSGA-II []) and a surrogate deep learning model using artificial neural networks. The field of multi-objective optimization is still an active research field (e.g., some recent multi-objective algorithms are proposed in [,]). As example, the PlatEMO framework [] available in (https://github.com/BIMK/PlatEMO, accessed on 31 October 2025), currently has (release 4.13) more than 300 multi-objective evolutionary algorithms. For this research, the state-of-the art multi-objective evolutionary algorithm NSGA-II was chosen because it is the most used and cited algorithm and still a state-of-the-art approach for solving optimization problems with two objectives [], as it is the handled problem.
From a database of thousands of structures, the results successfully demonstrate the application of the methodology in a reference test case from the scientific literature of frame steel structures []. They provide a set of non-dominated solutions with optimal accuracy and computational time, as well as useful insights about the hyperparameters of the neural network-based surrogate model.
The structure of the manuscript continues as follows: Section 2 presents the structural problem and test case. Section 3 then introduces the multi-objective optimization NAS approach, as well as the structure and configuration of the neural networks used. Section 4 presents the experimental results and discussion. Finally, Section 5 presents the conclusions.

2. Structural Engineering Problem and Test Case

The goal of the structural problem addressed is an optimal design problem where we want to obtain the lowest weight (equivalently mass) structure, which in turn can fulfill the mission for which it was designed. This mission consists of resisting the loads to which it is subjected without reaching the ultimate and serviceability limit states; that is, without exceeding the maximum constraints related to the limit stresses of the material and the limit deformations, both established in the regulations.
Usually, these evaluations of stress and displacements are carried out using a finite element method, which will be described in Section 2.1. Some considerations about skeletal frame structures are shown in Section 2.2. Finally, the handled test case is described in Section 2.3.

2.1. Finite Element Method

Structural calculation involves solving a finite element model (FEM) [] with Hermite approximation functions and its associated system of linear equations.
In the flat case (as is the test case to be solved) each node has three degrees of freedom in displacement: horizontal, vertical, and rotation, in addition to three types of associated force: horizontal force, vertical force, and moment. The vectors of forces (F) and displacements (U) are related through the stiffness matrix (K): F = KU. This matrix K depends on the geometrical properties of the section and the material properties and is created by an assembly process from each element (member) of the structure. To reduce the bandwidth of the matrix and consequently its calculation time, the inverse Cuthill–McKee renumbering method is applied, after which the system of equations is solved, obtaining the displacements of the nodes. From these, the stresses in any section of the structure can be calculated.
The next step is to obtain, by applying the direct method of stiffness described above, from the definition of the design variables of a skeletal structure (positions of the nodes, connectivity of the members, geometry of the section of each member, loads to which it is subjected, links, material), the weight/mass, and the values of stresses and displacements.
Finally, after calculating the stresses and displacements with FEM, they are compared with the limit values established for stresses (for each bar) and displacements (for certain specified nodes). If any of these are exceeded, their value is added to the violation of the global constraint, which is the value to be predicted by the surrogate model.

2.2. Skeletal Frame Structures

The geometrical properties of the cross-section that influence the calculation of stresses and strains explained in the previous section depend on the type of structure considered.
In the case of structures with articulated nodes (trusses), they are characterized by having loads only at the nodes and not resisting moments, and only the cross-section area is required. In the case of member structures with rigid nodes (frames), as is the test case to be solved, where the nodes between the members are rigid and moments must be considered, more geometrical magnitudes of the member must be considered in the calculation. In addition to the area, other magnitudes such as the section modulus, the section height ratio, and the moment of inertia must be considered.
In the design of structures in the real world, standardized cross-sections are used and are included in databases [] listed by cross-section shapes such as IPE in Europe or S in America (I-shaped sections), HEB in Europe, or W in America (H-shaped sections).

2.3. Test Case

The test case handled in this article is based on the one published in [], which deals with the optimization of continuous sections. This test case is utilized as a reference test case in the optimization of steel structures with rigid nodes by means of evolutionary algorithms from [], considering normalized cross-sections (discrete variables), as useful in the real engineering case. It is a four-bar structure subjected to loads and geometry (dimensions in meters), as described in Figure 1.
Figure 1. Test case. Frame structure dimensions and loads (lengths in meters; loads in T/m).
The density and modulus of elasticity correspond to the values of standard structural steel as follows: 7850 kg/m3 and 2.1 × 105 MPa, respectively. A maximum displacement at the midpoint of the longest beam (bar numbered 2 in Figure 1) of its length divided by 300 is also considered. All bars belonging to the IPE type, from IPE-080 to IPE-600 (eighteen types), with allowable stresses of 240 MPa, are considered.
All the data described above (design parameters) are necessary for the calculation of the structure as explained in previous subsections. In the case of the discrete sizing optimization problem, the design variables are exclusively the type of cross-section of each of the four members of the frame. These indices will be considered as input data of the surrogate model, with the value of the constraints of the problem (stresses and displacements) being its output. The description of the surrogate model is presented in the following section.

4. Experimental Results and Discussion

This section presents the experimental results obtained from the application of the NSGA-II algorithm to the multi-objective optimization of neural network surrogate models. The analysis focuses on the evolution of the hypervolume (HV) indicator, which reflects the quality of the non-dominated solutions in terms of training time and validation loss.
Figure 4 shows the evolution of the mean HV indicator across 9 independent runs of NSGA-II, using population sizes of 24 and 30, over 3000 fitness function evaluations. The experiments were conducted both with and without fitness function scaling. The left side of the figure corresponds to the case with a 5% training set size, while the right side shows results for a 15% training set size. The vertical scales are equivalent, allowing for a direct comparison between configurations.
Figure 4. Mean hypervolume convergence; training set sizes 5% (left) and 15% (right); dashed lines represent one standard deviation boundaries.
It can be observed that, with equivalent vertical scales, the initial hypervolume (HV) values in the 5% training set case—where training is faster—are slightly higher, ranging from 1.170 to 1.173, compared to the 15% case, which ranges from 1.158 to 1.165. However, after reaching the stopping criterion of the evolutionary process, the final HV values are superior in the 15% training set case, exceeding 1.185 in 3 out of 4 runs. This improvement is attributed to the larger training set, which enables the surrogate models to achieve greater accuracy.
A similar trend is observed in Figure 5, which shows the convergence of the median HV indicators. The final HV values in the scaled cases (represented by the blue and yellow lines) are consistently higher than those in the non-scaled cases. The configuration with the smallest population size (24 individuals and 125 generations) and no fitness scaling—represented by the red line—exhibits the lowest performance.
Figure 5. Median hypervolume convergence; training set sizes 5% (left) and 15% (right); dashed lines represent standard deviation boundaries.
Median non-dominated fronts at stopping criterion of the case of 30 population size scaling fitness functions versus random search are shown in Figure 6. NSGA-II optimization designs clearly dominate the random search solutions (yellow curves are lower and more left than black curves) in both training set sizes. Median non-dominated front of the 15% training set size dominates the one of the 5% training set size, also clearly appreciated in pink and gray lines of Figure 7.
Figure 6. Median non-dominated front (based on HV value, 9 independent executions); population size 30, scaled fitness functions versus random. Training set sizes 5% (left) and 15% (right).
Figure 7. Global accumulated non-dominated front versus median fronts of 5% and 15% training sizes (as in Figure 5).
The convergence data from the optimisation processes carried out (both populations and hypervolume indicator values) are available as a Supplementary File associated with this manuscript.
The global accumulated non-dominated front obtained from all tested experiments, consisting of 49 neural network designs, is shown as a red line in Figure 7. Detailed information on these designs—including hyperparameter configurations, corresponding experimental settings, and fitness function values—is provided in Table 2, sorted from lowest to highest training time (and inversely, from highest to lowest validation loss). It can be observed that the accumulated front dominates the pink and gray fronts, which correspond to the yellow curves previously shown in Figure 6.
Table 2. Accumulated non-dominated solutions (5% + 15% training set sizes), including hyperparameters values.
Figure 8 shows the boxplot distributions of the final HV for the 5% and 15% training set cases, shown on the left and right sides of the figure, respectively. The plots also include the distribution obtained from a random search, which performs significantly worse than the evolutionary optimization methods.
Figure 8. Boxplots of HV distribution (9 independent executions) at final population.
Table 3 and Table 4 present the results of the Friedman statistical significance test for the 5% and 15% training set cases, respectively. The p-values obtained—4.6 × 10−3 and 3.0 × 10−3—are below the significance threshold of 0.05, indicating statistically significant differences among the methods. The Bergmann–Hommel post hoc tests, shown in Table 5 and Table 6, reveal that random search performs worse than all NSGA-II optimization configurations, except for the case with a population size of 24, no fitness scaling, and a 15% training set size, where the p-value is 0.0683, slightly above the 0.05 significance level.
Table 3. Average ranking of the algorithms (comparison based on hypervolume at stopping criterion)—the lower, the better; Friedman test; 5 % training set size; p-value 0.004655.
Table 4. Average ranking of the algorithms (comparison based on hypervolume at stopping criterion)—the lower, the better; Friedman test; 15 % training set size; p-value 0.003019.
Table 5. Adjusted p values, Bergmann–Hommel’s post hoc procedure (comparison based on hypervolume at stopping criterion); 5 % training set size; statistical significance difference (p-value < 0.05) in italics.
Table 6. Adjusted p values, Bergmann–Hommel’s post hoc procedure (comparison based on hypervolume at stopping criterion); 15 % training set size; statistical significance difference (p-value < 0.05) in italics.
Among the NSGA-II configurations, although no statistically significant differences were found, the scaled fitness function cases consistently achieved better rankings than the non-scaled ones in both training set sizes (see Table 3 and Table 4).
In Table 7, scaled cases of the 5% and 15% training set sizes using a Friedman test are compared. The results indicate that there are no statistically significant differences between the two configurations (p-value = 0.61), although the 15% training set cases exhibit a more favorable ranking than those with 5%.
Table 7. Average ranking of the algorithms (comparison based on hypervolume at stopping criterion)—the lower, the better; Friedman test; fitness function scaled cases; p-value 0.6149.
Based on the analysis of Figure 6 and Figure 8, as well as the Friedman ranking tables and the Bergmann–Hommel post hoc tests, the superiority of the evolutionary approach using the NSGA-II algorithm over random search is clearly demonstrated.
Detailed information on the global accumulated non-dominated front obtained from all tested experiments is provided in Table 2, including hyperparameter values, corresponding configurations, and fitness function results. The table is sorted from lowest to highest training time (and inversely, from highest to lowest validation loss). The most accurate model achieves a validation loss of approximately 0.038 million with a training time of 9.7 min, while the least accurate model requires only 0.489 min of training time but yields a validation loss of 1315 million. Notably, three models located in the lower-left corner of the Pareto front exhibit validation losses below 0.1 million and training times under 2 min and can be considered compromise solutions (solutions ranked 35th, 36th, and 37th in Table 2).
Designs obtained using the 15% training set size account for 85% of the individuals in the non-dominated front, including extreme cases. In contrast, designs generated with the 5% training set size constitute less than 15% of the front and are associated with higher validation losses and shorter training times. This suggests that a 5% training set size (3275 individuals) is insufficient for this test case, whereas a 15% training set size (9830 individuals) appears to be more appropriate.
Additionally, among the 49 non-dominated individuals, 22 correspond to the configuration with a population size of 30 and scaled fitness functions, 11 to population size 30 without scaling, 9 to population size 24 without scaling, and the remaining 7 to population size 24 with scaling.
Regarding the number of hidden layers (with search values ranging from 1 to 7), a value of 1 is only present in the leftmost solution, which corresponds to the lowest training time and the highest validation loss. Except for this case, the eight individuals with better training time have between 2 and 3 hidden layers. Individuals ranked 9th to 35th in terms of training time exhibit hidden layer counts ranging from 4 to 7. Notably, the 14 non-dominated solutions with the lowest validation loss consistently feature three hidden layers, suggesting that this configuration offers a favorable balance between accuracy and training efficiency.
Regarding the number of nodes per layer (with search values ranging from 10 to 400), the top 22 individuals, in terms of training time, have between 11 and 55 nodes per layer. Solutions ranked 23rd to 35th show values between 75 and 158 nodes, while those ranked 36th to 49th range from 204 to 373 nodes per layer. Although the trend is not strictly monotonic, there is a noticeable tendency for higher accuracy (and correspondingly longer training times) to be associated with an increased number of nodes per layer within the recommended ranges.
Regarding batch size (with search values ranging from 32 to 1024), the 19 designs with the lowest training time exhibit batch sizes between 637 and 1023. Designs ranked 20th to 28th have batch sizes ranging from 205 to 985, while those ranked 29th to 37th fall between 124 and 199. The 11 designs with the best validation loss show batch sizes between 38 and 90. Although the trend is not strictly monotonically decreasing, there is a noticeable tendency for smaller batch sizes to be associated with higher accuracy and, correspondingly, longer training times.
Regarding the alpha parameter of the Leaky ReLU activation function (with search values ranging from 0.05 to 0.5), dominant values are concentrated between 0.05 and 0.08. Specifically, 20 solutions fall within the range 0.05–0.06, 17 within 0.06–0.07, 7 within 0.07–0.08, 3 within 0.08–0.10, and 2 between 0.10 and 0.23. Notably, 75% of the values lie within the interval from 0.051 to 0.069, which is close to the lower bound of the search space.
It is also observed that the 10 individuals with the best validation loss share common characteristics: they use three hidden layers, between 204 and 373 nodes per layer, and batch sizes ranging from 38 to 71. Among these, eight originate from experiments where the fitness function was scaled. Training times for these models range from 2.43 to 9.705 min.
Additionally, these 49 optimum designs from the multi-objective point of view (non-dominated solutions) were taken as reference designs and trained with the full training dataset (80%, 52,429 structures). The resulting training times, validation set losses, and test set mean squared errors (MSE) are represented in Figure 9. In this figure, blue dots refer to the test set results, while red dots correspond to validation set results. Non-dominated solutions have been highlighted with connecting lines, and their detailed metrics are shown in Table 8 and Table 9. Notably, eight solutions are common to both tables.
Figure 9. Accumulated non-dominated neural network designs (as in Table 2), retrained using 80% of the dataset, and results of validation set loss and test set MSE versus training time.
Table 8. Results obtained using 80% of the dataset for training, showing training time and validation loss for non-dominated solutions (highlighted with red lines in Figure 9).
Table 9. 80% training set size results in terms of training time and test set MSE, non-dominated solutions (blue line solutions in Figure 9).
Table 9 includes 11 solutions, among which the top five most accurate designs all share a configuration of three hidden layers. The remaining designs, which exhibit lower training times but higher validation losses, have between 5 and 7 hidden layers. The number of nodes per layer increases (though not monotonically) from 11 to 304 when moving from less accurate to more accurate designs. Similarly, batch size decreases (also non-monotonically) from 1012 to 38 along the same accuracy gradient. The alpha parameter of the Leaky ReLU activation function varies within the range from 0.0514 to 0.0728, without a clear trend.

5. Conclusions

From a multi-objective point of view, the importance of hyperparameter tuning in the performance of the neural network-based metamodel has been demonstrated, considering both the accuracy and complexity of the network. Extensive experimentation across different population sizes, training set proportions, and fitness scaling strategies revealed interesting insights: larger training sets (15%) provide a greater number of designs in the accumulated non-dominated front and scaled fitness functions contribute to improved convergence and final hypervolume indicators. Statistical tests confirm the superiority of evolutionary optimization over random search.
Instead of increasing the training set size during the optimization process (highly costly), in this manuscript only the non-dominated solutions were afforded to train with the full training set (80%), as shown in Table 8 and Table 9, with satisfactory outcomes.
The analysis of the 49 non-dominated solutions reveals clear trends in hyperparameter configurations in the handled test case. Models containing three hidden layers, approximately 300 neurons each, and small batch sizes (38–71) tend to achieve the best validation performance in less time during training. Additionally, the alpha parameter of the Leaky ReLU activation function is strongly concentrated near the lower bound of the search space (0.05–0.50), suggesting its relevance in fine-tuning model behavior.
In the handled test case, the final non-dominated front (e.g., Table 9) provides solutions that combine a high number of layers (6–7) with a low number of nodes per layer (11–55) for the fastest training solutions with the best accuracy. Alternatively, solutions with three layers and a high number of nodes per layer (200–300) provide the most accurate solutions and train more quickly. This combination is obviously not optimal in terms of both accuracy and network complexity simultaneously. Thus, in other referenced studies, even when solving larger structures, a limited number of network configurations that do not take this variability into account (either considering only a small number of layers with a small number of nodes per layer or a large number of layers with a large number of nodes per layer, or even limiting the number of layers in the network to a number shown here to be insufficient, such as four) have been tested.
The use of surrogate modeling in structural engineering is linked to those needs/applications that require a high number of structural evaluations, in which case the straightforward evaluation using numerical models (as the finite element method) would be unaffordable to available computational resources when repeated many times. They could typically comprise the cases of design, control, uncertainty quantification, and optimization []. Although in the first three former applications a surrogate covering the whole search space would be enough, in the latter case (optimization) the exploration of the variable space is not uniform but follows a search path. Then, it would be necessary to update the surrogate model during the optimization convergence, therefore requiring many trainings of the surrogate during the search. The multi-objective neural architecture search (NAS) approach proposed here allows us to obtain the most accurate surrogate model for every training time, or alternatively seen, the lower training time model for every accuracy. Therefore, low fidelity models (less accurate but faster to train) are available in the left part of the attained non-dominated set provided by the methodology of this manuscript, while high fidelity models (more accurate but slower to train) are available in the right part of the non-dominated set. The decision-maker/engineer is able to choose the surrogate model among the non-dominated set which mostly fits his preferences/needs. This also allows a more scientifically accurate selection when designing multi-fidelity strategies that combine simultaneously high fidelity and low fidelity methods (as in, e.g., []).
The generalization of these conclusions would be subject to a broader study with several structures of different sizes and/or types, which could be the subject of future studies. Among the potential future lines of research, the application of this methodology to analogous structural problems that are more complex or larger in scale could be explored, including other structural types, provided that an adequate database is available. Notwithstanding the elevated computational expense associated with the network training time during the optimization process, this article proposes a methodology that assists in resolving this issue. The methodology involves the execution of the optimization with a constrained training database size and the exclusive utilization of the complete database for the training of the optimal models that appear in the non-dominated solution front.
These findings provide valuable insights into designing efficient and accurate surrogate models in structural engineering and lay the groundwork for future multi-objective neural architecture search work in this field. Using this methodology for more complex issues is a promising avenue for future research, using either the same or alternative objective functions, depending on the nature of the problem.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/a18120754/s1. The data concerning the experimental results that are available consist of the populations and their hypervolumes along the convergence of the test cases, as well as at the stopping criterion. These data can be found in the Supplementary File associated with this manuscript.

Author Contributions

Conceptualization, N.L.-G., E.R. and D.G.; methodology, N.L.-G., E.R. and D.G.; software, N.L.-G., E.R. and D.G.; validation, N.L.-G., E.R. and D.G.; formal analysis, N.L.-G., E.R. and D.G.; investigation, N.L.-G., E.R. and D.G.; resources, E.R. and D.G.; data curation, N.L.-G.; writing—original draft preparation, N.L.-G., E.R., and D.G.; writing—review and editing, N.L.-G., E.R., and D.G.; visualization, N.L.-G. and E.R.; supervision, E.R. and D.G.; project administration, E.R. and D.G.; funding acquisition, E.R. and D.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research has been supported by: Ministerio de Ciencia, Innovación y Universidades, Gobierno de España, grant contract PID2024-57499OB-C31; and Consejería de Universidades, Ciencia e Innovación y Cultura del Gobierno de Canarias, grant contract PRECOMP02 SD-24/03. Also, first author acknowledges predoctoral grant FPU22/01933 of Ministry of Universities of Spanish Government.

Data Availability Statement

Data will be made available on request.

Acknowledgments

The computational equipment used in this research was acquired through the project “Infraestructura de Computación Científica para Aplicaciones de Inteligencia Artificial y Simulación Numérica en Medioambiente y Gestión de Energías Renovables (EIS 2021 04)”, awarded to the Instituto Universitario SIANI by the Consejería de Economía, Conocimiento y Empleo del Gobierno de Canarias, processed by the Agencia Canaria de Investigación, Innovación y Sociedad de la Información, and co-funded by the European Regional Development Fund (ERDF) under the Canary Islands ERDF Operational Program.

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

Abbreviations

The following abbreviations are used in this manuscript:
ADAMAdaptive Moment Estimation
ANNArtificial Neural Network
EAEvolutionary Algorithm
HVHypervolume
MSEMean Squared Error
NSGA-IINon-dominated Sorting Genetic Algorithm II

References

  1. Greiner, D.; Periaux, J.; Emperador, J.M.; Galván, B.; Winter, G. Game theory based Evolutionary Algorithms: A review with Nash applications in structural engineering optimization problems. Arch. Comput. Methods Eng. 2017, 24, 703–750. [Google Scholar] [CrossRef]
  2. Hajela, P.; Berke, L. Neurobiological computational models in structural analysis and design. Comput. Struct. 1991, 41, 657–667. [Google Scholar] [CrossRef]
  3. Waszczyszyn, Z. Some recent and current problems of neurocomputing in civil and structural engineering. In Advances in Computational Structures Technology; Topping, B.H.V., Ed.; CIVIL-COMP Press: Edinburgh, Scotland, 1996; pp. 43–58. [Google Scholar]
  4. Papadrakakis, M.; Lagaros, N.; Tsompanakis, Y. Structural optimization using evolution strategies and neural networks. Comput. Methods Appl. Mech. Eng. 1998, 156, 309–333. [Google Scholar] [CrossRef]
  5. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  6. Bäck, T.H.W.; Kononova, A.V.; van Stein, B.; Wang, H.; Antonov, K.A.; Kalkreuth, R.T.; de Nobel, J.; Vermetten, D.; de Winter, R.; Ye, F. Evolutionary algorithms for parameter optimization—Thirty years later. Evol. Comput. 2023, 31, 81–122. [Google Scholar] [CrossRef] [PubMed]
  7. Deb, K. Multi-Objective Optimization Using Evolutionary Algorithms; John Wiley & Sons: Chichester, UK, 2001. [Google Scholar]
  8. Hernández-Gil, M.; Ramos-de-Miguel, A.; Greiner, D.; Benítez, D.; Ramos-Macías, A.; Escobar, J.M. A computational model for multiobjective optimization of multipolar stimulation in cochlear implants: An enhanced focusing approach. Expert Syst. Appl. 2025, 280, 127472. [Google Scholar] [CrossRef]
  9. Greiner, D.; Cacereño, A. Enhancing the maintenance strategy and cost in systems with surrogate assisted multiobjective evolutionary algorithms. Dev. Built Environ. 2024, 19, 100478. [Google Scholar] [CrossRef]
  10. He, C.; Zhang, Y.; Gong, D.; Ji, X. A review of surrogate-assisted evolutionary algorithms for expensive optimization problems. Expert Syst. Appl. 2023, 217, 119495. [Google Scholar] [CrossRef]
  11. Xing, L.; Gardoni, P.; Song, G.; Zhou, Y. Deep learning-based surrogate capacity models and multi-objective fragility estimates for reinforced concrete frames. Comput. Methods Appl. Mech. Eng. 2025, 440, 117928. [Google Scholar] [CrossRef]
  12. Quevedo-Reina, R.; Álamo, G.; Padrón, L.A.; Aznárez, J.J. Surrogate model based on ANN for the evaluation of the fundamental frequency of offshore wind turbines supported on jackets. Comput. Struct. 2023, 274, 106917. [Google Scholar] [CrossRef]
  13. Quevedo-Reina, R.; Álamo, G.; Aznárez, J.J. ANN-based surrogate model for the structural evaluation of jacket support structures for offshore wind turbines. Ocean Eng. 2025, 317, 119984. [Google Scholar] [CrossRef]
  14. Elsken, T.; Hendrik-Metzen, J.; Hutter, F. Neural Architecture Search: A Survey. J. Mach. Learn. Res. 2019, 20, 1–21. [Google Scholar]
  15. Pour Avval, S.S.; Eskue, N.; Groves, R.; Yaghoubi, V. Systematic review of neural architecture search. Artif. Intell. Rev. 2025, 58, 73. [Google Scholar] [CrossRef]
  16. Negrin, I.; Kripka, M.; Yepes, V. Metamodel-assisted design optimization in the field of structural engineering: A literature review. Structures 2023, 52, 609–631. [Google Scholar] [CrossRef]
  17. Etim, B.; Al-Ghosoun, A.; Renno, J.; Seaid, M.; Mohamed, M.S. Machine Learning-Based Modeling for Structural Engineering: A Comprehensive Survey and Applications Overview. Buildings 2024, 14, 3515. [Google Scholar] [CrossRef]
  18. Saini, R. A Review on Artificial Neural Networks for Structural Analysis. J. Vib. Eng. Technol. 2025, 13, 142. [Google Scholar] [CrossRef]
  19. Samadian, D.; Muhit, I.B.; Dawood, N. Application of Data-Driven Surrogate Models in Structural Engineering: A Literature Review. Arch. Comput. Methods Eng. 2025, 32, 735–784. [Google Scholar] [CrossRef]
  20. Kim, Y.H.; Reddy, B.; Yun, S.; Seo, C. NEMO: Neuro-evolution with multi-objective optimization of deep neural network for speed and accuracy. In Proceedings of the International Conference on Machine Learning ICML 2017, Auto Machine Learning Workshop, Sydney, Australia, 6–11 August 2017. [Google Scholar]
  21. White, C.; Safari, M.; Sukthanker, R.; Ru, B.; Elsken, T.; Zela, A.; Dey, D.; Hutter, F. Neural Architecture Search: Insights from 1000 Papers. arXiv 2021, arXiv:2301.08727. [Google Scholar] [CrossRef]
  22. Deb, K.; Pratap, A.; Agarwal, S.; Meyarivan, T. A fast and elitist multiobjective genetic algorithm: NSGA-II. IEEE Trans. Evol. Comput. 2002, 6, 182–197. [Google Scholar] [CrossRef]
  23. Akopov, A.S. An Improved Parallel Biobjective Hybrid Real-Coded Genetic Algorithm with Clustering-Based Selection. Cybern. Inf. Technol. 2024, 24, 32–49. [Google Scholar] [CrossRef]
  24. Nguyen, T.L.; Nguyen, Q.A. A Multi-Objective PSO-GWO Approach for Smart Grid Reconfiguration with Renewable Energy and Electric Vehicles. Energies 2025, 18, 2020. [Google Scholar] [CrossRef]
  25. Tian, Y.; Cheng, R.; Zhang, X.Y.; Jin, Y.C. PlatEMO: A MATLAB Platform for Evolutionary Multi-Objective Optimization [Educational Forum]. IEEE Comput. Intell. Mag. 2017, 12, 73–87. [Google Scholar] [CrossRef]
  26. Deb, K. An interview with Kalyanmoy Deb 2022 ACM fellow. ACM SIGEVOlution 2023, 16, 1–6. [Google Scholar] [CrossRef]
  27. Greiner, D.; Winter, G.; Emperador, J.M. Optimising frame structures by different strategies of genetic algorithms. Finite Elem. Anal. Des. 2001, 37, 381–402. [Google Scholar] [CrossRef]
  28. Bathe, K.J. Finite Element Procedures; Prentice Hall: Hoboken, NJ, USA, 2006. [Google Scholar]
  29. EN 1993; Eurocode 3: Design of Steel Structures. European Community Standard: Luxembourg, 1993.
  30. Hernández-Ibáñez, S. Structural Optimum Design Methods. Colección Seinor; Colegio de Ingenieros de Caminos, Canales y Puertos: Madrid, Spain, 1990. (In Spanish) [Google Scholar]
  31. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  32. Wang, G.; Giannakis, G.B.; Chen, J. Learning ReLU networks on linearly separable data: Algorithm, optimality, and generalization. IEEE Trans. Signal Process. 2019, 67, 2357–2370. [Google Scholar] [CrossRef]
  33. Guerreiro, A.P.; Fonseca, C.M.; Paquete, L. The hypervolume indicator: Computational problems and algorithms. ACM Comput. Surv. 2021, 54, 1–42. [Google Scholar]
  34. Python Software Foundation. Python: Version 3.12.7 Documentation. 2024. Available online: https://docs.python.org/3/ (accessed on 10 November 2025).
  35. Blank, J.; Deb, K. Pymoo: Multi-objective optimization in Python. IEEE Access 2020, 8, 89497–89509. [Google Scholar] [CrossRef]
  36. Abadi, M.; Barham, P.; Chen, J.; Chen, Z.; Davis, A.; Dean, J.; Devin, M.; Ghemawat, S.; Irving, G.; Isard, M.; et al. TensorFlow: A System for Large-Scale Machine Learning. arXiv 2016, arXiv:1605.08695. [Google Scholar]
  37. Garcia, S.; Herrera, F. An extension on ‘statistical comparisons of classifiers over multiple data sets’ for all pairwise comparisons. J. Mach. Learn. Res. 2007, 9, 2677–2694. [Google Scholar]
  38. Benner, P.; Gugercin, S.; Willcox, K. A survey of projection-based model reduction methods for parametric dynamical systems. SIAM Rev. 2015, 57, 483–531. [Google Scholar] [CrossRef]
  39. Porta-Ko, A.; González-Horcas, S.; Pons-Prats, J.; Bugeda, G. Development of a multi-fidelity optimisation strategy based on hybrid methods. Struct. Multidiscip. Optim. 2024, 67, 163. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Article metric data becomes available approximately 24 hours after publication online.