Open Access
This article is

- freely available
- re-usable

*Computers*
**2019**,
*8*(1),
2;
https://doi.org/10.3390/computers8010002

Article

Neural Network-Based Formula for the Buckling Load Prediction of I-Section Cellular Steel Beams

^{1}

Research & Development, Abambres’ Lab, 1600-275 Lisbon, Portugal

^{2}

School of Civil Engineering, University of Leeds, LS2 9JT, Leeds, UK

^{3}

Tal Projecto, 1350-252 Lisboa, Portugal

^{*}

Author to whom correspondence should be addressed.

Received: 29 November 2018 / Accepted: 21 December 2018 / Published: 26 December 2018

## Abstract

**:**

Cellular beams are an attractive option for the steel construction industry due to their versatility in terms of strength, size, and weight. Further benefits are the integration of services thereby reducing ceiling-to-floor depth (thus, building’s height), which has a great economic impact. Moreover, the complex localized and global failures characterizing those members have led several scientists to focus their research on the development of more efficient design guidelines. This paper aims to propose an artificial neural network (ANN)-based formula to precisely compute the critical elastic buckling load of simply supported cellular beams under uniformly distributed vertical loads. The 3645-point dataset used in ANN design was obtained from an extensive parametric finite element analysis performed in ABAQUS. The independent variables adopted as ANN inputs are the following: beam’s length, opening diameter, web-post width, cross-section height, web thickness, flange width, flange thickness, and the distance between the last opening edge and the end support. The proposed model shows a strong potential as an effective design tool. The maximum and average relative errors among the 3645 data points were found to be 3.7% and 0.4%, respectively, whereas the average computing time per data point is smaller than a millisecond for any current personal computer.

Keywords:

elastic buckling; cellular steel beams; perforated beams; artificial neural networks; finite element analysis## 1. Introduction

The use of cellular beams (i.e., perforated beams with circular web openings) in the construction sector has significantly increased over the past decade on account of the distinct and discreet advantages they offer. Cellular beams are applicable for long span structures, where integration of services such as ventilation ducts and lightings systems within the beam is attained, but also for short spans, where spatial interference among concentrated mechanical devices and structural elements may require a compromised solution. Cellular beams allow reducing the height of buildings to fit a required number of floors, otherwise fitting more floors in a given height limit, thus having a significant economic impact to the whole structure’s budget. Furthermore, cellular beams offer practical advantages such as the possibility of (i) fixing the ceilings directly to the beams’ lower flanges instead of requiring additional suspension elements, and (ii) allowing future equipment addition or replacement within the existent void holes. In fact, with the wider adoption of Building Information Modeling (BIM), the knowledge of those expansion possibilities is becoming a valuable asset for the building management.

Long span and lightweight structures also benefit from flexible designs with the fewer number of columns and foundations, and thus from the reduced construction time [1]. The increase in beam depth due to the castellation process (i.e., profile cutting manufacturing) also provides greater flexural stiffness having a final section with larger section modulus [2]. However, the presence of web openings significantly influences the structural performance of the beams, which in particular is dependent on the geometry (shape, diameter, and critical opening length), location (shear-moment interaction), and spacing (closely and widely spaced) between perforations. The perforations lead to complex structural behaviors, attributed to the distribution of forces and stresses in the vicinity of the openings. This results in rather complicated and conservative design procedures ([2,3]).

The first analytical model proposed by Uenoya and Redwood [4] provided critical baseline data relating to the behavior of perforated beams by studying the in-plane stress distribution pattern using perforated plates. Later, Lucas and Darwin [5] proposed a design process based on the identification of the maximum bending and shear capacities at the web openings. It was afterwards suggested that the nominal capacities for the combinations of the bending moment and shear at each opening were determined. This method was accepted by the AISC [6] and the ASCE 23-97 [7]. However, the method only provided a reasonable accurate load estimate for beams with small height, whereas for greater heights becomes conservative. The method is also restrictive to a maximum opening height of 0.7 h, since the average errors were found to increase significantly above this range. Also, in 1990, Ward [8] proposed a simplified semi-empirical web-post model using finite element Modeling; however, this model was restrictive as it was based on a limited number of geometric configurations and best results were found with an error of 30%. Following, Chung et al. [9] studied the Vierendeel mechanism and derived moment-shear interaction curves for various types of perforated beams which ultimately led to the development of a generalized moment-shear curve to assess the load carrying capacities of beams with various openings. In 2003, Chung at al. [10] further analyzed this moment-shear interaction curve and concluded that different shapes and sizes of openings can affect a beam differently and that this curve is more relevant for beams with large openings. In the same year, Chung et al. [10] reported that the moment-shear curve can be somewhat conservative when analyzing beams with small web openings. However, Chung et al.’s findings were identified as a relatively good method as it provided a good approximation with mean errors of 15–25% (for various shapes of web openings), where cellular beams indicated an error of 15.8%. Since 2009, Tsavdaridis et al. published studies [1,11,12,13,14] on thin-walled perforated beams with circular and other novel non-standard web openings investigating the web-post buckling failure mode for closely spaced web openings as well as the Vierendeel mechanism when large isolated web openings. The studies revealed that the Vierendeel bending is influenced by both the shape and size of an opening and the load carrying capacity of beams with large web openings can be found by examining the formation (the order and position) of plastic hinges. The extreme opening diameter (0.8 h) as opposed to the maximum value of 0.75 h presented in earlier literature, was introduced by Tsavdaridis and D’Mello [13] in order to comprehensively develop an understanding of the parameters which affect the structural behavior of perforated beams. Akrami and Erfani (2016), in a comparative analysis of the design methodologies for perforated beams, found that the works of Chung et al [10] and Tsavdaridis and D’Mello [13] were less restrictive as compared to other design methods (ASCE 23-97, SCI-P100, SCI-P355) and produce the lowest errors. In 2011, Lawson and Hicks [15] published the SCI-P355 design guidelines, an update to SCI-P068 [16] and SCI-P100 [8] which proposed that the Vierendeel bending resistance is dependent on the classification of the web of the T-beams. This approach produced acceptable approximations for openings of specific dimensions where the best results were found with an error of 25–30%. It is worth noting that Chung et al. [10], Verweij [17] and Morkhade and Gupta [2] have reported that the current guidelines, specifically SCI-P100 [8] and SCI-P355 [15], are inadequate, complicated, and conservative when it comes to the design of perforated steel beams.

Artificial neural networks (ANN) have become a popular method to predict the response of structures. Gholizadeh et al. [18] presented a study relating the use of ANN in the evaluation of the load carrying capacity of the web-post of castellated steam beams based on 140 FE models. The computational technique generated predictions with great accuracy when compared to other methods. Sharifi and Tohidi [19] also illustrated the application of ANN to accurately estimate the elastic buckling load capacity of steel bridge girders that have rectangular openings at the bottom zone in the web. This is considered as the worse possible location to place an opening to resist lateral torsional buckling. The ANN formula was derived from 21 FE models which managed to accurately predict the elastic buckling load. In 2014, Tohidi and Sharifi [20] demonstrated the versatility of ANN by studying the buckling capacity of steel beams with rectangular web openings that has experienced corrosion in the web. In addition, Tohidi and Sharifi [21] developed an ANN model to estimate the bearing capacity of steel girders with corrosion at the bearing region. The ANN empirical formulas obtained were reported to be accurate in predicting the residual capacity of deteriorated steel beams. Examples of recent and relevant works concerning the application of ANN to other types of structures within civil engineering read [22,23,24,25].

The current study was motivated by the lack of rational (simple, efficient, and accurate) design procedures relating to the buckling response of cellular beams. This paper proposes an ANN-based formula to precisely compute the critical elastic buckling load of simply supported cellular beams under uniformly distributed vertical loads, as function of eight independent geometrical parameters. This research is the first step of an ongoing investigation that aims to propose a novel and simple analytical design method to accurately compute the inelastic resistance of cellular steel beams. A FE-based dataset comprising 3645 points was generated for this study, in order to allow the ANN model to have a significant generalization ability and be considered as a powerful tool for structural engineers and scientists to (i) estimate the elastic buckling load of cellular steel beams, and (ii) efficiently perform sensitivity analyses to further assess the behavior of those members.

## 2. Data Generation

#### 2.1. FE Modeling

Three-dimensional FE models were developed using ABAQUS [26,27], which were then parametrized to generated 3645 simulations. Typical values for the modulus of elasticity and Poisson’s ratio were adopted (E = 210 GPa, ν = 0.3). All models are simply supported where one end allows in-plane rotations but not translations and the other admits translations along the beam axis, beyond in-plane rotations. End twisting rotations were prevented by restraining both the top and bottom flange tips against out-of-plane displacements at the supports. A unitary load was applied to the top flange as a uniformly distributed pressure (then converted to a line load for ANN simulation purposes—see Table 1). The FE mesh adopted was quad-dominated using shell elements of type SR8, which was tested against experimental work conducted by Tsavdaridis and D’Mello [12]), and Surtees and Liu [28], providing accurate and reliable results [29]. The mesh sizes recommended in [30] for web and flanges were adopted. Figure 1a illustrates the various parameters considered in the parametric analysis, whereas Figure 1b illustrates one application of these type of structural members.

#### 2.2. Parametric Analysis

The parametric models were submitted to the ABAQUS Lanczos Eigensolver using Python scripts. Table 1 presents the possible values taken for each independent (parametric) variable (see Figure 1a) considered in the FEA. The ‘first’ web opening was placed at the centre of the beam whereas the remaining ones were offset from the former until (for a fixed beam’s length, opening diameter, and web-post width) no more circular openings could fit within member’s length. This approach resulted in 135 different distances from the end opening edge to the centre-line of the endplate (distance named ‘opening-support end distance’ in Table 1). Thus, combining all values of variables 1 and 3 to 8, presented in Table 1, one has (3^6) × 5 lengths = 3645 distinct steel beams for FEA (also called data points or examples in this manuscript). The 3645-point dataset considered in ANN simulations is available in [31].

## 3. Artificial Neural Networks

#### 3.1. Introduction

Machine learning, one of the six disciplines of artificial intelligence (AI) without which the task of having machines acting humanly could not be accomplished, allows us to ‘teach’ computers how to perform tasks by providing examples of how they should be done [32]. When there is abundant data (also called examples or patterns) explaining a certain phenomenon, but its theory richness is poor, machine learning can be a perfect tool. The world is quietly being reshaped by machine learning, being the artificial neural network (also referred in this manuscript as ANN or neural net) its (i) oldest [33] and (ii) most powerful [34] technique. ANNs also lead the number of practical applications, virtually covering any field of knowledge ([35,36]). In its most general form, an ANN is a mathematical model designed to perform a particular task, based in the way the human brain processes information, i.e. with the help of its processing units (the neurons). ANNs have been employed to perform several types of real-world basic tasks. Concerning functional approximation, ANN-based solutions are frequently more accurate than those provided by traditional approaches, such as multi-variate nonlinear regression, besides not requiring a good knowledge of the function shape being modeled [37].

The general ANN structure consists of several nodes disposed in L vertical layers (input layer, hidden layers, and output layer) and connected between them, as depicted in Figure 2. Associated to each node in layers 2 to L, also called neuron, is a linear or nonlinear transfer (also called activation) function, which receives the so-called net input and transmits an output (see Figure 5). All ANNs implemented in this work are called feedforward, since data presented in the input layer flows in the forward direction only, i.e. every node only connects to nodes belonging to layers located at the right-hand-side of its layer, as shown in Figure 2. ANN’s computing power makes them suitable to efficiently solve small to large-scale complex problems, which can be attributed to their (i) massively parallel distributed structure and (ii) ability to learn and generalize, i.e., produce reasonably accurate outputs for inputs not used during the learning (also called training) phase.

#### 3.2. Learning

Each connection between 2 nodes is associated to a synaptic weight (real value), which, together with each neuron’s bias (also a real value), are the most common types of neural net unknown parameters that will be determined through learning. Learning is nothing else than determining network unknown parameters through some algorithm in order to minimize network’s performance measure, typically a function of the difference between predicted and target (desired) outputs. When ANN learning has an iterative nature, it consists of three phases: (i) training, (ii) validation, and (iii) testing. From previous knowledge, examples or data points are selected to train the neural net, grouped in the so-called training dataset. Those examples are said to be ‘labelled’ or ‘unlabeled’, whether they consist of inputs paired with their targets, or just of the inputs themselves—learning is called supervised (e.g., functional approximation, classification) or unsupervised (e.g., clustering), whether data used is labelled or unlabeled, respectively. During an iterative learning, while the training dataset is used to tune network unknowns, a process of cross-validation takes place by using a set of data completely distinct from the training counterpart (the validation dataset), so that the generalization performance of the network can be attested. Once ‘optimum’ network parameters are determined, typically associated to a minimum of the validation performance curve (called early stop—see Figure 3), many authors still perform a final assessment of model’s accuracy, by presenting to it a third fully distinct dataset called ‘testing’. Heuristics suggests that early stopping avoids overfitting, i.e. the loss of ANN’s generalization ability. One of the causes of overfitting might be learning too many input-target examples suffering from data noise, since the network might learn some of its features, which do not belong to the underlying function being modeled [38].

#### 3.3. Implemented ANN features

The ‘behavior’ of any ANN depends on many ‘features’, having been considered 15 ANN features in this work (including data pre/post processing ones). For those features, it is important to bear in mind that no ANN guarantees good approximations via extrapolation (either in functional approximation or classification problems), i.e. the implemented ANNs should not be applied outside the input variable ranges used for network training. Since there are no objective rules dictating which method per feature guarantees the best network performance for a specific problem, an extensive parametric analysis (composed of nine parametric sub-analyses) was carried out to find ‘the optimum’ net design. A description of all implemented methods, selected from state of art literature on ANNs (including both traditional and promising modern techniques), is presented next—Table 2, Table 3 and Table 4 present all features and methods per feature. The whole work was coded in MATLAB [39], making use of its neural network toolbox when dealing with popular learning algorithms (1–3 in Table 4). Each parametric sub-analysis (SA) consists of running all feasible combinations (also called ‘combos’) of pre-selected methods for each ANN feature, in order to get performance results for each designed net, thus allowing the selection of the best ANN according to a certain criterion. The best network in each parametric SA is the one exhibiting the smallest average relative error (called performance) for all learning data.

It is worth highlighting that, in this manuscript, whenever a vector is added to a matrix, it means the former is to be added to all columns of the latter (valid in MATLAB).

#### 3.3.1. Qualitative Variable Representation (Feature 1)

A qualitative variable taking n distinct ‘values’ (usually called classes) can be represented in any of the following formats: one variable taking n equally spaced values in [0, 1], or 1-of-n encoding (boolean vectors —e.g., n = 3: [1 0 0] represents class 1, [0 1 0] represents class 2, and [0 0 1] represents class 3). After transformation, qualitative variables are placed at the end of the corresponding (input or output) dataset, in the same original order.

#### 3.3.2. Dimensional Analysis (Feature 2)

The most widely used form of dimensional analysis is the Buckingham’s π-theorem, which was implemented in this work as described in [40].

#### 3.3.3. Input Dimensionality Reduction (Feature 3)

When designing any ANN, it is crucial for its accuracy that the input variables are independent and relevant to the problem ([18,41]). There are two types of dimensionality reduction, namely (i) feature selection (a subset of the original set of input variables is used), and (ii) feature extraction (transformation of initial variables into a smaller set). In this work, dimensionality reduction is never performed when the number of input variables is less than six. The implemented methods are described next.

#### Linear Correlation

In this feature selection method, all possible pairs of input variables are assessed with respect to their linear dependence, by means of the Pearson correlation coefficient R
where (i) Var(X) and Cov(X, Y) are the variance of X and covariance of X and Y, respectively, and (ii) $\overline{x}$ and $\overline{y}$ are the mean values of each variable. In this work, cases where $\left|{R}_{XY}\right|\ge 0.99$ indicate that one of the variables in the pair must be removed from the ANN modeling. The one to be removed is the one appearing less in the remaining pairs $\left(X,Y\right)$ where $\left|{R}_{XY}\right|\ge 0.99$. Once a variable is selected for removal, all pairs $\left(X,Y\right)$ involving it must be disregarded in the subsequent steps for variable removal.

_{XY}, where X and Y denote any two distinct input variables. For a set of n data points (x_{i}, y_{i}), the Pearson correlation is defined by
$${R}_{XY}=\frac{{\displaystyle \sum _{i=1}^{n}\left({x}_{i}-\overline{x}\right)\left({y}_{i}-\overline{y}\right)}}{\sqrt{{\displaystyle \sum _{i=1}^{n}{\left({x}_{i}-\overline{x}\right)}^{2}}{\displaystyle \sum _{i=1}^{n}{\left({y}_{i}-\overline{y}\right)}^{2}}}}=\frac{Cov(X,Y)}{\sqrt{Var(X)Var(Y)}},$$

#### Auto-Encoder

This feature extraction technique uses itself a 3-layer feedforward ANN called auto-encoder (AE). After training, the hidden layer output (y
Concerning the learning algorithm used for all AEs, no L

_{2p}) for the presentation of each problem’s input pattern (y_{1p}) is a compressed vector (Q_{2}x 1) that can be used to replace the original input layer by a (much) smaller one, thus reducing the size of the ANN model. In this work, Q_{2}=round(Q_{1}/2) was adopted, being round a function that rounds the argument to the nearest integer. The implemented AE was trained using the ‘trainAutoencoder(…)’ function from MATLAB’s neural net toolbox. In order to select the best AE, 40 AEs were simulated, and their performance compared by means of the performance variable defined in Section 3.4. Each AE considered distinct (random) initialization parameters, half of the models used the ‘logsig’ hidden transfer functions, and the other half used the ‘satlin’ counterpart, being the identity function the common option for the output activation. In each AE, the maximum number of epochs—number of times the whole training dataset is presented to the network during learning, was defined (regardless the amount of data) by
$$\mathrm{max}epochs=\{\begin{array}{l}3000,{Q}_{1}>8\\ 1500,{Q}_{1}\le 8\end{array}.$$

_{2}weight regularization was employed, which was the only default specification not adopted in ‘trainAutoencoder (…)’.#### Orthogonal and Sparse Random Projections

This is another feature extraction technique aiming to reduce the dimension of input data Y

_{1}(Q_{1}× P) while retaining the Euclidean distance between data points in the new feature space. This is attained by projecting all data along the (i) orthogonal or (ii) sparse random matrix A (Q_{1}× Q_{2}, Q_{2}< Q_{1}), as described by Kasun et al. [41].#### 3.3.4. Training, Validation and Testing Datasets (Feature 4)

Four distributions of data (methods) were implemented, namely p

_{t}-p_{v}-p_{tt}= {80-10-10, 70-15-15, 60-20-20, 50-25-25}, where p_{t}-p_{v}-p_{tt}represent the amount of training, validation and testing examples as % of all learning data (P), respectively. Aiming to divide learning data into training, validation and testing subsets according to a predefined distribution p_{t}-p_{v}-p_{tt}, the following algorithm was implemented (all variables are involved in these steps, including qualitative ones after converted to numeric—see Section 3.3.1):- For each variable q (row) in the complete input dataset, compute its minimum and maximum values.
- Select all patterns (if some) from the learning dataset where each variable takes either its minimum or maximum value. Those patterns must be included in the training dataset, regardless what p
_{t}is. However, if the number of patterns ‘does not reach’ p_{t}, one should add the missing amount, providing those patterns are the ones having more variables taking extreme (minimum or maximum) values. - In order to select the validation patterns, randomly select p
_{v}/(p_{v}+ p_{tt}) of those patterns not belonging to the previously defined training dataset. The remainder defines the testing dataset.

It might happen that the actual distribution p

_{t}-p_{v}-p_{tt}is not equal to the one imposed a priori (before step 1), which is due to the minimum required training patterns specified in step 2.#### 3.3.5. Input Normalization (Feature 5)

The progress of training can be impaired if training data defines a region that is relatively narrow in some dimensions and elongated in others, which can be alleviated by normalizing each input variable across all data patterns. The implemented techniques are the following:

#### Linear Max Abs

Lachtermacher and Fuller [42] proposed a simple normalization technique given by
where {Y

$${\left\{{Y}_{1}\right\}}_{n}(i,:)=\frac{{Y}_{1}(i,:)}{\mathrm{max}\left\{\left|{Y}_{1}(i,:)\right|\right\}},$$

_{1}}_{n}(i, :) and Y_{1}(i, :) are the normalized and non-normalized values of the i^{th}input variable for all learning patterns, respectively. Notation ‘:’ in the column index, indicate the selection of all columns (learning patterns).#### Linear [0, 1] and [−1, 1]

A linear transformation for each input variable (i), mapping values in Y

_{1}(i,:) from [a*, b*] = [min(Y_{1}(i,:)), max(Y_{1}(i,:))] to a generic range [a, b], is obtained from
$${\left\{{Y}_{1}\right\}}_{n}\left(i,:\right)=a+\frac{\left({Y}_{1}\left(i,:\right)-a\ast \right)}{\left(b\ast -a\ast \right)}(b-a),$$

Ranges [a, b] = [0, 1] and [a, b] = [−1, 1] were considered.

#### Nonlinear

Proposed by Pu and Mesbahi [43], although in the context of output normalization, the only nonlinear normalization method implemented for input data reads
where (i) Y

$${\left\{{Y}_{1}\right\}}_{n}\left(i,j\right)=sign\left({Y}_{1}\left(i,j\right)\right)\sqrt{\frac{\left|{Y}_{1}\left(i,j\right)\right|}{{10}^{t}}}+C(i),$$

_{1}(i, j) is the non-normalized value of input variable i for pattern j, (ii) t is the number of digits in the integer part of Y_{1}(i, j), (iii) sign(…) yields the sign of the argument, and (iv) C(i) is the average of two values concerning variable i, C_{1}(i) and C_{2}(i), where the former leads to a minimum normalized value of 0.2 for all patterns, and the latter leads to a maximum normalized value of 0.8 for all patterns.#### Linear Mean Std

Tohidi and Sharifi [20] proposed the following technique
where ${\mu}_{{Y}_{1}\left(i,:\right)}$ and ${\sigma}_{{Y}_{1}\left(i,:\right)}$ are the mean and standard deviation of all non-normalized values (all patterns) stored by variable i.

$${\left\{{Y}_{1}\right\}}_{n}\left(i,:\right)=\frac{{Y}_{1}\left(i,:\right)-{\mu}_{{Y}_{1}\left(i,:\right)}}{{\sigma}_{{Y}_{1}\left(i,:\right)}},$$

#### 3.3.6. Output Transfer Functions (Feature 6)

#### Logistic

The most usual form of transfer functions is called Sigmoid. An example is the logistic function given by

$$\phi (s)=\frac{1}{1+{e}^{-s}}.$$

#### Hyperbolic Tang

The Hyperbolic Tangent function is also of sigmoid type, being defined as

$$\phi (s)=\frac{{e}^{s}-{e}^{-s}}{{e}^{s}+{e}^{-s}}.$$

#### Bilinear

The implemented Bilinear function is defined as

$$\phi (s)=\{\begin{array}{l}s,s\ge 0\\ 0,s<0\end{array}.$$

#### Identity

The Identity activation is often employed in output neurons, reading

$$\phi (s)=s.$$

#### 3.3.7. Output Normalization (Feature 7)

Normalization can also be applied to the output variables so that, for instance, the amplitude of the solution surface at each variable is the same. Otherwise, training may tend to focus (at least in the earlier stages) on the solution surface with the greatest amplitude [44]. Normalization ranges not including the zero value might be a useful alternative since convergence issues may arise due to the presence of many small (close to zero) target values [45]. Four normalization methods were implemented. The first three follow Equation (4), where (i) [a, b] = 70% [φ

_{min}, φ_{max}], (ii) [a, b] = 60% [φ_{min}, φ_{max}], and (iii) [a, b] = 50% [φ_{min}, φ_{max}], being [φ_{min}, φ_{max}] the output transfer function range, and [a, b] determined to be centered within [φ_{min}, φ_{max}] and to span the specified % (e.g., (b − a) = 0.7 (φ_{max}− φ_{min})). Whenever the output transfer functions are unbounded (Bilinear and Identity), it was considered [a, b] = [0, 1] and [a, b] = [−1, 1], respectively. The fourth normalization method implemented is the one described by Equation (6).#### 3.3.8. Network Architecture (Feature 8)

#### Multi-Layer Perceptron Network (MLPN)

This is a feedforward ANN exhibiting at least one hidden layer. Figure 2 depicts a 3-2-1 MLPN (3 input nodes, 2 hidden neurons and 1 output neuron), where units in each layer link only to some nodes located ahead. At this moment, it is appropriate to define the concept of partially- (PC) and fully-connected (FC) ANNs. In this work a FC feedforward network is characterized by having each node connected to every node in a different layer placed forward—any other type of network is said to be PC (e.g., the one in Figure 2). According to Wilamowski [46], PC MLPNs are less powerful than MLPN where connections across layers are allowed, which usually lead to smaller networks (less neurons).

Figure 4 represents a generic MLFN composed of L layers, where l (l = 1,…, L) is a generic layer and ‘ql’ a generic node, being q = 1,…, Q
where y
where φ

_{l}its position in layer l (1 is reserved to the top node). Figure 5 represents the model of a generic neuron (l = 2,…, L), where (i) p represents the data pattern presented to the network, (ii) subscripts m = 1,…, Q_{n}and n = 1,…, l − 1 are summation indexes representing all possible nodes connecting to neuron ‘ql’ (recall Figure 4), (iii) b_{ql}is neuron’s bias, and (iv) w_{mnql}represents the synaptic weight connecting units ‘mn’ and ‘ql’. Neuron’s net input for the presentation of pattern p (Sqlp) is defined as
$${S}_{qlp}={y}_{mnp}{w}_{mnql}+{b}_{ql},\hspace{1em}{y}_{mnp}{w}_{mnql}\equiv {\displaystyle \sum _{m=1}^{{Q}_{n}}{\displaystyle \sum _{n=1}^{l-1}{y}_{mnp}{w}_{mnql}}},$$

_{m1p}is the value of the m^{th}network input concerning example p. The output of a generic neuron can then be written as (l = 2,…, L)
$${y}_{qlp}={\phi}_{l}({S}_{qlp}),$$

_{l}is the transfer function used for all neurons in layer l.#### Radial-Basis Function Network (RBFN)

Although having similar topologies, RBFN and MLPN behave very differently due to distinct hidden neuron models—unlike the MLPN, RBFN have hidden neurons behaving differently than output neurons. According to Xie et al. [47], RBFN (i) are specially recommended in functional approximation problems when the function surface exhibits regular peaks and valleys, and (ii) perform more robustly than MLPN when dealing with noisy input data. Although traditional RBFN have 3 layers, a generic multi-hidden layer (see Figure 4) RBFN is allowed in this work, being the generic hidden neuron’s model concerning node ‘l
whereas the RBFNs implemented through MATLAB neural net toolbox (involving learning algorithms 1–3 in Table 4) are based on the following definitions

_{1}l_{2}’ (l_{1}= 1,…,Q_{l2}, l_{2}= 2,…, L − 1) presented in Figure 6. In this model, (i) ${v}_{{l}_{1}{l}_{2}p}$ and ${\xi}_{{l}_{1}{l}_{2}}$ (called RBF center) are vectors of the same size (${\xi}_{z{l}_{1}{l}_{2}}$ denotes de z component of vector ${\xi}_{{l}_{1}{l}_{2}}$, and it is a network unknown), being the former associated to the presentation of data pattern p, (ii) ${\sigma}_{{l}_{1}{l}_{2}}$ is called RBF width (a positive scalar) and also belongs, along with synaptic weights and RBF centers, to the set of network unknowns to be determined through learning, (iii) ${\phi}_{{l}_{2}}$ is the user-defined radial basis (transfer) function (RBF), described in Equations (20)–(23), and (iv) ${y}_{{l}_{1}{l}_{2}p}$ is neuron’s output when pattern p is presented to the network. In ANNs not involving learning algorithms 1–3 in Table 4, vectors ${v}_{{l}_{1}{l}_{2}p}$ and ${\xi}_{{l}_{1}{l}_{2}}$ are defined as (two versions of ${v}_{{l}_{1}{l}_{2}p}$ where implemented and the one yielding the best results was selected)
$$\begin{array}{l}{v}_{{l}_{1}{l}_{2}p}=\left[\begin{array}{cccc}{y}_{1({l}_{2}-1)p}{w}_{1({l}_{2}-1){l}_{1}{l}_{2}}& \dots & \begin{array}{cc}{y}_{z({l}_{2}-1)p}{w}_{z({l}_{2}-1){l}_{1}{l}_{2}}& \dots \end{array}& {y}_{Q{l}_{2}-1({l}_{2}-1)p}{w}_{Q{l}_{2}-1({l}_{2}-1){l}_{1}{l}_{2}}\end{array}\right]\\ \mathrm{or}\\ {v}_{{l}_{1}{l}_{2}p}=\left[\begin{array}{cccc}{y}_{1({l}_{2}-1)p}& \dots & \begin{array}{cc}{y}_{z({l}_{2}-1)p}& \dots \end{array}& {y}_{Q{l}_{2}-1({l}_{2}-1)p}\end{array}\right]\\ \mathrm{and}\\ {\xi}_{{l}_{1}{l}_{2}}=\left[\begin{array}{cccc}{\xi}_{1{l}_{1}{l}_{2}}& \dots & \begin{array}{cc}{\xi}_{z{l}_{1}{l}_{2}}& \dots \end{array}& {\xi}_{Q{l}_{2}-1{l}_{1}{l}_{2}}\end{array}\right]\end{array}$$

$$\begin{array}{l}{v}_{{l}_{1}{l}_{2}p}=\left[\begin{array}{cccc}{y}_{1({l}_{2}-1)p}& \dots & \begin{array}{cc}{y}_{z({l}_{2}-1)p}& \dots \end{array}& {y}_{Q{l}_{2}-1({l}_{2}-1)p}\end{array}\right]\\ {\xi}_{{l}_{1}{l}_{2}}=\left[\begin{array}{cccc}{w}_{1({l}_{2}-1){l}_{1}{l}_{2}}& \dots & \begin{array}{cc}{w}_{z({l}_{2}-1){l}_{1}{l}_{2}}& \dots \end{array}& {w}_{Q{l}_{2}-1({l}_{2}-1){l}_{1}{l}_{2}}\end{array}\right]\end{array}$$

Lastly, according to the implementation carried out for initialization purposes (described in Section 3.3.12), (i) RBF center vectors per hidden layer (one per hidden neuron) are initialized as integrated in a matrix (termed RBF center matrix) having the same size of a weight matrix linking the previous layer to that specific hidden layer, and (ii) RBF widths (one per hidden neuron) are initialized as integrated in a vector (called RBF width vector) with the same size of a hypothetic bias vector.

#### 3.3.9. Hidden Nodes (Feature 9)

Inspired by several heuristics found in the literature for the determination of a suitable number of hidden neurons in a single hidden layer net ([48,49,50]), each value in hntest, defined in Equation (15), was tested in this work as the total number of hidden nodes in the model, ie the sum of nodes in all hidden layers (initially defined with the same number of neurons). The number yielding the smallest performance measure for all patterns (as defined in Section 3.4, with outputs and targets not postprocessed), is adopted as the best solution. The aforementioned hntest is defined by
where (i) Q

$$\begin{array}{l}incr=[4,4,4,10,10,10,10]\\ minimum=[1,1,1,10,10,10,10]\\ ma{x}_{1}=\mathrm{min}\left(\mathrm{round}\right(\mathrm{max}\left(2{Q}_{1}+{Q}_{L},4{Q}_{1},\sqrt{\frac{P}{{Q}_{1}\mathrm{ln}(P)}}\right)\left),1500\right)\\ ma{x}_{2}=\mathrm{max}(\mathrm{min}(\mathrm{round}(0.1P),1500),300)\\ maximum=[ma{x}_{1},ma{x}_{1},ma{x}_{1},ma{x}_{2},ma{x}_{2},ma{x}_{2},ma{x}_{2}]\\ hntest=minimum({\mathrm{F}}_{13}):incr({\mathrm{F}}_{13}):maximum({\mathrm{F}}_{13})\end{array}$$

_{1}and Q_{L}are the number of input and output nodes, respectively, (ii) P and P_{t}are the number of learning and training patterns, respectively, and (iii) F_{13}is the number of feature 13’s method (see Table 4).#### 3.3.10. Connectivity (Feature 10)

For this ANN feature, three methods were implemented, namely (i) adjacent layers—only connections between adjacent layers are made possible, (ii) adjacent layers + input-output—only connections between (ii

_{1}) adjacent and (ii_{2}) input and output layers are allowed, and (iii) fully-connected (all possible feedforward connections).#### 3.3.11. Hidden Transfer Functions (Feature 11)

Besides functions (i) Logistic—Equation (7), (ii) Hyperbolic Tangent—Equation (8), and (iii) Bilinear—Equation (9), defined in Section 3.3.6, the ones defined next were also implemented as hidden transfer functions. During software validation it was observed that some hidden node outputs could be infinite or NaN (not-a-number in MATLAB—e.g., 0/0 = Inf/Inf = NaN), due to numerical issues concerning some hidden transfer functions and/or their calculated input. In those cases, it was decided to convert infinite to unitary values and NaNs to zero (the only exception was the bipolar sigmoid function, where NaNs were converted to −1). Other implemented trick was to convert possible Gaussian function’s NaN inputs to zero.

#### Identity-Logistic

In [51], issues associated with flat spots at the extremes of a sigmoid function were eliminated by adding a linear function to the latter, reading

$$\phi (s)=\frac{1}{1+{e}^{-s}}+s.$$

#### Bipolar

The so-called bipolar sigmoid activation function mentioned in [52], ranging in [−1, 1], reads

$$\phi (s)=\frac{1-{e}^{-s}}{1+{e}^{-s}}.$$

#### Positive Saturating Linear

In MATLAB neural net toolbox, the so-called Positive Saturating Linear transfer function, ranging in [0, 1], is defined as

$$\phi (s)=\{\begin{array}{cc}1,& s\ge 1\\ s,& 0<s<1\\ 0,& s\le 0\end{array}.$$

#### Sinusoid

Concerning less popular transfer functions, reference is made in [53] to the sinusoid, which in this work was implemented as

$$\phi (s)=\mathrm{sin}\left(\frac{\pi}{2}s\right).$$

#### Radial Basis Functions (RBF)

Although Gaussian activation often exhibits desirable properties as a RBF, several authors (e.g., [54]) have suggested several alternatives. Following nomenclature used in Section 3.3.8, (i) the Thin-Plate Spline function is defined by
(ii) the next function is employed as Gaussian-type function when learning algorithms 4–7 are used (see Table 4)
(iii) the Multiquadratic function is given by
and (iv) the Gaussian-type function (called ‘radbas’ in MATLAB toolbox) used by RBFNs trained with learning algorithms 1–3 (see Table 4), is defined by
where || … || denotes the Euclidean distance in all functions.

$${\phi}_{{l}_{2}}\left(s\right)=s\mathrm{ln}\left(\sqrt{s}\right),\hspace{1em}s={\Vert {v}_{{l}_{1}{l}_{2}p}-{\xi}_{{l}_{1}{l}_{2}}\Vert}^{2},$$

$${\phi}_{{l}_{2}}\left(s\right)={e}^{-0.5s},\hspace{1em}s={\Vert {v}_{{l}_{1}{l}_{2}p}-{\xi}_{{l}_{1}{l}_{2}}\Vert}^{2}/{\sigma}_{{l}_{1}{l}_{2}}^{2},$$

$${\phi}_{{l}_{2}}\left(s\right)=\sqrt{s},\hspace{1em}s={\Vert {v}_{{l}_{1}{l}_{2}p}-{\xi}_{{l}_{1}{l}_{2}}\Vert}^{2}+{\sigma}_{{l}_{1}{l}_{2}}^{2},$$

$${\phi}_{{l}_{2}}\left(s\right)={e}^{-{s}^{2}},\hspace{1em}s=\Vert {v}_{{l}_{1}{l}_{2}p}-{\xi}_{{l}_{1}{l}_{2}}\Vert {\sigma}_{{l}_{1}{l}_{2}},$$

#### 3.3.12. Parameter Initialization (Feature 12)

The initialization of (i) weight matrices (Q

_{a}× Q_{b}, being Q_{a}and Q_{b}node numbers in layers a and b being connected, respectively), (ii) bias vectors (Q_{b}× 1), (iii) RBF center matrices (Q_{c-1}× Q_{c}, being c the hidden layer that matrix refers to), and (iv) RBF width vectors (Q_{c}× 1), are independent and in most cases randomly generated. For each ANN design carried out in the context of each parametric analysis combo, and whenever the parameter initialization method is not the ‘Mini-Batch SVD’, ten distinct simulations varying (due to their random nature) initialization values are carried out, in order to find the best solution. The implemented initialization methods are described next.#### Midpoint, Rands, Randnc, Randnr, Randsmall

These are all MATLAB built-in functions. Midpoint is used to initialize weight and RBF center matrices only (not vectors). All columns of the initialized matrix are equal, being each entry equal to the midpoint of the (training) output range leaving the corresponding initial layer node—recall that in weight matrices, columns represent each node in the final layer being connected, whereas rows represent each node in the initial layer counterpart. Rands generates random numbers with uniform distribution in [−1, 1]. Randnc (only used to initialize matrices) generates random numbers with uniform distribution in [−1, 1], and normalizes each array column to 1 (unitary Euclidean norm). Randnr (only used to initialize matrices) generates random numbers with uniform distribution in [−1, 1], and normalizes each array row to 1 (unitary Euclidean norm). Randsmall generates random numbers with uniform distribution in [−0.1, 0.1].

#### Rand [−lim, lim]

This function is based on the proposal in [55], and generates random numbers with uniform distribution in [−lim, lim], being lim layer-dependent and defined by
where a and b refer to the initial and final layers integrating the matrix being initialized, and L is the total number of layers in the network. In the case of a bias or RBF width vector, lim is always taken as 0.5.

$$lim=\{\begin{array}{ll}{{Q}_{b}}^{1/{Q}_{a}},& b<L\\ 0.5,& b=L\end{array},$$

#### SVD

Although Deng et al. [56] proposed this method for a 3-layer network, it was implemented in this work regardless the number of hidden layers.

#### Mini-Batch SVD

Based on [56], this scheme is an alternative version of the former SVD. Now, training data is split into min{Q

_{b}, P_{t}} chunks (or subsets) of equal size P_{ti}= max{floor(P_{t}/Q_{b}), 1}—floor rounds the argument to the previous integer (whenever it is decimal) or yields the argument itself, being each chunk aimed to derive Q_{bi}= 1 hidden node.#### 3.3.13. Learning Algorithm (Feature 13)

The most popular learning algorithm is called error back-propagation (BP), a first-order gradient method. Second-order gradient methods are known to have higher training speed and accuracy [57]. The most employed is called Levenberg-Marquardt (LM). All these traditional schemes were implemented using MATLAB toolbox [39].

#### Back-Propagation (BP, BPA), Levenberg-Marquardt (LM)

Two types of BP schemes were implemented, one with constant learning rate (BP)—’traingd’ in MATLAB, and another with iteration-dependent rate, named BP with adaptive learning rate (BPA)—‘traingda’ in MATLAB. The learning parameters set different than their default values are:

- (i)
- Learning Rate = 0.01/cs
^{0.5}, being cs the chunk size, as defined in Section 3.3.15. - (ii)
- Minimum performance gradient = 0.

Concerning the LM scheme—‘trainlm’ in MATLAB, the only learning parameter set different than its default value was the abovementioned (ii).

#### Extreme Learning Machine (ELM, mb ELM, I-ELM, CI-ELM)

Besides these traditional learning schemes, iterative and time-consuming by nature, four versions of a recent, powerful and non-iterative learning algorithm, called Extreme Learning Machine (ELM), were implemented (unlike initially proposed by the authors of ELM, connections across layers were allowed in this work), namely: (batch) ELM [58], Mini-Batch ELM (mb ELM) [59], Incremental ELM (I-ELM) [60], Convex Incremental ELM (CI-ELM) [61].

#### 3.3.14. Performance Improvement (Feature 14)

A simple and recursive approach aiming to improve ANN accuracy is called Neural Network Composite (NNC), as described in [62]. In this work, a maximum of 10 extra ANNs were added to the original one, until maximum error was not improved between successive NNC solutions. Later in this manuscript, a solution given by a single neural net might be denoted as ANN, whereas the other possible solution is called NNC.

#### 3.3.15. Training Mode (Feature 15)

Depending on the relative amount of training patterns, with respect to the whole training dataset, that is presented to the network in each iteration of the learning process, several types of training modes can be used, namely (i) batch or (ii) mini-batch. Whereas in the batch mode all training patterns are presented (called an epoch) to the network in each iteration, in the mini-batch counterpart the training dataset is split into several data chunks (or subsets) and in each iteration a single and new chunk is presented to the network, until (eventually) all chunks have been presented. Learning involving iterative schemes (e.g., BP- or LM-based) might require many epochs until an ‘optimum’ design is found. The particular case of having a mini-batch mode where all chunks are composed by a single (distinct) training pattern (number of data chunks = P

_{t}, chunk size = 1), is called online or sequential mode. Wilson and Martinez [63] suggested that if one wants to use mini-batch training with the same stability as online training, a rough estimate of the suitable learning rate to be used in learning algorithms such as the BP, is η_{online}/$\sqrt{cs}$, where cs is the chunk size and η_{online}is the online learning rate—their proposal was adopted in this work. Based on the proposal of Liang et al. [59], the constant chunk size (cs) adopted for all chunks in mini-batch mode reads cs = min{mean(hn) + 50, P_{t}}, being hn a vector storing the number of hidden nodes in each hidden layer in the beginning of training, and mean(hn) the average of all values in hn.#### 3.4. Network Performance Assessment

Several types of results were computed to assess network outputs, namely (i) maximum error, (ii) % errors greater than 3%, and (iii) performance, which are defined next. All above-mentioned errors are relative errors (expressed in %) based on the following definition, concerning a single output variable and data pattern,
where (i) d

$${e}_{qp}=100\left|\frac{{d}_{qp}-{y}_{qLp}}{{d}_{qp}}\right|,$$

_{qp}is the q^{th}desired (or target) output when pattern p within iteration i (p = 1,…, P_{i}) is presented to the network, and (ii) y_{qLp}is net’s q^{th}output for the same data pattern. Moreover, denominator in Equation (25) is replaced by 1 whenever |d_{qp}| < 0.05 − d_{qp}in the nominator keeps its real value. This exception to Equation (25) aims to reduce the apparent negative effect of large relative errors associated to target values close to zero. Even so, this trick may still lead to (relatively) large solution errors while groundbreaking results are depicted as regression plots (target versus predicted outputs).#### 3.4.1. Maximum Error

This variable measures the maximum relative error, as defined by Equation (25), among all output variables and learning patterns.

#### 3.4.2. Percentage of Errors > 3%

This variable measures the percentage of relative errors, as defined by Equation (25), among all output variables and learning patterns, that are greater than 3%.

#### 3.4.3. Performance

In functional approximation problems, network performance is defined as the average relative error, as defined in Equation (25), among all output variables and data patterns being evaluated (e.g., training, all data).

#### 3.5. Software Validation

Several benchmark datasets/functions were used to validate the developed software, involving low- to high-dimensional problems and small to large volumes of data. Due to paper length limit, validation results are not presented herein but they were made public in [64].

#### 3.6. Parametric Analysis Results

Aiming to reduce the computing time by cutting in the number of combos to be run—note that all features combined lead to hundreds of millions of combos, the whole parametric simulation was divided into nine parametric SAs, where in each one feature 7 only takes a single value. This measure aims to make the performance ranking of all combos within each ‘small’ analysis more ‘reliable’, since results used for comparison are based on target and output datasets as used in ANN training and yielded by the designed network, respectively (they are free of any postprocessing that eliminates output normalization effects on relative error values). Whereas (i) the 1st and 2nd SAs aimed to select the best methods from features 1, 2, 5, 8, and 13 (all combined), while adopting a single popular method for each of the remaining features (F

_{3}: 6, F_{4}: 2, F_{6}: {1 or 7}, F_{7}: 1, F_{9}: 1, F_{10}: 1, F_{11}: {3, 9 or 11}, F_{12}: 2, F_{14}: 1, F_{15}: 1—see Table 2, Table 3 and Table 4)—SA 1 involved learning algorithms 1–3 and SA 2 involved the ELM-based counterpart, (ii) the 3rd–7th SAs combined all possible methods from features 3, 4, 6, and 7, and concerning all other features, adopted the methods integrating the best combination from the aforementioned first SA, (iii) the 8th SA combined all possible methods from features 11, 12, and 14, and concerning all other features, adopted the methods integrating the best combination (results compared after postprocessing) among the previous five sub-analyses, and lastly (iv) the 9th SA combined all possible methods from features 9, 10 and 15, and concerning all other features, adopted the methods integrating the best combination from the previous analysis. Summing up the ANN feature combinations for all parametric SAs, a total of 475 combos were run for this work.ANN feature methods used in the best combo from each of the abovementioned nine parametric sub-analyses, are specified in Table 5 (the numbers represent the method number as in Table 2, Table 3 and Table 4). Table 6 shows the corresponding relevant results for those combos, namely (i) maximum error, (ii) % errors > 3%, (iii) performance (all described in Section 3, and evaluated for all learning data), (iv) total number of hidden nodes in the model, and (v) average computing time per example (including data pre- and post-processing). All results shown in Table 6 are based on target and output datasets computed in their original format, i.e. free of any transformations due to output normalization and/or dimensional analysis. The microprocessors used in this work have the following features: OS: Win10Home 64bits, RAMs: 48/128 GB, Local Disk Memory: 1 TB, CPUs: Intel® Core™ i7 8700K @ 3.70–4.70 GHz/i9 7960X @ 2.80-4.20 GHz.

#### 3.7. Proposed ANN-Based Model

The proposed model is the one, among the best ones from all parametric SAs, exhibiting the lowest maximum error (SA 9). That model is characterized by the ANN feature methods {1, 2, 1, 3, 1, 7, 5, 1, 3, 3, 1, 5, 3, 1, 3} in Table 2, Table 3 and Table 4. Aiming to allow implementation of this model by any user, all variables/equations required for (i) data preprocessing, (ii) ANN simulation, and (iii) data postprocessing, are presented in Section 3.7.1, Section 3.7.2 and Section 3.7.3, respectively. The proposed model is a single MLPN with 5 layers and a distribution of nodes/layer of 8-11-11-11-1. Concerning connectivity, the network is fully-connected, and the hidden and output transfer functions are all Logistic (Equation (7)) and Identity (Equation (10)), respectively. The network was trained using the LM algorithm (1500 epochs). After design, the average network computing time concerning the presentation of a single example (including data pre/postprocessing) is 6.93 × 10

^{−}^{5}s—Figure 7 depicts a simplified scheme of some of network key features. Lastly, all relevant performance results concerning the proposed ANN are illustrated in Section 3.7.4. The obtained ANN solution for every data point can be found in [31].It is worth recalling that, in this manuscript, whenever a vector is added to a matrix, it means the former is to be added to all columns of the latter (valid in MATLAB).

#### 3.7.1. Input Data Preprocessing

For future use of the proposed ANN to simulate new data Y

_{1,sim}(8 × P_{sim}matrix) concerning P_{sim}patterns, the same data preprocessing (if any) performed before training must be applied to the input dataset. That preprocessing is defined by the methods used for ANN features 2, 3 and 5 (respectively 2, 6 and 1—see Table 2), which should be applied after all (eventual) qualitative variables in the input dataset are converted to numerical (using feature 1’s method). Next, the necessary preprocessing to be applied to Y_{1,sim}, concerning features 2, 3 and 5, is fully described.#### Dimensional Analysis and Dimensionality Reduction

Since dimensional analysis (d.a.) and dimensionality reduction (d.r.) were not carried out, one has

$${\left\{{Y}_{1,sim}\right\}}_{d.r.}^{after}={\left\{{Y}_{1,sim}\right\}}_{d.a.}^{after}={Y}_{1,sim},$$

#### Input Normalization

After input normalization, the new input dataset ${\left\{{Y}_{1,sim}\right\}}_{n}^{after}$ is defined as function of the previously determined ${\left\{{Y}_{1,sim}\right\}}_{d.r}^{after}$, and they have the same size, reading
where one recalls that operator ‘.x’ multiplies component i in vector INP by all components in row i of ${\left\{{Y}_{1,sim}\right\}}_{d.r}^{after}$.

$$\begin{array}{l}{\left\{{Y}_{1,sim}\right\}}_{n}^{after}=\mathrm{INP}.\mathrm{x}{\left\{{Y}_{1,sim}\right\}}_{d.r}^{after}\\ \mathrm{INP}=\left[\begin{array}{l}0.125\\ 0.00139275766016713\\ 0.00178571428571429\\ 0.00364431486880467\\ 0.00142857142857143\\ 0.0666666666666667\\ 0.0037037037037037\\ 0.04\end{array}\right]\end{array}$$

#### 3.7.2. ANN-Based Analytical Model

Once determined the preprocessed input dataset {Y
where
Arrays W

_{1,sim}}_{n}^{after}(8 × P_{sim}matrix), the next step is to present it to the proposed ANN to obtain the predicted output dataset {Y_{5,sim}}_{n}^{after}(1 × P_{sim}vector), which will be given in the same preprocessed format of the target dataset used in learning. In order to convert the predicted outputs to their ‘original format’ (i.e., without any transformation due to normalization or dimensional analysis—the only transformation visible will be the (eventual) qualitative variables written in their numeric representation), some postprocessing is needed, as described in detail in Section 3.7.3. Next, the mathematical representation of the proposed ANN is given, so that any user can implement it to determine {Y_{5,sim}}_{n}^{after}, thus eliminating all rumors that ANNs are ‘black boxes’.
$$\begin{array}{l}{Y}_{2}={\phi}_{2}\left({W}_{1-2}^{T}{\left\{{Y}_{1,sim}\right\}}_{n}^{after}+{b}_{2}\right)\\ {Y}_{3}={\phi}_{3}\left({W}_{1-3}^{T}{\left\{{Y}_{1,sim}\right\}}_{n}^{after}+{W}_{2-3}^{T}{Y}_{2}+{b}_{3}\right)\\ {Y}_{4}={\phi}_{4}\left({W}_{1-4}^{T}{\left\{{Y}_{1,sim}\right\}}_{n}^{after}+{W}_{2-4}^{T}{Y}_{2}+{W}_{3-4}^{T}{Y}_{3}+{b}_{4}\right)\\ {\left\{{Y}_{5,sim}\right\}}_{n}^{after}={\phi}_{5}\left({W}_{1-5}^{T}{\left\{{Y}_{1,sim}\right\}}_{n}^{after}+{W}_{2-5}^{T}{Y}_{2}+{W}_{3-5}^{T}{Y}_{3}+{W}_{4-5}^{T}{Y}_{4}+{b}_{5}\right)\end{array}$$

$$\begin{array}{l}{\phi}_{2}={\phi}_{3}={\phi}_{4}=\phi (s)=\frac{1}{1+{e}^{-s}}\\ {\phi}_{5}={\phi}_{5}(s)=s\end{array}$$

_{j-s}and b_{s}are stored online in [65], aiming to avoid an overlong article and ease model’s implementation by any interested reader.#### 3.7.3. Output Data Postprocessing

In order to transform the output dataset obtained by the proposed ANN, {Y

_{5,sim}}_{n}^{after}(1 × P_{sim}vector), to its original format (Y_{5,sim}), i.e. without the effects of dimensional analysis and/or output normalization (possibly) taken in target dataset preprocessing prior training, the postprocessing addressed next must be performed.Once obtained {Y
since no output normalization nor dimensional analysis were carried out.

_{5,sim}}_{n}^{after}, the following relations hold for its relation to its non-normalized (${\left\{{Y}_{5,sim}\right\}}_{d.a.}^{after}$) and original (Y_{5,sim}) formats (just after the dimensional analysis stage, and free of any pre-processing effects, respectively), reading
$${Y}_{5,sim}={\left\{{Y}_{5,sim}\right\}}_{d.a.}^{after}={\left\{{Y}_{5,sim}\right\}}_{n}^{after},$$

#### 3.7.4. Performance Results

Finally, results yielded by the proposed ANN, in terms of performance variables defined in Section 3.4, are presented in this section in the form of several graphs: (i) a regression plot per output variable (Figure 8), where network target and output data are plotted, for each data point, as x- and y- coordinates, respectively—a measure of quality is given by the Pearson Correlation Coefficient (R), as defined in Equation (1); (ii) a performance plot (Figure 9), where performance (average error) values are displayed for several learning datasets; and (iii) an error plot (Figure 10), where values concern all data (iii

_{1}) maximum error and (iii_{2}) % of errors greater than 3%. It is worth highlighting that all graphical results just mentioned are based on effective target and output values, i.e. computed in their original format.## 4. Design Considerations

The bar chart in Figure 11, where beams numbers in the horizontal axis are referenced to the data point ID in the ANN dataset [31], presents (i) the design loads given by SCI-P355 [15], and (ii) the FEA-based elastic buckling loads obtained from this work. As expected, SCI-P355 yielded significantly lower loads and it is worth noting that the differences between the two approaches do not produce the same percentage variance in load estimates for the 8 randomly selected beams. In particular, for slender web-posts (i.e., closely spaced web openings) such as in beams 82, 136, 163, and 217, buckling of the web-post will always govern the design in SCI-P355, hence the reduced design load. As for the FEAs, parameter l

_{ed}(end web-post distance to the support—see Figure 1a) governs the design for widely spaced web openings. SCI-P355 does not consider the distance l_{ed}and recommends it to be greater than 50% of the opening diameter. In this work, the opening diameter and web-post width were taken within the recommended design limits (1.25 < H/Φ < 1.75 and 1.08 < (b_{wp}+ Φ)/Φ < 1.5, respectively).## 5. Concluding Remarks

An ANN-based analytical model is proposed to effectively predict the elastic buckling load of simply supported cellular steel beams subjected to a uniformly distributed load. Finite element solutions from 3645 distinct beams were used for ANN design (training, validation, and testing). The independent variables adopted as ANN inputs are the following: beam’s length, opening diameter, web-post width, cross-section height, web thickness, flange width, flange thickness, and the distance between the last opening edge and the end support. The maximum and average relative errors yielded by the proposed ANN among the 3645 data points were 3.7% and 0.4%, respectively. Moreover, that model is able to compute the buckling load of a single beam in less than a millisecond, for any current personal computer. These facts make the proposed model a potential tool for structural engineers and researchers who aim to accurately estimate the elastic buckling load of cellular steel beams (i) within the ranges of the input variables adopted in this study (see Table 1), and (ii) without the burden of the costly resources associated to FEA.

This research is the first step of an ongoing investigation that aims to propose a novel and simple analytical design method to accurately compute the inelastic resistance of cellular steel beams.

## Author Contributions

Conceptualization, M.A.; Methodology, M.A.; Software, M.A.; Data Curation, K.R., K.D.T.; Writing-Original Draft Preparation, M.A., K.R., K.D.T., T.P.R.; Writing-Review & Editing, M.A., K.R., K.D.T., T.P.R.

## Funding

This research received no external funding.

## Acknowledgments

The authors would like to thank the UK Research Council (EPSRC: EP/L504993/1) and the University of Leeds for their financial support.

## Conflicts of Interest

The authors declare no conflict of interest.

## References

- Tsavdaridis, K.D. Structural Performance of Perforated Steel Beams with Novel Web Openings and with Partial Concrete Encasement. Ph.D. Thesis, City University of London, London, UK, 2010. [Google Scholar]
- Morkhade, S.G.; Gupta, L.M. An experimental and parametric study on steel beams with web openings. Int. J. Adv. Struct. Eng.
**2015**, 7, 249–260. [Google Scholar] [CrossRef][Green Version] - Akrami, V.; Erfani, S. Review and assessment of design methodologies for perforated steel beams. J. Struct. Eng.
**2016**, 142, 1–14. [Google Scholar] [CrossRef] - Uenoya, M.; Redwood, R.G. Buckling of webs with openings. Comput. Struct.
**1978**, 9, 191–199. [Google Scholar] [CrossRef] - Lucas, W.K.; Darwin, D. Steel and Composite Beams with Web Openings; The American Iron and Steel Institute: City of Wichita, KS, USA, 1990. [Google Scholar]
- Darwin, D. Steel and Composite Beams with Web Opening; Steel Design Guide Series 2; American Institute of Steel Construction (AISC): Chicago, IL, USA, 1990. [Google Scholar]
- SEI/ASCE. Specifications for Structural Steel Beams with Openings; SEI/ASCE 23-97; ASCE: Reston, VA, USA, 1998. [Google Scholar]
- Ward, J.K. Design of Composite and Non-Composite Cellular Beams SCI P100; Steel Construction Institute: Berkshire, UK, 1990. [Google Scholar]
- Chung, K.F.; Lui, T.C.H.; Ko, A.C.H. Investigation on vierendeel mechanism in steel beams with circular web openings. J. Constr. Steel Res.
**2001**, 57, 467–490. [Google Scholar] [CrossRef] - Chung, K.F.; Liu, C.H.; Ko, A.C.H. Steel beams with large web openings of various shapes and sizes: An empirical design method using a generalized moment-shear interaction curve. J. Constr. Steel Res.
**2003**, 59, 1177–1200. [Google Scholar] [CrossRef] - Tsavdaridis, K.D.; D’Mello, C. Finite Element Investigation of Perforated Beams with Different Web Opening Configurations. In Proceedings of the 6th International Conference on Advances is Steel Structures (ICASS 2009), Hong Kong, China, 16–18 December 2009; pp. 213–220. [Google Scholar]
- Tsavdaridis, K.D.; D’Mello, C. Web buckling study of the behaviour and strength of perforated steel beams with different novel web opening shapes. J. Constr. Steel Res.
**2011**, 67, 1605–1620. [Google Scholar] [CrossRef][Green Version] - Tsavdaridis, K.D.; D’Mello, C. Vierendeel bending study of perforated steel beams with various novel web opening shapes through non-linear finite element analyses. J. Struct. Eng.
**2012**, 138, 1214–1230. [Google Scholar] [CrossRef] - Tsavdaridis, K.D.; Kingman, J.J.; Toropov, V.V. Application of Structural Topology Optimisation to Perforated Steel Beams. Comput. Struct.
**2015**, 158, 108–123. [Google Scholar] [CrossRef] - Lawson, R.M.; Hicks, S.J. Design of Composite Beams with Large Openings SCI P355; Steel Construction Institute: Berkshire, UK, 2011. [Google Scholar]
- Lawson, R.M. Design for Openings in the Webs of Composite Beams SCI P068; Steel Construction Institute: Berkshire, UK, 1987. [Google Scholar]
- Verweij, J.G. Cellular Beam-Columns in Portal Frame Structures. Master’s Thesis, Delft University of Technology, Delft, The Netherlands, 2010. [Google Scholar]
- Gholizadeh, S.; Pirmoz, A.; Attarnejad, R. Assessment of load carrying capacity of castellated steel beams by neural networks. J. Constr. Steel Res.
**2011**, 67, 770–779. [Google Scholar] [CrossRef] - Sharifi, Y.; Tohidi, S. Lateral-torsional buckling capacity assessment of web opening steel girders by artificial neural networks—Elastic investigation. Front. Struct. Civ. Eng.
**2014**, 8, 167–177. [Google Scholar] [CrossRef] - Tohidi, S.; Sharifi, Y. Inelastic lateral-torsional buckling capacity of corroded web opening steel beams using artificial neural networks. IES J. Part A Civ. Struct. Eng.
**2014**, 8, 24–40. [Google Scholar] [CrossRef] - Tohidi, S.; Sharifi, Y. Load-carrying capacity of locally corroded steel plate girder ends using artificial neural network. Thin-Walled Struct.
**2016**, 100, 48–61. [Google Scholar] [CrossRef] - Asteris, P.G.; Kolovos, K.G.; Douvika, M.G.; Roinos, K. Prediction of self-compacting concrete strength using artificial neural networks. Eur. J. Environ. Civ. Eng.
**2016**, 20 (Suppl. 1), S102–S122. [Google Scholar] [CrossRef] - Cascardi, A.; Micelli, F.; Aiello, M.A. An Artificial Neural Networks model for the prediction of the compressive strength of FRP-confined concrete circular columns. Eng. Struct.
**2017**, 140, 199–208. [Google Scholar] [CrossRef] - Marović, I.; Androjić, I.; Jajac, N.; Hanák, T. Urban Road Infrastructure Maintenance Planning with Application of Neural Networks. Complexity
**2018**, 2018, 1–10. [Google Scholar] [CrossRef] - Wong, E.W.C.; Choi, H.S.; Kim, D.K.; Hashim, F.M. Development of ANN Model for the Prediction of VIV Fatigue Damage of Top-tensioned Riser. MATEC Web Conf.
**2018**, 203, 1013. [Google Scholar] [CrossRef] - Dassault Systèmes. ABAQUS 6.11, Abaqus/CAE User’s Manual; Dassault Systemes: Vélizy-Villacoublay, France, 2011. [Google Scholar]
- Dassault Systèmes Simulia Corp. ABAQUS CAE (2017); Software; Dassault Systèmes Simulia Corp.: Vélizy-Villacoublay, France, 2017. [Google Scholar]
- Surtees, J.O.; Lui, Z. Report of Loading Tests on Cellform Beams; Research Report; University of Leeds: Leeds, UK, 1995. [Google Scholar]
- Rajana, K. Advanced Computational Parametric Study of the Linear Elastic and Non-Linear Post Buckling Behaviour of Non-Composite Cellular Steel Beams. Master’s Thesis, University of Leeds, Leeds, UK, 2018. [Google Scholar] [CrossRef]
- El-Sawhy, K.L.; Sweedan, A.M.I.; Martini, M.I. Moment gradient factor of cellular steel beams under inelastic flexure. J. Constr. Steel Res.
**2014**, 98, 20–34. [Google Scholar] [CrossRef] - Developer. Dataset ANN [Data Set]. Zenodo. 2018. Available online: http://doi.org/10.5281/zenodo.1486181 (accessed on 29 November 2018).
- Hertzmann, A.; Fleet, D. Machine Learning and Data Mining; Lecture Notes CSC 411/D11; Computer Science Department, University of Toronto: Toronto, ON, Canada, 2012. [Google Scholar]
- McCulloch, W.S.; Pitts, W. A logical calculus of the ideas immanent in nervous activity. Bull. Math. Biophys.
**1943**, 5, 115–133. [Google Scholar] [CrossRef] - Hern, A. Google Says Machine Learning Is the Future. So I Tried It Myself. Available online: www.theguardian.com/technology/2016/jun/28/all (accessed on 2 November 2016).
- Prieto, A.; Prieto, B.; Ortigosa, E.M.; Ros, E.; Pelayo, F.; Ortega, J.; Rojas, I. Neural networks: An overview of early research, current frameworks and new challenges. Neurocomputing
**2016**, 214, 242–268. [Google Scholar] [CrossRef] - Wilamowski, B.M.; Irwin, J.D. The Industrial Electronics Handbook: Intelligent Systems; CRC Press: Boca Raton, FL, USA, 2011. [Google Scholar]
- Flood, I. Towards the next generation of artificial neural networks for civil engineering. Adv. Eng. Inform.
**2008**, 228, 4–14. [Google Scholar] [CrossRef] - Haykin, S.S. Neural Networks and Learning Machines; Prentice Hall/Pearson: New York, NY, USA, 2009. [Google Scholar]
- The Mathworks, Inc. MATLAB R2017a, User’s Guide; The Mathworks, Inc.: Natick, MA, USA, 2017. [Google Scholar]
- Bhaskar, R.; Nigam, A. Qualitative physics using dimensional analysis. Artif. Intell.
**1990**, 45, 111–173. [Google Scholar] [CrossRef] - Kasun, L.L.C.; Yang, Y.; Huang, G.-B.; Zhang, Z. Dimension reduction with extreme learning machine. IEEE Trans. Image Process.
**2016**, 25, 3906–3918. [Google Scholar] [CrossRef] [PubMed] - Lachtermacher, G.; Fuller, J.D. Backpropagation in time-series forecasting. J. Forecast.
**1995**, 14, 381–393. [Google Scholar] [CrossRef] - Pu, Y.; Mesbahi, E. Application of artificial neural networks to evaluation of ultimate strength of steel panels. Eng. Struct.
**2006**, 28, 1190–1196. [Google Scholar] [CrossRef] - Flood, I.; Kartam, N. Neural Networks in Civil Engineering: I-Principals and Understanding. J. Comput. Civ. Eng.
**1994**, 8, 131–148. [Google Scholar] [CrossRef] - Mukherjee, A.; Deshpande, J.M.; Anmala, J. Prediction of buckling load of columns using artificial neural networks. J. Struct. Eng.
**1996**, 122, 1385–1387. [Google Scholar] [CrossRef] - Wilamowski, B.M. Neural Network Architectures and Learning algorithms. IEEE Ind. Electron. Mag.
**2009**, 3, 56–63. [Google Scholar] [CrossRef] - Xie, T.; Yu, H.; Wilamowski, B. Comparison between traditional neural networks and radial basis function networks. In Proceedings of the 2011 IEEE International Symposium on Industrial Electronics (ISIE), Gdansk University of Technology Gdansk, Gdansk, Poland, 27–30 June 2011; pp. 1194–1199. [Google Scholar]
- Aymerich, F.; Serra, M. Prediction of fatigue strength of composite laminates by means of neural networks. Key Eng. Mater.
**1998**, 144, 231–240. [Google Scholar] [CrossRef] - Rafiq, M.; Bugmann, G.; Easterbrook, D. Neural network design for engineering applications. Comput. Struct.
**2001**, 79, 1541–1552. [Google Scholar] [CrossRef] - Xu, S.; Chen, L. Novel approach for determining the optimal number of hidden layer neurons for FNN’s and its application in data mining. In Proceedings of the International Conference on Information Technology and Applications (ICITA), Cairns, Australia, 23–26 June 2008; pp. 683–686. [Google Scholar]
- Gunaratnam, D.J.; Gero, J.S. Effect of representation on the performance of neural networks in structural engineering applications. Comput.-Aided Civ. Infrastruct. Eng.
**1994**, 9, 97–108. [Google Scholar] [CrossRef] - Lefik, M.; Schrefler, B.A. Artificial neural network as an incremental non-linear constitutive model for a finite element code. Comput. Methods Appl. Mech. Eng.
**2003**, 192, 3265–3283. [Google Scholar] [CrossRef] - Bai, Z.; Huang, G.; Wang, D.; Wang, H.; Westover, M. Sparse extreme learning machine for classification. IEEE Trans. Cybern.
**2014**, 44, 1858–1870. [Google Scholar] [CrossRef] [PubMed] - Schwenker, F.; Kestler, H.; Palm, G. Three learning phases for radial-basis-function networks. Neural Netw.
**2001**, 14, 439–458. [Google Scholar] [CrossRef] - Waszczyszyn, Z. Neural Networks in the Analysis and Design of Structures; CISM Courses and Lectures No. 404; Springer: Wien, Austria; New York, NY, USA, 1999. [Google Scholar]
- Deng, W.-Y.; Bai, Z.; Huang, G.-B.; Zheng, Q.-H. A fast SVD-Hidden-nodes based extreme learning machine for large-scale data Analytics. Neural Netw.
**2016**, 77, 14–28. [Google Scholar] [CrossRef] - Wilamowski, B.M. How to not get frustrated with neural networks. In Proceedings of the 2011 IEEE International Conference on Industrial Technology (ICIT), Auburn University, Auburn, AL, USA, 14–16 March 2011. [Google Scholar]
- Huang, G.-B.; Zhu, Q.-Y.; Siew, C.-K. Extreme learning machine: Theory and applications. Neurocomputing
**2006**, 70, 489–501. [Google Scholar] [CrossRef][Green Version] - Liang, N.; Huang, G.; Saratchandran, P.; Sundararajan, N. A fast and accurate online Sequential learning algorithm for Feedforward networks. IEEE Trans. Neural Netw.
**2006**, 17, 1411–1423. [Google Scholar] [CrossRef] - Huang, G.; Chen, L.; Siew, C. Universal approximation using incremental constructive feedforward networks with random hidden nodes. IEEE Trans. Neural Netw.
**2006**, 17, 879–892. [Google Scholar] [CrossRef] - Huang, G.-B.; Chen, L. Convex incremental extreme learning machine. Neurocomputing
**2007**, 70, 3056–3062. [Google Scholar] [CrossRef] - Beyer, W.; Liebscher, M.; Beer, M.; Graf, W. Neural Network Based Response Surface Methods—A Comparative Study. In Proceedings of the 5th German LS-DYNA Forum, Ulm, Germany, 12–13 October 2006; pp. 29–38. [Google Scholar]
- Wilson, D.R.; Martinez, T.R. The general inefficiency of batch training for gradient descent learning. Neural Netw.
**2003**, 16, 1429–1451. [Google Scholar] [CrossRef][Green Version] - The Researcher. ANNSoftwareValidation-Report.pdf. 2018. Available online: https://www.researchgate.net/profile/Abambres_M/project/Applied-Artificial-Intelligence/attachment/5aff6a82b53d2f63c3ccbaa0/AS:[email protected]/download/ANN+Software+Validation+-+Report.pdf?context=ProjectUpdatesLog (accessed on 29 November 2018).
- Developer. W and b Arrays [Data Set]. Zenodo. 2018. Available online: http://doi.org/10.5281/zenodo.1486268 (accessed on 29 November 2018).

**Figure 1.**Steel cellular members: (

**a**) input (parametric) variables, and (

**b**) application in an office building floor system in Lisbon, Portugal.

**Figure 6.**Generic hidden neuron l

_{1}l

_{2}placed anywhere in the RBFN of Figure 4 (l

_{2}= 2,…, L − 1).

Inputs Variables—Figure 1a | ANN Node No. | Possible Values | |||||
---|---|---|---|---|---|---|---|

Beam’s length | L (m) | 1 | 4 | 5 | 6 | 7 | 8 |

Opening-support end distance | l_{ed} (mm) | 2 | 135 values in [12, 718] | ||||

Opening diameter | Φ (mm) | 3 | H/1.25 | H/1.5 | H/1.7 | - | - |

Web-post width | b_{wp} (mm) | 4 | Φ/10 | Φ/3.45 | Φ/2.04 | - | - |

Section height | H (mm) | 5 | 700 | 560 | 420 | - | - |

Web thickness | t_{w} (mm) | 6 | 15 | 12 | 9 | - | - |

Flange width | b_{f} (mm) | 7 | 270 | 216 | 162 | - | - |

Flange thickness | t_{f} (mm) | 8 | 25 | 20 | 15 | - | - |

Target/Output Variable | |||||||

Elastic Buckling Load | λ_{b} (kN/m) |

FEATURE METHOD | F1 | F2 | F3 | F4 | F5 |
---|---|---|---|---|---|

Qualitative Var Representation | Dimensional Analysis | Input Dimensionality Reduction | % Train-Valid-Test | Input Normalization | |

1 | Boolean vectors | Yes | Linear correlation | 80-10-10 | Linear max abs |

2 | Eq spaced in [0, 1] | No | Auto-encoder | 70-15-15 | Linear [0, 1] |

3 | - | - | - | 60-20-20 | Linear [−1, 1] |

4 | - | - | Ortho rand proj. | 50-25-25 | Nonlinear |

5 | - | - | Sparse rand proj. | - | Lin mean std |

6 | - | - | No | - | No |

FEATURE METHOD | F6 | F7 | F8 | F9 | F10 |
---|---|---|---|---|---|

Output Transfer | Output Normalization | Net Architecture | Hidden Layers | Connectivity | |

1 | Logistic | Lin [a, b] = 0.7 [φ_{min}, φ_{max}] | MLPN | 1 HL | Adjacent layers |

2 | - | Lin [a, b] = 0.6 [φ_{min}, φ_{max}] | RBFN | 2 HL | Adj layers + in-out |

3 | Hyperbolic tang | Lin [a, b] = 0.5 [φ_{min}, φ_{max}] | - | 3 HL | Fully-connected |

4 | - | Linear mean std | - | - | - |

5 | Bilinear | No | - | - | - |

6 | Compet | - | - | - | - |

7 | Identity | - | - | - | - |

FEATURE METHOD | F11 | F12 | F13 | F14 | F15 |
---|---|---|---|---|---|

Hidden Transfer | Parameter Initialization | Learning Algorithm | Performance Improvement | Training Mode | |

1 | Logistic | Midpoint (W) + Rands (b) | BP | NNC | Batch |

2 | Identity-logistic | Rands | BPA | - | Mini-Batch |

3 | Hyperbolic tang | Randnc (W) + Rands (b) | LM | - | Online |

4 | Bipolar | Randnr (W) + Rands (b) | ELM | - | - |

5 | Bilinear | Randsmall | mb ELM | - | - |

6 | Positive sat linear | Rand [−Δ, Δ] | I-ELM | - | - |

7 | Sinusoid | SVD | CI-ELM | - | - |

8 | Thin-plate spline | MB SVD | - | - | - |

9 | Gaussian | - | - | - | - |

10 | Multiquadratic | - | - | - | - |

11 | Radbas | - | - | - | - |

SA | F1 | F2 | F3 | F4 | F5 | F6 | F7 | F8 | F9 | F10 | F11 | F12 | F13 | F14 | F15 |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|

1 | 1 | 2 | 6 | 2 | 1 | 1 | 1 | 1 | 1 | 1 | 3 | 2 | 3 | 1 | 3 |

2 | 1 | 2 | 6 | 2 | 1 | 7 | 1 | 1 | 1 | 1 | 3 | 2 | 5 | 1 | 3 |

3 | 1 | 2 | 6 | 2 | 1 | 1 | 1 | 1 | 1 | 1 | 3 | 2 | 3 | 1 | 3 |

4 | 1 | 2 | 6 | 2 | 1 | 1 | 2 | 1 | 1 | 1 | 3 | 2 | 3 | 1 | 3 |

5 | 1 | 2 | 6 | 3 | 1 | 1 | 3 | 1 | 1 | 1 | 3 | 2 | 3 | 1 | 3 |

6 | 1 | 2 | 6 | 2 | 1 | 7 | 4 | 1 | 1 | 1 | 3 | 2 | 3 | 1 | 3 |

7 | 1 | 2 | 6 | 3 | 1 | 7 | 5 | 1 | 1 | 1 | 3 | 2 | 3 | 1 | 3 |

8 | 1 | 2 | 6 | 3 | 1 | 7 | 5 | 1 | 1 | 1 | 1 | 5 | 3 | 1 | 3 |

9 | 1 | 2 | 6 | 3 | 1 | 7 | 5 | 1 | 3 | 3 | 1 | 5 | 3 | 1 | 3 |

**Table 6.**Performance results for the best design from each parametric sub-analysis: (a) ANN, (b) NNC.

SA | ANN | ||||

Max Error (%) | Performance All Data (%) | Errors > 3% (%) | Total Hidden Nodes | Running Time/Data Point (s) | |

1 | 7.5 | 0.7 | 1.8 | 32 | 7.33 × 10^{−5} |

2 | 201.0 | 5.7 | 54.9 | 365 | 7.06 × 10^{−5} |

3 | 8.2 | 0.7 | 1.2 | 32 | 8.49 × 10^{−5} |

4 | 8.5 | 0.7 | 1.3 | 32 | 7.67 × 10^{−5} |

5 | 6.8 | 0.7 | 1.8 | 32 | 6.86 × 10^{−5} |

6 | 7.2 | 0.7 | 1.6 | 32 | 7.57 × 10^{−5} |

7 | 28.9 | 1.5 | 13.2 | 29 | 6.82 × 10^{−5} |

8 | 8.2 | 0.9 | 3.4 | 29 | 6.70 × 10^{−5} |

9 | 3.7 | 0.4 | 0.1 | 33 | 6.93 × 10^{−5} |

SA | NNC | ||||

Max Error (%) | Performance All Data (%) | Errors > 3% (%) | Total Hidden Nodes | Running Time/Data Point (s) | |

1 | - | - | - | - | - |

2 | 180.0 | 5.2 | 51.4 | 365 | 8.16 × 10^{−5} |

3 | - | - | - | - | - |

4 | 8.6 | 0.6 | 0.9 | 32 | 7.83 × 10^{−5} |

5 | - | - | - | - | - |

6 | 2.6 | 0.3 | 0.0 | 32 | 8.34 × 10^{−5} |

7 | 9.5 | 0.9 | 3.9 | 29 | 6.93 × 10^{−5} |

8 | 7.0 | 0.7 | 1.9 | 29 | 6.80 × 10^{−5} |

9 | - | - | - | - | - |

© 2018 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).