Next Article in Journal
Parametric Blending with Geodesic Curves on Triangular Meshes
Previous Article in Journal
Efficient 3D Model Simplification Algorithms Based on OpenMP
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Physics-Informed Feature Engineering and R2-Based Signal-to-Noise Ratio Feature Selection to Predict Concrete Shear Strength

by
Trevor J. Bihl
1,*,
William A. Young II
2 and
Adam Moyer
2
1
School of Electrical Engineering and Computer Science, Ohio University, Athens, OH 45701, USA
2
College of Business, Ohio University, Athens, OH 45701, USA
*
Author to whom correspondence should be addressed.
Mathematics 2025, 13(19), 3182; https://doi.org/10.3390/math13193182
Submission received: 23 August 2025 / Revised: 13 September 2025 / Accepted: 30 September 2025 / Published: 4 October 2025

Abstract

Accurate prediction of reinforced concrete shear strength is essential for structural safety, yet datasets often contain a mix of raw geometric and material properties alongside physics-informed engineered features, making optimal feature selection challenging. This study introduces a statistically principled framework that advances feature reduction for neural networks in three novel ways. First, it extends the artificial neural network-based signal-to-noise ratio (ANN-SNR) method, previously limited to classification, into regression tasks for the first time. Second, it couples ANN-SNR with a confidence-interval (CI)-based stopping rule, using the lower bound of the baseline ANN’s R2 confidence interval as a rigorous statistical threshold for determining when feature elimination should cease. Third, it systematically evaluates both raw experimental variables and physics-informed engineered features, showing how their combination enhances both robustness and interpretability. Applied to experimental concrete shear strength data, the framework revealed that many low-SNR features in conventional formulations contribute little to predictive performance and can be safely removed. In contrast, hybrid models that combined key raw and engineered features consistently yielded the strongest performance. Overall, the proposed method reduced the input feature set by approximately 45% while maintaining results statistically indistinguishable from baseline and fully optimized models (R2 ≈ 0.85). These findings demonstrate that ANN-SNR with CI-based stopping provides a defensible and interpretable pathway for reducing model complexity in reinforced concrete shear strength prediction, offering practical benefits for design efficiency without compromising reliability.

1. Introduction

Overly simplified models, limited generalizability, and uncertainty in design predictions are common consequences of data scarcity in civil engineering. Data scarcity exists due to the realities of civil engineering experimentation, which is both costly and time-consuming. Thus, traditional design equations are based on sparse datasets and rely heavily on conservative assumptions. While these models provide safe lower-bound estimates, they often fail to capture complex nonlinear interactions among geometric and material properties, potentially leading to inefficient or unconservative designs.
Mathematical models are central to structural engineering [1] and provide analytical frameworks to estimate performance characteristics. However, traditional models are often built on idealized assumptions and calibrated using limited datasets, which constrains their ability to generalize across a broad range of conditions. In many domains, this has led to a shift toward data-centric modeling, where models are driven by observed data rather than fixed assumptions. For this, this research explores a compiled database of 1200+ observations across 49 studies on concrete shear strength [2].
Due to their ability to model any measurable underlying function to an arbitrary degree of accuracy, artificial neural networks (ANNs) are of interest herein. Appropriately constructed, ANNs are universal approximators [3] however, their nonlinear and black-box nature has made their adoption slow in safety-critical applications, such as civil engineering [4] To address these competing limitations, black-box interpretability versus feature transparency, this study advances the mathematical modeling of reinforced concrete shear strength by integrating ANNs for modeling, feature engineering, feature selection, and hyperparameter optimization.
To address black-box concerns, this research evaluates both raw and physics-informed engineered features, assessing their combined influence on reinforced concrete shear strength prediction. Model configuration is systematically optimized through Bayesian hyperparameter search, ensuring robust selection of network architecture and learning parameters. For feature selection, we extend signal-to-noise ratio (SNR) screening, previously confined to classification [5] to regression tasks, and develop a novel R2 confidence-interval-based stopping rule that defines statistical equivalence to a full model. Applied to reinforced concrete data, this integrated framework enables systematic reduction in features while maintaining predictive performance.
Taken together, this paper contributes to the integration of feature engineering, mathematical learning models, optimization methods, and statistically principled feature selection to advance a long-standing, data-scarce problem in structural engineering. By balancing predictive accuracy with model parsimony, the ANN framework achieves accuracy surpassing traditional code-based formulas while clarifying the relative influence of key design variables through a transparent, data-driven approach.

2. Motivation: Concrete Shear Strength

Figure 1 illustrates a typical reinforced concrete beam subjected to shear loading, highlighting key geometric and reinforcement parameters used in shear strength estimation. Traditional methods for predicting shear strength rely on simplified models rooted in engineering mechanics and calibrated with empirical data from laboratory tests. These models typically aim to ensure safety by providing conservative lower-bound estimates of the shear capacity.
The traditional method used to calculate the lower bound on concrete shear strength is defined as standard ACI 318-14 [7]:
V n = V c + V S
where V c is the concrete contribution and V S is the transverse steel contribution [2,8] In (1), the concrete contribution is defined as
V c = 2 λ f c · b d
where λ is an effect size, f c is the compressive strength of concrete, b is the beam width, and d is the beam depth [2,8] The steel contribution is then defined as
V S = A v f y d s
where A v is the shear reinforcement, f y is the yield strength of the sheer reinforcement, d is the effective depth of the beam, and s is the spacing of the shear reinforcement [2,8]

2.1. Experimental Dataset

As mentioned across multiple studies of concrete shear strength, cf. [2,6,9] limited large datasets exist for modeling. Thus, aggregating data from multiple studies is common practice. Of interest herein is using the compiled dataset introduced in [2] which was taken from multiple publications, including [10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37,38,39,40,41,42,43,44,45,46,47,48,49,50]. This resulted in a total dataset of 1259 samples from 49 studies published between 1951 and 2004. Notably, this further creates a heterogeneous dataset as some studies were interested in different geometries, sizes, and questions related to constructing concrete beams. Overall, this aggregation provides both statistical diversity necessary for machine learning, albeit at the cost of possibly introducing additional variability.
A description of variables, their ranges, and units is presented in Table 1. Notably, the issues in traditional prediction methods are seen even in the data summary of Table 1, with the maximum predicted value of shear strength from the studies, V p r e d , being 60+% larger than the maximum tested value, V t e s t .

2.2. Performance Metrics

For regression tasks in concrete shear strength estimation, performance is typically evaluated by regressive [2,17] Consistent with [2,17] the coefficient of determination (R2) is the standard metric in concrete shear strength prediction studies, and it provides a direct measure of variance explained by the model. In practical terms, R2 measures how much of the variation in shear strength data is explained by the model. An R2 close to 1 means the model predictions closely follow the experimental results, while lower values indicate larger unexplained differences. R2 is defined as
R 2 = 1 i y i y ^ i 2 i y i μ 2
where y i are the actual values, y ^ i are the predicted values, and μ are the mean values [51]. Notably, R2 is not a perfect measure, e.g. it does not capture curvilinear relationships, and it can imply relationships that are not there [51]. As a result, a qualitative assessment of performance is needed by observing real versus predicted and parameter estimate plots.
Given the stochastic nature of many AI algorithms in randomized training and sensitivity to starting points, multiple replications and then computing confidence intervals (CIs) for the performance metrics are useful evaluation tools. For R2, the confidence interval used is computed by first computing the mean and standard deviation:
R ^ 2 = 1 n i = 1 n R i 2   a n d   s ^ = 1 n 1 i = 1 n R i 2 R ^ 2 2
where n is the number of replications. The mean confidence interval for R2 used herein employs a t-score to permit small replication number estimates of the mean R2. Consistent with general mean CIs, this was computed as
  C I = R ^ 2 t n 1,1 α 2 s ^ n ,   R ^ 2 + t n 1,1 α 2 s ^ n ,
with t n 1,1 α 2 being the Student t quantile.

2.3. ANNs for Concrete Shear Strength Modeling

ANNs have seen limited but focused applicability in civil engineering. ANNs are of interest due to their ability to model complex nonlinear data problems, but hesitation exists due to the general black-box nature of ANN algorithms. Londhe et al. [4] provide a comprehensive and recent review of ANN methodologies in civil engineering, highlighting applications ranging from structural health monitoring to concrete strength prediction, and emphasizing the flexibility of ANNs to handle incomplete or noisy data.
In the specific domain of reinforced concrete shear strength prediction, Sanad and Saka [6] pioneered the use of ANNs for shear strength prediction, and then Young et al. [2] pioneered the exploration of ANN hyperparameter design for more accurate shear strength prediction. Young et al. [2] further focused on how input variables such as beam dimensions and material properties influence predictions. Their work demonstrated that ANNs can achieve high predictive accuracy while also revealing variable importance, setting a precedent for knowledge discovery in model outputs.
Markou and Bakas [52] extended this work by applying AI algorithms, including ANNs, to a large synthetic database of slender reinforced concrete beams generated via 3D nonlinear finite element analysis. Their study underscored the potential of combining simulation-derived datasets with AI to improve prediction quality for configurations not well-represented in experimental data.
More recently, Murphy and Paal [8] introduced a hybrid transfer learning and physics-informed feature engineering approach for modeling the shear strength of concrete walls. By embedding physics-based data features as inputs to opaque machine learning neural networks, they demonstrated that feature engineering grounded in structural mechanics can improve both predictive accuracy and model interpretability.

2.4. Physics-Informed Feature Engineering

Murphy and Paal [8] developed a feature engineering strategy whereby physics-defined features were applied as input data to conventional ANNs. This concept aims to craft features that are more relevant and incorporate subject matter expertise for the ANN to model. Of interest herein is extending this consideration by incorporating these features and understanding their relevance for predictive tasks.
The engineered features from [8] include a calculation of the total shear capacity from concrete and reinforcement:
  p 0 = α c λ f c + ρ t f y   A c v
where A c v is a cross-sectional area and assumed herein as A c v = b d , α c λ f c is associated with concrete strength and ρ t f y   is associated with reinforcement strength. In addition to p 0 , ref. [8] defines the following additional features:
  p 1 = α c λ f c
  p 2 = α c λ f c A c v
  p 3 = ρ t f y  
p 4 = ρ t f y t   A c v
where p 1 is the contribution of strength from concrete, p 2 scales p 1 by area, p 3 represents the reinforcement ratio and yield stress product, and p 4 scales p 3 by area.

3. Methodology

3.1. Artificial Neural Networks

Artificial Neural Networks (ANNs) are a class of supervised machine learning methods inspired by the structure and learning processes of biological neural systems [53] They are typically nonlinear models that learn mappings between input and output data by adapting the weights of connections among layers (input, hidden, and output) during training. Through this process, ANNs construct a mathematical model that captures complex patterns and relationships within the data.
A wide variety of ANN approaches exist, e.g., see [53], however, given the relatively small scale of the dataset and the absence of temporal or spatial structure in the features, a feedforward ANN provides an appropriate balance between model complexity and predictive capacity. Notably, when appropriately constructed, a feedforward ANN can serve as a provably optimal universal approximator [3]. However, simplicity is relative, and even a relatively simple feedforward ANN has considerable hyperparameters to choose from, which greatly influence performance. These include the number of hidden nodes, the type of functions in each node, the training approach, the training objective, and the number of iterations.

3.2. Hyperparameter Determination

AI/ML algorithmic results are often only as good as the settings used to train these algorithms. Such hyperparameters can be numerous, and determining their appropriate values is a discipline in itself. While general taxonomies of these approaches exist [54] these methods can largely be separated into model-free and model-based approaches. Model-free approaches can range from scientific grid searches to experiential/haphazard and random methods. Model-based approaches have generally outperformed model-free approaches and employ a wrapper function around the algorithm of interest and use a concerted search strategy [54].
Bayesian optimization (BO) is one of the most competitive model-based hyperparameter determination methods [55]. BO employs a sequential search strategy for optimizing expensive, black-box functions, with the objective function (typically) being the validation performance of a learning algorithm as a function of its hyperparameters, a function which is usually noisy, non-convex, and lacks a closed-form expression or gradient information. BO addresses this by fitting a probabilistic surrogate model (also known as a response surface model) to predict the objective function and by using an acquisition function to decide which hyperparameter setting to evaluate next.

3.3. Feature Saliency in Artificial Neural Networks

In many data-driven applications, experiments often aim to collect large volumes of data under the assumptions that it will be difficult, if not impossible, to collect additional data later and that more data will lead to better modeling capabilities. However, this often results in having more data than necessary. The consequence of this can be not only larger datasets than needed but also the inclusion of redundant, irrelevant, and low-quality data [56]. Thus, feature selection and feature extraction methods are commonly employed to mitigate these issues by, respectively, identifying and retaining only the most informative variables or transforming data to a more useful representation [57].
Feature extraction is the process of transforming raw data into a reduced set of informative features that capture the most relevant characteristics of the original data. These transformations often aim to maximize data variance, improve class separability, or reduce dimensionality, and possibly enhance the performance and efficiency of learning algorithms. However, these methods can create less interpretable features and cause the resultant models to be less explainable. This is in contrast to feature selection, which retains a subset of the original variables and preserves their interpretability, which is thus of interest.
Feature selection for ANN applications can take the form of filter-based approaches, which aim to provide feature relevance ranking before the ANN model is run [58] post-classifier methods which incorporate a sensitivity analysis [2] and combined model-saliency approaches, which simultaneously train and compute saliency [5] Due to their ability to intrinsically capture the importance of each feature to a model, the combined ANN-SNR approach is of interest herein.
ANN-SNR feature selection is an extension of [59,60] and is a weight-based approach for feature selection. The ANN-SNR approach adds a noise feature to the dataset, and then the ANN is trained. After training, the weights of each node are compared against the trained weights of the noise-related nodes using an SNR calculation:
S N R i = 10 log 10 j = 1 J w i , j 1 2 j = 1 J w N , j 1 2
where S N R i is the SNR value in dB for the i-th feature, J is the number of hidden nodes, w i , j 1 is the first layer weight between node i and node j, and w N , j 1 is the first layer weight from the noise node N to node j [5,57].
Consistent with [57] when using ANN-SNR, the data set is standardized, and one random normal data vector, N(0,1), is added to the dataset as the noise feature. After training an ANN and computing, by (12), the SNR values for each feature, features with higher SNR values are considered like signals that stand out clearly above background noise, whereas features with comparatively low SNR values are like weak signals that add little to the prediction. Removing the weakest features step by step helps isolate the variables that truly matter. The process of (12) is applied in a backward selection manner for all data until only one non-noise feature remains, with the noise features retained. The accuracy of each iteratively explored model is evaluated, and the best set of data features is selected.
It should be noted that the SNR values reported in ANN-SNR applications are relative measures derived from ANN weights and input variability, and are not directly comparable to signal-to-noise ratios in communication systems. Their absolute magnitude is less important than their ranking, since the method eliminates features with the lowest relative SNR.

3.4. ANN-SNR Predictive Stopping Rules

In general, finding the smallest feature set that provides for a model statistically similar results (or better) to a model trained on the entire feature set is of interest. For regression tasks, this involves selecting the smallest feature set whose mean R2 remains within or above the full model’s 95% CI lower bound on the R2. This approach is two-stage: first exploring all subsets through iterative elimination and then validating the best candidate with full optimization to ensures that the final model balances parsimony and statistical equivalence, producing a reduced yet defensible representation of the original ANN model.
The process in Figure 2 captures the end-to-end approach, beginning with finding the baseline performance through BO to optimize an initial ANN with all k features to determine l best algorithmic settings. The ANN with the l optimized settings is then trained on all k features for p replications to compute the baseline mean R2 and its 95% confidence interval. Next, setup for the ANN-SNR approach is performed with data features that are standardized, and at the initialization of the ANN-SNR backward feature selection loop, a synthetic noise variable ( μ = 0 ,   σ = 1 ) is added. TRACE is initialized as an empty record to track subsets and their R2 values. An ANN with l settings is trained on [F, noise], and the resulting R2 is recorded in TRACE along with the current subset. Features are then ranked based on their signal-to-noise ratio (SNR), derived from the ANN’s first-layer weights to quantify relative feature importance as seen in Equation (12), and the lowest-ranked feature is eliminated from F. This cycle continues until only one feature remains in F.
After the loop terminates, TRACE contains the record of all candidate feature subsets and their associated R2 values. From TRACE, the smallest subset whose mean R2 remains above the baseline confidence interval threshold is selected as the final reduced feature set. The final ANN is then re-optimized with BO using this reduced F, and its predictive performance (R2 and CI) is reported. The best result from this provides m algorithm settings.

4. Results

Results are presented for the baseline dataset and architecture, first refined using Bayesian Optimization (BO) to identify optimal hyperparameters. The optimized model was then subjected to ANN-SNR feature selection to produce a parsimonious feature set while maintaining predictive accuracy. Once a parsimonious feature set is found, results are presented for algorithms and feature sets. For all analyses, MATLAB 2025a on a MacBook Pro (Apple M4 Max, 48 GB RAM) was used.

4.1. Baseline Results

The baseline results, Table 2, are drawn from Young et al. [2] and represent a two-hidden-layer feedforward neural network (7 and 4 neurons) trained using the Levenberg–Marquardt (LM) algorithm with mean squared error (MSE) as the performance function. Two feature sets were tested: (1) the full set of experimental variables, and (2) a reduced set obtained via post hoc sensitivity analysis feature selection. The reduced feature set achieved a slightly higher mean coefficient of determination (R2 = 0.877, range: 0.832–0.903) compared to the full set (R2 = 0.866, range: 0.841–0.892). Although these differences were not statistically significant, they indicate that targeted feature selection can yield practically meaningful improvements in model accuracy.
Because ANN performance is inherently stochastic, with outcomes sensitive to initialization, training order, and software platform, BO was used to refine hyperparameters for the baseline architecture rather than relying on fixed settings. BO was applied to identify the best-performing configuration for each feature subset, ensuring that subsequent SNR-based feature selection (ANN-SNR) operated on an optimized baseline model rather than one influenced by suboptimal tuning. The hyperparameters tuned in this process are summarized in Table 3.
The hyperparameters in Table 3 define the structure and learning behavior of the artificial neural network (ANN). The number of neurons in the first and second hidden layers controls the model’s capacity to detect and refine nonlinear relationships, with higher values allowing more complex feature representations at the risk of overfitting. Activation functions (tansig, logsig, poslin) introduce nonlinearity, enabling the network to model complex behaviors beyond simple linear mappings. Regularization applies an L2 penalty to the weight magnitudes, helping improve generalization to unseen data. Training algorithms (LM, BR, SCG) balance convergence speed, stability, and robustness differently, while the patience setting determines the maximum number of epochs before stopping to prevent excessive or premature termination of training. The choice of performance metric (MAE, MSE, MSEReg) directly influences the optimization process, shaping the trade-off between penalizing large deviations and ensuring overall robustness to variability in the data.
Once the BO process was completed, the BO recommendation was rerun for 5 iterations to get performance results. Table 4 presents the results from this analysis along with the features used, the architecture recommended, and R2 with a 95% t-score mean confidence interval. Results for the original feature set, BO-ANN, were statistically comparable to the baseline (mean R2 = 0.8564, 95% CI: [0.8208, 0.8920]), suggesting that the baseline network was already close to optimal for the given data and feature set. When BO was applied exclusively to the engineered features ( p 0 p 4 ) from Murphy and Paal [8]. BO-ANN Engineered, performance decreased substantially (mean R2 = 0.6518, 95% CI: [0.5524, 0.7512]). This reflects the apparent limited predictive power of these engineered features when used in isolation. Finally, combining the original experimental features with the engineered features in a single feature set and applying Bayesian optimization produced results from the algorithm BO-ANN Hybrid (mean R2 = 0.8428, 95% CI: [0.8181, 0.8675]), which was statistically indistinguishable from both the baseline and BO-only approaches. This indicates that, for this dataset, engineered features did not significantly enhance predictive accuracy when combined with the original variables.

4.2. ANN-SNR Feature Selection

The architectures identified by BO from Table 4 served as the starting point for the ANN–SNR procedure. As discussed previously, ANN-SNR introduces a synthetic noise variable, and the signal-to-noise ratio (SNR) of each feature was computed after training. Features were iteratively removed based on the lowest SNR, with the model retrained at each step.
Table 5 presents the results of this process. SNR values (in dB) are reported for each feature at the time of elimination, with negative values indicating a weaker and noisier contribution to output prediction. The corresponding R2 values quantify predictive accuracy after each removal. Performance peaked early in the elimination process, with iteration 2 (R2 = 0.867) exceeding the original full-feature BO–ANN model. Iterations 3 and 4 produced the most parsimonious feature sets while maintaining high accuracy, within or above the model’s CI (R2 = 0.843 and R2 = 0.822, respectively). Based on the criterion of selecting the smallest feature subset whose performance remained within the confidence interval (CI) of the baseline BO–ANN model, both iterations 3 and 4 qualified as final candidates. Beyond iteration 4, further removals caused a marked decline in accuracy, indicating that variables with moderate SNR still contributed essential predictive information. The finding that not all experimental variables are salient aligns with Young et al. [2]; however, this analysis specifically identifies f y and ρ l as unimportant, in contrast to [2], which flagged ρ v h and a / d as low-value features.
Table 6 summarizes the results of BO-ANN Engineered, which explored only engineered features and the ANN-SNR feature selection process to find a parsimonious set. Each row corresponds to an iteration that reports the signal-to-noise ratio (SNR, in dB) for individual engineered features ( p 0 p 4 ) and the corresponding model R2. Notably, R2 values start considerably lower than BO-ANN, as presented in Table 5, and results in Table 6 rapidly decline across all iterations. Overall, these results suggest that the engineered features, in isolation, have limited predictive value and that feature removal or degradation in SNR correlates with a loss in model fit. The analysis showed that the smallest feature set falling within the model’s confidence interval occurred in the first iteration, which retained all original features.
Table 7 summarizes the performance of the BO–ANN Hybrid model during iterative ANN–SNR feature selection. This model starts with both the original features and all engineered feature set. SNR values (in dB) for each feature reflect their relative contribution at the point of removal, with negative or near-zero SNR indicating limited signal strength relative to noise. Notably, performance initially improved with iteration 2 and several mid-stage iterations (3, 5, 6, 7, 9, 10) maintained R2 values around or above 0.80. The final stopping point was iteration 8, which represented the smallest feature set that still had performance (R2 = 0.832) within (or above) the model’s confidence interval. This subset ( b , ρ v , ρ v h , a / d , p 0 , p 4 ) achieved a balance between parsimony and accuracy as well as between raw data features and engineered features. Compared to models using only engineered or original features, the hybrid set demonstrated greater resilience to feature removal, indicating redundancy between engineered and original variables. However, excessive removal beyond the confidence-interval threshold caused sharp performance declines (R2 < 0.70), showing that certain moderate-SNR features still provided complementary predictive information.
With the final recommended feature sets and BO-optimized architectures, the algorithms were rerun for five independent replications, this time without the synthetic noise variable, to produce the final results summarized in Table 8. Similarly to Table 4, performance metrics were averaged across replications to provide stable estimates. For each algorithm–feature configuration, the table reports the BO-optimized ANN architecture (hidden layer sizes, activation function, and regularization parameter) along with the resulting mean R2 and 95% confidence interval (CI).
Figure 3 illustrates the BO search surface for the BO–ANN–SNR3 configuration, plotted over the number of hidden nodes in the first and second layers. Each blue marker represents an individual trial evaluated during the optimization, while the surface provides an interpolated landscape of the BO objective. The red marker denotes the best-performing configuration identified by BO. As shown, the BO–ANN–SNR3 model achieved its best performance with a two-layer configuration of 68 and 15 hidden nodes using a tansig activation function and LM with MAE regularization. This result not only validates the utility of BO for architecture selection but also demonstrates how Figure 3 can serve as a representative example of the tuning process applied across all feature sets reported in Table 8.
Figure 3 highlights two important aspects of the optimization process. First, the search surface is nonconvex and irregular, with multiple local peaks and valleys. Notably, this is also for only 2 of the 8 hyperparameters BO is optimizing over, Table 3, and the entire relationship is more complex and hyperdimensional. This underscores the difficulty of tuning ANN architectures using manual or grid search methods, since small changes in the number of hidden nodes can lead to large and unpredictable swings in performance. Second, BO is shown to be effective at navigating this complex landscape: rather than exhaustively testing every possibility, BO strategically samples promising regions, converging on an architecture that balances predictive accuracy and model simplicity.
Results, Table 8 and Figure 3, indicate that the BO–ANN–SNR3 model, using the feature set ( b , d , f c , ρ v , ρ v h , a , l n ,   a / d ), achieved a mean R2 of 0.8358 with a 95% CI of [0.8160, 0.8556], demonstrating strong predictive accuracy. Removing one additional feature, ρ v , to form the BO–ANN–SNR4 model with feature set ( b , d , f c , ρ v h , a , l n , a d ) reduced performance noticeably to a mean R2 of 0.7824, CI [0.7494, 0.8154], suggesting that ρ v contributed to improved predictions despite moderate SNR. By contrast, the engineered-only feature set ( p 0 ,   p 1 ,   p 2 ,   p 3 ,   p 4 ), yielded the weakest performance (mean R2 of 0.6518, CI [0.5524, 0.7512]), showing that, in isolation, composite physics-informed variables possibly remove nonlinear interactions that the ANN can otherwise exploit.
The BO tuned model BO-ANN-SNR Hybrid provided a hybrid configuration with a feature set containing both raw and engineered features ( b , ρ v ,   ρ v h , a d ,   p 0 , p 1 , p 4 ) achieved the highest performance, with a mean R2 of 0.8524, CI [0.8312, 0.8736], exceeding all other tested combinations. Notably, this result exceeded all other reduced-feature models and fell within the performance range of the baseline BO–ANN models. These findings reinforce that combining select raw and engineered features yields the most balanced outcome: preserving interpretability and parsimony while maintaining accuracy that is statistically indistinguishable from the full feature set.

5. Discussion

When comparing the original baseline ANN models (Table 2) with the BO-optimized ANN models (Table 4) and the BO–ANN–SNR reduced-feature models (Table 8), performance remains statistically comparable despite substantial reductions in the number of input features. The original baseline configurations achieved mean R2 values between 0.866 and 0.877, with confidence intervals overlapping those of the BO-optimized ANN models. For example, the top performing BO–ANN configuration in Table 4 reached a mean R2 of 0.8564 (95% CI: [0.8208, 0.8920]), fully overlapping the baseline intervals, indicating no statistically significant loss in predictive accuracy despite feature reduction.
Against prior literature (summary presented in Table 9), the proposed framework demonstrates equivalence while improving efficiency. Compared to the results of Young et al. [2] where we recomputed a mean best R2 of 0.877 with a wide range [0.832–0.903], the BO-ANN-SNR3 model achieved a similar accuracy (mean R2 = 0.852, CI [0.831–0.874]) with one fewer feature and a much narrower confidence interval. This suggests that while the maximum performance in Young et al. [2] may appear higher, it is less stable across replications. In contrast, BO-ANN-SNR3 offers comparable predictive capability with improved reproducibility and statistical transparency, making it a more defensible choice for engineering applications.
Likewise, results from Murphy and Paal [8] included one algorithm with a maximum R2 of 0.877; however, the algorithm of Murphy and Paal [8] used 18 input variables, including variables not under consideration herein, and a much more complex transfer learning algorithm. By comparison, the BO-ANN-SNR Hybrid model achieved statistically similar performance (mean R2 = 0.852, CI [0.831–0.874]) with fewer than half as many features (7). Importantly, the hybrid configuration confirms that near-equivalent accuracy can be achieved with substantially fewer inputs, improving parsimony and interpretability without sacrificing reliability.
Additional insights are provided by examining engineered features in isolation. When used alone, the engineered set produced substantially lower predictive accuracy (R2 ≈ 0.65). This likely reflects two factors: first, the formulations condense multiple raw variables into composite terms, which may remove nonlinear interactions that ANNs can otherwise exploit; and second, aggregating features such as f c into composite/hybrid features may obscure important relationships between material strength and geometric dimensions. Including these engineered features alongside raw variables, however, preserved that information and improved robustness. The SNR-based feature reduction further confirmed that many variables contribute little to modeling and can be removed without degrading performance. Importantly, the hybrid feature set achieved the strongest performance among reduced-feature models, underscoring two critical insights: firstly, that retaining salient variables with strong statistical signal and domain relevance is essential, and secondly, that combining raw and engineered features leverages the strengths of both, maximizing predictive accuracy while enhancing interpretability. Such a balanced feature set ensures the model remains not only a powerful predictor but also a transparent analytical tool, while enabling dimensionality reduction that improves efficiency and practicality for engineering applications.
Some limitations should be acknowledged. The primary contribution of this work is the adaptation of ANN-SNR for regression, making feature saliency intrinsic to the ANN itself. Accordingly, comparisons are drawn primarily to other ANN-based studies (e.g., Young et al. [2] and Murphy & Paal [8]), which provide the most relevant context. External feature selection methods such as principal component analysis (PCA) could be an alternative; however, PCA would involve a further rendering of features to be more abstract and less explainable. Further, LASSO (Least Absolute Shrinkage and Selection Operator) could provide an additional alternative to feature saliency; however, LASSO assumes linearity in relationships whereas the ANN approach can capture nonlinear relationships. Finally, methods such as random forests could provide external measures of feature importance through impurity-based rankings or permutation testing; however, their interpretations would not be intrinsic to the neural network itself. In contrast, the ANN-SNR framework derives feature saliency directly from ANN weights and integrates it into the feature reduction process, producing interpretability that is consistent with the deployed model and grounded in the nonlinear structure of the ANN. Another limitation lies in dataset heterogeneity: the 1259 samples were aggregated from 49 experimental programs, each with its own testing protocols and material characterizations. While such heterogeneity could introduce bias, aggregated datasets are the accepted practice in AI/ML applications to reinforced concrete, see [2,6,8], precisely because individual programs rarely provide sufficient observations for data-driven modeling.

6. Conclusions

The results demonstrate the value of engineered features, and selecting salient features is critical to achieving models that are both accurate and interpretable. The analysis shows that eliminating low-value variables can significantly reduce dimensionality without compromising accuracy. Furthermore, the superior performance of the hybrid feature set highlights the value of combining original raw variables with thoughtfully engineered features; this approach not only preserves physical interpretability for domain experts but also leverages transformations that can capture nonlinear relationships more effectively. The findings suggest that feature engineering should be approached as a complementary process to statistical selection, rather than as a replacement, to maximize both predictive strength and interpretability.
The work further introduced ANN-SNR feature selection for prediction and developed a stopping rule for feature elimination whereby one stops at the smallest set of features whose performance remains within or above the full model’s confidence interval. This criterion ensures that dimensionality reduction does not come at the expense of statistically significant performance loss, providing a defensible balance between parsimony and accuracy.
Future work should focus on developing physics-informed ANN methodologies, e.g., [61], that more directly integrate engineered features derived from governing equations and domain principles, such as detailed physics models as seen in [62]. Such approaches could yield models that are not only high-performing but also grounded in physically meaningful relationships, enhancing both trustworthiness and explanatory power. A further understanding of data heterogeneity and developing stratified models within more homogeneous subsets to further isolate any effects of heterogeneity is a direction of possible interest to the field. Additionally, exploring automated pipelines that merge physics-based transformations with modern optimization strategies, such as Bayesian optimization and SNR-based filtering, could provide a systematic way to design compact yet powerful models for complex engineering prediction tasks.

Author Contributions

Conceptualization, T.J.B. and W.A.Y.II; methodology, T.J.B.; software, T.J.B.; validation, T.J.B., W.A.Y.II and A.M.; formal analysis, T.J.B.; investigation, T.J.B., W.A.Y.II and A.M.; data curation, W.A.Y.II; original draft preparation, T.J.B. and W.A.Y.II; review and editing, T.J.B., W.A.Y.II and A.M.; visualization, T.J.B. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Dataset is available on request from the authors.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ANNArtificial Neural Network
CIConfidence Interval
BOBayesian Optimization
SNRSignal-to-Noise Ratio

References

  1. Bihl, T.; Sands, T. Dynamic System Modeling for Flexible Structures. Online J. Robot. Autom. Technol. 2024, 3, 1–4. [Google Scholar]
  2. Young, W.; Weckman, G.; Brown, M.; Thompson, J. Extracting knowledge of concrete shear strength from artificial neural networks. Int. J. Ind. Eng. Theory Appl. Pract. 2008, 15, 26–35. [Google Scholar]
  3. Hornik, K.; Stinchcombe, M.; White, H. Universal approximation of an unknown mapping and its derivatives using multilayer feedforward networks. Neural Netw. 1990, 3, 551–560. [Google Scholar] [CrossRef]
  4. Londhe, S.; Kulkarni, P.; Dixit, P.; Shah, S.; Joshi, S.; Sahu, P. Unveiling the Artificial Neural Network Mystery with Special Reference to Applications in Civil Engineering. Comput. Eng. Phys. Model. 2024, 7, 18–44. [Google Scholar]
  5. Bauer, K.; Alsing, S.; Greene, K. Feature screening using signal-to-noise ratios. Neurocomputing 2000, 31, 29–44. [Google Scholar] [CrossRef]
  6. Sanad, A.; Saka, M. Prediction of ultimate shear strength of reinforced-concrete deep beams using neural networks. J. Struct. Eng. 2001, 127, 818–828. [Google Scholar] [CrossRef]
  7. Ghosh, S.K.; Brewe, J. The most notable changes from ACI 318-14 to ACI 318-19 for precast concrete. PCI J. 2024, 69, 23–35. [Google Scholar] [CrossRef]
  8. Kuchma, D.; Wei, S.; Sanders, D.; Belarbi, A.; Novak, L. Development of the One-Way Shear Design Provisions of ACI 318-19 for Reinforced Concrete. ACI Struct. J. 2019, 116, 285–296. [Google Scholar] [CrossRef]
  9. Murphy, J.; Paal, S. A combined transfer learning physics informed interpretable machine learning approach to modelling the shear strength of concrete walls. In Proceedings of the EG-ICE 2025: AI-Driven Collaboration for Sustainable and Resilient Built Environments Conference, Glasgow, UK, 1–4 July 2025. [Google Scholar]
  10. Smith, K.; Vantsiotis, A. Shear strength of deep beams. J. Proc. 1982, 79, 201–213. [Google Scholar]
  11. Rogowsky, D.; MacGregor, J. Shear Strength of Deep Reinforced Concrete Continuous Beams; University of Alberta: Edmonton, AB, Canada, 1983. [Google Scholar]
  12. Tan, K.; Kong, F.; Teng, S.; Guan, L. High-strength concrete deep beams with effective span and shear span variations. Struct. J. 1995, 92, 395–405. [Google Scholar]
  13. Shin, S.; Lee, K.; Moon, J.; Ghosh, S. Shear strength of reinforced high-strength concrete beams with shear span-to-depth ratios between 1.5 and 2.5. Struct. J. 1999, 96, 549–556. [Google Scholar]
  14. Tan, K.; Lu, H. Shear behavior of large reinforced concrete deep beams and code comparisons. Struct. J. 1999, 96, 836–846. [Google Scholar]
  15. Oh, J.; Shin, S. Shear strength of reinforced high-strength concrete deep beams. Struct. J. 2001, 98, 164–173. [Google Scholar]
  16. Bresler, B.; Scordelis, A. Shear strength of reinforced concrete beams. J. Proc. 1963, 60, 51–74. [Google Scholar]
  17. Moody, K.; Viest, I.; Elstner, R.; Hognestad, E. Shear strength of reinforced concrete beams Part 1—Tests of simple beams. J. Proc. 1954, 51, 317–332. [Google Scholar]
  18. Ramakrishnan, V.; Ananthanarayana, Y. Ultimate strength of deep beams in shear. J. Proc. 1968, 65, 87–98. [Google Scholar]
  19. Kong, F.; Robins, P.; Cole, D. Web reinforcement effects on deep beams. J. Proc. 1970, 67, 1010–1018. [Google Scholar]
  20. Subedi, N.; Vardy, A.; Kubotat, N. Reinforced concrete deep beams some test results. Mag. Concr. Res. 1986, 38, 206–219. [Google Scholar] [CrossRef]
  21. Yoon, Y.; Cook, W.; Mitchell, D. Minimum shear reinforcement in normal, medium, and high-strength concrete beams. Struct. J. 1996, 93, 576–584. [Google Scholar]
  22. Foster, S.; Gilbert, R. Experimental studies on high-strength concrete deep beams. Struct. J. 1998, 95, 382–390. [Google Scholar]
  23. de Paiva, H.; Siess, C. Strength and behavior of deep beams in shear. J. Struct. Div. 1965, 91, 19–41. [Google Scholar] [CrossRef]
  24. Angelakos, D.; Bentz, E.; Collins, M. Effect of concrete strength and minimum stirrups on shear strength of large members. Struct. J. 2001, 98, 291–300. [Google Scholar]
  25. Kani, M.; Huggins, M.; Wittkopp, R. Shear in Reinforced Concrete; University of Toronto: Toronto, ON, Canada, 1979. [Google Scholar]
  26. Krefeld, W.; Thurston, C. Studies of the shear and diagonal tension strength of simply supported reinforced concrete beams. J. Proc. 1966, 63, 451–476. [Google Scholar]
  27. De Cossio, R.; Siess, C. Behavior and strength in shear of beams and frames without web reinforcement. J. Proc. 1960, 56, 695–736. [Google Scholar]
  28. Hsuing, W.; Frantz, G. Transverse Stirrup Spacing in RC Beam. J. Struct. Eng. 1985, 11, 353–362. [Google Scholar] [CrossRef]
  29. Watstein, D.; Mathey, R. Strains in beams having diagonal cracks. J. Proc. 1958, 55, 717–728. [Google Scholar]
  30. Morrow, J.; Viest, I. Shear strength of reinforced concrete frame members without web reinforcement. J. Proc. 1957, 53, 833–869. [Google Scholar]
  31. Ferguson, P. Some implications of recent diagonal tension tests. J. Proc. 1956, 53, 157–172. [Google Scholar]
  32. Van Den Berg, F. Shear strength of reinforced concrete beams without web reinforcement. J. Proc. 1962, 59, 1587–1600. [Google Scholar]
  33. Clark, A. Diagonal tension in reinforced concrete beams. J. Proc. 1951, 48, 145–156. [Google Scholar]
  34. Ahmad, S.L.D. Flexure-shear interaction of reinforced high strength concrete beams. Struct. J. 1987, 84, 330–341. [Google Scholar]
  35. Rajagopalan, K.; Ferguson, P. Exploratory shear tests emphasizing percentage of longitudinal steel. J. Proc. 1968, 65, 634–638. [Google Scholar]
  36. Ozcebe, G.; Ersoy, U.; Tankut, T. Minimum flexural reinforcement for T-beams made of higher strength concrete. Can. J. Civ. Eng. 1999, 26, 525–534. [Google Scholar] [CrossRef]
  37. Kong, P.; Rangan, B. Shear strength of high-performance concrete beams. Struct. J. 1998, 95, 677–688. [Google Scholar]
  38. Tan, K.; Kong, F.; Teng, S.; Weng, L. Effect of web reinforcement on high-strength concrete deep beams. Struct. J. 1997, 94, 572–582. [Google Scholar]
  39. Chang, T.; Kesler, C. Correlation of Sonic Properties of Concrete with Creep and Relaxation; University of Illinois: Champaign, IL, USA, 1956. [Google Scholar]
  40. Xie, Y.; Ahmad, S.; Yu, T.; Hino, S.; Chung, W. Shear ductility of reinforced concrete beams of normal and high-strength concrete. Struct. J. 1994, 91, 140–149. [Google Scholar]
  41. Sarsam, K.; Al-Musawi, J. Shear design of high-and normal strength concrete beams with web reinforcement. Struct. J. 1992, 89, 658–664. [Google Scholar]
  42. Johnson, M.; Ramirez, J. Minimum shear reinforcement in beams with higher strength concrete. Struct. J. 1989, 86, 376–382. [Google Scholar]
  43. Roller, J.J.; Russell, H.G. Shear Strength of High-Strength Concrete Beams with Web Reinforcement. ACI Struct. J. 1990, 87, 191–198. [Google Scholar] [CrossRef]
  44. Rigotti, M. Diagonal Cracking in Reinforced Concrete Deep Beams: An Experimental Investigation. Ph.D. Thesis, Concordia University, Montreal, QC, Canada, 2002. [Google Scholar]
  45. Yoshida, Y. Shear Reinforcement for Large Lightly Reinforced Concrete Members. Ph.D. Thesis, University of Toronto, Toronto, ON, Canada, 2000. [Google Scholar]
  46. Cao, S. Size Effect and the Influence of Longitudinal Reinforcement on the Shear Response of Large Reinforced Concrete Members. Ph.D. Thesis, University of Toronto, Toronto, ON, Canada, 2001. [Google Scholar]
  47. Laupa, A.; Siess, C.; Newmark, N. The Shear Strength of Simple-Span Reinforced Concrete Beams Without Web Reinforcement; Civil Engineering Studies SRS-052; University of Illinois at Urbana-Champaign: Champaign, IL, USA, 1953. [Google Scholar]
  48. Bazant, Z.; Kazemi, M. Size effect on diagonal shear failure of beams without stirrups. ACI Struct. J. 1991, 88, 268–276. [Google Scholar] [CrossRef]
  49. Yang, K.; Chung, H.; Lee, E.; Eun, H. Shear characteristics of high-strength concrete deep beams without shear reinforcements. Eng. Struct. 2003, 25, 1343–1352. [Google Scholar] [CrossRef]
  50. Shioya, T.; Iguro, M.; Nojiri, Y.; Akiyama, H.; Okada, T. Shear strength of large reinforced concrete beams. Fract. Mech. Appl. Concr. 1990, 118, 259–280. [Google Scholar]
  51. Kutner, M.; Nachtsheim, C.; Neter, J.; Li, W. Applied Linear Statistical Models; McGraw-Hill Irwin: New York, NY, USA, 2005. [Google Scholar]
  52. Markou, G.; Bakas, N. Prediction of the shear capacity of reinforced concrete slender beams without stirrups by applying artificial intelligence algorithms in a big database of beams generated by 3D nonlinear finite element analysis. Comput. Concr. 2021, 28, 533–547. [Google Scholar]
  53. Bihl, T.; Young, W., II; Moyer, A.; Frimel, S. Artificial neural networks and data science. In Encyclopedia of Data Science and Machine Learning; IGI Global: Hershey, PA, USA, 2023; pp. 899–921. [Google Scholar]
  54. Bihl, T.; Schoenbeck, J.; Steeneck, D.; Jordan, J. Easy and efficient hyperparameter optimization to address some artificial intelligence “ilities”. In Proceedings of the Hawaii International Conference on System Sciences, Maui, HI, USA, 7–10 January 2020; pp. 6965–6974. [Google Scholar]
  55. Wang, X.; Jin, Y.; Schmitt, S.; Olhofer, M. Recent advances in Bayesian optimization. ACM Comput. Surv. 2023, 55, 287. [Google Scholar] [CrossRef]
  56. Guyon, I.; Elisseeff, A. An introduction to variable and feature selection. J. Mach. Learn. Res. 2003, 3, 1157–1182. [Google Scholar]
  57. Moore, K.; Bihl, T.; Bauer, K., Jr.; Dube, T. Feature extraction and feature selection for classifying cyber traffic threats. J. Def. Model. Simul. 2017, 14, 217–231. [Google Scholar] [CrossRef]
  58. Bihl, T.; Bauer, K.; Temple, M. Feature selection for RF fingerprinting with multiple discriminant analysis and using ZigBee device emissions. IEEE Trans. Inf. Forensics Secur. 2016, 11, 1862–1874. [Google Scholar] [CrossRef]
  59. Steppe, J.; Bauer, K.J. Improved feature screening in feedforward neural networks. Neurocomputing 1996, 13, 47–58. [Google Scholar] [CrossRef]
  60. Belue, L.; Bauer, K.J. Determining input features for multilayer perceptrons. Neurocomputing 1995, 7, 111–121. [Google Scholar] [CrossRef]
  61. Raissi, M.; Perdikaris, P.; Karniadakis, G. Physics-informed neural networks: A deep learning framework for solving forward and inverse problems involving nonlinear partial differential equations. J. Comput. Phys. 2019, 378, 686–707. [Google Scholar] [CrossRef]
  62. Angelakos, D. The Influence of Concrete Strength and Longitudinal Reinforcement Ratio on the Shear Strength of Large-Size Reinforced Concrete Beams With, and Without, Transverse Reinforcement. Ph.D. Thesis, University of Toronto, Toronto, ON, Canada, 1999. [Google Scholar]
Figure 1. Visual description of basic parameters used to estimate shear strength from a general deep beam model (a) elevation (b) cross-section (adapted from [2,6]).
Figure 1. Visual description of basic parameters used to estimate shear strength from a general deep beam model (a) elevation (b) cross-section (adapted from [2,6]).
Mathematics 13 03182 g001
Figure 2. Pseudocode process for Predictive ANN-SNR with CI stopping rule.
Figure 2. Pseudocode process for Predictive ANN-SNR with CI stopping rule.
Mathematics 13 03182 g002
Figure 3. BO search space example, for BO-ANN-SNR3, with blue dots indicating explored points and red star indicating best result.
Figure 3. BO search space example, for BO-ANN-SNR3, with blue dots indicating explored points and red star indicating best result.
Mathematics 13 03182 g003
Table 1. Concrete Shear Strength Database Variables, from [2].
Table 1. Concrete Shear Strength Database Variables, from [2].
VariableDescription[Range] (S.D.) and Units
b Width of the beam[1.5, 79.1] (4.45) inches
d Effective depth of the beam[0.8, 118.1] (10.12) inches
f c Compressive strength of concrete used in the beam[880, 18,170] (0.31) psi
f y Yield strength of the shear reinforcement[20.3, 258] (20.54) ksi
ρ l Longitudinal reinforcement ratio (steel reinforcement by area of concrete)[0.001, 0.070] (0.012)
ρ v Vertical shear reinforcement ratio[0.001, 0.029] (0.003)
ρ v h Longitudinal shear reinforcement ratio[0, 0.077] (0.004)
a Shear span (distance from the applied load and the support)[2.4, 708.6] (40.78) inches
a d Shear span-to-depth ratio[0.27, 9.7] (1.82)
l n Clear span of the beam[5.6, 1417.2] (81.5) inches
V t e s t Measured shear strength of the specimen at the face of the support[0.33, 502.35] (53.78) kips
V p r e d Predicted shear strength at the face of the support (by sources)[0.20, 828.21] (44.49) kips
Table 2. Baseline results summary from [2].
Table 2. Baseline results summary from [2].
AlgorithmFeaturesArchitectureR2
Baseline [2] b , d , f c , f y , ρ l , ρ v , ρ v h , a , l n , a d logsig [7, 4]
LM with MSE
Mean R2: 0.866
Range R2: [0.841, 0.892]
Baseline [2] b , d , f c , f y , ρ l , ρ v , a , l n logsig [7, 4]
LM with MSE
Mean R2: 0.877
Range R2: [0.832, 0.903]
Table 3. Bayesian Optimization (BO) search space.
Table 3. Bayesian Optimization (BO) search space.
HyperparameterDescriptionRange
Layer 1 nodesNeurons in the first hidden layer; controls initial feature extraction.[3, 70]
Layer 2 nodesNeurons in the second hidden layer; refines and combines features.[1, 70]
Activation functionNonlinear function used for each node in the hidden layers[tansig, logsig, poslin]
RegularizationL2 penalty to improve generalization and reduce overfitting[0, 0.4]
TrainingTraining algorithm used[LM, BR, SCG]
PatienceMaximum epochs before training stops[4000, 5000]
Performance MetricError function for optimization[MAE, MSE, MSEReg]
BO IterationsIterations for BO to search over this space100
Table 4. Results summary for BO determined optimal settings for given features after 100 iterations, with R2 results from 5 replications of each feature and algorithmic setting combination.
Table 4. Results summary for BO determined optimal settings for given features after 100 iterations, with R2 results from 5 replications of each feature and algorithmic setting combination.
AlgorithmFeaturesArchitectureR2
BO-ANN b , d , f c , f y , ρ l , ρ v , ρ v h , a , l n , a d tansig [68, 15]
LM with MAE
Reg = 0.005395
Mean R2: 0.8564
95% CI: [0.8208, 0.8920]
BO-ANN Engineered p 0 ,   p 1 ,   p 2 ,   p 3 ,   p 4 tansig [61, 14]
LM with MAE
Reg = 0.0017861
Mean R2: 0.6518
95% CI: [0.5524, 0.7512]
BO-ANN Hybrid b , d , f c , f y , ρ l , ρ v , ρ v h , a , l n , a d , p 0 ,   p 1 ,   p 2 ,   p 3 ,   p 4 tansig [64, 24]
LM with MAE
Reg = 0.0015075
Mean R2: 0.8428
95% CI: [0.8181, 0.8675]
Table 5. Results summary for BO-ANN with iterative ANN-SNR feature selection process.
Table 5. Results summary for BO-ANN with iterative ANN-SNR feature selection process.
IterationFeatures and SNR (dB)R2
b d f c f y ρ l ρ v ρ v h a l n a / d
10.16−0.030.31−1.60.21−1.07−0.85−0.51−0.58−0.340.841
21.08−0.420.63-−1.440.09−0.77−0.19−0.64−0.570.867
3 1−0.41−0.570.74--−0.09−1.06−0.180.25−0.550.843
4 11.160.531.02--−0.05-0.451.321.230.822
51.990.793.24----1.291.140.970.722
6−0.86-−0.51----−0.53−0.72−0.740.541
70.96-1.33----1.410.861.380.682
8−1.17-−1.40----−0.17-−1.040.754
9−0.82------−0.05-0.980.547
10-------0.98-0.590.576
11---------−0.720.022
1 Final selected sets, in bold, that balanced parsimony and accuracy.
Table 6. Results summary for BO-ANN Engineered with ANN-SNR feature selection process.
Table 6. Results summary for BO-ANN Engineered with ANN-SNR feature selection process.
IterationFeatures and SNR (dB)R2
p 0 p 1 p 2 p 3 p 4
1 11.140.37−0.580.16−0.490.591
2−0.51−0.97-−1.330.170.526
30.44−1.07--0.260.481
4−0.01---0.410.356
5----−0.420.202
1 Final selected set, in bold, to be within the original CI has no features removed.
Table 7. Results summary for BO-ANN Hybrid with iterative ANN-SNR feature selection process.
Table 7. Results summary for BO-ANN Hybrid with iterative ANN-SNR feature selection process.
IterFeatures and SNR (dB)R2
b d f c f y ρ l ρ v ρ v h a l n a / d p 0 p 1 p 2 p 3 p 4
114.312.19.9213.711.723.120.3511.515.225.115.716.415.714.819.50.78
211.510.4-9.95.6917.514.513.414.321.49.6618.79.652.895.350.841
311.715.7-9.572.4113.298.5211.7310.3515.75.3418.985.34-6.910.804
45.44−0.61-−0.99-9.744.944.46−1.5411.551.385.951.38-8.570.786
54.01--−0.81-3.183.473.44-6.704.716.851.83-4.360.799
66.97----11.839.638.13-15.876.5913.106.59-12.20.818
711.89----17.0412.478.02-18.2714.0716.08--16.830.812
8 1−1.46----0.372.74--6.15−0.371.60--4.110.832
9-----18.8916.41--20.3915.3918.37--21.240.801
10-----22.8016.95--23.11-21.82--21.350.805
11-----3.28---5.73-4.67--5.860.763
12---------21.64-10.54--17.880.659
13---------16.3----14.530.652
14---------1.9-----0.162
1 Final selected set, in bold, that balanced parsimony and accuracy and is within the original CI.
Table 8. Results summary for BO determined optimal algorithms for given features after 100 iterations, with R2 results from 5 replications of each feature and algorithmic setting combination.
Table 8. Results summary for BO determined optimal algorithms for given features after 100 iterations, with R2 results from 5 replications of each feature and algorithmic setting combination.
AlgorithmFeaturesArchitectureR2
BO-ANN-SNR3 b , d , f c , ρ v , ρ v h , a , l n ,   a / d tansig [68, 15]
LM with MAE
Reg = 0.005395
Mean R2: 0.8358
95% CI: [0.8160, 0.8556]
BO-ANN-SNR4 b , d , f c , ρ v h , a , l n , a d tansig [68, 15]
LM with MAE
Reg = 0.005395
Mean R2: 0.7824
95% CI: [0.7494, 0.8154]
BO-ANN-SNR
Engineered
p 0 ,   p 1 ,   p 2 ,   p 3 ,   p 4 tansig [61, 14]
LM with MAE
Reg = 0.0017861
Mean R2: 0.6518
95% CI: [0.5524, 0.7512]
BO-ANN-SNR
Hybrid
b , ρ v ,   ρ v h , a d ,   p 0 , p 1 , p 4 tansig [64, 24]
LM with MAE
Reg = 0.015075
Mean R2: 0.8524
95% CI: [0.8312, 0.8736]
Table 9. Results summary for best methods explored in this work along with published baselines, results from this work in bold.
Table 9. Results summary for best methods explored in this work along with published baselines, results from this work in bold.
AlgorithmFeaturesR2
BO-ANN-SNR38 raw data featuresMean R2: 0.8358
95% CI: [0.8160, 0.8556]
BO-ANN10 raw data featuresMean R2: 0.8564
95% CI: [0.8208, 0.8920]
BO-ANN-SNR
Hybrid
7 features: 4 raw, 3 engineeredMean R2: 0.8524
95% CI: [0.8312, 0.8736]
Best-Young [2]8 raw data featuresMean R2: 0.877
Range R2: [0.832, 0.903]
Best-Murphy and Paal [8]18 data features: 17 raw, 1 engineeredMax R2: 0.877
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Bihl, T.J.; Young, W.A., II; Moyer, A. Physics-Informed Feature Engineering and R2-Based Signal-to-Noise Ratio Feature Selection to Predict Concrete Shear Strength. Mathematics 2025, 13, 3182. https://doi.org/10.3390/math13193182

AMA Style

Bihl TJ, Young WA II, Moyer A. Physics-Informed Feature Engineering and R2-Based Signal-to-Noise Ratio Feature Selection to Predict Concrete Shear Strength. Mathematics. 2025; 13(19):3182. https://doi.org/10.3390/math13193182

Chicago/Turabian Style

Bihl, Trevor J., William A. Young, II, and Adam Moyer. 2025. "Physics-Informed Feature Engineering and R2-Based Signal-to-Noise Ratio Feature Selection to Predict Concrete Shear Strength" Mathematics 13, no. 19: 3182. https://doi.org/10.3390/math13193182

APA Style

Bihl, T. J., Young, W. A., II, & Moyer, A. (2025). Physics-Informed Feature Engineering and R2-Based Signal-to-Noise Ratio Feature Selection to Predict Concrete Shear Strength. Mathematics, 13(19), 3182. https://doi.org/10.3390/math13193182

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop