Next Article in Journal
Digital Images Authentication Technique Based on DWT, DCT and Local Binary Patterns
Next Article in Special Issue
An Integrated Machine Learning Algorithm for Separating the Long-Term Deflection Data of Prestressed Concrete Bridges
Previous Article in Journal
Toward High Throughput Core-CBCM CMOS Capacitive Sensors for Life Science Applications: A Novel Current-Mode for High Dynamic Range Circuitry
Previous Article in Special Issue
Performance of Rayleigh-Based Distributed Optical Fiber Sensors Bonded to Reinforcing Bars in Bending
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Probabilistic Damage Detection of a Steel Truss Bridge Model by Optimally Designed Bayesian Neural Network

1
School of Civil Engineering, Wuhan University, Wuhan 430072, China
2
School of Civil Engineering & Mechanics, Huazhong University of Science and Technology, Wuhan 430074, China
*
Author to whom correspondence should be addressed.
Sensors 2018, 18(10), 3371; https://doi.org/10.3390/s18103371
Submission received: 12 September 2018 / Revised: 3 October 2018 / Accepted: 6 October 2018 / Published: 9 October 2018
(This article belongs to the Special Issue Bridge Structural Health Monitoring and Damage Identification)

Abstract

:
Excellent pattern matching capability makes artificial neural networks (ANNs) a very promising approach for vibration-based structural health monitoring (SHM). The proper design of the network architecture with the suitable complexity is vital to the ANN-based structural damage detection. In addition to the number of hidden neurons, the type of transfer function used in the hidden layer cannot be neglected for the ANN design. Neural network learning can be further presented in the framework of Bayesian statistics, but the issues of selection for the hidden layer transfer function with respect to the Bayesian neural network has not yet been reported in the literature. In addition, most of the research works in the literature for addressing the predictive distribution of neural network output is only for a single target variable, while multiple target variables are rarely involved. In the present paper, for the purpose of probabilistic structural damage detection, Bayesian neural networks with multiple target variables are optimally designed, and the selection of the number of neurons, and the transfer function in the hidden layer, are carried out simultaneously to achieve a neural network architecture with suitable complexity. Furthermore, the nonlinear network function can be approximately linear by assuming the posterior distribution of network parameters is a sufficiently narrow Gaussian, and then the input-dependent covariance matrix of the predictive distribution of network output can be obtained with the Gaussian assumption for the situation of multiple target variables. Structural damage detection is conducted for a steel truss bridge model to verify the proposed method through a set of numerical case studies.

1. Introduction

Recent structural damage or component failure has raised public awareness of the need for improved infrastructure safety and maintenance. It is of great practical significance to regularly evaluate and assess the service status of civil infrastructure. There has been great interest in the development of structural health monitoring (SHM) methodology, based on vibration measurements in recent decades [1]. Structural damage generally leads to the reduction of local stiffness, which has an influence on the dynamic characteristic parameters, and changes of these parameters before and after damage are further utilized by the vibration-based structural damage detection methods to identify the damage location and related extent. Different methods have been developed in various aspects of this area so far, such as dynamic analysis methods [2,3,4], probabilistic methods [5,6,7,8,9,10], model reduction-based methods [11,12,13], artificial intelligence methods [14,15,16], modal identification methods [17,18,19,20], wavelet-based methods [21,22], optimal sensor configurations [23,24,25], and signal processing techniques [26], et al.
Among these methods, artificial neural networks (ANNs) have not been specifically developed for structural damage detection, but their pattern matching capability makes them very promising to be employed as a tool for this purpose. It is obvious that the complexity of the ANN models, including the number of hidden layers and the number of hidden neurons, impacts significantly on the training process of the ANN, as well as the performance of the trained ANN, especially for simulating a complicated functional relationship. If these numbers are too small, the obtained ANN might not be able to capture the true behavior of the training data. In contrast, if they are too large, the trained ANN may produce outputs that will fluctuate in the region between training data points. Thus, proper design of the structure of ANN models, with suitable complexity, is necessary to ensure the successful application of ANN-based structural damage detection methods. However, the ANN model is usually determined by experience or rule of thumb only, and very limited research works exist in the area of ANN-based damage detection, addressing the issue of designing ANN models properly [27,28,29]. Other than the number of hidden neurons, the type of transfer (activation) function utilized in the hidden layer is also a non-negligible factor with respect to the ANN design, since the nonlinearity of the transfer function has an effect on the generalization ability of neural networks of the trained ANN.
In the traditional ANN approach as mentioned above, the values of the network parameters, including weights and biases, are estimated from the training data set by minimizing the sum of squared errors, representing the error between the target variables and the output of the neural network. More importantly, neural network learning can also be presented in the framework of Bayesian inference. Starting from the early research works related to Bayesian neural network [30,31], there has been a growing interest in the application of Bayesian inference theory in the field of neural networks [32,33,34,35,36,37]. By noting the significance of the ANN design as mentioned previously, the selection of the number of neurons in the hidden layer for the Bayesian neural network has been addressed [28]; however, to the best of our knowledge, the issues of selection of the transfer function, with respect to the Bayesian neural network, has not yet been addressed in previous studies. In addition, one of the significant features of the Bayesian neural network is the capability of quantifying the error or uncertainty of the network output, by employing the Bayesian inference approach. Nevertheless, most of the research activities in the literature for addressing the predictive distribution of neural network output only deal with a single target variable [38,39,40], whereas the distribution over network output for multiple target variables is seldom mentioned.
In this paper, the Bayesian neural networks with multiple target variables is optimally designed for the purpose of probabilistic structural damage detection; additionally, the simultaneous selection of numbers of neurons and the corresponding transfer function in the hidden layer to obtain a neural network with suitable complexity is well addressed. Furthermore, by assuming that the posterior distribution of network parameters is approximated as a sufficiently narrow Gaussian, so that the nonlinear network function is approximately linear with respect to the network parameters over the region of parameter space, the input-dependent covariance matrix of predictive distribution, being approximated also as a Gaussian, can be obtained for the case of multiple target variables. The feasibility and validity of the proposed method is verified by the numerical case studies of a steel truss bridge model.

2. Theoretical Development

The multi-layer feedforward neural network is commonly employed for ANN-based structural damage detection in the literature [41], and it has been proved to be capable of approximating any functional relationship between inputs and outputs with only one hidden layer [42]. In the present paper, without loss of generality, the proposed method focuses on the design of the structure of the hidden layer for a single-hidden-layer feedforward Bayesian neural network, since a linear function is always utilized as the transfer function in the output layer. This thus involves the selection of the number of neurons in the hidden layer, as well as the corresponding transfer function of the neurons in the same layer. The Bayesian neural network investigated in this paper updates the weight and bias values according to Levenberg–Marquardt optimization routine through the Bayesian regularization procedure by minimizing a linear combination of squared errors and weights.
The set of model classes for the Bayesian neural network in the present paper are defined as M 1 , M 2 , , M N M , representing the network model with the different number of neurons and various types of transfer functions in the hidden layer, and N M is the total number of model classes considered. Consider the problem of predicting multiple target variables t R N O × 1 from a vector x R N I × 1 of inputs by utilizing the Bayesian neural network approach. N I and N O are the number of neurons in the input and output layers. Let D N = { { x 1 , t 1 } , { x 2 , t 2 } , , { x N , t N } } denote the input–output training data set which is generated from the finite-element analysis of target structure, and N is the total number of training data sets. For the structural damage detection problem investigated in this paper, the changes of natural frequencies and normalized partial mode shapes corresponding to the measured degrees of freedom (DOF) for the first few modes before and after damage are utilized as pattern features for the damage, which correspond to the ANN input. The damaged members that are to be identified relate to the ANN output. Thus, the aim of the Bayesian neural network design is to select the most plausible class of models by using data D N from N M prescribed classes of Bayesian neural network models to approximate the functional relationship defined by the input and output data set.
It is assumed that for the j th class of model M j , that the conditional distribution p ( t | x , w j ; M j ) is Gaussian with an x -dependent mean given by the output of a neural network model y ( x ,   w j ; M j ) . Also, the multiple target variables t are assumed to be independent conditional on the inputs x and network parameters w j including weights and biases with shared noise precision parameter β j , which represents the precision of the Gaussian noise. Then, the conditional distribution of the target values with respect to the j th class of models is given by [38]:
p ( t | x ,   w j , β j , M j ) = N ( t | y ( x ,   w j ; M j ) ,   β j 1 I N O )
where I N O is an identity matrix in dimension N O .
Given the input−output training data set D N , one can construct the corresponding likelihood function as:
p ( D N | w j , β j , M j ) = n = 1 N N ( t n | y ( x n ,   w j ; M j ) ,   β j 1 I N O ) = = ( β j 2 π ) N N o / 2 exp ( β j 2 n = 1 N y ( x n ,   w j ; M j ) t n 2 )
where · represents the Euclidean norm.
Similarly, a prior distribution can be chosen to be Gaussian over the uncertain network parameters w j as the following:
P ( w j | α j , M j ) = N ( w j | 0 ,   α j 1 I W j ) = ( α j 2 π ) W j / 2 exp ( α j 2 w j 2 )
where I W j is a W j × W j identity matrix. W j = N I + N O + N H is the dimension of the weight vector, and N H is the number of neurons in the hidden layer.
By following the Bayes’ theorem, the posterior distribution of the network parameters w j for the model class M j is given by:
p ( w j | D N , α j , β j , M j ) = p ( D N | w j , β j , M j ) p ( w j | α j , M j ) p ( D N | α j , β j , M j )
which, however, is non-Gaussian due to the consequence of the nonlinear dependence of the network function y ( x ,   w j ; M j ) on the network parameters w j .
By utilizing the Laplace approximation, one can seek a Gaussian approximation to the posterior distribution p ( w j | D N , α j , β j , M j ) in Equation (4) at a (local) maximum of the posterior, say w j , MP , which can be obtained through the standard nonlinear optimization routine. As usual, it is convenient to maximize the logarithm of the posterior, which can be written as:
ln [ p ( w j | D N , α j , β j , M j ) ] β j 2 n = 1 N y ( x n ,   w j ; M j ) t n 2 α j 2 w j 2
which corresponds to a regularized sum-of-squares error function, while the two hyperparameters α j and β j are fixed and known at this moment.
Then, from Equation (5), a local Gaussian approximation can be built in by evaluating the matrix of second derivatives of the negative log posterior distribution at the maximum of the posterior, and it is given by:
p ( w j | D N , α j , β j , M j ) = N ( w j | w j , MP ,   H j 1 ( w j , MP ) )
where H j ( w j , MP ) = ln [ p ( w j | D N , α j , β j , M j ) ] | w j = w j , MP is the Hessian matrix evaluated at w j , MP .
By making use of the evidence framework, together with the Gaussian approximation to the posterior utilizing the Laplace approximation, the marginal likelihood or evidence for the two hyperparameters is obtained by integrating the network parameters as:
p ( D N | α j , β j , M j ) = p ( D N | w j , β j , M j ) p ( w j | α j , M j ) d w j p ( D N | w j , MP , β j , M j ) p ( w j , MP | α j , M j ) ( 2 π ) W j / 2 | H j ( w j , MP ) | 1 / 2
In the evidence framework, the point estimates for hyperparameters α j and β j can be obtained by maximizing ln [ p ( D N | α j , β j , M j ) ] with respect to α j and β j at the maximum of the posterior w j , MP [38], respectively, as:
1 α j , MP = 1 γ j , MP w j , MP 2 , 1 β j , MP = 1 N N O γ j , MP n = 1 N y ( x n ,   w j , MP ; M j ) t n 2
where γ j , MP = W j α j , MP tr ( H j 1 ( w j , MP ) ) represents the effective number of parameters corresponding to the j th model class. Equation (8) presents a practical iterative procedure for estimating the hyperparameters during the training process of Bayesian neural network.
To select the most plausible model class among N M prescribed model classes for the Bayesian neural network, the probability of a model class M j conditional on the given set of input-target training data D N should be calculated. This can be obtained by following the Bayes’ theorem as:
p ( M j | D N ) = p ( D N | M j ) π ( M j ) j = 1 N c p ( D N | M j ) π ( M j )
where the prior probability π ( M j ) on the model class M j , for j = 1 to N M , satisfies j = 1 N c π ( M j ) = 1 . As there is generally no prior information about each class of models for the purpose of structural damage detection, it is simply assumed hereafter that each individual model class possesses the same initial plausibility, i.e., π ( M j ) = 1 / N M . The factor p ( D N | M j ) is the most important term in Equation (9), and it is known as the evidence for the model class M j giving the set of input–output training data D N . Generally, the class of models to be used is the one that maximizes the posterior probability p ( M j | D N ) , or equivalently, maximizes the model evidence p ( D N | M j ) with respect to M j .
It is noted that instead of utilizing the usual Bayesian treatment for hyperparameters involving marginalization over all possible values, the model evidence p ( D N | M j ) presented in Equation (9) can be approximated by substituting the values of hyperparameters α j , MP and β j , MP obtained at the maximum of the posterior w j , MP from the iterative optimization procedure given in Equation (8) into the marginal likelihood in Equation (7), i.e.,
p ( D N | M j ) p ( D N | w j , MP , β j , MP , M j ) p ( w j , MP | α j , MP , M j ) ( 2 π ) W j / 2 | H j ( w j , MP ) | 1 / 2
It is important to note that the form of the evidence p ( D N | M j ) given in Equation (10) is consistent with that given by the authors of [43], that is, the evidence consists of two terms, p ( D N | w j , MP , β j , MP , M j ) and p ( w j , MP | α j , MP , M j ) ( 2 π ) W j / 2 | H j ( w j , MP ) | 1 / 2 , namely, the likelihood factor and the Ockham factor. Specifically, the likelihood factor favors more complex model classes. Thus, it will be higher for those model classes making the probability of the data D N higher, implying a better fit to the data. The Ockham factor, however, imposes a penalty against the complexity of the specified model class. The balance between these two factors allows one to select the most probable model class through a mathematically rigorous and robust way, which is just complex enough to fit the given data. In this study, the class of models to be selected is the one possessing the highest value of evidence, i.e., maximum posterior probability, among the entire set of model classes for the given set of train data. It is also noted that because the corresponding numerical values are usually very large, the logarithmic form of the evidence is taken in order to avoid the computational problem during the procedure of model class selection. This yields that:
ln [ p ( D N | M j ) ] N N O 2 ln β j , MP β j , MP 2 n = 1 N y ( x n ,   w j , MP ; M j ) t n 2 + W j 2 ln α j , MP α j , MP 2 w j , MP 2 + W j 2 ln ( 2 π ) 1 2 ln | H j ( w j , MP ) |
where the sum of the terms in the first row gives the logarithm of the likelihood factor, and that in the second row represents the logarithmic expression of the Ockham factor.
In addition, it should be noted that if Equation (9) is applied for identifying the ‘optimal’ class of Bayesian neural network models by direct comparison of the conditional probability p ( M j | D N ) , the total number of model classes N M to be considered in the identification process should be specified at the beginning. If N M is too small, the ‘optimal’ model class might not be included in the study. However, if this number is too large, the computational consumption would be unaffordable. Thus, instead of directly comparing the all N M model classes and picking up the ‘best’ one, the number of neurons as well as the transfer function in the hidden layer is identified by following a computationally efficient algorithm [29]. Denote N T as the number of transfer functions involved for comparison, and the main procedure of the algorithm is summarized as following:
  • Initialize the index of the type of hidden layer transfer function i = 1 (outer loop).
  • Initialize the index of number of hidden neurons k = 1 (inner loop), and calculate the corresponding log evidence ln [ p ( D N | M j ( i , k ) ) ] in Equation (11), where j depends on both i and k .
  • Increase the index k by 1 , and calculate the log evidences ln [ p ( D N | M j ( i , k + 1 ) ) ] , which is compared with ln [ p ( D N | M j ( i , k ) ) ] ; If ln [ p ( D N | M j ( i , k + 1 ) ) ] < ln [ p ( D N | M j ( i , k ) ) ] , then M j ( i , k ) is the model class with ‘optimal’ number of hidden neurons as for the i th transfer function, and the inner loop will stop; Or, the algorithm will increase the index k by 1 and repeat this step; Record the log evidence ln [ p ( D N | M j ( i , k i ) ) ] related to the ‘optimal’ number of hidden neurons for the present transfer function.
  • Increase the index of transfer function i by 1, repeat steps 2 and 3, and record the log evidence ln [ p ( D N | M j ( i + 1 , k i + 1 ) ) ] related to the ‘optimal’ number of hidden neurons for the ( i + 1 ) th transfer function; Compare ln [ p ( D N | M j ( i + 1 , k i + 1 ) ) ] with ln [ p ( D N | M j ( i , k i ) ) ] , and record the larger one.
  • If i < N T , set i = i + 1 , and repeat step 4; Otherwise, if i = N T , the whole algorithm will stop, and output the ‘optimal’ number of hidden neurons k i * as well as the related hidden layer transfer function index i * corresponding to the ‘optimal’ model class M j ( i * , k i * ) with recorded largest log evidence ln [ p ( D N | M j ( i * , k i * ) ) ] .
On the other hand, it is assumed that the identified most plausible model class through the above design procedure is M K , and K = j ( i * , k i * ) . Then, the posterior of the network parameters given in Equation (6) can be employed to produce a distribution over the network outputs, and the predictive distribution is obtained by marginalizing with respect to this posterior distribution, given as:
p ( t | x , D N , M K ) = p ( t | x ,   w K , β K , MP , M K ) p ( w K | D N , α K , MP , β K , MP , M K ) d w K
It is noted that this integration is analytically intractable due to the nonlinearity of the network function y ( x ,   w K ; M K ) as a function of the network parameters w K . Under such circumstances, an approximation to evaluate this integral is generally required. To make progress, it is assumed that the covariance of posterior distribution of network parameters is small, so that the network function is approximately linear with respect to the parameters over the region of parameter space for which the posterior probability is significantly nonzero [38]. By making a Taylor series expansion of the network function around the maximum of the posterior w K , MP and retaining only the linear terms, the conditional distribution of the target values can be written as:
p ( t | x ,   w K , β K , MP , M K ) N ( t | y ( x ,   w K , MP ; M K ) + J K ( x , w K , MP ) ( w K w K , MP ) ,   β K , MP 1 I N O )
where the mean is a linear function of w K , and J K ( x , w K , MP ) = y ( x ,   w K ; M K ) | w K = w K , MP is the x-dependent Jacobian matrix of the vector-valued network function evaluated at the maximum of the posterior w K , MP , and can be calculated by utilizing the finite difference approach.
Then, by noticing that the posterior distribution of weights is approximated as a sufficiently narrow Gaussian, one arrives at a Gaussian distribution over the outputs of the network as the following:
p ( t | x , D N , M K ) N ( t | y ( x ,   w K , MP ; M K ) ,   Σ K ( x , w K , MP ) )
where the predictive distribution is approximated as a multivariate Gaussian, the mean of which is given by the network function y ( x ,   w K , MP ; M K ) with the network model parameters set to their maximum of the posterior -estimate. The corresponding x-dependent covariance matrix is given by:
Σ K ( x , w K , MP ) = β K , MP 1 I N O + J K ( x , w K , MP ) H K 1 ( w K , MP ) J K T ( x , w K , MP )
It is apparent that the x-dependent covariance matrix Σ K ( x , w K , MP ) consists of two terms. The first one reflects the intrinsic noise on the target variable, whereas the second is an x-dependent term expressing uncertainty, due to the uncertainty of the weights.

3. Case Studies

To verify the proposed methodology, numerical case studies are conducted in this section for a simply-supported steel truss bridge model, as shown in Figure 1. The space steel truss bridge, including 11 crossbeams, has a total length of 2.8 m, a width of 0.48 m and a height of 0.4 m, respectively. Except for the cross-section of the crossbeams being the I-steel type, the other components, including the upper and lower chords, and the diagonal and vertical rods, are all made of a pair of angle irons with the same size. Sectional and material properties of the truss bridge model are shown in Table 1. Each component of the truss bridge model is discretized into one beam finite element, and the total number of such elements for the entire finite element model is 85. The configuration of measurement points on the bridge deck is shown in Figure 2, and the ten measurement points, evenly located on the both sides of bridge deck, measure the vertical motion of the bridge structure. The first six mode shapes of the steel truss bridge deck in intact status are shown in Figure 3, including the first three vertical vibration and torsional modes.
The damage is considered on the lower chord of the steel truss bridge model by reducing the bending stiffness, which is achieved by reducing the Young’s Modulus of the potentially damaged structural member with a scaling factor in the present study, and there is a total of 20 potential damage locations, which are denoted by E1 to E20, as shown in Figure 4, respectively. However, the excessive number of damaged locations to be identified will result in too many output neurons, which is not conducive to Bayesian neural network training and subsequent damage identification procedures. Thus, one can consider combining these 20 possible damage locations into 5 groups, represented by EG1 to EG5, respectively, and each group consists of 4 lower chord elements, as shown in Figure 4 and Table 2. In this way, the number of output neurons in the neural network is greatly reduced, from 20 to 5. For example, the element group EG2 includes four individual elements E3, E4, E13, and E14. In such circumstances, if any one or more of the four elements in the same element group are damaged, the damage is only reflected by the overall bending stiffness reduction of the element group EG2. In addition, considering the fact that in the initial stage of the damage development, the number of simultaneously damaged elements would be very few, it is reasonably assumed herein that there are, at most, two damaged elements at the same time for the truss bridge model. Furthermore, three damage extents of 0, 20%, and 40% with respect to the element group, respectively, are utilized to generate the training data for the Bayesian neural network. It is also noted that for each sample of training data, the differences of the natural frequencies and mode shapes before and after the damage for the steel truss bridge model is stacked into a vector and used as the network input. It is noted that although all modes can be calculated from the finite element model, only modal parameters of the first two modes are utilized for damage identification of the steel truss bridge model, in order to evaluate the proposed method with limited information to mimic reality, and the total number of input neurons is 22.
This example considers two commonly used types of hidden layer transfer functions, i.e., tansig and satlins [44], labeled as TF1 and TF2 respectively. Figure 5 shows the iterative curve of the proposed Bayesian neural network design procedure. The abscissa represents the number of iterations, while the ordinate denotes the logarithm of evidence. This curve can be clearly divided into two segments, related to the hidden layer transfer functions TF1 and TF2, respectively. It is clear from the results corresponding to TF1 that, as the number of neurons in the hidden layer increases, the log-evidence value gradually increases. Specifically, as shown in Figure 5, when the number of hidden neurons is less than five, the increase of log evidence along with the number of hidden neurons is particularly significant, whereas after that, this increase becomes very gentle. It can thus be seen that, the number of neurons in the hidden layer has a great influence on the performance of the Bayesian neural network, especially when the number of hidden neurons is small; however, as the number increases to a certain extent, the influence becomes less obvious. This clearly indicates that the number of neurons in the hidden layer cannot be too small; otherwise, the Bayesian neural network might not work well. As the number of hidden neurons increases to 12, the log-evidence value reaches the maximum value with respect to the hidden layer transfer function TF1, and the continuous increase of the number of hidden neurons results in a decrease in the log evidence. Thus, the network design algorithm switches to the next candidate transfer function TF2. Similarly, with the increase of number of hidden neurons, the log evidence increases significantly when the number of hidden neurons is relatively small, and gradually reaches its local maximum value. This, however, is smaller than the largest log-evidence value found previously for TF1, and the algorithm stops and finally determines TF1 as the best transfer function under consideration, and the corresponding optimal number of hidden layer neurons is found to be 12 at the same time.
In order to interpret the principle of the proposed network design approach more clearly, taking the hidden layer transfer function TF1 as an example, Figure 6 shows the full picture of the log evidence, the likelihood factor, and the Ockham factor when the number of neurons in the hidden layer gradually increases from 1 to 20, with the number of input and output neurons remaining unchanged. It can clearly be seen from this figure that, as the number of neurons in the hidden layer gradually increases, the structure of the Bayesian neural network model becomes more complex, and network model is expected to better fit the training data, which is reflected by the increase of likelihood factor. At the same time, the absolute value of the Ockham factor gradually increases, which represents a gradual increase in penalty for the complexity of the network model, and the detailed numerical results are also presented in Table 3 for reference (the results corresponding to the ‘optimal’ numbers of hidden neurons for both TF1 and TF2 are shown in bold). Therefore, under the joint influence of these two factors, the Bayesian network model with the optimal number of hidden neural can be reasonably determined. In addition, it should be realized here that the network performance is sharply increased for a small number of hidden neurons, and then gradually decreased after the log evidence reaches its maximum value, but the magnitude of the decrease is not too obvious. This clearly indicates that the normal operation of the Bayesian neural network requires a sufficient number of hidden neurons, and it also shows that the network optimization design is particularly important in the case of fewer neurons in the hidden layer.
In this paper, three damage cases are considered for the steel truss bridge model. As shown in Figure 7 and Table 4, the damage is simulated by reducing bending stiffness of the corresponding individual element. Specifically, Case 1 considers the single damage situation, and the damage only occurs in element E3, which belongs to the 2nd element group EG2. Case 2 relates to a double-damage condition, and the damage exists simultaneously in E3 and E14. The last one is a multi-damage case, based on Case 2, where the simultaneous damage on E19 and E20 is further involved. Referring to the element group configuration as shown in Table 2 and Figure 4, both Case 1 and Case 2 represent damage that occurred in EG2, whereas the latter shows a relatively greater damage extent. At the same time, Case 3 assumes that EG2 and EG5 are simultaneously damaged. Without loss of generality, the modal parameters utilized for the training data of the neural network involve 1% Gaussian white noise, and the previously optimized Bayesian neural network is employed for the network training and subsequent probabilistic damage detection process. The prediction results of the network output are shown in Table 5. It can be seen from this table that the results of each damage case and the corresponding degree of uncertainty can be successfully identified. Comparing the damage identification results of Case 1 and Case 2, it is clearly seen that the uncertainty of identified damage is not much different under both cases, but the identified damage extent of EG2 for Case 2 is significantly larger than Case 1. This is consistent with the prescribed damage for each case as defined in Table 4, i.e., in Case 2, element group EG2 contains two damaged individual elements at the same time. For the last case, the proposed method also successfully identifies the actually damaged element groups EG2 and EG4, and the uncertainty associated with the damage identification results in such a case is also successfully quantified.
Additionally, the damage probabilities [45] for this truss bridge model are further quantified to seek a better interpretation of the damage detection results, the data of which are shown in Figure 8 for all three cases under consideration. It can be found for Case 1 that the 2nd element group of the truss bridge model has the greatest probability of damage, as indicated in Figure 8(a1−e1), since the corresponding probability value is somewhat more significant than others. But it is also realized that, as compared to the relatively small extent of damage in EG2, EG4 also possesses a certain degree of damage probability due to the symmetry of the steel truss bridge model. This might not be virtually revealed by directly inspecting the results presented in Table 5, and it will possibly cause some interference to the damage judgment for the present case. As for Case 2, there are also similar observations derived from Figure 8(a2–e2), i.e., EG2 is also the most likely damaged, as compared to the remaining element groups. The probability of damage in EG2 is much more significant than others, since it is a more serious damage configuration as compared to Case 1. For Case 3, it is very obvious from Figure 8(a3−e3) that both EG2 and EG5 have a relatively large possibility of damage, which coincides well with the actual damage configuration prescribed for this case.

4. Conclusions

For the purpose of probabilistic damage detection based on vibration measurements, the artificial neural network embedded with Bayesian inference for multiple target variables is optimally designed through simultaneous selection of the number of hidden neurons and the type of transfer functions in the hidden layer to obtain a network model with suitable complexity. The validity of the proposed methodology is verified by the numerical case studies conducted for a steel truss bridge model. The obtained results clearly show that under the joint influence of the likelihood and Ockham factors, the proposed network design procedure allows one to determine the optimal architecture of Bayesian neural network model, including a suitable hidden layer transfer function, as well as the optimal number of hidden neurons at the same time via a mathematically rigorous and robust way, which is just complex enough to fit the training data. In addition, it is also revealed that for the present example, regardless of the specific form of the hidden layer transfer function, the normal operation of the Bayesian neural network requires a sufficient number of neurons in the hidden layer, and the network optimization design is particularly important in the situation of fewer neurons in the hidden layer. Furthermore, in addition to the successful identification of damage locations and relative extents, the multivariate predictive distribution of the Bayesian neural network output derived in this paper can be efficiently utilized for quantifying the uncertainty associated with the statistical damage identification results. It should be pointed out that both the Hessian matrix of the logarithm of the posterior and the input-dependent Jacobian matrix of the vector-valued network function evaluated at the maximum of the posterior of weight vector, are calculated by the finite difference approach, which is very time consuming, especially for the weight vector with high dimension. Thus, one of the directions of future work will focus on reducing such computational cost for the purpose of practice application. Furthermore, only numerical examples are employed in the present paper, and the proposed method will be further validated by the experimental case studies in the coming publications.

Author Contributions

Writing-Original Draft Preparation, T.Y.; Writing-Review & Editing, H.Z.

Funding

The authors gratefully acknowledge the financial support provided by the National Natural Science Foundation of China (Grants No. 51778506, 51838006), and a scholarship from the China Scholarship Council (File No. 201806275091) while the first author visiting the Center for Engineering Dynamics and Institute for Risk and Uncertainty in the University of Liverpool.

Acknowledgments

The authors would also like to thank the Editor and the anonymous reviewers for their constructive comments and valuable suggestions to improve the quality of the article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Qarib, H.; Adeli, H. Recent advances in health monitoring of civil structures. Sci. Iran. 2014, 21, 1733–1742. [Google Scholar]
  2. Fei, Q.G.; Xu, Y.L.; Ng, C.L.; Wong, K.Y.; Chan, W.Y.; Man, K.L. Structural health monitoring oriented finite element model of Tsing Ma bridge tower. Int. J. Struct. Stab. Dyn. 2007, 7, 647–668. [Google Scholar] [CrossRef]
  3. Yin, T.; Lam, H.F. Vibration analysis of a rectangular thin plate with two parallel cracks. Adv. Struct. Eng. 2010, 13, 741–753. [Google Scholar] [CrossRef]
  4. Yin, T.; Lam, H.F. Dynamic analysis of finite-length circular cylindrical shells with a circumferential surface crack. J. Eng. Mech. 2013, 139, 1419–1434. [Google Scholar] [CrossRef]
  5. Beck, J.L.; Au, S.K.; Vanik, M.W. Monitoring structural health using a probabilistic measure. Comput. Aided Civ. Inf. 2001, 16, 1–11. [Google Scholar] [CrossRef]
  6. Huang, Y.; Beck, J.L.; Li, H. Hierarchical sparse Bayesian learning for structural damage detection: Theory, computation and application. Struct. Saf. 2017, 64, 37–53. [Google Scholar] [CrossRef]
  7. Mu, H.Q.; Yuen, K.V. Novel outlier-resistant extended Kalman filter for robust online structural identification. J. Eng. Mech. 2014, 141, 04014100. [Google Scholar] [CrossRef]
  8. Yin, T.; Lam, H.F.; Chow, H.M. A Bayesian probabilistic approach for crack characterization in plate structures. Comput. Aided Civ. Inf. 2010, 25, 375–386. [Google Scholar] [CrossRef]
  9. Mu, H.Q.; Yuen, K.V. Ground motion prediction equation development by heterogeneous Bayesian learning. Comput. Aided Civ. Inf. 2016, 31, 761–776. [Google Scholar] [CrossRef]
  10. Huang, Y.; Beck, J.L.; Wu, S.; Li, H. Robust Bayesian compressive sensing for signals in structural health monitoring. Comput. Aided Civ. Inf. 2014, 29, 160–179. [Google Scholar] [CrossRef]
  11. Yin, T.; Lam, H.F.; Chow, H.M.; Zhu, H.P. Dynamic reduction-based structural damage detection of transmission tower utilizing ambient vibration data. Eng. Struct. 2009, 31, 2009–2019. [Google Scholar] [CrossRef]
  12. Lam, H.F.; Yin, T. Dynamic reduction-based structural damage detection of transmission towers: Practical issues and experimental verification. Eng. Struct. 2011, 33, 1459–1478. [Google Scholar] [CrossRef]
  13. Yin, T.; Jiang, Q.H.; Yuen, K.V. Vibration-based damage detection for structural connections using incomplete modal data by Bayesian approach and model reduction technique. Eng. Struct. 2017, 132, 260–277. [Google Scholar] [CrossRef]
  14. Lam, H.F.; Yuen, K.V.; Beck, J.L. Structural health monitoring via measured Ritz vectors utilizing artificial neural networks. Comput. Aided Civ. Infrastruct. Eng. 2006, 21, 232–241. [Google Scholar] [CrossRef]
  15. Green, P.L. Bayesian system identification of a nonlinear dynamical system using a novel variant of Simulated Annealing. Mech. Syst. Signal. Proc. 2015, 52–53, 133–146. [Google Scholar] [CrossRef]
  16. Kourehli, S.S. Damage assessment in structures using incomplete modal data and artificial neural network. Int. J. Struct. Stab. Dyn. 2015, 15, 1450087. [Google Scholar] [CrossRef]
  17. Au, S.K.; Zhang, F.L.; Ni, Y.C. Bayesian operational modal analysis: Theory, computation, practice. Comput. Struct. 2013, 126, 3–14. [Google Scholar] [CrossRef]
  18. Yang, J.H.; Lam, H.F.; Hu, J. Ambient vibration test, modal identification and structural model updating following Bayesian framework. Int. J. Struct. Stab. Dyn. 2015, 15, 1540024. [Google Scholar] [CrossRef]
  19. Zhang, F.L.; Au, S.K. Fundamental two-stage formulation for Bayesian system identification, Part II: Application to ambient vibration data. Mech. Syst. Signal. Proc. 2016, 66, 43–61. [Google Scholar] [CrossRef]
  20. Zhang, F.L.; Xiong, H.B.; Shi, W.X.; Ou, X. Structural health monitoring of Shanghai Tower during different stages using a Bayesian approach. Struct. Control Health Monit. 2016, 23, 1366–1384. [Google Scholar] [CrossRef]
  21. Zhu, B.; Leung, A.Y.T.; Wong, C.K.; Lu, W.Z. On-line health monitoring and damage detection of structures based on the wavelet transform. Int. J. Struct. Stab. Dyn. 2008, 8, 367–387. [Google Scholar] [CrossRef]
  22. Alvandi, A.; Bastien, J.; Jolin, M. Bridge integrity assessment by continuous wavelet transforms. Int. J. Struct. Stab. Dyn. 2009, 9, 11–43. [Google Scholar] [CrossRef]
  23. Papadimitriou, C.; Beck, J.L.; Au, S.K. Entropy-based optimal sensor location for structural model updating. J. Vib. Control 2000, 6, 781–800. [Google Scholar] [CrossRef]
  24. Yi, T.H.; Li, H.N.; Gu, M.; Zhang, X.D. Sensor placement optimization in structural health monitoring using niching monkey algorithm. Int. J. Struct. Stab. Dyn. 2014, 14, 1440012. [Google Scholar] [CrossRef]
  25. Yin, T.; Yuen, K.V.; Lam, H.F.; Zhu, H.P. Entropy-based optimal sensor placement for model identification of periodic structures endowed with bolted joints. Comput. Aided Civ. Inf. 2017, 32, 1007–1024. [Google Scholar] [CrossRef]
  26. Qarib, H.; Adeli, H. A new adaptive algorithm for automated feature extraction in exponentially damped signals for health monitoring of smart structures. Smart Mater. Struct. 2015, 24, 125040. [Google Scholar] [CrossRef]
  27. Yuen, K.V.; Lam, H.F. On the complexity of artificial neural networks for smart structures monitoring. Eng. Struct. 2006, 28, 977–984. [Google Scholar] [CrossRef]
  28. Arangio, S.; Beck, J.L. Bayesian neural networks for bridge integrity assessment. Struct. Control Health Monit. 2012, 19, 3–21. [Google Scholar] [CrossRef]
  29. Lam, H.F.; Ng, C.T. The selection of pattern features for structural damage detection using an extended Bayesian ANN algorithm. Eng. Struct. 2008, 30, 2762–2770. [Google Scholar] [CrossRef] [Green Version]
  30. MacKay, D.J.C. A practical Bayesian framework for back-propagation networks. Neural Comput. 1992, 4, 448–472. [Google Scholar] [CrossRef]
  31. Buntine, W.; Weigend, A. Bayesian back-propagation. Complex Syst. 1991, 5, 603–643. [Google Scholar]
  32. MacKay, D.J.C. Bayesian Methods for Backpropagation Networks. Model of Neural Networks III; Chapter 6; Springer: Berlin, Germany, 1994; pp. 211–254. [Google Scholar]
  33. Lampinen, J.; Vethari, A. Bayesian approach for neural networks—Review and case studies. Neural Netw. 2001, 14, 257–274. [Google Scholar] [CrossRef]
  34. Barber, D. Bayesian Methods for Supervised Neural Networks. In Handbook of Brain Theory and Neural Networks; MIT Press: Cambridge, UK, 2002. [Google Scholar]
  35. Neal, R.M. Bayesian Learning for Neural Networks; Lecture Notes in Statistics; Springer: Berlin, Germany, 1996; Volume 118. [Google Scholar]
  36. Lee, H.K.H. Bayesian Nonparametrics via Neural Networks; ASA−SIAM Series on Statistics and Applied Probability; SIAM: Society for Industrial and Applied Mathematics: Philadelphia, PA, USA, 2004. [Google Scholar]
  37. Nabney, I.T. NETLAB—Algorithms for Pattern Recognition; Springer: Berlin, Germany, 2004. [Google Scholar]
  38. Bishop, C.M. Pattern Recognition and Machine Learning; Springer: Berlin, Germany, 2006. [Google Scholar]
  39. Iruansi, O.; Guadagnini, M.; Pilakoutas, K.; Neocleous, K. Predicting the shear resistance of RC beams without shear reinforcement using a Bayesian neural network. Int. J. Reliab. Saf. 2012, 6, 82–109. [Google Scholar] [CrossRef]
  40. Kocadağlı, O.; Aşıkgil, B. Nonlinear time series forecasting with Bayesian neural networks. Expert Syst. Appl. 2014, 41, 6596–6610. [Google Scholar] [CrossRef]
  41. Sohn, H.; Farrar, C.R.; Hernez, F.M.; Czrnecki, J.J.; Shunk, D.D.; Stinemates, D.W.; Nadler, B.R. A Review of Structural Health Monitoring Literature: 1996–2001; Report No. LA-13976-MS; Los Alamos National Laboratory: Los Alamos, NM, USA, 2004. [Google Scholar]
  42. Cybenko, G. Approximation by superpositions of a sigmoidal function. Math. Control Signal 1989, 2, 303–314. [Google Scholar] [CrossRef]
  43. Beck, J.L.; Yuen, K.V. Model selection using response measurement: A Bayesian probabilistic approach. J. Eng. Mech. 2004, 130, 192–203. [Google Scholar] [CrossRef]
  44. Neural Network Toolbox User’s Guide; Mathworks Inc.: Natick, MA, USA, 2016.
  45. Yin, T.; Zhu, H.P.; Fu, S.J. Damage identification of periodically-supported structures following the Bayesian probabilistic approach. Int. J. Struct. Stab. Dyn. 2018. [Google Scholar] [CrossRef]
Figure 1. Steel truss bridge model.
Figure 1. Steel truss bridge model.
Sensors 18 03371 g001
Figure 2. Configuration of ten measurement points on the bridge deck.
Figure 2. Configuration of ten measurement points on the bridge deck.
Sensors 18 03371 g002
Figure 3. First six mode shapes: (a) Mode 1 (74.91 Hz); (b) Mode 2 (83.67 Hz); (c) Mode 3 (221.13 Hz); (d) Mode 4 (240.72 Hz); (e) Mode 5 (312.06 Hz); (f) Mode 6 (370.63 Hz).
Figure 3. First six mode shapes: (a) Mode 1 (74.91 Hz); (b) Mode 2 (83.67 Hz); (c) Mode 3 (221.13 Hz); (d) Mode 4 (240.72 Hz); (e) Mode 5 (312.06 Hz); (f) Mode 6 (370.63 Hz).
Sensors 18 03371 g003
Figure 4. Configuration of element groups for generating training data.
Figure 4. Configuration of element groups for generating training data.
Sensors 18 03371 g004
Figure 5. Iteration history of the model class selection procedure.
Figure 5. Iteration history of the model class selection procedure.
Sensors 18 03371 g005
Figure 6. Logarithm value of likelihood factor, Ockham factor and evidence.
Figure 6. Logarithm value of likelihood factor, Ockham factor and evidence.
Sensors 18 03371 g006
Figure 7. Damage configurations for the truss bridge model: (a) Case 1 (E3); (b) Case 2 (E3 and E14); (c) Case 3 (E3 and E14 and E19 and E20).
Figure 7. Damage configurations for the truss bridge model: (a) Case 1 (E3); (b) Case 2 (E3 and E14); (c) Case 3 (E3 and E14 and E19 and E20).
Sensors 18 03371 g007
Figure 8. Probability of damage: (a1)−(e1) Case 1; (a2)−(e2) Case 2; (a3)−(e3) Case 3.
Figure 8. Probability of damage: (a1)−(e1) Case 1; (a2)−(e2) Case 2; (a3)−(e3) Case 3.
Sensors 18 03371 g008
Table 1. Sectional and material properties of the steel truss bridge model.
Table 1. Sectional and material properties of the steel truss bridge model.
Parameter NameValues
Young’s modulus2.06 × 1011 N/m2
Mass density7.85 × 103 kg/m3
Cross-sectional area (Angle iron)0.47 × 10−3 m2
Cross-sectional area (I-steel)1.57 × 10−3 m2
Moment of inertia (Angle iron)0.11 × 10−6 m4
Moment of inertia (I-steel)2.21 × 10−6 m4
Bridge span2.8 m
Bridge width0.48 m
Bridge height0.4 m
Table 2. Definition of element groups for the steel truss bridge model.
Table 2. Definition of element groups for the steel truss bridge model.
Element GroupsNumber of ElementsIncluded Elements
EG14{E1, E2, E11, E12}
EG24{E3, E4, E13, E14}
EG34{E5, E6, E15, E16}
EG44{E7, E8, E17, E18}
EG54{E9, E10, E19, E20}
Table 3. The results of model class selection for the Bayesian neural network.
Table 3. The results of model class selection for the Bayesian neural network.
Transfer FunctionsNumber of Hidden Neurons NHLogarithm of
EvidenceLikelihood FactorOckham Factor
TF1 (tansig)81172.421724.16−551.74
91187.231780.93−593.70
101197.861849.99−652.12
111199.741905.59−705.85
121199.931986.27−786.34
131195.722037.02−841.30
141186.622083.90−897.28
151178.152183.16−1005.01
TF2 (satlins)1560.74588.01−27.27
2584.43633.46−49.03
3625.17705.89−80.73
4754.87895.72−140.84
51002.491265.83−263.34
6999.301280.24−280.94
7995.601290.60−295.00
8993.111302.23−309.12
Table 4. Damage cases considered for the steel truss bridge model.
Table 4. Damage cases considered for the steel truss bridge model.
CasesDamaged Elements (Stiffness Reduction)Element Groups
Case 1E3 (30%)EG2
Case 2E3 (30%) & E14 (50%)EG2
Case 3E3 (30%) & E14 (50%) & E19 (30%) & E20 (50%)EG2 & EG5
Table 5. Prediction results of the Bayesian neural network.
Table 5. Prediction results of the Bayesian neural network.
Network OutputCase 1Case 2Case 3
IdentifiedSTD ValueIdentifiedSTD ValueIdentifiedSTD Value
t1 (EG1)0.00950.060.00520.06−0.00490.10
t2 (EG2)0.14080.140.29380.150.26890.29
t3 (EG3)0.01750.050.01610.060.01820.10
t4 (EG4)0.05230.100.02870.12−0.08120.20
t5 (EG5)0.00470.060.00660.070.42020.10

Share and Cite

MDPI and ACS Style

Yin, T.; Zhu, H.-p. Probabilistic Damage Detection of a Steel Truss Bridge Model by Optimally Designed Bayesian Neural Network. Sensors 2018, 18, 3371. https://doi.org/10.3390/s18103371

AMA Style

Yin T, Zhu H-p. Probabilistic Damage Detection of a Steel Truss Bridge Model by Optimally Designed Bayesian Neural Network. Sensors. 2018; 18(10):3371. https://doi.org/10.3390/s18103371

Chicago/Turabian Style

Yin, Tao, and Hong-ping Zhu. 2018. "Probabilistic Damage Detection of a Steel Truss Bridge Model by Optimally Designed Bayesian Neural Network" Sensors 18, no. 10: 3371. https://doi.org/10.3390/s18103371

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop