Structural Damage Features Extracted by Convolutional Neural Networks from Mode Shapes

This paper aims to locate damaged rods in a three-dimensional (3D) steel truss and reveals some internal working mechanisms of the convolutional neural network (CNN), which is based on the first-order modal parameters and CNN. The CNN training samples (including a large number of damage scenarios) are created by ABAQUS and PYTHON scripts. The mode shapes and mode curvature differences are taken as the inputs of the CNN training samples, respectively, and the damage locating accuracy of the CNN is investigated. Finally, the features extracted from each convolutional layer of the CNN are checked to reveal some internal working mechanisms of the CNN and explain the specific meanings of some features. The results show that the CNN-based damage detection method using mode shapes as the inputs has a higher locating accuracy for all damage degrees, while the method using mode curvature differences as the inputs has a lower accuracy for the targets with a low damage degree; however, with the increase of the target damage degree, it gradually achieves the same good locating accuracy as mode shapes. The features extracted from each convolutional layer show that the CNN can obtain the difference between the sample to be classified and the average of training samples in shallow layers, and then amplify the difference in the subsequent convolutional layer, which is similar to a power function, finally it produces a distinguishable peak signal at the damage location. Then a damage locating method is derived from the feature extraction of the CNN. All of these results indicate that the CNN using first-order modal parameters not only has a powerful damage location ability, but also opens up a new way to extract damage features from the measurement data.


Introduction
Structural damage detection, as an important part of structural health monitoring (SHM), aims to timely and accurately find out structural problems so that corresponding maintenances can be taken to ensure the safety and reliability of a structure during its service period. At present, the commonly used damage detection methods include appearance inspection, sample surveys, ultrasonic wave and X-ray, which are used to detect local internal and external defects of structures [1]; there are also other methods like static test methods, which are used to obtain static parameters of structures, such as structural stiffness, deflection, strain and moment of inertia. These detecting methods not only need a large amount of manpower and material resources, but also require a high professional skills and vast engineering experience of detection, in addition to the long detecting cycle and subjective assessment [2]. Dynamic detection methods based on structural modal parameters become a research hotspot in recent In this paper, the modal data of a steel truss with various damage scenarios are obtained through finite element simulations, they are used as the inputs of a CNN to train the network and investigate the prediction accuracy of the CNN. Meanwhile, the output of each convolutional layer is extracted to study the internal working mechanism of the CNN.

Finite Element Simulations
This paper uses a steel truss as the structural model ( Figure 1); its length, width and height are 9.912 m, 0.354 m, and 0.354 m. Its left and right ends are simply supported. The truss consists of 355 rods; each rod has a hollow circular cross section with an external radius of 0.005 m and a thickness of 0.002 m. The Young's modulus, Poisson's ratio, density, and modal damping of the model are 211 GPa, 0.288, 7800, and 0.02 kg/m 3 , respectively.
The finite element software package, ABAQUS (SIMULIA Inc., Providence, RI, USA), is used to obtain the modal parameters of the steel truss, while PYTHON scripts are used to prepare the input files of the finite element analyses for a variety of designed damage scenarios and submit the analyses to ABAQUS solver. The first order mode shapes and the mode curvature differences are used respectively to train the CNN. There are 16 damage degrees and 355 damage locations used in this study. The structural damage is defined as changes to the elastic modulus of the rod concerned. It is assumed that the damage degree of a rod has a linear relation with the reduction of its elastic modulus, so that this paper uses 16 damage degrees, the elastic modulus of which is respectively reduced by 0%, 2%, 4%, 5%, 6%, 8%, 10%, 20%, 30%, 40%, 50%, 60%, 70%, 80%, 90%, 95% from the intact state, while for each damage degree, there are 355 damage locations to represent the damage in each of the 355 rods respectively, hence, there are a total of 5326 damage scenarios. This paper only considers single damage scenarios, i.e., each scenario has only one rod damaged.

Training Samples
The input data of the CNN training samples are derived from the vertical displacement of the rods on the front lower edge of the truss (the red line in Figure 1). There are 28 rods along this line; each rod has 8 observation points, the observation points at both ends of the adjacent rods coincide with each other, so there are a total of 197 observation points, the vertical displacement in the first order mode of which are obtained to form the mode shapes samples. The mode shape curvatures can be calculated by the following formula: where k φ is the vertical displacement of the k-th point, and = 0.05 m l Δ , it is the distance between two adjacent points. The mode curvature difference is the difference between the curvatures of a damage scenario and the intact scenario. Thus, each mode curvature difference sample contains 195 The finite element software package, ABAQUS (SIMULIA Inc., Providence, RI, USA), is used to obtain the modal parameters of the steel truss, while PYTHON scripts are used to prepare the input files of the finite element analyses for a variety of designed damage scenarios and submit the analyses to ABAQUS solver. The first order mode shapes and the mode curvature differences are used respectively to train the CNN.
There are 16 damage degrees and 355 damage locations used in this study. The structural damage is defined as changes to the elastic modulus of the rod concerned. It is assumed that the damage degree of a rod has a linear relation with the reduction of its elastic modulus, so that this paper uses 16 damage degrees, the elastic modulus of which is respectively reduced by 0%, 2%, 4%, 5%, 6%, 8%, 10%, 20%, 30%, 40%, 50%, 60%, 70%, 80%, 90%, 95% from the intact state, while for each damage degree, there are 355 damage locations to represent the damage in each of the 355 rods respectively, hence, there are a total of 5326 damage scenarios. This paper only considers single damage scenarios, i.e., each scenario has only one rod damaged.

Training Samples
The input data of the CNN training samples are derived from the vertical displacement of the rods on the front lower edge of the truss (the red line in Figure 1). There are 28 rods along this line; each rod has 8 observation points, the observation points at both ends of the adjacent rods coincide with each other, so there are a total of 197 observation points, the vertical displacement in the first order mode of which are obtained to form the mode shapes samples. The mode shape curvatures can be calculated by the following formula: where φ k is the vertical displacement of the k-th point, and ∆l= 0.05 m, it is the distance between two adjacent points. The mode curvature difference is the difference between the curvatures of a damage scenario and the intact scenario. Thus, each mode curvature difference sample contains 195 data.
To ensure these two kinds of samples have the same dimension, the first and the last datum of the mode shape samples are removed. Therefore, each sample is a 1 × 195 row vector and is arranged into a 15 × 13 matrix . Since there are 16 damage degrees and 355 damage locations, there are 5326 samples  for each index. The samples for damage degrees of 0%, 2%, 5%, 8%, 20%, 40%, 60%, 80%, and 95% were selected as the training set and the samples for damage degrees of 0%, 4%, 6%, 10%, 30%, 50%, 70%, and 90% were selected as the test set.

Structural Damage Detection
In this paper, the classification function of the CNN is used to locate damaged rods. As a different damage scenario will produce a different structural mode, therefore, the classification function of the CNN can be used to classify different damage scenario by matching the mode samples and damage scenarios one to one, so as to realize the detection of damage degree and location.
The CNN input is the modal parameters of different damage scenarios, while the CNN output is the label of different damage scenarios. That is, each damage scenario is marked with a label, starting with the damage of 2% in Rod 1 being marked as Label 1, then the damage of 2% in Rod 2 will be marked as Label 2, and so on. The label number increases consecutively for next damage degree, for example, the damage of 4% in Rod 1 will be marked as Label 356. In the end, the damage of 95% in Rod 355 is marked as Label 5325. Label 5326 is for the intact truss.
The whole training process is divided into training phase and testing phase. Firstly, the CNN conducts network training according to the one-to-one input and output relationship. The training is mainly to update the CNN weights and reduce the error between the actual outputs and the target outputs. When the training phase is completed, the trained network can be saved and the test samples are inputted to the network. By comparing the target outputs with the actual outputs of the network, the prediction accuracy of damage location can be obtained. Figure 2 is the seventh span of the steel truss, which shows the naming convention of the rods.

Structural Damage Detection
In this paper, the classification function of the CNN is used to locate damaged rods. As a different damage scenario will produce a different structural mode, therefore, the classification function of the CNN can be used to classify different damage scenario by matching the mode samples and damage scenarios one to one, so as to realize the detection of damage degree and location.
The CNN input is the modal parameters of different damage scenarios, while the CNN output is the label of different damage scenarios. That is, each damage scenario is marked with a label, starting with the damage of 2% in Rod 1 being marked as Label 1, then the damage of 2% in Rod 2 will be marked as Label 2, and so on. The label number increases consecutively for next damage degree, for example, the damage of 4% in Rod 1 will be marked as Label 356. In the end, the damage of 95% in Rod 355 is marked as Label 5325. Label 5326 is for the intact truss.
The whole training process is divided into training phase and testing phase. Firstly, the CNN conducts network training according to the one-to-one input and output relationship. The training is mainly to update the CNN weights and reduce the error between the actual outputs and the target outputs. When the training phase is completed, the trained network can be saved and the test samples are inputted to the network. By comparing the target outputs with the actual outputs of the network, the prediction accuracy of damage location can be obtained. Figure 2 is the seventh span of the steel truss, which shows the naming convention of the rods.

CNN Architecture
The architecture of the CNN (Figure 3) is designed in MATLAB (MathWorks Inc., Natick, MA, USA). It mainly includes an input layer, four convolutional layers, a fully-connected layer, a softmax layer and an output layer.

CNN Architecture
The architecture of the CNN (Figure 3) is designed in MATLAB (MathWorks Inc., Natick, MA, USA). It mainly includes an input layer, four convolutional layers, a fully-connected layer, a softmax layer and an output layer.
The convolution process is shown in Figure 4. A fixed-size convolution kernel is used to multiply the corresponding elements of each sub-region of the input matrix then sum them up to be an element in the new matrix. Then the convolution kernel slides with a fixed stride (the stride in Figure 4 is 2); the process is repeated until all elements of the input elements are involved; finally, it forms a new matrix and then it will be mapped nonlinearly by the activation function and normalized in the batch normalization layer. The convolution process is shown in Figure 4. A fixed-size convolution kernel is used to multiply the corresponding elements of each sub-region of the input matrix then sum them up to be an element in the new matrix. Then the convolution kernel slides with a fixed stride (the stride in Figure 4 is 2); the process is repeated until all elements of the input elements are involved; finally, it forms a new matrix and then it will be mapped nonlinearly by the activation function and normalized in the batch normalization layer. After the input data passes through the four convolutional layers, it is fully expanded into a one-dimensional column vector by the flatten layer and then enters the fully-connected layer. The probability that a sample belongs to a target class is calculated by the softmax layer. The softmax is a function of the following form:  The convolution process is shown in Figure 4. A fixed-size convolution kernel is used to multiply the corresponding elements of each sub-region of the input matrix then sum them up to be an element in the new matrix. Then the convolution kernel slides with a fixed stride (the stride in Figure 4 is 2); the process is repeated until all elements of the input elements are involved; finally, it forms a new matrix and then it will be mapped nonlinearly by the activation function and normalized in the batch normalization layer. After the input data passes through the four convolutional layers, it is fully expanded into a one-dimensional column vector by the flatten layer and then enters the fully-connected layer. The probability that a sample belongs to a target class is calculated by the softmax layer. The softmax is a function of the following form: After the input data passes through the four convolutional layers, it is fully expanded into a one-dimensional column vector by the flatten layer and then enters the fully-connected layer. The probability that a sample belongs to a target class is calculated by the softmax layer. The softmax is a function of the following form: where P(i) is the probability that a sample belongs to the i-th class, x is the input in a column vector form, n is the number of classes, and θ is the undetermined coefficients in a column vector form.
When the sample finally enters the class output layer, the sample will be classified into the category with the highest probability.

First Order Mode Shape as CNN Input
Through many trial calculations, it is found that the damage locating accuracy becomes higher as the convolutional layer increases. Figure 5 shows the training process for the convolutional layers being 1, 2, 3 and 4, respectively. Finally, the network having four convolutional layers is used. The kernel size of each layer is 3 × 3, 2 × 2, 1 × 1 and 1 × 1, respectively and the stride is 1; no pooling layers are used.
where ( ) P i is the probability that a sample belongs to the i-th class, x is the input in a column vector form, n is the number of classes, and θ is the undetermined coefficients in a column vector form.
When the sample finally enters the class output layer, the sample will be classified into the category with the highest probability.

First Order Mode Shape as CNN Input
Through many trial calculations, it is found that the damage locating accuracy becomes higher as the convolutional layer increases. Figure 5 shows the training process for the convolutional layers being 1, 2, 3 and 4, respectively. Finally, the network having four convolutional layers is used. The kernel size of each layer is 3 × 3, 2 × 2, 1 × 1 and 1 × 1, respectively and the stride is 1; no pooling layers are used. As shown in Figure 5, the CNN reaches a high accuracy after 1000 iterations. As for the test sets of different damage scenarios, the prediction accuracies are shown in Table 1: As shown in Figure 5, the CNN reaches a high accuracy after 1000 iterations. As for the test sets of different damage scenarios, the prediction accuracies are shown in Table 1: Table 1 shows that a damage scenario can be distinguished from the intact state by the CNN trained with the first order mode shape as the CNN input, the damage prediction accuracy reaches 79% in low damage degree (4%), and increases to the maximum value of 87% for the damage degree of 70%. As for the rods around the observation point, such as the front lower rods, the front upper rods, the front vertical rods and the front diagonal rods, the prediction accuracy can reach 100% when the damage degree is only 4%.

Mode Curvature Difference as CNN Input
Because the mode curvature difference is more distinguishable at the damage location than the mode shape, it is used as the CNN input and the locating accuracy is investigated.
Through many trial calculations, it is found that the increase of the convolutional layers has slight effect on the training results. Figure 6 shows the training process for the convolutional layers being 1, 2, 3, and 4, respectively, the stride is 1 and there are no pooling layers. Finally, the CNN with only one convolutional layer is selected for damage detection and the convolution kernel size is 1 × 1. As shown in Figure 6, the network reached a high accuracy after 100 iterations. As for its test sets, the prediction accuracies are shown in Table 2: Table 2. Prediction accuracies of the mode curvature differences.

Type
Damage Degree 0% 4% 6% 10% 30% 50% 70% 90% Figure 6. Training processes. Finally, the CNN with only one convolutional layer is selected for damage detection and the convolution kernel size is 1 × 1. As shown in Figure 6, the network reached a high accuracy after 100 iterations. As for its test sets, the prediction accuracies are shown in Table 2:  Table 2 shows that, selecting the first order mode curvature difference as the CNN input is unable to correctly distinguish the intact and damage scenarios sometimes. Its damage prediction accuracy is only 33% in the low damage degree (4%), but the prediction accuracy increases with the increase of damage degree, the accuracy reaches the maximum value of 90% when the damage degree increases to 50%. For the rods around the observation points, such as the front lower rod, the front upper rod and the front diagonal rod, the damage prediction accuracy is also significantly higher than other rods.

Sample Features
Some first order mode shape samples are shown in Figure 7. There are 14 curves of different damage scenarios in Figure 7 and these curves have different colors; but they are so similar that they look like only one curve. Some first order mode curvature difference samples are shown in Figure 8.  Table 2 shows that, selecting the first order mode curvature difference as the CNN input is unable to correctly distinguish the intact and damage scenarios sometimes. Its damage prediction accuracy is only 33% in the low damage degree (4%), but the prediction accuracy increases with the increase of damage degree, the accuracy reaches the maximum value of 90% when the damage degree increases to 50%. For the rods around the observation points, such as the front lower rod, the front upper rod and the front diagonal rod, the damage prediction accuracy is also significantly higher than other rods.

Sample Features
Some first order mode shape samples are shown in Figure 7. There are 14 curves of different damage scenarios in Figure 7 and these curves have different colors; but they are so similar that they look like only one curve. Some first order mode curvature difference samples are shown in Figure 8.    Figure 7 shows that, the first order mode shapes of various damage scenarios are almost the same, they are not distinguishable for naked eyes; on the other hand, the mode curvature differences show evident fluctuations in the damage locations ( Figure 8).
As for the input data having distinguishable features, such as the mode curvature differences, the CNN can locate the damaged rods using the shallow learning rather than using the deep learning, as shown in Figure 6, the prediction accuracy is no longer improved when the layer number increases. While if the input data have no obvious fluctuations at the damage locations, such as the first order mode shapes, it requires deep learning to achieve a higher prediction accuracy, as shown in Figure 5, the prediction accuracy is gradually improved as the layer number increases.

Features Extracted by the CNN
The function "activations" in the MATLAB Deep Learning Toolbox is used to get the features extracted by each convolutional layer. Figure 9 shows the output data of 16 channels of each layer, they represent the features extracted from the 3202-th damage scenario (the front lower rod in the seventh span with a damage of 50%) by the CNN. For all the curves, the x-coordinate is the location and the y-coordinate is its amplitude.  Figure 7 shows that, the first order mode shapes of various damage scenarios are almost the same, they are not distinguishable for naked eyes; on the other hand, the mode curvature differences show evident fluctuations in the damage locations ( Figure 8).
As for the input data having distinguishable features, such as the mode curvature differences, the CNN can locate the damaged rods using the shallow learning rather than using the deep learning, as shown in Figure 6, the prediction accuracy is no longer improved when the layer number increases. While if the input data have no obvious fluctuations at the damage locations, such as the first order mode shapes, it requires deep learning to achieve a higher prediction accuracy, as shown in Figure 5, the prediction accuracy is gradually improved as the layer number increases.

Features Extracted by the CNN
The function "activations" in the MATLAB Deep Learning Toolbox is used to get the features extracted by each convolutional layer. Figure 9 shows the output data of 16 channels of each layer, they represent the features extracted from the 3202-th damage scenario (the front lower rod in the seventh span with a damage of 50%) by the CNN. For all the curves, the x-coordinate is the location and the y-coordinate is its amplitude.
Because the convolution kernel size greater than 1 × 1 will decrease the data dimension, which makes the data difficult to compare, thus the kernels of the four convolutional layers are re-sized to 1 × 1 in the following context and the CNN is re-trained. The prediction accuracy reduced from 84% to 74% after changing the convolution kernel sizes. Figure 10 shows the features extracted from the 3202-th damage scenario by the CNN after the change of the convolution kernel size. As shown in Figures 9 and 10, the features extracted by the two CNNs have certain similarity, especially in the shallow layers, for example, the features marked by red circles in Figures 9 and 10 are very similar to each other. They are normalized and re-drawn in Figures 11 and 12. Because the convolution kernel size greater than 1 × 1 will decrease the data dimension, which makes the data difficult to compare, thus the kernels of the four convolutional layers are re-sized to 1 × 1 in the following context and the CNN is re-trained. The prediction accuracy reduced from 84% to 74% after changing the convolution kernel sizes. Figure 10 shows the features extracted from the 3202-th damage scenario by the CNN after the change of the convolution kernel size. As shown in Figures 9 and 10, the features extracted by the two CNNs have certain similarity, especially in the shallow layers, for example, the features marked by red circles in Figure 9 and 10 are very similar to each other. They are normalized and re-drawn in Figure 11 and 12. Firstly, try to explain the specific meaning of the feature extracted from the first convolutional layer (Figure 12). It is re-drawn on Figure 13 (Curve 1). At the same time, the difference between the average of the training samples and the sample of the 3202-th scenario is calculated and the normalized result is shown in Figure 13 (Curve 2), then the difference between the sample of intact scenario and the sample of the 3202-th scenario is calculated and the normalized result is shown in Figure 13 (Curve 3).
These figures show that the red-circled feature is very similar to the difference between the average of training samples and the sample of the 3202-th scenario, especially at the peak position, they almost overlap each other, but there are still some deviations in other position. The red-circled feature and the difference between the intact scenario and the 3202-th scenario show a better fitting, they almost overlap each other in all positions.
Secondly, the feature marked by green-circled in the third layer is considered. The normalized result is shown in Figure 14 (Curve 1), at the same time, the features extracted by the first convolutional layer (Figure 12) are mapped by a quadratic function and the result is shown in Figure 14 (Curve 2). The comparison shows that the two curves are almost identical, especially at the peak position.    The features in Figure 15 are extracted from the convolutional layers after the first order mode shape samples of other damage scenarios are inputted into the CNN. The features of the top row are extracted from the first convolutional layer for various damage scenarios and the features of the bottom row are from the third convolutional layer. All these features have the same performance as above, the first layer features are similar to the difference between the sample to be classified and the average of training samples, and the third layer features are similar to the mapping of the first layer features via a power function.

264
Firstly, try to explain the specific meaning of the feature extracted from the first convolutional 265 layer ( Figure 12). It is re-drawn on Figure 13

264
Firstly, try to explain the specific meaning of the feature extracted from the first convolutional 265 layer ( Figure 12). It is re-drawn on Figure 13 (Curve 1). At the same time, the difference between the 266 average of the training samples and the sample of the 3202-th scenario is calculated and the 267 normalized result is shown in Figure 13

287
Curve 2 is the square of the feature extracted from the first layer.

288
The features in Figure 15 are extracted from the convolutional layers after the first order mode     Through comprehensive observations of the features extracted from the first to the third layers, it can be found that the input samples produce distinguishable peak features at the damage location after three convolutional layers; the normalized result is very similar to the square of the difference between the sample to be classified and the average of training samples.

Discussions and Conclusions
This paper uses the CNN method to locate the damaged rod of a complex steel truss based on the first order mode shapes and mode curvature differences. It shows the prediction accuracies for a variety of damage scenarios and reveals some internal working mechanisms of the network. Sections 3.1 and 3.2 show that the CNN method based on the first order mode shape has a higher accuracy when locating the single damaged rod in all damage scenarios, even if the damage degree is only 4%, the prediction accuracies of some rods reach 100%, while the method based on the mode curvature difference can achieve the same prediction accuracy in locating the rod with a high damage degree, but the prediction accuracy is poor when locating the rod with a low damage degree. Section 3.3. shows the features of the two kinds of input data, compared with the first order mode shape (unprocessed data), the curvature difference (processed data) has obvious numerical variations at the damage location, but it does not have a better accuracy as expected, the central difference method may result in the loss of some information, while the original data like the first order mode shape may contain more damage information and the damage detection accuracy based on it is higher. Furthermore, to extract the inconspicuous features from the data, the CNN needs a deep learning architecture to get more powerful features extraction ability, hence, the CNN, using the first order mode shape as the inputs, needs a deep learning architecture to achieve higher accuracy; while the CNN, using the mode curvature difference as the inputs, only needs a shallow learning architecture to achieve its maximum prediction accuracy. However, even for the method based on the first order mode shape, there are always some rods with poor prediction accuracy. The reason may be that, only the vertical displacements of the truss are used as the inputs, they may not contain enough damage information.
By observing the features extracted from the mode shapes, it can be seen that the CNN can obtain the difference between the sample to be classified and the weighted average of the training samples in shallow layers, and then amplify the difference in the subsequent convolutional layer. Thus, it can be inferred that the partial work of the CNN to locate a damaged rod is to render the input data remarkable at the damage segment through updating the internal weights of the network, it appears like the result of calculating the difference between the sample to be classified and the weighted average of the training samples and then map it via a quadratic function. After the calculation of multiple convolutional layers, the mode shapes will have a distinguishable peak feature in the damaged segment like the curvature differences. This kind of data processing is very similar to the existing damage indexes based on the mode curvature differences, mode shape coefficient [21] and mode shape difference function [22], which were used traditionally to overcome the disadvantages of low-order mode shapes.
As the Figure 9 shows, there are certainly many other features extracted by the CNN, but their specific meanings are to be explored. The work above can only identify the damage existing in which segment; however, the CNN should be able to locate the damage existing in which rod. Besides, this paper ignored the function of the large size convolution kernel in the feature extractions; therefore, what the CNN does inside still needs a further exploration.
A neural network is a method simulating the biological neuron through mathematical algorithm model. Through the previous discussions, it can be found that it does something very similar to the human mind in the process of damage locating. By observing the features extracted by the CNN, a simple method of damage location can be derived, that is to calculate the difference between the sample to be classified and the sample of intact scenario to get the mode shape difference function and then map it via a power function to magnify the fluctuation feature, so as to produce a distinguishable peak in the damaged segment and finally locate the damaged segment. The formula is as follows: f (x) = mode shape difference function φ = mode shape of the intact structure φ d = mode shape of the damaged structure x = spatial location n = power factor From the above discussion, it can be concluded that: 1. the CNN method based on the mode shape has a good prediction accuracy of damaged rods, but there are always some rods with low prediction accuracy. Due to the different sensitivity of modal data to the damage of different types of rods, their prediction accuracies are also different; 2.
the CNN method based on the mode curvature difference has similar good prediction accuracy to the mode shape when locating the rod with high damage degree. However, the prediction accuracy is low when locating the rods with small damage degree. Like the mode shape, there are always some rods with low prediction accuracy due to the different sensitivity of modal data to the damage of different rods; 3.
the origin data has more damage information for the CNN. Though the first order mode shapes look indistinguishable, the multi-layer CNN can extract the damage information from the data, which are indistinguishable for naked eyes; 4.
using the first order mode shape as the input of the CNN to locate the damaged rod, the partial work of the CNN is to update its internal weights to get the difference between the input sample and the weighted average of the training samples in shallow layers, then the data will produce a distinguishable peak feature in the damaged segment after processed by multiple convolutional layers behind. This CNN process is similar to calculating the difference between the sample to be classified and the weighted average of the training samples and then mapping it via a quadratic function.