Random ‒ Forest Machine Learning Approach for High ‒ Speed Railway Track Slab Deformation Identification Using Track-Side Vibration Monitoring

Featured Application: The work can be applied to identification of defects based on data-driven methods of structures. Abstract: High ‒ speed railways (HSRs) are established all over the world owing to their advantages of high speed, ride comfort, and low vibration and noise. A ballastless track slab is a crucial part of the HSR, and its working condition directly affects the safe operation of the train. With increasing train operation time, track slabs suffer from various defects such as track slab warping and arching as well as interlayer disengagement defect. These defects will eventually lead to the deformation of track slabs and thus jeopardize safe train operation. Therefore, it is important to monitor the condition of ballastless track slabs and identify their defects. This paper proposes a method for monitoring track slab deformation using fiber optic sensing technology and an intelligent method for identifying track slab deformation using the random ‒ forest model. The results show that track ‒ side monitoring can effectively capture the vibration signals caused by train vibration, track slab deformation, noise, and environmental vibration. The proposed intelligent algorithm can identify track slab deformation effectively, and the recognition rate can reach 96.09%. This paper provides new methods for track slab deformation monitoring and intelligent identification.


Introduction
High-speed railways (HSRs) are flourishing in various countries owing to their efficiency, convenience, comfort, and low maintenance and repair workload [1].By the end of 2020, China's HSR mileage reached 38,000 km [2], ranking first in the world.A ballastless track slab is an important part of various types of track structures, as it plays a "top-down" role (i.e., bearing the train load and transferring it down to the foundation) [3].However, prolonged operation could cause various defects in track slabs, such as warping deformation of track slab [4,5], wide and narrow joint damages [6,7], interlayer debonding [3,8,9], and arching deformation of track slab [10][11][12].Under temperature load [13,14] and high-frequency train load [14], these damages may cause track slab deformation and abnormal vibrations in wheel-rail system.Considering the complex external factors, the performance of track slab is spatially and temporally variable, so the occurrence of track slab damage is sudden and unpredictable.Timely detection of track slab defects prevents track slab failure and ensures safe train operation.Therefore, the real-time monitoring of track slab performance and the identification of the damage condition of track slabs are essential for long-term safe operation of the HSR.Existing track slab defect detection methods can be divided into three categories: visual inspection [15,16], nondestructive testing (NDT) [4,11,[17][18][19][20][21][22], and structural health monitoring (SHM) [23][24][25][26][27].Most of the visual inspection methods rely on a variety of track inspection equipment, such as road rulers, gauges, and various customized equipment.Although visual inspection methods are flexible and simple, they have limitations that most of them require working on the track structure and must be conducted in skylight.HSR skylight time is short and it is difficult to detect the track structure damage in a short time by using this method.
By contrast, NDT methods mainly use noncontact measurement to detect track defects.Some scholars use the differences between rail inspection vehicles' static and dynamic inspection data [17] and the geometric data [4,11] to detect track slab deformation.The track inspection data are derived from various sensing devices installed on the track inspection car.However, owing to the high cost and limited number of track inspection vehicles, track inspection vehicle measurement cycle of approximately half a month can be conducted, resulting in a long detection cycle and a large number of missing effective data.In addition, according to the field survey results of railway management department, track slab deformation mostly occurs during daytime and tends to be stable at night.However, most of the track inspection data in track slab deformation detection methods are collected at night [28]; thus, the information during the day, which is significant for the safe operation of the track, is omitted.Therefore, such track slab deformation detection methods have limitations.In addition, some scholars used the impact-echo waveform [18][19][20] and ground penetrating radar [21,22] methods to detect track slab defects.However, these methods require on-site inspection by technicians for a specific section of track, resulting in delayed detection of damage, which not only takes up skylight time but also affects the normal operation of high-speed trains.Therefore, there is an urgent need for a real-time monitoring method for track slab defects that can monitor without affecting the normal operation of trains.
Therefore, with the aid of advanced sensing technology, real-time monitoring methods for track structure defects have been developed.Some approaches used fiber optic sensors for the real-time monitoring of track structure deformation caused by temperature [23], and for measuring the lateral displacement of the track structure [24].Another approach used distributed fiber optics to monitor and locate cracks in the track slab [25] as well as to monitor the load deflection behavior of the track [26] and the arching deformation of the track slab [27].This paper proposes a method for monitoring track slab deformation using fiber optic sensing technology and the results show that track-side monitoring can effectively capture the vibration signals caused by train vibration, track slab deformation, noise, and environmental vibration.The complex and weak information of track slab deformation hidden in the track-side monitoring data makes it difficult to identify the track defects from the measurement data that contain a lot of redundant information.As the track monitoring data in the railway field include information of train vibration, track deformation, noise, and environmental vibration, it is challenging to extract the desired track slab defect information from the complex signals.
Feature extraction of track slab deformation is crucial to accurate identification.Wavelet analysis is considered to be an effective tool with strong analysis capability for signal localization, but the basis function needs to be defined in advance, and the choice of the basis function directly affects the analysis performance.Thus, the analysis method lacks adaptability [29].To address the shortcomings of the traditional time-frequency analysis method, Huang et al. proposed empirical mode decomposition (EMD) [30], which is widely used in the field of rail damage feature extraction [31], track irregularity feature extraction [32,33], cement asphalt (CA) mortar layer damage feature extraction [34], and other track defect feature extraction.Although this method has the ability to decompose the signal adaptively, its algorithm itself suffers from modal mixing, endpoint effect, and negative frequency problem.Therefore, to avoid these problems, the local mean decomposition (LMD) method [35] was proposed.This method has extensive application in the field of bearing fault identification [36][37][38].However, LMD method for feature extraction in the field of railway track defect identification is rarely reported.
Since damage identification methods based on machine learning do not require a priori knowledge of system dynamics, they not only are applicable to linear mapping systems but also can represent arbitrarily complex nonlinear mapping systems.They also have the advantages of nonparametric methods for damage identification, thus avoiding some of the complications associated with traditional damage identification methods, which must establish a functional relationship between structural response and physical parameters.More importantly, high noise and large uncertainty of the monitoring data in the railway track field make identifying the track damage difficult by constructing a simple indicator; machine learning algorithms have, therefore, been introduced.In recent years, machine learning methods have had many applications in track defect identification and prediction.Machine learning methods were used to detect defects of important components of the track structure that are directly related to safe train operation, including the defects of rails [39,40], fasteners [41][42][43][44], rail pads [45], and turnout [46].The classification of three types of rail defects-surface defect, cross level defect and depression in track profiles-by track geometry data on the basis of logistic regression and decision tree [39] and the classification of rail crack by acoustic emission waves on the basis of a multibranch convolutional neural network (CNN) [40] were explored.The rail fastener defects were detected from images on the basis of a CNN [41], generative adversarial network, residual network [42], point cloud deep learning [43], and Faster region-CNN [44].The dynamic stiffness of rail pads was predicted using several machine learning methodologies (multilinear regression, K nearest neighbors, regression tree, random forest, gradient boosting, multilayer perceptron, and support vector machine (SVM)) [45].The fault detection approach for the HSR turnout based on a deep denoising auto-encoder and one-class SVM [46] was also proposed.Being subjected to repeated temperature load and train load, the track structure suffers from track slab cracks, track slab deformation, CA mortar layer damage, and so on.Although these damages do not necessarily affect the safe train operation directly, they could pose considerable threat to the safe operation if they are not identified in time.For the timely detection of such damages, machine learning methods have been used.The track slab cracks were detected on the basis of the region proposal network and fully convolutional networks [47] and the different severity levels of track slab cracks were studied on the basis of the Inception-ResNet-v2 network [48].The arching deformation of track slabs was detected on the basis of CNN [11] using track geometry data.The damage of CA mortar layer was identified on the basis of SVM by using the acceleration of wheelset [9], the back propagation neural network by using vibration data obtained from modal analysis [49], and CNN by using Hilbert spectrum [34].In summary, intelligent identification techniques based on random-forest model for track slab deformation identification by using track-side vibration acceleration signals acquired from HSRs field have not been sufficiently explored.
This study uses track-side vibration signals for track slab deformation identification without interrupting the normal operation of HSRs.Track-side vibration signals contain information of not only train vibration and track slab deformation but also noise and environmental vibration, which can be eliminated by adopting data preprocessing methods.The train group, running speed, and load of high-speed trains are different, and the wheelrail vibration signal decays layer by layer from the rail through track slab, CA mortar layer, supporting concrete layer and bridge deck slab to reach the cable trough.Therefore, many factors influence track-side vibration signals [50], resulting in many attributes; hence, more parameters are required for using machine learning methods [51].Compared with other traditional models, such as SVMs and neural networks, a random-forest model requires fewer parameters, i.e., only the number of decision trees and split features, in the modeling process regardless of the number of attributes [52].Although the number of parameters of the SVM model (i.e., kernel function parameter and penalty parameter [53]) is the same as that of the random-forest model, in practical applications, different kernel functions and corresponding kernel function parameters must be selected according to the actual situation, and the parameters need to be optimized several times.Neural networks have two types of parameters, weights and thresholds, and the number of parameters increases exponentially with the number of neurons [54].Both the traditional methods require complex parameter optimization algorithms to obtain the appropriate parameters.Owing to the simpler parameter determination process and greater advantages in practical applications of a random-forest model compared with the two traditional methods, we use it in this study for track slab deformation identification.In addition, Ref. [55] has demonstrated through both theoretical and extensive empirical studies that a randomforest model has lower generalization errors, higher prediction accuracies, and better robustness to noise compared with other methods such as SVMs, neural networks, and k-nearest neighbor.In summary, a random-forest model is suitable for the identification of track slab deformation.
The current research on operating HSR track slab deformation identification is summarized as follows: This paper proposes a novel strategy that combines advanced sensing technique based on fiber optic sensing and an intelligent recognition model based on random forest for identifying track slab deformation by using on-site track-side vibration acceleration data.First, an on-site real-time monitoring system for track-side vibration monitoring to facilitate track slab deformation was developed (Section 2).Second, an automatic preprocessing system for data interception, denoising and fixed window segmentation on the raw data was established (Section 3).Then, the feature extraction of time-domain as well as time-frequency-domain of the selected characteristic product functions (PFs) after LMD was proposed (Section 4).Next, the architecture of random-forest model was developed (Section 5).Finally, the model parameters were then optimized and the performance of random-forest model and the comparison decision-tree model were evaluated (Section 6).The study mainly contributes to building an intelligent method that can quickly and accurately detect the deformation of track slabs.The overview of the proposed identification method is shown in Figure 1.
The study mainly contributes to building an intelligent method that can quickly and accurately detect the deformation of track slabs.The overview of the proposed identification method is shown in Figure 1.

Track-Side Real-Time Monitoring for Track Slab Deformation
We established a real-time monitoring system that facilitates track slab deformation detection, as illustrated in Figure 2. The proposed track-side monitoring system consists of eight fiber optic accelerometers, an interrogation based on the phase generation carrier demodulation technique, an optical cable, and a desktop computer.The selected sensors are interferometric fiber optic accelerometers [56] with a suitable bandwidth (0.1-1000 Hz) and high sensitivity (500 rad/g).The sampling frequency of the proposed system is 1000 Hz.The monitoring system was installed in a CRTS II HSR on viaduct in China.The red box in Figure 2 is the cross section of the CRTS II ballastless track structure on viaduct, and the components from top to bottom are the rail, track slab, CA mortar layer, supporting concrete layer, and viaduct.The blue box in the cross section is the cable trough, which is the transversely nearest place to the deformation-prone area in the safety-securing zone allowed for sensor placement.The field image of the cable trough where the sensors were installed is shown in the blue box in Figure 2. The eight sensors in Figure 2 are illustrated by eight blue circles with a spacing of 3.2 m, which is approximately half the length of the track slab (6.45 m).The red lines between the sensors represent the transmission optical cable.The sensor is connected to the demodulator via the optical cable, and the data acquisition and data demodulation are performed by a computer continuously.The deformation of the track slab occurred near a certain sensor in the monitoring section during the monitoring time, so the proposed real-time monitoring system captures the vibration data of the normal and deformed states of track slab.Thus, a solid foundation for the subsequent intelligent identification is laid.

Track-Side Real-Time Monitoring for Track Slab Deformation
We established a real-time monitoring system that facilitates track slab deformation detection, as illustrated in Figure 2. The proposed track-side monitoring system consists of eight fiber optic accelerometers, an interrogation based on the phase generation carrier demodulation technique, an optical cable, and a desktop computer.The selected sensors are interferometric fiber optic accelerometers [56] with a suitable bandwidth (0.1-1000 Hz) and high sensitivity (500 rad/g).The sampling frequency of the proposed system is 1000 Hz.The monitoring system was installed in a CRTS II HSR on viaduct in China.The red box in Figure 2 is the cross section of the CRTS II ballastless track structure on viaduct, and the components from top to bottom are the rail, track slab, CA mortar layer, supporting concrete layer, and viaduct.The blue box in the cross section is the cable trough, which is the transversely nearest place to the deformation-prone area in the safety-securing zone allowed for sensor placement.The field image of the cable trough where the sensors were installed is shown in the blue box in Figure 2. The eight sensors in Figure 2 are illustrated by eight blue circles with a spacing of 3.2 m, which is approximately half the length of the track slab (6.45 m).The red lines between the sensors represent the transmission optical cable.The sensor is connected to the demodulator via the optical cable, and the data acquisition and data demodulation are performed by a computer continuously.The deformation of the track slab occurred near a certain sensor in the monitoring section during the monitoring time, so the proposed real-time monitoring system captures the vibration data of the normal and deformed states of track slab.Thus, a solid foundation for the subsequent intelligent identification is laid.

Preprocessing of Monitoring Data
Since the sensors are installed on the operating HSR, the vibration signals captured by sensors contain the information of train vibration, track slab deformation, noise and environmental vibration.In addition, the train grouping, running speed, and load of different trains are all different.All these factors are manifested in the vibration signals, thus making obtaining track deformation status information from a single monitoring data considerably challenging.Hence, the raw data must be preprocessed in order to accurately identify the deformation state of track slab.Figure 3 illustrates the preprocessing process.First, data interception is performed on the raw data.When the train passes by a sensor, the information captured by the sensor is the wheel-rail vibration signal containing track slab deformation information; when no train passes by, the information captured by the sensor is environmental noise.Therefore, the effective information for the identification of the track slab deformation is the wheel-rail vibration signal when the train passes.

Preprocessing of Monitoring Data
Since the sensors are installed on the operating HSR, the vibration signals captured by sensors contain the information of train vibration, track slab deformation, noise and environmental vibration.In addition, the train grouping, running speed, and load of different trains are all different.All these factors are manifested in the vibration signals, thus making obtaining track deformation status information from a single monitoring data considerably challenging.Hence, the raw data must be preprocessed in order to accurately identify the deformation state of track slab.Figure 3 illustrates the preprocessing process.

Preprocessing of Monitoring Data
Since the sensors are installed on the operating HSR, the vibration signals captured by sensors contain the information of train vibration, track slab deformation, noise and environmental vibration.In addition, the train grouping, running speed, and load of different trains are all different.All these factors are manifested in the vibration signals, thus making obtaining track deformation status information from a single monitoring data considerably challenging.Hence, the raw data must be preprocessed in order to accurately identify the deformation state of track slab.Figure 3 illustrates the preprocessing process.First, data interception is performed on the raw data.When the train passes by a sensor, the information captured by the sensor is the wheel-rail vibration signal containing track slab deformation information; when no train passes by, the information captured by the sensor is environmental noise.Therefore, the effective information for the identification of the track slab deformation is the wheel-rail vibration signal when the train passes.First, data interception is performed on the raw data.When the train passes by a sensor, the information captured by the sensor is the wheel-rail vibration signal containing track slab deformation information; when no train passes by, the information captured by the sensor is environmental noise.Therefore, the effective information for the identification of the track slab deformation is the wheel-rail vibration signal when the train passes.In this regard, we employed a data interception method to derive vibration signals containing the track slab deformation information.The results of data interception are shown in Figure 3.In the data interception process, the vibration acceleration threshold δ is defined.When the vibration acceleration x(t) ≥ δ at time t, the vibration acceleration x(t) at time t is recorded.x(t + 1), x(t + 2), and x(t + 3) are also recorded until the vibration acceleration x(t + n − 1) ≥ δ, x(t + n) ≤ δ, and x(t + n + 1) < δ; then recording is stopped.Thus, a new vibration acceleration set X = [x(t), x(t + 1), x(t + 2), . . .x(t + n − 1)], in which the environmental vibration information is eliminated, is formed.The threshold value in this paper is set to 0.01 g.
Second, the denoising process is performed on the new vibration acceleration set.The signal acquired in the field often comprises noise and requires denoising to effectively display the useful information in the original signal.This paper adopts "sym3" wavelet function for five-layer wavelet threshold denoising of the new vibration acceleration set.The signals before and after denoising are shown in Figure 3.The red curves represent the denoised signal.It can be seen that the "burr" of the denoised signal disappears, indicating that the method can effectively remove the high-frequency noise from the original signal.The correlation coefficient between the original signal and the denoised signal reaches 0.98, so the denoised signal can characterize the original signal and the subsequent analysis can be performed on it.
Finally, data segments are implemented on the denoised signals.When trains of different groupings and speeds pass by a sensor, the time duration of the vibration signal is different.To eliminate the effect of the time difference on the recognition effect, a fixed window is applied to realize the data segmentation of the denoised signal.Since the sampling frequency of the monitoring system is 1000 Hz, in this paper, the length of the fixed window is set to 1000 and the time duration is 1 s.The time durations of the examples in Figure 3 are 6 s, 4 s, and 4 s, so they can be divided into 6, 4, and 4 data subsets, respectively.
By using data interception, wavelet threshold denoising, and fixed window segmentation, the raw data are transformed into a subset of data with the same time length.These algorithms are all coded in the MATLAB environment to form an automated data preprocessing system.

Time-Domain and Time-Frequency-Domain Feature Extraction Based on the LMD Method 4.1. LMD Method
Due to the dynamic working conditions and environmental factors, the vibration signals obtained from the field monitoring are usually nonstationary.The analysis tools for nonstationary signals include wavelet decomposition, the EMD method, and EMD-like methods.LMD [35] adaptively decomposes a series of components that can describe the characteristics of the signal by multiple iterations based on the local characteristics of the signal.Compared with EMD, LMD uses a smoothing method to obtain the local mean function and the envelope estimation function, which can effectively avoid modal mixing, end-point effects, and negative frequency problems of EMD [57].In this regard, this paper uses LMD to decompose the data.LMD essentially separates the pure frequency modulation signal and the envelope signal from the original signal and multiplies these signals to obtain an instantaneous frequency with the physical significance of PFs; iterations are performed until all PFs are separated out.Subsequently, the time-frequency distribution of the original signal can be obtained.For any signal x(t), the decomposition process is as follows [35]: All local extremum points N i of the original signal x(t) are calculated and the average m i of all adjacent local extremum points N i and N i+1 is derived: The local mean line can be obtained by connecting all mean points m i between the corresponding extreme moment points t(N i ) and t(N i+1 ) by a straight line, and then the local mean line is smoothed by the sliding average method to obtain the local mean function m 11 (t).

2.
The local amplitude a i is obtained from the adjacent local extrema N i and N i+1 : Similarly, the local amplitude line can be obtained by connecting all local amplitudes a i between the corresponding extreme moment points t(N i ) and t(N i+1 ) by a straight line, and then the local amplitude line is smoothed by the sliding average method to obtain the envelope estimation function a 11 (t).

3.
By separating the local mean function m 11 (t) from the original signal x(t), Equation ( 3) can be obtained:

4.
By dividing h 11 (t) with the envelope estimation function a 11 (t) to demodulate h 11 (t), Equation ( 4) can be obtained: The above steps are repeated for s 11 (t) to obtain the envelope estimation function a 12 (t) of s 11 (t).The above iterative process is repeated until a 12 (t) equals 1, implying that s 11 (t) is a pure frequency modulation signal.Assuming that the pure frequency modulation signal s 1n (t) is obtained by n iterations, the envelope estimation function a 1(n+1) (t) of s 1n (t) satisfies a 1(n+1) (t) = 1.Equation ( 5) can be obtained as follows: where The termination condition of the iteration is: 1.
The envelope signal a 1 (t) is obtained by multiplying all the envelope estimation functions obtained in the iterative process. 2.
The first PF of the original signal is obtained by multiplying the envelope signal a 1 (t) with the pure frequency modulation signal s 1n (t): 3. PF 1 (t) is separated from the original signal x(t) to obtain a residual signal u(t), and the residual signal u(t) is repeated as the original signal for k iterations until u k (t) is a monotonic function.
So far, the original signal x(t) is decomposed into k component signals and a residual function u k (t): As explained above, the decomposition step of LMD is a multi-loop iterative process, first, a pure frequency modulation signal is obtained by iterating through the local extreme of the original signal, and then a PF component is obtained by further iteration, and so on, until all PFs are separated out.The decomposition flow chart of LMD is shown in Figure 4. FM signal in the figure represents the frequency modulation signal.

Selection of Characteristic PFs Based on Pearson Correlation Coefficient
In general, a series of PFs characterizing different feature scales can be obtained by LMD of the vibration signals.To quantify the correlation degree between PFs and the original signal, the Pearson correlation coefficient is introduced, as shown in Equation (12).The Pearson correlation coefficient is a value between −1 and 1, with the value tending to −1 or 1 when the correlation between the two variables strengthens and to 0 when the correlation between the two variables weakens.In general, the correlation coefficient less than or equal to 0.2 [58] implies that the two variables are not correlated; this principle is adopted in this paper.If the correlation coefficient between a PF component and the original data is less than or equal to 0.2, the PF component is discarded and the retained PFs are called the characteristic PFs.After the selection of Pearson correlation coefficients, the noise and the environmental vibration of the original signal were discarded and the components useful for damage identification were retained.

Selection of Characteristic PFs Based on Pearson Correlation Coefficient
In general, a series of PFs characterizing different feature scales can be obtained by LMD of the vibration signals.To quantify the correlation degree between PFs and the original signal, the Pearson correlation coefficient is introduced, as shown in Equation ( 12).The Pearson correlation coefficient is a value between −1 and 1, with the value tending to −1 or 1 when the correlation between the two variables strengthens and to 0 when the correlation between the two variables weakens.In general, the correlation coefficient less than or equal to 0.2 [58] implies that the two variables are not correlated; this principle is adopted in this paper.If the correlation coefficient between a PF component and the original data is less than or equal to 0.2, the PF component is discarded and the retained PFs are called the characteristic PFs.After the selection of Pearson correlation coefficients, the noise and the environmental vibration of the original signal were discarded and the components useful for damage identification were retained.
where X represents the original data and Y represents the PFs.

Time-Domain Feature Extraction
When the track structure is in the deformed state, the waveform of the vibration signal definitely changes [59].The time-domain analysis method analyzes the system intuitively and accurately in the time domain and can be used to conduct structural damage identification studies by calculating the time-domain characteristic parameters of the vibration signal to reflect the relevant information of the vibration signal.It has been shown that the hierarchical function obtained by the adaptive decomposition method carries more damage information and is more sensitive to damage after decomposition [60].Therefore, this paper employs the peak-to-peak acceleration, variance, root mean square, shape factor, crest factor, and skewness of the characteristic PFs [52] as time-domain feature indicators.Time-domain feature analysis is performed separately for the characteristic PFs and then the average value is taken as the feature vector.Table 1 lists the specific information of the time-domain features.

Name Expression Description
Peak-to-Peak The difference between the maximum and minimum values. Variance The average of the sum of squares of the difference between the data and the mean value.
Root mean square Indicates the magnitude of the energy of the vibration signal.

Shape factor
Refers to a value that is affected by the shape of waveforms.
Crest factor C = max PF q /PF rms Detects the existence of shocks.Measures the skewness of the data distribution (Note: PF q represents the qth characteristic PF, PF represents the mean value of characteristic PFs, and r represents the number of characteristic PFs.)

Skewness
The amplitude of the vibration can only describe the instantaneous value of the vibration magnitude and does not include the time duration of vibration, and its root mean square value considering the time duration can characterize the magnitude of the vibration signal energy.The numerator in time-domain feature indicators of peak-to-peak acceleration, shape factor, and crest factor are the maximum value of vibration amplitude, highlighting the role of large amplitude.In addition, shape factor and crest factor can detect the presence of shocks in the vibration response signal.Variance can effectively characterize the extent to which a random variable deviates from its mean and, to a certain extent, can reflect the stability of the signal.Skewness can reflect the distribution characteristics of the vibration signal waveform.Due to the limitation of length, this paper only shows the scatter plot of the skewness of the track slab in normal and deformed states, and the results are shown in Figure 5.In this paper, 100 data were randomly selected from the sample pool of normal and deformed states, respectively, for comparative analysis.The black boxes and red dots in the figure represent the skewness in normal and deformed states, respectively.It can be clearly seen that when the track slab is in the deformation state, the vibration waveform asymmetry increases.In the range of coordinates both less or greater than 0, the skewness of the damage condition is greater than that in the normal state.This phenomenon indicates that the vibration waveform asymmetry increases when track slab deformation occurs and the skewness index increases.This is consistent with the existing literature research results [61].
track slab deformation occurs and the skewness index increases.This is consistent with the existing literature research results [61].

Time-Frequency-Domain Feature Extraction
When damage occurs to the track structure, the vibration energy distribution of the structure will change, and, therefore, the energy of the characteristic PFs of the track structure will change simultaneously.A certain location of the track structure at which damage occurs can be considered the location with reduced frequency softening to absorb more vibration energy, which leads to an increase in the energy entropy of the vibration response.Therefore, the energy entropy of characteristic PFs can be selected as one of the damage identification features [59].The energy entropy of characteristic PFs is calculated by the following steps: 1. Calculate the energy of the qth characteristic PF: where l is the length of a characteristic PF.
2. Calculate the total energy of these r efficient characteristic PFs: 3. Calculate the energy entropy of characteristic PFs: where is the percentage of the energy of the qth characteristic PF relative to the total energy entropy, and To better understand the model formulation, the variables and parameters that appear in this section are classified and named.The results are shown in Table 2.

Time-Frequency-Domain Feature Extraction
When damage occurs to the track structure, the vibration energy distribution of the structure will change, and, therefore, the energy of the characteristic PFs of the track structure will change simultaneously.A certain location of the track structure at which damage occurs can be considered the location with reduced frequency softening to absorb more vibration energy, which leads to an increase in the energy entropy of the vibration response.Therefore, the energy entropy of characteristic PFs can be selected as one of the damage identification features [59].The energy entropy of characteristic PFs is calculated by the following steps: 1.
Calculate the energy of the qth characteristic PF: where l is the length of a characteristic PF.

2.
Calculate the total energy of these r efficient characteristic PFs: Calculate the energy entropy of characteristic PFs: where p q = E q E is the percentage of the energy of the qth characteristic PF relative to the total energy entropy, and r ∑ q=1 p q = 1.
To better understand the model formulation, the variables and parameters that appear in this section are classified and named.The results are shown in Table 2.
Figure 6 is a scatter plot of the energy entropy of the vibration response around track slab deformation.The black squares and red dots in the figure represent the energy entropy of the characteristic PFs in the normal and deformed state, respectively.Figure 6 indicates that the energy entropy of the characteristic PFs is larger in the deformed state, which is due to the damage of the track slab.When the track slab is in the deformation state, the structural stiffness decreases and the position needs to absorb more energy to achieve the equilibrium state; thus, the energy entropy increases.This is consistent with the existing literature results [62].Length of characteristic product function Figure 6 is a scatter plot of the energy entropy of the vibration response around track slab deformation.The black squares and red dots in the figure represent the energy entropy of the characteristic PFs in the normal and deformed state, respectively.Figure 6 indicates that the energy entropy of the characteristic PFs is larger in the deformed state, which is due to the damage of the track slab.When the track slab is in the deformation state, the structural stiffness decreases and the position needs to absorb more energy to achieve the equilibrium state; thus, the energy entropy increases.This is consistent with the existing literature results [62].In this section, LMD is performed on the vibration data into several PFs, and the characteristic PFs are then selected with Pearson correlation coefficients.Six time-domain statistics and one time-frequency-domain feature are extracted from the characteristic PFs, which form a characteristic parameter vector matrix.

Establishment of Random-Forest Model
Random forest [63] is an ensemble machine learning algorithm with CART decision trees as the base classifier, where each decision tree is generated independently and completely split.It can solve both classification and regression problems by combining multiple independent decision trees based on voting or taking the average value; this approach tends to have higher accuracy and greater stability than using a single decision tree.Each decision tree is generated according to the following rules: (1) For a training set size of N, for each decision tree, N training samples that are put back are taken randomly from the training set as the training set.Approximately two-thirds of the observations are selected more than once, while the rest will not be chosen.The remaining one-third is called "out of bag" (OOB).( 2) If the feature dimension of each sample is M, a subset of mtry features is randomly selected from M features (mtry << M), and mtry is called the split features.When the decision tree is split, the best of these mtry features is selected for splitting.Each tree grows to the greatest extent and without the pruning process.(3) To classify the samples of the testing set, the classification label is the majority vote decision of the decision trees.The introduction of random samples and random variables is crucial to the classification performance of the random-forest model.The former enables the model to have a more stable resistance to overfitting, and the latter enables it to have a higher accuracy rate than the decision tree method.The architecture of the random-forest model is shown in Figure 7.The working process is summarized as follows: random samples are input to each independent decision tree for classification, and each decision tree provides independent results.For the classification problem, the final decision is made by considering the majority of votes to determine the classification result.
In this section, LMD is performed on the vibration data into several PFs, and the characteristic PFs are then selected with Pearson correlation coefficients.Six time-domain statistics and one time-frequency-domain feature are extracted from the characteristic PFs, which form a characteristic parameter vector matrix.

Establishment of Random-Forest Model
Random forest [63] is an ensemble machine learning algorithm with CART decision trees as the base classifier, where each decision tree is generated independently and completely split.It can solve both classification and regression problems by combining multiple independent decision trees based on voting or taking the average value; this approach tends to have higher accuracy and greater stability than using a single decision tree.Each decision tree is generated according to the following rules: (1) For a training set size of N, for each decision tree, N training samples that are put back are taken randomly from the training set as the training set.Approximately two-thirds of the observations are selected more than once, while the rest will not be chosen.The remaining one-third is called "out of bag" (OOB).( 2) If the feature dimension of each sample is M, a subset of mtry features is randomly selected from M features (mtry << M), and mtry is called the split features.When the decision tree is split, the best of these mtry features is selected for splitting.Each tree grows to the greatest extent and without the pruning process.(3) To classify the samples of the testing set, the classification label is the majority vote decision of the decision trees.The introduction of random samples and random variables is crucial to the classification performance of the random-forest model.The former enables the model to have a more stable resistance to overfitting, and the latter enables it to have a higher accuracy rate than the decision tree method.The architecture of the random-forest model is shown in Figure 7.The working process is summarized as follows: random samples are input to each independent decision tree for classification, and each decision tree provides independent results.For the classification problem, the final decision is made by considering the majority of votes to determine the classification result.The random-forest model classification performance is related to two factors [52]: the correlation of any two trees (the higher the correlation, the higher the error rate) and the classification ability of each tree (the stronger the classification ability, the lower the error rate).The classification ability of a tree and the correlation between two trees increases with the split features mtry.Therefore, the key problem is to determine the optimal split features mtry such that the classification ability and correlation are balanced, which is a parameter of the random-forest model.In addition, the more the number of decision trees, the better the randomization of the model, and thus the smaller the testing error of the model.Theoretically, the more the number of decision trees T, the better the randomization.With increasing number of decision trees, the test error variation decreases; hence, a The random-forest model classification performance is related to two factors [52]: the correlation of any two trees (the higher the correlation, the higher the error rate) and the classification ability of each tree (the stronger the classification ability, the lower the error rate).The classification ability of a tree and the correlation between two trees increases with the split features mtry.Therefore, the key problem is to determine the optimal split features mtry such that the classification ability and correlation are balanced, which is a parameter of the random-forest model.In addition, the more the number of decision trees, the better the randomization of the model, and thus the smaller the testing error of the model.Theoretically, the more the number of decision trees T, the better the randomization.With increasing number of decision trees, the test error variation decreases; hence, a reasonable number of trees can be determined.Hence, the split features mtry and the number of decision trees T are the two parameters of the random-forest model.The number of decision trees in the random-forest model can be determined by modifying the split features mtry and OOB error, while the error can be adjusted by changing the split features mtry and the number of decision trees T.

Random-Forest Model Results
In this section, the results of the random-forest model and the model parameter optimization are presented.As mentioned in the previous section, fiber optic accelerometers are installed adjacent to the operating HSR to continuously acquire vibration data of the track slab in the normal and deformed states continuously.First, raw data are preprocessed with data interception, denoising, and segmentation.Then, the data are decomposed by LMD, and the characteristic PFs are selected on the basis of Pearson correlation coefficients.Finally, six time-domain feature parameters and one time-frequency-domain feature parameter are calculated of the selected characteristic PFs.Therefore, seven features can be extracted from the dataset to form the characteristic parameter vector matrix.For conducting evaluation analysis after the model development, the total dataset is divided into training and testing datasets.In this study, the dataset comprises 2534 samples, that is, 1875 and 659 samples for the normal and deformed states of track slab, respectively.Of the total data, 70% is used for the training set and the rest for the testing and validation of the results.During the development of the random-forest model, the number of decision trees T and the split features mtry are determined.In general, the split features mtry = (M) 1/2 ; in this paper, M 7, so we let mtry = 3.The number of decision trees in random-forest model is determined according to the relationship between the number of decision trees and the classification accuracy, as shown in Figure 8.The classification accuracy tends to be stable when the number of decision trees reaches 1300.Given that the more decision trees, the better the randomization performance of the model, the number of decision trees in this paper is set to be 2000.
reasonable number of trees can be determined.Hence, the split features mtry and the number of decision trees T are the two parameters of the random-forest model.The number of decision trees in the random-forest model can be determined by modifying the split features mtry and OOB error, while the error can be adjusted by changing the split features mtry and the number of decision trees T.

Random-Forest Model Results
In this section, the results of the random-forest model and the model parameter optimization are presented.As mentioned in the previous section, fiber optic accelerometers are installed adjacent to the operating HSR to continuously acquire vibration data of the track slab in the normal and deformed states continuously.First, raw data are preprocessed with data interception, denoising, and segmentation.Then, the data are decomposed by LMD, and the characteristic PFs are selected on the basis of Pearson correlation coefficients.Finally, six time-domain feature parameters and one time-frequency-domain feature parameter are calculated of the selected characteristic PFs.Therefore, seven features can be extracted from the dataset to form the characteristic parameter vector matrix.For conducting evaluation analysis after the model development, the total dataset is divided into training and testing datasets.In this study, the dataset comprises 2534 samples, that is, 1875 and 659 samples for the normal and deformed states of track slab, respectively.Of the total data, 70% is used for the training set and the rest for the testing and validation of the results.During the development of the random-forest model, the number of decision trees T and the split features mtry are determined.In general, the split features mtry = (M) 1/2 ; in this paper, M = 7, so we let mtry = 3.The number of decision trees in randomforest model is determined according to the relationship between the number of decision trees and the classification accuracy, as shown in Figure 8.The classification accuracy tends to be stable when the number of decision trees reaches 1300.Given that the more decision trees, the better the randomization performance of the model, the number of decision trees in this paper is set to be 2000.After the number of decision trees in the random-forest model is determined, the split features mtry need to be determined.When the classification accuracy reaches the maximum, the number of split features is considered to be optimal.Table 3 shows the effect of the number of split features mtry on the classification accuracy when the number of decision trees T is equal to 2000.After the number of decision trees in the random-forest model is determined, the split features mtry need to be determined.When the classification accuracy reaches the maximum, the number of split features is considered to be optimal.Table 3 shows the effect of the number of split features mtry on the classification accuracy when the number of decision trees T is equal to 2000.The best performance of the random-forest model to recognize Class 1 reaches 98.18% accuracy when mtry is set to 1. Similarly, the random-forest model using 3 split features (mtry = 3) can well identify Class 2 with 95.59% accuracy.Considering the average classifying accuracy, the random-forest models employing 3, 4, and 7 split features can achieve relatively higher accuracy (96.09%, 95.79%, and 94.58%) compared with other cases.It is also worth noting that more attention should be paid to the recognition accuracy of Class 2, because the deformed state may affect the safe operation of the HSR.From the model perspective, the split features mtry balances the strength of the trees as well as the correlation of the trees in the random-forest model.Therefore, this paper, the split features mtry = 3 and the number of decision trees T = 2000 are used as the optimization parameters of the random-forest model.Figure 9 illustrates the results of random-forest classification using these optimized parameters (mtry = 3, T = 2000).The best performance of the random-forest model to recognize Class 1 reaches 98.18% accuracy when mtry is set to 1. Similarly, the random-forest model using 3 split features (mtry = 3) can well identify Class 2 with 95.59% accuracy.Considering the average classifying accuracy, the random-forest models employing 3, 4, and 7 split features can achieve relatively higher accuracy (96.09%, 95.79%, and 94.58%) compared with other cases.It is also worth noting that more attention should be paid to the recognition accuracy of Class 2, because the deformed state may affect the safe operation of the HSR.From the model perspective, the split features mtry balances the strength of the trees as well as the correlation of the trees in the random-forest model.Therefore, in this paper, the split features mtry = 3 and the number of decision trees T = 2000 are used as the optimization parameters of the random-forest model.Figure 9 illustrates the results of random-forest classification using these optimized parameters (mtry = 3, T = 2000).The points in Figure 9 represent the number of decision trees for both classifications, the horizontal coordinate represents the number of decision trees whose voting result is Class 1, and the vertical coordinate represents the number of decision trees whose voting result is Class 2. Since the total number of decision trees is 2000, all points are on the diagonal (x + y = 2000, x represents the number of decision trees whose voting result is Class 1, and y represents the number of decision trees whose voting result is Class 2).The red asterisks in the figure represent misclassified samples and the blue circles represent correctly classified samples.The samples in the blue boxes in Figure 9 represent close voting results and are prone to misclassification.Therefore, the closer the sample points are to the coordinates (2000, 0) and (0, 2000), the better is the performance of the model.It can be seen that, in the figure, most of the samples are concentrated at the two coordinates mentioned above, which indicates that the proposed random-forest model has a good classification performance.From a quantitative perspective, the total number of samples in the blue box is 7, accounting for 0.92% of the total testing set; this is a small percentage of the data, indicating that this random-forest model has a good classification effect.

Decision-Tree Model Results
To verify the superiority of the proposed framework, we adopt a decision-tree model for a comparison study.The decision-tree model is a predictive model with an attribute structure that represents a mapping relationship between object attributes and object values [51].A decision tree contains a root node and a number of internal nodes and leaf nodes.The root node contains the entire set of samples, the internal node represents a feature or attribute, and the leaf node represents a class.The path from the root node to each leaf node corresponds to a sequence of decision tests.Decision-tree learning essentially generalizes a set of classification rules from a training set to obtain a decision tree that is less contradictory to the dataset, while having good generalization ability [64].In this paper, the ID3 algorithm and information gain principle are adopted to calculate and select the attribute with the greatest information gain as the classification attribute.In general, a number of attempts are needed to determine the sample number contained in leaf nodes to achieve good performance.Even a small sample number contained in leaf nodes will considerably affect the generalization ability of the decision-tree model.In this study, the cross-validation method is used to determine the sample number contained in leaf nodes to optimize the decision-tree model.Figure 10 shows the relationship between the number of samples contained in leaf nodes and the cross-validation error.The smaller the crossvalidation error, the better is the generalization ability of the decision-tree model.As shown by the red circle in the figure, when the sample number contained in leaf nodes is 28, the cross-validation error is the smallest at 0.0523 and the decision-tree model generalization ability is the best.The decision-tree model is optimized for improved performance using the cross-validation method.in the blue box is 7, accounting for 0.92% of the total testing set; this is a small percentage of the data, indicating that this random-forest model has a good classification effect.

Decision-Tree Model Results
To verify the superiority of the proposed framework, we adopt a decision-tree model for a comparison study.The decision-tree model is a predictive model with an attribute structure that represents a mapping relationship between object attributes and object values [51].A decision tree contains a root node and a number of internal nodes and leaf nodes.The root node contains the entire set of samples, the internal node represents a feature or attribute, and the leaf node represents a class.The path from the root node to each leaf node corresponds to a sequence of decision tests.Decision-tree learning essentially generalizes a set of classification rules from a training set to obtain a decision tree that is less contradictory to the dataset, while having good generalization ability [64].In this paper, the ID3 algorithm and information gain principle are adopted to calculate and select the attribute with the greatest information gain as the classification attribute.In general, a number of attempts are needed to determine the sample number contained in leaf nodes to achieve good performance.Even a small sample number contained in leaf nodes will considerably affect the generalization ability of the decision-tree model.In this study, the cross-validation method is used to determine the sample number contained in leaf nodes to optimize the decision-tree model.Figure 10 shows the relationship between the number of samples contained in leaf nodes and the cross-validation error.The smaller the cross-validation error, the better is the generalization ability of the decision-tree model.As shown by the red circle in the figure, when the sample number contained in leaf nodes is 28, the cross-validation error is the smallest at 0.0523 and the decision-tree model generalization ability is the best.The decision-tree model is optimized for improved performance using the cross-validation method.During decision-tree model learning, the node partitioning process will be repeated for classifying the training samples as accurately as possible.This sometimes results in overfitting because of too many branches in the decision-tree model.Therefore, the risk of overfitting can be reduced by actively removing some branches.In this study, the original decision-tree is pruned for comparison.The classification accuracy and cross-validation error of the original, optimized and pruned decision-trees are compared to determine the best performance of decision-tree model.Table 4 shows the classification accuracy and cross-validation error of the original, optimized, and pruned decision-tree.During decision-tree model learning, the node partitioning process will be repeated for classifying the training samples as accurately as possible.This sometimes results in overfitting because of too many branches in the decision-tree model.Therefore, the risk of overfitting can be reduced by actively removing some branches.In this study, the original decision-tree is pruned for comparison.The classification accuracy and cross-validation error of the original, optimized and pruned decision-trees are compared to determine the best performance of decision-tree model.Table 4 shows the classification accuracy and cross-validation error of the original, optimized, and pruned decision-tree.As shown in Table 4, the cross-validation error reduced from 6.76% to 5.23% and the average classification accuracy increased from 91.49% to 93.06% by optimizing the original decision-tree; the cross-validation error reduced from 6.76% to 6.26% and the average classification accuracy increased from 91.49% to 92.51% by pruning the original decisiontree.Therefore, both optimizing and pruning of the original decision-tree can improve the generalization performance of the decision tree, with increased Class 1, Class 2, and average classification accuracy.In particular, the cross-validation error of the optimized decision-tree is smaller than that of the pruned decision-tree, indicating that optimization makes the decision-tree model generalize better than pruning.Therefore, the optimized decision-tree classification accuracy is higher than that of pruning, which consistent with the above results.In summary, for the recognition of track slab deformation in this study, the original decision-tree needs to be optimized without pruning, and the classification accuracies of Class 1 and Class 2 are 94.96% and 91.15%, respectively, with an average classification accuracy of 93.06%.Comparison of the classification results shown in Table 3 indicates that the average classification accuracy of the random-forest model is higher than that of decision-tree model regardless of the value of split features.This confirms that the proposed random-forest model is better than the decision-tree model for the identification of track slab deformation.

Conclusions
In this paper, we proposed a method to analyze the health state of HSR using datadriven approaches based on on-site monitoring data.A novel random-forest-based strategy that combines LMD, time-domain feature extraction, and time-frequency-domain feature extraction for identification of track slabs deformation was proposed.The main conclusions are as follows.

1.
A track-side vibration monitoring method based on fiber optic vibration sensing technology can effectively capture vibration signals that contain the information of train vibration, track slab deformation, noise, and environmental vibration.

2.
The preprocessing methods of data interception, denoising, and data segmentation can effectively eliminate the effects of environmental vibration, noise, and time differences on the recognition effects.In addition, the time-domain and time-frequencydomain feature extraction methods based on LMD effectively extract the crucial information for detecting the deformation of track slabs.

3.
The proposed intelligent recognition algorithm based on random-forest model can accurately identify the deformation of track slabs.The verification test results showed that the recognition accuracy reached 96.09%.
This paper provides not only a new method for operational HSR deformation monitoring but also an intelligent identification algorithm for track slab deformation identification.More importantly, the proposed method is applicable to the monitoring and identification of other structural defects.In the future, we will research the random-forest model on the deformation degree of track slabs.

Figure 1 .
Figure 1.Flowchart of the deformation identification of a track slab.

Figure 1 .
Figure 1.Flowchart of the deformation identification of a track slab.
the original signal x(t) is decomposed into k component signals and a residual function uk(t): explained above, the decomposition step of LMD is a multi-loop iterative process, first, a pure frequency modulation signal is obtained by iterating through the local extreme of the original signal, and then a PF component is obtained by further iteration, and so on, until all PFs are separated out.The decomposition flow chart of LMD is shown in Figure 4. FM signal in the figure represents the frequency modulation signal.

Figure 5 .
Figure 5. Scatter plot of the skewness of the vibration response.

Figure 5 .
Figure 5. Scatter plot of the skewness of the vibration response.

Figure 6 .
Figure 6.Scatter plot of the energy entropy of the vibration response.

Figure 6 .
Figure 6.Scatter plot of the energy entropy of the vibration response.

Figure 7 .
Figure 7. Architecture of a random-forest classification model.

Figure 7 .
Figure 7. Architecture of a random-forest classification model.

Figure 8 .
Figure 8. Relationship between the number of decision trees and classification accuracy.

Figure 8 .
Figure 8. Relationship between the number of decision trees and classification accuracy.

Figure 9 .
Figure 9. Results of random-forest classification.The points in Figure9represent the number of decision trees for both classifications, the horizontal coordinate represents the number of decision trees whose voting result is Class 1, and the vertical coordinate represents the number of decision trees whose voting result is Class 2. Since the total number of decision trees is 2000, all points are on the diagonal (x + y = 2000, x represents the number of decision trees whose voting result is Class 1, and y represents the number of decision trees whose voting result is Class 2).The red asterisks in the figure represent misclassified samples and the blue circles represent correctly classified samples.The samples in the blue boxes in Figure9represent close voting results and are prone to misclassification.Therefore, the closer the sample points are to the coordinates (2000, 0) and (0, 2000), the better is the performance of the model.It can be seen that, in the figure, most of the samples are concentrated at the two coordinates mentioned above, which indicates that the proposed random-forest model has a good classification performance.From a quantitative perspective, the total number of samples

Figure 10 .
Figure 10.Relationship between the number of samples contained in leaf nodes and the crossvalidation error.

Figure 10 .
Figure 10.Relationship between the number of samples contained in leaf nodes and the crossvalidation error.

Table 1 .
Summary of time domain features.

Table 2 .
Nomenclature table of the model variables and parameters.

Table 2 .
Nomenclature table of the model variables and parameters.

Table 3 .
Classification accuracy (%) with different number of split features.Class 1 stands for the normal state of the track slab and Class 2 for the deformed state of the track slab).

Table 3 .
Classification accuracy (%) with different number of split features.Class 1 stands for the normal state of the track slab and Class 2 for the deformed state of the track slab).
(Note: Error stands for the cross-validation error).