Beam Deflection Monitoring Based on a Genetic Algorithm Using Lidar Data

The Light Detection And Ranging (LiDAR) system has become a prominent tool in structural health monitoring. Among such systems, Terrestrial Laser Scanning (TLS) is a potential technology for the acquisition of three-dimensional (3D) information to assess structural health conditions. This paper enhances the application of TLS to damage detection and shape change analysis for structural element specimens. Specifically, estimating the deflection of a structural element with the aid of a Lidar system is introduced in this study. The proposed approach was validated by an indoor experiment by inducing artificial deflection on a simply supported beam. A robust genetic algorithm method is utilized to enhance the accuracy level of measuring deflection using lidar data. The proposed research primarily covers robust optimization of a genetic algorithm control parameter using the Taguchi experiment design. Once the acquired data is defined in terms of plane, which has minimum error, using a genetic algorithm and the deflection of the specimen can be extracted from the shape change analysis.


Introduction
In the 4th industrial revolution, Structural Health Monitoring(SHM) is becoming a hot issue in the construction industry [1,2] in which nondestructive evaluation and serviceability monitoring are essential features. Nondestructive evaluation plays a critical role in assuring that structural components and systems perform in a reliable and cost-effective fashion. This mechanism does not affect the future usefulness of the object or material. On the other hand, serviceability monitoring is performed in real time throughout its service life span [3,4]; this is directly related to controlling the structural responses caused by either deflection, cracks, vibration, creep, or a combination of them. Various techniques have been utilized for the assessment of structure performance [5]. Sensors are essential components and have different purposes based on the nature of the techniques. In general, we can classify sensors into two types based on the connection. Contact sensors which are commonly known for having physical interactions with the target structure. A Linear Variable Differential Transformer(LVDT), piezoelectric transducer, fiber optic sensor and acoustic emission sensor are common examples of this type [6][7][8]. On the other hand, non-contact sensors are known for acquiring responses from the target material without making direct or indirect contact. Laser sensor systems, drones with vision-based sensors using cameras, wireless rechargeable sensor networks, radar sensor networks and lidar sensor systems are grouped as this type [9][10][11]. Recently, non-contact sensors are being commonly utilized due to their portability, easy use in harsh surroundings, and so on. Knowing about the deflection of a beam or any structure using TLS has a big advantage from the perspective of site conditions. For the sake of describing the benefits of this study in the real world, aged structures, considering their spatial positions and site conditions, are not safe due to perilous structural conditions, inconvenience, insecurities and slippery site conditions. Consequently, the Light Detection And Ranging (LiDAR) system has become prominent in structural health monitoring. Despite such sensors being used for detection, measurement and characterization of hidden and/or apparent defects using advanced techniques, there are still many questions regarding the dimensions of the sensors for structural health monitoring.
Terrestrial Laser Scanning (TLS), or lidar, is a crucial non-contact optical sensor that analyzes the structural Three Dimensional(3D) shape in terms of a very dense 3D point cloud. This technology is a recent innovation in the spatial information of data acquisition, allowing for a scanned area to be digitally captured with unprecedented resolution and precision. However, the acquired data are influenced by several factors. Among them, random errors due to inherent physical properties are often difficult to eliminate while error due to environmental and scanning geometry issues can be removed through analysis [12]. Consequently, attention should be given to data acquisition by considering the position, incidence angle, and color of a specimen [13]. As a result, it is possible to dwindle noises tremendously.
In recent years, various approaches have been used to estimate the elements for serviceability limit states such as accumulated deflection, crack and dynamic displacement [14][15][16]. Several researchers have published papers concerning the serviceability assessment in support of structural health monitoring [17][18][19]. Park et al. conducted an experiment to define the deflection of a steel structure element via TLS using the geometrical shape of the specimen [20]. Even though the model of Cabaleiro et al. is not valid due to torsion, they tried to model a deformed beam caused by concentric loads and torsional forces on the specimen after scrutinizing the max deflection with respect to the allowable building codes [21]. Some researchers proposed how to estimate the deflection of a structural element, i.e., a beam, by integrating photogrammetry with TLS. For instance, Gordon measured the beam deflection with respect to benchmark photogrammetric data [22]. Zogg and Ingensand tried to monitor a deformation of real structure called Felsenau Viaduct(CH) bridge which is a part of Swiss highway via with TLS. The deformation on their study is mainly caused by settlement and tilting of a structure. Eventually, they showed that TLS could replace the area wide precise levelling in monitoring of a structure deformation since the maximum difference in between these two system is less than 1 mm [23]. Olsen et al. conducted a damage assessment for full scale structural test specimen by identifying volumetric change and its deformation [24]. Cabaleiro et al. utilized TLS data for checking of a deflection and stresses caused by torsion especially for open cross section which have a very low torsional strength. They compared the result obtained from the proposed methodology with the measurement taken during an experiment and Finite Element Modeling(FEM). As a result, they could show that their proposed algorithm is much closer to the measurement taken directly which is considered as a ground truth data [25].
Naturally, Lidar data are highly vulnerable and easily affected by noise [26]. Different scholars have been studying the factors which affect the Lidar data regarding structural health monitoring [13,27]. On the other hand, some researchers have used different types of denoising methods and optimization techniques in order to diminish their effect [28,29]. Notice that, outliers and noise are differing conceptually. Indeed, these two terms have really ambiguous meaning to differentiate by the researchers. However, Sagado et al. defined outliers and noise individually. According to their definition, an outlier is a data point which is different from the remaining data whereas noise can be defined as mislabeled examples (class noise) or errors in the values of attributes (attribute noise) [30]. Genetic Algorithm(GA) is also one of the crucial optimization methods used in order to attain the target effectively without the noise affection. GA has been conducted in different topics of structural health monitoring by different scholars. Optimization of sensor placement which provides the best possible performance is one of many critical topics [31]. However, estimating a structural deflection for the purpose of assessing the performance via a genetic algorithm is not prominently conducted so far. Rather, many researchers have performed this optimization method to determine a structure deflection both analytically and numerically [32]. GA can be also incorporated with a regression methodology to increase the robustness of curve fitting. Indeed, there are many techniques that are capable to increase the robustness of a curve fitting technique when we have numerous numbers of outliers in the data; Least Absolute Deviation (LAD), M-estimation, and S estimation are popular schemes which can considerably improve estimation precision [33].
This study presents an effective algorithm for measuring the structural deflection by enriching previous studies. Improvements are made according to the following five main steps of the procedure listed; the acquisition of TLS data for the loading and unloading scenarios, fitting of a plane for the point cloud acquired during the unloading scenario using a robust genetic algorithm, transformation of the scanner coordinates into local structural coordinates, curve fitting of transformed data for the loading case, and eventually, estimation and comparison of deflection between contact sensors. Furthermore, our research illustrates the performance of the proposed procedure with a validating experiment in which deflection measurements are simulated based on the loading scenario.

Least Square Regression
A plane can be described by a normal vector n = [A, B, D] T perpendicular to the plane, and a vector on the plane connected to a known point p 1 = (x 1 , y 1 , z 1 ) and an arbitrary point p 2 = (x, y, z) is described by p 2 − p 1 since the normal vector and vector in the plane are perpendicular to each other, the dot product of these two vectors should be null [34].
For the sake of simplifying the over-determined problem, begin by removing one component by constraining the solution space. Thus we assume the coefficient D is the unit: Keeping the assumption that the z-component of the data is functionally dependent on the x and y-components and given a set of samples (x i , y i , z i ), determine A, B, and C so that the plane z = Ax + By + C best fits the samples in the sense that the sum of the squared errors between the z i and the plane values Ax i + By i + C is minimized [35]. This can be written via Equation (3).

Genetic Algorithm
Stochastic optimization is a process of seeking a maximum or minimum value of a mathematical or statistical function in the presence of randomness. GA is one of the common method which include in this kind of optimization techniques. It is used to find the optimal solution(s) to a given computational problem maximizing or minimizing a function. A genetic algorithm is a random-based classical evolutionary algorithm [36]. This principle of continual improvement over generations is utilized by evolutionary algorithms to optimize solutions to a problem. In the initial generation, a population composed of different individuals is generated randomly or by other methods. An individual is a solution to the problem, which can vary in quality: the quality of an individual with regard to the problem is called the fitness, which reflects the adequacy of the solution to the problem to be solved. The higher the fitness of an individual, the more likely it is to pass some or all of its genotype to individuals of the next generation. Increasing the robustness of an algorithm improves the output. The primary distinguishing features of such algorithms are encoding, a selection mechanism, a crossover mechanism, a mutation mechanism, and a culling mechanism. Such algorithms can optimize multiple objectives simultaneously and can be used as black boxes since they do not assume any properties of the mathematical model to be optimized [37]; their only real limitation is the computational complexity.

Proposed Approach for Computation of Deflection
The proposed method is applied to measure artificial deflection of a structural element from TLS data and compare this with the contact sensor mounted during the experiment. The deflection is estimated by taking the difference in shape change between the loading and unloading scenarios. In the unloading case, first the resulting point cloud data are represented by plane and an extracted line, which help to set the spatial position of a local specimen coordinate system. Once the coordinate system is transferred to the new system, we can define and visualize a deflection curve for each loading case. The flowchart in Figure 1 shows the steps needed to estimate the deflection of a structure. The subsequent subsections explain the proposed method in detail.

Acquisition of a 3d point cloud Unloading Epoch
Pre-processing

Acquisition of the Point Cloud and Pre-Processing
Representing or modeling real-world scenarios with virtual worlds makes it easy to apply analytical theories and visualize the effects. Among them, capturing an object using a 3D digital device, i.e., TLS in our case, enables us to measure accurately any changes in shape with magnificent resolution. Basically, The TLS point cloud tells us the spatial position (X, Y, Z) of a point with respect to its own scanner coordinate system. Additionally, TLS provides us with information about the scalar field that represents intensity, RGB color information, and time in some cases. Indeed, not every sensors can produce RGB color information for the users. Some brands like, RIEGL LMS-Z420i, has a camera system that detached from the main system which is used to acquired the color data of an objects. whereas in our case the system has in-built camera system which helps to produce RGB color data with the corresponding X, Y, and Z coordinates. In this study, only the spatial information of a point cloud is utilized for the proposed analysis to achieve the required solution. Pre-processing is a technique that improves the accuracy and resolution of data, along with filtering techniques that serve to enhance or highlight the spatial characteristics of an image data set. Manual trimming and segmentation are employed in this study since they are considered enough for analyzing the data. After removing the outliers and segmentation, the number of points on the flange became 39,730 while those on the web entity become 163,024 points. Figure 2 depicts the resulting scanned point cloud representation of a specimen with the corresponding solid 3D representation of an object.

Characterizing the Point Cloud in Terms of Definable Mathematical Elements
Our data form a point cloud set which has information about its spatial position. We only have a collection of point cloud data which describe the specimen. Therefore, representing this data by a plane and/or a surface, which can be defined by some mathematical equation, is significant for further analysis. Extracting the line which helps us to locate the local structural coordinate system, transforming the coordinate of a point from scanner coordinate system to local structural coordinate system and computing the deflection of a specimen anywhere along with the span are some tasks that can be computed whenever the acquired point cloud data is defined by mathematical plane and/or surface.

Least Squares Genetic Algorithm(LS-GA) Based Plane Fitting
The principle of least squares sum in the curve fitting problem, which minimizes the difference between the data and model of the output, is useful. However, in this study, the basic principle of least squares is incorporated with a notable stochastic optimization known as a genetic algorithm for the purpose of decreasing the difference between the data and model [38]. Consequently, this is particularly useful if we utilize an error metric in terms of the functional form and optimize it using a GA. Holland and his student stated that GA works with an encoding of the individuals throughout the algorithm [36]. Originally, they constructed a GA based on binary numbers. Once we have fixed the number of chromosomes in the population, the basic algorithm operators (Selection, Crossover, and Mutation) start to conduct iteratively until it converges to optimal solutions. Currently, all these complication and processes are simplified using different platforms. Among them, the global optimization toolbox in MATLAB software is utilized for our studies. The essential parameters for running the GA in this study are [37]: • Fitness function: The fitness function is a mathematical formulation of the desired optimization problem. It determines how suitable a solution is. The magnitude of the residual, which defines the difference between the model and data examines the individuals formed throughout all generation. This can be described in Equation (4): where F(A, B, C) is the point model, normally defined as the nearest data point P • Population: The population is a set of individuals that have a chance to be the fittest among them. An individual is characterized by a set of variables called its genes. It is a basic building block of for the algorithm. In our case, this refers to the parameter of the regression model. These genes are joined into a string to form a chromosome (individuals). It is simply a series of binary numbers (0 and 1), which encode all values of parameter of the regression model. I 1 , I 2 up to I 200 that are shown in Figure 3 are list of chromosomes in the population which have a binary representation of all the parameters A, B, and C of the plane. Our GA is one of the unique behaviors of this optimization techniques; it works only on the number of chromosomes inside the population. The number of chromosomes in the population defined by the user is labeled as the population size. The building block and their corresponding chromosomes in the population is shown in the Figure 3 . • Fitness scaling function: If we consider only multiplying the fitness by some factor, then as the name implies, this does not change the relationships among the population. Rather, a much more sensible fitness scaling is an affine transformation, where scaled(f) = a * f + b, of the fitness value, is seen via formulae, the values are multiplied by some number and then offset by another either up or down. This parameter plays a crucial role in GA by limiting the tendency of the strongest solution to overwhelm the weaker ones thus avoiding premature convergence. • Selection: The selection is a parameter for choosing two parents from the population for the purpose of mating. This process continues in every single generation until one of the stopping criteria is attained. Parents, i.e., a pair of individuals, are chosen based on their fitness score. • Crossover/Reproduction: This parameter mainly depicts how the selected individuals mate with each other and create new offspring for the next generation. In this process, part of the individuals is exchanged with its mate based on the user-defined crossover probability.

•
Elite Count: This is the number of individuals with the best fitness values in the current generation that have been retained for the next generation. Because of their robust fitness score, more copies of these individuals are achieved in the subsequent generation. • Crossover fraction: The fraction of individuals in the next generation, other than the elite children, that are created via crossover.

•
Mutation: A process of changing of a bit (gen) within a bit string (chromosome). This is done to maintain diversity within the population and prevent premature convergence.

•
Migration: An exchange of information (exchange of individuals) between the sub-populations. Figure 4 illustrates how the parameter of genetic algorithm is functioning in one generation. This feature of a genetic algorithm allows us to find the optimal parameter of a plane with respect to an error metric.
Offspring Population

Taguchi Experimental Design
Even though the prevalent method of modeling a plane fitting is based on deterministic ordinary differential equations, a stochastic method such as a genetic algorithm is more appropriate and well established whenever the amount of noise is significant [38][39][40]. However, GAs have limitations regarding the optimality of individuals which satisfy the given fitness function unless an appropriate input parameter is selected. Consequently, to overcome this drawback, a different scheme has been utilized for tuning a suitable parameter for a genetic algorithm [41][42][43]. Among various techniques, the Taguchi experimental design approach is used for tuning the parameters of the GA solver. Dr. Taguchi developed the design of an experiment based on well-defined guidelines. Taguchi's design uses orthogonal arrays that estimate the effects of factors on the response mean and variation. An orthogonal array means the design is balanced such that the factor levels are weighted equally. Taguchi uses the following convention for naming the orthogonal arrays: L a (b c ), where L stands for Latin square, a is the number of experimental runs, b is the number of levels of each factor, and c is the number of variables [44]. We can easily assess each factor individually because of utilizing an Orthogonal Array on the design experiment. As a result, the effect of one factor does not affect the estimation of a different factor. This can reduce the time and cost associated with the experiment when fractionated designs are used. Therefore, Taguchi's design of experiments plays a crucial role in the reduction of number of trial experiments [45]. The parameters in Taguchi's design are selected by considering the effects of individual value on the algorithm output. Here, nine parameters and their levels are selected and summarized in Table 1. Among the various preferences for experiment design, a robust design is identified using the signal-to-noise ratio (SNR). This is because the higher values SNR identify the control settings that minimize the effects of the noise factors. Basically, the Taguchi design of experiment is used for two optimization processes: (a) use the signal-to-noise ratio to identify those control factors that reduce variability. (b) identify control factors that shift the mean to the target and have a small or no effect on the signal-to-noise ratio [45]. A SNR is suggested by Taguchi for cases in which the response standard deviation is related to the mean. The purpose of SNR in Taguchi's approach to robust parameter design is to provide an easy-to-use performance criterion that takes the process mean and variance into account. Among the numerous SNR values, "the smaller-the-better" philosophy used in this study, as shown in Equation (5).
where; SNR is signal to noise ratio, y i is the response variables, and n is number of response values. ∑ n i=1 implies the summation over n response values at the outer array points.

Extraction of the 3D Intersection Line
Defining the coordinates of points using the new basis is important because it allows for visualization and quantification of the real shape changes of the specimen. However, it is not easy to find out the appropriate position of a new basis (local structural coordinate system) since we have a collection of points. Therefore, considering one axis (in this case, the X-axis) coincides with intersection 3D line between the flange fitted plane P f and the web fitted plane P w is an efficient way of determining the position with the new basis as shown in the Figure 5. Suppose the direction vectors for the flange and web planes in the Figure 5 are N f (i f , j f , k f ) and N w (i w , j w , k w ), respectively. Since these two planes cross each other perpendicularly, the vector product of their normal vectors is equivalent to the direction vector S of their line of intersection, L i (6).
However, this information is not enough to extract the equation of the line. Thus, picking one arbitrary known point for our case results in the intersection of the line of intersection of the two planes L i with the left edge of the plane L l , and this point is inserted in to the following symmetric form in Equation (7) for 3D lines.
where Accordingly, the real spatial position of the new basis (x',y',z') can be used to determine the coordinate transformation, followed by setting the x'-axis of the local structural coordinates on the extracted intersection line of the planes, as shown in Figure 5.

Transformation of Coordinates
To easily evaluate the real changes in shape of the specimen by visualizing the effect of loading, the reference frame is transformed from the scanner coordinate system to the local structural coordinate system. This implies that the acquired data set coordinates should be defined using the local structural coordinate system. A rigid transformation (also called an isometry) is a transformation which preserves the shape and size of the object. Reflections, translations, rotations, and/or combinations of these three transformations can be categorized as rigid transformations [46]. A translation T O O is a transformation which slides an object by a fixed range from point O to point O , as in Figure 5. All the other points move the same distance in the same direction whenever the coordinates of the point cloud are multiplied by the translation matrix, as shown by Equation (8). After a translation, which is defined as moving the origin of the scanner coordinates to the hypothetical spatial position of the local structural coordinates, a rotation is applied. A rotation R θ_i,O is a transformation that twirls an object about a fixed point O', called the center of rotation, with an anlge θ in either the x, y, z, or a combination of them; the rotation directions are called the Euler angles [47].
The rotation matrices for rotating a vector about the x-axis by an angle α, about y-axis by an angle β and about the z-axis by an angle γ are given by Equations (9a)-(9c), respectively.
Once the transformation of a rotation angle into a rotation matrix is successfully completed, the transformation matrix in Equation (10) is obtained by combining the above-described rotation matrices with the translation matrix as follows: where, [M] is 4 × 4 transformation matrix; R (α, x ) is a rotation matrix for rotating of a vector by an angle α with respect to the x axis; R (β, y ) is Rotation matrix for rotation of a vector by angle β with respect to y' axis; and R (γ, z ) is Rotation matrix for rotation of a vector by angle γ with respect to z axis. In the three-dimensional world, four coordinates are necessary when considering the perspective of a scene. In projective space, two parallel lines appear to meet at the horizon, which is not the case in Euclidean space. Therefore, mathematicians use homogeneous coordinates which can represent the N-dimensional coordinates using N + 1 numbers [48]. Equation (11) shows, how Cartesian coordinates can be described in terms of homogeneous coordinates. Consequently, the transformation matrix [M] is defined as a 4 × 4 matrix by redefining the rotation and translation matrices in terms of adding an additional dimension to the coordinates.
where, (X, Y, Z) are the Cartesian coordinates and (x, y, z, w) are the homogeneous coordinates. Now, the coordinates of a point cloud acquired in a different loading are multiplied by the transformation matrix described in Equation (10) to obtain the new coordinates so as to identify and quantify the actual shape change of a specimen. This can be described with the following Equation (12): where, (x , y , z , 1) is the point coordinates w.r.t. the new basis; (x, y, z, 1) is the point coordinates w.r.t. the old basis; [M] is the transformation matrix.

Estimation of the Deflection Curve from the Loading Scenario
After obtaining the coordinate transformation, it is possible to consider changes in the shape of an object formed by the point cloud that are equivalent to the actual object. In the last phase of the proposed algorithm, we fit the curved surface as a 2nd degree curve on the longitudinal axis and a linear curve on the transverse axis for each individual loading scenario using the transformed point clouds. The GA fitness function appears more complicated than the previous one since the error metric formulation is now considered as a two degree curve caused by loading. This can be described by the following Equation (13): where d 2 i is the squared error which depicts the difference between the actual data and the model; P i is the measured data of the z-coordinate of a point cloud; F (A, B, C, D, E) = Ax i 2 + Bx i + Cy i + Dx i y i + E ; x i is the measured x-coordinate of a point cloud; y i is the measured y-coordinate of a point cloud; A, B, C, D & E are the coefficient parameter of the model; and N is the number of points in the point cloud data.
Like the unloading scenario, the resulting point cloud from loading is fitted as a curved surface according to the parameters obtained from genetic optimization. As discussed earlier, the method of finding suitable parameters for the optimization function, i.e., Taguchi's Design, is also utilized here. Because we assumed that lateral torsional buckling of a specimen is negligible due to the stiffeners, the longitudinal profile deflected shape of the specimen is our main focus. As a result, taking the longitudinal line which lies on the curved surface as a deflection curve makes it easy to estimate the deflection of the required position. Even though we set the local structural coordinates according to the intersection between the upper flange and the front side of the web, the center line of the deflected shape of a curved surface (flange) accurately describes the actual effect of the loading on the specimen. According to our experimental setup, which is a fixed-fixed support of beam, it is known that there is zero deflection at the end point of the specimen. Finally, the results obtained from the actual LVDT measurements and the analysis of what is proposed are compared and contrasted.

Experimental Study
The proposed method has been validated through experiment carried out at Sungkyunkwan university. The experiment involved a steel box girder section specimen under different loadings. The experiment took place at the Concrete Material Lab (Suwon), where the University is located. Figure 6 shows how all the entities are synchronized during the experiment. As shown in Figure 6, the main entities that play essential roles during the experiment are:

Design of the Experiment
The indoor experiment is setup as follows: • Setting the position of a specimen through the UTM machine keeps all the necessary alignments both horizontally and vertically. The scanner device stands 2.5 m away from the front web face while considering which factors affect the accuracy of the data. Target color, incidence angle, range and intensity are the main factors that affect the point cloud noise and parametric model fitting [13]. Consequently, this setup considers all the results from Bolkas et al. in accordance with the available space in the laboratory • The contact sensor (LVDT) is mounted below the specimen at three different positions. One is at the center and the other two LVDTs are fixed 55 cm from the left and right edges individually, as shown in Figure 7. These LVDTs and the UTM machine are connected to a computer.

•
The specimen is scanned without any loading by setting the required field of view. The field of view for scanning an object should reduce the outliers caused by objects out side of the target. • Once we are done scanning the specimen scene without loading, we apply loading via the UTM until the center LVDT reading reaches 1mm. The 1 mm sag is attained at a 57.33 KN loading. Again, the scanning process starts over by pausing the applied load and keeping the 1 mm sag.

•
In this fashion, specimen scanning is carried out for different deflection sizes for the corresponding loadings. Table 2 summarizes the induced load for each case along with their corresponding LVDT sensor readings.

•
After capturing all the necessary data with the USB, which was plugged in to the scanner during the scanning process, we changed the file format from .PLY to .PTS using cyclone which is a software module of Leica, for the purpose of using the cloud compare software. Once we have the data file format which is capable to utilize via with cloud compare, it is easy to apply manual segmentation of an object entity, removing the outliers and preparing the data for further analysis. one of the advantage of this software regarding with removal of an outliers, it provides segmentation command in different shape using polylines. Figure 2a,b, which depict the data with and with out the outliers respectively, are obtained from this software.

Validation Results
As described in previous sections, the specimen is scanned exactly after excitation, which leads to changes in shape followed by scanning without any loading. Once all specimen shape information is obtained for each load case, removal of the outliers and segmentation are carried out. The point cloud obtained from the experiments was processed using the proposed algorithm. Figure 2b depicts the resulting cloud data after preprocessing.

Selection of Optimal Parameters for GA
As shown in Table 1, eight control parameters at four values and one with two are identified in this study. Due to the number of parameters and their values considered in here, L32 (2 1 x4 8 ) orthogonal array is suitable for the experiment design. Orthogonal Arrays (OAs) provide a set of well balanced and minimal experiments. This array assumes that there is no interaction between any two factors. The experiments are repeated five times to increase the consistency of the experiment response. The analysis for the proposed Taguchi design is analyzed using a statistical software MINITAB19. Signal-to-noise ratios (SNR), which are log functions of the desired output as described in Equation (5), serve as objective functions for optimization, as well as help in data analysis and the prediction of optimum results. In the Taguchi method, the word "optimization" means determining the best values for the control factors. In turn, the best values for the control factors are those that maximize the signal-to-noise ratios. This can also be described as the resulting best values for the control factors so that the fitness functions are negligible, since the fitness functions are directly proportional to the error metric during plane fitting. Based on Taguchi two-step method rules, performance of a design is checked by maximizing the SN ratio and adjusting the mean to the target values. The Higher values of the Signal-to-noise ratio identify control factor settings that minimize the effects of the noise factor. In this study, the fitness value which is a response for the design experiment has a smaller the better characteristic. Figure 8 depicts the graphical representation of a robust level of parameter for the Genetic Algorithm. As a result, the optimal combination of control factors for the genetic algorithm to minimize the error is depicted in Table 3.

Main Effects Plot for SN ratios Data Means
Signal-to-noise: Smaller is better I Figure 8. Signal to noise ratio plot for each of the GA parameters.

Computation of Deflection Based on the Genetic Algorithm
Because of the nature of a lidar system, which is a collection of points, it is difficult to point out the edge of a specimen that would be considered as the "x-axis" of the local structural coordinate system. Therefore, representing the point cloud data by definable mathematical elements, planes, and lying one axis of a coordinate on a line which is obtained by correlating these planes are the easiest and optimal approach. A stochastic optimization, the genetic algorithm in this case, is utlized in this study to find the optimal planes that are most compatible with the data acquired. Even though this kind of optimization is prominently used for nonlinear forms and forms for which no derivative information exists [39], it also plays an enormous role in finding the optimal solution even in heavily noisy spaces [40]. Accordingly, the fitness function for the genetic algorithm is directly derived from the sum of the squares of the residuals, as shown in Equation (3).
Using the optimal control parameters shown in Table 3, the individual plane entity was fitted using the fitness function for the GA from the least square error (Equation (3)), as shown in Figure 5. The coefficients of the plane equation, i.e., the best individuals in genetic algorithm language, are obtained through a built-in MATLAB code, the Global Optimization toolbox, after determining the optimal parameter. By utilizing the Taguchi experimental design and obtaining the optimal parameter, the fitness values started to converge in the early generations (around the 3rd) and stopped at the 68th generation by satisfying the 1st stopping criteria (i.e., the average change in fitness value is less than the function tolerance). Figure 9 depicts the mean and best fitness values with respect to each generation and the best individual values for the flange entity plane.These best individuals are 0.0083 , −0.0257 and −0.5946. Therefore, the resulting planes which represent the acquired data for each entity of the specimen looks as follows: After obtaining the plane equation that represents the web part in a similar way to the flange part, the 3D intersection line L i , which is defined as the longitudinal axis of the local beam structural coordinates, is developed by considering Equation (7) parametrically. This gives us: Now, it is possible to position at least one axis of the new coordinate system, the x-axis in our case, along the intersection line described in the above figure. From Figure 5, the difference between points O and Q gives us the vector which lies on the new x-axis, called X . Before performing the rotation, we should determine the geometric correlation between the vector O Q and the old basis (X, Y, Z) after translating of point O to point O . Figure 10 shows the resulting point cloud data after performing the required transformation. This information is helpful in quantifying the deflection of a structure by applying the transformation matrix to the acquired point cloud data during the loading scenario.
Similar to the loading scenario, the fitness function for the loading case is obtained using Equation (13). A 2nd degree of polynomial curve is constructed for the fitness function considering the complexity degree of the structure based on given experimental study. This leads us to compute the best individuals similar as before, but here, we have five individual parameters for the curved surface, as illustrated in Figure 11. This provides the deflection shape of the top flange specimen as depicted in Figure 12. Even though our x-axis lies on the intersection between the flange and web as shown in Figure 5, the center line in Figure 12 indicates a larger deflection shape of the structure induced by the point load. Therefore, squeezing the plane towards the center transversely from each side gives us the required hypothetical deflection curve, which enables us to calculate the deflection of a specimen throughout its entire length. Here, only the deflection curve, which appears 4 mm at the center LVDT sensor, is shown diagrammatically. The rest of the deflections constructed using the proposed model along with their corresponding load contact sensor measurements are depicted in Figure 13. The resulting equation of center line C L from Figure 12 was calibrated by considering its 2nd degree curve behaviour. Some boundary conditions appear because of the support condition. As shown in Figure 6a, the support conditions used in this experiment lead us to determine the displacement and the moments at the support (i.e., δ (x=0,x=L) = 0 & M (x=0,x=L) = 0), such that the moment at the midpoint is a maximum.This calibration is needed because of the inherent random error of the instrument and the natural difference between the theoretical and experimental solutions. In the following, we illustrate how a signal equation can be calibrated by employing the boundary equation for the 4 mm loading scenario. The deflection curve obtained from the proposed analysis is as follows : Z = −2.138 × 10 −3 − 7.092 × 10 −3 · X + 3.925 × 10 −3 · X 2 According to basic beam theory, the max deflection for a simply supported beam must attained at the middle of the span. This implies that dZ/dx = 0, dZ/dx = 7.85 × 10 −3 · x − 7.092 × 10 −3 = 0 x = 0.90344 ......while, x must be 1. This tells us the resulting signal deviates by (1 − 0.90344) = 0.09656 from the real value. Therefore, the signal needs to be adjusted by +0.09656 along the x-direction and −0.002138 along the z-direction.

•
Since the specimen has transverse stiffener at 45 cm apart along both sides, it is very stiff along with the loading. However, the stiffness of a specimen is not uniform throughout the span with increasing load. A part of the specimen in between two stiffeners has not been equally disturbed with the part where exactly stiffener is welded. Consequently, the global deflection curve may not be expected exactly as a 2nd degree parabolic. • Secondly, the nature of a specimen has also its part in affecting the result for the edge LVDTs. Hence, the beam is labelled as a deep beam because of its span-to-depth ratio and the concentrated load with it. Therefore, the shear effect is predominant than flexural in our specimen. This implies that the deflection curve which is expected from the flexural effect is affected to some extent. Even the data shown for edge LVDTs are biased by this effect.
As a result, the proposed approach of representing the deflection shape in terms of a parabolic curve may only accurately capture the deflection formed throughout the whole span length except for around the mid-span. For emphasis, in the case of edge LVDTs in Figure 13a,c, the resulting deflection from the LVDT in each load scenario is almost the same as with the nominal one. However, this situation is incompatible with the proposed model since we are dealing with a deflection curve considered as a 2nd degree polynomial curve. It is obvious that if we consider a deflection curve as explained above while keeping the max deflection (at the center) equal to the nominal deflection, the deflection at the edge LVDT is less than the nominal one. Consequently, the proposed methodology is validated only for the middle LVDT sensor. As shown in Figure 13d, the proposed algorithm is accurate and effective for nominal deflections of 2 and 3 mm since the error is between ±1. For the case of nominal deflections of 1 and 4 mm, the error increases to ±4 and ±2.6, respectively.

Conclusions
In this work, we proposed a method for estimating the deflection of a beam based on a non-contact sensor, TLS. We simultaneously applied regression curve fitting with a genetic algorithm. The accustomed method, least squares regression fitting, is used to determine the fitness function which measures the suitability of the solution generated by the GA. The GA minimizes the fitness function obtained from the error equation. In addition, transformation of the coordinates is a crucial step in this study. This is because in order to define any geometrical changes in a structure with respect to its own axes, the scan data obtained from the device must also visualize the target structure with all its parts and elements so that periodic maintenance is predetermined. The equation for the curve is extracted from the surface fitting after transforming the coordinates of the data. This equation represents the deflection curve and, therefore, allows us to estimate the deflection value of a beam at any position across its span. The deflection value at various points was validated using the corresponding direct contact sensor, LVDT.
The proposed method is practical in applying the concept of structural health monitoring in different structures using non-contact sensors. Besides computation of a deflection, the geometric shape of a structure at the time of scanning can be visualized throughout this proposed methodology. This leads to regulating a structure health condition all over the area. As shown in the result, this algorithm is suitable for long-spanned beam. However, it is also effective in determining the deflection around the mid-span.

Conflicts of Interest:
The authors declare no conflict of interest.

Abbreviations
The following abbreviations are used in this manuscript: