Next Article in Journal
Underwater Spiral Wave Sound Source Based on Phased Array with Three Transducers
Next Article in Special Issue
Real-Time Water Surface Object Detection Based on Improved Faster R-CNN
Previous Article in Journal
Differential Run-Length Encryption in Sensor Networks
Previous Article in Special Issue
A Machine-Learning Approach to Distinguish Passengers and Drivers Reading While Driving
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Full-Waveform LiDAR Point Clouds Classification Based on Wavelet Support Vector Machine and Ensemble Learning

1
School of Remote Sensing and Information Engineering, Wuhan University, Wuhan 430079, China
2
Key Laboratory for National Geographic Census and Monitoring, National Administration of Surveying, Mapping and Geoinformation, Wuhan 430079, China
3
Beijing Key Laboratory of Urban Spatial Information Engineering, Beijing 100038, China
4
Institute of Geological Survey, China University of Geosciences, Wuhan 430074, China
*
Author to whom correspondence should be addressed.
Sensors 2019, 19(14), 3191; https://doi.org/10.3390/s19143191
Submission received: 13 June 2019 / Revised: 16 July 2019 / Accepted: 17 July 2019 / Published: 19 July 2019

Abstract

:
Light Detection and Ranging (LiDAR) produces 3D point clouds that describe ground objects, and has been used to make object interpretation in many cases. However, traditional LiDAR only records discrete echo signals and provides limited feature parameters of point clouds, while full-waveform LiDAR (FWL) records the backscattered echo in the form of a waveform, which provides more echo information. With the development of machine learning, support vector machine (SVM) is one of the commonly used classifiers to deal with high dimensional data via small amount of samples. Ensemble learning, which combines a set of base classifiers to determine the output result, is presented and SVM ensemble is used to improve the discrimination ability, owing to small differences in features between different types of data. In addition, previous kernel functions of SVM usually cause under-fitting or over-fitting that decreases the generalization performance. Hence, a series of kernel functions based on wavelet analysis are used to construct different wavelet SVMs (WSVMs) that improve the heterogeneity of ensemble system. Meanwhile, the parameters of SVM have a significant influence on the classification result. Therefore, in this paper, FWL point clouds are classified by WSVM ensemble and particle swarm optimization is used to find the optimal parameters of WSVM. Experimental results illustrate that the proposed method is robust and effective, and it is applicable to some practical work.

1. Introduction

The remote sensing technique is based on the sensor recorded energy that is reflected or emitted from the Earth’s surface [1]. Therefore, remote sensing can acquire information about ground objects without physical contact and can be divided into two categories: passive and active. Passive remote sensing data mostly exist in the form of spectral images, which makes it difficult for them to describe 3D space features and they usually require geometric correction.
Light Detection and Ranging (LiDAR) is a type of active remote sensing technique. Different from the 2D information displayed by spectral images, the advantage of LiDAR is to directly generate point clouds with large area coverage and high-precision coordinates [2]. In most cases, LiDAR point clouds are used to produce elevation data, such as the Digital Elevation Model (DEM) and Digital Surface Model (DSM), which is typically employed as ancillary information to assist passive remote sensed data in classification [3,4,5,6]. With the hardware equipment of LiDAR, more and more features are extracted for point cloud classification. According to the necessity of training samples, classification methods can be summarized as two types: unsupervised and supervised. Unsupervised methods are based on a suitable definition of similarity between data without any prior knowledge, and have been employed in the field of LiDAR point cloud classification [7], including K-means, iterative self-organizing data analysis (ISODATA), fuzzy c-means (FCM) [8,9,10], etc. However, the classification results may not correspond to the classes of ground objects and the classification accuracy of these methods has difficulty meeting practical requirements in some situations. Supervised methods, such as decision tree, random forest (RF), artificial neural network (ANN), and k-nearest neighbors (KNN) algorithm, have also been applied in the field of LiDAR point cloud classification [11,12,13,14]. They define class labels with prior knowledge deduced from training samples and the classification accuracy is improved by repeatedly modifying the training samples. However, these methods may have difficulty solving the classification problems with a small number of samples and are sensitive to feature correlation.
Different from traditional LiDAR, full-waveform LiDAR (FWL) is a new type, which records the backscattered echo in very small intervals and obtains a continuous echo waveform [15]. It is able to obtain more footprints, echoes and the actual situation of data collection. By decomposing full-waveform data, point clouds are calculated and waveform features such as the amplitude, width, peak location and intervals of peak location are extracted. The classification of LiDAR point clouds determines their categories and the commonly used features are intensity, geometry, texture and elevation. For FWL, except for the conventional features above, waveform features can be used in the classification process of point clouds, and the methods mentioned above are also applicable for FWL point cloud classification [16,17,18]. Moreover, the affiliation of waveform features in convention is helpful to improve the classification accuracy [17,19].
As a supervised method in machine learning, support vector machine (SVM) shows potential for effective and efficient classification of different types of data, and has a strong ability to solve a series of problems [20,21,22]. Because of its optimal hyper-plane in feature space based on structural risk minimization theory, SVM has become a widely used classification tool in need of a small amount of training data and fewer computational efforts [23,24,25], and it has been utilized in the field of FWL point cloud classification [26,27]. In SVM, the kernel function is an important factor affecting the classification results, while the commonly used liner, polynomial and sigmoid kernel functions may easily cause over-fitting or under-fitting in the lack of localization ability. Wavelet analysis is a powerful estimation technique for the time–frequency analysis of a signal [28], representing a signal by different resolutions. Therefore, wavelet functions, which have a good localization property, are used to build kernel functions employed in the construction of wavelet SVM (WSVM). In addition, the penalty factor and kernel function parameter play an important role in the classification result via SVM. To obtain better classification accuracy, parameter optimization is necessary, whose essence is a combinatorial optimization problem, and swarm intelligence algorithm can be used to solve this kind of problem because of its ability to obtain a satisfactory solution.
Moreover, under some circumstances, a single classifier is difficult to solve the case when the differences between feature values are very small and it may not label data correctly. In this case, ensemble learning is a promising research direction, and it is composed of a set of individual component classifiers (base classifiers) whose predictions are combined to determine the final results, improving the overall classification accuracy [29,30]. When SVMs are used as base classifiers, the ensemble system is called the SVM ensemble, which has been applied in many fields, such as hyperspectral data classification, dynamic financial distress prediction, credit card evaluation [31,32,33], etc. An ensemble learning system with different base classifiers highlights the heterogeneity, but ignores the homogeneity, while a system with the same base classifiers only focuses on the homogeneity and neglects the heterogeneity. In such a system based on WSVM, the base classifiers are composed of SVMs using wavelet kernel functions, which are in different forms but all considered on the basis of wavelet analysis and theoretically related. Therefore, the classification process for FWL point clouds are completed by WSVM ensemble in this paper, employing five variable wavelet kernel functions, and each classifier constructed by them, respectively, is used twice. At the same time, particle swarm optimization (PSO), which is a swarm intelligence algorithm with a good performance and stable convergence to the optimal solution, is utilized to optimize the SVM parameters.

2. Feature Extraction for FWL

According to the type of device, LiDAR point cloud classification can be conducted on the basis of many features, such as intensity, geometric, textural, and multispectral features. Compared with traditional LiDAR, FWL is able to extract more points, showing more details on structure. Therefore, geometric features are adopted to show the positional relationship between point clouds. However, it reflects the characteristics of discrete point clouds and neglects the interaction of the entire laser pulse with the ground objects. Hence, waveform features can be seen as a powerful complement to the classification process. For FWL, waveform decomposition of original data, whose essence is to fit the waveform with a function, is a useful technique to extract data information and should be performed before the process of classification. In this paper, before waveform decomposition, filtration of noise in the raw data [34] is conducted to reduce the noise interference for classification results, and then a set of Gaussian functions [35] is employed to fit the waveform. The relevant parameters of the function thus obtained are called waveform features of point clouds, reflecting the position and backscattering property of the targets during laser transmission.
Geometric features chosen for classification include the elevation (h), elevation standard deviation ( σ h ), volume density ( ρ ), curve (c), vertical angle ( N Z ), and vertical angle variance ( σ N Z ). Before the extraction of geometric features, a restricted 3D neighborhood ω is defined. A covariance matrix of 3D coordinates and its eigenvalues λ 1 , λ 2 , and λ 3 ( λ 1 < λ 2 < λ 3 ) are computed on the basis of 3D points in ω . The corresponding eigenvector of λ 1 is the normal vector ( N X , N Y , N Z ). Geometric features take advantage of the geometric shape intuitively and show geometric relationships between point clouds. However, these features cannot give the radiometric information expressing the reflectance properties of ground objects. While waveform features provide both geometric information and radiometric information, which is used to distinguish different surface materials. Furthermore, compared with the geometric features extracted from a specific neighborhood, the waveform features can reflect the characteristics of the pulse emission direction directly.
Waveform features extracted in this paper are the echo amplitude (A), width ( σ ), peak location (u) and intensity (I). Amplitude is the peak value of the echo signal, width is the standard deviation of each Gaussian component, peak location is the maximum position of the Gaussian function and intensity is the backscattered energy of a single Gaussian component, which is calculated by integrating each Gaussian component. All used features and their explanations are shown in Table 1. [19].

3. Methodology of LiDAR Point Clouds Classification

3.1. Construction of WSVM Model

SVM is a machine learning method which classifies data by creating a hyper-plane. Assume a set of training data X = { ( x 1 , y 1 ) , ( x 2 , y 2 ) , , ( x n , y n ) } , where X i R d , and y i { 1 , + 1 } is a class label. Then, the training process can be transformed into the optimization of the following expressions [36,37]:
m i n ϕ ( ω , ϵ ) = 1 2 | | ω | | 2 + C i = 1 n ϵ i
y i ( ω T x i + b ) 1 ϵ i
where ω is the normal vector of the hyper-plane, and ϵ i ( ϵ i 0 ) is a slack variable used for measuring classification errors. C is the penalty parameter, b is the bias or threshold, and ϕ is a function that maps the input data to a higher-dimensional space where data can be linearly separable. To avoid high-dimensional calculations, the mapping step can be skipped by kernel function, which is able to calculate the mapping directly. Additionally, the kernel function is defined as follows:
K ( x i , x j ) = ϕ ( x i ) T · ϕ ( x j )
The series of commonly used kernel functions are global functions rather than local functions, however SVM is an approximation model where a local function basis is better than a global basis [38,39]. The essence of wavelet analysis is to approximate signals using a set of functions generated by dilations and translations of a mother wavelet function [40], which means that a good localization property in the time and frequency domain is presented and the details of a signal are extracted [41]. It also has the advantages of a low redundancy, high stability, and well adaptability to high-dimensional data. The construction process of the wavelet kernel function is as follows [42,43]:
The one-dimensional wavelet function can be described as:
h ( x ) = i = 1 d h ( x i )
Then, the dot-product of the wavelet kernel is
K ( x , x ) = i = 1 d h ( x i c i a ) h ( x i c i a )
where h ( x ) is a mother wavelet, x , x R d , a is a dilation factor, and c is a translation factor. According to the translation invariant kernel theorem K ( x , x ) = K ( x x ) , Equation (5) can be transformed as:
K ( x , x ) = i = 1 d h ( x i x i a )
Then, the Gaussian, Shannon, Mexican Hat, Morlet, and Harmonic wavelet functions are used as mother wavelets to construct wavelet kernel functions and the wavelet kernel functions are defined as follows [44,45,46]:
Gaussian:
K ( x , x ) = i = 1 d e x p [ ( x i x i ) 2 2 σ 2 ]
Shannon:
K ( x , x ) = i = 1 d s i n ( π 2 × x i x i σ ) π 2 × x i x i σ c o s ( 3 π 2 × x i x i σ )
Mexican Hat:
K ( x , x ) = i = 1 d [ 1 ( x i x i ) 2 σ 2 e x p [ ( x i x i ) 2 2 σ 2 ]
Morlet:
K ( x , x ) = i = 1 d c o s ( 1.75 × x i x i σ ) e x p [ ( x i x i ) 2 2 σ 2 ]
Harmonic:
K ( x , x ) = i = 1 d e i 4 π x i x i σ e i 2 π x i x i σ i 2 π x i x i σ
Wavelet kernel functions are able to approximate arbitrary functions with a good localization property and have the ability of multi-scale analysis [47]. They amplify the difference between the values of the samples’ features and improve the stability of classifier model. Moreover, unlike the commonly used kernel functions that are correlative and redundant, the wavelet kernel functions are orthogonal [48]. These advantages mean that WSVM has a better generalization ability, higher accuracy, and lower computational complexity.

3.2. Parameter Optimization

SVM parameters have a great influence on the classification results. SVMs constructed by the above five wavelet kernel functions all have the penalty parameter C and kernel function parameter σ . C controls the generalization capacity of the classifier, while σ determines the distribution of data after mapping it into a new feature space. PSO is a population-based parallel search algorithm using a group of particles. It has been noticed that members of a group seem to share information among them, a fact that leads to increase the efficiency of the current group. A particle moves toward the optimum according to its present velocity, its previous experience, and the experience of its neighbors. In a n-D search space, the position and velocity of the ith particle are represented as vectors X i = x i , 1 , , x i , n and V i = v i , 1 , , v i , n , where each element is coded by real values. Let P b e s t i and G b e s t be the best position of the ith particle and the group’s best position thus far, respectively. The velocity and position of each particle are updated as follows [49,50]:
V i k + 1 = ω · V i k + r 1 · c 1 · ( P b e s t i k X i k ) + r 2 · c 2 · ( G b e s t k X i k )
X i k + 1 = X i k + V i k + 1
where V i k is the velocity of the ith particle at iteration k, ω is the inertia weight factor, c 1 and c 2 are the acceleration coefficients, r 1 and r 2 are random numbers between 0 and 1, and X i k is the position of the ith particle at iteration k. In the velocity updating process, the parameters such as ω , c 1 and c 2 should be determined in advance, which makes it cumbersome to solve large-scale optimization problems.
As PSO is used for SVM parameter optimization, whereby the particle whose vector value can acquire the highest classification accuracy is the optimal solution. Each dataset is divided into two parts for base classifiers: training and testing. Each iteration, the training samples are used for the construction of a base classifier, then a set of parameters is generated, and the testing samples are used to prove the validity of the parameters. After that, the classification accuracy of these parameters can be achieved by comparing the predicted and original labels. The optimal classification accuracy obtained in the current generation is recorded and compared with the highest classification accuracy in the last iteration. If the former wins, the optimal solution is replaced with the current solution. When the maximum number of iterations is reached, the final parameters are obtained.

3.3. WSVM Ensemble

Ensemble learning trains a series of base classifiers and combines their outputs by a fusion strategy, improving the generalization performance [51]. As mentioned in Section 1, an ensemble system that uses SVMs as base classifiers is called SVM ensemble [52]. Considering the heterogeneity and homogeneity of the system [53], while reducing the running time, the bagging algorithm is adopted to aggregate the base classifiers and five kernel functions mentioned in Section 3.1 are utilized to construct SVMs. Furthermore, each kernel function is used twice and a total of 10 WSVMs are utilized for the ensemble learning.
After each WSVM has output a predicted label, it is important to adopt a fusion strategy that determines the final label of the ensemble system. In this paper, majority voting is employed as the fusion strategy. Each classifier generates a predicted label, which is taken as a vote, and the final result is determined by the vote of each classifier. For example, there is a classification problem with n classes ( Y = 1 , 2 , , n ) , a training dataset, an instance x to be classified and T classifiers. Each classifier outputs a label L i of x, where i = 1 , 2 , , T and L i Y . Let V be the final label of x, then
V = a r g m a x i = 1 T p ( L i = y )
where y Y ; if a is true, p ( a ) = 1 , otherwise p ( a ) = 0 and a is the output of base classifier.
Majority voting treats the output of each classifier equally, determining the final result by counting the number of occurrences for each class. When encountering the same number of occurrences, in cases where it is difficult to decide, a random selection of these classes is adopted to determine the final result.
Following the feature extraction in Section 2, bootstrap sampling is used to generate sub-datasets from the original data, and the number of instances in each sub-dataset is not greater than that in the original data. Then, on the basis of the corresponding sub-dataset, each SVM obtains the optimal parameters by the continuous process of training and testing samples combined with PSO. After that, each base classifier predicts the labels with these parameters. Finally, majority voting is utilized to generate the final result. The schematic diagram of the proposed method is shown in Figure 1.

3.4. Implementation of the Proposed Method

The classification process of FWL point clouds is completed in this paper, employing WSVM ensemble. The detailed steps of the proposed method are as follows:
  • Step 1: Acquire FWL data and filter noise in the data.
  • Step 2: Decompose full-waveform LiDAR data and extract the features displayed in Table 1.
  • Step 3: Use bootstrap sampling to generate sub-datasets and assign them to each base classifier.
  • Step 4: Train base classifiers with the parameters of each particle in the population.
  • Step 5: Obtain the classification accuracy of each particle and update the population using Equations (12) and (13).
  • Step 6: Update global optimal accuracy and corresponding parameters.
    Step 6.1:
    Compare the classification accuracy of the particles in the current generation with the global optimal accuracy.
    Step 6.2:
    Determine whether to update the global optimal accuracy and the corresponding parameters on the basis of comparison result.
    Step 6.3:
    Return to Step 4 until the number of iterations has reached the maximum value.
  • Step 7: Save the parameters achieved by the global optimal accuracy of each base classifier according to Steps 4–6, and take them as the parameters of each classifier.
  • Step 8: Each base classifier predicts the labels of data with their parameters.
  • Step 9: Output the final results with majority voting.

4. Experimental Results and Discussion

The experimental environment in this study was a computer with a 2.30GHz CPU and 8G of RAM. The data-processing operation was realized using MATLAB 2016a and VS2017 software. The manual classification process was accomplished using LiDAR software and visual interpretation by researchers with relevant working experience.

4.1. Experimental Platform and Data Information

The data used in this paper were acquired by airborne LiDAR system ALS60, in 2009, and in the form of Las 1.3. There were three study areas and the experimental data were colored according to elevation values, as shown in Figure 2a–c. Study Area 1 is flat and open, with a low point cloud density; Study Area 2 is a dense residential area; and Study Area 3 is mainly distributed with large buildings. The number of points, training and testing samples of the study areas and some other information about the study areas are shown in Table 2, and training and testing samples were generated by manual classification. To quicken the process, the number of points in each training subset generated by bootstrap sampling was one-tenth of the whole dataset. After the classification for all of the point clouds was completed, testing samples of these three study areas were used to validate the classification accuracy of the ensemble system.

4.2. Classification Results for Point Clouds

The classification results of the proposed classification method for FWL point clouds were compared with those of basic SVM, optimal single WSVMs, RF, ISODATA, and WSVM ensemble without parameter optimization. The classification results are shown in Figure 3, Figure 4 and Figure 5, while the CPU time of each method is shown in Table 3 (unit: s), and the classification accuracy of each method is shown in Table 4. The ground objects were classified into low vegetations (blue color), trees (green color), buildings (yellow color) and others (red color). Low vegetations includes grass and crops, while others includes mainly artificial surfaces such as roads and concrete floors.
As shown in Table 3, the CPU time of WSVM ensemble without parameter optimization was lower than any other compared methods, and those of RF and optimal WSVMs were relatively high. The CPU time of the proposed method was 85.0167, 78.4931 and 97.0145 s, respectively, which was significantly reduced when compared to optimal single WSVM. As shown in Table 4, the proposed method performed better than RF and ISODATA; it also obviously outperformed basic SVM and WSVM ensemble without parameter optimization, which illustrates the importance of SVM kernel function and parameter optimization. In addition, the classification accuracy of optimal single WSVMs proved the applicability of WSVM in such classification problems. Finally, from the perspective of the ensemble, the classification accuracy of WSVM ensemble was higher than optimal single WSVMs used in the ensemble system. The classification accuracies via optimal single WSVMs of the three study areas were 96.5720%, 94.5937% and 92.1143%, which were 1.1757%, 0.5018% and 1.7179% lower, respectively, than those of the proposed method. Considering its CPU time, as presented in Table 3, this method has good application prospect.
As shown in Figure 3, Figure 4 and Figure 5, the results of ISODATA and basic SVM had obvious misclassifications, which proved that unsupervised classification and improper kernel functions of SVM have difficulty distinguishing the objects with approximate feature values. It was also found that the results of WSVM ensemble without parameter optimization had some noise (see, for example, left center of Figure 3e and the left half of Figure 5e), which illustrates that the random parameters may be not suitable in such classification problems. Optimal single WSVMs and the proposed method had similar results, but further comparison of these two methods’ results revealed that the latter method had better classification ability for details. As shown in Figure 3, the road in the lower left corner extracted by the proposed method is more complete than optimal single WSVM. As shown in Figure 5, the low vegetations in the lower left corner are misclassified as others by optimal single WSVM, while these errors do not exist in the result of the proposed method.

5. Conclusions

A FWL point cloud classification method employing WSVM ensemble is proposed in this paper, which utilizes five different wavelet kernel functions and ensemble learning. In the process of classification, geometric and waveform features are adopted, and bagging algorithm is used to integrate the base classifiers. Further, in each base classifier, PSO is employed to optimize WSVM parameters to obtain satisfactory classification results. The classification accuracy of all study areas was over 93%, which proved the feasibility of the proposed method. In addition, it was higher than that of basic SVM, optimal single WSVMs, RF, ISODATA and WSVM ensemble without parameter optimization. Moreover, in the case of a similar classification accuracy, the CPU time of the WSVM ensemble was less than single WSVMs. In conclusion, the proposed method can acquire accurate and effective results of FWL point cloud classification and has strong application potential for large-scale classification problems. On the basis of this work, future research such as the optimization of FWL feature extraction will be conducted.

Author Contributions

Writing–original draft preparation, X.L.; supervision, X.L.; methodology, Y.Y.; investigation, Y.L.; Validation, M.W.; writing–review and editing, X.L.

Funding

This work was funded by the National Key Research & Development Program of China under Grant No. 41771368, the Beijing Key Laboratory of Urban Spatial Information Engineering under Grant No. 2019204, the Technical Research Service of Airborne LiDAR Data Acquisition and Digital Elevation Model Updating Project in Guangdong Province under Grant No. 0612-1841D0330175 and the Airborne LiDAR Data Acquisition and Digital Elevation Model Updating in Guangdong Province under Grant No. GPCGD173109FG317F.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Magiera, J. Can Satellite Remote Sensing Be Applied in Geological Mapping in Tropics? E3S Web Conf. 2018, 35, 1–9. [Google Scholar] [CrossRef]
  2. Maltezos, E.; Doulamis, A.; Doulamis, N. Building Extraction From LiDAR Data Applying Deep Convolutional Neural Networks. IEEE Geosci. Remote Sens. Lett. 2018, 16, 155–159. [Google Scholar] [CrossRef]
  3. Ekhtari, N.; Glennie, C.; Fernandez-Diaz, J.C. Classification of Airborne Multispectral LiDAR Point Clouds for Land Cover Mapping. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 2068–2078. [Google Scholar] [CrossRef]
  4. Elaksher, A.F. Fusion of Hyperspectral Images and LiDAR-based Dems for Coastal Mapping. Opt. Lasers Eng. 2008, 46, 493–498. [Google Scholar] [CrossRef]
  5. Bork, E.W.; Su, J.G. Integrating LIDAR Data and Multispectral Imagery for Enhanced Classification of Rangeland Vegetation: A Meta Analysis. Remote Sens. Environ. 2007, 111, 11–24. [Google Scholar] [CrossRef]
  6. Campos-Taberner, M.; Romero-Soriano, A.; Gatta, C.; Camps-Valls, G.; Lagrange, A.; Saux, B.L.; Beaupère, A.; Boulch, A.; Chan-Hon-Tong, A.; Herbin, S.; et al. Processing of extremely high-resolution LiDAR and RGB data: Outcome of the 2015 IEEE GRSS Data Fusion Contest–Part A: 2-D Contest. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2016, 9, 5547–5559. [Google Scholar] [CrossRef]
  7. Langer, H.; Falsaperla, S.; Masotti, M.; Campanini, R.; Spampinato, S.; Messina, A. Synopsis of Supervised and Unsupervised Pattern Classification Techniques Applied to Volcanic Tremor Data at Mt Etna, Italy. Geophys. J. Int. 2009, 178, 1132–1144. [Google Scholar] [CrossRef]
  8. Chehataa, N.; Davidb, N.; Bretar, F. LiDAR Data Classification Using Hierarchical K-means Clustering. In Proceedings of the ISPRS Congress, Beijing, China, 3–11 July 2008; pp. 325–330. [Google Scholar]
  9. Hou, T.; Liu, J.; Liu, Y. Algorithmic Clustering of LiDAR Point Cloud Data for Textural Damage Identifications of Structural Elements. Measurement 2017, 108, 77–90. [Google Scholar] [CrossRef]
  10. Gu, Z.; Cao, S.; Sanchez-Azofeifa, G.A. Using LiDAR Waveform Metrics to Describe and Identify Successional Stages of Tropical dry forests. Int. J. Appl. Earth Obs. Geoinf. 2018, 73, 482–492. [Google Scholar] [CrossRef]
  11. Vo, A.; Linh, T.; Laefer, D.F.; Tiede, D.; Oleire-Oltmanns, S.; Baraldi, A.; Shimoni, M.; Moser, G.; Tuia, D. Processing of Extremely High Resolution LiDAR and RGB Data: Outcome of the 2015 IEEE GRSS Data Fusion Contest—Part B: 3-D Contest. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2015, 9, 5560–5575. [Google Scholar] [CrossRef]
  12. Ni, H.; Lin, X.; Zhang, J. Classification of ALS Point Cloud with Improved Point Cloud Segmentation and Random Forests. Remote Sens. 2017, 9, 288. [Google Scholar] [CrossRef]
  13. Qiao, J.; Liu, X.; Zhang, Y.H. Land Cover Classification Using LiDAR Height Texture and ANNs. J. Remote Sens. 2011, 15, 539–553. [Google Scholar]
  14. Penner, M.; Pitt, D.G.; Woods, M.E. Parametric vs. Nonparametric LiDAR Models for Operational Forest Inventory In Boreal Ontario. Can. J. Remote Sens. 2013, 39, 426–443. [Google Scholar]
  15. Mallet, C.; Bretar, F. Full-waveform topographic LiDAR: State-of-the-art. ISPRS J. Photogramm. Remote Sens. 2009, 64, 1–16. [Google Scholar] [CrossRef]
  16. Wang, J.; Li, C.; Tang, L.; Zhou, M.; Li, J.M. A Comparison of Two Different Approaches of Point Cloud Classification Based on Full-waveform LIDAR Data. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2012, 39, 179–182. [Google Scholar] [CrossRef]
  17. Guo, L.; Chehata, N.; Mallet, C.; Boukir, S. Relevance of Airborne Lidar and Multispectral Image Data for Urban Scene Classification Using Random Forests. ISPRS J. Photogramm. Remote Sens. 2011, 66, 56–66. [Google Scholar] [CrossRef]
  18. Höfle, B.; Hollaus, M.; Hagenauer, J. Urban Vegetation Detection Using Radiometrically Calibrated Small-footprint Full-waveform Airborne LiDAR Data. ISPRS J. Photogramm. Remote Sens. 2012, 67, 134–147. [Google Scholar] [CrossRef]
  19. Mallet, C.; Bretar, F.; Roux, M.; Soergel, U.; Heipke, C. Relevance Assessment of Full-waveform Lidar Data for Urban Area Classification. ISPRS J. Photogramm. Remote Sens. 2011, 66, 71–84. [Google Scholar] [CrossRef]
  20. Maxwell, A.E.; Warner, T.A.; Fang, F. Implementation of Machine-learning Classification in Remote Sensing: An Applied Review. Int. J. Remote Sens. 2018, 39, 2784–2817. [Google Scholar] [CrossRef]
  21. Phu, V.N.; Chau, V.T.N.; Tran, V.T.N. SVM for English Semantic Classification in Parallel Environment. Int. J. Speech Technol. 2017, 20, 487–508. [Google Scholar] [CrossRef]
  22. Sukawattanavijit, C.; Jie, C.; Zhang, H. GA-SVM Algorithm for Improving Land-Cover Classification Using SAR and Optical Remote Sensing Data. IEEE Geosci. Remote Sens. Lett. 2017, 14, 284–288. [Google Scholar] [CrossRef]
  23. Santos, P.; Villa, L.F.; Reñones, A.; Bustillo, A.; Maudes, J. An SVM-Based Solution for Fault Detection in Wind Turbines. Sensors 2015, 15, 5627–5648. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  24. Dihkan, M.; Guneroglu, N.; Karsli, F.; Guneroglu, A. Remote Sensing of Tea Plantations Using An SVM Classifier and Pattern-based Accuracy Assessment Technique. Int. J. Remote Sens. 2013, 34, 8549–8565. [Google Scholar] [CrossRef]
  25. Seifi, M.R.; Ghassemian, H. A Probabilistic SVM Approach for Hyperspectral Image Classification Using Spectral and Texture Features. Int. J. Remote Sens. 2017, 38, 4265–4284. [Google Scholar] [CrossRef]
  26. Zhou, M.; Li, C.; Ma, L.; Guan, H.C. Land Cover Classification from Full-waveform LiDAR Data Based on Support Vector Machines. In Proceedings of the International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences, ISPRS Congress, Prague, Czech Republic, 12–19 July 2016; pp. 447–452. [Google Scholar]
  27. Tseng, Y.; Wang, C.; Chu, H.J.; Hung, Y.C. Waveform-based Point Cloud Classification in Land-cover Identification. Int. J. Appl. Earth Obs. Geoinf. 2015, 34, 78–88. [Google Scholar] [CrossRef]
  28. Kedzierski, M.; Fryskowska, A. Terrestrial and Aerial Laser Scanning Data Integration Using Wavelet Analysis for the Purpose of 3D Building Modeling. Sensors 2014, 14, 12070–12092. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  29. Krawczyk, B.; Minku, L.L.; Gama, J.; Stefanowski, J.; Woźniak, M. Ensemble Learning for Data Stream Analysis: A Survey. Inf. Fusion 2017, 37, 132–156. [Google Scholar] [CrossRef]
  30. Dou, P.; Chen, Y. Remote Sensing Imagery Classification Using AdaBoost with A Weight Vector. Remote Sens. Lett. 2017, 8, 733–742. [Google Scholar] [CrossRef]
  31. Waske, B.; Van, L.S.; Benediktsson, J.A.; Rabe, A.; Hostert, P. Sensitivity of Support Vector Machines to Random Feature Selection in Classification of Hyperspectral Data. IEEE Trans. Geosci. Remote Sens. 2010, 48, 2880–2889. [Google Scholar] [CrossRef] [Green Version]
  32. Sun, J.; Fujita, H.; Chen, P.; Li, H. Dynamic Financial Distress Prediction with Concept Drift Based on Time Weighting Combined with Adaboost Support Vector Machine Ensemble. Knowl. Based Syst. 2017, 120, 4–14. [Google Scholar] [CrossRef]
  33. Yu, L.; Yue, W.; Wang, S.; Lai, K.K. Support Vector Machine Based Multiagent Ensemble Learning for Credit Risk Evaluation. Expert Syst. Appl. 2010, 37, 1351–1360. [Google Scholar] [CrossRef]
  34. Lai, X.; Zheng, M. A Denoising Method for LiDAR Full-Waveform Data. Math. Probl. Eng. 2015, 2015, 164318. [Google Scholar] [CrossRef]
  35. Lai, X.; Qin, N.N.; Han, X.S.; Wang, J.H.; Hou, W.G. Iterative Decomposition Method for Small Foot-print LiDAR Waveform. J. Infrared Millim. Waves 2013, 32, 319–324. [Google Scholar] [CrossRef]
  36. Wang, M.; Wan, Y.; Ye, Z.; Lai, X.D. Remote Sensing Image Classification Based on the Optimal Support Vector Machine and Modified Binary Coded Ant Colony Optimization Algorithm. Inf. Sci. 2017, 402, 50–68. [Google Scholar] [CrossRef]
  37. Tharwat, A. Parameter Investigation of Support Vector Machine Classifier with Kernel Functions. Knowl. Inf. Syst. 2019, 3, 1–34. [Google Scholar] [CrossRef]
  38. Xie, Z.P.; Chen, D.S.; Chen, S.C.; Qiao, L.S.; Yang, B. A Tight Support Kernel for Support Vector Machine. In Proceedings of the International Conference on Wavelet Analysis and Pattern Recognition, Hong Kong, China, 30–31 August 2008; pp. 460–464. [Google Scholar]
  39. Wang, A.; Zhao, Y.; Hou, Y.T.; Li, Y.L. A Novel Construction of SVM Compound Kernel Function. In Proceedings of the International Conference on Logistics Systems and Intelligent Management (ICLSIM), Harbin, China, 9–10 January 2010; pp. 1462–1465. [Google Scholar]
  40. Zhang, L.; Zhou, W.; Jiao, L. Wavelet Support Vector Machine. IEEE Trans. Syst. Man Cybern. 2004, 34, 34–39. [Google Scholar] [CrossRef]
  41. Xia, C.; Zhang, M.; Cao, J. A Hybrid Application of Soft Computing Methods with Wavelet SVM and Neural Network to Electric Power Load Forecasting. J. Electr. Syst. Inf. Technol. 2018, 5, 681–696. [Google Scholar] [CrossRef]
  42. García, P.J.; García-Gonzalo, E.; Alonso, J.R.; Muñiz, C.D. A Hybrid Wavelet Kernel SVM-based Method Using Artificial Bee Colony Algorithm for Predicting the Cyanotoxin Content from Experimental Cyanobacteria Concentrations in the Trasona Reservoir (Northern Spain). J. Comput. Appl. Math. 2017, 309, 587–602. [Google Scholar] [CrossRef]
  43. Liu, Z.; Cui, Y.; Li, W. A Classification Method for Complex Power Quality Disturbances Using EEMD and Rank Wavelet SVM. IEEE Trans. Smart Grid 2017, 6, 1678–1685. [Google Scholar] [CrossRef]
  44. Tang, B.; Song, T.; Li, F.; Deng, L. Fault Diagnosis for A Wind Turbine Transmission System Based on Manifold Learning and Shannon Wavelet Support Vector Machine. Renew. Eng. 2014, 62, 1–9. [Google Scholar] [CrossRef]
  45. Wang, P.; Gan, X.; Yao, D. Anomaly Intrusion Detection Based on Support Vector Machine with Mexico Hat Wavelet Kernel Function. Appl. Mech. Mater. 2014, 687, 3897–3900. [Google Scholar] [CrossRef]
  46. Domenico, C.; Alessandro, P. Spectrum-compatible Accelerograms with Harmonic Wavelets. Comput. Struct. 2015, 147, 26–35. [Google Scholar]
  47. Yang, G. License Plate Character Recognition Based on Wavelet Kernel LS-SVM. In Proceedings of the International Conference on Computer Research and Development, Shanghai, China, 11–13 March 2011; pp. 222–226. [Google Scholar]
  48. Wu, D.; Gu, X.; Guo, Q. A New Scheme of Automatic Modulation Classification Using Wavelet and WSVM. In Proceedings of the International Conference on Mobile Technology, Guangzhou, China, 15–17 November 2005. [Google Scholar]
  49. Eberhart, R.; Kennedy, J. A New Optimizer Using Particle Swarm Theory. In Proceedings of the Sixth International Symposium on Micro Machine and Human Science, Nagoya, Japan, 4–6 October 1995; pp. 39–43. [Google Scholar]
  50. Wang, D.; Tan, D.; Liu, L. Particle Swarm Optimization Algorithm: An Overview. Soft Comput. 2018, 22, 387–408. [Google Scholar] [CrossRef]
  51. Li, L.; Hu, Q.; Wu, X.; Yu, D. Exploration of Classification Confidence in Ensemble Learning. Pattern Recognit. 2014, 47, 3120–3131. [Google Scholar] [CrossRef]
  52. Wang, S.; Mathew, A.; Yan, C.; Li, F.X.; Ma, L.; Lee, J. Empirical Analysis of Support Vector Machine Ensemble Classifiers. Expert Syst. Appl. 2009, 36, 6466–6476. [Google Scholar] [CrossRef]
  53. Bi, Y. The Impact of Diversity on the Accuracy of Evidential Classifier Ensembles. Int. J. Approx. Reason. 2012, 53, 584–607. [Google Scholar] [CrossRef]
Figure 1. Schematic diagram of the proposed method.
Figure 1. Schematic diagram of the proposed method.
Sensors 19 03191 g001
Figure 2. Original data of the three study areas.
Figure 2. Original data of the three study areas.
Sensors 19 03191 g002
Figure 3. Classification results of Study Area 1.
Figure 3. Classification results of Study Area 1.
Sensors 19 03191 g003
Figure 4. Classification results of Study Area 2.
Figure 4. Classification results of Study Area 2.
Sensors 19 03191 g004
Figure 5. Classification results of Study Area 3.
Figure 5. Classification results of Study Area 3.
Sensors 19 03191 g005
Table 1. Features for classification.
Table 1. Features for classification.
Feature TypeFeature NameFormulaExplanation
Geometrich/In general, elevation can effectively distinguish between ground and off-ground points, but, for trees, houses, and hillsides with similar elevations, absolute elevation may not work.
σ h σ h = i = 1 n ( Z Z a v e ) 2 n 1 Z denotes the current point, Z a v e denotes the average elevation of all points in ω , and n is the number of points in ω . High vegetation and the edges of buildings often have a greater height difference.
ρ ρ = n / v ω v ω is the volume of ω . Generally speaking, ρ of building walls and trees is lower than others.
c c = λ 1 / ( λ 1 + λ 2 + λ 3 ) c reflects the shape of the surface of the object, and the canopy usually has a high value.
N Z /Deviation angle of a normal vector from the vertical direction, reflecting the flatness of the ground object.
σ N Z σ N Z = N X 2 + N Y 2 N Z The variance of the vertical angles of 3D points in ω , reflecting the shape of the ground object.
WaveformA/The value of the natural surface and building is the highest, and that of the asphalt surface and trees is low, thus it can distinguish between vegetation and artificial objects.
σ / σ reflects the time that the laser pulse interacts with the ground object. Due to the scattering effect of the canopy on the laser, σ can distinguish between non-vegetation and vegetation.
u/u can be used to calculate the distance between the laser emission location and the target.
I/I is the amount of energy returned by the laser pulse interacting with the ground objects whose characteristic is similar to A.
Table 2. Experimental data information.
Table 2. Experimental data information.
ExperimentalData AreaTotalTrainingTestingPoint Cloud
Data ( m 2 ) PointsSamplesSamplesDensity
Study Area 1203,833227,078645056831.11
Study Area 2180,030283,315648950701.57
Study Area 3131,767226,123673953991.72
Table 3. CPU time of different methods (s).
Table 3. CPU time of different methods (s).
ExperimentalBasicOptimalRFISODATANon-OptimizationProposed
DataSVMWSVM Method
Study Area 178.6326123.0621139.952782.209869.664285.0167
Study Area 273.5228111.6443123.894376.375466.002978.4931
Study Area 388.9489130.5389152.011293.082477.413197.0145
Table 4. Classification accuracy of different methods (%).
Table 4. Classification accuracy of different methods (%).
ExperimentalBasicOptimal SingleRFISODATANon-OptimizationProposed
DataSVMWSVMs Method
Study Area 166.074396.572094.823655.794393.495697.7477
Study Area 255.570294.593788.201777.683790.332495.0955
Study Area 366.850592.114393.146975.788484.428293.8322

Share and Cite

MDPI and ACS Style

Lai, X.; Yuan, Y.; Li, Y.; Wang, M. Full-Waveform LiDAR Point Clouds Classification Based on Wavelet Support Vector Machine and Ensemble Learning. Sensors 2019, 19, 3191. https://doi.org/10.3390/s19143191

AMA Style

Lai X, Yuan Y, Li Y, Wang M. Full-Waveform LiDAR Point Clouds Classification Based on Wavelet Support Vector Machine and Ensemble Learning. Sensors. 2019; 19(14):3191. https://doi.org/10.3390/s19143191

Chicago/Turabian Style

Lai, Xudong, Yifei Yuan, Yongxu Li, and Mingwei Wang. 2019. "Full-Waveform LiDAR Point Clouds Classification Based on Wavelet Support Vector Machine and Ensemble Learning" Sensors 19, no. 14: 3191. https://doi.org/10.3390/s19143191

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop