Next Article in Journal
Extraction of Information about Individual Trees from High-Spatial-Resolution UAV-Acquired Images of an Orchard
Previous Article in Journal
Antarctic Supraglacial Lake Detection Using Landsat 8 and Sentinel-2 Imagery: Towards Continental Generation of Lake Volumes
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-View Features Joint Learning with Label and Local Distribution Consistency for Point Cloud Classification

1
College of Information Science and Engineering, Northeastern University, Shenyang 110819, China
2
College of Civil Engineering, Nanjing Forestry University, Nanjing 210037, China
3
Department of Geomatics Engineering, University of Calgary, Calgary, AB T2N 1N4, Canada
4
Department of Mathematics and Computing Science, Saint Mary’s University, Halifax, NS B3P 2M6, Canada
5
School of Land Science and Technology, China University of Geosciences, Beijing 100083, China
*
Authors to whom correspondence should be addressed.
The first three authors share the co-first authorship, and contribute equally to the manuscript.
Remote Sens. 2020, 12(1), 135; https://doi.org/10.3390/rs12010135
Submission received: 19 November 2019 / Revised: 26 December 2019 / Accepted: 26 December 2019 / Published: 1 January 2020

Abstract

:
In outdoor Light Detection and Ranging (lidar)point cloud classification, finding the discriminative features for point cloud perception and scene understanding represents one of the great challenges. The features derived from defect-laden (i.e., noise, outliers, occlusions and irregularities) and raw outdoor LiDAR scans usually contain redundant and irrelevant information which adversely affects the accuracy of point semantic labeling. Moreover, point cloud features of different views have a capability to express different attributes of the same point. The simplest way of concatenating these features of different views cannot guarantee the applicability and effectiveness of the fused features. To solve these problems and achieve outdoor point cloud classification with fewer training samples, we propose a novel multi-view features and classifiers’ joint learning framework. The proposed framework uses label consistency and local distribution consistency of multi-space constraints for multi-view point cloud features extraction and classification. In the framework, the manifold learning is used to carry out subspace joint learning of multi-view features by introducing three kinds of constraints, i.e., local distribution consistency of feature space and position space, label consistency among multi-view predicted labels and ground truth, and label consistency among multi-view predicted labels. The proposed model can be well trained by fewer training points, and an iterative algorithm is used to solve the joint optimization of multi-view feature projection matrices and linear classifiers. Subsequently, the multi-view features are fused and used for point cloud classification effectively. We evaluate the proposed method on five different point cloud scenes and experimental results demonstrate that the classification performance of the proposed method is at par or outperforms the compared algorithms.

Graphical Abstract

1. Introduction

In recent years, with the rapid advancement of computer vision and Light Detection and Ranging (lidar) technology, an increasing number of point clouds are acquired and widely used in various remote-sensing applications. In applications such as autonomous driving, understanding the outdoor scenes by semantic labeling of point clouds has become a hot topic [1,2,3,4]. Point cloud classification is to mark a specific semantic attribute label for each point in the point cloud [2], which is a key step in environmental perception and scene understanding. Due to disorder, sparsity and irregularity of point clouds as well as the possible presence of uncertainty in point clouds (noise, outliers and missing data), the intra-class points in the same scene are quite different and the difference of inter-class points is not obvious [3]. Therefore, effective classification of point clouds is a challenging problem.
For the past few years, the classification algorithms proposed in [3,4,5,6,7,8,9] have achieved good performances for classifying images and point clouds. For example, Zhang et al. [10] proposed DKSVD (discriminative K-SVD [11]) algorithm, which introduces classification error to optimize feature extraction and classifier, simultaneously. To explore prior knowledge of label information, Jiang et al. [12] introduced label consistency constraints into the objective functions of LCKSVD1 (Label Consistent K-SVD) and LCKSVD2 algorithms, and showed better classification results. Zhang et al. [1] used discriminant dictionary learning to construct multi-level point set features for point cloud classification. Li et al. [4] proposed a deep-learning network based on multi-level voxel features fusion for point cloud classification. The feature dimensions used by the above methods are relatively high, and are found to be carrying noise and redundant information [13]. To overcome this drawback, dimensionality reduction and sparse representation are widely used. Dimensionality reduction can be seen as a special case of subspace learning, which projects high-dimensional data to low-dimensional subspace through algorithms, such as ICA (independent component analysis) [14], PCA (principal component analysis) [15], optimal mean robust PCA [16] and other variants.
In addition, most supervised classification methods usually require a huge amount of training samples to learn features and classifiers to achieve very high classification accuracy. As the training set generation through point cloud labeling is time-consuming, it significantly lowers the algorithmic efficiency [17]. Therefore, if we can successfully classify a large volumes of point clouds using only a small percentage of training samples, it should have a great practical applicability because the time and labor cost will be significantly reduced [5]. To solve this problem, [13,17,18,19] proposed semi-supervised or supervised classification methods for feature transformation matrix and classifier joint learning. For example, Mei et al. [17] concatenated multiple single point features to form high-dimensional features for each point, and then used the joint constraints of margin, graph of adjacency and labels to train the model with a small portion of samples based on a semi-supervised framework. Zhu et al. [18] used the feature vector composed of multi-scale features in a series of images to express each image, and then the constraints of local connection relations of labels and samples were introduced to jointly learn the features projection matrix and classifier with local and global consistency. These methods directly fuse different types of sample features or the same features defined at multiple scales for classification. However, this kind of feature fusion and its variants have relatively limited expression ability of sample attributes and classification performances. In this case, the effectiveness of feature fusion cannot be guaranteed.
To express and classify multimedia data effectively, researchers have proposed a variety of multi-view learning algorithms. Each point can be described by high-dimensional features from multiple views, e.g., eigenvalue features of covariance matrix [1], spin image features [20], normal vector, FPFH (fast-point feature histogram) [21] and VFH (viewpoint feature histogram) [22]. The created features in each view contain unique information that is different from the features derived from other views. Meanwhile, it should be noted that multi-view features also include some overlapping information to a certain extent, although they are generalized from different views. Different view features describe properties of different aspects of a point, but they commonly represent the same point. Typically, the multi-view learning methods [23,24,25,26,27,28,29,30,31] can effectively fuse features from different views. The authors leverage this diversity and consistency of different view features to obtain more discriminative feature representation. More specifically, Nie et al. [25] proposed an adaptive weighted multi-view learning algorithm (MLAN) for image clustering and semi-supervised classification. This algorithm introduces different view weights and learns the local structure from different view data based on a manifold learning method. In [28], the features’ low-rank representations of each view are jointly optimized by introducing the exclusivity and category consistency constraints. In contrast to [28], the method in [31] jointly learned the low-rank representations of multi-view by introducing the error term with adaptive weight of each view and the diversity regular term for reducing redundant information between different views. Then, the joint projection graph of the multi-view was constructed by the low-rank representations for clustering/classification. All these methods outperform the single-view feature-learning methods, however, they are not applicable for classifying outdoor point clouds. By contrast with the above multi-view learning methods, some multi-view Convolutional Neural Networks (CNNs) based methods (deep learning mechanism) for point cloud processing are proposed in recent years [32]. Generally, this kind of method relies on 2D rendered views instead of on the 3D data. For example, Su et al. [33] use multi-view CNNs to extract the features of 2D renderings from a 3D object, which shows good performance for 3D object model classification. For point cloud semantic segmentation of outdoor scenes, multi-view CNN-based methods, e.g., [34,35,36], need to render point cloud to generate multi-view images with multi-modal representation, which can include the information of depth, color, normal and other features. Then the generated multiple images are semantically segmented by the networks, e.g., U-Net [37], SegNet [38] and Fully Convolutional Network (FCN) [39]. After that, the semantic segmentation results of multiple images are projected to meshes to jointly determine the label of each mesh vertex. Finally, the labeled vertices are projected to the original point cloud. Although these deep learning-based methods have been obtaining good results, they rely on full 3D meshes to generate multi-view rendering, which is difficult to enable the reliable 3D meshes for outdoor point clouds. These multi-view methods are processed based on images, not 3D point clouds. In addition, to the best of our knowledge, there does not exist any multi-view learning method that is directly applied to point cloud classification.
To fill this gap, we propose a features extraction and point clouds classification model based on multiple views and space representation consistency under constraints of label consistency (MvsRCLC). The overall flowchart of the proposed algorithm is shown in Figure 1. Firstly, the multi-view features of each point are extracted. Then, the features of each view are joined to learn the subspace of each view to remove redundant information, making the features representation more suitable for classification tasks. Secondly, the local distribution consistency constraint of feature space and position space is used to express the adjacency graph of each point in the local neighborhood. After that, the label consistency constraint is used to ensure the consistency between the predicted labels of all views and the ground truth label, as well as the consistency between the predicted labels within multiple views. Label consistency includes label consistency of grouped points (LCG) and label consistency of single point (LCS). Finally, an iterative optimization algorithm of objective function is proposed to progressively learn a subspace projection matrix and optimal linear classifier via solving a minimization problem. In the experimental section, two airborne laser-scanning (ALS) data scenes, two mobile laser-scanning (MLS) scenes and a terrestrial laser-scanning (TLS) scene with different complexities are used to evaluate the proposed algorithm.
We state the original contributions explicitly as below:
(1) A multi-view features and classifiers fusion framework for point cloud classification is proposed. By introducing the multiple constraints among different views and combining the classification error terms, the feature subspaces of multi-view can be jointly learned. This subspace learning method can be used to remove redundant information or noise effectively. Moreover, by simultaneously optimizing the feature projection matrices and linear classifiers of all views on the unified objective function, different view features can be fully explored, and more discriminant feature subspaces and optimal linear classifiers can be obtained.
(2) Unlike previous methods that simply concatenate multi-view features for classification, we propose a multi-view subspace learning method using diversity and consistency constraints between multi-view features, and then multi-view features and the classifier are coupled to classify point clouds, thereby improving labeling accuracy.
(3) Our algorithm takes multiple constraints including, local distribution consistency in feature space and position space, LCG constraints and predicted labels consistency constraints of multi-view features into account. It enhances subspace representations of point clouds and improves classification accuracy of point clouds. The proposed method has better classification performances particularly for using a small portion of training samples.
(4) The joint optimization method of multi-view objective function based on an iterative algorithm proposed in this paper can rapidly converge on the point cloud scenes used in this paper. The proposed algorithm is superior to other state-of-the-art methods by labeling ubiquitous points, ALS, MLS and TLS point clouds.

2. Materials and Methods

In this section, the multi-view point features, i.e., normal vector, covariance eigenvalue features and spin image features, used in our method are firstly presented. Then, the joint learning of the multi-view point feature extraction and classification model is constructed, which mainly includes: subspace learning, local distribution consistency constraint of multi-view feature and position spaces, and multi-view label constraints. Next, the proposed model is optimized before it is used to classify point clouds.

2.1. Multi-View Point Cloud Feature Extraction

The different types of features represent different attributes of point clouds. Thus, we extract different types of point cloud features to fully express point cloud attributes. In this paper, point cloud features of different views are constructed by using normal vector F n o r , covariance eigenvalue features F c o v [1] and spin image features F s p i n [20]. Here, we also generate a series of single point features extracted at multi-scale regions to enhance object recognition abilities. The features extraction processes are described as follows:
The neighborhood points of point p in radius r is regarded as the support region of point p. We construct the features defined at different scales by progressively changing radius r. Different radii r are selected to construct features of different scales for each view. The normal vector and covariance eigenvalue are calculated according to the methods in [3]. Since the normal vector and covariance eigenvalue features mainly represent the geometric features of point cloud, we take them as the same view features. Therefore, the multi-scale features of point clouds belonging to different views constructed by our method is X = [ X 1 , X 2 ] T , where, X 1 = [ F c o v , F n o r ]   and   X 2 = F s p i n . X 1 and X 2 represent the features of view 1 and the view 2, respectively. It is to be noted that normal vector, covariance eigenvalue features and spin image features are taken as an example for the proposed method. This means other kinds of features can be used to construct the multi-view features, i.e., the multi-view features X = [ X 1 , , X m ] T (m is the number of views, m > 2) include m kind of features. X v (v [ 1 , m ] ) is the vth view feature.

2.2. Multiple Views and Space Representation Consistency under Constraints of Label Consistency (MvsRCLC)

In this section, we discuss multiple views and spaces representations consistency with label consistency constraints for features extraction and point clouds classification (MvsRCLC).

2.2.1. Reconstruction Independent Component Analysis (RICA) Subspace Learning

In the multi-scale point cloud features representation, redundant and/or noisy information is usually present. To solve this problem, the feature projection technique is usually employed. It transforms the data in the high-dimensional space into a lower-dimensional space to achieve dimensionality reduction and subspace learning. To project the high-dimensional features into subspace by the feature transformation matrix with less reconstruction error, MvsRCLC minimizes the following objective function using RICA (reconstruction independent component analysis) algorithm [14] to extract the optimal feature transformation matrix.
Θ 1 ( W ) = W T W X X F 2 + α . g ( W X )
where W d × d is the feature transformation matrix that projects X d × n onto a d -dimensional feature space ( d < d ). In Equation (1), g ( · ) = log ( cosh ( · ) ) [ 13 ] , where   cos h ( W X ) = ( exp ( W X ) + exp ( W X ) ) / 2 , and α is a trade-off factor that is used to balance the reconstruction error term and sparse representation term.
For multi-view features, different views can be projected to the same subspace through different feature transformation matrices. Therefore, the subspace learning objective function of the multi-view features can be defined as:
Θ 1 v ( W v ) = v = 1 m W v T W v X v X v F 2 + α . g ( W v X v )
where, W v is the vth view feature transformation matrix, and X v is the vth feature. m is the number of views.

2.2.2. Multi-View Local Distribution Consistency Constraints

(1) Point Cloud Spatial Position Constraint Term
Although each point has different view features, the intrinsic spatial relationships are explicitly embedded in the point clouds. Therefore, the constructed spatial position constraint is applicable to subspace learning of all the views. Intuitively, for point cloud data, K neighbor points in the spatial position space tend to belong to the same category, and points of the same category also have similar data distribution in the feature space. Based on this observation, a coordinate distance constraint from each point to its K neighbor points is imposed, which is expressed by a spatial position weight matrix. Thus, the spatial position weight matrix of the point cloud is expressed as follows:
V i j = { exp ( P i P j 2 2 σ ) , if p i   ϰ K v (     p j   )   or   p j   ϰ K v (     p i   ) 0              otherwise
where, P i is the coordinate. ϰ K v ( · ) represents the set of the K nearest neighbor points in the spatial position space.
(2) Point Cloud Feature Space Constraint Term
Similarly, in the feature space, K neighbor points tend to express the objects of the same category. Thus, the feature distance constraint from each point to its K neighbor points in feature space can be constructed, which is expressed by a feature space weight matrix. The point cloud feature space weight matrix is constructed below:
U i j = { exp ( ( X ) i ( X ) j 2 2 ) ,     if    p i   ϰ K u (    p j   )   o r    p j   ϰ K u (    p i   ) 0              otherwise
where, ϰ K u (   p j   ) represents the nearest K neighbor of point pj in feature space.
Considering points of the same category also have similar data distribution in the spatial position space and feature space, spatial position space and feature space constraints in the same view can be jointly expressed as Equation (5). More details can be found in [17,18,19].
Θ 2 ( W ) = 1 2 i , j = 1 n ( V i j + β U i j ) ( W X ) i ( W X ) j 2 2 = t r ( ( W X ) ( L V + β L U ) ( W X ) T )
where, D U and D V are diagonal matrices, d i i U = j U i j , d i i V = j V i j , L U = D U U i j , L V = D V V i j . Operator   tr(.) is the trace of the matrix. Parameter β is the trade-off parameter.
As shown in Figure 2, the same color points belong to the same category in Figure 2a. Figure 2b shows the 5 neighbor points of p1 (the red point) in spatial position space. Figure 2c shows the different view relationships of the p1 and its neighbor points in feature space. The weight of each point in the blue circle (Figure 2b) and red circle (Figure 2c) depends on the distance to the p1. According to Equations (3) and (4), the weight of neighbor points in different spaces can be described by V i j and U i j . If the points are not neighbors, the weight values are assigned to 0. Although features of different views have a certain degree of difference, the points from the same category need to have similar adjacent relationships in the feature space of different views. In order to ensure that local points have a similar relationship graph in multi-view features, the relationship graph can be constrained by the features of different views. Figure 2d is the relationship of neighbor points for projection features, which is projected by the constraints of spatial position space and feature space. Generally, the neighbor points in spatial position space and feature space usually belong to the same category, thus, these constraints can guarantee the similarity of projection features for neighbor points. To minimize the feature discrepancy of points in subspace, the joint constraints of spatial position space and feature space are used to construct the multi-view objective function, which is shown as below:
Θ 2 v ( W v ) = v = 1 m t r ( ( W v X v ) ( L V   + β L U v ) ( W v X v ) T )  
where, L U v is the Laplacian matrix of vth in the feature space.

2.2.3. Label Consistency

(1) Label Consistency of Grouped Points (LCG)
It should be remembered that the corresponding labels need to be consistent before and after the feature transformation of the same category points. Based on this constraint, a label matrix Q of grouped points is built. Assuming that, x 1   and   x 2 belong to the first category, x 3 and   x 4 belong to the second category and x 5   and   x 6 belong to the third category, the matrix Q can be expressed as:
Q   =   [ 1 1 0 0 0 0 1 1 0 0 0 0 0 0 1 1 0 0 0 0 1 1 0 0 0 0 0 0 1 1 0 0 0 0 1 1 ]
After defining Q, the corresponding objective function of LCG can be expressed as follows:
Θ 3 ( G ) = Q G W X F 2 + γ G F 2
where G is the weight matrix, and the term γ G F 2 is the constraint used to prevent overfitting.
In addition, the ground truth labels of grouped points should be consistent with the predicted labels of grouped points at each view, and predicted labels of grouped points between different views are consistent. Therefore, we introduce the difference between the predicted labels of grouped points at different views as a constraint, which can be calculated by G v W v X v G w W w X w F 2 . Then, the corresponding multi-view objective function can be expressed as follows:
Θ 3 v ( G v ) = v = 1 m Q G v W v X v F 2 + γ G v F 2   + δ v w G v W v X v G w W w X w F 2
where, γ   and   δ are trade-off parameters.
(2) Label Consistency of Single Point (LCS)
While the classification results between the different views need to be as consistent as possible, each view classification result needs to be close to the ground truth label. After the point cloud feature transformation, the classification results obtained by the linear classifier using the projected features should be consistent with the ground truth. The LCS can be expressed as:
Θ 4 ( H ) = F H W X F 2 + γ H F 2
where H is the linear classification matrix, and the term γ H F 2 is the constraint used to prevent overfitting. F is the ground truth label matrix, which is same for all views. To make the predicted labels between the different views consistent, we introduce the discrepancy between the predicted labels of each point at different views as a constraint, which can be calculated by H v W v X v H w W w X w F 2 . Then, the objective function of the multi-view LCS can be expressed as follows:
Θ 4 v ( H v ) = v = 1 m F H v W v X v F 2 + γ H v F 2 + δ v w H v W v X v H w W w X w F 2
where, γ   and   δ are trade-off parameters.

2.2.4. Objective Function of MvsRCLC

To make the point cloud features with stronger expression ability, MvsRCLC learns the discriminative optimal feature expression through the subspace learning method. To leverage the diversity of subspace feature expressions from different views, joint constraint terms of position space and feature space are introduced. To make sure that the subspace feature expressions of the same category can be consistent in different views, and the subspace feature expression has the best discriminability, MvsRCLC introduces the LCG and LCS constraints. The labels of grouped points Q and the label of single point F are used to optimize the discriminability of the multi-view subspace expression for each category. Therefore, the objective function of MvsRCLC is as follows:
arg   min H v , W v , G v Θ ( H v , W v , G v ) = Θ 1 v + λ 1 Θ 2 v + λ 2 Θ 3 v + λ 3 Θ 4 v
where,   λ 1 ,   λ 2   and   λ 3 are trade-off parameters. In Equation (11), H v , W v and G v need to be optimized. After it has been solved, more discriminative features of point, i.e., Z v = W v X v , and linear classifier H v of each view can be obtained.

2.3. Optimization Technique

Since the objective function is highly non-linear, conventional techniques such as the gradient descent method or Newton method cannot be directly used in our situation. Alternatively, an iterative optimized strategy is utilized to solve Equation (11). The detailed pseudocode is elaborated in Algorithm 1. For convenience, we remove the subscript of sub-view in Equation (11).
As mentioned before, we use the MvsRCLC optimization algorithm listed in Algorithm 1 to solve the objective function. More precisely, in each iteration, we update only one variable and the rest of variables are fixed. By doing this, W, G and H can be optimized individually.

2.3.1. Update of W

Once we fix G and H, Equation (11) can be transformed into Equation (12) with a variable of W as input:
L ( W ) = W T W X X 2 2 + α g ( W X ) + λ 1 t r ( ( W X ) ( L V + β L U ) ( W X ) T ) + λ 2 ( Q G W X F 2 + δ v w G W X G w W w X w F 2 ) + λ 3 ( F H W X F 2 + δ v w H W X H w W w X w F 2 )
Note that Equation (12) can be considered as an unconstrained optimization problem. We obtain the derivative of Equation (12) with respect to W.
L ( W ) W = 2 ( W W T W X X T + W X X T W T W 2 W X X T ) + α g ( W X ) W + 2 λ 1 W X ( L v + β L U ) T X T +   2 λ 2 ( G T ( G W X Q ) X T + δ v w G T ( G W X G w W w X w ) X T ) + 2 λ 3 ( H T ( H W X F ) X T + δ v w H T ( H W X H w W w X w ) X T )
where, g ( W X ) W i j = k = 1 n tan h ( W i . X . k ) X j k ,   W i . represents the ith row of the matrix W. Given the original feature matrix X of the training data, the unconstrained optimization method L-BFGS [40] is used to update W.

2.3.2. Update of G

Similarly, we fix H and W, Equation (11) can be turned into Equation (15) with a variable of G as input:
L ( G ) = λ 2 ( Q G W X F 2 + γ G F 2 + δ v w G W X G w W w X w F 2 )
Equation (14) is also an unconstrained optimization problem. We obtain the derivative of Equation (14) with respect to G. Then by setting the derivative L ( G ) G = 0 , we obtain the solution of G:
= ( Q X T W T + δ v w G w W w X w X T W T ) / ( ( 1 + v w δ ) W X X T W T + γ I ) 1
Thus, the weight matrix of G can be updated by Equation (15).

2.3.3. Update of H

When fix G and W, Equation (11) can be converted into Equation (16) with a variable of H as input:
L ( H ) = λ 3 F H W X F 2 + λ 3 δ v w H W X H w W w X w F 2
Equation (16) is an unconstrained optimization problem. We obtain the derivative of Equation (16) with respect to H. By setting the derivative L ( H ) H = 0 , we obtain the solution of H:
H = ( F X T W T + δ v w H w W w X w X T W T ) ( ( 1 + v w δ ) W X X T W T + γ I ) 1
Thus, the weight matrix of H can be updated by Equation (17).
After optimization, the optimal weight matrices of W, G and H can be calculated. More details regarding the detailed optimization process are shown in Algorithm 1.

2.4. Point Cloud Labeling

After the objective function has been solved, the feature transformation matrix and the label projection matrix (linear classifier) of each view can be learned. For the testing set, the point cloud can be classified using learned projection matrices and linear classifiers. Since H v and     W v are the optimal solutions, feeding the features of new points X v , the classification result of the point cloud can be obtained:
  y i = arg min j ( v = 1 m ( ϱ v H v W v X v ) 1 × j )   ( 1 j c )
where c is the number of categories.   y i is the predicted classification label of multi-view, and ϱ v is the weight of each view for the point cloud classification, v = 1 m ϱ v = 1 and ϱ v is determined by the ratio of X v in each view.
Algorithm 1: MvsRCLC optimization algorithm (The pseudocode of the multiple views and space representation consistency under constraints of label consistency (MvsRCLC) optimization algorithm.)
Input: multi-view feature matrix: X = { X 1 , , X m }, ground truth label matrix of single point: F, ground truth label matrix of grouped points: G
Parameters: α, β, γ, δ , λ 1 , λ 2 , λ 3 , convergence error: ε 1 and the maximum number of iterations: T
Initialization: W v , H v   and   L U v , v = 1 , , m ,   t   =   0 , iter = 0
Calculating Laplacian matrix L V of spatial position
while not converged do
for each view v { 1 , , m }
    while iterT do
  Update W v : Fixed H v   and   G v , W v can be solved by the unconstrained optimization operator L-BFGS according to Equation (13).
  Update G v : Fixed W v   and   H v , update according to Equation (15)
  Update H v : Fixed W v   and   G v , update according to Equation (17)
      Update: iter = iter + 1
     end
   end
   Convergence condition: | Θ t Θ t + 1 | | Θ t | ε 1
   If not converge, update: t = t + 1
end
Output: projection matrix: { W 1 , , W m }, linear classifier: { H 1 , , H m }.

3. Performance Evaluation

In this section, we first briefly describe the experimental data and evaluation metrics. Afterwards, we compare the classification results of the proposed algorithm with other related algorithms in two different experiments. Finally, we analyze the parameters and convergence of the proposed method.

3.1. Experiment Data and Evaluation Metrics

Five different point cloud scenes (see Figure 3) are used to evaluate the performance of the proposed algorithm. The five scenes are mainly divided into three types according to the different platforms where the lidar sensor are mounted. The first type is ALS point cloud data, including Scene1 scanned in residential area and Scene2 scanned in urban area. These two scenes are collected by a Leica ALS50 system with a mean flying height of 500 m above ground and a 45° field of view in Tianjin, China [41]. The point density is approximately 20–30 points/m2. The ground points of these two scenes have been manually filtered out. As shown in Figure 3a,b, the data of these two scenes only contain non-ground points, i.e., buildings, trees and cars. Note that the training samples and testing samples are defined in [41]. The second type of data used in this paper is MLS point clouds, including Scene3 and Scene4. The point clouds in Scene3 and Scene4 are acquired by the backpacked mobile acquisition device [42], which is also a mobile laser scanner mounted on a person, in Shenyang and Beijing, China. As shown in Figure 3c,d, the data (filtered out ground points manually) of these two scenes are manually labeled four and nine categories by ourselves, respectively. Moreover, Scene4 is complex data published in [4], which have a large fluctuation of the point number among the nine classes, i.e., buildings, trees, cars, pedestrians, wire poles, street lamps, traffic signs, wires and pylons. The third type is TLS point cloud, i.e., Scene5, which is acquired by the terrestrial laser scanner (RIEGL MS-Z620) in urban area [5]. Affected by the distance of objects to the scanner, point density of the Scene5 varies a lot, and many objects are incomplete and noisy. As shown in Figure 3e, similar to ALS point clouds, the ground points of Scene5 have been filtered out manually. Four categories, i.e., Cars, Trees, Pedestrians, and Buildings, are used to evaluate the classification performance. The specific information of the five scenes is shown in Table 1. The data of Scene1, Scene2 and Scene5 can be downloaded at the author’s website (http://geogother.bnu.edu.cn/teacherweb/zhangliqiang/). The point cloud of Scene3 (https://pan.baidu.com/s/1WA_YwOACBcy5jArUAmd6xA) and Scene4 (https://pan.baidu.com/s/1lOCe39sfPvkpPDTY1-TOrw) are also public datasets and are extensively used in the previously published works [4,43].
In our experiment, the proposed algorithm is implemented in MATLAB 2017b. All the experiments were run on a personal computer, equipped with a 4.20 GHz Intel Core i7-7700k CPU, 24 GB of main memory. In order to be more comprehensive and effective, we use four popular evaluation indicators to evaluate the classification performance of each category, including precision/recall, intersection over union (IoU), and F1-score. Overall classification results of each scene are evaluated by four other popular metrics, namely overall accuracy (OA), mean intersection over union (mIoU), Kappa and mF1. The detailed definitions of these metrics are presented in [3,4].

3.2. Experimental Results

In order to prove the effect of the proposed algorithm from different perspectives, two groups of experiments are conducted using different types of point clouds. The first experiment group is mainly used to verify the classification results of the proposed method by selecting a small number of training samples from point cloud scenes with a large volume of point sets. The second experiment group is mainly conducted to compare the proposed method with the popular multi-view joint classification methods. It proves the classification performance of the proposed algorithm using a small number of training samples from the point cloud scenes with a small-volume of point sets. To avoid ambiguity, the term large and small volume point set defined in our context refer to point clouds greater than and less than 50 thousand points.

3.2.1. The First Experimental Group

Experiments were carried out on five different point cloud scenes shown in Table 1 to test the classification performance of the proposed algorithm on point clouds with a small percentage of training samples. A small number of points are selected for model training from each scene. The remaining points are regarded as testing points that need to be classified.
(1) Comparison methods
To highlight the performance of the proposed algorithm, we compare nine algorithms, which can be divided into two categories, i.e., single view feature based method and multiple features fusion based method. To prove the effectiveness of the multi-view features fusion, we compare our method with FC(our) and FSI(our). To show the advantages of the multi-view features joint learning, we compare the proposed method, FC(our) and FSI(our), with FC(SVM) and FSI(SVM). These four comparison methods are based on a single view feature. Adaboost is a well-known classifier. RICA-SVM is a well-known classifier that employs features learning. D-KSVD, LC-KSVD1 and LC-KSVD2 are three typical joint features learning and classification methods. These five methods are simply based on multiple features fusion. To show the performance of different classifiers, we compare our method with Adaboost and RICA-SVM. We compare our method with RICA-SVM, D-KSVD, LC-KSVD1 and LC-KSVD2 to highlight the performance of different features learning methods. In addition, we also compare our method with D-KSVD, LC-KSVD1 and LC-KSVD2 to prove the effectiveness of the multi-view features learning and fusion. The comparison methods are described as follows:
• Classification methods of single-view feature:
1) FC(our): Based on the view of covariance eigenvalue features, the trained single view classification model by our method is used for classification, that is,   Y = H v W v X v , where, v is the view of covariance eigenvalue features.
2) FSI(our): Based on the view of spin image features, the trained single view classification model of our method is used for classification, that is,   Y = H w W w X w , where, w is the view of spin image features.
3) FC(SVM): A classification method based on covariance eigenvalue features and SVM classifier.
4) FSI(SVM): A classification method based on spin image feature and SVM classifier.
• Classification methods with multi-view features fusion:
5) Adaboost: This configuration is based on the direct fusion of covariance eigenvalue and spin image features. Adaboost classifier is used for classification. In our case, we set the number of trees to 100.
6) RICA-SVM: Covariance eigenvalue feature and spin image feature are firstly fused directly, and then the feature projection is conducted with RICA. The point cloud classification is done by support vector machine (SVM) classifier.
7) D-KSVD [10]: Based on the direct fusion of covariance eigenvalue features and spin image features, D-KSVD is used to classify point clouds. The value range of dictionary words is set to {128,256,512}, the value range of α is set to {0.0001, 0.001, …, 10}, and the remaining parameters are adopted default values used in [10]. Setting 128 words and α = 0.1 can obtain the optimal classification results in this configuration.
8) LC-KSVD1 [11]: Based on the direct fusion of covariance eigenvalue features and spin image features, LC-KSVD1 is used for classification. The objective function of the model is shown in Equation (19). The value range of dictionary words is {128, 256, 512}, and the value range of α is {0.001, 0.01, 0.1, 1, 10}.
arg   min D , A , S Θ L C 1 ( D , A , S ) = arg min D , A , S X D S 2 2 + α Q A S 2 2 ,    s . t .    i , s i 0 T
where, T is the sparse factor, D is the dictionary, A is the linear transformation matrix, S is the sparse code, and Q is the significant sparse code.
9) LC-KSVD2 [11]: Based on the direct fusion of covariance eigenvalue features and spin image features, LC-KSVD2 is used for classification. The classification model of this method is shown in Equation (20). According to [11], the model can obtain the optimal classification results when α = β is taken in multiple groups of experiments. In our experiment, α and β are tuned from the set {0.001, 0.01, 0.1, 1,10}, respectively. The remaining parameters are the default parameters of [11]. We adopt 128 words, α = β = 0.001 in the LC-KSVD1 and LC-KSVD2 experiments to obtain the optimal results.
arg   min D , A , S , H Θ L C 2 ( D , A , S , H ) = arg min D , A , S , H X D S 2 2 + α Q A S 2 2 + β Y H S 2 2 ,    s . t .   i , s i 0 T
where, Y is the predicted label matrix, and H is the linear classifier matrix.
(2) Parameters setting
Before analyzing the quality of the labeling methodology, we first list all the relevant parameters and give their recommended values. Generally, the value of K nearest neighbor is suggested to be selected from the set {3, 4, 5, 6, 7, 8, 9, 10}. Parameter σ is selected from the set {10−6, 10−5, …, 105, 106}. Weight balancing parameters α , β ,   γ   and   δ are suggested to be set from {10−4, 10−3, …, 103, 104} and balancing parameters λ 1 , λ 2 and λ 3 are tuned from {10−5, 10−4, …, 104, 105}.
(3) The results
For ALS point clouds, we use Scene1 and Scene2 shown in Table 1 to carry out experiments on the same hardware specs. This experiment group verifies the algorithms’ labeling performances under the premise of using a small number of samples for training and a larger number of points for testing. The number of the selected training points is less than 5% of the testing points. More specifically, 10,000 and 9602 points are selected from the training samples of Scene1 and Scene2 for training, all of the testing point clouds are used for testing samples. To make an unbiased comparison with other methods, the same training and testing samples are used for testing other methods. The classification result statistics of various types of evaluation measures are shown in Table 2 and Table 3. To verify the efficiency of our method for point cloud classification, we compare the classification running time of different methods with multiple features on the test set of Scene1 (422,355 points) and Scene2 (236,802 points). The running time comparison is shown in Table 4.
From the results listed in Table 2, Table 3 and Table 4, it is easily to draw the conclusions below:
1) The classification performance of our method outperforms other comparison methods on ALS point clouds. FC(our) and FSI(our) can also achieve good classification performance. Moreover, not all multiple feature-based methods are superior to single feature-based methods. In this experiment group, for multiple feature-based methods, the OA and Kappa of our method are at least 4.3% higher than other comparison methods, and the mIoU and mF1 of our method are at least 1.2% higher than other comparison methods. Compared with the single feature-based method, the classification performance of our method outperforms FC(our), FSI(our), FC(SVM) and FSI(SVM), which indicates that the proposed method can effectively fuse multi-view features and it is effective to fuse multi-view features for the improvement of point cloud classification.
2) The classification performances of LC-KSVD1 and LC-KSVD2 outperform DKSVD, which proves the effectiveness of the constraints of LCG and LCS. Adaboost cannot achieve a better classification performance by directly concatenating different view features. However, we also adopt the strategy of concatenating different view’ features, RICA-SVM has achieved relatively good classification results because it learns the transformation matrix of fused features, thereby making the projected features more distinguishable. The proposed method has achieved the highest value in terms of OA, mIoU, kappa, and mF1 for Scene1 and the best accuracy in terms of OA, mIoU and kappa for Scene2. These qualitative values demonstrate that the effectiveness of joint learning for feature projection matrices and multi-view classifiers under the constraints of labels.
3) As shown in Table 4, for the point cloud classification of Scene1 and Scene2, our method requires less than 4.5%, 2.7%, 1.3%, 0.9% and 44.2% running time of the Adaboost, LC-KSVD1, LC-KSVD2, DKSVD and RICA-SVM, respectively. Thus, the speed of our method outperforms the compared methods. Although the running time of RICA-SVM is relatively close to our method, the OA and Kappa of our method are at least 5.4% higher than RICA-SVM according to the classification accuracy performance of Scene1 and Scene2 in Table 2 and Table 3. The above comparisons demonstrate that our method is superior to Adaboost, LC-KSVD1, LC-KSVD2, DKSVD and RICA-SVM in terms of accuracy and efficiency.
To show the advantages of the proposed method vividly, we also compared the performance of our method with RICA-SVM, which is the second-ranked classification approach based on multi-view fusion features. Figure 4b,c give the comparisons of labeling results of Scene2. As shown in Figure 4, the cars classification of our method and RICA-SVM are bad. This is mainly due to the fewer training points of cars and the great similarity of extracted features between cars and buildings. From the enlarged black boxes, we can find that a large number of trees are misclassified as buildings using RICA-SVM. This is probably because the discriminability of trees features is weakened by the direct fusion of multiple features. However, the classification performance of our method is more accurate due to the effective fusion of multiple features.
We used Scene3–Scene5 depicted in Table 1 to verify the labeling performances by using MLS and TLS point clouds. More precisely, 3200, 7200 and 3200 points are randomly chosen as training points from Scene3, Scene4 and Scene5. This only counts for just 0.7%, 0.8% and 0.8% of the corresponding scene’s points. The rest of point clouds of each scenes are all regarded as testing data. To make an unbiased comparison with other methods, the same training and testing samples were used in this group. The classification results of different methods on Scene3–Scene5 are shown in Table 5, Table 6 and Table 7.
As shown in Table 5, Table 6 and Table 7, our method can achieve the similar classification performance on MLS point cloud and TLS point cloud, and our method has certain advantages over other compared algorithms. In all three scenes, the OA and Kappa of our method are at least 1% and 0.47% higher than the other compared algorithms, respectively.
FC(SVM), FSI(SVM), Adaboost, DKSVD, LC-KSVD1 and LC-KSVD2 can classify most points (OA ≥51%), which proves that the features of each view are relatively effective. RICA-SVM can obtain one or two of the highest values of all the metrics in each scene, which indicates that simple fusion of features to learn projection features through RICA is helpful to the expression of discriminant features. In addition, our method is jointly trained with multi-view features, which makes FC(our) and FSI(our) obtain better classification results. In the two MLS and one TLS point cloud scenes, the classification performances of FC(our) and FSI(our) outperform FC(SVM) and FSI(SVM). This proves the classification model of each view jointly learned with multi-view features is applicable.
To vividly show the superiority of the proposed method, we compared the performance of our method with LCKSVD1, LCKSVD2, DKSVD and RICA-SVM. Figure 5 gives the comparisons of labeling results of Scene4. From the enlarged black boxes, we can find that the proposed method has obvious advantages in the classification of buildings and trees. In terms of the overall classification performance, our method is closer to the ground truth due to the more discriminant features extraction by multi-view features joint learning and the effective fusion of the multiple features.

3.2.2. The Second Experimental Group

(1) Comparison methods
To evaluate the performance of our method, we compared results of nine related algorithms. To highlight the performance of the proposed multi-view features joint learning and fusion method, we compared it with three representative multi-view learning methods, i.e., adaptively weighted procrustes (AWP) [26], automatic multi-view graph and weights learning (AMGL) [23] and multi-view learning with adaptive neighbors (MLAN) [25]. AWP is an unsupervised method, which is compared with AMGL, MLAN and our method to show the performance of multi-view unsupervised and supervised classification methods. To prove the effectiveness of multi-view learning, we compared our method with non-negative sparse graph (NNSG) [19] and SVM [44]. Besides, we compare our method with FC(our), FSI(our), FC(SVM) and FSI(SVM) to show the effectiveness of multi-view features fusion. The compared algorithms and their recommended parameters are described below:
1) AWP 2018 [26]: a multi-view unsupervised classification algorithm based on adaptively weighted procrustes (AWP).
2) AMGL 2016 [23]: a new framework for automatic multi-view graph and weights learning. This method uses a small number of samples to train the classifier and further to predict unlabeled samples. In this paper, the neighborhood parameter of AMGL is set to 5, and the remaining parameters are assigned with default values of the source code that was released to the public in [23].
3) MLAN 2018 [25]: an effective multi-view model for adaptive learning of local data structure. The adaptive neighborhood value of MLAN is set to 9, and the remaining parameters are assigned with default parameters of the source codes in [25].
4) NNSG 2015 [19]: a non-negative sparse graph method for learning linear regression. We set λ 1 = 0.001 , λ 2 = 0.5 and λ 3 = 0.01 .
5) SVM [44]: multi-view features are directly concatenated together, which are then provided as input to SVM classifier for training and classification.
Apart from the comparisons with the above five methods, we also compared our results with FC(our), FSI(our), FC(SVM) and FSI(SVM) (see Section 3.2.1 for more detailed descriptions regarding these four configurations.)
(2) The results
For testing ALS point clouds, we randomly selected 9062 points from Scene2, from which 300 points (3.31%) were randomly selected for training. The remaining points were used as testing samples. Figure 6 shows a qualitative comparison result of the four multi-view methods. The quantitative comparison on all evaluation metrics are shown in Table 8.
From the results listed in Table 8 and illustrated in Figure 6, we can make the following observations:
1) In Figure 6, the results of the proposed method and AMGL are closer to ground truth, the rest of qualitative results have relatively large errors. From Table 8, overall classification measures of 79.7%, 56.6%, 64.3% and 69.4% with regard to OA, mIoU, Kappa and mF1 maintain the highest values, which demonstrates the advantages/superiority of the proposed method.
2) AWP has a high OA, but it cannot classify cars because the sample number of cars is relatively small in the experimental data. In addition, it is rare to see the categorized car points, as demonstrated in Figure 6c. In contrast, for AMGL algorithm, numerous tree points are mistakenly classified as cars (see Figure 6d). Although the classification precision of trees and the recall of buildings reached 100% using MLAN algorithm, numerous building points are misclassified as trees. In addition, the car points cannot be recognized, as demonstrated in Figure 6e. NNSG has capability to classify most cars, but a large number of building points and trees are falsely classified as cars, as evident in Figure 6f. The overall labeling results of NNSG is relatively poor. Although the SVM algorithm has the best classification performance of trees, the overall labeling results are inferior to ours.
3) For the covariance eigenvalues feature, the four overall evaluation metrics of the classification model obtained by our joint training method (FC(our)) are at least 4% higher than the FC(SVM). For the spin image feature, the classification results of our model trained by joint learning (FSI(our)) also have certain advantages over FSI(SVM). Therefore, we can make safe conclusions that the classification model jointly trained by our method is more effective than SVM-based model using single-view features. In addition, the classification performance of our multi-view fusion method outperforms the single-view feature classification.
For testing MLS point clouds, we randomly selected 12,000 points from Scene3 as experimental data, from which 1200 points (10%) were randomly selected for training. The remaining points are as testing samples. Figure 7 shows a qualitative comparison result of the multi-view methods. The quantitative comparison on all evaluation metrics are shown in Table 9.
From the results listed in Table 9 and illustrated in Figure 7, we can make the following observations:
1) In the qualitative and quantitative comparison, our method has better performance in the classification of point cloud in Scene3 than other methods. For the four classification evaluation metrics, our method has achieved the highest classification accuracy.
2) AWP, AMGL and NNSG have poor classification performances on each class of points, and there is a large number of misclassification points as demonstrated in Figure 7c–f. MLAN has the best overall classification result for the poles. Although this method has a highest recall value for tree classification, there are more misclassifications for the other class (see Figure 7e). SVM has the capability to distinguish trees and cars, achieving 36.0%/52.9% and 67.6%/80.6% with regard to IoU/F1-score measures. Although the MLAN and SVM algorithms show the best labeling result for a certain class, we still have absolute advantages in terms of overall measure evaluations as we obtained 70.0%, 49.7%, 56.0% and 65.7% with regard to overall measures of OA, mIoU, Kappa and mF1.
3) The classification results obtained by our multi-view fusion method are significantly improved compared with the single-view feature classification methods. In addition, the model of multi-view feature joint training and classification outperforms the multi-view feature direct fusion classification.

3.3. Effectiveness on ISPRS 3D Semantic Labeling Dataset

To prove the effectiveness of the proposed method for the International Society for Photogrammetry and Remote Sensing (ISPRS) 3D Semantic Labeling Dataset [45], we conducted an experiment on this dataset. Here, we add three kinds of features, i.e., fast-point feature histograms (FPFH), normal angle distribution histogram (NAD) and latitude sampling histogram (LSH) [3] as multi-view features. The experimental results are shown in Figure 8 and Table 10, which can demonstrate that the proposed method achieves a promising classification performance for the cars, trees, buildings and grounds.

3.4. Effectiveness of Multiple Constraints

In order to verify the contribution of each independent constraint (IC) in the objective function of our method, we compared the different ICs for point cloud classification. We use three configurations to prove the effectiveness of individual constraints:
(1) IC1: IC1 is the subspace learning term, which removes the redundant information by joint learning the multi-view feature transformation matrices and linear classifiers. Then the point cloud classification model can be obtained. The objective function is as follows:
Θ I C 1 ( H v , W v ) = arg min H v , W v v = 1 m W v T W v X v X v F 2 + α g ( W v X v ) + λ 3 ( F H v W v X v F 2 + γ H v F 2 )
(2) IC2: Based on IC1, we introduce spatial position and feature space constraints into the objective function, as shown in Equation (22). The point cloud classification model is obtained by learning H v and W v .
Θ I C 2 ( H v , W v ) = arg min H v , W v v = 1 m W v T W v X v X v F 2 + α g ( W v X v ) + λ 1 t r ( ( W v X v ) (   L U v + β L V ) ( W v X v ) T )   + λ 3 ( F H v W v X v F 2 + γ H v F 2 )
(3) IC3: In order to verify the contributions of multi-view joint training and multi-view classifier joint classification, we construct the classification model as shown in Equation (23), which directly concatenate the derived features of multi-views.
Θ IC 3 ( H , W ) = arg min H , W W T W X X F 2 + α g ( W X ) + λ 1 t r ( ( W X ) ( L U + β L V ) ( W X ) T ) + λ 3 ( F H W X F 2 + γ H F 2 )
Here, to show the effectiveness of spatial position and feature space constraints, we compared IC1 with IC2. To show the advantages of multi-view learning, we compared our method and IC2 with IC3. Besides, we also compare our method with IC2 to prove the effectiveness of LCG. We arbitrarily select Scene1 (ALS point clouds with 3 classes) and Scene4 (MLS point clouds with 9 classes) to verify the effectiveness of individual constraints. In the experiment, 10,000 and 7200 points are selected from Scene1 and Scene4, respectively, which amounts to 0.8% points of testing sets. For IC1, IC2, IC3 and the proposed model, we keep the sum dimension number of the transformation features consistent (the sum dimension number of the new feature is 86), and the maximum iteration number is set to 20. The comparisons of the above three modified configurations with our classification results of each classification model are shown in Table 11.
From Table 11, we have the following observations.
(1) IC2 outperforms IC1. The main reason is that the consistency constraint of position and feature spaces is introduced in different views, so the learned subspace can better reflect the local geometry of the data and the consistency of the local space distribution among different views.
(2) Our method achieves 77.49%/80.93% and 57.15%/68.62% with regarding to OA and Kappa measures on Scene1/Scene4, which are 4.54%/1.64% and 6.66%/1.25% higher than IC1, and 3.03%/0.25% and 3.73%/0.32% higher than IC2. The experimental results prove that the constraints of LCG and label consistency among different views are effective. The joint learning of the feature transformation matrix and classifier can effectively improve the accuracy of point cloud classification. The main reason is that LCG makes the learned subspace more discriminative, and the constraints among different views make the trend of the feature expression and classification results of different views more consistent.
(3) The overall performance of our method outperforms IC3. The experimental results show that the multi-view joint learning classification method for features projecting and fusion is more effective than the direct fusion classification methods. The main reason is that different features reflect the different attributes of the same point. The features of multi-view can make full use of the attributes of each point through joint constraint, which can achieve better classification results than the method of multiple features direct fusion.

3.5. Parameters Analysis

In the proposed method, there are mainly 7 key parameters that need to be tuned, including α , β ,   γ , δ , λ 1 , λ 2 and λ 3 . In the experiment, we set α = 0.1, β =0.01, and δ =0.001.
Parameters of λ 1 , λ 2 and λ 3 correspond to the terms of position space and feature space constraint, multi-view LCG constraint and multi-view LCS constraint, respectively. We jointly discuss λ 2 and λ 3 , which affect the weight of the LCG term and LCS term. Figure 9a,b show the classification accuracy of different λ 2 and λ 3 on Scene1 and Scene4 when λ 1 is fixed. It can be seen that the difference of accuracy with different λ 2 and λ 3 is small, which is within 0.8% and 0.05% for Scene1 and Scene4, respectively. This demonstrates that λ 2 and λ 3 have a small influence on the classification result when they vary within the given values. Moreover, the accuracy fluctuations are smooth, i.e., the classification accuracy is relatively less affected by the varying λ 2 and λ 3 when λ 1 is fixed. In addition, as shown in Figure 9c,d, when λ 3 is fixed, and λ 1 and λ 2 varies within given ranges, the classification accuracy of our method on different parameter values has larger changes, which is within 5.5% and 0.25% for Scene1 and Scene4, respectively. However, the accuracy fluctuation caused by the change of λ 1 and λ 2 (fixed λ 1 ) is relatively stable, except for some values of parameters. As shown in Figure 9e,f, when λ 2 is fixed, and λ 1 and λ 3 varies within given values, and the classification accuracy difference of our method on different parameter values is within 2.5% and 0.2% for Scene1 and Scene4, respectively. Besides, we can also find that the accuracy fluctuations are relatively small, i.e., the proposed method is relatively stable, which has less influence by the varying λ 1 and λ 3 when λ 2 is fixed. From Figure 9, we can observe that the accuracy fluctuation trends of the two scenes are similar for each situation (two parameter values vary when a parameter value fixes). Besides, the comparisons of these three situations demonstrate that λ 1 has the largest influence on classification accuracy, and λ 2 and λ 3 have relatively small influence on classification accuracy. From the above analysis we can draw a safe conclusion that the parameters of our method have a low impact on the point cloud classification in a certain range. That is, our method is relatively insensitive to parameters λ 1 , λ 2 and λ 3 .

3.6. Convergence Analysis

The objective function shown in Equation (11) is highly non-linear, although it is solvable, it is very difficult to simultaneously optimize the variables W v , G v and H v . Figure 10 reports the objective function values within the 25 iterations. The objective function is optimized by the proposed update rules depicted in Algorithm 1. Figure 10a,b are the objective function value curves corresponding to Scene1 and Scene4. We can see that our optimization method can quickly converge to a local or even global minimum for our two scenes.

4. Conclusions

In this paper, we have proposed a multi-view joint learning framework for point cloud classification, which includes multi-view subspace learning term for removing redundant information and representing low-dimensional features; the local distribution consistency constraint of feature space and position space term for the adjacency expression of neighborhood points; and the label consistency terms with the consistency between the predicted labels of all views and the ground truth label and the consistency among the predicted labels of each view. These terms are joined to learn the transformation matrices and optimal classifiers by the iterative optimization algorithm of objective function, and the convergence is very fast. Experiments performed on two ALS point cloud scenes, two MLS point cloud scenes and a TLS point cloud scene with fewer training points clearly confirm that our method outperforms the compared algorithms. Although our method achieves more promising classification accuracy than the compared algorithms, there are some drawbacks and interesting ideas that can be further explored to extend the research reported in this paper. Currently, the proposed method cannot achieve accurate semantic labeling of city-level point clouds having complex geometric shapes and diverse objects. In addition, the derived multiple features are relatively simple. Point set features and high-level features are not fully explored. In future work, the multi-view joint learning method and deep-learning technologies can be combined to learn a multi-view deep-learning network for improving the semantic labeling of point clouds.

Author Contributions

Y.L. and D.C. analyzed the data and wrote the Matlab source code. G.T., and S.X. helped with the project and study design, paper writing, and analysis of the results. J.P. and Y.W. helped with the data analysis, experimental analysis, and comparisons. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China under Grant 41971415.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhang, Z.; Zhang, L.; Tong, X. Discriminative-dictionary-learning-based multilevel point-cluster features for ALS point-cloud classification. IEEE Trans. Geosci. Remote Sens. 2016, 54, 7309–7322. [Google Scholar] [CrossRef]
  2. Farabet, C.; Couprie, C.; Najman, L. Learning hierarchical features for scene labeling. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 35, 1915–1929. [Google Scholar] [CrossRef] [Green Version]
  3. Li, Y.; Tong, G.; Du, X. A single point-based multilevel features fusion and pyramid neighborhood optimization method for ALS point cloud classification. Appl. Sci. 2019, 9, 951. [Google Scholar] [CrossRef] [Green Version]
  4. Li, Y.; Tong, G.; Li, X. MVF-CNN: Fusion of multilevel features for large-scale point cloud classification. IEEE Access. 2019, 7, 46522–46537. [Google Scholar] [CrossRef]
  5. Wang, Z.; Zhang, L.; Fang, T. A multiscale and hierarchical feature extraction method for terrestrial laser scanning point cloud classification. IEEE Trans. Geosci. Remote Sens. 2014, 53, 2409–2425. [Google Scholar] [CrossRef]
  6. Guo, B.; Huang, X.; Zhang, F. Classification of airborne laser scanning data using JointBoost. ISPRS J. Photogramm. Remote Sens. 2015, 100, 71–83. [Google Scholar] [CrossRef]
  7. Weinmann, M.; Jutzi, B.; Hinz, S. Semantic point cloud interpretation based on optimal neighborhoods, relevant features and efficient classifiers. ISPRS J. Photogramm. Remote Sens. 2015, 105, 286–304. [Google Scholar] [CrossRef]
  8. Chen, D.; Peethambaran, J.; Zhang, Z. A supervoxel-based vegetation classification via decomposition and modelling of full-waveform airborne laser scanning data. Int. J. Remote Sens. 2018, 39, 2937–2968. [Google Scholar] [CrossRef]
  9. Li, Y.; Chen, D.; Du, X. Higher-order conditional random fields-based 3D semantic labeling of airborne laser-scanning point clouds. Remote Sens. 2019, 11, 1248. [Google Scholar] [CrossRef] [Green Version]
  10. Zhang, Q.; Li, B. Discriminative K-SVD for dictionary learning in face recognition. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Francisco, CA, USA, 13–18 June 2010. [Google Scholar]
  11. Aharon, M.; Elad, M.; Bruckstein, A. K-SVD: An algorithm for designing overcomplete dictionaries for sparse representation. IEEE Trans. Signal Process. 2006, 54, 4311–4322. [Google Scholar] [CrossRef]
  12. Jiang, Z.; Lin, Z.; Davis, L. Label consistent K-SVD: Learning a discriminative dictionary for recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 2651–2664. [Google Scholar] [CrossRef] [PubMed]
  13. Mei, J.; Wang, Y.; Zhang, L. PSASL: Pixel-level and superpixel-level aware subspace learning for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 4278–4293. [Google Scholar] [CrossRef]
  14. Le, Q.; Karpenko, A.; Ngiam, J. ICA with reconstruction cost for efficient overcomplete feature learning. In Proceedings of the Advances in Neural Information Processing Systems, Granada, Spain, 12–14 December 2011. [Google Scholar]
  15. Martínez, A.; Kak, A. PCA versus LDA. IEEE Trans. Pattern Anal. Mach. Intell. 2001, 23, 228–233. [Google Scholar] [CrossRef] [Green Version]
  16. Nie, F.; Yuan, J.; Huang, H. Optimal mean robust principal component analysis. In Proceedings of the International Conference on Machine Learning, Beijing, China, 21–26 June 2014. [Google Scholar]
  17. Mei, J.; Zhang, L.; Wang, Y. Joint margin, cograph, and label constraints for semisupervised scene parsing from point clouds. IEEE Trans. Geosci. Remote Sens. 2018, 56, 3800–3813. [Google Scholar] [CrossRef]
  18. Zhu, P.; Zhang, L.; Wang, Y. Projection learning with local and global consistency constraints for scene classification. ISPRS J. Photogramm. Remote Sens. 2018, 144, 202–216. [Google Scholar] [CrossRef]
  19. Fang, X.; Xu, Y.; Li, X. Learning a nonnegative sparse graph for linear regression. IEEE Trans. Image Process. 2015, 24, 2760–2771. [Google Scholar] [CrossRef]
  20. Johnson, A. A Representation for 3D Surface Matching. Ph.D. Thesis, Robotics Institute, Carnegie Mellon University, Pittsburgh, PA, USA, 1997. [Google Scholar]
  21. Rusu, R.; Blodow, N.; Beetz, M. Fast point feature histograms (FPFH) for 3D registration. In Proceedings of the IEEE International Conference on Robotics and Automation, Kobe, Japan, 12–17 May 2009. [Google Scholar]
  22. Rusu, R.; Bradski, G.; Thibaux, R. Fast 3D recognition and pose using the Viewpoint Feature Histogram. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, Taipei, Taiwan, 18–22 October 2010. [Google Scholar]
  23. Nie, F.; Li, J.; Li, X. Parameter-free auto-weighted multiple graph learning: A framework for multiview clustering and semi-supervised classification. In Proceedings of the International Joint Conference on Artificial Intelligence, New York, NY, USA, 9–15 July 2016. [Google Scholar]
  24. Nie, F.; Cai, G.; Li, X. Multi-view clustering and semi-supervised classification with adaptive neighbours. In Proceedings of the AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017. [Google Scholar]
  25. Nie, F.; Cai, G.; Li, J. Auto-weighted multi-view learning for image clustering and semi-supervised classification. IEEE Trans. Image Process. 2018, 27, 1501–1511. [Google Scholar] [CrossRef]
  26. Nie, F.; Tian, L.; Li, X. Multiview clustering via adaptively weighted procrustes. In Proceedings of the ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, London, UK, 19–23 August 2018. [Google Scholar]
  27. Zhang, C.; Hu, Q.; Fu, H. Latent multi-view subspace clustering. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  28. Wang, X.; Guo, X.; Lei, Z. Exclusivity-consistency regularized multi-view subspace clustering. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  29. Cao, X.; Zhang, C.; Fu, H. Diversity-induced multi-view subspace clustering. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015. [Google Scholar]
  30. Wang, Y.; Wu, L.; Lin, X. Multiview spectral clustering via structured low-rank matrix factorization. IEEE Trans. Neural Netw. Learn. Syst. 2018, 29, 4833–4843. [Google Scholar] [CrossRef] [Green Version]
  31. Tang, C.; Zhu, X.; Liu, X. Learning joint affinity graph for multi-view subspace clustering. IEEE Trans. Multimedia 2019, 21, 1724–1736. [Google Scholar] [CrossRef]
  32. Griffiths, D.; Boehm, J. A Review on deep learning techniques for 3D sensed data classification. Remote Sens. 2019, 11, 1499. [Google Scholar] [CrossRef] [Green Version]
  33. Su, H.; Maji, S.; Kalogerakis, E.; Learned-Miller, E. Multi-view convolutional neural networks for 3D shape recognition. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015. [Google Scholar]
  34. Boulch, A.; Guerry, J.; Le Saux, B.; Audebert, N. SnapNet: 3D point cloud semantic labeling with 2D deep segmentation networks. Comput. Graph. 2018, 71, 189–198. [Google Scholar] [CrossRef]
  35. Felix, J.; Martin, D.; Patrik, T.; Goutam, B.; Fahad, S.; Michael, F. Deep projective 3D semantic segmentation. In International Conference on Computer Analysis of Images and Patterns; Lecture Notes in Computer Science Series Volume 10424; Springer: Cham, Switzerland, 2017; pp. 95–107. [Google Scholar]
  36. Qin, N.; Hu, X.; Dai, H. Deep fusion of multi-view and multimodal representation of ALS point cloud for 3D terrain scene recognition. ISPRS J. Photogramm. Remote Sens. 2018, 143, 205–212. [Google Scholar] [CrossRef]
  37. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional networks for biomedical image segmentation. In International Conference on Medical Image Computing and Computer-Assisted Intervention; Lecture Notes in Computer Science Series Volume 9351; Springer International Publishing: Cham, Switzerland, 2015; pp. 234–241. [Google Scholar]
  38. Badrinarayanan, V.; Kendall, A.; Cipolla, R. SegNet: A deep convolutional encoder-decoder architecture for image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 39, 2481–2495. [Google Scholar] [CrossRef]
  39. Shelhamer, E.; Long, J.; Darrell, T. Fully convolutional networks for semantic segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 640–651. [Google Scholar] [CrossRef] [PubMed]
  40. Malouf, R. A comparison of algorithms for maximum entropy parameter estimation. In Proceedings of the 6th Conference on Computational Natural Language Learning CoNLL 2002, Taipei, Taiwan, 31 August–1 September 2002. [Google Scholar]
  41. Zhang, Z.; Zhang, L.; Tong, X.; Mathiopoulos, P.; Guo, B.; Huang, X.; Wang, Z.; Wang, Y. A multilevel point-cluster-based discriminative feature for ALS point cloud classification. IEEE Trans. Geosci. Remote Sens. 2016, 54, 3309–3321. [Google Scholar] [CrossRef]
  42. Huang, H.; Wang, L.; Jiang, B. Precision verification of 3D SLAM backpacked mobile mapping robot. Bull. Surv. Mapp. 2016, 12, 68–73. [Google Scholar]
  43. Tong, G.; Li, Y.; Zhang, W.; Chen, D.; Zhang, Z.; Yang, J.; Zhang, J. Point set multi-level aggregation feature extraction based on multi-scale max pooling and LDA for point cloud classification. Remote Sens. 2019, 11, 2846. [Google Scholar] [CrossRef] [Green Version]
  44. Chang, C.; Lin, C. LIBSVM: A library for support vector machines. ACM Trans. Intell. Syst. Technol. 2011, 2, 27. [Google Scholar] [CrossRef]
  45. Niemeyer, J.; Rottensteiner, F.; Sörgel, U. Contextual classification of lidar data and building object detection in urban areas. ISPRS J. Photogramm. Remote Sens. 2014, 87, 152–165. [Google Scholar] [CrossRef]
Figure 1. The flowchart of the proposed method. (a) Training data; (b) testing data, (c) feature extraction; (d) the process of multiple views and space representation consistency under constraints of label consistency (MvsRCLC); (e) classification results. X1 and Xv represent the first view and the vth view features of the point cloud, respectively.
Figure 1. The flowchart of the proposed method. (a) Training data; (b) testing data, (c) feature extraction; (d) the process of multiple views and space representation consistency under constraints of label consistency (MvsRCLC); (e) classification results. X1 and Xv represent the first view and the vth view features of the point cloud, respectively.
Remotesensing 12 00135 g001
Figure 2. The schematic diagram of feature space and spatial position space joint constraints. (a) ground truth, (b) 5 neighbor points of p1 in spatial position space, (c) 5 neighbor points of point p1 for different view features in feature space, (d) the projection results (subspace features) with the joint constraints of spatial position space and feature space. Note that different numbers represent different point clouds. Different colors represent different classes. The red point labeled 1 is the selected point p1.
Figure 2. The schematic diagram of feature space and spatial position space joint constraints. (a) ground truth, (b) 5 neighbor points of p1 in spatial position space, (c) 5 neighbor points of point p1 for different view features in feature space, (d) the projection results (subspace features) with the joint constraints of spatial position space and feature space. Note that different numbers represent different point clouds. Different colors represent different classes. The red point labeled 1 is the selected point p1.
Remotesensing 12 00135 g002
Figure 3. Five different point cloud scenes used in this paper. Subfigures from (a) to (e) represent the scanned Scene1 to Scene5.
Figure 3. Five different point cloud scenes used in this paper. Subfigures from (a) to (e) represent the scanned Scene1 to Scene5.
Remotesensing 12 00135 g003aRemotesensing 12 00135 g003bRemotesensing 12 00135 g003c
Figure 4. Classification results of Scene2. (a) Ground truth; (b) our method; (c) RICA-SVM.
Figure 4. Classification results of Scene2. (a) Ground truth; (b) our method; (c) RICA-SVM.
Remotesensing 12 00135 g004
Figure 5. Classification results of Scene4. (a) Ground truth; (b) our method; (c) LCKSVD1; (d) LCKSVD2; (e) DKSVD; (f) RICA-SVM.
Figure 5. Classification results of Scene4. (a) Ground truth; (b) our method; (c) LCKSVD1; (d) LCKSVD2; (e) DKSVD; (f) RICA-SVM.
Remotesensing 12 00135 g005aRemotesensing 12 00135 g005b
Figure 6. Classification results of ALS point clouds selected from Scene2. (a) Ground truth; (b) our method; (c) AWP; (d) AMGL; (e) MLAN; (f) NNSG.
Figure 6. Classification results of ALS point clouds selected from Scene2. (a) Ground truth; (b) our method; (c) AWP; (d) AMGL; (e) MLAN; (f) NNSG.
Remotesensing 12 00135 g006
Figure 7. Classification results of MLS point cloud selected from Scene3. (a) Ground truth; (b) our method; (c) AWP; (d) AMGL; (e) MLAN; (f) NNSG. Note that our method obviously outperforms other methods in black box regions.
Figure 7. Classification results of MLS point cloud selected from Scene3. (a) Ground truth; (b) our method; (c) AWP; (d) AMGL; (e) MLAN; (f) NNSG. Note that our method obviously outperforms other methods in black box regions.
Remotesensing 12 00135 g007
Figure 8. Classification results on ISPRS 3D Semantic Labeling Dataset.
Figure 8. Classification results on ISPRS 3D Semantic Labeling Dataset.
Remotesensing 12 00135 g008
Figure 9. The influence of different parameters on the classification accuracy. (a,b) are the classification accuracy corresponding to different λ 2 and λ 3 in Scene1 and Scene4 when λ 1 is fixed, respectively. (c,d) are the classification accuracy corresponding to different λ 1 and λ 2 in Scene1 and Scene4 when λ 3 is fixed, respectively. (e,f) are the classification accuracy corresponding to different λ 1 and λ 3 in Scene1 and Scene4 when λ 2 is fixed, respectively.
Figure 9. The influence of different parameters on the classification accuracy. (a,b) are the classification accuracy corresponding to different λ 2 and λ 3 in Scene1 and Scene4 when λ 1 is fixed, respectively. (c,d) are the classification accuracy corresponding to different λ 1 and λ 2 in Scene1 and Scene4 when λ 3 is fixed, respectively. (e,f) are the classification accuracy corresponding to different λ 1 and λ 3 in Scene1 and Scene4 when λ 2 is fixed, respectively.
Remotesensing 12 00135 g009
Figure 10. Convergence processes for Scene1 and Scene4 in subfigure (a,b).
Figure 10. Convergence processes for Scene1 and Scene4 in subfigure (a,b).
Remotesensing 12 00135 g010
Table 1. Statistics of five point cloud datasets used in this paper. For airborne laser-scanning (ALS) point clouds, the number of training set and the test set are separated by symbol “/”. The symbol “--” represents the corresponding class does not exist.
Table 1. Statistics of five point cloud datasets used in this paper. For airborne laser-scanning (ALS) point clouds, the number of training set and the test set are separated by symbol “/”. The symbol “--” represents the corresponding class does not exist.
TypesALSMLSTLS
ScenesScene1Scene2Scene3Scene4Scene5
Trees68,802/213,99039,743/73,20765,295516,960214,151
Buildings37,128/200,54964,952/156,186312,475230,91088,818
Cars5380/78164584/740991,967103,98331,026
Pedestrians------278051,163
Wire poles----93522286--
Street lamps------32,713--
Traffic signs------1556--
Wires------3875--
Pylons------5196--
Total points111,310/422,355109,279/236,802479,089900,200385,158
Table 2. Classification results of Scene1 (%). F1-score corresponds to the accuracy in terms of buildings, cars and trees. Note that the highest accuracy values are highlighted in bold.
Table 2. Classification results of Scene1 (%). F1-score corresponds to the accuracy in terms of buildings, cars and trees. Note that the highest accuracy values are highlighted in bold.
Scene1MethodsOAmIoUKappa F1-ScoremF1
Multiple featuresOur method77.4945.0457.1581.65/10.93/75.3155.96
Adaboost62.0436.2539.5681.53/4.99/54.4246.98
LC-KSVD169.0641.1347.6478.39/6.41/71.4952.1
LC-KSVD268.9541.0647.5478.82/6.23/71.1751.98
DKSVD68.5740.6546.978.33/6.22/70.3951.65
RICA-SVM72.9443.7852.8582.18/8.17/72.8754.41
Single featureFC(our)77.244.6256.5881.39/10.25/74.9255.52
FSI(our)68.9338.344.7679.10/4.74/64.0349.29
FC(SVM)73.5143.3952.9582.22/7.87/71.9854.02
FSI(SVM)72.0543.2151.6980.97/6.72/73.5653.75
Table 3. Classification results of Scene2 (%). F1-score corresponds to the accuracy in terms of buildings, cars and trees. Note that the highest accuracy values are highlighted in bold.
Table 3. Classification results of Scene2 (%). F1-score corresponds to the accuracy in terms of buildings, cars and trees. Note that the highest accuracy values are highlighted in bold.
Scene2MethodsOAmIoUKappa F1-ScoremF1
Multiple featuresOur method84.8450.6667.3779.48/8.90/89.7559.38
Adaboost69.3235.9542.6581.53/4.99/54.4246.61
LC-KSVD175.3143.6650.5168.63/16.16/82.3655.71
LC-KSVD261.9234.4534.761.53/12.26/68.7947.53
DKSVD59.931.7627.5555.84/4.69/70.2543.59
RICA-SVM78.5642.1551.0466.44/3.61/85.6251.89
Single featureFC(our)84.4449.0266.9579.31/0.00/89.6856.33
FSI(our)79.1249.0558.173.10/22.37/87.0060.82
FC(SVM)80.8343.0253.6868.49/0.00/87.0251.83
FSI(SVM)82.5645.9960.5374.52/0.00/88.0053.71
Table 4. Running time of point cloud classification on Scene1 and Scene2 (Unit: second). Note that the best results are highlighted in bold.
Table 4. Running time of point cloud classification on Scene1 and Scene2 (Unit: second). Note that the best results are highlighted in bold.
ScenesOur MethodAdaboostLC-KSVD1 LC-KSVD2DKSVDRICA-SVM
Scene11.0332.85127.15125.82201.552.33
Scene20.7917.7929.0157.0685.661.35
Table 5. Classification results of Scene3 (%). F1-score corresponds to the accuracy in terms of poles/buildings/cars/trees. Note the highest accuracy values are highlighted in bold.
Table 5. Classification results of Scene3 (%). F1-score corresponds to the accuracy in terms of poles/buildings/cars/trees. Note the highest accuracy values are highlighted in bold.
Scene3MethodOA mIoUKappa F1-ScoremF1
Multiple featuresOur method70.6841.6551.1421.88/79.41/42.69/76.1755.01
Adaboost65.2338.745.8717.99/73.90/41.36/75.1552.1
LC-KSVD169.5642.2650.6720.56/78.40/43.54/79.0455.39
LC-KSVD269.3642.0150.4820.29/78.27/43.72/78.4855.19
DKSVD62.235.8639.8615.00/73.60/34.22/72.1448.74
RICA-SVM68.4141.7249.518.84/77.31/42.50/79.8554.63
Single featureFC(our)61.9935.0139.912.23/72.81/29.53/74.1947.19
FSI(our)69.7141.0549.9421.59/78.76/42.48/75.1554.49
FC(SVM)59.7834.938.4712.23/70.20/32.26/74.8347.38
FSI(SVM)68.0839.7847.8819.56/77.94/39.74/74.6952.98
Table 6. Classification results of Scene4 (%). F1-score corresponds to the accuracy in terms of poles/buildings/trees/street lights/pedestrians/cars/wires/towers/traffic signs. Note that the highest accuracy values are highlighted in bold.
Table 6. Classification results of Scene4 (%). F1-score corresponds to the accuracy in terms of poles/buildings/trees/street lights/pedestrians/cars/wires/towers/traffic signs. Note that the highest accuracy values are highlighted in bold.
Scene4MethodOA mIoUKappa F1-ScoremF1
Multiple featuresOur method80.9330.9268.6212.71/81.55/91.67/40.95/6.96
/60.15/28.71/24.87/25.48
41.46
Adaboost63.2227.1846.3523.16/69.86/77.96/49.25/2.16
/44.60/60.92/12.57/3.90
38.27
LC-KSVD177.4229.9763.9411.40/77.34/90.65/2.41/4.07
/55.49/39.90/7.74/17.15
40.68
LC-KSVD277.8930.1364.5312.09/78.51/90.68/41.31/3.88
/55.84/39.64/26.58/18.79
40.81
DKSVD76.3727.6162.1611.24/77.24/89.84/36.91/4.76
/52.58/32.87/22.36/9.59
37.49
RICA-SVM77.0831.1963.7811.77/78.31/89.64/45.98/5.19
/56.20/55.17/18.70/16.19
41.91
Single featureFC(our)63.2317.7444.782.35/67.37/81.94/5.29/2.12
/32.84/9.69/14.80/3.40
24.42
FSI(our)77.127.8863.49.79/80.31/88.91/39.52/5.51
/52.91/28.91/15.54/17.88
37.69
FC(SVM)59.6120.3842.122.55/68.57/77.75/13.52/1.57
/34.80/43.57/14.30/3.06
28.85
FSI(SVM)76.5528.7363.069.82/81.07/87.92/42.08/2.89
/54.28/34.85/15.86/20.89
38.85
Table 7. Classification results of Scene 5(%). F1-score corresponds to the accuracy in terms of pedestrians/trees/cars/buildings. Note that the highest accuracy values are highlighted in bold.
Table 7. Classification results of Scene 5(%). F1-score corresponds to the accuracy in terms of pedestrians/trees/cars/buildings. Note that the highest accuracy values are highlighted in bold.
Scene5MethodOA mIoUKappa F1-ScoremF1
Multiple featuresOur method69.739.3935.0831.93/84.29/16.28/72.5051.25
Adaboost64.6629.9318.0228.44/80.31/0.87/52.5640.55
LC-KSVD167.0337.5532.0432.49/80.75/22.98/66.7650.74
LC-KSVD267.1637.9132.0733.36/80.80/22.23/67.1450.89
DKSVD54.8427.3120.3719.20/72.57/18.75/47.7239.56
RICA-SVM67.9638.532.1736.71/80.81/18.24/69.8451.4
Single featureFC(our)6634.9228.1926.29/80.72/18.78/63.4747.31
FSI(our)51.3525.1820.7424.55/70.32/21.52/33.9537.59
FC(SVM)67.1634.728.826.04/81.65/15.86/63.1946.68
FSI(SVM)45.6817.5612.351.40/63.37/36.76/1.2425.69
Table 8. Classification results of Scene2 (%). Precision/recall/intersection over union (IoU)/F1-score, OA, mIoU, Kappa and mF1 (%). Note that the highest accuracy values are highlighted in bold.
Table 8. Classification results of Scene2 (%). Precision/recall/intersection over union (IoU)/F1-score, OA, mIoU, Kappa and mF1 (%). Note that the highest accuracy values are highlighted in bold.
MethodBuildingCarTreeOAmIoUKappa mF1
Our method83.3/76.5
/66.3/79.8
29.9/67.1
/26.1/41.4
91.6/83.1
/77.2/87.2
79.756.664.369.4
AWP69.2/78.6
/58.2/73.6
0.0/0.0
/0.0/0.0
84.4/84.3
/70.0/82.3
75.942.753.352
AMGL81.9/70.3
/60.9/75.7
23.0//77.3
/21.5/35.5
95.1/77.4
/74.4/85.3
74.952.358.465.5
MLAN48.7/100.0/
/48.7/65.5
0.0/0.0
/0.0/0.0
100.0/59.9
/59.9/74.9
64.836.241.446.8
NNSG73.3/64.3
/52.1/68.5
12.0/87.9
/11.8/21.1
99.2/34.2
/34.1/50.9
48.432.730.546.8
SVM84.6/66.7
/59.5/74.6
24.7/72.4/
/22.6/36.8
91.9/83.2
/77.5/87.3
76. 753.26066.3
FC(our)84.7/74.2
/65.5/79.1
27.1/65.4
/23.7/38.3
90.8/82.8
/76.4/86.6
78.655.262.768
FSI(our)77.2/64.8
/54.3/70.5
22.8/59.8
/19.7/33.0
88.0/80.1
/72.2/83.9
73.448.853.862.5
FC(SVM)83.8/65.4
/58.0/73.5
22.1/67.5
/20.0/33.3
90.1/80.7
/74.1/85.1
74.550.756.564
FSI(SVM)81.3/55.8
/49.5/66.2
21.0/52.3
/17.6/30.0
84.5/86.0
/74.3/85.2
73.147.15260.5
Table 9. Classification results of Scene3 (%). Precision/recall/IoU/F1-score, OA, mIoU, Kappa and mF1 (%). Note the highest accuracy values are highlighted in bold.
Table 9. Classification results of Scene3 (%). Precision/recall/IoU/F1-score, OA, mIoU, Kappa and mF1 (%). Note the highest accuracy values are highlighted in bold.
MethodPoleBuildingCarTreeOAmIoUKappamF1
Our method70.7/64.1
/50.7/67.3
62.1/70.1
/49.1/65.9
67.5/40.8
/34.1/50.9
68.3/92.9
/65.0/78.7
7049.75665.7
AWP32.4/74.5
/29.2/45.2
0.0/0.0
/0.0/0.0
18.9/32.1
/13.5/23.8
0.0/0.0
/0.0/0.0
26.610.72.217.2
AMGL31.1/22.4
/15.0/26.0
31.4/49.6
/23.8/38.5
31.0/16.9
/12.3/21.9
33.0/38.0
/21.5/35.3
31.718.1930.4
MLAN100.0/57.2
/57.2/72.8
69.2/57.8
/46.0/63.0
44.6/34.1
/24.0/38.7
60.3/100.0
/60.3/75.2
62.846.949.362.4
NNSG53.7/49.1
/34.5/51.3
69.7/27.7
/24.8/39.6
74.1/13.0
/12.5/22.1
36.6/91.9
/35.4/52.4
45.526.827.341.4
SVM63.0/66.0
/47.6/64.5
64.4/62.1
/46.2/63.2
61.1/46.7
/36.0/52.9
73.2/89.7
/67.6/80.6
66.149.354.865.3
FC(our)54.5/59.1
/39.6/56.7
55.5/61.0
/41.0/58.1
56.3/30.8
/24.9/39.8
67.4/85.7
/60.6/75.5
6547.853.364.1
FSI(our)68.5/62.2
/48.4/65.2
60.6/68.7
/47.5/64.4
61.2/42.6
/33.5/50.2
68.4/86.4
/61.7/76.4
59.241.545.557.5
FC(SVM)53.6/59.2
/39.2/56.3
57.3/55.9
/39.5/56.6
52.6/39.7
/29.2/45.3
69.8/81.3
/60.1/75.1
594245.458.3
FSI(SVM)66.4/60.6
/46.4/63.4
57.1/69.0
/45.5/62.5
56.3/36.0
/28.2/43.9
69.4/86.0
/62.3/76.8
62.945.650.561.7
Table 10. Classification results of ISPRS 3D Semantic Labeling Dataset (%). Precision/Recall/IoU/F1-score, OA, mIoU, Kappa and mF1 (%).
Table 10. Classification results of ISPRS 3D Semantic Labeling Dataset (%). Precision/Recall/IoU/F1-score, OA, mIoU, Kappa and mF1 (%).
MetricsGrounds CarsBuildingsTrees
Precision91.492.193.485
Recall94.248.387.390.8
F1-Score92.863.490.287.8
OA89.5
mIoU73.3
Kappa84.4
mF183.5
Table 11. Classification results of classification models with different configurations on Scene1 and Scene4 (%). Note the highest accuracy values are highlighted in bold.
Table 11. Classification results of classification models with different configurations on Scene1 and Scene4 (%). Note the highest accuracy values are highlighted in bold.
SceneMethodOA mIoUKappa
Scene1
(ALS)
IC172.9541.650.49
IC274.4643.3253.42
IC372.4243.0651.82
Ours77.4945.0457.15
Scene 4
(MLS)
IC179.2929.3566.37
IC280.6830.8668.3
IC379.3931.3867.05
Ours80.9330.9268.62

Share and Cite

MDPI and ACS Style

Tong, G.; Li, Y.; Chen, D.; Xia, S.; Peethambaran, J.; Wang, Y. Multi-View Features Joint Learning with Label and Local Distribution Consistency for Point Cloud Classification. Remote Sens. 2020, 12, 135. https://doi.org/10.3390/rs12010135

AMA Style

Tong G, Li Y, Chen D, Xia S, Peethambaran J, Wang Y. Multi-View Features Joint Learning with Label and Local Distribution Consistency for Point Cloud Classification. Remote Sensing. 2020; 12(1):135. https://doi.org/10.3390/rs12010135

Chicago/Turabian Style

Tong, Guofeng, Yong Li, Dong Chen, Shaobo Xia, Jiju Peethambaran, and Yuebin Wang. 2020. "Multi-View Features Joint Learning with Label and Local Distribution Consistency for Point Cloud Classification" Remote Sensing 12, no. 1: 135. https://doi.org/10.3390/rs12010135

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop