Next Article in Journal
Strategies for Soil Salinity Mapping Using Remote Sensing and Machine Learning in the Yellow River Delta
Previous Article in Journal
Estimation of Footprint-Scale Across-Track Slopes Based on Elevation Frequency Histogram from Single-Track ICESat-2 Photon Data of Strong Beam
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Transfer Learning Based on Multi-Branch Architecture Feature Extractor for Airborne LiDAR Point Cloud Semantic Segmentation with Few Samples

1
School of Remote Sensing and Information Engineering, Wuhan University, Wuhan 430072, China
2
Department of Oceanography, Dalhousie University, Halifax, NS B3H 4R2, Canada
3
Faculty of Resources and Environmental Science, Hubei University, Wuhan 430062, China
4
Tianjin Key Laboratory of Rail Transit Navigation Positioning and Spatio-Temporal Big Data Technology, Tianjin 300251, China
5
School of Electrical and Electronic Engineering, Wuhan Polytechnic University, Wuhan 430023, China
6
School of Resources Environment Science and Technology, Hubei University of Science and Technology, Xianning 437100, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2025, 17(15), 2618; https://doi.org/10.3390/rs17152618
Submission received: 30 May 2025 / Revised: 20 July 2025 / Accepted: 24 July 2025 / Published: 28 July 2025

Abstract

The existing deep learning-based Airborne Laser Scanning (ALS) point cloud semantic segmentation methods require a large amount of labeled data for training, which is not always feasible in practice. Insufficient training data may lead to over-fitting. To address this issue, we propose a novel Multi-branch Feature Extractor (MFE) and a three-stage transfer learning strategy that conducts pre-training on multi-source ALS data and transfers the model to another dataset with few samples, thereby improving the model’s generalization ability and reducing the need for manual annotation. The proposed MFE is based on a novel multi-branch architecture integrating Neighborhood Embedding Block (NEB) and Point Transformer Block (PTB); it aims to extract heterogeneous features (e.g., geometric features, reflectance features, and internal structural features) by leveraging the parameters contained in ALS point clouds. To address model transfer, a three-stage strategy was developed: (1) A pre-training subtask was employed to pre-train the proposed MFE if the source domain consisted of multi-source ALS data, overcoming parameter differences. (2) A domain adaptation subtask was employed to align cross-domain feature distributions between source and target domains. (3) An incremental learning subtask was proposed for continuous learning of novel categories in the target domain, avoiding catastrophic forgetting. Experiments conducted on the source domain consisted of DALES and Dublin datasets and the target domain consists of ISPRS benchmark dataset. The experimental results show that the proposed method achieved the highest OA of 85.5% and an average F1 score of 74.0% using only 10% training samples, which means the proposed framework can reduce manual annotation by 90% while keeping competitive classification accuracy.

1. Introduction

Airborne Laser Scanning (ALS) point clouds acquired by an airborne Light Detection and Ranging (LiDAR) system have demonstrated widespread applications in high-accuracy Digital Elevation Model (DEM) production [1], three-dimensional (3D) city modeling [2], vegetation detection and forest parameters retrieval [3], power line detection [4], and many others. One of the prerequisites to apply a point cloud to the aforementioned fields is the extraction of the thematic information from the original laser scanning data. In recent years, deep learning-based methods have achieved superior results in mobile laser scanning (MLS) and terrestrial laser scanning (TLS) point cloud semantic segmentation [5]. These methods can be summarized into three main categories [6], which are shown in Table 1. A recent comprehensive survey on deep learning techniques for 3D point cloud analysis was provided by Ref. [7]. Though they originated from the semantic segmentation of MLS and TLS point clouds, these methods have been extended to ALS point cloud analysis with minor modifications [8,9,10,11].
While deep learning techniques outperform traditional classifiers in various applications, they rely heavily on large volumes of manually annotated data for training. However, annotation remains a labor-intensive task, particularly for ALS point clouds. ALS point cloud data is not only costly to collect but also requires specialized expertise for annotation, incurring substantial time and financial expenses. Furthermore, like other machine learning methods, deep learning demands consistency in data distribution and feature space between the training dataset and the target application dataset. Failure to meet this requirement leads to degraded model performance, making it challenging to extend pre-trained models to datasets with inconsistent distributions. This issue is prevalent in airborne LiDAR data acquisition projects, where different LiDAR systems are deployed across diverse topographies. Data from various sensors may share common parameters such as 3D coordinates (X, Y, Z) while differing in sensor-specific attributes like intensity and multi-echo information. Additionally, category definitions vary across projects, resulting in datasets with distinct distributions. Traditional machine learning classifiers address such distribution discrepancies by retraining models from scratch using new data that matches the testing dataset’s distribution [22]. In contrast, a more efficient approach is to transfer knowledge learned from one task to other similar tasks with limited training data—a concept at the core of transfer learning.
The origins of transfer learning can be traced back to Bozinovski’s work in the 1970s [23]. Since then, numerous studies have applied this framework across diverse domains [24], including natural language processing, audio processing, image processing, and multi-source data fusion. In these fields, transfer learning has consistently demonstrated advantages in both efficiency and performance [25].
Transfer learning has also been investigated for ALS point cloud semantic segmentation and classification, aiming at reducing the requirements of annotation. Ma H et al. [26] first introduced model transfer for ALS point cloud ground filtering, which is used to segment ground points and non-ground points. In this study, the trained model was transferred to new data without additional training. The effectiveness of traditional classifiers such as random forest was verified. However, the performance of deep learning methods is often superior to that of traditional classifiers and does not require manual feature design and selection [27]. Some works extend transfer learning to deep learning methods. Peng et al. [28] and Xie et al. [29] introduced Unsupervised Domain Adaptation (UDA), where models pre-trained in the source domain were transferred to a target domain through correlation alignment and domain adversarial learning. However, these methods assume that the source and target data share a consistent label and feature space, which is often not the reality. Zhao et al. [30] converted point clouds into 2D representations and utilize pre-trained 2D networks for classification. Due to the loss of complex 3D geometric information, the performance for large scenes was limited [31].
In order to address the issue of inconsistent labels between source and target domains while directly acting on 3D points to avoid information loss caused by projection, Dai et al. [32] proposed a cross-domain incremental learning method with few samples. In this study, a photogrammetric point cloud and ALS point clouds were selected as the source domain and target domain, respectively. A conditional domain adversarial network (CDAN) [33] was introduced for domain adaptation of base categories consisting of ground, vegetation, and building. Incremental learning was used to acquire knowledge of novel categories in the target domain. Although the authors provided a novel approach of transfer learning for ALS point cloud learning using few samples, the performance of the transferred model was not as perfect as expected, which may have been due to the significant discrepancies between the source and target domains. These discrepancies can be summarized as follows:
(1)
Different scenes from where a LiDAR system collects data. This affects the number of categories and the feature distribution of ground objects. More complex scenes contain more object categories, leading to a harder semantic segmentation task.
(2)
Inconsistent parameters recorded in the point clouds. In addition to the common XYZ values, photogrammetric point clouds generally include the colors RGB, while ALS data may contain intensity and echo-related parameters. When the parameters of the target domain and the source domain are different, the transferability of the features learned by the pre-trained model is poor, and directly applying the pre-trained model to the target domain may result in unsatisfactory performance.
(3)
Inconsistencies in point density. Due to the variety of flight altitudes and LiDAR systems used in practice, data collected can have significant differences in point density, ranging from a few points to several hundred points per square meter. The same ground object may exhibit various features under inconsistent point densities; ideally, learned features should be density-invariant.
To address these challenges and to utilize multi-source ALS data as the source domain, a novel Multi-branch Feature Extractor and a novel transfer learning strategy were proposed, aimed at improving the performance of the model and reducing the amount of annotated data. The proposed method addresses the following issues: (1) It fully utilizes the parameters contained in ALS point clouds to learn salient features for semantic segmentation. (2) It overcomes difficulties in model transfer caused by differences in types of parameters between source domain and target domain. (3) It deals with inconsistent categories between source domain and target domain during domain adaptation. The main contributions include:
(1)
A novel multi-branch architecture was proposed to learn heterogeneous features (such as geometric features, reflection features, and internal structural features). Based on this architecture, a feature extraction network named Multi-branch Feature Extractor (MFE) was proposed, where each branch consists of the proposed Neighborhood Embedding Block (NEB) and the existing Point Transformer Block (PTB). A pre-training strategy is expounded to address the difference in parameters between source and target domains: A specific branch was pre-trained and transferred through domain adaptation. The number of branches is adjustable to fit the scenarios where the source domain consists of multiple point cloud datasets with inconsistent parameters.
(2)
A Multi-class Domain Adversarial Network (MDAN) was proposed to transfer the pre-trained feature extractor from source domain to target domain. By setting up independent discriminators for each base category, it promotes positive transfer of the base categories between source domain and target domain, avoiding negative transfer caused by unique categories.
(3)
An incremental learning strategy consisting of a Knowledge Transfer Module (KTM) and a Feature Separation Module (FSM) was designed to enable the model to segment novel categories while avoiding catastrophic base category data loss. For base categories, a Knowledge Transfer Module (KTM) consisting of Kullback–Leibler (KL) divergence and feature similarity was used to transfer knowledge. For novel categories, an FSM that utilized orthogonal loss functions as constraints was proposed to learn unique features for distinguishing base categories.
The remainder of this paper is organized as follows: Section 2 describes the proposed Multi-branch Feature Extractor and transfer learning strategy for ALS point cloud semantic segmentation with few samples. Section 3 describes the experimental results, including datasets, evaluation metrics, segmentation results, and method comparison. Section 4 presents the ablation experiments and discussions. Section 5 provides the conclusions and future directions for our work.

2. Method

In this section, the proposed method is described from five aspects: problem definition, Multi-branch Feature Extractor (MFE), pre-training subtask, domain adaption subtask, and incremental learning subtask.
The workflow of the proposed method is illustrated in Figure 1. To provide a clearer description, the same categories in the source and target domains are defined as base categories, and the unique categories in the target domain are defined as novel categories. The detailed definition is presented in Section 2.1. Existing studies have shown that asymmetric mapping can better simulate the differences in low-level features than symmetric mapping [34]; therefore, the pre-training subtask was separated from the domain adaptation subtask. Another advantage is that when facing a new dataset, there is no need for additional pre-training. At the same time, due to the presence of novel categories in the target domain, the model should continue to learn and acquire the ability to segment novel categories. However, introducing novel categories into the model may lead to a sudden loss of knowledge regarding base categories—a phenomenon known as catastrophic forgetting in transfer learning [35]. To mitigate this issue, an incremental learning subtask is proposed to enable the model to segment novel categories while avoiding catastrophic forgetting.
Hence, the proposed method consists of three modules corresponding to three subtasks: pre-training subtask, domain adaptation subtask, and incremental learning subtask. In the pre-training subtask, multi-source ALS data was used to pre-train the Multi-branch Feature Extractor F s , and semantic segmentation loss was used to constrain the learning process. In the domain adaptation subtask, target subdomain D t _ b with base categories were used to train F t _ b , and weights of F t _ b were initialized by F s . Multi-class Domain Adversarial Network (MDAN) was applied to learn domain invariant features. Semantic segmentation loss was used to learn the target domain’s unique features and train the classifier G t _ b of the target domain. In the incremental learning subtask, through KTM and FSM, target domain feature extractor F t and classifier G t were able to segment target domain data D t _ n with novel categories while retaining its ability to segment D t _ b . The finally obtained F t and G t had the abilities of semantic segmentation for the whole target domain D t .

2.1. Problem Definition

The benefits of transfer learning are generally believed to come from reusing the pre-trained feature hierarchy [36]. In the present study, features will be learned from a source domain that consists of inconsistent training datasets, and the learned features will be transferred to the downstream segmentation task, where there are novel categories. Therefore, the objective is to identify a feature representation that maximizes task relevance between the source and target domain while minimizing domain discrepancies. Considering the source domain D s = X s ,   C s and target domain D t = X t ,   C t , both comprise the parameters X and labels C. The purpose of transfer learning is to transfer the pre-trained feature extractor   F s to the target domain as F t . The same categories in both source and target domains are regarded as base categories C b = C s C t , and the unique categories of the target domain are regarded as novel categories C n = C t C b . Therefore, the target domain can be decomposed into subdomains D t _ b = X t _ b ,   C b and D t _ n = X t _ n ,   C n .
Armed with the above definition, the three subtasks shown in Figure 1 can be summarized and formulated as:
(1)
Pre-training subtask: Feature extractor F s was pre-trained in source domain D s .
(2)
Domain adaptation subtask: For the base categories C b , domain adaptation was employed to transfer the pre-trained F s adapted to the target domain as F t _ b in the subdomain D t _ b : F s D t _ b F t _ b , and target classifier G t _ b was trained in the subdomain D t _ b .
(3)
Incremental learning subtask: The feature extractor F t and classifier G t should be able to segment C n while retaining the ability to segment for C b in D t : F t _ b D t F t , G t _ b D t G t .

2.2. Multi-Branch Feature Extractor (MFE)

Given that the source domain may comprise multi-source datasets with inconsistent categories, point densities, and other attributes, a Multi-branch Feature Extractor (MFE) based on the Point Transformer was designed, as illustrated in Figure 2. The input data was divided into different branches, such as a coordinate branch, an intensity branch, and an echo-related branch, according to the types of parameters they contained. Feature aggregation within each branch was conducted by extended Point Transformer with proposed Neighborhood Embedding Block (NEB). Finally, the features learned from multiple branches were concatenated and fed into subsequent processing, such as classifiers.

2.2.1. Neighborhood Embedding Block (NEB)

In the classical Point Transformer, embedded features are encoded via a shared-weight MLP before being fed into the Point Transformer Block, which is effective for the TLS and MLS point clouds. However, this encoding method is hard to extend to ALS point clouds. For instance, ground points may be distributed arbitrarily across a scene, meaning two points that are far apart in coordinate space could belong to the same category. In contrast, directly encoding coordinate parameters using a shared-weight MLP would lead to significantly different feature representations for points of the same category.
In traditional classifiers for airborne LiDAR point clouds, it is necessary to handcraft features from raw data and input them into a classifier [37]. Neighborhood-based features, such as eigenvalues, play a crucial role in point cloud classification. Drawing on this insight, we hypothesize that incorporating point neighborhood information into the embedding module can effectively enhance the overall performance of the network. As shown in Figure 3, the NEB includes three steps: Firstly, parameters of the current point were embedded using MLP1. Secondly, parameters of the K-nearest-neighboring points were embedded by MLP2. Finally, the two encoded features were aggregated to obtain the finally embedded features through MLP3.
The pseudo code for the NEB is shown in Algorithm 1:
Algorithm 1. Neighborhood Embedding Block
Input: ALS point cloud data X N × D i n , where N is the number of points, and   D i n is the dimension of the input. For instance, if the input is X, Y, and Z, then D i n = 3 .
Output: Embedded features F N × D o u t , where D o u t is the dimension of the embedded features.
Initialization: P is point sets, K is the number of neighboring points.
1: For  p i i = 1 , 2 , 3 , , N in P:
2:  Search for the K-nearest-neighboring points of the current point ( p i ); the points within the neighborhood are denoted by X p i K × D i n .
3:  Obtain the original parameters of the current point ( p i ) from point cloud data ( X N × D i n ) as X p i D i n .
4:  Encode the current point using MLP1: X p i D i n M L P 1 F o p i D o u t .
5:  Encode the neighborhood of the current point using MLP2: X p i K × D i n M L P 2 F n p i D o u t .
6:  Combine the encoded features to obtain the final features through MLP3: [ F o p i D o u t   F n p i D o u t ] M L P 3 F p i D o u t .
7: End for

2.2.2. Concept of Multi-Branch Architecture

In the existing deep learning methods for point cloud semantic segmentation, the most commonly selected parameters include coordinate values, intensity, echo-related parameters, etc. These parameters are fed to the network simultaneously for feature learning. Though simple in network design, due to the different characteristics of ground objects reflected by different parameters, it might lead the inconspicuous features weighing out the salient ones, thereby decreasing the final segmentation accuracy. Numerous studies explore the relationship between input parameters and segmentation accuracy [38,39]. However, these studies also adopt the simultaneous input of parameters, leading to severe coupling of the influences exerted by different parameters. This coupling hinders the ability to discern the relationship between learned features and ground objects.
Thus, to our knowledge, there is currently limited research on the impact of different parameters on ALS point cloud semantic segmentation. In order to more intuitively demonstrate the impact, t-Distributed Stochastic Neighbor Embedding (t-SNE) was employed to evaluate the effectiveness of the learned features. The experiments were conducted on the ISPRS benchmark dataset and Point Transformer. The preprocessing of input data was the same as that of existing deep learning methods. The network was trained using the training set and was applied to the testing set. The features input into the classifier were taken out for visualization, as shown in Figure 3. Four combinations were conducted separately:
(a)
Mixed parameters (including XYZ values, intensity, and echo-related parameters), which are also commonly input into current deep learning methods for ALS point cloud analysis;
(b)
Echo-related parameters, which reflect the internal structure of ground objects;
(c)
Intensity, which characterizes the backscattering coefficient of ground objects;
(d)
XYZ values, which record the three-dimensional coordinate information of ground objects and are also the most commonly used parameters in traditional machine learning methods.
From Figure 4, it can be seen that when mixed parameters are input, the learned features are somewhat separable among low_veg, imp_surf, tree, and roof. However, the boundaries between these categories remain indistinct, and the intra-class distances are large. For the other five categories, the learned features are not clearly distinguishable. When the echo-related parameter is used, low_veg and imp_surf can be distinguished from tree and roof, which may be due to differences in internal structure. When intensity is used, low_veg and imp_surf can be distinguished, mainly due to their different backscattering coefficients. When XYZ values are used, tree and roof can be distinguished, which may be due to the differences in their geometric structures. Additionally, advantageous features of cars and fac can be learned when intensity and XYZ values are input separately.
As revealed by the t-SNE, features learned from different parameters are capable of distinguishing between various ground objects. Furthermore, since source domain data originates from diverse datasets with inconsistent parameter configurations, the number of inputs for feature learning remains variable. To address this, a multi-branch architecture is introduced. A key characteristic of this structure is that each specific branch is assigned distinct inputs and tasked with learning corresponding specific features, with the final output features being utilized for subsequent processing.

2.2.3. Construction of MFE Based on Multi-Branch Architecture

With the aforementioned facts in mind, MFE was proposed based on extended Point Transformer and multi-branch architecture. MFE consists of three branches corresponding to three types of parameters. Each branch consists of NEB and PTB. The three branches are:
(a)
Coordinate branch: the input includes X, Y, and Z values. In the NEB, besides X, Y, and Z values, relative Z values of the K-nearest-neighboring points of the current point were input into the embedding features. Relative Z values were calculated because relative Z values are significant features to distinguish ground objects [40].
(b)
Intensity branch: The input is the intensity. In the NEB, the intensity of the current point and that of the K-nearest-neighboring points were input into the embedding features.
(c)
Echo-related branch: The input is the preprocessed echo-related parameters. In the NEB, echo-related parameters of the current point and those of the K-nearest-neighboring points were input into embedding features.
Preprocessing is necessary for the coordinate values, intensity and echo-related parameters, as was described in detail in Section 3.2. The value K was determined by trial and error.

2.3. Pre-Training Subtask

The purpose of the pre-training subtask is to pre-train the feature extractor with multi-source ALS data. The larger the amount of pre-training data, the better the generalization performance of the pre-trained model [41]. Thus, we aimed to utilize as much data as possible for pre-training. However, multi-source ALS data involves diverse types of parameters. To address this issue, we constructed a three-branch architecture for the feature extractor, specifically comprising a coordinate branch, an intensity branch, and an echo-related branch. Each branch consists of an MFE and an MLP-based classifier.
The overall framework of the pre-training subtask is illustrated in Figure 5. First, the source domain data is divided into square blocks. Given the potential impact of point density variations on model transferability, voxelization is performed prior to feature learning. During voxelization, one point is randomly selected from each voxel, and these selected points are fed into the MFE. After voxelization, the point cloud density primarily depends on the voxel size—an approach that effectively mitigates differences in point cloud density across multi-source ALS data. To achieve this, the voxel size is determined based on the lowest point density among the datasets in both the source and target domains.
The feature extractor was optimized through semantic segmentation loss function in each branch:
a r g m i n G s ,   F s   L s e g _ s = i = 1 C b y i l o g G s F s x i ,
where y i is the true label, x i is the input of each branch, F s is feature extractor, and G s is the MLP-based classifier.
The semantic segmentation loss in the pre-training subtask can be defined as:
L s e g _ s = L s e g _ C + L s e g _ I + L s e g _ E ,
Furthermore, appropriate branches are selected based on the shared parameter types between the target domain and source domain to learn transferable features. For instance, if the source domain lacks intensity data, only the coordinate branch and echo-related branch undergo pre-training in the subtask before being transferred to the target domain. As for the intensity branch, supervised training is conducted using the target domain training set.

2.4. Domain Adaption Subtask

In order to transfer the pre-trained F s to the target domain to obtain F t _ b , two modules are proposed: (1) the Domain Adaptation Module, which aims to learn domain-invariant features, and (2) the Semantic Segmentation Module, which is designed to learn target domain-specific features and training target classifier G t _ b . A detailed description of these two modules is provided below.
(1) Domain Adaptation Module
Domain adaptation is recognized as a special case of transfer learning [42]. In recent years, deep domain adaptation has been proposed to integrate the properties of deep networks with domain adaptation techniques. Domain-Adversarial Training of Neural Networks (DANN) [43] was the first to introduce adversarial learning into deep domain adaptation, which maps source and target domain data symmetrically into a new feature space to learn domain-invariant features. CDAN [33] introduced category information to allow the network to learn domain invariant features corresponding to categories. Dai [32] applied CDAN for ALS data to achieve domain adaptation. However, CDAN performs optimally when the source and target domains share identical categories. Thus, directly using CDAN may cause the following issues: (1) Classifying source unique categories and target novel categories into one category will reduce the ability of the feature extractor. (2) The source’s unique categories and target’s novel categories may lead to negative transfer. (3) It cannot be guaranteed that all samples used for domain adaptation obtain the correct category, and the misclassified samples may also lead to negative transfer.
To address the above issues, inspired by Multi-Adversarial Domain Adaptation (MADA) [44], a separate adversarial network named Multi-class Domain Adversarial Network (MDAN) was proposed, as shown in Figure 6 and Figure 7. From Figure 6, domain-adversarial training was applied to each branch of the pre-trained Multi-branch Feature Extractor, aligning the feature distribution through MDAN. As shown in Figure 7, in MDAN, one discriminator corresponds to one base category; hence, there is an equal number of discriminators and base categories. The advantage of setting multiple discriminators is to alleviate negative transfer by preventing false alignment of modes in different distributions across domains. In the implementation of MDAN, the output features f s of the source feature extractor and the output features f t of the target feature extractor were selected by labels separately to obtain f s _ b , f t _ b as the input of MDAN. In MDAN, multiple domain discriminators were designed with the same number as C b .   f s _ b , f t _ b were input into the MLP-based domain discriminator to obtain the domain output, and adversarial loss was calculated and backpropagated through Gradient Reverse Layer (GRL). Compared to MADA, our method has the following advantages: (1) freezing the pre-trained feature extractor helps to preserve the feature extraction abilities of the source domain and improve the ability to extract domain invariant features, and (2) using labels instead of category probabilities to determine which discriminator the sample should input can help with positive transfer, avoiding negative transfer caused by misclassification.
Based on the above analysis, the loss function within each domain discriminator is calculated as follows:
a r g m i n D   a r g m a x F t _ b       L M D A N i = x j D s ,     x k D t _ b L D ( D F t b x j ,   d j + L D ( D F t b x k , d k
where i represents the i-th domain discriminator, L D is the loss function, D is the MLP-based domain discriminator, and d j and d k are the domain labels.
Then, the total adversarial loss is:
a r g m i n D   a r g m a x F t _ b         L a d v _ M D A N = j = 1 B i C b L M D A N i j ,
where B is the number of branches.
To reduce computational complexity and avoid domain label imbalance caused by significant differences in data volume between the source and target domains, an equal number of blocks are randomly selected from both domains in each epoch for domain adversarial learning.
(2) Semantic Segmentation Module
After domain adaptation, F t _ b obtains the ability to extract domain invariant features. However, the target domain and the source domain are not exactly the same. In order to learn target domain specific features and train the target classifier G t _ b , semantic segmentation loss is used, defined as follows:
a r g m i n G t _ b ,   F t _ b   L s e g _ b a s e = x i D t _ b y i l o g G t _ b F t _ b x i ,
Therefore, the overall loss function of domain adaptation subtask can be defined as:
L D A =   L s e g _ b a s e +   λ 1 L a d v _ M D A N ,
where λ 1 is a learnable weight parameter.

2.5. Incremental Learning Subtask

After the domain adaptation subtask, F t _ b obtains the ability of feature extraction for C b . Due to the presence of novel categories, the feature extractor should continue to learn for C n . If only C n is used for learning, the feature extractor would inevitably fall into catastrophic forgetting of C b . If C n and C b are used to train the F t and G t together, the model would lose its generalization ability to C b because the source domain is no longer visible. To address this issue, an incremental learning strategy was proposed that integrates knowledge transfer and feature separation. This strategy allows the model to learn the segmenting ability of C n while avoiding catastrophic forgetting for C b and retaining the generalization ability learned from the source domain. This subtask consists of two modules: (1) the Knowledge Transfer Module for C b and (2) the Feature Separation Module for C n . Next, we will provide a detailed description of these two modules.
(1) Knowledge Transfer Module (KTM)
The purpose of knowledge transfer is to enable the student model (SM) to retain the feature extraction capability of the teacher model (TM). Dai [32] introduced KL divergence as L K L into incremental learning to avoid catastrophic forgetting. However, the dimension of the classifier’s output layer was expanded, which could lead to the two identical points having different category probability distributions, and it could reduce the performance of knowledge transfer. Therefore, in addition to L K L , feature similarity was introduced as a constraint to overcome these differences. According to [45], Manhattan distance has better similarity measurement in prototype networks than Euclidean distance, cosine distance, and Chebyshev distance. At the same time, it is more stable. Therefore, for feature similarity measurement, Manhattan distance was chosen instead of other distances. Then, the loss function of feature similarity can be defined as:
a r g m i n F t ,     G t   L f e a t = 1 N x i D t _ b N F t x i F t _ b x i ,
Then, the loss function in KTM consists of KL divergence L K L and feature similarity L f e a t .
L K T = L K L + λ 2 L f e a t ,
where λ 2 is a learnable weight parameter.
(2) Feature Separation Module (FSM)
Due to the presence of novel categories C n , Dai [32] introduced an adversarial learning module, hoping for the model to learn distinguished features between C b and C n . However, the adversarial learning is not stable, and additional domain discriminators will increase the computational complexity. To solve this problem, the idea of feature separation was introduced using orthogonal constraint loss functions as a constraint to expect the network to learn unique and separable features for C n . Assuming that the features of C b are orthogonal to C n , the classifier can independently activate both types of features through linear transformation. Orthogonal constraints are established between the output features of F t and F t _ b .
a r g m i n   F t   L O r t h = F t _ b x i T F t x i F r o 2 ,     x i D t _ n ,
In order to make the learned features more effective for C n , semantic segmentation loss was used as an auxiliary:
a r g m i n G t ,   F t   L s e g _ n o v e l = x i D t _ n y i l o g G t F t x i
Then, the loss function in FSM consists of L O r t h and L s e g _ n o v e l :
L F S = L O r t h +   λ 3 L s e g _ n o v e l ,
where λ 3 is a learnable weight parameter.
In summary, the total loss function of incremental learning subtask can be defined as:
L I L = L K T + λ 4 L F S ,
where λ 4 is a learnable weight parameter.
Pseudocodes of the three-stage transfer learning strategy are provided as shown in Algorithm 2:
Algorithm 2. Three-stage transfer learning strategy
Pre-training subtask
Input: Source domain X N s , where N s is the number of total blocks after gridding the source domain.
Output: Pre-trained MFE P consists of Multi-branch Feature Extractor F s and an MLP-based classification layer within each branch. The number of branches is set to 3, namely coordinate branch MFE-C, intensity branch MFE-I, and echo-related branch MFE-E.
1: For each block X i in  X N s :
2:  Train MFE P -C, MFE P -I and MFE P -E using X i and calculate L s e g _ s .
3:  Backpropagate L s e g _ s and update MFE P .
4: End for
Domain adaptation subtask
Input: Source domain X N s and target domain with labeled base categories X N t _ b , where N s and N t _ b   are the number of blocks after gridding the source and target domain, respectively. Frozen pre-trained MFE P . MDAN consists of k MLP-based domain discriminators, where k is the number of base categories.
Output: Domain-adapted MFE D A consists of Multi-branch Feature Extractor F t _ b and MLP-based classification layer G t _ b .
1: For  i  in [1, 2, 3, …, N t _ b ]:
2:  Given a block X s i from source domain and a block X t i from target domain.
3:  For each branch j in {MFE-C, MFE-I, MFE-E}:
4:   Features f t , f s are obtained by: X t i F t _ b j f t ,  X s i F s j f s .
5:   For each base category k:
6:    Domain labels d k are obtained by M D A N k : [ f t _ k , f s _ k ]   M D A N k   d k .
7:    Loss L M D A N k j is calculated.
8:   End for
9:  Loss L a d v _ M D A N is calculated by: L a d v _ M D A N = j k L M D A N k j .
10: End for
11: Train MFE D A using X t i and calculate the segmentation loss L s e g _ b a s e .
12:  L D A =   L s e g _ b a s e +   λ 1 L a d v _ M D A N .
13: Backpropagate L D A , updating MFE D and MDAN.
14: End for
Incremental learning subtask
Input: Target domain X N t , where N t is the number of total blocks after gridding the target domain. Frozen domain adaptated MFE D A .
Output: MFE T consists of Multi-branch Feature Extractor F t and MLP-based classification layer G t .
1: For each block  X i  in  X N t :
2:  For points with labeled base categories:
3:   Features f t _ b and logists g t _ b are obtained by MFE D A : X i MFE D A f t _ b ,   g t _ b .
4:   Features f t _ b and logists g t _ b are obtained by MFE T : X i MFE T f t _ b ,   g t _ b .
5:    L K L is calculated by g t _ b and g t _ b ; L f e a t is calculated by f t _ b and f t _ b .
6:    L K T = L K L +   λ 2 L f e a t .
7:  End for
8:  For points with labeled novel categories:
9:   Features f t _ n are obtained by MFE D A : X i F t _ b f t _ n .
10:   Features f t _ n and logists g t _ b are obtained by M F E T : X t i M F E T f t _ b ,   g t _ b .
11:    L O r t h is calculated by f t _ n and f t _ n ; segmentation loss L s e g _ n o v e l is calculated.
12:    L F S = L O r t h +   λ 3 L s e g _ n o v e l .
13: End for
14:     L I L = L K T + λ 4 L F S
15: Backpropagate L I L and update MFE T .
16: End for

3. Experiments

In this chapter, we will introduce the experimental results from three aspects: datasets, evaluation metrics, and experimental results.

3.1. Datasets and Evaluation Metrics

Three datasets were used to evaluate the proposed model: the International Society for Photogrammetry and Remote Sensing (ISPRS) benchmark dataset [46,47], the Dublin City Annotated LiDAR Point Cloud dataset [48], and the Dayton Annotated LiDAR Earth Scan (DALES) dataset [49]. All these datasets were acquired via airborne LiDAR systems. Specifically, the DALES and Dublin datasets served as the source domains; the training set of the ISPRS benchmark dataset was used for domain adaptation and incremental learning, while its testing set was employed to evaluate the proposed method. Detailed information on these three datasets is presented in Table 2.
Three evaluation metrics were adopted to assess the segmentation results: F1 score (F1), average F1 (averF1), intersection over union (IoU), and overall accuracy (OA). Specifically, F1, averF1, and OA are used to compare with other fully supervised deep learning methods, while IoU is employed to compare with other few sample learning methods.
A paired t-test was used to examine significant differences between experimental results. In the ablation experiments, multiple trials were conducted, and the p-values of two-tailed tests were reported. If p < 0.05, it indicates that there is a statistically significant difference between the two compared models.

3.2. Experimental Results

3.2.1. Processing Details

1.
Data preprocessing
All three datasets were divided into blocks of 50 × 50 m, and coordinate values were centralized in each block. Due to the large range of the centralized coordinate values, intensity, and echo-related parameters, they were preprocessed by the following methods before being input into the network in order for the network to learn normally:
  • Coordinate values: Normalized after centralization, after which the range of value is 0~1.
  • Intensity: Normalization, after which the range of value is 0~1.
  • Echo-related parameters: Existing deep learning methods normalize the return number (RN) and number of returns (NR) separately [50]. Other studies pointed out that NR is positively correlated with the complexity of the internal structure [38]. Therefore, by combining RN and NR, a novel normalization method for echo-related parameters was proposed as follow:
F e =           0       ,     w h e n   N R = 1 R N N R ,     o t h e r w i s e   ,
The range of F e is [0, 1], where 0 indicates that the ground object cannot be penetrated. The closer the value is to 1, the more complexity there is of the internal structure of the object from where the laser beam reflected.
2.
Settings of pre-training subtask
Both the training set and testing set of the DALES dataset and Dublin dataset were used to pre-train MFE. A voxel size of 0.5 m was selected. The pre-training situation for each branch of MFE was as follows:
(1)
Coordinate branch: Pre-trained using the DALES dataset and Dublin dataset.
(2)
Echo-related branch: Pre-trained using the DALES dataset and Dublin dataset.
(3)
Intensity branch: Pre-trained using the Dublin dataset.
3.
Settings of domain adaptation subtask and incremental learning subtask
The training set of the ISPRS benchmark dataset was used in the domain adaptation subtask and incremental learning subtask. In order to evaluate the impact of different sampling proportions on the model, 10%, 1%, and 0.1% samples of the training set were selected through random sampling in each category, and the number of samples in each category was not less than 1. The number of samples is shown in Table 3.
For the domain adaptation subtask, samples with base categories were used. For the incremental learning subtask, samples with base categories were used in the Knowledge Transfer Module, and samples with novel categories were used in the Feature Separation Module.
After the three subtasks, feature extractor F t and classifier G t were evaluated on the testing set of the ISPRS benchmark dataset.
4.
Other settings
In all three subtasks, the Adam optimizer with an initial learning rate of 0.0005 was used, the momentum value was 0.9, and the batch size was 4. The learning rate was iteratively reduced based on the current epoch by a factor of 0.7. In the pre-training subtask, the training epoch was 30. In the domain adaptation subtask and incremental learning subtask, the training epoch was 300.

3.2.2. Comparison of Experimental Results

The proposed method was evaluated from the following two aspects: (1) comparison with other fully supervised deep learning methods and (2) comparison with other few sample learning methods.
1.
Comparison with other fully supervised deep learning methods
Table 4 shows the semantic segmentation results of the proposed method in comparison with other fully supervised deep learning methods published in recent literature. The thematic map of the segmentation result using 10% training samples and the ground truth are shown in Figure 8, while Figure 9 illustrates the difference between them. Figure 10 displays the zoomed-in map, which shows that some facades were misclassified as roofs, and some confusion also occurred between low_veg and fences, as well as between trees and roofs. These errors can be attributed to the close similarity in geometric and internal structural characteristics between these classes, making it difficult to learn distinctive features for their differentiation.
From Table 4, it can be seen that the proposed method achieved the highest OA and averF1, and achieved the highest F1 score in five out of nine categories. Among the six base categories, the segmentation performance was highly competitive, except for fence. This may have been due to the low point density of the ISPRS benchmark dataset, which resulted in incomplete morphology of the fence, and the proposed MFE relied on local features, leading to an inferior result for this category. Notably, the proposed method achieved the best performance in three novel categories (low_veg, fac, and shrub). This indicates that the method can effectively transfer knowledge learned from the source domain to the target domain, thereby enhancing the model’s generalization ability.
Meanwhile, the impact of various proportions of samples on semantic segmentation was compared. When using 1% of training samples, the model still achieved competitive results compared to other fully supervised deep learning methods. This was mainly due to the strong generalization ability of the model acquired from the source domain, which allowed the model to avoid overfitting, thereby maintaining the performance of the model with fewer target training samples. This greatly reduced the cost of manual annotation. This means that when using 0.1% of training samples, each category had only a few to less than 200 training samples, and the results deteriorated. Nevertheless, the method still achieved favorable performance for major ground objects such as imp_surf, tree, and roof, demonstrating its practical value in large-scale applications. By comparing the results of different sampling proportions, it can be seen that compared to using 100% training samples, using 10% training samples kept competitive classification accuracy while reducing manual annotations by 90%.
2.
Comparison with other few sample learning methods
To evaluate the proposed method for few sample learning, it was compared with Thr-MPRNet and Dai’s method [32]. In order to obtain more objective comparisons, the same dataset, SensatUrban [58], and sampling method as Dai [32] were adopted. A brief overview of the sampling method in Ref. [32] is as follows: (1) The training set is divided into grids with overlapping regions. (2) The class of each “representative block” is defined based on the proportion of points within each grid. If the number of points belonging to the selected category within a block exceeds a certain threshold, then the block is considered the “representative block” of this category. (3) In each “representative block,” N points are sampled and their annotation information is retained. A more detailed description of the sampling method can be found in the original paper.
Ground, vegetation, and building were taken as base categories. As the SensatUrban dataset contains coordinate values XYZ and color information RGB, in the pre-training subtask, only the coordinate branch was set up and pre-trained using coordinate values XYZ, while RGB information was abandoned. The sampled training set of the ISPRS benchmark dataset was used to train the intensity branch and echo-related branch in the domain adaptation subtask. The detailed settings are shown in Table 5.
As shown in Table 6, Thr-MPRNet and Dai’s methods achieved inferior results, which may have been due to (1) difficulty in transferring the pre-trained model to the target domain due to the difference in parameter type between the source and target domains. Simply embedding the parameters through the MLP-based embedding layer can lead to poor transferability of the learned features. (2) There was a significant difference in point density between the source and target domains, which may have resulted in significant differences in local features. This is not conducive to the transfer of features. The proposed Multi-branch Feature Extractor and transfer learning strategy seem likely to solve these two problems effectively.

4. Ablation Experiments and Discussions

The experimental results show the effectiveness of the proposed method. Ablation experiments were conducted to further study the impact of each component of the proposed method. We will separately discuss the proposed NEB, multi-branch architecture, and transfer learning strategy.

4.1. Discussions of Neighborhood Embedding Block

Experiments were performed to assess the effectiveness of NEB and compare the impact of different K-nearest-neighbor sizes on model performance. To eliminate interference from the multi-branch architecture, only the coordinate branch was adopted. The experimental results are presented in Table 7, where “Base” denotes the original Point Transformer utilizing an MLP-based embedding block. For each K-nearest-neighbor setting, the experiment was repeated five times, with the averages and ranges of F1 scores, OA, and averF1 reported, along with p-values for OA and averF1 in comparison to the “Base”. As shown in the table, incorporating neighborhood information into NEB can effectively enhance model performance with increasing K values. However, when K exceeds 16, the model performance starts to decline. This phenomenon may be attributed to the fact that a larger K might encode features from other categories, leading to inconsistent embedded features among points of the same category and thus hindering subsequent feature learning and overall model performance. When K < 64, both p-values for OA and averF1 are less than 0.05, indicating statistically significant improvements. In contrast, when K = 64, the p-value for averF1 exceeds 0.05, suggesting that the improvement is not statistically significant.

4.2. Discussions of Multi-Branch Architecture

As mentioned before, the two purposes of multi-branch architecture in the feature extractor are (a) to learn more representative features and (b) to deal with variable parameters contained in a point cloud dataset. In this subsection, an ablation experiment was conducted to show the effectiveness of the first purpose. In the ablation experiment, the descriptions of the two models used are shown in Table 8. To avoid the impact of NEB, an MLP-based embedding layer was used for feature embedding. The flowchart of the model is shown in Figure 11.
The experimental results are presented in Table 9. Each experiment was conducted five times, with the averages and ranges of the F1 scores, OA, and averF1 reported, along with p-values for OA and averF1 comparing M1 and M2. Both p-values (for OA and averF1) were less than 0.05, indicating a statistically significant difference between M1 and M2. As observed, the multi-branch architecture significantly improved performance across all nine categories. Among them, the segmentation accuracy of powerlines and fences showed the most substantial increase, which can be attributed to the separate echo-related branch effectively extracting their internal structural features. There was also a noticeable improvement in low_veg segmentation, possibly due to the intensity branch’s capability to extract distinguishable features. For buildings, the enhanced segmentation accuracy may stem from the unique geometric features extracted by the coordinate branch. Combined with the t-SNE analysis in Section 2.3 and the experimental results, the multi-branch architecture is confirmed to be superior to the single-branch counterpart. This superiority may be attributed to the fact that different parameters correspond to distinct physical characteristics, enabling the multi-branch architecture to more effectively learn features that distinguish various ground objects.
To evaluate the time complexity of the proposed architecture, M1 and M2 were executed on the testing set of ISPRS benchmark dataset five times and the running time was separately recorded, as shown in Table 10. Though the multi-branch architecture almost doubles the running time in average compared to the single branch architecture, it is still in the acceptable range. But the most important is that it can effectively solve the issue of parameters differences between the source and target domains, therefore, greatly improve the performance of model transfer.

4.3. Discussion of Domain Adaptation Subtask and Incremental Learning Subtask

To demonstrate the effectiveness of domain adaptation subtask and incremental learning subtask, comparative experiment and ablation experiment were conducted. 10% samples of the target training set were used for domain adaptation and incremental learning. The trained model was evaluated on target testing set through segmentation performance.
(1) Discussion of Domain Adaptation Subtask
To verify the effectiveness of the proposed multi-class domain adversarial network (MDAN), comparative experiment was conducted where Fine-tuning (FT) method was used as baseline and MADA as comparison. In Fine-tuning, pre-trained Multi-branch Feature Extractor was frozen and only MLP-based classifier was trained in target subdomain D t _ b . For MADA, due to the need for category probability to determine which domain discriminator the sample was input, the pre-trained G s in the pre-training subtask was connected to F t _ b as G s to obtain category probability, and G s was updated by semantic segmentation loss.
The experimental results are shown in Table 11. Since domain adaptation subtask was applied for C b , only power, imp_surf (ground), car, fence, roof (building) and tree (vegetation) were considered. The experiments were conducted five times, the average and ranges of F1 scores and OA were reported, along with p-values (OA) compared to MDAN. The p-values (OA) of FT and MADA were both less than 0.05, indicating there were significant differences. From the table, it can be seen that MDAN could transfer learned knowledge more effectively compared with MADA, which likely contributed to the MDAN strategy of solely using points of the same category for domain adaptation, avoiding negative transfer. Specifically, for categories with fewer samples of power, fences, and cars, the improvement was more pronounced compared with FT. This may be due to the fact that the proposed method can avoid overfitting and improve the segmentation performance.
(2) Discussion of Incremental Learning Subtask
As for the incremental learning subtask, ablation experiments on KTM and FSM were carried out separately. To evaluate KTM, an MLP-base classifier was fine-tuned using the labeled samples from C n to obtain the ability to segment novel categories. The experimental results are shown in Table 12. “Base” represents the segmentation results of C b after the domain adaptation subtask. The experiments were conducted five times, and the average and ranges of F1 scores and OA were reported, along with the p-values (OA) compared to the proposed KTM + FSM. The p-values (OA) of FT and MADA were both less than 0.05, indicating there were significant differences. It can be seen that when only using FSM, compared to fine-tuning, the model performed better in segmenting novel categories C n , but it would fall into catastrophic forgetting, resulting in poor performance for base categories C b . The combination of the proposed KTM and FSM improved the segmentation performance of novel categories while avoiding catastrophic forgetting the knowledge learned from the base categories.
In order to better evaluate the proposed FSM, experiments were conducted five times to compare with the Semantic Adversarial Learning Module (SALM) proposed by Dai [25]. The average and ranges of F1 scores and OA were reported, along with p-values (OA) compared between SALM and FSM. The p-value (OA) was less than 0.05, indicating there was significant difference. The experimental results in Table 13 show that the proposed FSM achieved better results in novel categories and was more stable.

4.4. Analysis of the Impact of Random Sampling

As mentioned in Section 3.2, few labeled samples were selected via random sampling. Experiments were conducted to evaluate the impact of this process. For proportions of 10%, 1%, and 0.1%, five random samplings were performed for each, with OA and averF1 recorded. The experimental results are presented in Table 14. In general, both OA and averF1 exhibited stability, despite minor fluctuations in averF1. This indicates that random sampling had little impact on segmentation performance, thereby confirming the feasibility of this sampling method.

5. Conclusions

In this study, a novel Multi-branch Feature Extractor and transfer learning strategy were proposed, aiming at improving the performance of the model and reducing the amount of annotated data. MFE was based on multi-branch architecture, which is able to extract more representative features and deal with inconsistent parameter types between the source domain and target domain. A three-stage transfer learning strategy was proposed to cope with feature transfer and catastrophic forgetting: (1) The pre-training subtask effectively utilizes multi-source ALS data to obtain pre-trained feature extractor. The source domain consists of datasets collected by different LiDAR systems; therefore, there are data discrepancies in term of parameters recorded, point density, topography, and landforms. MFE plays a core role in this stage to learn features. (2) In the domain transfer subtask, MDAN aligns the distributions of the source and target domains for each category separately through multiple domain discriminators, avoiding negative transfer caused by unique categories of the source domain and novel categories of the target domain. (3) The incremental learning subtask, consisting of KTM and FSM, was designed to learn the knowledge of novel categories while avoiding catastrophic forgetting for base categories. The experiments conducted on the ISPRS benchmark dataset validated the effectiveness of the proposed method. Through the proposed method, using only 10% labeled training samples can achieve competitive segmentation results comparable to other fully supervised methods. Meanwhile, the proposed method is also superior to other few sample learning methods. By fully utilizing existing annotated ALS data, the need for manual annotation is greatly reduced when facing a new dataset.
Although the proposed method reduced the need for manual annotation, the cost was the increase in computational complexity, especially for the domain transfer subtask. The existing few sample learning methods require retraining for a new dataset. The proposed method solves this problem by separating the pre-training subtask from the domain adaptation subtask. However, in the domain adaptation subtask, it is still necessary to extract features from the source domain to align the features, which will increase the computational load. Reducing the computational load of the domain adaptation subtask is still a problem that needs further research. Airborne LiDAR data is difficult to collect compared to spectral images in Earth observation. Exploring the possibility of cross-modal transfer from spectral images to LiDAR point clouds is meaningful and listed as the future research direction of the authors’ R&D. Meanwhile, MFE has demonstrated its application value in semantic segmentation. Extending it to other 3D data processing tasks (e.g., object detection) is also a potential research direction.

Author Contributions

Conceptualization, J.Y. and H.M.; data curation, J.Y. and J.D.; funding acquisition, L.Z., K.L. and J.D.; investigation, J.D. and W.L.; methodology, J.Y.; project administration, H.M.; software, L.Z., W.L., K.L. and Z.C.; supervision, H.M.; validation, J.Y., K.L. and L.Z.; visualization, J.Y., K.L. and L.Z.; writing—original draft, J.Y.; writing—review and editing, H.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Tianjin Key Laboratory of Rail Transit Navigation Positioning and Spatio-temporal Big Data Technology, grant number TKL2023B13, and Natural Science Foundation of Hubei Province, grant number 2025AFB833.

Data Availability Statement

The data presented in this study is available in the DALES dataset at “https://udayton.edu/engineering/research/centers/vision_lab/research/was_data_analysis_and_processing/dale.php (accessed on 30 May 2025)”, the Dublin dataset at “https://v-sense.scss.tcd.ie/DublinCity (accessed on 30 May 2025)”, and the ISPRS benchmark dataset at “https://isprs.org/resources/datasets/benchmarks/UrbanSemLab/3d-semantic-labeling.aspx (accessed on 30 May 2025)”.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Liu, X.; Zhang, Z.; Peterson, J.; Chandra, S. Large area DEM generation using airborne LiDAR data and quality control. In Accuracy in Geomatics: Proceedings of the 8th International Symposium on Spatial Accuracy Assessment in Natural Resources and Environmental Sciences, Shanghai, China, 25–27 June 2008; World Academic Union: Liverpool, UK, 2008; Volume 2. [Google Scholar]
  2. Murakami, H.; Nakagawa, K.; Hasegawa, H.; Shibata, T.; Iwanami, E. Change detection of building using an airborne laser scanner. ISPRS J. Photogramm. Remote Sens. 1999, 54, 148–152. [Google Scholar] [CrossRef]
  3. Ivanovs, J.; Lazdins, A.; Lang, M. The influence of forest tree species composition on the forest height predicted from airborne laser scanning data—A case study in Latvia. Balt. For. 2023, 29, id663. [Google Scholar] [CrossRef]
  4. Matikainen, L.; Lehtomäki, M.; Ahokas, E.; Hyyppä, J.; Karjalainen, M.; Jaakkola, A.; Kukko, A.; Heinonen, T. Remote sensing methods for power line corridor surveys. ISPRS J. Photogramm. Remote Sens. 2016, 119, 10–31. [Google Scholar] [CrossRef]
  5. Guo, Y.; Wang, H.; Hu, Q.; Liu, H.; Liu, L.; Bennamoun, M. Deep learning for 3d point clouds: A survey. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 43, 4338–4364. [Google Scholar] [CrossRef] [PubMed]
  6. An, Z.; Sun, G.; Liu, Y.; Liu, F.; Wu, Z.; Wang, D.; Van Gool, L.; Belongie, S. Rethinking few-shot 3d point cloud semantic segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 3996–4006. [Google Scholar]
  7. Sarker, S.; Sarker, P.; Stone, G.; Gorman, R.; Tavakkoli, A.; Bebis, G.; Sattarvand, J. A comprehensive overview of deep learning techniques for 3D point cloud classification and semantic segmentation. Mach. Vis. Appl. 2024, 35, 67. [Google Scholar] [CrossRef]
  8. Huang, R.; Xu, Y.; Stilla, U. GraNet: Global relation-aware attentional network for semantic segmentation of ALS point cloud. ISPRS J. Photogramm. Remote Sens. 2021, 177, 1–20. [Google Scholar] [CrossRef]
  9. Liang, Z.; Lai, X. Multilevel geometric feature embedding in transformer network for ALS point cloud semantic segmentation. Remote Sens. 2024, 16, 3386. [Google Scholar] [CrossRef]
  10. Liu, T.; Wei, B.; Hao, J.; Li, Z.; Ye, F.; Wang, L. A multi-point focus transformer approach for large-scale ALS point cloud ground filtering. Int. J. Remote Sens. 2025, 46, 979–999. [Google Scholar] [CrossRef]
  11. Zhao, J.; Zhou, H. Cross-layer Features Fusion Network with Attention MLP for ALS Point Cloud Segmentation. In Proceedings of the 5th International Conference on Computer Vision, Zhuhai, China, 19–21 April 2024; Image and Deep Learning (CVIDL). IEEE: New York, NY, USA, 2024; pp. 1110–1114. [Google Scholar]
  12. Hu, Q.; Yang, B.; Xie, L.; Rosa, S.; Guo, Y.; Wang, Z.; Trigoni, N.; Markham, A. Randla-net: Efficient semantic segmentation of large-scale point cloud. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 11108–11117. [Google Scholar]
  13. Qi, C.R.; Su, H.; Mo, K.; Guibas, L.J. Pointnet: Deep learning on point sets for 3d classification and segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 652–660. [Google Scholar]
  14. Qi, C.R.; Yi, L.; Su, H.; Guibas, L.J. Pointnet++: Deep hierarchical feature learning on point sets in a metric space. Adv. Neural Inf. Process Syst. 2017, 30, 5105–5114. [Google Scholar]
  15. Phan, A.V.; Le Nguyen, M.; Nguyen, Y.L.H.; Bui, L.T. Dgcnn: A convolutional neural network over large-scale labeled graphs. Neural Netw. 2018, 108, 533–543. [Google Scholar] [CrossRef]
  16. Zhou, Y.; Tuzel, O. Voxelnet: End-to-end learning for point cloud based 3d object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 4490–4499. [Google Scholar]
  17. Wang, Y.; Sun, Y.; Liu, Z.; Sarma, S.E.; Bronstein, M.M.; Solomon, J.M. Dynamic graph cnn for learning on point cloud. ACM Trans. Graph. (TOG) 2019, 38, 1–12. [Google Scholar] [CrossRef]
  18. Zhao, H.; Jiang, L.; Jia, J.; Torr, P.H.; Koltun, V. Point transformer. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 16259–16268. [Google Scholar]
  19. Park, C.; Jeong, Y.; Cho, M.; Park, J. Fast point transformer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 16949–16958. [Google Scholar]
  20. Wang, P.S. Octformer: Octree-based transformers for 3d point clouds. ACM Trans. Graph. (TOG) 2023, 42, 1–11. [Google Scholar] [CrossRef]
  21. Guo, M.H.; Cai, J.X.; Liu, Z.N.; Mu, T.J.; Martin, R.R.; Hu, S.M. Pct: Point cloud transformer. Comput. Vis. Media 2021, 7, 187–199. [Google Scholar] [CrossRef]
  22. Kheddar, H.; Himeur, Y.; Al-Maadeed, S.; Amira, A.; Bensaali, F. Deep transfer learning for automatic speech recognition: Towards better generalization. Knowl.-Based Syst. 2023, 277, 110851. [Google Scholar] [CrossRef]
  23. Bozinovski, S. Reminder of the first paper on transfer learning in neural networks, 1976. Informatica 2020, 44, 291. [Google Scholar] [CrossRef]
  24. Iman, M.; Arabnia, H.R.; Rasheed, K. A review of deep transfer learning and recent advancements. Technologies 2023, 11, 40. [Google Scholar] [CrossRef]
  25. Niu, S.; Liu, Y.; Wang, J.; Song, H. A decade survey of transfer learning (2010–2020). IEEE Trans. Artif. Intell. 2021, 1, 151–166. [Google Scholar] [CrossRef]
  26. Ma, H.; Cai, Z.; Zhang, L. Comparison of the filtering models for airborne LiDAR data by three classifiers with exploration on model transfer. J. Appl. Remote Sens. 2018, 12, 016021. [Google Scholar] [CrossRef]
  27. Ahmed, S.F.; Alam, M.S.B.; Hassan, M.; Rozbu, M.R.; Ishtiak, T.; Rafa, N.; Mofijur, M.; Shawkat Ali, A.; Gandomi, A. Deep learning modelling techniques: Current progress, applications, advantages, and challenges. Artif. Intell. Rev. 2023, 56, 13521–13617. [Google Scholar] [CrossRef]
  28. Peng, S.; Xi, X.; Wang, C.; Xie, R.; Wang, P.; Tan, H. Point-based multilevel domain adaptation for point cloud segmentation. IEEE Geosci. Remote Sens. Lett. 2020, 19, 1–5. [Google Scholar] [CrossRef]
  29. Xie, Y.; Schindler, K.; Tian, J.; Zhu, X.X. Exploring cross-city semantic segmentation of ALS point clouds. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2021, 43, 247–254. [Google Scholar] [CrossRef]
  30. Zhao, C.; Yu, D.; Xu, J.; Zhang, B.; Li, D. Airborne LiDAR point cloud classification based on transfer learning. In Proceedings of the Eleventh International Conference on Digital Image Processing (ICDIP 2019), Guangzhou, China, 10–13 May 2019; SPIE: Bellingham, WA, USA, 2019; Volume 11179, pp. 550–556. [Google Scholar]
  31. Dai, M.; Xing, S.; Xu, Q.; Li, P.; Pan, J.; Zhang, G.; Wang, H. Multiprototype Relational Network for Few-Shot ALS Point Cloud Semantic Segmentation by Transferring Knowledge From Photogrammetric Point Clouds. IEEE Trans. Geosci. Remote Sens. 2024, 62, 1–17. [Google Scholar] [CrossRef]
  32. Dai, M.; Xing, S.; Xu, Q.; Li, P.; Pan, J.; Wang, H. Cross-Domain Incremental Feature Learning for ALS Point Cloud Semantic Segmentation with Few Samples. IEEE Trans. Geosci. Remote Sens. 2024, 63, 1–14. [Google Scholar] [CrossRef]
  33. Long, M.; Cao, Z.; Wang, J.; Jordan, M.I. Conditional Adversarial Domain Adaptation. In Proceedings of the Computer Vision—ECCV 2018—15th European Conference, Munich, Germany, 8–14 September 2018; Springer: Cham, Switzerland, 2018. [Google Scholar] [CrossRef]
  34. Tzeng, E.; Hoffman, J.; Saenko, K.; Darrell, T. Adversarial Discriminative Domain Adaptation. arXiv 2017. [Google Scholar] [CrossRef]
  35. Kirkpatrick, J.; Pascanu, R.; Rabinowitz, N.; Veness, J.; Desjardins, G.; Rusu, A.A.; Milan, K.; Quan, J.; Ramalho, T.; Grabska-Barwinska, A. Overcoming catastrophic forgetting in neural networks. Proc. Natl. Acad. Sci. USA 2017, 114, 3521–3526. [Google Scholar] [CrossRef]
  36. Neyshabur, B.; Sedghi, H.; Zhang, C. What is being transferred in transfer learning? Adv. Neural Inf. Process. Syst. 2020, 33, 512–523. [Google Scholar]
  37. Ni, H.; Lin, X.; Zhang, J. Classification of ALS point cloud with improved point cloud segmentation and random forests. Remote Sens. 2017, 9, 288. [Google Scholar] [CrossRef]
  38. Reymann, C.; Lacroix, S. Improving LiDAR point cloud classification using intensities and multiple echoes. In Proceedings of the 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Hamburg, Germany, 28 September–3 October 2015; IEEE: New York, NY, USA, 2015; pp. 5122–5128. [Google Scholar]
  39. Stanislas, L.; Nubert, J.; Dugas, D.; Nitsch, J.; Sünderhauf, N.; Siegwart, R.; Cadena, C.; Peynot, T. Airborne particle classification in lidar point clouds using deep learning. In Field and Service Robotics: Results of the 12th International Conference; Springer: Singapore, 2021; pp. 395–410. [Google Scholar]
  40. Zhang, K.; Ye, L.; Xiao, W.; Sheng, Y.; Zhang, S.; Tao, X.; Zhou, Y. A dual attention neural network for airborne LiDAR point cloud semantic segmentation. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–17. [Google Scholar] [CrossRef]
  41. Mahajan, D.; Girshick, R.; Ramanathan, V.; He, K.; Paluri, M.; Li, Y.; Bharambe, A.; Van Der Maaten, L. Exploring the limits of weakly supervised pretraining. In Proceedings of the European conference on computer vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 181–196. [Google Scholar]
  42. Panigrahi, S.; Nanda, A.; Swarnkar, T. A survey on transfer learning. In Intelligent and Cloud Computing: Proceedings of ICICC 2019, Bhubaneswar, India, 20 December 2019; Springer: Singapore, 2020; Volume 1, pp. 781–789. [Google Scholar]
  43. Ganin, Y.; Ustinova, E.; Ajakan, H.; Germain, P.; Larochelle, H.; Laviolette, F.; March, M.; Lempitsky, V. Domain-Adversarial Training of Neural Networks. J. Mach. Learn. Res. 2016, 17, 1–35. [Google Scholar] [CrossRef]
  44. Pei, Z.; Cao, Z.; Long, M.; Wang, J. Multi-Adversarial Domain Adaptation. In Proceedings of the AAAI conference on artificial intelligence, New Orleans, LA, USA, 2–7 February 2018; AAAI Press: Palo Alto, CA, USA, 2018. [Google Scholar] [CrossRef]
  45. Yu, Z.; Wang, K.; Xie, S.; Zhong, Y.; Lv, Z. Prototypical Network Based on Manhattan Distance. CMES-Comput. Model. Eng. Sci. 2022, 131, 655. [Google Scholar] [CrossRef]
  46. Cramer, M. The dgpf-test on digital airborne camera evaluation–overview and test design. Photogramm.-Fernerkund.-Geoinf. 2010, 73–82. [Google Scholar] [CrossRef]
  47. Rottensteiner, F.; Sohn, G.; Jung, J.; Gerke, M.; Baillard, C.; Benitez, S.; Breitkopf, U. The isprs benchmark on urban object classification and 3D building reconstruction. ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci. 2012, 1–3, 293–298. [Google Scholar] [CrossRef]
  48. Iman Zolanvari, S.M.; Ruano, S.; Rana, A.; Cummins, A.; da Silva, R.E.; Rahbar, M.; Smolic, A. Dublin City: Annotated LiDAR Point Cloud and its Applications. arXiv 2019. [Google Scholar] [CrossRef]
  49. Varney, N.; Asari, V.K.; Graehling, Q. DALES: A Large-scale Aerial LiDAR Data Set for Semantic Segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Seattle, WA, USA, 14–19 June 2020; IEEE: New York, NY, USA, 2020. [Google Scholar] [CrossRef]
  50. Wang, J.; Li, H.; Xu, Z.; Xie, X. Semantic segmentation of urban airborne LiDAR point clouds based on fusion attention mechanism and multi-scale features. Remote Sens. 2023, 15, 5248. [Google Scholar] [CrossRef]
  51. Thomas, H.; Qi, C.R.; Deschaud, J.E.; Marcotegui, B.; Goulette, F.; Guibas, L.J. KPConv: Flexible and Deformable Convolution for Point Clouds. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; IEEE: New York, NY, USA, 2020. [Google Scholar] [CrossRef]
  52. Li, X.; Wang, L.; Wang, M.; Wen, C.; Fang, Y. Dance-net: Density-aware convolution networks with context encoding for airborne lidar point cloud classification. ISPRS J. Photogramm. Remote Sens. 2020, 166, 128–139. [Google Scholar] [CrossRef]
  53. Li, W.; Wang, F.-D.; Xia, G.-S. A geometry-attentional network for ALS point cloud classification. ISPRS J. Photogramm. Remote Sens. 2020, 164, 26–40. [Google Scholar] [CrossRef]
  54. Mao, Y.; Chen, K.; Diao, W.; Sun, X.; Lu, X.; Fu, K.; Weinmann, M. Beyond single receptive field: A receptive field fusion-and-stratification network for airborne laser scanning point cloud classification. ISPRS J. Photogramm. Remote Sens. 2022, 188, 45–61. [Google Scholar] [CrossRef]
  55. Chen, Y.; Xing, Y.; Li, X.; Gao, W. DGCN-ED: Dynamic graph convolutional networks with encoder–decoder structure and its application for airborne LiDAR point classification. Int. J. Remote Sens. 2023, 44, 3489–3506. [Google Scholar] [CrossRef]
  56. Mao, Y.Q.; Bi, H.; Li, X.; Chen, K.; Wang, Z.; Sun, X.; Fu, K. Twin deformable point convolutions for airborne laser scanning point cloud classification. ISPRS J. Photogramm. Remote Sens. 2025, 221, 78–91. [Google Scholar] [CrossRef]
  57. Wang, Z.; Chen, H.; Liu, J.; Qin, J.; Sheng, Y.; Yang, L. Multilevel intuitive attention neural network for airborne LiDAR point cloud semantic segmentation. Int. J. Appl. Earth Obs. Geoinf. 2024, 132, 104020. [Google Scholar] [CrossRef]
  58. Hu, Q.; Yang, B.; Khalid, S.; Xiao, W.; Trigoni, N.; Markham, A. Towards Semantic Segmentation of Urban-Scale 3D Point Clouds: A Dataset, Benchmarks and Challenges (CVPR’2021). In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
Figure 1. The flowchart of the proposed model. Among them, D represents domain, F represents the proposed MFE, G represents multilayer perceptron (MLP)-based classifier, C b represents base categories, and C n represents novel categories. Subscripts s, t, and t_b represent source domain, target domain, and base categories from target domain, respectively.
Figure 1. The flowchart of the proposed model. Among them, D represents domain, F represents the proposed MFE, G represents multilayer perceptron (MLP)-based classifier, C b represents base categories, and C n represents novel categories. Subscripts s, t, and t_b represent source domain, target domain, and base categories from target domain, respectively.
Remotesensing 17 02618 g001
Figure 2. Flowchart of Multi-branch Feature Extractor, where N is the number of points, B is the number of branches, K is the number of neighborhoods, and D i n represents the dimension of the input parameter.
Figure 2. Flowchart of Multi-branch Feature Extractor, where N is the number of points, B is the number of branches, K is the number of neighborhoods, and D i n represents the dimension of the input parameter.
Remotesensing 17 02618 g002
Figure 3. Illustration of NEB, where N is the number of points and K is the number of neighborhoods, D i n is the dimension of the input, and D o u t is the dimension of the embedded features.
Figure 3. Illustration of NEB, where N is the number of points and K is the number of neighborhoods, D i n is the dimension of the input, and D o u t is the dimension of the embedded features.
Remotesensing 17 02618 g003
Figure 4. Visualization of features learned from different parameter combinations using t-SNE.
Figure 4. Visualization of features learned from different parameter combinations using t-SNE.
Remotesensing 17 02618 g004
Figure 5. Pre-training subtask based on multi-branch architecture. N is the number of points.
Figure 5. Pre-training subtask based on multi-branch architecture. N is the number of points.
Remotesensing 17 02618 g005
Figure 6. The flowchart of domain adaptation through Multi-class Domain Adversarial Network (MDAN).
Figure 6. The flowchart of domain adaptation through Multi-class Domain Adversarial Network (MDAN).
Remotesensing 17 02618 g006
Figure 7. Description of the proposed MDAN.
Figure 7. Description of the proposed MDAN.
Remotesensing 17 02618 g007
Figure 8. Segmentation results of ISPRS benchmark dataset. (a) Results of our methods. (b) Ground truth.
Figure 8. Segmentation results of ISPRS benchmark dataset. (a) Results of our methods. (b) Ground truth.
Remotesensing 17 02618 g008
Figure 9. Comparison between the results of our method and ground truth.
Figure 9. Comparison between the results of our method and ground truth.
Remotesensing 17 02618 g009
Figure 10. Zoomed-in thematic map of ISPRS benchmark dataset. (Left) Results of our methods. (Right) Ground truth.
Figure 10. Zoomed-in thematic map of ISPRS benchmark dataset. (Left) Results of our methods. (Right) Ground truth.
Remotesensing 17 02618 g010
Figure 11. Flowchart of the used models in the ablation experiment of multi-branch architecture.
Figure 11. Flowchart of the used models in the ablation experiment of multi-branch architecture.
Remotesensing 17 02618 g011
Table 1. Summarization of point cloud semantic segmentation methods.
Table 1. Summarization of point cloud semantic segmentation methods.
MethodsBrief Descriptions Advantages and Limitations
MLP-based methods [12,13,14]A shared multi-layer perceptron (MLP) is adopted as the core building block. Aggregating features are complemented by symmetric functions.Simple and efficient, but underperforms when the point cloud contains complex local structures.
Convolution-based methods [15,16,17]Inspired by the success of convolution kernels in image segmentation, they are adopted for point clouds in order to extract local geometries.Mature deep learning models can be adopted, but this type of method requires converting the point cloud into a 2D/3D regular structure first by gridding/voxelization, and information loss is unavoidable during this process. Moreover, the input order of the point cloud may cause differences in the segmentation results.
Attention-based methods [18,19,20,21]Long-range dependencies are modeled by attention mechanisms, which are suitable for handling point cloud irregularities.Attention-based methods have achieved superior results, but this type of method has a high demand for computing resources, and a large amount of high-quality annotated data is required to ensure its high performance.
Table 2. Detailed information of the three datasets.
Table 2. Detailed information of the three datasets.
DomainTarget DomainSource Domain
DatasetISPRS benchmark datasetDALES datasetDublin dataset
DensityAbove 4 pts/m2Above 50 pts/m2>150 pts/m2
Parameter typeXYZ values
Intensity
Echo-related parameters
XYZ values
Echo-related parameters
XYZ Values
intensity
Echo-related parameters
CategoriesPowerlines, low_veg, imp_sur (ground), cars, fence/hedge (fences), roof (building), fac, shrub, tree (vegetation)Ground, vegetation, cars, trucks, powerlines, poles, fences, buildingBuilding, ground and vegetation
Base categoriesGround, vegetation, car, building, fence, powerlines
Novel categorieslow_veg, fac, shrub--
Table 3. Different sampling proportions in the training set of the ISPRS benchmark dataset.
Table 3. Different sampling proportions in the training set of the ISPRS benchmark dataset.
ProportionBase CategoriesNovel Categories
PowerImp_SurfCarFenceRoofTreeLow_VegFacShrub
100%546193,723461412,070152,045135,173180,58027,25047,605
10%5519,383462120715,20513,51818,05827254761
1%6193947121152113521806273477
0.1%11945131531361812848
Table 4. Comparison of segmentation results using the ISPRS benchmark dataset with other fully supervised deep learning methods. The segmentation accuracy of each category was evaluated by F1 and is expressed in percentage. The highest values of F1, OA, and averF1 are marked with bold text, and the second highest values are marked with underlines.
Table 4. Comparison of segmentation results using the ISPRS benchmark dataset with other fully supervised deep learning methods. The segmentation accuracy of each category was evaluated by F1 and is expressed in percentage. The highest values of F1, OA, and averF1 are marked with bold text, and the second highest values are marked with underlines.
MethodPowerLow_VegImp_SurfCarFenceRoofFacShrubTreeOAaverF1
PointNet++ (2017) [14]57.979.690.666.131.591.654.341.677.081.265.6
KPConv (2019) [51]63.182.391.472.525.294.460.344.981.283.768.4
DANCE-NET (2020) [52]68.481.692.877.238.693.960.247.281.483.971.2
GANet (2020) [53]75.482.091.677.844.294.461.549.682.684.573.2
GraNet (2021) [18]67.782.791.780.951.194.562.049.982.084.573.6
RFFS-NET (2022) [54]75.580.090.578.545.592.757.948.375.782.171.6
DGCN-ED (2023) [55]72.682.992.375.941.892.160.943.479.583.571.3
TDConvs (2024) [56]67.082.491.684.748.794.263.346.981.784.573.4
MIA-Net (2024) [57]65.879.589.771.126.294.063.548.182.883.371.8
Ours (10%)72.983.191.983.442.195.463.951.782.885.574.0
Ours (1%)65.281.690.366.131.694.160.143.582.683.768.4
Ours (0.1%)59.680.289.762.926.192.558.336.180.781.865.1
Ours (100%)76.383.491.883.237.695.568.452.883.985.774.7
Table 5. Experimental settings for comparison with other few sample learning methods.
Table 5. Experimental settings for comparison with other few sample learning methods.
DomainTarget DomainSource Domain
DatasetISPRS benchmark datasetSensatUrban
Parameter typeXYZ values
Intensity
Echo-related parameters
XYZ values
RGB
Number of categories912
Base categoriesGround, vegetation, building
Novel categoriesPower, low_veg, car, fence, fac, shrub
Settings of Multi-branch
Feature
Extractor
Components(1) Coordinate branch, (2) intensity branch, (3) echo-related branch
Pre-training subtask_Coordinate branch: pre-training using XYZ values of SensatUrban dataset
Domain adaptation subtask(1) Coordinate branch: domain adaptation using sampled training set with base categories
(2) Intensity branch and echo-related branch: supervised learning using sampled training set with base categories
-
Incremental learning subtaskCoordinate branch, intensity branch, and echo-related branch: incremental learning using sampled training set with base categories and novel categories-
Table 6. Comparison of segmentation results with other few sample learning methods. The segmentation accuracy of each category was evaluated by IoU. The highest values of IoU, OA, and mIoU are marked with bold text.
Table 6. Comparison of segmentation results with other few sample learning methods. The segmentation accuracy of each category was evaluated by IoU. The highest values of IoU, OA, and mIoU are marked with bold text.
MethodPowerLow_VegImp_SurfCarFenceRoofFacShrubTreeOAmIoU
Thr-MPRNet (2024) [37]0.060.5870.68.61.757.812.0817.444.744.132.4
Dai (2025) [32]0.043.984.012.01.547.215.618.259.648.137.3
Ours51.870.384.658.827.789.144.328.067.984.258.1
Table 7. Effect of different K-nearest neighboring on the NEB measured by OA, averF1, and p-values.
Table 7. Effect of different K-nearest neighboring on the NEB measured by OA, averF1, and p-values.
MethodOAp-Values (OA)averF1p-Value (averF1)
Base80.1 (±0.18)-62.9 (±0.74)-
NEB (K = 4)80.9 (±0.26)0.00965.6 (±0.72)<0.0001
NEB (K = 8)81.2 (±0.24)0.000566.7 (±0.83)0.001
NEB (K = 16)81.8 (±0.28)0.000867.1 (±0.76)0.0006
NEB (K = 32)81.7 (±0.21)0.000666.53 (±0.77)0.005
NEB (K = 64)80.5 (±0.22)0.0463.2 (±0.72)0.49
Table 8. Models with different architecture.
Table 8. Models with different architecture.
ModelDescription
M1Model consists of an MLP-based embedding layer, Point Transformer Block (PTB), and MLP-based classifier. The input data is a 5-dimensional vector that includes 3 coordinate values, intensity, and an echo-related parameter.
M2Model consists of a three-branch feature extractor that includes a coordinate branch, an intensity branch, and an echo-related branch. Each branch consists of an MLP-based embedding layer and PTB. Features from each branch were concatenated and then input into an MLP-based classifier to obtain segmentation results.
Table 9. Ablation experiment of multi-branch architecture and single-branch architecture. The segmentation effectiveness of each category was evaluated by F1 score and is expressed in percentage.
Table 9. Ablation experiment of multi-branch architecture and single-branch architecture. The segmentation effectiveness of each category was evaluated by F1 score and is expressed in percentage.
MethodPowerLow_VegImp_SurfCarFenceRoofFacShrubTreeOAaverF1p-Value
(OA)
p-Value
(averF1)
M150.2
(±5.41)
80.4
(±0.28)
91.0
(±0.25)
67.4
(±1.94)
21.6
(±5.15)
92.8
(±0.25)
56.7
(±0.48)
46.7
(±0.43)
80.5
(±0.28)
82.4
(±0.28)
65.3
(±0.82)
--
M267.1
(±4.82)
82.3
(±0.34)
91.2
(±0.23)
70.5
(±2.17)
33.6
(±3.24)
94.5
(±0.21)
60.7
(±0.52)
49.3
(±0.38)
80.8
(±0.31)
83.7
(±0.26)
69.8
(±0.78)
0.00080.0007
Table 10. Time complexity measured by running time of M1 and M2 on the testing set of ISPRS dataset.
Table 10. Time complexity measured by running time of M1 and M2 on the testing set of ISPRS dataset.
MethodMaximum (Second)Minimum (Second)Average (Second)
M128.731.930.7
M255.361.458.6
Table 11. Comparison of different domain adaptation methods. The segmentation results of each category were evaluated by F1 score and are expressed in percentage.
Table 11. Comparison of different domain adaptation methods. The segmentation results of each category were evaluated by F1 score and are expressed in percentage.
MethodPowerImp_SurfCarFenceRoofTreeOAp-Value (OA)
FT50.4 (±5.89)97.9 (±0.14)66.3 (±1.91)48.9 (±5.41)90.7 (±0.21)84.1 (±0.25)90.3 (±0.28)<0.0001
MADA57.1 (±4.76)98.3 (±0.17)78.6 (±2.41)55.2 (±4.68)97.5 (±0.12)93.7 (±0.18)95.8 (±0.24)<0.0001
MDAN (Ours)76.9 (±3.46)98.5 (±0.11)82.6 (±1.81)61.4 (±4.32)98.0 (±0.15)93.9 (±0.2)96.5 (±0.17)-
Table 12. Ablation experiments of KTM and FSM. The segmentation effectiveness of each category was evaluated by F1 and is expressed in percentage.
Table 12. Ablation experiments of KTM and FSM. The segmentation effectiveness of each category was evaluated by F1 and is expressed in percentage.
Method C b C n OAp-Value (OA)
OALow_VegFacShrubOA
Base96.5------
KTM + Fine-tuning91.7 (±0.25)74.5 (±0.37)42.8 (±0.92)41.6 (±1.27)62.7 (±0.45)82.2 (±0.34)<0.0001
FSM81.0 (±0.19)84.0 (±0.28)69.5 (±0.86)65.1 (±0.93)78.1 (±0.38)80.1 (±0.23)0.0001
Ours (KTM + FSM)91.7 (±0.25)83.2 (±0.19)64.2 (±0.73)52.1 (±0.84)73.2 (±0.26)85.5 (±0.25)-
Table 13. Comparative experiments of SALM and the proposed FSM. The segmentation accuracy of each category was evaluated by F1 score and is expressed in percentage.
Table 13. Comparative experiments of SALM and the proposed FSM. The segmentation accuracy of each category was evaluated by F1 score and is expressed in percentage.
Method C n  
Low_VegFacShrubOAp-Value (OA)
SALM80.4 (±0.81)58.3 (±1.45)56.8 (±0.96)77.1 (±0.9)-
FSM84.0 (±0.28)69.5 (±0.86)65.1 (±0.93)78.1 (±0.38)0.008
Table 14. The impact of random sampling. OA and averF1 are expressed in percentage.
Table 14. The impact of random sampling. OA and averF1 are expressed in percentage.
ProportionMaximumMinimumAverage
10%OA85.785.285.5
averF174.372.173.8
1%OA84.183.483.8
averF167.169.268.4
0.1%OA81.980.781.6
averF165.662.164.1
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yuan, J.; Ma, H.; Zhang, L.; Deng, J.; Luo, W.; Liu, K.; Cai, Z. Transfer Learning Based on Multi-Branch Architecture Feature Extractor for Airborne LiDAR Point Cloud Semantic Segmentation with Few Samples. Remote Sens. 2025, 17, 2618. https://doi.org/10.3390/rs17152618

AMA Style

Yuan J, Ma H, Zhang L, Deng J, Luo W, Liu K, Cai Z. Transfer Learning Based on Multi-Branch Architecture Feature Extractor for Airborne LiDAR Point Cloud Semantic Segmentation with Few Samples. Remote Sensing. 2025; 17(15):2618. https://doi.org/10.3390/rs17152618

Chicago/Turabian Style

Yuan, Jialin, Hongchao Ma, Liang Zhang, Jiwei Deng, Wenjun Luo, Ke Liu, and Zhan Cai. 2025. "Transfer Learning Based on Multi-Branch Architecture Feature Extractor for Airborne LiDAR Point Cloud Semantic Segmentation with Few Samples" Remote Sensing 17, no. 15: 2618. https://doi.org/10.3390/rs17152618

APA Style

Yuan, J., Ma, H., Zhang, L., Deng, J., Luo, W., Liu, K., & Cai, Z. (2025). Transfer Learning Based on Multi-Branch Architecture Feature Extractor for Airborne LiDAR Point Cloud Semantic Segmentation with Few Samples. Remote Sensing, 17(15), 2618. https://doi.org/10.3390/rs17152618

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop