Next Article in Journal
Water Vapor Correction in Measurements of Aerosol Backscatter Coefficients Using a 910 nm Vaisala CL51 Ceilometer
Previous Article in Journal
MFA-SCDNet: A Semantic Change Detection Network for Visible and Infrared Image Pairs
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

PSNet: Patch-Based Self-Attention Network for 3D Point Cloud Semantic Segmentation

1
College of Geographical Sciences, Harbin Normal University, Harbin 150025, China
2
Guangdong Urban-Rural Planning and Design Research Institute Technology Group Co., Ltd., Guangzhou 510290, China
3
Inspur Cloud Information Technology Co., Ltd., Jinan 250101, China
4
School of Remote Sensing and Information Engineering, Wuhan University, Wuhan 430079, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Remote Sens. 2025, 17(12), 2012; https://doi.org/10.3390/rs17122012
Submission received: 18 March 2025 / Revised: 15 May 2025 / Accepted: 5 June 2025 / Published: 11 June 2025

Abstract

:
LiDAR-captured 3D point clouds are widely used in self-driving cars and smart cities. Point-based semantic segmentation methods allow for more efficient use of the rich geometric information contained in 3D point clouds, so it has gradually replaced other methods as the mainstream deep learning method in 3D point cloud semantic segmentation. However, existing methods suffer from limited receptive fields and feature misalignment due to hierarchical downsampling. To address these challenges, we propose PSNet, a novel patch-based self-attention network that significantly expands the receptive field while ensuring feature alignment through a patch-aggregation paradigm. PSNet combines patch-based self-attention feature extraction with common point feature aggregation (CPFA) to implicitly model large-scale spatial relationships. The framework first divides the point cloud into overlapping patches to extract local features via multi-head self-attention, then aggregates features of common points across patches to capture long-range context. Extensive experiments on Toronto-3D and Complex Scene Point Cloud (CSPC) datasets validate PSNet’s state-of-the-art performance, achieving overall accuracies (OAs) of 98.4% and 97.2%, respectively, with significant improvements in challenging categories (e.g., +32.1% IoU for fences). Experimental results on the S3DIS dataset show that PSNet attains competitive mIoU accuracy (71.2%) while maintaining lower inference latency (7.03 s). The PSNet architecture achieves a larger receptive field coverage, which represents a significant advantage over existing methods. This work not only reveals the mechanism of patch-based self-attention for receptive field enhancement but also provides insights into attention-based 3D geometric learning and semantic segmentation architectures. Furthermore, it provides substantial references for applications in autonomous vehicle navigation and smart city infrastructure management.

Graphical Abstract

1. Introduction

Benefiting from the synergistic development of sensor technology and artificial intelligence, the 3D point cloud, as the main data format of LiDAR, is playing an increasingly significant role in the field of autonomous driving [1]. Compared with binocular cameras and RGB-D cameras, LiDAR is not easily affected by some external conditions such as weather and illumination, and it can provide a 3D point cloud containing clear ground objects and rich spatial details [2]. The latent features of the 3D point cloud can be extracted by the 3D point cloud semantic segmentation method, which can assign a specific semantic label to each point and help the machine to discriminate the region of interest (ROI). As a hot research topic for scholars in the 3D field, 3D point cloud semantic segmentation methods have been widely used in many ubiquitous receptive fields such as robotics, autonomous driving, smart cities, and augmented reality [3].
Semantic segmentation in 3D point clouds is more challenging than semantic segmentation in 2D images. First of all, the 3D point cloud is distributed in a Euclidean space, which makes the amount of data contained in the 3D point cloud much larger than that of the 2D image. Second, unlike gridded images, the distribution of the 3D point cloud is irregular. Its data density is not uniform, which easily leads to a long-tailed distribution [4]. Third, the 3D point cloud is discrete, and its structural characteristics are not clear. This makes 3D point clouds lack texture information similar to that evident in 2D images. The above characteristics will bring significant challenges to the task of semantic segmentation of point clouds. Traditional 3D point cloud semantic segmentation methods usually rely on hand-crafted features, in which it is difficult to fully express the shape and texture information contained in the point cloud [5,6]. This is insufficient in terms of speed and accuracy to meet the semantic segmentation requirements brought by massive point clouds. Fortunately, deep learning has gradually been applied to 3D vision tasks due to its powerful feature expression capabilities [2]. At present, 3D point cloud semantic segmentation methods based on deep learning mainly include projection-based methods, discretization-based methods, and point-based methods [1]. Figure 1 shows some typical 3D point cloud semantic segmentation methods based on deep learning in chronological order.
Projection-based semantic segmentation methods mainly include multi-view projection and spherical projection methods. Usually, two-dimensional image processing methods, such as convolution, are adopted to process the projected image of the point cloud, and then the two-dimensional semantic segmentation result is back-projected to the point cloud feature space. Tatarchenko et al. [7] proposed the TangentConv method, which projects the local surface around each point onto a tangent plane to obtain a tangent image, and then uses convolution to extract the features of the tangent image. Milioto et al. [8] proposed RangeNet++, which performs spherical projection on the input point cloud to obtain a range image, and then performs semantic segmentation on the range image. Projection-based methods are prone to loss of geometric features, and their performance depends on the choice of viewpoint and the angle of projection.
Discretization-based semantic segmentation methods usually convert point clouds into dense or sparse discrete representations, commonly known as voxel and sparse lattice methods. Graham et al. [9] proposed a sparse convolution operation, where the convolution operation is performed only on occupied voxels. Su et al. [10] proposed SPLATNet, which uses bilateral convolutional layers as the base module to transform point clouds into a sparse lattice space and extract features using bilateral convolutional layers. Discretization-based methods tend to have a high computational cost and memory footprint [2].
The point-based semantic segmentation method directly uses the raw point cloud as the input of the model, and completely retains the rich geometric information in the 3D point cloud [1]. It can be classified into multilayer perceptron (MLP)-based methods, convolution-based methods, recurrent neural network (RNN)-based methods, graph-based methods, and self-attention-based methods. Qi et al. [11] proposed PointNet++ to extract and integrate features over nested partitions of the input point set. Engelmann et al. [12] introduce a dilated point convolution (DPC) operation to aggregate nearby point features. Ye et al. [13] capture long-range spatial dependencies based on bidirectional recurrent neural networks. Ma et al. [14] adopted an undirected graph to capture global contextual information. Point Transformer V2 [15] introduces novel group vector attention and lightweight partition-based pooling methods to achieve better spatial alignment and more efficient sampling. LCPFormer [16] is equipped with the Local Context Propagation (LCP) module, which enhances the ability of the model to capture comprehensive and discriminative local structural information. Sufficient studies have shown that the point-based semantic segmentation method is more helpful in achieving satisfactory segmentation results. As the mainstream 3D point cloud semantic segmentation method, it is gradually replacing the projection-based semantic segmentation method and the voxel-based semantic segmentation method.
However, existing point-based 3D point cloud semantic segmentation methods still have some limitations. On the one hand, these methods need to find a certain number of neighbors. However, the comparative experiments of Engelmann et al. [12] found that when the number of nearest neighbors is fixed at 16, the receptive field of the model is still small even if many layers are stacked. The restricted receptive field makes the model unable to directly model large-scale spatial relationships. On the other hand, the downsampling process of mainstream semantic segmentation frameworks inevitably results in the loss of local information of point clouds. Moreover, the interpolation operation during the upsampling process may also cause misalignment of features. Addressing these two major challenges, the core objective of this study is to enhance the model’s receptive field and prevent feature misalignment during downsampling through a novel patch-aggregation paradigm. Therefore, this paper proposes a novel self-attention network for 3D point cloud semantic segmentation, which adopts a new paradigm, patch-based self-attention feature extraction and common point feature aggregation (CPFA), instead of stepwise downsampling and upsampling operations. It can learn local features and long-range dependencies more fully. The main contributions of the paper are briefly summarized as follows: (1) We propose a patch-based self-attention network (PSNet) to segment 3D point clouds. It uses a self-attention module to extract local feature information within patches and losslessly exploits the spatial information of the raw data. (2) We design a CPFA component. It can aggregate common point features belonging to different patches to capture long-range spatial dependencies, and eliminate feature bias. (3) Based on Toronto-3D and CSP datasets and a variety of classic point cloud semantic segmentation algorithms, extensive performance evaluations and comparisons are carried out to verify the advancement and effectiveness of the proposed method.

2. Related Works

2.1. Self-Attention Mechanism

The self-attention mechanism [17,18] is a variant of the attention mechanism [19,20,21], which is the core module of the Transformer [22,23]. It was originally proposed to solve the problem of modeling long-distance dependencies in the field of natural language processing [24]. The self-attention mechanism is based on the soft attention mechanism [25] in the key–value pair mode, and maps the input information to three different feature vector spaces of key (K), value (V), and query (Q), thereby calculating the attention coefficient and capturing contextual information. In particular, the multi-head self-attention mechanism [22] is popular due to its performance advantages. The essence of the multi-head self-attention mechanism is the parallel mapping of multiple independent attention modules, which allows the complex relationships of input information to be modeled in different feature subspaces [26]. DeepSeek [27] employs the multi-head latent attention mechanism, which integrates low-rank key–value joint compression to enhance inference efficiency. These advancements collectively demonstrate the adaptability of self-attention frameworks in modeling complex spatial and semantic relationships, laying the theoretical foundation for their extension to 3D geometric learning tasks such as point cloud segmentation.

2.2. Downsampling Method

The mainstream point cloud semantic segmentation methods introduce a downsampling process in order to expand the receptive field and improve the computational efficiency [28]. In general, in the encoder of the model, a series of downsampling methods such as random sampling [29], farthest point sampling [26,30] or learning-based sampling [31,32] are used to discard some points in the input data in order to map features layer by layer to a higher features space to gradually acquire a larger receptive field and extract higher-level abstract semantic features. In the decoder of the model, upsampling is performed by interpolation to restore the abstract features obtained in downsampling to the original size of the input data [11]. However, the downsampling process continuously loses the local information of the input data, and the upsampling process may lead to feature bias in the point cloud data. These operations will generate a negative effect on the final semantic segmentation result. The recently proposed SliceSamp [33] algorithm integrates feature slicing with depthwise separable convolutions, achieving enhanced computational efficiency while preserving information integrity. Swin Transformer [30] attempts to mitigate information loss through patch-based operations, yet it still faces challenges in global feature aggregation.

3. Methods

To address the limitations of existing methods caused by traditional sampling paradigms—such as insufficient receptive fields and feature misalignment—we propose a patch-based self-attention network (PSNet) for 3D point cloud semantic segmentation. The backbone (depth of 4) of the model consists of four key components: intersecting patch generation (IPG), position encoding (PE), self-attention feature encoding (SFE, depth of 2), and common point feature aggregation (CPFA). PSNet employs the multi-head self-attention layer of the Vision Transformer as its core feature extraction component. In this study, we introduce a novel paradigm that replaces the conventional stepwise downsampling and upsampling operations with intersecting patch-based self-attention feature extraction and common point feature aggregation. This paradigm effectively addresses challenges such as difficulty in modeling large-scale spatial relationships, information loss, and feature misalignment commonly encountered in downsampling–upsampling architectures. Figure 2 illustrates the overall architecture of PSNet for the task of 3D point cloud semantic segmentation.

3.1. Intersecting Patch Generation (IPG)

To address the limited receptive field in existing methods and reduce the computational effort in the self-attention feature extraction process, we innovatively propose an intersecting patch generation strategy. We introduce the patching mechanism [34] in the Visual Transformer model to generate multiple intersecting point cloud patches from the input of the model. The execution workflow of this component is as follows: First, a probabilistic sampling algorithm is applied to obtain a set of input points (81,920 points per batch). Subsequently, the farthest point sampling (FPS) algorithm is employed to select N s a m p l e ( N s a m p l e = 256) center points from the input point cloud. By utilizing farthest point sampling (FPS) for center point selection, representative coverage across both sparse and dense regions is ensured. For each of these centers, N n e i g h ( N n e i g h = 400) neighboring points are retrieved to form a compact point set—referred to as a point cloud patch—enabling subsequent local self-attention operations within each patch. During the patch generation process, the indices of each point in the original input point cloud are recorded for every generated patch. This facilitates the mapping of per-patch features back to the original input space in the later common point feature aggregation stage. This component replaces the conventional approach of progressively downsampling the entire input point cloud, thereby reducing computational complexity and mitigating issues of information loss and feature misalignment commonly associated with downsampling procedures. Moreover, during the intersecting patch generation phase, it is essential to ensure that every point in the input point cloud is included in at least one patch to prevent information loss. Specifically, this study ensures full coverage by setting the product of N s a m p l e and N n e i g h to be greater than the number of input points per batch. Under this configuration, certain points may belong to multiple patches—these are referred to as common points. The farthest point sampling ensures representative coverage across density variations, while fixed-size neighborhoods with learned attention weights maintain consistent feature extraction. The intersecting patch generation component in Figure 2 illustrates the process of generating intersecting patches.

3.2. Position Encoding (PE)

Compared with the algorithms of recurrent neural network series, self-attention does not easily learn the sequence of data. The position encoding component can encode the position of points inside each point cloud patch to alleviate the above problem [22]. As depicted in Figure 2, for each point cloud patch, the input to the position encoding component comprises the 3D coordinates (x, y, z) of the points within the patch, which has a dimension of ( N n e i g h , 3), and the corresponding point cloud patch features with a dimension of ( N n e i g h , C). During the position encoding process, the relative position ( p r e l , i ) of each query point ( p i ) with respect to the center point ( p c ) is firstly calculated according to Equation (1). This relative position is defined as the coordinate offset between the 3D coordinate ( p c o o r , i ) of query point ( p i ) and the 3D coordinate ( p c o o r , c ) of the center point ( p c ). Concurrently, the Euclidean distance from each point to the center point is calculated based on Equation (2). Subsequently, the fused position vector for each point is obtained by adding the absolute coordinates, the relative position vector, and the Euclidean distance of p i and p c . Finally, as illustrated in Equation (4), a linear layer is utilized to encode this fused position vector into the same dimension as the input patch features. The resulting positional encoding vector is then added to the feature vector ( p f , i ) of each point, thereby yielding the output f p o s , i of the position encoding component.
p r e l ,   i = p c o o r ,   i p c o o r ,   c
d i s t i = | |   p c o o r ,   i p c o o r ,   c   | |
r i = p c o o r ,   i p r e l ,   i d i s t i
f p o s , i = L i n e a r r i + p f ,   i
where p r e l , i denotes the relative coordinates of each query point ( p i ) and center point ( p c ). p c o o r , i and p c o o r , c denote the coordinates of p i and p c , respectively. | | | | denotes the calculation of the Euclidean distance d i s t i of p i , and p c . r i denotes the position vector of p i . denotes the sum operation. f p o s , i denotes the position encoding output of p i . C o n v denotes the convolution operation. p f , i denotes the feature vector of p i .

3.3. Self-Attention Feature Encoding (SFE)

Figure 2 illustrates the internal structure of the self-attention feature encoding component, which is used to encode the internal features of each point cloud patch. The component receives as input the features f p o s of each point fused with position information. It first updates the features of each point inside the point cloud patch through a normalization layer and a multi-head self-attention layer. The output f a g g of the feature encoding component is then obtained through a normalization layer and a fully connected layer, while a residual connection is introduced during the feature encoding process to fuse the input features with the intermediate features generated during the feature encoding process.
Figure 3 presents the detailed architecture of the multi-head self-attention module within the self-attention feature encoding component. In this study, the self-attention feature encoding component is designed with a depth of 2 layers, and the number of heads in the multi-head self-attention module is set to 8. The module performs multiple independent self-attention computations on the input feature tensor ( f p o s ). Subsequently, the results from these parallel computations are concatenated. Finally, a linear layer is applied to project the concatenated features back to the same dimension as the input, resulting in the output tensor of shape ( N n e i g h , C), which serves as the final output of the self-attention feature encoding component.

3.4. Common Point Feature Aggregation (CPFA)

The CPFA component is responsible for aggregating the attention features of each point cloud patch to update per-point features, thereby obtaining comprehensive features for all points in the current batch of input point cloud data. This aggregation process consists of three key steps: (1) Feature summation of common points across different patches based on point indices obtained during the intersecting patch generation phase. As illustrated in Figure 4, f1, f2, and f3 represent three distinct patch features from the same input batch, where f1 + f2 + f3 denotes the elementwise addition of features from different patches according to corresponding point indices, resulting in aggregated features for common points. (2) Projection of the aggregated features back to the original point cloud using point indices to overwrite the initial features, yielding updated features for each point. (3) Processing the updated features through an MLP module to further enhance the fused features of the current batch. This component effectively combines features of common points across different patches while preserving the intrinsic characteristics of point sets within individual patches. It achieves long-range feature aggregation for the current point cloud batch while simultaneously learning local spatial representations of point sets within different patches. Moreover, it can also further enhance robustness by aggregating features across density-varying regions.

3.5. Loss Function

In the 3D point cloud semantic segmentation task, multicategory cross-entropy loss is often used to measure the learning effectiveness of the model. We use a weighted cross-entropy loss (WCEL) function to address the potential impact of the category imbalance problem on the experimental results. To address the phenomenon of unbalanced samples of each category in the dataset, we introduce a category weight prior in the multicategory cross-entropy loss function. The specific calculation formula is as follows.
r c = n c c = 1 M n c
w c = 1 r c
y i c = 1     i f   l a b e l i = c 0     e l s e                
W C E L = 1 N i = 1 N c = 1 M w c y i c log p i c
where r c denotes the proportion of samples in category c. n c denotes the number of samples in category c, and M denotes the number of categories. w c denotes the weight of samples in category c. y i c is an indicator function that takes 1 if the ground truth of sample i is c, and 0 otherwise. N denotes the number of samples, and p i c denotes the probability that sample i is predicted to be in category c.

4. Results

4.1. Datasets

Three open-access datasets were selected for algorithm validation experiments: the Toronto-3D dataset [35], the Complex Scene Point Cloud (CSPC) dataset [36], and the Stanford Large-Scale 3D Indoor Spaces (S3DIS) dataset.
Toronto-3D is a large urban outdoor point cloud dataset collected by the Mobile Laser System (MLS), which includes four areas, L001, L002, L003, and L004, with a total of 78.3 million points. It is divided into eight categories, including road, road markings, natural, building, utility line, poles, car, and fence. It has 10 attributes, including 3D location, RGB color, intensity, GPS time, scan angle level, and category.
CSPC is a LiDAR dataset for semantic segmentation of large-scale complex outdoor scenes, including Scene1, Scene2, Scene3, Scene4, and Scene5, with a total of about 68 million points. It is divided into six categories, including ground, car, building, vegetation, bridge, and utility pole. The points in the dataset have six attributes, which represent 3D position and RGB color. The category imbalance problem in these two datasets poses a challenge to the semantic segmentation approach.
S3DIS is a widely adopted benchmark dataset designed to facilitate research in indoor point cloud semantic segmentation. It comprises six large-scale indoor environments, labeled as Area 1 through Area 6, encompassing a total of 271 rooms across 11 scene types, including offices, conference rooms, corridors, and other common indoor spaces. The dataset contains over 695 million annotated points, each represented by its 3D spatial coordinates and corresponding RGB color values. S3DIS defines a comprehensive set of 13 semantic categories, which include both architectural elements—such as ceilings, walls, and floors—and movable objects—such as tables, chairs, sofas, and clutter. These semantic annotations enable fine-grained understanding and analysis of complex indoor scenes at the point level. Due to its high object density, significant variability in spatial layouts, and extensive annotation coverage, S3DIS offers a robust and realistic benchmark for evaluating semantic segmentation algorithms on large-scale indoor point cloud data. It has become an essential resource for advancing research in 3D scene understanding, particularly in applications such as indoor navigation, augmented reality, and intelligent environment perception.

4.2. Implementation Details

In the experiments on the Toronto-3D dataset, three regions, L001, L003, and L004, were used for model training, and L002 was used for testing. We compared the experimental results of six semantic segmentation methods, PointNet++ [11], DGCNN [37], KPFCNN [38], MS-PCNN [14], TGNet [39], and MS-TGNet [35], from the paper on the Toronto-3D dataset. In the experiments on the CSPC dataset, Scene1, Scene3, and Scene4 were used for training, while Scene2 and Scene5 were used for testing. We compared the experimental results of SnapNet [40], PointNet++ [11], 3D CNN [41], DeepNet [5], and KPConv [38], the five semantic segmentation methods in the CSPC dataset paper. In addition, comparative experiments were conducted in the above two datasets using representative RandLA-Net [29] and Point Transformer [42] methods. In the experiments on the indoor scene dataset S3DIS, Areas 1, 2, 3, 4, and 6 were used for training, and Area 5 was reserved for testing. We compare the point cloud semantic segmentation performance in terms of mAcc and mIoU, model complexity in terms of the number of parameters (Params), and average wall-time latency of several recent state-of-the-art (SOTA) methods, including KPConv [38], PAConv [43], Point Transformer [42], Fast Point Transformer [44], Stratified Transformer [45], and PointNeXt-XL [46]. In this study, the model was trained on preprocessed point clouds that underwent voxel grid downsampling. Following established practices in the existing studies, we carefully selected grid resolutions of 0.01m for Toronto-3D, 0.03m for CSPC, and 0.04m for S3DIS to optimally balance point cloud density with computational efficiency.
For the model parameter setting, the number of points N in each input batch was set to 81,920. The number of center points ( N s a m p l e ), selected using the farthest point sampling algorithm, was set to 256. The number of neighboring points ( N n e i g h ) for each centroid was set to 400. A critical implementation requirement is that the product of the number of center points and neighboring points must be greater than the number of points (N = 81,920) per input batch, which guarantees sufficient patch overlap for effective feature aggregation via the CPFA module. The initial learning rate was set to 0.01.

4.3. Evaluation Metrics

To verify the effectiveness of our method, three representative evaluation metrics are used to evaluate the performance of our method and other methods, including the Intersection over Union (IoU), the Mean Intersection over Union (mIoU), and the overall accuracy (OA) of each category. mIoU represents the average value of the IoU for all categories. The calculation methods of the above evaluation metrics are shown below.
O A = i = 1 C m a t i , i i = 1 C j = 1 C m a t i , j
I o U c = m a t c , c i = 1 C m a t i , c + j = 1 C m a t c , j m a t c , c , c = 1,2 , C
m I o U = c = 1 C I o U c C
where C denotes the number of categories. mat denotes the confusion matrix used for evaluation, whose rows denote the true labels and whose columns denote the predicted results. m a t i , j denotes the number of points in category i that are predicted to be in category j.

4.4. Performance Comparison

Table 1 shows the semantic segmentation results of different methods on the L002 region of the Toronto-3D dataset. The results demonstrate our method’s superior performance, achieving a SOTA performance with 98.4% OA and 88.9% mIoU, representing significant improvements of 1.6% and 9.0% over Point Transformer, respectively. Particularly noteworthy are the dramatic gains in challenging categories: In the road marking (75.7% IoU) category, compared with Point Transformer (64.6%), PSNet achieved an IoU increase of +11.1%. This stems from our patch-based attention mechanism’s ability to preserve fine and small structures that typically get lost during conventional downsampling, underscoring how traditional approaches struggle with such fine-scale elements. For the fence (69.0% IoU, +32.1%) category, this remarkable enhancement demonstrates our method’s proficiency in handling elongated, discontinuous objects. The common point feature aggregation (CPFA) mechanism effectively connects spatially separated fence segments that would be processed independently in other frameworks.
Unlike methods that achieve strong performance on certain object categories at the expense of others (e.g., RandLA-Net, which attains a high IoU of 96.6% on the road category but performs poorly on the fence category, with only 29.5%), our approach consistently improves performance across all object types. The significant 9% gain in mIoU over Point Transformer can be attributed to our method’s superior preservation of local structures through intersecting patch generation and the effective prevention of feature dilution via CPFA. These results collectively validate our method’s ability to overcome three key limitations of existing approaches, information loss during downsampling, restricted receptive fields, and feature misalignment, while demonstrating robust performance across diverse urban objects in complex outdoor scenarios.
The experimental results on the CSPC Scene2 dataset (Table 2) demonstrate that our method achieves superior performance, with PSNet attaining an OA of 97.2% and an mIoU of 80.3%. These results represent substantial improvements of 1.8% in OA and 9.3% in mIoU over Point Transformer, a SOTA model. One particularly compelling case underscores the architectural advantages of our approach. On the pole category, which presents a challenging class due to its slender and vertically oriented structure, our method achieves an IoU of 44.5%, surpassing Point Transformer by a significant margin of 22.4% (Point Transformer: 22.1%). This improvement can be attributed to the synergistic combination of our position encoding scheme and patch-based processing strategy. Specifically, the explicit encoding of relative heights and geometric characteristics unique to thin structures enables more accurate localization and classification, while the patch-based attention mechanism preserves fine-grained spatial details that are often lost during conventional downsampling procedures. This demonstrates the superior capability of our architecture in modeling complex, geometry-sensitive object categories. Notably, the CPFA module’s demonstrated capability across overlapping patches suggests strong potential for instance segmentation tasks, where precise object boundary preservation is paramount. Furthermore, our quantitative results on slender pole detection indicate particular promise for challenging object detection scenarios, as the architecture inherently addresses key limitations in current detection frameworks regarding elongated structures. This demonstrates the superior capability of our architecture in modeling complex, geometry-sensitive object categories, while simultaneously revealing its adaptability to broader 3D understanding tasks beyond semantic segmentation.
The experimental results presented in Table 3 demonstrate that our proposed method achieves advanced performance on the CSPC Scene5 dataset, with significant improvements in both overall accuracy (OA) and mean Intersection-over-Union (mIoU) metrics. The superior performance (OA of 96.2%, mIoU of 74.6%) can be primarily attributed to our novel patch-based point cloud processing paradigm, which fundamentally differs from conventional multi-scale downsampling approaches. By replacing the conventional hierarchical downsampling strategy with our patch-based input generation, we effectively eliminate the feature distortion that typically occurs during repeated downsampling operations, thereby preserving more discriminative geometric information throughout the network. Notably, significant improvements are observed in challenging object categories, particularly those involving fine structures and small objects. For instance, our method achieves 46.5% mIoU for poles, a 6.0% improvement over Point Transformer, demonstrating enhanced capability in capturing slender and sparse structures. Similarly, we obtain 69.6% mIoU for cars, which represents a 6.4% gain over KPConv, indicating superior performance in segmenting compact yet discriminative regions. A potential explanation for these improvements may lie in the patch-based approach’s ability to preserve the resolution and geometric integrity of dense objects and slender structures, which would otherwise be degraded through conventional downsampling operations. At the same time, the model maintains strong performance on large-scale semantic categories such as the ground (93.7%) and buildings (94.6%), highlighting its effectiveness across diverse geometric scales.
The quantitative experimental results of the Toronto-3D and CSPC datasets show that our patching-based self-attention method has better prediction results for all types of features in complex scenes than other methods, and the IoU of some categories is significantly improved compared with existing methods. By abandoning the upsampling and downsampling operations and adopting the new patching-aggregation paradigm, the model can learn the local and long-range details more comprehensively and accurately, which further improves the prediction ability of the model for various ground objects in complex scenes.
To further evaluate the generalizability and scalability of our proposed method in indoor scenarios, we conducted comprehensive experiments on the S3DIS Area5 dataset, comparing PSNet with several SOTA approaches. As shown in Table 4, our method achieves competitive performance, with 71.2% mIoU and 77.8% mAcc, while maintaining superior efficiency—requiring only 21.87M parameters and an average latency of 7.03 s. This represents a good balance between accuracy and computational cost. Quantitatively, compared to the SOTA model PointNeXt-XL, PSNet improves inference speed by a factor of 5.43, albeit with a 5.2% drop in mAcc and a 3.7% drop in mIoU. In contrast to the fastest method, Fast Point Transformer (1.13s of inference latency), PSNet improves the mAcc by 0.5% and the mIoU by 1.1%, but at the cost of a 6.22× increase in inference time. Specifically, PSNet replaces traditional multi-stage downsampling with a patch-based feature extraction paradigm, significantly reducing model complexity compared to PointNeXt-XL. Furthermore, the incorporation of a common point feature aggregation component enhances local topological modeling, enabling PSNet to surpass conventional convolution-based methods such as KPConv (+4.1% in mIoU) and PAConv (+4.6% in mIoU), while preserving its lightweight nature. Nonetheless, PSNet exhibits limitations in capturing the global context compared to attention-based models such as Stratified Transformer (78.1% mAcc and 72.0% mIoU) and PointNeXt-XL. The patch-based local interaction mechanism introduces noticeable information decay when modeling large-scale semantic relationships, which constrains segmentation accuracy on complex indoor scene datasets like S3DIS. Therefore, future work should focus on developing lightweight global modeling strategies to overcome the current accuracy bottleneck. Overall, the proposed method demonstrates competitive accuracy and efficiency on the S3DIS dataset. Although the performance gains are not as pronounced as those observed in outdoor scenarios, the method still exhibits strong scalability and effectiveness in indoor point cloud scenes.

4.5. Result Visualization

To compare the prediction results of the models more visually, the prediction results of the classical Point Transformer, RandLA-Net, and our method are visualized on the Toronto-3D and CSPC datasets. Figure 5 shows the prediction results of each method on the ROI of the Toronto-3D dataset. The three rows from top to bottom represent the three test examples in turn, and each column from left to right represents the raw point cloud in natural color (RGB), the label of the point cloud, and the prediction results of our method, Point Transformer, and RandLA-Net, in turn.
As shown in Example 1 (Figure 5), the prediction results of Point Transformer and RandLA-Net methods are not satisfactory for some of the dense road markings, and they can only predict some of the road markings in the highlighted part circled in the figure. Comparing the RGB point clouds, we can know that the misclassification of roads and road markings mainly exists in the boundary areas of the two classes. In contrast, our method has a smoother and more complete prediction of road markings. From the prediction results of Example 2, we can find that for a long strip of fence spanning a large spatial extent, each method misclassifies part of its area into the case of natural (trees and shrubs) or roads. For the highlighted part, the results obtained by our method are closer to the ground truth than the other methods. From the prediction results of Example 3, it can be observed that our method accurately identifies all the highlighted natural areas. In contrast, the Point Transformer and RandLA-Net methods exhibit significant misclassification between natural areas and fences. Analysis of RGB point clouds reveals that the misclassification phenomenon of other methods mainly occurs for two ground objects with close spatial relationships in the real world. This is because it is difficult for the model to learn the global features of large-scale ground objects from a macroscopic view under the restricted receptive field. The feature learning process is susceptible to interference from other ground objects in close proximity. The visualization results of each method on the Toronto-3D dataset show that the prediction results of the Point Transformer and RandLA-Net methods are misclassified in some regions. Compared with them, our method has fewer misclassifications, more complete and accurate prediction results, and better ability to capture a large range of spatial relationships.
Figure 6 shows the prediction results of our method with Point Transformer and RandLA-Net on CSPC. In the prediction result of Example 1, the prediction result of the RandLA-Net method incorrectly predicts the whole bridge area into building and vegetation categories. In contrast, the prediction results of our method and Point Transformer only misclassify the bridge piers into building categories. Comparing the RGB point clouds, we can find that the color and shape features of the bridge piers are too similar to those of the buildings, resulting in poor feature differentiation, so there may be misclassification due to inaccurate localization of edge regions. In Example 2, Point Transformer and our method achieve good predictions on the highlighted car region, while the RandLA-Net method predicts the whole region as a ground class. These two examples above demonstrate the advantages of self-attention models over convolutional models for feature extraction. For the prediction results of Example 1 and Example 3, the Point Transformer and RandLA-Net methods easily predict tall utility poles with a similar shape to vegetation as vegetation. In contrast, our method correctly and completely predicts such long utility poles even if they are labeled as buildings in the ground truth labels. Visual comparison of RGB point clouds can confirm that these ground objects are more appropriately classified as utility poles. In conclusion, the visual comparison of prediction results on CSPC shows that the prediction results of our method are closer to the ground truth than Point Transformer and RandLA-Net methods. It is experimentally demonstrated that our method can capture long-range information by further expanding the receptive field and has more accurate prediction results on large-scale ground objects. It can significantly improve the model performance, and some of its prediction results are even better than the ground truth in terms of reasonableness.
The expanded receptive field enables PSNet to capture long-range context, as evidenced by the coverage in Figure 7. Figure 7 shows the range of receptive fields in different methods with multiple ROI examples. Each row represents one example, and each column from left to right represents the raw point cloud, the label of the point cloud, the receptive field of our method, and the receptive field of the Point Transformer. The red area indicates the range of the receptive field, and the blue area indicates the background. Unlike conventional architectures that progressively enlarge the receptive field through repeated downsampling, this study replaces this mechanism with a patch-based strategy where the receptive field size is determined by the number of neighboring points within each patch. This design ensures that PSNet’s receptive field remains consistent across all feature extraction layers, independent of network depth. Comparing the receptive field distribution of multiple examples, it is obvious that our method establishes larger receptive fields of the neural network. Comparing the receptive field of our method with the ground truth labels, it can be seen that the receptive field of our method covers a wide range of different classes of ground objects in the surrounding area, which allows the model to capture a larger range of spatial relationships and thus achieve a scene-level understanding of the input point cloud.

4.6. Ablation Study

To investigate the impact of different patch configurations on model performance, we conduct an ablation study on the number of center points/patches ( N s a m p l e ) and neighbor points ( N n e i g h ) in the patch partitioning strategy. This study has conducted a detailed comparison of the different combinations of N s a m p l e = [128, 256, 512] and N n e i g h = [200, 400, 800]. The results (Table 5) reveal several key observations regarding patch overlap and receptive field size. The ablation study on patch partitioning strategies reveals a critical trade-off between receptive field size and patch overlap. When the number of center points ( N s a m p l e ) and neighbors ( N n e i g h ) are both too small (e.g., 128/200), the model suffers from insufficient point coverage and limited contextual understanding, resulting in poor performance (63.5% mIoU). Expanding the receptive field by increasing N n e i g h (e.g., 128/800) significantly improves accuracy (85.9% mIoU) by enhancing local geometric modeling. The optimal balance is achieved with moderate settings (256/400, 88.9% mIoU), where patch coverage slightly exceeds the input point count to ensure comprehensive sampling without excessive redundancy. However, overly dense partitioning (e.g., 512/800) leads to performance degradation (82.6% mIoU) due to feature overfitting and computational redundancy from extreme patch overlap. These findings demonstrate that effective point cloud segmentation requires careful calibration of patch density to maximize receptive field benefits while minimizing overlap-induced inefficiencies.
To further evaluate the effectiveness and contributions of each proposed component, we conducted an ablation study. This study compares different combinations of our key design elements: intersecting patch generation (IPG), position encoding (PE), self-attention feature encoding (SFE), and common point feature aggregation (CPFA). As shown in Table 6, compared with the baseline Point Transformer (ID 1, mIoU = 79.9%), our full model (ID 4) achieved the highest performance (88.9% mIoU) by integrating all components. The intersecting patch generation (IPG) and self-attention feature encoding (SFE) combination establishes fundamental improvements (+5.7% mIoU) through adaptive point cloud partitioning and long-range dependency capture within patches. Position encoding (PE) further enhances geometric awareness (+1.2%) by explicitly modeling spatial relationships, particularly benefiting complex structural regions. The critical 2.1% improvement from common point feature aggregation (CPFA) highlights the importance of resolving feature conflicts in overlapping patches through scatter-reduce-based consolidation. These components collectively address four key challenges: IPG ensures comprehensive patch coverage, SFE enables contextual modeling, PE strengthens geometric sensitivity, and CPFA maintains feature consistency across patches. While the current IPG and PE design enhances geometric sensitivity, the fixed encoding strategy may struggle with highly irregular structures. In the future, we will consider designing adaptive patch partitioning and position encoding strategies to further enhance the generalization capability of this component.

5. Discussion

The experimental results demonstrate that our proposed patch-based self-attention network (PSNet) achieves advanced performance across multiple large-scale 3D point cloud datasets, including urban outdoor (Toronto-3D, CSPC) and indoor (S3DIS) environments. The key advantage of PSNet lies in its ability to address three fundamental limitations of existing point cloud segmentation methods: (1) restricted receptive fields due to hierarchical downsampling, (2) progressive information loss during feature propagation, and (3) feature misalignment caused by inconsistent local aggregation.
On the Toronto-3D dataset, PSNet outperforms strong baselines such as Point Transformer and RandLA-Net by significant margins (e.g., +9.0% mIoU and +1.6% OA over Point Transformer). The most notable improvements occur in geometrically challenging categories such as road markings (+11.1% IoU) and fences (+32.1% IoU), where conventional methods struggle due to their reliance on fixed-grid downsampling or limited receptive fields. This underscores the effectiveness of our intersecting patch generation (IPG) strategy, which ensures comprehensive spatial coverage while preserving fine-grained structures. Similarly, on the CSPC dataset, PSNet achieves a 22.4% IoU improvement for utility poles, demonstrating its superiority in modeling slender and sparse objects—a task where local geometric awareness and long-range context integration are critical. The success of PSNet can be attributed to two core innovations: Patch-based self-attention with CPFA: By replacing traditional downsampling with patchwise feature extraction and common point aggregation, PSNet mitigates information loss and maintains feature consistency across overlapping regions. The ablation study (Table 6) confirms that CPFA contributes a 2.1% mIoU gain by resolving feature conflicts in patch intersections. Expanded receptive fields: As visualized in Figure 7, PSNet’s patch partitioning enables scene-level context modeling, reducing misclassification in spatially extended objects (e.g., fences, bridges) that require long-range dependency capture.
While PSNet excels in outdoor scenes, its performance on fine-grained indoor objects (e.g., furniture in S3DIS) is slightly inferior to specialized architectures like PointNeXt-XL. This suggests that static patch sizing may not optimally adapt to highly variable indoor object scales. Future work could explore dynamic patch partitioning and hierarchical patch fusion to enhance multi-scale reasoning. Additionally, integrating lightweight global attention mechanisms could further bridge the gap between local detail preservation and scene-wide coherence.

6. Conclusions

In this work, we present PSNet, a novel patch-based self-attention network for large-scale 3D point cloud semantic segmentation. Departing from conventional downsampling paradigms, this study introduces a novel point cloud patch-aggregation framework to address the challenges of limited receptive fields and feature misalignment. Our approach leverages intersecting patch generation (IPG) to enable lossless partitioning of input point clouds into spatially overlapping patches, ensuring comprehensive coverage while preserving fine-grained structural details. To enhance geometric sensitivity, we incorporate explicit position encoding (PE) that captures spatial relationships, essential for accurately segmenting complex urban objects such as poles and road markings. Within each patch, local features are extracted through self-attention feature encoding (SFE), while common point feature aggregation (CPFA) synergistically integrates localized attention with long-range contextual modeling, enabling robust feature consolidation and preventing feature misalignment.
Extensive experiments on three large-scale benchmarks (Toronto-3D, CSPC Scene2, and S3DIS) demonstrate that PSNet achieves advanced performance, setting new benchmarks with 88.9% mIoU on Toronto-3D and 80.3% mIoU on CSPC, surpassing prior advanced models by up to 9.3% in mIoU. Qualitative visual comparisons further highlight its superior robustness in segmenting challenging object categories (such as fences and utility poles) with minimal misclassification. In terms of computational efficiency, PSNet strikes an effective balance between accuracy and speed. Compared to global attention-based methods, PSNet achieves competitive accuracy (1.1% higher mIoU than Fast Point Transformer) while offering significantly faster inference latency (5.43× faster than PointNeXt-XL).
This work effectively bridges the gap between local geometric precision and long-range contextual modeling, providing a scalable and efficient solution for real-world applications including autonomous driving, urban mapping, and indoor navigation. Future research will focus on adaptive patch optimization, enhanced position encoding strategies, and cross-dataset generalization to further advance large-scale and fine-grained 3D scene understanding. Additionally, future work could explore temporal encoding schemes based on LSTM or Transformer networks for processing time-sequential point clouds, enabling dynamic object tracking and trajectory prediction in applications such as autonomous driving.

Author Contributions

Conceptualization, H.Y. and M.W.; methodology, H.Y. and Y.L.; software, Y.L. and M.W.; validation, H.Y., Y.L. and M.W.; formal analysis, H.Y. and Y.L.; investigation, Y.L.; resources, Y.L.; data curation, M.W.; writing—original draft preparation, H.Y. and Y.L.; writing—review and editing, H.Y. and M.W.; visualization, M.W.; supervision, M.W.; project administration, H.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Shandong Province Technology Innovation Guidance Plan (Central Guidance for Local Scientific and Technological Development Project: Research and Industrialization of Artificial Intelligence Large Model Service Platform), grant number [YDZX2024088].

Data Availability Statement

All datasets utilized in this work are publicly accessible. The Toronto-3D LiDAR point cloud dataset is available at https://1drv.ms/u/s!Amlc6yZnF87psX6hKS8VOQllVvj4?e=yWhrYX, accessed on 1 June 2025. The CSPC (Crowd-Sourced Point Cloud) dataset is available under the data sharing provisions specified in its foundational publication (IEEE Access, 2020, DOI: 10.1109/ACCESS.2020.2992612). The S3DIS dataset is available at https://docs.google.com/forms/d/e/1FAIpQLScDimvNMCGhy_rmBA2gHfDu3naktRm6A8BPwAWWDv-Uhm6Shw/viewform?c=0&w=l, accessed on 1 June 2025.

Acknowledgments

We gratefully acknowledge the open-access Toronto-3D and CSPC datasets, whose availability significantly advanced this research. We also acknowledge the support from the Central Guidance for Local Scientific and Technological Development Project.

Conflicts of Interest

Author Yaru Liu was employed by the company Guangdong Urban-Rural Planning and Design Research Institute Technology Group Co., Ltd. Author Ming Wang was employed by the company Inspur Cloud Information Technology Co., Ltd. The remaining author declares that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Guo, Y.; Wang, H.; Hu, Q.; Liu, H.; Liu, L.; Bennamoun, M. Deep Learning for 3D Point Clouds: A Survey. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 4338–4364. [Google Scholar] [CrossRef] [PubMed]
  2. Xie, Y.; Tian, J.; Zhu, X.X. Linking Points With Labels in 3D: A Review of Point Cloud Semantic Segmentation. IEEE Geosci. Remote Sens. Mag. 2020, 8, 38–59. [Google Scholar] [CrossRef]
  3. Qiu, S.; Anwar, S.; Barnes, N. PU-Transformer: Point Cloud Upsampling Transformer. In Proceedings of the Asian Conference on Computer Vision, Macao, China, 4–8 December 2022; pp. 2475–2493. [Google Scholar]
  4. Zhang, Y.; Zhou, Z.; David, P.; Yue, X.; Xi, Z.; Gong, B.; Foroosh, H. PolarNet: An Improved Grid Representation for Online LiDAR Point Clouds Semantic Segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 9601–9610. [Google Scholar]
  5. Hackel, T.; Savinov, N.; Ladicky, L.; Wegner, J.D.; Schindler, K.; Pollefeys, M. Semantic3D.Net: A New Large-Scale Point Cloud Classification Benchmark. arXiv 2017, arXiv:1704.03847. [Google Scholar] [CrossRef]
  6. Landrieu, L.; Raguet, H.; Vallet, B.; Mallet, C.; Weinmann, M. A Structured Regularization Framework for Spatially Smoothing Semantic Labelings of 3D Point Clouds. ISPRS J. Photogramm. Remote Sens. 2017, 132, 102–118. [Google Scholar] [CrossRef]
  7. Tatarchenko, M.; Park, J.; Koltun, V.; Zhou, Q.-Y. Tangent Convolutions for Dense Prediction in 3D. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3887–3896. [Google Scholar]
  8. Milioto, A.; Vizzo, I.; Behley, J.; Stachniss, C. RangeNet ++: Fast and Accurate LiDAR Semantic Segmentation. In Proceedings of the 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Macao, China, 3–8 November 2019; pp. 4213–4220. [Google Scholar] [CrossRef]
  9. Graham, B.; Engelcke, M.; van der Maaten, L. 3D Semantic Segmentation With Submanifold Sparse Convolutional Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 9224–9232. [Google Scholar]
  10. Su, H.; Jampani, V.; Sun, D.; Maji, S.; Kalogerakis, E.; Yang, M.-H.; Kautz, J. SPLATNet: Sparse Lattice Networks for Point Cloud Processing. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 2530–2539. [Google Scholar]
  11. Qi, C.R.; Yi, L.; Su, H.; Guibas, L.J. PointNet++: Deep Hierarchical Feature Learning on Point Sets in a Metric Space. In Advances in Neural Information Processing Systems; Curran Associates, Inc.: San Francisco, CA, USA, 2017; Volume 30. [Google Scholar]
  12. Engelmann, F.; Kontogianni, T.; Leibe, B. Dilated Point Convolutions: On the Receptive Field Size of Point Convolutions on 3D Point Clouds. In Proceedings of the 2020 IEEE International Conference on Robotics and Automation (ICRA), Paris, France, 31 May–31 August 2020; pp. 9463–9469. [Google Scholar] [CrossRef]
  13. Ye, X.; Li, J.; Huang, H.; Du, L.; Zhang, X. 3D Recurrent Neural Networks with Context Fusion for Point Cloud Semantic Segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 403–417. [Google Scholar]
  14. Ma, Y.; Guo, Y.; Liu, H.; Lei, Y.; Wen, G. Global Context Reasoning for Semantic Segmentation of 3D Point Clouds. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, 2020, Snowmass Village, CO, USA, 1–5 March 2020; pp. 2931–2940. [Google Scholar]
  15. Wu, X.; Lao, Y.; Jiang, L.; Liu, X.; Zhao, H. Point Transformer V2: Grouped Vector Attention and Partition-Based Pooling. Adv. Neural Inf. Process. Syst. 2022, 35, 33330–33342. [Google Scholar]
  16. Huang, Z.; Zhao, Z.; Li, B.; Han, J. LCPFormer: Towards Effective 3D Point Cloud Analysis via Local Context Propagation in Transformers. IEEE Trans. Circuits Syst. Video Technol. 2023, 33, 4985–4996. [Google Scholar] [CrossRef]
  17. Wang, B.; Liu, K.; Zhao, J. Inner Attention Based Recurrent Neural Networks for Answer Selection. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers); Erk, K., Smith, N.A., Eds.; Association for Computational Linguistics: Berlin, Germany, 2016; pp. 1288–1297. [Google Scholar] [CrossRef]
  18. Lin, Z.; Feng, M.; dos Santos, C.N.; Yu, M.; Xiang, B.; Zhou, B.; Bengio, Y. A Structured Self-Attentive Sentence Embedding. International Conference on Learning Representations. arXiv 2017, arXiv:1703.03130. [Google Scholar] [CrossRef]
  19. Niu, Z. A Review on the Attention Mechanism of Deep Learning. Neurocomputing 2021, 452, 48–62. [Google Scholar] [CrossRef]
  20. Luong, M.-T.; Pham, H.; Manning, C.D. Effective Approaches to Attention-Based Neural Machine Translation. arXiv 2015, arXiv:1508.04025. [Google Scholar] [CrossRef]
  21. Miller, A.; Fisch, A.; Dodge, J.; Karimi, A.-H.; Bordes, A.; Weston, J. Key-Value Memory Networks for Directly Reading Documents. arXiv 2016, arXiv:1606.03126. [Google Scholar] [CrossRef]
  22. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention Is All You Need. In Advances in Neural Information Processing Systems; Curran Associates, Inc.: San Francisco, CA, USA, 2017; Volume 30. [Google Scholar]
  23. Ruan, L. Survey: Transformer Based Video-Language Pre-Training. AI Open 2022, 3, 1–13. [Google Scholar] [CrossRef]
  24. Daniluk, M.; Rocktäschel, T.; Welbl, J.; Riedel, S. Frustratingly Short Attention Spans in Neural Language Modeling. arXiv 2017, arXiv:1702.04521. [Google Scholar] [CrossRef]
  25. Gulcehre, C.; Chandar, S.; Cho, K.; Bengio, Y. Dynamic Neural Turing Machine with Soft and Hard Addressing Schemes. arXiv 2017, arXiv:1607.00036. [Google Scholar] [CrossRef]
  26. Li, J.; Tu, Z.; Yang, B.; Lyu, M.R.; Zhang, T. Multi-Head Attention with Disagreement Regularization. arXiv 2018, arXiv:1810.10183. [Google Scholar] [CrossRef]
  27. DeepSeek-AI; Liu, A.; Feng, B.; Xue, B.; Wang, B.; Wu, B.; Lu, C.; Zhao, C.; Deng, C.; Zhang, C.; et al. DeepSeek-V3 Technical Report. arXiv 2025, arXiv:2412.19437. [Google Scholar] [CrossRef]
  28. Han, X.; Dong, Z.; Yang, B. A Point-Based Deep Learning Network for Semantic Segmentation of MLS Point Clouds. ISPRS J. Photogramm. Remote Sens. 2021, 175, 199–214. [Google Scholar] [CrossRef]
  29. Hu, Q.; Yang, B.; Xie, L.; Rosa, S.; Guo, Y.; Wang, Z.; Trigoni, N.; Markham, A. RandLA-Net: Efficient Semantic Segmentation of Large-Scale Point Clouds. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 11108–11117. [Google Scholar]
  30. Yang, Z.; Sun, Y.; Liu, S.; Shen, X.; Jia, J. STD: Sparse-to-Dense 3D Object Detector for Point Cloud. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 1951–1960. [Google Scholar]
  31. Nezhadarya, E.; Taghavi, E.; Razani, R.; Liu, B.; Luo, J. Adaptive Hierarchical Down-Sampling for Point Cloud Classification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 12956–12964. [Google Scholar]
  32. Zhang, Y.; Hu, Q.; Xu, G.; Ma, Y.; Wan, J.; Guo, Y. Not All Points Are Equal: Learning Highly Efficient Point-Based Detectors for 3D LiDAR Point Clouds. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 18953–18962. [Google Scholar]
  33. He, L.; Wang, M. SliceSamp: A Promising Downsampling Alternative for Retaining Information in a Neural Network. Appl. Sci. 2023, 13, 11657. [Google Scholar] [CrossRef]
  34. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin Transformer: Hierarchical Vision Transformer Using Shifted Windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 2021, Montreal, BC, Canada, 11–17 October 2021; pp. 10012–10022. [Google Scholar]
  35. Tan, W.; Qin, N.; Ma, L.; Li, Y.; Du, J.; Cai, G.; Yang, K.; Li, J. Toronto-3D: A Large-Scale Mobile LiDAR Dataset for Semantic Segmentation of Urban Roadways. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 13–19 June 2020; pp. 202–203. [Google Scholar]
  36. Tong, G.; Li, Y.; Chen, D.; Sun, Q.; Cao, W.; Xiang, G. CSPC-Dataset: New LiDAR Point Cloud Dataset and Benchmark for Large-Scale Scene Semantic Segmentation. IEEE Access 2020, 8, 87695–87718. [Google Scholar] [CrossRef]
  37. Wang, Y.; Sun, Y.; Liu, Z.; Sarma, S.E.; Bronstein, M.M.; Solomon, J.M. Dynamic Graph CNN for Learning on Point Clouds. ACM Trans. Graph. 2019, 38, 1–12. [Google Scholar] [CrossRef]
  38. Thomas, H.; Qi, C.R.; Deschaud, J.-E.; Marcotegui, B.; Goulette, F.; Guibas, L.J. KPConv: Flexible and Deformable Convolution for Point Clouds. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 6411–6420. [Google Scholar]
  39. Li, Y.; Ma, L.; Zhong, Z.; Cao, D.; Li, J. TGNet: Geometric Graph CNN on 3-D Point Cloud Segmentation. IEEE Trans. Geosci. Remote Sens. 2020, 58, 3588–3600. [Google Scholar] [CrossRef]
  40. Boulch, A.; Guerry, J.; Le Saux, B.; Audebert, N. SnapNet: 3D Point Cloud Semantic Labeling with 2D Deep Segmentation Networks. Comput. Graph. 2018, 71, 189–198. [Google Scholar] [CrossRef]
  41. Huang, J.; You, S. Point Cloud Labeling Using 3D Convolutional Neural Network. In Proceedings of the 2016 23rd International Conference on Pattern Recognition (ICPR), Cancun, Mexico, 4–8 December 2016; pp. 2670–2675. [Google Scholar] [CrossRef]
  42. Zhao, H.; Jiang, L.; Jia, J.; Torr, P.H.S.; Koltun, V. Point Transformer. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 16259–16268. [Google Scholar]
  43. Xu, M.; Ding, R.; Zhao, H.; Qi, X. PAConv: Position Adaptive Convolution with Dynamic Kernel Assembling on Point Clouds. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 26 April 2021. [Google Scholar] [CrossRef]
  44. Park, C.; Jeong, Y.; Cho, M.; Park, J. Fast Point Transformer. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 4 April 2022. [Google Scholar] [CrossRef]
  45. Lai, X.; Liu, J.; Jiang, L.; Wang, L.; Zhao, H.; Liu, S.; Qi, X.; Jia, J. Stratified Transformer for 3D Point Cloud Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 28 March 2022. [Google Scholar] [CrossRef]
  46. Qian, G.; Li, Y.; Peng, H.; Mai, J.; Hammoud, H.A.A.K.; Elhoseiny, M.; Ghanem, B. PointNeXt: Revisiting PointNet++ with Improved Training and Scaling Strategies. In Proceedings of the Conference and Workshop on Neural Information Processing Systems, New Orleans, LA, USA, 12 October 2022. [Google Scholar] [CrossRef]
Figure 1. Timeline of the evolution of deep learning-based 3D point cloud semantic segmentation methods.
Figure 1. Timeline of the evolution of deep learning-based 3D point cloud semantic segmentation methods.
Remotesensing 17 02012 g001
Figure 2. Patch-based self-attention network architecture.
Figure 2. Patch-based self-attention network architecture.
Remotesensing 17 02012 g002
Figure 3. Multi-head self-attention module.
Figure 3. Multi-head self-attention module.
Remotesensing 17 02012 g003
Figure 4. Feature aggregation of common point.
Figure 4. Feature aggregation of common point.
Remotesensing 17 02012 g004
Figure 5. Prediction results of different models on Toronto-3D. The red circle indicates regions where the baseline model exhibits incorrect segmentation.
Figure 5. Prediction results of different models on Toronto-3D. The red circle indicates regions where the baseline model exhibits incorrect segmentation.
Remotesensing 17 02012 g005
Figure 6. Prediction results of different models on CSPC. The red circle indicates regions where the baseline model exhibits incorrect segmentation.
Figure 6. Prediction results of different models on CSPC. The red circle indicates regions where the baseline model exhibits incorrect segmentation.
Remotesensing 17 02012 g006
Figure 7. Receptive field of different methods.
Figure 7. Receptive field of different methods.
Remotesensing 17 02012 g007
Table 1. Semantic segmentation performance comparison on the L002 region of Toronto-3D.
Table 1. Semantic segmentation performance comparison on the L002 region of Toronto-3D.
MethodsIoUmIoUOA
RoadRoad MarkingNaturalBuildingUtility LinePoleCarFence
PointNet++91.47.689.874.068.659.554.07.556.691.2
PointNet++ (MSG)90.70.086.775.856.260.944.510.253.190.6
DGCNN90.60.481.364.047.156.949.37.349.689.0
KPFCNN90.20.086.886.881.173.142.921.660.391.7
MS-PCNN91.03.590.577.362.368.553.617.158.091.5
TGNet91.410.691.076.968.366.354.18.258.391.6
MS-TGNet90.918.892.280.669.471.251.113.661.091.7
RandLA-Net96.666.796.289.385.981.078.829.578.096.6
Point Transformer96.764.695.991.087.679.087.536.979.996.8
ours98.175.798.596.490.586.796.569.088.998.4
Table 2. Semantic segmentation performance comparison on the Scene2 component of the CSPC dataset.
Table 2. Semantic segmentation performance comparison on the Scene2 component of the CSPC dataset.
MethodsIoUmIoUOA
GroundBuildingCarVegetationPoleBridge
SnapNet42.843.96.010.80.00.017.354.8
PointNet++92.890.565.672.526.469.669.694.1
3D CNN78.290.51.35.40.50.219.258.4
DeepNet79.935.38.78.60.30.022.261.2
KPConv94.187.866.677.50.00.054.393.6
RandLA-Net92.087.979.967.317.06.058.393.3
Point Transformer93.092.874.993.822.149.871.095.4
ours95.895.188.998.644.558.980.397.2
Table 3. Semantic segmentation performance comparison on the Scene5 component of the CSPC (%) dataset.
Table 3. Semantic segmentation performance comparison on the Scene5 component of the CSPC (%) dataset.
MethodsIoUmIoUOA
GroundBuildingCarVegetationPoleBridge
SnapNet40.238.40.28.40.0-17.552.3
PointNet++89.887.225.134.522.8-51.991.2
3D CNN71.056.51.39.11.5-27.969.9
DeepNet71.344.90.910.60.5-25.663.3
KPConv87.588.763.254.80.5-58.892.4
RandLA-Net90.290.354.548.034.4-63.593.4
Point Transformer88.290.254.156.540.5-65.994.6
ours93.794.669.668.646.5-74.696.2
Table 4. Semantic segmentation performance comparison on the Area5 component of the S3DIS dataset.
Table 4. Semantic segmentation performance comparison on the Area5 component of the S3DIS dataset.
MethodmAccmIoUParams (M)Latency (s)
KPConv72.867.114.9105.15
PAConv73.066.690.128.13
Point Transformer76.570.412.818.07
Fast Point Transformer77.370.137.91.13
Stratified Transformer78.172.092.043.02
PointNeXt-XL83.074.941.638.16
PSNet (ours)77.871.221.877.03
Table 5. Partition strategy ablation on the L002 region of Toronto-3D.
Table 5. Partition strategy ablation on the L002 region of Toronto-3D.
ID N s a m p l e N n e i g h mIoU (%)
112820063.5
212840072.2
312880085.9
425620076.7
525640088.9
625680087.5
751220086.3
851240088.7
951280087.8
Table 6. Module design ablation on the L002 region of Toronto-3D.
Table 6. Module design ablation on the L002 region of Toronto-3D.
IDIPGSFEPECPFAmIoU (%)
1----79.9
2--85.6
3-86.8
488.9
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yi, H.; Liu, Y.; Wang, M. PSNet: Patch-Based Self-Attention Network for 3D Point Cloud Semantic Segmentation. Remote Sens. 2025, 17, 2012. https://doi.org/10.3390/rs17122012

AMA Style

Yi H, Liu Y, Wang M. PSNet: Patch-Based Self-Attention Network for 3D Point Cloud Semantic Segmentation. Remote Sensing. 2025; 17(12):2012. https://doi.org/10.3390/rs17122012

Chicago/Turabian Style

Yi, Hong, Yaru Liu, and Ming Wang. 2025. "PSNet: Patch-Based Self-Attention Network for 3D Point Cloud Semantic Segmentation" Remote Sensing 17, no. 12: 2012. https://doi.org/10.3390/rs17122012

APA Style

Yi, H., Liu, Y., & Wang, M. (2025). PSNet: Patch-Based Self-Attention Network for 3D Point Cloud Semantic Segmentation. Remote Sensing, 17(12), 2012. https://doi.org/10.3390/rs17122012

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop