Next Article in Journal
Improving UAV Remote Sensing Photogrammetry Accuracy Under Navigation Interference Using Anomaly Detection and Data Fusion
Previous Article in Journal
A Fire Segmentation Method with Flame Detail Enhancement U-Net in Multispectral Remote Sensing Images Under Category Imbalance
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Density-Aware Tree–Graph Cross-Message Passing for LiDAR Point Cloud 3D Object Detection

1
School of Electronics and Information Technology, Sun Yat-sen University, Guangzhou 510275, China
2
School of Information and Communication Engineering, University of Electronic Science and Technology of China, Chengdu 611731, China
3
School of Cyber Science and Technology, Sun Yat-sen University, Shenzhen 518107, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2025, 17(13), 2177; https://doi.org/10.3390/rs17132177
Submission received: 30 April 2025 / Revised: 16 June 2025 / Accepted: 19 June 2025 / Published: 25 June 2025

Abstract

LiDAR-based 3D object detection is fundamental in autonomous driving but remains challenging due to the irregularity, unordered nature, and non-uniform density of point clouds. Existing methods primarily rely on either graph-based or tree-based representations: Graph-based models capture fine-grained local geometry, while tree-based approaches encode hierarchical global semantics. However, these paradigms are often used independently, limiting their overall representational capacity. In this paper, we propose density-aware tree–graph cross-message passing (DA-TGCMP), a unified framework that exploits the complementary strengths of both structures to enable more expressive and robust feature learning. Specifically, we introduce a density-aware graph construction (DAGC) strategy that adaptively models geometric relationships in regions with varying point density and a hierarchical tree representation (HTR) that captures multi-scale contextual information. To bridge the gap between local precision and global contexts, we design a tree–graph cross-message-passing (TGCMP) mechanism that enables bidirectional interaction between graph and tree features. The experimental results of three large-scale benchmarks, KITTI, nuScenes, and Waymo, show that our method achieves competitive performance. Specifically, under the moderate difficulty setting, DA-TGCMP outperforms VoPiFNet by approximately 2.59%, 0.49%, and 3.05% in the car, pedestrian, and cyclist categories, respectively.

1. Introduction

With the rapid advancement of autonomous driving, LiDAR has emerged as a critical sensing modality for perceiving complex outdoor environments. Unlike image-based [1] and radar-based [2,3] sensors, LiDAR provides illumination-invariant and weather-resilient depth measurements, generating accurate and dense 3D point clouds that encode rich geometric structures [4]. These properties make LiDAR essential for a wide range of perception tasks, including 3D object detection, semantic segmentation, and scene reconstruction [1,5]. Despite its advantages, processing raw point clouds remains challenging due to their unordered, irregular, and non-uniform nature [5,6,7,8]. These characteristics hinder the direct use of convolutional operations and complicate neighborhood definition, necessitating specialized architectures capable of modeling both fine-grained local geometry and global contextual dependencies.
To address these challenges, recent methods primarily adopt either graph-based or tree-based representations. Graph-based methods [9,10,11,12] construct local neighborhoods via spatial proximity, effectively capturing fine geometric details and local variations, as shown in Figure 1a. Most use fixed-radius or k-nearest neighbor (KNN) strategies [13,14], which assume uniform point distributions. This assumption makes them sensitive to point density variations—leading to information loss in sparse regions and noisy redundancy in dense ones [15]. In contrast, tree-based approaches [16,17] exploit hierarchical structures through progressive downsampling to efficiently capture multi-scale global semantics, as shown in Figure 1b. However, such coarse-to-fine hierarchies can discard crucial local details, and static tree structures often lack adaptability to irregular spatial layouts [18]. These two paradigms exhibit complementary advantages: Graph-based models excel at local geometric encoding, while tree-based models are better suited for global semantic reasoning. Nonetheless, existing methods typically treat them in isolation, missing opportunities for joint optimization.
Recent efforts have improved graph construction and hierarchical aggregation independently [7,13], but two core limitations persist: (1) hierarchical downsampling often leads to local information loss and semantic gaps across scales; (2) there is limited interaction between local graph-based features and global tree-based context, restricting the expressive capacity of hierarchical feature learning.
Inspired by the theoretical framework of Hierarchical Support Graphs (HSGs) [19], we argue that integrating hierarchical (tree) structures with graph-based message passing can significantly enhance feature learning. Specifically, the multi-level structure combined with cross-message passing effectively reduces path resistance within the graph, thereby improving connectivity and facilitating more efficient information flow.
Therefore, we propose a novel framework: density-aware tree–graph cross-message passing for LiDAR-based 3D object detection. Our method integrates the complementary strengths of graph and tree structures to achieve both local fidelity and global context understanding. Specifically, we introduce a density-aware graph construction module that dynamically incorporates local density cues when forming neighborhoods, enabling more robust feature extraction in non-uniform regions. We then design a hierarchical tree representation module to encode multi-scale semantics while preserving spatial consistency, as shown in Figure 1c. To bridge the local–global gap, we propose a tree–graph cross-message-passing mechanism, which facilitates bi-directional feature interactions between the graph and tree branches. This design leads to stronger feature expressiveness, as it enables both precise local encoding and robust global context reasoning through complementary interactions between tree and graph representations.
In summary, the main contributions of this work are summarized as follows:
  • The DAGC strategy adaptively builds local neighborhoods by considering spatial proximity and local point density, thus avoiding redundant connections in dense regions and maintaining connectivity in sparse areas.
  • HTR models multi-scale semantic features through structured downsampling, preserving the global context while being aware of spatial structure.
  • The TGCMP mechanism enables bi-directional feature exchange between graph and tree domains, allowing local geometric patterns and global semantic cues to reinforce each other.

2. Related Works

2.1. Overview of LiDAR-Based Methods

Voxelization is a common approach that transforms irregular point clouds into structured grids, enabling 3D convolutional processing. Early methods like VoxelNet [20] pioneered end-to-end voxel feature encoding but incurred high computational costs. Subsequent works—including SECOND [21], PointPillars [22], Part- A 2  [23], Voxel-RCNN [24], HVNet [25], and CIA-SSD [26]—have focused on improving efficiency and accuracy by leveraging sparse convolutions, pseudo-BEV representations, multi-scale feature extraction, and confidence-aware refinement. VoTr [27] introduces transformers to better model contextual dependencies, addressing the locality bias of traditional CNNs. Swin3D [28] introduces a generalized contextual relative positional embedding mechanism to capture various irregularities in point cloud signals while enabling efficient self-attention on sparse voxels, thereby allowing the model to scale to larger sizes and datasets.
Point-based detectors operate directly on raw point clouds to preserve detailed geometric structures. Point-RCNN [29] applies PointNet++ [30] for hierarchical feature learning and two-stage refinement. To improve efficiency, 3DSSD [31] and IA-SSD [32] utilize progressive downsampling and fusion schemes to retain informative foreground points with reduced overhead. STD [33] introduces a point-level proposal generation mechanism with spherical anchors to enhance object recall. Point-GNN [9] formulates detection as a graph reasoning problem, aligning features to point positions via graph neural networks.
To harness both fine-grained point features and the global voxel context, hybrid approaches have gained popularity. SA-SSD [34] incorporates auxiliary supervision to enrich voxel features with point-level cues. PV-RCNN [35] introduces a voxel-set abstraction mechanism to project voxel features to keypoints, achieving fine-to-coarse information fusion. CT3D [36] leverages channel-wise transformers to model inter-channel geometric relationships, and Pyramid-RCNN [37] proposes a scale-adaptive spherical query strategy to enhance RoI feature coverage.

2.2. Methods Based on Graph Network

Graph neural networks (GNNs) have become an increasingly influential framework for processing 3D point clouds, thanks to their ability to handle irregular, unordered data and model complex geometric relationships. Unlike convolutional operations on structured grids, GNNs naturally encode local topology and inter-point dependencies, making them well-suited for point cloud representation learning [10,38,39,40]. To better capture local spatial interactions, Point-GNN [9] constructs hierarchical graphs from point clouds and performs feature propagation along edges, followed by graph-aware NMS to improve proposal refinement. DRG-CNN [12] dynamically builds neighborhood graphs per layer, allowing receptive fields to adaptively expand and enabling more flexible context encoding. SVGA-Net [41] introduces voxel-guided multi-scale graph construction and leverages attention mechanisms to selectively fuse geometric features across scales. DGT-Det3D [18] further combines graph reasoning with semantic-aware modules, enhancing both the structural and contextual aspects of point features. Collectively, these graph-based approaches demonstrate the strength of GNNs in learning fine-grained geometric representations and neighborhood-aware context, contributing to more accurate and robust 3D object detection.

2.3. Methods Based on Multi-Scale Feature Learning

Existing multi-scale feature learning methods, such as FPN-based architectures [42], PointNet++ [30], hybrid voxel encoders [25], and KD-Net [43], have advanced 3D detection by hierarchically aggregating features at different resolutions. However, they often process local and global contexts separately or lack adaptability to non-uniform density and irregular geometry. Recent adaptive hierarchies improve flexibility but still underexploit cross-scale interactions. To bridge this gap, we propose a tree–graph cross-message-passing mechanism that enables bi-directional interactions between hierarchical (tree-based) and local geometric (graph-based) features, unifying fine-grained detail preservation with global semantic reasoning for enhanced multi-scale 3D representation.

2.4. Transformer-Based 3D Object Detection Methods

The introduction of Transformers [44] has fundamentally reshaped 3D visual understanding by enabling long-range dependency modeling via self-attention mechanisms. Inspired by their breakthroughs in 2D computer vision—such as the Vision Transformer (ViT) [45,46], DETR [47], and Swin Transformer [48]—researchers have actively explored their extensions to the 3D domain. In 3D perception, Point-BERT [49] pioneers masked point modeling for unsupervised pretraining, treating point clouds as tokenized sequences and effectively capturing global semantics. Other approaches, like Point Transformer [50], propose coupling the strengths of attention mechanisms with local geometric priors, enabling precise context aggregation at both fine and coarse levels. DVST [51] integrates deformable voxel-set attention into a Transformer backbone to adaptively aggregate informative regions and enhance 3D object detection from LiDAR point clouds. Meanwhile, FusionRCNN [52] effectively fuses LiDAR geometric information and camera texture information through cross-modal attention, achieving more precise feature complementarity and enhancement, which significantly improves multi-modal 3D object detection performance. SGF3D [53] applies cross-modal attention modules to bridge modality gaps and improve fusion consistency in multi-sensor scenarios. These transformer-inspired architectures have advanced the state of the art across key 3D tasks, including object classification, semantic segmentation, and 3D detection, by bridging local detail preservation with global feature reasoning.

3. Methodology

3.1. The Overview of Our Method

The pipeline of DA-TGCMP is shown in Figure 2. The framework consists of four main stages. Initially, voxelization combined with local density estimation is employed to capture the inherent spatial non-uniformity of the point cloud. DPCG is then constructed to enhance local neighborhood modeling, especially in sparse regions. Subsequently, a hierarchical tree representation module is introduced to encode multi-scale semantic features in a structured and spatially aware manner. To bridge local and global representations, the TGCMP mechanism is developed, enabling bidirectional feature exchange between the tree and graph views. This unified design facilitates expressive hierarchical feature learning by integrating fine-grained geometric cues with high-level semantics, ultimately leading to more robust and accurate 3D detection.

3.2. Local Density Estimation

In 3D point cloud processing, the inherent non-uniformity in point distributions arises due to various factors such as sensor limitations, occlusions, and varying distances from the sensor to objects. This results in the phenomenon of dense regions near the sensor and sparse regions farther away, which introduces significant challenges for effective local feature extraction. Traditional methods that rely on fixed-radius or fixed-k neighborhood queries often fail to adapt to these density variations, which may lead to suboptimal performance in downstream tasks.
To address this issue, we propose a novel local density estimation (LDE) module that explicitly models local density variations, enabling more adaptive and robust feature extraction. Given a point cloud P = { p 1 , p 2 , , p N } , where each point p i R 3 represents a 3D coordinate, the first step is to identify the local neighborhood for each point. This is achieved by performing search, which selects the k closest points to each point p i . Formally, for each point p i , we define the neighborhood N k ( p i ) as a set of its KNN:
N k ( p i ) = { p j | j Top - k ( p i ) }
The KNN search strategy ensures that each point has a fixed number of neighbors, which provides a consistent basis for subsequent density estimation, irrespective of the point’s position in the point cloud. Compared to fixed-radius approaches, this method offers a more stable representation, particularly in point clouds with significant density variations. Once the local neighborhood for each point is established, the next step is to estimate the local density ρ i for each point using Kernel Density Estimation (KDE). KDE provides a smooth and continuous estimate of the local density by assigning weights to the neighboring points based on their distances from p i . The KDE for point p i is given by
ρ i = p j N k ( p i ) exp p i p j 2 2 σ 2
where p i p j denotes the Euclidean distance between points p i and p j , and σ is the bandwidth parameter that controls the kernel’s spread. This kernel function ensures that closer neighbors contribute more significantly to the density estimate than more distant neighbors, which allows for fine-grained local density modeling.
To further enhance the density estimation process, we introduce an adaptive bandwidth strategy that dynamically adjusts bandwidth σ i based on the local distribution of points. The bandwidth is computed as the average distance between the point p i and its k nearest neighbors:
σ i = 1 k p j N k ( p i ) p i p j
This adaptive strategy ensures that in dense regions, the kernel bandwidth is smaller (i.e., the kernel becomes more concentrated around each point), while in sparse regions, the kernel bandwidth becomes larger. This adaptability allows the method to better handle point clouds with varying density, ensuring that the kernel can effectively capture local structures regardless of the region’s sparsity or density. After estimating the local densities, we perform a density normalization step to rescale the density values to a consistent range. This normalization step is crucial for enabling meaningful comparisons between points across point clouds with varying global densities. The normalized density ρ i for each point is computed as follows:
ρ i = ρ i min ( ρ ) max ( ρ ) min ( ρ )
where min ( ρ ) and max ( ρ ) are the minimum and maximum density values across the entire point cloud, respectively. This normalization step ensures that the local density values are rescaled to the range [ 0 , 1 ] , providing a uniform density scale across point clouds with different global density distributions. Finally, based on the normalized density values, we partition the point cloud into dense and sparse regions. Points with ρ i > ρ dense are classified as part of a dense region, while points with ρ i < ρ sparse are classified as part of a sparse region. The thresholds ρ dense and ρ sparse can either be manually defined or adaptively determined based on the statistical properties of the density distribution, such as the quantiles of the density values. This step enables the segmentation of the point cloud into regions that exhibit similar density characteristics, allowing for region-specific processing strategies in subsequent steps. For example, in dense regions, fine-grained feature extraction can be performed, whereas in sparse regions, broader receptive fields may be required to capture important structures.

3.3. Dynamic Point Cloud Grouping

To effectively capture local geometric structures in point clouds, we propose a novel method termed dynamic point cloud grouping (DPCG). As shown in Algorithm 1, by integrating local density indices with adaptive distance thresholds, DPCG enables efficient point cloud segmentation and feature aggregation. The dynamic distance threshold r i for each point is defined as follows:
r i = r min + ( r max r min ) ( 1 D i ) ,
where r min and r max denote the minimum and maximum distance thresholds, respectively. Typically, r min is set to a small value (e.g., 0.1) to capture fine-grained structures in high-density regions, while r max is assigned a larger value (e.g., 0.5) to ensure robust grouping in sparse areas. This dynamic adjustment balances the trade-off between geometric fidelity and computational efficiency.
Grouping is performed using a breadth-first search (BFS) strategy. At each iteration, the ungrouped point with the highest density D i is selected as the seed x seed . A queue is initialized with x seed , and points are iteratively dequeued to explore their unvisited neighbor x k . A neighbor x k is added to the current group if the distance x j x k is less than its adaptive threshold r k . The process repeats until the queue is empty. New groups are initialized by selecting the next highest-density ungrouped point, and the procedure continues until all points are assigned to groups.
Algorithm 1 Dynamic Point Cloud Grouping (DPCG)
Input: 
Point cloud X , density indices { D i }
Output: 
Groups G
   1:
for each point x i X  do
   2:
    Compute dynamic threshold r i using Equation (5)
   3:
end for
   4:
Mark all points as unvisited
   5:
while there exist unvisited points do
   6:
    Select seed point x seed with the highest D i
   7:
    Initialize a queue with x seed
   8:
    Mark x seed as visited
   9:
    while queue is not empty do
 10:
        Dequeue point x j
 11:
        for each unvisited neighbor x k of x j  do
 12:
            if  x j x k r k  then
 13:
                Add x k to the current group
 14:
                Mark x k as visited
 15:
                Enqueue x k
 16:
            end if
 17:
        end for
 18:
    end while
 19:
    Save the current group to G
 20:
end while
 21:
return  G

3.4. Density Graph Construction

Dynamic point cloud grouping enables adaptive neighborhood formation, but conventional graph construction often overlooks underlying density variations. This can result in missing fine-grained patterns in sparse areas or overemphasizing redundant information in dense regions.
To address this, we propose a density graph construction strategy that integrates local density cues for more reliable graph formation. This enhances geometric structure preservation and facilitates robust local feature extraction across varying density regions.
Formally, given a point cloud P = { ( x i , f i ) } i = 1 N with coordinates x i R 3 and features f i R d , we define nodes as
V = { g i = ( x i , f i ) i = 1 , , N } .
For each node g i , we estimate an adaptive radius r i and define its neighborhood:
N ( i ) = { g j   x i x j r i , j i } .
The adjacency matrix A encodes these relationships, with A i j = 1 if g j N ( i ) . The edge features are
e i j = ϕ ( f i , f j , x i x j ) ,
where ϕ ( · ) is a learnable transformation (e.g., MLP).
Node features are projected for attention computation:
q i = W q f i , k j = W k f j ,
with the following attention scores:
s i j = softmax j q i k j + δ mul ( x i x j ) .
The updated node features are
f ^ i = j N ( i ) s i j · e i j .
This density-aware graph enables effective local feature learning by jointly modeling semantic and geometric relationships across variable-density regions.

3.5. Hierarchical Tree Representation

We introduce a novel hierarchical tree representation during the point cloud downsampling process to effectively organize points at multiple scales. Given a single-frame point cloud P R N × 3 , where N is the number of points and each point consists of 3D coordinates, we progressively construct a tree hierarchy. At each layer, the point cloud is represented at a coarser resolution by dividing the 3D space into cubic voxels. The centroid of each voxel acts as the parent node of the tree, enabling multi-scale feature extraction and efficient attention aggregation.

3.5.1. Voxel Grid Construction and Parent–Child Mapping

To build the tree, the 3D space is partitioned into a voxel grid at each layer. For each voxel v i , we compute its centroid C i d R 3 , which serves as the representative point for that voxel. At the coarsest level l = L τ , the centroids directly correspond to the input points of the point cloud X, acting as the leaf nodes of the tree. As we move to finer layers l, the voxel size increases, representing a coarser version of the point cloud. Each centroid at the finer layer is mapped to a parent voxel at the coarser layer, forming the tree hierarchy.
Let V l represent the set of voxels at layer l, where each voxel contains a subset of points. The centroid C i d of a voxel at layer l is computed by averaging the coordinates of the points within that voxel:
C i d = 1 | P i d | p j P i d p j
where P i d is the set of points in the i-th voxel at layer l, and | P i d | is the number of points in that voxel.

3.5.2. Parent–Child Relationship and Layer-Wise Indexing

The essence of the tree construction lies in mapping the parent and child nodes across layers. For each child point in a finer layer, we define the index ρ d c that maps it to its parent point in the coarser layer:
ρ d c = i if p j v i d
where p j is a point in the finer layer, and v i d is the voxel containing p j . Conversely, for each parent point at layer l, the set of child points P i d in the next layer l + 1 is defined as
P i d = { p j | p j v i c }
where v i c is the voxel containing the parent point at layer l.
The voxel size at each layer l increases according to
s l = s 0 × N l 1
where s 0 is the initial voxel size, N is the scaling factor, and l is the layer index. As l increases, the voxel’s size grows, leading to progressively coarser representations of the point cloud.

3.5.3. Coarse-Level Node Computation

At each layer, the parent node’s coordinates are computed as the average of the child nodes’ coordinates:
C i c = 1 | I i | j I i C j d
where C i c is the centroid of the i-th parent node at the coarser level, I i is the set of indices for the child points at the finer level, and C j d is the coordinate of the j-th child point.
This tree construction process continues until we obtain a multi-layered tree, where the leaf nodes represent the fine-grained point cloud X, and the root node represents the coarsest point representation. As shown in Algorithm 2, this tree structure facilitates efficient feature aggregation across scales, enabling effective point cloud representation learning.
Algorithm 2 Tree Construction for Point Cloud Downsampling
  1:
Input: Point cloud X R N × 3 , initial voxel size s 0 , scaling factor N
  2:
Output: Tree structure with parent-child relationships
  3:
Initialize tree T
  4:
Set voxel size s s 0
  5:
for each layer l from 1 to L max  do
  6:
    Divide 3D space into voxel grid with size s
  7:
    for each voxel v i at layer l do
  8:
        Compute centroid C i l by averaging points in v i
  9:
    end for
10:
    if  l = L max  then
11:
        Add centroids C i l as leaf nodes in T
12:
    else
13:
        for each centroitd C i l  do
14:
             Map C i l to its parent centroid C parent ( l + 1 ) in layer l + 1
15:
             Add parent-child relationship to tree T
16:
        end for
17:
    end if
18:
    Update voxel size s s × N
19:
end for
20:
Return tree T

3.6. Tree–Graph Cross-Message Passing

In hierarchical point cloud processing, downsampling is essential for building multi-scale representations but inevitably causes the loss of fine-grained local structures. Density-aware graphs provide a way to preserve local geometric and semantic relationships; however, they often suffer from information degradation across hierarchical levels. On the other hand, tree structures, constructed in a dense-to-sparse manner, effectively retain global contextual cues but are less sensitive to local variations. By drawing inspiration from the theoretical framework of hierarchical support graphs (HSGs) [19], to bridge the gap between local precision and the global context, we propose a TGCMP mechanism, which enables bi-directional feature interaction between tree and graph representations, leading to more expressive hierarchical feature learning. TGCMP implicitly increases the number of node-disjoint paths between node pairs through the integration of tree and graph structures. This design is analogous to how the vertical and horizontal edges in HSGs enhance graph connectivity and facilitate more robust information exchange. Specifically, TGCMP comprises the following components:
  • The tree features at level l as F l tree R N l × D , where N l is the number of nodes and D is the feature dimension.
  • The graph features at level l as F l graph R N l × D .
  • The 3D coordinates of nodes as X l R N l × 3 .

3.6.1. Top-Down Message Passing (Tree → Graph)

The top-down branch aims to inject global contextual information from the tree structure into the local graph features at the same level l. First, the tree feature f i tree of node i is projected into a query vector:
q i = W q f i tree ,
where W q R D × d is a learnable weight matrix, and d denotes the query dimension. Similarly, the graph features are projected into keys and values:
k j = W k f j graph , e j = W e f j graph ,
where W k , W e R D × d denotes learnable parameters. To encode spatial relationships, we compute a relative positional encoding:
δ i j = ϕ ( X i X j ) ,
where ϕ ( · ) is a lightweight neural network, such as an MLP, that maps relative coordinates into feature space. The attention weight between a tree node i and a neighboring graph node j N ( i ) is computed as follows:
s i j = softmax j q i k j + δ i j .
The aggregated message at node i is computed by applying an attention-weighted aggregation over its neighbors:
f ˜ i graph = j N ( i ) s i j · e j .
where e j denotes the transformed features of neighboring graph nodes, and s i j denotes the attention weights. The updated feature at node i is then obtained by combining its original feature with the aggregated message using a residual connection followed by a multilayer perceptron (MLP):
f i graph f i graph + MLP f ˜ i graph .
This design provides several theoretical advantages:
  • Enhanced connectivity: Cross-layer aggregation increases node-disjoint paths:
    c a b TGCMP > c a b graph
    thus alleviating over-squashing and improving message robustness.
  • Reduced effective resistance: Cross-layer connections reduce the effective resistance between node pairs:
    R a b TGCMP = ( e a e b ) L TGCMP + ( e a e b ) < R a b graph
    leading to more efficient long-range communication.
  • Expanded receptive field: The tree-to-graph cross-message passing expands the receptive field logarithmically without increasing the network’s depth:
    receptive_field TGCMP = O ( log N )
These properties ensure that top-down message passing provides both global context injection and computational efficiency.

3.6.2. Bottom-Up Message Passing (Graph → Tree)

The bottom-up branch aggregates fine-grained local features from graph nodes and propagates them to their corresponding parent tree nodes.
Let C ( i ) denote the set of child graph nodes associated with a tree node i. The feature of the parent node is updated as follows:
f i tree = 1 | C ( i ) | j C ( i ) MLP Concat f j graph , X j X i ,
where Concat ( · ) denotes feature concatenation, and the MLP learns to adaptively combine the child features with their relative positional offsets.

3.7. Loss Functions

Our network is structured as a two-stage detector and optimized by a unified multi-task loss that concurrently supervises the RPN and R-CNN stages.
L = L rpn + L rcnn .
Following [23], the RPN loss L rpn consists of object classification loss, box regression loss, and corner regression loss:
L rpn = α 1 L cls + α 2 L reg + α 3 L corner ,
where α 1 , α 2 , and α 3 are weighting coefficients. The classification loss L cls is computed using the focal loss [54]:
L cls = α ( 1 p a ) γ log p a ,
where p a is the predicted objectness score, and α = 0.25 and γ = 2 are hyperparameters. The box regression loss L reg employs the Smooth L1 loss [55] over the encoded box residuals:
L reg = b ( x , y , z , w , l , h , θ ) SmoothL 1 ( Δ b ) ,
where the residuals are defined as
Δ x = x g t x a d a , Δ y = y g t y a d a , Δ z = z g t z a d a , Δ w = log w g t w a , Δ l = log l g t l a , Δ h = log h g t h a , Δ θ = θ g t θ a ,
with d a = ( w a ) 2 + ( l a ) 2 . The corner loss L corner is computed based on the sine error [21] between the predicted and ground truth box corners. In the refinement stage, L rcnn supervises the region-of-interest (ROI) heads by optimizing classification, bounding box regression, and corner prediction:
L rcnn = L rcls + L rloc + L rcorner .
Here, L rcls is the cross-entropy loss for classification, L rloc is the Smooth L1 loss for box regression, and L rcorner is the sine-based corner loss.

4. Experiment

4.1. Datasets and Evaluation Metric

To comprehensively evaluate the performance of our proposed DA-TGCMP method, we conduct experiments on three large-scale LiDAR-based 3D object detection benchmarks: KITTI, nuScenes, and Waymo Open Dataset. These datasets vary in complexity, sensor settings, and evaluation metrics, thus enabling a holistic assessment.
The KITTI dataset is one of the earliest and most widely adopted benchmarks for 3D detection. We follow the standard evaluation protocol that partitions the official 7481 LiDAR scans into a training set (3712 samples) and a validation set (3769 samples), ensuring minimal sequence overlap to maintain evaluation fairness. The detection range is constrained within [0 m, 70.4 m] along the X-axis, [−40 m, 40 m] along the Y-axis, and [−3 m, 1 m] along the Z-axis in the LiDAR coordinate frame. KITTI defines evaluation difficulty levels—easy, moderate, and hard—based on object size, occlusion, and truncation, with performance reported using average precision (AP) under the 40 recall position metrics.
The nuScenes dataset introduces a more challenging urban scenario, offering 1000 driving scenes captured at 2 Hz and covering diverse weather, lighting, and traffic conditions. It includes 700 scenes for training, 150 for validation, and 150 for testing. Rich annotations span 23 object categories and are aligned across multiple modalities, including LiDAR, cameras, and radar. Detection performance is assessed using both the mean average precision (mAP) and the nuScenes detection score (NDS), the latter being a compound metric accounting for classification, translation, scale, orientation, velocity, and attribute estimation.
The Waymo Open Dataset is known for its dense and high-fidelity point cloud data, capturing 360° environments with a fine voxel resolution of 0.1 m, 0.1 m, and 0.15 m. It comprises 798 training sequences (around 158,000 frames) and 202 validation sequences (around 40,000 frames). The detection space spans from −75.2 m to 75.2 m in both X and Y directions and from −2 m to 4 m vertically. For fair evaluation, Waymo differentiates predictions into LEVEL_1 (≥5 LiDAR points) and LEVEL_2 (≥1 LiDAR point). The metrics used include the standard mean average precision (mAP) and its orientation-sensitive variant mAPH, which reflects heading prediction quality.

4.2. Implementation Details

Recent works [21,22,33,35] used 11 recall positions to calculate average precision (AP) on the validation set, while the KITTI benchmark adopts 40 recall positions for the test set. We follow this approach, using 40 recall positions for the test set and 11 for the validation set. For consistency, we train DA-TGCMP on the complete set of KITTI training samples when providing submissions to the KITTI test benchmark.
For region proposal generation, we downsample the sparse voxel tensor by factors of 1×, 2×, 4×, and 8×. The voxel feature tensor sizes are set to [16, 32, 64, 64] at each scale. We place six anchors at each pixel of the BEV feature map for three categories, each with two heading angles, and compute pixel features for both classification and box regression tasks.
We use 3D average precision (3D AP) and bird’s-eye-view average precision (BEV AP) as our performance metrics. Training follows standard practices with an Adam optimizer [56], a weight decay of 0.01, and a momentum of 0.9, and we use over 70 epochs with an initial learning rate of 0.003. During inference, we apply non-maximum suppression (NMS) with a threshold of 0.7 to filter out redundant proposals and keep the top 100 for input into the refinement module.

5. Results and Analysis

5.1. Comparison on the KITTI Test Set

Table 1 summarizes the quantitative performance of DA-TGCMP on the KITTI 3D detection test set across three object categories, car, pedestrian, and cyclist, evaluated under easy, moderate, and hard settings. The results show that DA-TGCMP achieves the highest AP scores among all listed methods in the car category: 91.01% (easy), 83.56% (moderate), and 79.14% (hard). Compared to previous leading methods such as PVT-SSD (90.65%, 82.29%, and 76.85%) and PV-RCNN (90.25%, 81.43%, and 76.82%), the proposed method shows moderate yet consistent improvements, particularly under the moderate and hard settings, suggesting the effective handling of challenging driving scenes. For the pedestrian category, DA-TGCMP reaches AP scores of 53.53% (easy), 47.92% (moderate), and 45.96% (hard). These results are marginally better than existing voxel-based methods such as VoFiNet (53.07%, 47.43%, and 45.22%) and HVPR (53.47%, 43.96%, and 40.64%), indicating slight gains in detecting small-scale or occluded pedestrian objects. However, the improvements are relatively limited, suggesting that the detection of pedestrians remains a challenging task even for advanced methods. In the cyclist category, DA-TGCMP achieves 81.62% (easy), 67.15% (moderate), and 61.73% (hard). These values are close to or slightly above the best-performing baselines, such as VoFiNet (77.64%, 64.10%, and 58.00%) and Part-A² (79.17%, 63.52%, and 56.93%), indicating competitive performance. The gains are more pronounced under higher difficulty levels, reflecting the method’s robustness when facing occlusion or limited point density. Overall, DA-TGCMP demonstrates better performance across all categories, particularly excelling in the car detection task. The relatively smaller margins in pedestrian and cyclist detection suggest room for further improvement, particularly in handling fine-grained localization under occlusion.

5.2. Comparison on the nuScenes Test Set

The DA-TGCMP algorithm demonstrates exceptional performance on the nuScenes 3D detection test set shown in Table 2, as evidenced by its high mean average precision (mAP) of 66.6 and NuScenes detection score (NDS) of 71.7. These scores indicate that the algorithm can effectively detect various objects in complex traffic scenarios, providing a robust foundation for 3D object detection tasks. Car Detection: With an accuracy of 86.9, DA-TGCMP excels in detecting cars, which are the most common objects in traffic scenarios. This high accuracy ensures the reliable detection of primary road users, which is crucial for autonomous driving safety. Pedestrian Detection: Achieving an accuracy of 86.9, the algorithm also performs outstandingly in identifying pedestrians. Given the complexity and variability of human shapes and poses, this result highlights the algorithm’s strong feature extraction and generalization capabilities. Construction Vehicle (C.V.) Detection: Achieving an accuracy of 32.1, the algorithm can manage the complex shapes of construction vehicles. Although there is room for improvement, this accuracy reflects its capability to detect these less common but important objects in specific environments. Traffic Cone (T.C.) Detection: At 82.3, the algorithm demonstrates good performance in detecting traffic cones. Their unique shape is effectively captured by the algorithm, aiding in road guidance and obstacle avoidance. Compared to other methods, DA-TGCMP shows competitive performance across most categories. It performs particularly well in car and pedestrian detection, indicating its strength in handling common and complex traffic participants. However, there is room for improvement in detecting objects like bicycles and motorcycles.

5.3. Comparison on Waymo Validation Set

As detailed in Table 3, we validate the multi-class detection results of the Waymo Open Dataset. The DA-TGCMP algorithm exhibits remarkable performance on the Waymo Open Dataset. Specifically, in vehicle detection, it attains a LEVEL_1 mAP of 76.64, marking a 4.37% improvement over SECOND’s 72.27 and a substantial 20.02% enhancement compared to PointPillars’ 56.62. In pedestrian detection, the LEVEL_1 mAP of 74.46 represents a 15.21% improvement relative to PointPillars’ 59.25. Moreover, for cyclist detection at LEVEL_2, it achieves an mAPH of 64.78, surpassing PointPillars’ 58.34 by 6.44%. In vehicle detection, it achieves a high LEVEL_1 mAP and a LEVEL_2 mAPH of 76.69, outperforming CenterPoint (72.15) and VoTr-TSD (74.95) by 4.49% and 1.69%, respectively. For pedestrian detection, it demonstrates robust performance at both difficulty levels, achieving a LEVEL_1 mAP of 74.46, which underscores its effective handling of challenges such as diverse poses and occlusions. In cyclist detection, while the LEVEL_1 mAP is slightly lower than some methods, the LEVEL_2 mAPH of 64.78 indicates the efficient detection of dynamic cyclists. In pedestrian detection at LEVEL_1, it outperforms CenterPoint (69.04) and VoTr-TSD (71.36) by a significant margin. These results highlight DA-TGCMP’s effectiveness in managing complex traffic scenes with high accuracy and good generalization.

5.4. Ablation Studies

In this section, we perform an ablation study to evaluate the contribution of key components in the DA-TGCMP framework. The main goal is to investigate how each individual module impacts the overall performance of our model in the task of LiDAR-based 3D object detection.

5.4.1. Effect of Each Module

The baseline model provides a robust foundation for 3D object detection, achieving strong performance across both the KITTI and nuScenes datasets. As detailed in Table 4, On the KITTI dataset, the baseline model achieves 91.19%, 82.63%, and 82.01% in the easy, moderate, and hard categories, respectively. On the nuScenes dataset, the model records an NDS of 65.4 and a mean average precision (mAP) of 0.592. These results demonstrate the baseline’s ability to effectively capture the key features required for accurate 3D object detection, establishing it as a strong starting point for further enhancements.
Integrating the DAGC module with the baseline model leads to significant performance improvements, especially in the moderate and hard categories on KITTI. The model achieves 84.86% and 82.61% in these categories, respectively, compared to the baseline’s 82.63% and 82.01%. This demonstrates DAGC’s ability to address varying point densities within LiDAR point clouds by dynamically adapting the graph’s structure. The improved results highlight DAGC’s effectiveness in enhancing feature extraction, particularly in complex scenarios where point cloud density varies. Additionally, on the nuScenes dataset, the model shows slight increases in both NDS and mAP, further confirming DAGC’s role in improving feature representation.
The addition of the hierarchical tree representation (HTR) further refines the model’s ability to capture multi-scale semantic features. While the improvement in the easy category on KITTI is marginal, the model experiences notable gains in the moderate and hard categories, with 84.24% and 82.24%, respectively, compared to the baseline’s 82.63% and 82.01%. This indicates that HTR enhances the model’s ability to preserve global context and capture multi-scale features through its structured downsampling approach. These improvements are also reflected in a slight increase in NDS on the nuScenes dataset, further demonstrating the effectiveness of hierarchical downsampling in strengthening the model’s performance.
When combining DAGC and HTR with the tree–graph cross-message-passing (TGCM) module, the model experiences the most significant performance improvements. On KITTI, the model achieves 92.52%, 86.87%, and 82.59% for the easy, moderate, and hard categories, respectively, a substantial improvement over the previous configurations. Similarly, on nuScenes, the model records an NDS of 68.1 and an mAP of 0.630, further emphasizing the importance of TGCM. This module facilitates bi-directional feature interaction between the graph and tree structures, enabling the model to effectively aggregate local features with the global context. The enhanced feature learning process through cross-message passing results in significant boosts across all performance metrics on both datasets.
Each module contributes uniquely to the performance improvements observed in the experimental results: DAGC proves to be highly effective in handling varying point densities within LiDAR point clouds. Its adaptive graph construction allows the model to better capture geometric structures, especially in the moderate and hard categories of KITTI. This improvement is crucial for preserving relationships across regions with varying densities and ensuring robust feature extraction.
HTR strengthens the model’s ability to capture global context and multi-scale semantic features. The structured downsampling approach of HTR helps preserve crucial global information, as demonstrated by the improved performance across different difficulty levels on KITTI and the enhanced NDS on nuScenes.
TGCMP stands out as the most impactful module, refining local features and effectively aggregating global context through cross-message passing between tree and graph representations. This dual enhancement significantly boosts performance across all metrics on both datasets, demonstrating the importance of incorporating both local and global feature interactions.

5.4.2. Effect of DAGC

To further validate the effectiveness of the proposed density-aware graph construction (DAGC) module, we conduct a comparative study against two widely used neighborhood construction strategies, fixed-radius and kNN, on the KITTI dataset. The evaluation spans three object categories—car, pedestrian, and cyclist—across all standard difficulty levels (easy, moderate, and hard).
DAGC consistently achieves the highest accuracy across all difficulty levels. In particular, as follow in Table 5, it exhibits substantial gains in the moderate and hard categories, which typically involve cluttered scenes or partially occluded vehicles. These improvements highlight DAGC’s ability to construct more informative graph structures that better capture the local geometric variations in sparse or complex regions.The performance of DAGC surpasses both fixed-radius and KNN methods, with notable advantages in the hard category. Pedestrian detection in this category is challenging due to factors such as occlusion and long-range perception. The superior results of DAGC suggest that its density-adaptive neighborhood construction enables the more effective representation of pedestrians under such adverse conditions. DAGC also leads in the easy and hard categories for cyclist detection. The improvements in the hard setting reflect the module’s robustness in handling ambiguous or incomplete point cloud inputs, while the gains in the easy setting demonstrate that even in less complex scenarios, DAGC facilitates more accurate feature extraction compared to traditional methods.

5.4.3. Effect Between Dense and Sparse Regions

As shown in Table 6, our proposed DA-TGCMP consistently demonstrates superior performance in small-object detection across both dense and sparse regions. Specifically, on the KITTI dataset, DA-TGCMP achieves a notable 92.38 mAP for bicycles (Cyc) in dense regions and 17.36 mAP in sparse regions (IoU = 0.5), outperforming competitive baselines such as PointGNN and VoxelGraph-RCNN by a substantial margin. This corresponds to a 68% relative improvement in sparse-region small-object detection compared to the best alternative method.
Crucially, DA-TGCMP exhibits remarkable resilience against sparsity-induced performance degradation. While existing baselines suffer severe accuracy declines in sparse regions (e.g., PointGNN’s Cyc mAP drops by 91.9% from dense to sparse regions), DA-TGCMP significantly mitigates this effect, reducing the performance drop by 10.7 percentage points and achieving a retention rate of 81.2%. These results underline the effectiveness of our dynamic graph convolution in preserving discriminative features in challenging sparse environments.

5.4.4. Effect of Different Dynamic Radius Designs

We systematically evaluate the impact of different dynamic distance threshold designs on small-object detection performance, as shown in Table 7. The fixed-radius method performs relatively poorly (56.21 AP) due to its inability to adapt to variations in point cloud density. In contrast, the density-aware linear adjustment strategy significantly improves detection accuracy (60.82 AP), indicating that adapting to local sparsity enables more effective feature aggregation. Further incorporating nonlinear adjustment functions, the exponential form ( exp ( D i ) ) demonstrates better inclusiveness in sparse regions and achieves 61.47 AP, outperforming the logarithmic form (59.35 AP), which responds more slowly to density changes. Moreover, broadening the threshold range (from 0.05 to 0.6) enhances the ability to capture structures in sparse regions while maintaining accuracy in dense areas, leading to 62.03 AP. The best performance (63.40 AP) is achieved using an MLP-based strategy to learn adaptive thresholds, suggesting that data-driven threshold learning combining semantic and structural cues is a promising direction for future research.
As shown in Figure 3, the BEV visualization reveals the effectiveness of our dynamic point cloud processing pipeline. Figure 3a displays the rasterized point cloud density distribution. Red points depict the spatial arrangement of point clouds after rasterization, reflecting the raw density characteristics of the scene. From Figure 3b, this adaptive clustering reorganizes sparse point cloud data into semantically meaningful groups, facilitating downstream tasks like object detection and segmentation.

5.4.5. Effect of Tree Structure

To evaluate the contribution of hierarchical tree structure to point cloud representation, we conduct a comparison with two baseline methods, single-scale voxel aggregation and PointNet-style global pooling, as shown in Table 8. In our experiments comparing single-scale voxel aggregation, PointNet-style global pooling, and the proposed hierarchical tree structure, the results demonstrate that the tree structure outperforms the other methods across all metrics and datasets. Single-scale voxel aggregation performed decently on simpler scenarios but struggled with complex cases, as it fails to capture multi-scale features. PointNet-style global pooling showed slight improvements over voxel aggregation, particularly in more complex scenarios, but still lacked the effective handling of multi-scale information. Hierarchical tree structure consistently outperformed the other methods, achieving the highest accuracy on both KITTI and nuScenes. Its ability to capture multi-scale relationships and hierarchical structures contributed to its superior performance, especially in challenging scenarios. Overall, the hierarchical tree structure proves to be the most effective in handling diverse 3D object detection tasks, particularly by leveraging multi-scale feature aggregation and hierarchical relationships, which are crucial for complex, real-world applications.

5.4.6. Effect of TGCMP

To validate the effectiveness of our method, we design a series of ablation experiments focusing on evaluating the advantages of TGCMP in multi-scale feature learning, especially in capturing both global contextual information and fine-grained local features. Specifically, our ablation experiments validate the following key aspects: role of top-down message passing, role of bottom-up message passing, and the importance of tree–graph interaction.
As shown in Table 9, The TGCMP mechanism significantly improves 3D object detection performance in point clouds by integrating both top-down and bottom-up message-passing strategies. By jointly capturing global context and local details, TGCMP outperforms the use of either tree-based or graph-based message passing alone. It achieves superior detection accuracy across all object categories (Car.AP 83.05, Pedestrian.AP 62.88, and Cyclist.AP 77.55), demonstrating its strong representational capacity and adaptability for multi-scale feature learning.

5.4.7. Memory–Performance Trade-Off Analysis

To investigate the relationship between detection performance and resource consumption, we conduct a comparative study of three architectures—DAGC, HTR, and TGCMP—focusing on their memory usage and detection accuracy across three object categories: car, pedestrian, and bicycle. As shown in Table 10, DPCG and HIT exhibit low memory footprints (11.854 GB and 8.949 GB, respectively) and reduced FLOPs (near-zero and −2.875 GFLOPs). However, these efficiency gains come at the cost of suboptimal detection performance, particularly for the pedestrian and bicycle categories.
In contrast, TPCMP consistently achieves the highest detection accuracy across all object classes—82.56 for cars, 63.31 for pedestrians, and 78.38 for bicycles. This improvement is accompanied by the highest memory usage (14.372 GB) and a significant increase in computational overhead (+3.581 GFLOPs) compared to the implicit baseline. These findings highlight a clear trade-off: while TPCMP is ideal for accuracy-critical applications, HTR may be more suitable for resource-constrained settings, and DAGC serves as a balanced general-purpose alternative.

6. Visualization

Figure 4 shows the visualization results of the KITTI dataset. The visualization results on the KITTI dataset vividly demonstrate the effectiveness of our method, particularly in detecting small objects at a distance. In the first and third rows of the figure, our method successfully detects vehicles at a significant distance from the sensor. These distant vehicles, appearing small in the point cloud, are accurately identified and enclosed within green bounding boxes. This indicates that our approach can effectively capture and recognize the sparse and subtle point clusters representative of small, distant objects. The successful detection of these distant vehicles underscores the robustness and sensitivity of our method in handling challenging scenarios involving small objects. This capability is crucial for enhancing the safety and reliability of autonomous driving systems, as it ensures that even objects far away from the vehicle can be promptly and accurately detected.
We compare voxel-based SECOND, point-voxel fusion PV-RCNN++, and our DA-TGCMP on highway and urban scenes, as shown in the Figure 5. In highway scenes, SECOND struggles with distant small vehicles, showing missed detections and inaccurate boxes. PV-RCNN++ improves this slightly, but localization remains suboptimal. DA-TGCMP delivers the precise and complete detection of small distant objects, benefiting from the tree–graph cross-message passing that enhances features in sparse regions. In urban scenes, SECOND misses many small objects (e.g., cyclists, pedestrians). PV-RCNN++ detects more objects but suffers from overlap and inaccuracy. DA-TGCMP achieves the accurate and comprehensive detection of small objects by effectively capturing fine-grained features and context. These results highlight that DA-TGCMP significantly improves small-object detection without sacrificing large-object performance.

7. Discussion

The DA-TGCMP framework shows more pronounced improvements in detecting cars compared to pedestrians and cyclists due to the larger size and higher density of point clouds for cars, which allows for more effective feature extraction and modeling. In contrast, the smaller size and sparsity of point clouds for pedestrians and cyclists, along with their more complex and variable shapes, make it harder to achieve the same level of accuracy.
There is a clear trade-off between accuracy and computational efficiency. The TGCMP mechanism, which provides the highest accuracy by enabling bidirectional feature interaction between tree and graph structures, requires the most memory and computational resources. On the other hand, simpler modules like DAGC and HTR offer lower accuracy but are more computationally efficient, making them suitable for resource-constrained settings.
The method is expected to perform well in scenarios with dense point clouds, structured environments, and clear, unoccluded scenes, where both local and global features can be effectively captured. However, it may struggle in situations with sparse point clouds, complex and variable object shapes, and significant occlusion, which can limit the framework’s ability to extract robust features for accurate detection.
Although DA-TGCMP demonstrates strong performance in detecting large objects such as vehicles, challenges remain for small objects like pedestrians and cyclists due to sparse point clouds and complex, variable shapes. In future work, we aim to enhance small-object detection by designing lighter and more efficient feature interaction modules that better capture fine-grained details in sparse and cluttered scenes. We also plan to improve the computational efficiency of the TGCMP mechanism to achieve a better balance between accuracy and real-time performance.
In addition, we intend to extend our framework for the automatic annotation of unlabeled datasets. By leveraging the high accuracy and strong generalization of DA-TGCMP in structured and dense environments, we aim to enable large-scale 3D dataset construction with reduced reliance on manual labeling.

8. Conclusions

In this paper, we present DA-TGCMP, a unified framework that integrates density-aware graphs and hierarchical trees via a novel tree–graph cross-message-passing mechanism. This design bridges the gap between local geometric precision and global contextual understanding, enabling more expressive and robust feature learning for LiDAR-based 3D object detection. Through extensive experiments on three large-scale benchmarks—KITTI, nuScenes, and Waymo—we demonstrate the effectiveness and generalizability of our approach across diverse driving scenarios. Moreover, comprehensive ablation studies confirm the individual contributions of the density-aware graph construction, hierarchical tree representation, and cross-message-passing module, each playing a vital role in boosting detection accuracy. Together, these components make DA-TGCMP a promising solution for advancing 3D perception in autonomous driving systems.

Author Contributions

Methodology, W.Z. and Y.L.; Validation, J.L.; Formal analysis, H.H.; Investigation, H.R.; Writing—original draft, J.Z.; Supervision, H.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Southern Marine Science and Engineering Guangdong Laboratory (Zhuhai) under Grant SML2021SP406.

Data Availability Statement

The original contributions presented in the study are included in the article, further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wu, T.J.; He, R.; Peng, C.C. Real-Time Environmental Contour Construction Using 3D LiDAR and Image Recognition with Object Removal. Remote Sens. 2024, 16, 4513. [Google Scholar] [CrossRef]
  2. Ren, H.; Zhou, R.; Zou, L.; Tang, H. Hierarchical Distribution-Based Exemplar Replay for Incremental SAR Automatic Target Recognition. IEEE Trans. Aerosp. Electron. Syst. 2025, 61, 6576–6588. [Google Scholar] [CrossRef]
  3. Zhang, X.; Zhang, S.; Sun, Z.; Liu, C.; Sun, Y.; Ji, K.; Kuang, G. Cross-sensor SAR image target detection based on dynamic feature discrimination and center-aware calibration. IEEE Trans. Geosci. Remote Sens. 2025, 63, 5209417. [Google Scholar] [CrossRef]
  4. Xia, X.; Meng, Z.; Han, X.; Li, H.; Tsukiji, T.; Xu, R.; Zhang, Z.; Ma, J. Automated driving systems data acquisition and processing platform. arXiv 2022, arXiv:2211.13425. [Google Scholar]
  5. Liao, Z.; Dong, X.; He, Q. Calculating the Optimal Point Cloud Density for Airborne LiDAR Landslide Investigation: An Adaptive Approach. Remote Sens. 2024, 16, 4563. [Google Scholar] [CrossRef]
  6. Guo, Y.; Wang, H.; Hu, Q.; Liu, H.; Liu, L.; Bennamoun, M. Deep learning for 3d point clouds: A survey. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 43, 4338–4364. [Google Scholar] [CrossRef]
  7. Gupta, A.; Jain, S.; Choudhary, P.; Parida, M. Dynamic object detection using sparse LiDAR data for autonomous machine driving and road safety applications. Expert Syst. Appl. 2024, 255, 124636. [Google Scholar]
  8. Li, Y.; Ren, H.; Yu, X.; Zhang, C.; Zou, L.; Zhou, Y. Threshold-free open-set learning network for SAR automatic target recognition. IEEE Sens. J. 2024, 24, 6700–6708. [Google Scholar] [CrossRef]
  9. Shi, W.; Rajkumar, R. Point-gnn: Graph neural network for 3d object detection in a point cloud. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 1711–1719. [Google Scholar]
  10. Zhang, Y.; Huang, D.; Wang, Y. PC-RGNN: Point cloud completion and graph neural network for 3D object detection. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtually, 2–9 February 2021; Volume 35, pp. 3430–3437. [Google Scholar]
  11. Yin, J.; Shen, J.; Gao, X.; Crandall, D.J.; Yang, R. Graph neural network and spatiotemporal transformer attention for 3D video object detection from point clouds. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 45, 9822–9835. [Google Scholar] [CrossRef]
  12. Yue, C.; Wang, Y.; Tang, X.; Chen, Q. DRGCNN: Dynamic region graph convolutional neural network for point clouds. Expert Syst. Appl. 2022, 205, 117663. [Google Scholar] [CrossRef]
  13. Zhang, R.; Wang, L.; Guo, Z.; Shi, J. Nearest neighbors meet deep neural networks for point cloud analysis. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 2–7 January 2023; pp. 1246–1255. [Google Scholar]
  14. Pei, Y.; Zhao, X.; Li, H.; Ma, J.; Zhang, J.; Pu, S. Clusterformer: Cluster-based transformer for 3d object detection in point clouds. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 2–3 October 2023; pp. 6664–6673. [Google Scholar]
  15. Zhang, M.; Ma, Y.; Li, J.; Zhang, J. A density connection weight-based clustering approach for dataset with density-sparse region. Expert Syst. Appl. 2023, 230, 120633. [Google Scholar] [CrossRef]
  16. Chen, G.; Wang, M.; Yang, Y.; Yuan, L.; Yue, Y. Fast and Robust Point Cloud Registration with Tree-based Transformer. In Proceedings of the 2024 IEEE International Conference on Robotics and Automation (ICRA), Yokohama, Japan, 13–17 May 2024; pp. 773–780. [Google Scholar]
  17. Pang, J.; Bui, K.; Tian, D. PIVOT-Net: Heterogeneous point-voxel-tree-based framework for point cloud compression. In Proceedings of the 2024 International Conference on 3D Vision (3DV), Davos, Switzerland, 18–21 March 2024; pp. 1270–1279. [Google Scholar]
  18. Ren, S.; Pan, X.; Zhao, W.; Nie, B.; Han, B. Dynamic graph transformer for 3D object detection. Knowl.-Based Syst. 2023, 259, 110085. [Google Scholar] [CrossRef]
  19. Vonessen, C.; Grötschla, F.; Wattenhofer, R. Next Level Message-Passing with Hierarchical Support Graphs. arXiv 2024, arXiv:2406.15852. [Google Scholar]
  20. Zhou, Y.; Tuzel, O. Voxelnet: End-to-end learning for point cloud based 3d object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 4490–4499. [Google Scholar]
  21. Yan, Y.; Mao, Y.; Li, B. Second: Sparsely embedded convolutional detection. Sensors 2018, 18, 3337. [Google Scholar] [CrossRef]
  22. Lang, A.H.; Vora, S.; Caesar, H.; Zhou, L.; Yang, J.; Beijbom, O. Pointpillars: Fast encoders for object detection from point clouds. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 12697–12705. [Google Scholar]
  23. Shi, S.; Wang, Z.; Shi, J.; Wang, X.; Li, H. From points to parts: 3d object detection from point cloud with part-aware and part-aggregation network. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 43, 2647–2664. [Google Scholar] [CrossRef]
  24. Deng, J.; Shi, S.; Li, P.; Zhou, W.; Zhang, Y.; Li, H. Voxel r-cnn: Towards high performance voxel-based 3d object detection. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtually, 2–9 February 2021; Volume 35, pp. 1201–1209. [Google Scholar]
  25. Ye, M.; Xu, S.; Cao, T. Hvnet: Hybrid voxel network for lidar based 3d object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 1631–1640. [Google Scholar]
  26. Zheng, W.; Tang, W.; Chen, S.; Jiang, L.; Fu, C.W. Cia-ssd: Confident iou-aware single-stage object detector from point cloud. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtually, 19–21 May 2021; Volume 35, pp. 3555–3562. [Google Scholar]
  27. Mao, J.; Xue, Y.; Niu, M.; Bai, H.; Feng, J.; Liang, X.; Xu, H.; Xu, C. Voxel transformer for 3d object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Virtually, 2–9 February 2021; pp. 3164–3173. [Google Scholar]
  28. Yang, Y.Q.; Guo, Y.X.; Xiong, J.Y.; Liu, Y.; Pan, H.; Wang, P.S.; Tong, X.; Guo, B. Swin3d: A pretrained transformer backbone for 3d indoor scene understanding. Comput. Vis. Media 2025, 11, 83–101. [Google Scholar] [CrossRef]
  29. Shi, S.; Wang, X.; Li, H. Pointrcnn: 3d object proposal generation and detection from point cloud. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 770–779. [Google Scholar]
  30. Qi, C.R.; Yi, L.; Su, H.; Guibas, L.J. Pointnet++: Deep hierarchical feature learning on point sets in a metric space. Adv. Neural Inf. Process. Syst. 2017, 30, 5105–5114. [Google Scholar]
  31. Yang, Z.; Sun, Y.; Liu, S.; Jia, J. 3dssd: Point-based 3d single stage object detector. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 11040–11048. [Google Scholar]
  32. Zhang, Y.; Hu, Q.; Xu, G.; Ma, Y.; Wan, J.; Guo, Y. Not all points are equal: Learning highly efficient point-based detectors for 3d lidar point clouds. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 18953–18962. [Google Scholar]
  33. Yang, Z.; Sun, Y.; Liu, S.; Shen, X.; Jia, J. STD: Sparse-to-Dense 3D Object Detector for Point Cloud. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 1951–1960. [Google Scholar] [CrossRef]
  34. He, C.; Zeng, H.; Huang, J.; Hua, X.S.; Zhang, L. Structure aware single-stage 3d object detection from point cloud. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 11873–11882. [Google Scholar]
  35. Shi, S.; Guo, C.; Jiang, L.; Wang, Z.; Shi, J.; Wang, X.; Li, H. Pv-rcnn: Point-voxel feature set abstraction for 3d object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 10529–10538. [Google Scholar]
  36. Sheng, H.; Cai, S.; Liu, Y.; Deng, B.; Huang, J.; Hua, X.S.; Zhao, M.J. Improving 3d object detection with channel-wise transformer. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 2743–2752. [Google Scholar]
  37. Mao, J.; Niu, M.; Bai, H.; Liang, X.; Xu, H.; Xu, C. Pyramid r-cnn: Towards better performance and adaptability for 3d object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 2723–2732. [Google Scholar]
  38. Li, Y.; Ma, L.; Tan, W.; Sun, C.; Cao, D.; Li, J. GRNet: Geometric relation network for 3D object detection from point clouds. ISPRS J. Photogramm. Remote Sens. 2020, 165, 43–53. [Google Scholar] [CrossRef]
  39. Chen, J.; Lei, B.; Song, Q.; Ying, H.; Chen, D.Z.; Wu, J. A hierarchical graph network for 3d object detection on point clouds. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 392–401. [Google Scholar]
  40. Cheng, H.; Zhu, J.; Lu, J.; Han, X. EDGCNet: Joint dynamic hyperbolic graph convolution and dual squeeze-and-attention for 3D point cloud segmentation. Expert Syst. Appl. 2024, 237, 121551. [Google Scholar] [CrossRef]
  41. He, Q.; Wang, Z.; Zeng, H.; Zeng, Y.; Liu, Y. Svga-net: Sparse voxel-graph attention network for 3d object detection from point clouds. In Proceedings of the AAAI Conference on Artificial Intelligence, Online, 22 February–1 March 2022; Volume 36, pp. 870–878. [Google Scholar]
  42. Sun, Z.; Leng, X.; Zhang, X.; Zhou, Z.; Xiong, B.; Ji, K.; Kuang, G. Arbitrary-direction SAR ship detection method for multi-scale imbalance. IEEE Trans. Geosci. Remote Sens. 2025, 63, 5208921. [Google Scholar]
  43. Wang, R.; Wang, Z.; Zhang, G.; Kang, H.; Luo, F. KD-Net: A novel knowledge-and data-driven network for SAR target recognition with limited data. In Proceedings of the IET International Radar Conference (IRC 2023), IET, Chongqing, China, 3–5 December 2023; Volume 2023, pp. 2623–2627. [Google Scholar]
  44. Han, K.; Xiao, A.; Wu, E.; Guo, J.; Xu, C.; Wang, Y. Transformer in transformer. Adv. Neural Inf. Process. Syst. 2021, 34, 15908–15919. [Google Scholar]
  45. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  46. Zhou, W.; Jiang, W.; Zheng, Z.; Li, J.; Su, T.; Hu, H. From grids to pseudo-regions: Dynamic memory augmented image captioning with dual relation transformer. Expert Syst. Appl. 2025, 273, 126850. [Google Scholar] [CrossRef]
  47. Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; Zagoruyko, S. End-to-end object detection with transformers. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 213–229. [Google Scholar]
  48. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 10012–10022. [Google Scholar]
  49. Yu, X.; Tang, L.; Rao, Y.; Huang, T.; Zhou, J.; Lu, J. Point-bert: Pre-training 3d point cloud transformers with masked point modeling. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 19313–19322. [Google Scholar]
  50. Pan, X.; Xia, Z.; Song, S.; Li, L.E.; Huang, G. 3d object detection with pointformer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 19–25 June 2021; pp. 7463–7472. [Google Scholar]
  51. Ning, Y.; Cao, J.; Bao, C.; Hao, Q. DVST: Deformable voxel set transformer for 3D object detection from point clouds. Remote Sens. 2023, 15, 5612. [Google Scholar] [CrossRef]
  52. Xu, X.; Dong, S.; Xu, T.; Ding, L.; Wang, J.; Jiang, P.; Song, L.; Li, J. Fusionrcnn: Lidar-camera fusion for two-stage 3d object detection. Remote Sens. 2023, 15, 1839. [Google Scholar] [CrossRef]
  53. Chen, Z.; Pham, K.T.; Ye, M.; Shen, Z.; Chen, Q. Cross-Cluster Shifting for Efficient and Effective 3D Object Detection in Autonomous Driving. In Proceedings of the 2024 IEEE International Conference on Robotics and Automation (ICRA), Yokohama, Japan, 13–17 May 2024; pp. 4273–4280. [Google Scholar]
  54. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
  55. Shi, S.; Jiang, L.; Deng, J.; Wang, Z.; Guo, C.; Shi, J.; Wang, X.; Li, H. PV-RCNN++: Point-voxel feature set abstraction with local vector representation for 3D object detection. Int. J. Comput. Vis. 2023, 131, 531–551. [Google Scholar] [CrossRef]
  56. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  57. Qi, C.R.; Liu, W.; Wu, C.; Su, H.; Guibas, L.J. Frustum pointnets for 3d object detection from rgb-d data. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 918–927. [Google Scholar]
  58. Vora, S.; Lang, A.H.; Helou, B.; Beijbom, O. Pointpainting: Sequential fusion for 3d object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June2020; pp. 4604–4612. [Google Scholar]
  59. Yoo, J.H.; Kim, Y.; Kim, J.; Choi, J.W. 3d-cvf: Generating joint camera and lidar features using cross-view spatial feature fusion for 3d object detection. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Proceedings, Part XXVII 16. Springer: Berlin/Heidelberg, Germany, 2020; pp. 720–736. [Google Scholar]
  60. Xie, T.; Wang, L.; Wang, K.; Li, R.; Zhang, X.; Zhang, H.; Yang, L.; Liu, H.; Li, J. FARP-Net: Local-global feature aggregation and relation-aware proposals for 3D object detection. IEEE Trans. Multimed. 2023, 26, 1027–1040. [Google Scholar] [CrossRef]
  61. Yang, H.; Wang, W.; Chen, M.; Lin, B.; He, T.; Chen, H.; He, X.; Ouyang, W. Pvt-ssd: Single-stage 3d object detector with point-voxel transformer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 13476–13487. [Google Scholar]
  62. Noh, J.; Lee, S.; Ham, B. Hvpr: Hybrid voxel-point representation for single-stage 3d object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 19–25 June 2021; pp. 14605–14614. [Google Scholar]
  63. Wang, C.H.; Chen, H.W.; Chen, Y.; Hsiao, P.Y.; Fu, L.C. VoPiFNet: Voxel-pixel fusion network for multi-class 3D object detection. IEEE Trans. Intell. Transp. Syst. 2024, 25, 8527–8537. [Google Scholar] [CrossRef]
  64. Yin, T.; Zhou, X.; Krähenbühl, P. Multimodal virtual point 3d detection. Adv. Neural Inf. Process. Syst. 2021, 34, 16494–16507. [Google Scholar]
  65. Xu, S.; Zhou, D.; Fang, J.; Yin, J.; Bin, Z.; Zhang, L. Fusionpainting: Multimodal fusion with adaptive attention for 3d object detection. In Proceedings of the 2021 IEEE International Intelligent Transportation Systems Conference (ITSC), Indianapolis, IN, USA, 19–22 September 2021; pp. 3047–3054. [Google Scholar]
  66. Nagesh, S.; Baig, A.; Srinivasan, S.; Rangesh, A.; Trivedi, M. Structure Aware and Class Balanced 3D Object Detection on nuScenes Dataset. arXiv 2022, arXiv:2205.12519. [Google Scholar]
  67. Chen, Y.; Liu, J.; Zhang, X.; Qi, X.; Jia, J. Largekernel3d: Scaling up kernels in 3d sparse cnns. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June2023; pp. 13488–13498. [Google Scholar]
  68. Bai, X.; Hu, Z.; Zhu, X.; Huang, Q.; Chen, Y.; Fu, H.; Tai, C.L. Transfusion: Robust lidar-camera fusion for 3d object detection with transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 1090–1099. [Google Scholar]
  69. Shi, G.; Li, R.; Ma, C. Pillarnet: Real-time and high-performance pillar-based 3d object detection. In Proceedings of the European Conference on Computer Vision, Milan, Italy, 23–27 October 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 35–52. [Google Scholar]
  70. Zhou, Y.; Sun, P.; Zhang, Y.; Anguelov, D.; Gao, J.; Ouyang, T.; Guo, J.; Ngiam, J.; Vasudevan, V. End-to-end multi-view fusion for 3d object detection in lidar point clouds. In Proceedings of the Conference on Robot Learning, PMLR, Virtual, 16–18 November 2020; pp. 923–932. [Google Scholar]
  71. Ge, R.; Ding, Z.; Hu, Y.; Wang, Y.; Chen, S.; Huang, L.; Li, Y. Afdet: Anchor free one stage 3d object detection. arXiv 2020, arXiv:2006.12671. [Google Scholar]
  72. Wang, Y.; Fathi, A.; Kundu, A.; Ross, D.A.; Pantofaru, C.; Funkhouser, T.; Solomon, J. Pillar-based object detection for autonomous driving. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 18–34. [Google Scholar]
  73. Yin, T.; Zhou, X.; Krahenbuhl, P. Center-based 3d object detection and tracking. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 19–25 June 2021; pp. 11784–11793. [Google Scholar]
  74. Zhu, Z.; Meng, Q.; Wang, X.; Wang, K.; Yan, L.; Yang, J. Curricular object manipulation in lidar-based object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 1125–1135. [Google Scholar]
  75. Nie, M.; Xue, Y.; Wang, C.; Ye, C.; Xu, H.; Zhu, X.; Huang, Q.; Mi, M.B.; Wang, X.; Zhang, L. Partner: Level up the polar representation for lidar 3d object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 2–3 October 2023; pp. 3801–3813. [Google Scholar]
  76. Lu, B.; Sun, Y.; Yang, Z. Voxel graph attention for 3-D object detection from point clouds. IEEE Trans. Instrum. Meas. 2023, 72, 5023012. [Google Scholar] [CrossRef]
Figure 1. Illustration of the different structures of point cloud data. (a) Graph structure captures local neighborhoods through edges between nodes but has limited global representation. (b) Tree structure provides hierarchical organization and multi-scale analysis but may be lacking in local detail capture. (c) Tree–graph structure combines the strengths of both, enabling the effective integration of local and global features through bidirectional message passing.
Figure 1. Illustration of the different structures of point cloud data. (a) Graph structure captures local neighborhoods through edges between nodes but has limited global representation. (b) Tree structure provides hierarchical organization and multi-scale analysis but may be lacking in local detail capture. (c) Tree–graph structure combines the strengths of both, enabling the effective integration of local and global features through bidirectional message passing.
Remotesensing 17 02177 g001
Figure 2. Architecture of the proposed DA-TGCMP framework for LiDAR-based 3D object detection. The framework comprises four stages: (1) density-aware voxelization, where local point densities are estimated to reflect spatial variation; (2) graph construction using the proposed DPCG strategy to build robust local neighborhoods; (3) hierarchical tree construction via structured downsampling to capture multi-scale semantic context; and (4) tree–graph cross-message passing, which enables bi-directional feature interaction between tree and graph representations for enhanced feature learning.
Figure 2. Architecture of the proposed DA-TGCMP framework for LiDAR-based 3D object detection. The framework comprises four stages: (1) density-aware voxelization, where local point densities are estimated to reflect spatial variation; (2) graph construction using the proposed DPCG strategy to build robust local neighborhoods; (3) hierarchical tree construction via structured downsampling to capture multi-scale semantic context; and (4) tree–graph cross-message passing, which enables bi-directional feature interaction between tree and graph representations for enhanced feature learning.
Remotesensing 17 02177 g002
Figure 3. Bird’s eye view (BEV) of point cloud feature. (a) The point cloud density distribution after rasterization. (b) The result of dynamic point cloud grouping.
Figure 3. Bird’s eye view (BEV) of point cloud feature. (a) The point cloud density distribution after rasterization. (b) The result of dynamic point cloud grouping.
Remotesensing 17 02177 g003
Figure 4. Extra qualitative results achieved by our method on the validation set of the KITTI dataset. We also show the corresponding projected 3D bounding boxes on images. The Red boxes indicate detected bounding boxes, and green boxes denote ground truth annotations.
Figure 4. Extra qualitative results achieved by our method on the validation set of the KITTI dataset. We also show the corresponding projected 3D bounding boxes on images. The Red boxes indicate detected bounding boxes, and green boxes denote ground truth annotations.
Remotesensing 17 02177 g004
Figure 5. Qualitative comparison of 3D object detection results on highway and urban scenes. Blue circles highlight small or distant objects. Compared to SECOND and PV-RCNN++, our DA-TGCMP detects more small objects with more accurate and tighter bounding boxes. The blue circles highlight critical regions for comparative analysis.
Figure 5. Qualitative comparison of 3D object detection results on highway and urban scenes. Blue circles highlight small or distant objects. Compared to SECOND and PV-RCNN++, our DA-TGCMP detects more small objects with more accurate and tighter bounding boxes. The blue circles highlight critical regions for comparative analysis.
Remotesensing 17 02177 g005
Table 1. Performance comparison of the KITTI test set. The results are reported by the AP and 40 recall points.
Table 1. Performance comparison of the KITTI test set. The results are reported by the AP and 40 recall points.
ModalityMethodCar-3D.APPedestrian-3D.APCyclist-3D.AP
EasyModHardEasyModHardEasyModHard
Fusion-based
Camera + LiDARF-PointNet [57]82.1969.7960.5950.5342.1538.0872.2756.1249.01
PointPainting [58]88.9378.2777.4850.3240.9737.8777.6363.7855.89
3D-CVF [59]89.2080.0573.11
FusionRCNN [52]88.1281.9877.53
Point-based
LiDARPointRCNN [29]86.9675.6470.747.9839.3736.0174.9658.8252.53
Point-GNN [9]88.3379.4672.2951.9243.7740.1478.6063.4857.08
SA-SSD  [34]88.7579.7974.16
PV-RCNN [35]90.2581.4376.8252.1743.2940.2978.5763.7157.65
IA-SSD [32]88.3480.1375.0446.5139.0335.6078.3561.9455.70
FARP-Net [60]88.3681.5578.98
PVT-SSD [61]90.6582.2976.85
Voxel-based
LiDARSECOND [21]84.6575.9668.71
PointPillars [22]82.5874.3168.9951.4541.9238.8977.158.6551.92
Part- A 2  [23]87.8178.4973.5153.1043.3540.0679.1763.5256.93
CIA-SSD [26]89.5980.2872.87
HVPR [62]86.3877.9273.0453.4743.9640.64
VoPiFNet [63]88.5180.9776.7453.0747.4345.2277.6464.1058.00
DA-TGCMP91.0183.5679.1453.5347.9245.9681.6267.1561.73
Table 2. Performance comparison of the nuScenes 3D detection test set. C.V, Motor., Ped., and T.C. are abbreviations for construction vehicle, motorcycle, pedestrian, and traffic cones, respectively.
Table 2. Performance comparison of the nuScenes 3D detection test set. C.V, Motor., Ped., and T.C. are abbreviations for construction vehicle, motorcycle, pedestrian, and traffic cones, respectively.
MethodsModalitymAPNDSCarTruckC.VBusTrailerBarrierMotor.BikePed.T.C.
PointPainting [58]L + C46.458.177.935.815.836.237.360.241.524.173.362.4
3D-CVF [59]L + C52.762.383.045.015.948.849.665.951.230.474.262.9
MVP  [64]L + C66.470.586.858.526.167.457.374.870.049.389.185.0
FusionPainting [65]L + C68.171.687.160.830.068.561.771.874.753.588.385.0
PointPillars [22]L30.545.368.423.04.128.223.438.927.41.159.730.8
CBGS  [66]L52.863.381.148.510.554.942.965.751.522.380.170.9
LargeKernel3D [67]L65.370.585.955.326.866.260.274.372.546.685.680.0
Transfusion-L  [68]L65.570.286.256.728.266.358.878.268.344.286.182.0
PillarNet-34  [69]L66.071.487.657.527.963.663.177.270.142.387.383.3
DA-TGCMPL66.671.786.959.532.167.562.177.574.347.286.982.3
Table 3. Performance comparison on the Waymo Open Dataset with 202 validation sequences for 3D vehicle (IoU = 0.7), pedestrian (IoU = 0.5), and cyclist (IoU = 0.5) detection.
Table 3. Performance comparison on the Waymo Open Dataset with 202 validation sequences for 3D vehicle (IoU = 0.7), pedestrian (IoU = 0.5), and cyclist (IoU = 0.5) detection.
MethodVeh. (LEVEL_1)Veh. (LEVEL_2)Ped. (LEVEL_1)Ped. (LEVEL_2)Cyc. (LEVEL_1)Cyc. (LEVEL_2)
mAPmAPHmAPmAPHmAPmAPHmAPmAPHmAPmAPHmAPmAPH
SECOND [21]72.2771.6963.8563.3368.758.1860.7251.3160.6259.2858.3457.05
PointPillars [22]56.6259.25
MVF [70]62.9365.33
AFDet [71]63.6965.33
Pillar-based [72]69.8072.51
Part-A2-Net  [23]74.8274.3265.8865.4271.7663.6462.5355.367.3566.1565.0563.89
Voxel-RCNN  [24]75.5966.5972.51
CT3D [36]76.3069.0465.33
Pyramid-RCNN [37]76.375.6867.2366.68
CenterPoint [73]76.776.268.868.379.072.971.065.3
Curricular [74]72.1571.0464.6464.2973.6264.4765.8460.39
VoTr-TSD [27]74.9574.2565.9165.29
CenterFormer [21]72.2771.6963.8563.3368.758.1860.7251.3160.6259.2858.3457.05
PV-RCNN  [35]75.1774.666.3565.8472.6563.5263.4255.2967.2665.8264.8863.48
PARTNER [75]76.0575.5268.5868.11
PV-RCNN++ [55]76.1475.6268.0567.5673.9765.4365.6457.8268.3867.0665.9264.65
DA-TGCMP76.6476.6968.8668.2374.4666.3564.4357.5468.5267.3665.5264.78
Table 4. Results of the proposed DAGC, HTR, and TGCMP modules on the KITTI and nuScenes validation sets.
Table 4. Results of the proposed DAGC, HTR, and TGCMP modules on the KITTI and nuScenes validation sets.
ModuleKITTI (recall_40 mAP)nuScenes
EasyModHardNDSmAP
Baseline91.1982.6382.0165.40.592
Baseline + DAGC91.5684.8682.6166.70.622
Baseline + HTR91.3584.2482.2465.90.610
Baseline + DAGC + HTR92.5286.8782.5968.10.630
Baseline + DAGC + HTR + TGCMP92.6687.8282.7568.50.638
Table 5. Effects of DAGC module, fixed-radius, and kNN for DA-TGCMP. Results on the KITTI val set for car, pedestrian, and cyclist classes.
Table 5. Effects of DAGC module, fixed-radius, and kNN for DA-TGCMP. Results on the KITTI val set for car, pedestrian, and cyclist classes.
Fixed-Radius (r = 0.5)kNNDAGCCar-3D.AP (IoU = 0.7)Ped-3D.AP (IoU = 0.5)Cyc-3D.AP (IoU = 0.5)
EasyModHardEasyModHardEasyModHard
89.6379.5378.5568.4461.8755.5288.8775.4768.12
91.6180.3378.1269.1262.3558.8989.1276.7868.22
94.34 82.53 80.53 69.08 63.26 58.96 89.53 78.4 69.61
Table 6. Performance comparison on KITTI dataset across dense and sparse regions.
Table 6. Performance comparison on KITTI dataset across dense and sparse regions.
Method ρ dense ρ sparse
KITTI (recall_40 mAP)KITTI (recall_40 mAP)
Car. (0.7)Ped. (0.5)Cyc. (0.5)Car. (0.7)Ped. (0.5)Cyc. (0.5)
PointGNN  [9]93.2386.2189.0223.1212.367.15
VoxelGraph-RCNN [76]95.3988.4391.6025.0314.5910.34
DA-TGCMP95.7287.3192.3828.2615.1117.36
Table 7. Effect of different dynamic radius designs of small-object (pedestrian) detection on KITTI Val.
Table 7. Effect of different dynamic radius designs of small-object (pedestrian) detection on KITTI Val.
DesignRadius FormPedestrian AP (Mod)Remarks
Fixed Radius r = 0.3 56.21No adaptation to density
Linear Mapping r i = 0.1 + 0.4 ( 1 D i ) 60.82Balanced performance
Exponential r i = 0.1 + 0.4 exp ( D i ) 61.47Enhanced recall in sparse areas
Logarithmic r i = 0.1 + 0.4 log ( 1 + D i ) 59.35Conservative adaptation
Extended Range r i = 0.05 + 0.55 ( 1 D i ) 62.03Finer granularity in dense regions
Learnable Radius r i = MLP ( F i ) 63.40Task-adaptive, best performance
Table 8. Results on the KITTI and nuScenes validation sets using different baseline methods.
Table 8. Results on the KITTI and nuScenes validation sets using different baseline methods.
Baseline MethodKITTI (recall_40 mAP)nuScenes
EasyModHardNDSmAP
Single-Scale Voxel90.1081.2277.3661.420.517
PointNet-Style Global Pooling91.1181.2579.5362.90.568
Hierarchical Tree Structure91.5384.3383.2765.80.593
Table 9. Performance comparison on the KITTI val playing the role of top-down message passing and bottom-up message passing.
Table 9. Performance comparison on the KITTI val playing the role of top-down message passing and bottom-up message passing.
Message Passing MethodCar.APPedestrian.APCyclist.AP
Mod (IoU = 0.7)Mod (IoU = 0.5)Mod (IoU = 0.5)
Top-Down82.2362.5476.13
Bottom-Up79.6560.9873.45
Top-Down and Bottom-Up83.0562.8877.55
Table 10. Comparison of detection accuracy, memory usage, and computational cost across different architectures.
Table 10. Comparison of detection accuracy, memory usage, and computational cost across different architectures.
StructureCar. (Mod)Ped. (Mod)Cyc. (Mod)GFLOPsMem. (GB)
DAGC77.5659.6974.35−011.854
HTR78.3658.3771.60−2.8758.949
TGCMP82.5663.3178.38+3.58114.372
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhao, J.; Li, J.; Zhou, W.; Ren, H.; Long, Y.; Hu, H. Density-Aware Tree–Graph Cross-Message Passing for LiDAR Point Cloud 3D Object Detection. Remote Sens. 2025, 17, 2177. https://doi.org/10.3390/rs17132177

AMA Style

Zhao J, Li J, Zhou W, Ren H, Long Y, Hu H. Density-Aware Tree–Graph Cross-Message Passing for LiDAR Point Cloud 3D Object Detection. Remote Sensing. 2025; 17(13):2177. https://doi.org/10.3390/rs17132177

Chicago/Turabian Style

Zhao, Jingwen, Jianchao Li, Wei Zhou, Haohao Ren, Yunliang Long, and Haifeng Hu. 2025. "Density-Aware Tree–Graph Cross-Message Passing for LiDAR Point Cloud 3D Object Detection" Remote Sensing 17, no. 13: 2177. https://doi.org/10.3390/rs17132177

APA Style

Zhao, J., Li, J., Zhou, W., Ren, H., Long, Y., & Hu, H. (2025). Density-Aware Tree–Graph Cross-Message Passing for LiDAR Point Cloud 3D Object Detection. Remote Sensing, 17(13), 2177. https://doi.org/10.3390/rs17132177

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop