Next Article in Journal
Land Surface Temperature Retrieval from Fengyun-3D Medium Resolution Spectral Imager II (FY-3D MERSI-II) Data with the Improved Two-Factor Split-Window Algorithm
Previous Article in Journal
The Global Water Body Layer from TanDEM-X Interferometric SAR Data
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

HCNET: A Point Cloud Object Detection Network Based on Height and Channel Attention

School of Telecommunication Engineering, Xidian University, Xi’an 610100, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2021, 13(24), 5071; https://doi.org/10.3390/rs13245071
Submission received: 20 October 2021 / Revised: 9 December 2021 / Accepted: 10 December 2021 / Published: 14 December 2021

Abstract

:
The use of LiDAR point clouds for accurate three-dimensional perception is crucial for realizing high-level autonomous driving systems. Upon considering the drawbacks of the current point cloud object-detection algorithms, this paper proposes HCNet, an algorithm that combines an attention mechanism with adaptive adjustment, starting from feature fusion and overcoming the sparse and uneven distribution of point clouds. Inspired by the basic idea of an attention mechanism, a feature-fusion structure HC module with height attention and channel attention, weighted in parallel, is proposed to perform feature-fusion on multiple pseudo images. The use of several weighting mechanisms enhances the ability of feature-information expression. Additionally, we designed an adaptively adjusted detection head that also overcomes the sparsity of the point cloud from the perspective of original information fusion. It reduces the interference caused by the uneven distribution of the point cloud from the perspective of adaptive adjustment. The results show that our HCNet has better accuracy than other one-stage-network or even two-stage-network RCNNs under some evaluation detection metrics. Additionally, it has a detection rate of 30FPS. Especially for hard samples, the algorithm in this paper has better detection performance than many existing algorithms.

1. Introduction

In the 21st century, automatic driving has gradually broken through the limitation of hardware, which has sped up its research process. Its application and popularization will bring great changes to human society, such as the transformation of public transportation, infrastructure, and urban appearance. Currently, there are numerous high-tech companies, automobile manufacturers, and startup companies working on autonomous-driving technologies in order to build a smarter and safer system.
As one of the key technologies of automatic driving, object detection has made great progress in recent years [1]. Before the deployment of automatic driving by LiDAR, the camera was the main sensor. Since the image data can perceive color information, it can play a vital role in tasks such as traffic light recognition. However, the lack of accurate three-dimensional information cannot meet the safety requirements of autonomous driving scenarios.
In order to meet the needs of three-dimensional information for target detection, new technologies represented by LiDAR have made major breakthroughs in the real-time acquisition of multi-grade three-dimensional spatial targets in recent years. This system can partially block through the woods to directly obtain high-precision three-dimensional information on the real surface, which cannot be replaced by traditional photogrammetry. Therefore, the sources of information of the autonomous driving perception system [2] are mainly images and point cloud data.
The high-precision point cloud data generated by LiDAR makes up for this shortcoming, as it can obtain more detailed target-shape and three-dimensional position information and provide a more robust perception effect. On one hand, how to use point cloud information more effectively has become a recent research hotspot; on the other hand, as point clouds are usually disordered, sparse, and uneven, the task of point cloud target detection still poses a great challenge.
In order to interpret and understand a 3D point cloud, deep learning has become the mainstream approach in the field of object detection. Based on the two-dimensional deep-learning research, including images and videos, researchers are gradually expanding the powerful tool of deep neural networks to 3D point clouds.
Due to the irregular format, one of the biggest challenges in designing learning algorithms is to develop efficient data structures to represent 3D point clouds. To better apply mature deep-learning tools to the point cloud field, most methods first convert the disordered point cloud data into a regular voxel grid and then use 3D convolution [3] for processing.
In order to directly process the original point cloud, PointNet [4,5] takes the lead in using point level multi-layer perceptrons (MLPs) and maximum pooling to ensure the invariance of the arrangement. Following that, a series of 3D deep-learning methods are based on PointNet.
As the most representative point cloud target-detection network, VoxelNet [6] lays the foundation for the voxel-based point cloud target-detection algorithms. However, due to the high computational cost of 3D convolution, it is difficult to run in real-time. Thus, Yan et al. introduced sparse 3D convolution to reduce the computational cost of the network in SECOND [7].
Later, a new network, PointPillars [8], proposed to use the concept of pillars to regularize the point cloud data, to use PointNet to learn the features in the pillars, and then to project the point cloud into a 2D pseudo image. However, due to the limited resolution, the division method of pillars would cause a certain degree of information loss.
This paper proposes a new point cloud object detection network, HCNet, based on a 2D convolutional [9] network. To avoid the high computational cost of 3D convolution [10,11] and reduce the coding and quantization loss of the PointPillars method, we propose a coding method that can generate multi-layer pseudo images based on the attention mechanism. Specifically, for the divided voxel grid, we first gather them at the same height to generate multiple pseudo images containing point cloud information of different heights. Through the dual attention weighting processing of the height dimension and the channel dimension, we merge multiple pseudo images into a feature map.
Simultaneously, to overcome the sparseness and uneven distribution of the point cloud, we propose an adaptively adjusted detection head that, first, compensates for the sparseness of the point cloud by supplementing the original feature information. A distance-based adaptive weighting strategy is proposed to overcome the uneven distribution of the point cloud, so that the network can focus on key information.
The main contributions of this paper are as follows:
  • The proposed HC module: This avoids the loss of information caused by quantifying point cloud data while enhancing the ability to express features;
  • A self-adjusting detection head: This overcomes the impact of the sparseness and uneven distribution of the point cloud on the object detection task to a certain extent;
  • The proposed algorithm has an inference speed of 30 fps and is comparable to the performance of the most advanced methods.

2. Related Work

Traditional point cloud perception methods usually use Euclidean clustering [12] or region generation [13] to group point clouds, and use ground-culling filtering [14] and map filtering algorithms [15] to increase the detection accuracy. In recent years, with the rapid development of deep learning, the research of point cloud target detection, based on deep learning, has also burst into new vitality. The existing deep-learning target-detection methods are divided mainly into single-stage networks and two-stage networks.

2.1. Two-Stage Network

The two-stage [16] method first proposes several regions [17] that contain objects, and then determines the category label of each proposal by extracting the features in the regions [18]. Chen [19] et al. have generated a set of high-precision 3D candidate frames from the BEV map and projected them onto the feature maps of multiple views. They then combined the regional features of multiple views to predict the oriented 3D bounding box. Shi et al. put forth a PointRCNN [20] framework. Specifically, they segmented the three-dimensional point cloud directly to obtain the front scenic spot, and then fused the semantic features and local spatial features to generate high-quality 3D bounding boxes. They performed pioneering work on 3D object detection using graph convolutional networks based on PointRCNN. The F-PointNET [21] uses existing 2D object detectors to generate 2D candidate regions of the object and then extracts 3D frustum proposals for each 2D candidate region. The PointNet is applied to learn the point cloud features of each three-dimensional frustum to realize the estimation of the 3D bounding box.

2.2. Single-Stage Network

The single-stage method [22] does not require region-proposal generation. It uses a single-stage network directly to predict the target category probability and return to the 3D bounding box [23] of the object. Yang et al. [24] discretized the point cloud with equally spaced units and, similarly, encoded the reflectivity to obtain a regular representation. They then applied a full convolutional network to estimate the target’s position and heading angle. This method is not only superior to the single-stage methods (including VeloFCN [25], 3D-FCN [26], and Vote3Deep [27]) in precision, but it is also greatly improved in speed. Zhou et al. proposed a voxel-based end-to-end trainable framework named VoxelNet. They segmented the point cloud into equally spaced voxels and encoded the features within each voxel into 4D tensors. They connected a region proposal network to produce test results. Although this framework achieved good performance, this method is very slow due to the sparsity of voxels and 3D convolution. Later, Yan et al. used sparse convolutional networks [28] to improve inference efficiency. Based on this network, Lang et al. proposed a 3D object detector called PointPillars, which uses PointNet to learn point cloud features in vertical columns and encode the learned features into a pseudo image. Later, the pipeline [29] of 2D target detection is applied to predict the 3D bounding box.

2.3. Attention Mechanism

Human visual attention enables us to pay more attention to specific areas in the picture, then gradually adjust the focus, move to other specific areas, and finally, infer the information of the whole picture.
In the various elements given in Figure 1, the red area indicates where attention is more focused. People will pay more attention to people’s faces, the title of the text, and the first sentence of the article. This visualization shows how human beings efficiently allocate limited attention resources when viewing an image.
In order to allow the computer to choose the information that is more critical to the current task goal from a large amount of information, researchers have transferred this idea to deep learning.
The attention mechanism [30] in a neural network is a mechanism that allows the network to learn to focus on key information and ignore irrelevant information. It can make the neural network [31] focus on a subset of its inputs. The attention mechanism module can be divided into spatial domain attention [32], channel domain attention [33], and hybrid domain attention [34]. These classic divisions inspired us to apply the attention mechanism to the feature fusion of multi-layer pseudo images. For example, Point-SENet [35] utilized the channel attention of the SE module to predict the 3D detection box. Point2Sequence [36] proposed an attention-based sequence-to-sequence structure to aggregate the sequential features of multi-scale areas.

3. HCNet

In this section, we introduce HCNet, which outputs accurate classification and 3D bounding boxes based on the LiDAR point cloud.
The HCNet is a single-stage end-to-end trainable network. As shown in Figure 2, HCNet consists of three blocks:
(1)
The HC module fully aggregating the original features of the point cloud through the dual attention of height dimension and channel dimension;
(2)
a backbone network that uses a 2D CNN to extract features;
(3)
an adaptive detection-head network that adjusts the feature map and then outputs the labels and bounding boxes of the objects.

3.1. HC Module

In consistence with the existing voxel-based methods, we first divide the point cloud space into voxel grids. For the three-dimensional point cloud space with the range of (D, H, W), we divide it multiple times at equal intervals from three dimensions. Firstly, it is divided into N and M times in the horizontal direction (distance), and then P times in the vertical direction (height). The obtained voxel size is D′ = D/N, H′ = H/M, and W′ = W/P. According to the division position, each voxel has a division coordinate (x, y, z). We used a strategy similar to PointNet to extract the features within the voxel and aggregated the features at the same height into pseudo images according to the coordinate z to generate pseudo images.
In Figure 3, taking P = 3 as an example, the amount of vertical cutting can be adjusted according to different needs.
The HC module aggregates multiple pseudo images by introducing an attention mechanism to focus on the point cloud features in different height ranges. Its structure is shown in Figure 4, which consists of three parts: Feature Aggregation Block, HC Block, and Attention-Weighted Block.

3.2. Feature Aggregation Block

The feature aggregation part is used mainly to aggregate the features of the pseudo images.
Here, the un-squeeze operation is used first to expand the size of each pseudo image to (B, C, H, W, 1), where B represents the batch size, H and W represent the height and width of the feature map, C represents the number of channels, and 1 represents the height dimension from the extension. Then, we use the concat operation to splice the N feature maps to obtain a tensor P of size (B, C, H, W, N).

3.3. HC Block

The HC block realizes the fusion of multi-layer feature maps mainly through two parallel attention mechanisms. It contains two branches: a height attention branch (Figure 5) and a channel attention branch (Figure 6).

3.3.1. Height Attention Branch

Given a tensor P, we perform max-pooling to aggregate the features across channel-wise dimensions. The output size is (B, 1, H, W, N) for the channel-wise feature. Then, the reshape operation is used to generate an attention map F h of size (B, H, W, N). To explore the feature correlation between different heights, the height attention branch uses a convolution module to generate the final height global response. Assuming that F i n is the feature map output by the max-pooling operation, the height attention branch can be expressed by (1). The size of the output F a t t h will be (B, H, W, N).
F in = max ( P ) F h = Re s h a p e ( F i n ) F a t t h = Re l u ( C o n v a t t h ( F h ) )
Here, the operation of max ( ) means that the max-pooling operation is performed to aggregate the height features across their channel-wise dimensions, the operation Re s h a p e ( ) is for reshaping the feature map into (B, H, W, N), Re l u ( ) represents the ReLU activation function, and C o n v ( ) represents the convolution module, which contains two 1 × 1 convolutional layers.

3.3.2. Channel Attention Branch

The role of the channel attention branch is to enable the network to focus on more important channel features and reduce focus on unimportant channels. Primarily, given the input tensor P, it first needs to pass the max-pooling operation to complete the feature aggregation by height and output the feature map of size (B, C, H, W). Then, a global channel response is also generated through the convolution module. The channel attention can be formulated as represented in (2). The size of the F a t t c would be (B, C, H, W).
F in = max ( P ) F c = Re s h a p e ( F i n ) F a t t c = Re l u ( C o n v a t t c ( F c ) )
Here, the operation of max ( ) means that the max-pooling operation is performed to aggregate the height features across their height-wise dimensions, the operation Re s h a p e ( ) is for reshaping the feature map into (B, H, W, C), Re l u ( ) represents the ReLU activation function, and C o n v ( ) represents the convolution module, which contains two 1 × 1 convolutional layers.

3.4. Attention-Weighted Block

The attention-weighted block mainly uses F a t t h and F a t t c to complete the attention adjustment of the feature tensor. It realizes the dual adjustment of the height dimension and the channel dimension, and then obtains richer feature information. A max-pooling operation is used to aggregate the feature from the backbone. The attention-weighted block can be formulated as given in (3). The size of the F o u t will be (B, C, H, W).
F a t t h c = s i g m o i d ( F a t t h F a t t c ) F = F a t t h c P F o u t = max ( F )
Here, ⨂ represents the matrix multiplication, s i g m o i d ( ) represents the sigmoid function employed to normalize the values of the attention matrix to a range of (0,1), the second equation represents the element-wise multiplication (that is, the completion of the tensor P in the height and channel dimensions’ parallel adjustment), and max ( ) means the max-pooling operation preformed to aggregate the features across their height-wise dimensions.

3.5. Backbone

We use a similar backbone as the YOLO [37], and the structure is shown in Figure 7.
The backbone has three parts:
(1)
the top-down part, which generates features at smaller and smaller resolutions;
(2)
the down-top part, which up-samples the feature map from the bottom up and stitches it with the features of the upper layer;
(3)
the third part, which up-samples the features of each layer of the network to the same size through up-sampling, and then splices to generate the final output features.
The top-down part consists of a series of convolutional blocks; each block contains a convolutional layer with a convolution kernel size of 3 × 3, each followed by BatchNorm and a ReLU. The down-top part comprises a series of deblocks. Each deblock contains a 2D transposed convolution with a 3 × 3 convolution kernel used to adjust the size of the current feature map to the same size as the previous feature. The final feature maps from each down-top deblock are combined through up-sampling and concatenation to generate the input of the detection head.

3.6. Detection Head

In this paper, we propose an adaptively adjusted detection head that enables the network to compensate for the detailed information and adaptively adjust the sparse feature map, thereby improving the object detection performance of the network structure.
This section proposes an adaptively adjusted detection head that includes three parts: a detection head backbone network, an original information-fusion module, and an adaptive-adjustment module. The network structure in this section is presented in Figure 8. The original information-fusion module mainly weakens the sparsity of the point cloud from the perspective of compensating for information, whereas the adaptive-adjustment module implements adaptive adjustment on the feature map to avoid the uneven distribution.

3.6.1. Detection Head Backbone Network

The main frame of the detection head, designed in this section, uses a top-down multi-scale network, which can enhance the ability of the network model through the fusion of information at different scales.
As shown in Figure 7, the backbone network consists of a block module and a deblock module, and the results of the two modules are aggregated through the concat operation. Thus, the network has features that contain different receptive fields. By merging these multi-scale features, a higher quality target-prediction regression effect is achieved.

3.6.2. Original Information-Fusion Module

This mainly incorporates the features of the original point cloud pseudo images into the input features of the block and deblock modules of the backbone network of the detection head. Then, the merged features are adjusted from the channel dimension and space dimension.
First, through the convolution operation, the sizes of the two layers of the pseudo feature maps are adjusted to the same size as the input features of the block module and the deblock module, and the information is then aggregated through the concat operation. A similar strategy is also used by CBAM [38] to perform feature weighting.

3.6.3. Adaptive-Adjustment Module

The adaptive-adjustment module adjusts the width and height of the feature map so that the information in the key area of the feature map can be emphasized. Thus, the network can pay attention to the sparse object point cloud information from a distance. First, we aggregate the features through strip pooling. For an input feature map F in R H × W , with height H and width W, the pooling-window size of the strip pooling is usually H × 1 or 1 × W. While performing feature aggregation in the width direction of the feature map, its pooling window should be set to 1 × W, and the specific pooling process can be expressed as (4).
y i h = 1 W 0 j W x i , j
Here, y i h is the feature pooled by the horizontal pooling window representing the aggregate feature of the current row.
Similarly, when performing strip pooling in the height direction of the feature map, the pooling window should be set to H × 1, and the specific calculation process is as shown in (5).
y j w = 1 H 0 i H x i , j
Here, y j w is the feature pooled by the vertical pooling window, which represents the aggregate feature of the current column.
Through the horizontal and vertical strip-pooling operations, the features in the strip area can be coded collectively, and the key, detailed features can be captured at the same time. We thus generate row-compression information of size H × 1, and column-compression information of size 1 × W.
Then, the convolution component is used to extract the adaptive information, in both horizontal and vertical directions, from the two sets of obtained compressed information. The basic structure of the convolution component comprises a 2D convolution layer, a BatchNorm layer, and a ReLU function layer. Consequently, the weight adjustment of the feature map is realized by multiplying element by element. F i n and F out , respectively, represent the adaptively adjusted input and output feature maps. F in s t r i p represents the feature map after the strip-pooling process. F a s t r i p represents the feature map after convolution and up-sampling.
The overall calculation process can be expressed as (6).
F in s t r i p = s t r i p ( F i n ) F a s t r i p = u p ( C o n v ( F i n s t r i p ) ) F o u t = F a s t r i p F i n
In the above formula, s t r i p ( ) represents the strip pooling process. The function C o n v ( ) uses convolution to extract the hidden information from the pooled features, while the function u p ( ) represents an up-sampling operation used to generate global adaptive information of the same size.

3.7. Loss Function

We use the same loss function as SECOND. This further consists of three different kinds of loss functions: regression loss, classification loss, and directional loss.
Regression loss function—In 3D target detection, a 3D bounding box is usually defined by the parameters ( x , y , z , h , w , l , θ ) . The residual positioning coding of the ground truth and anchors are defined by (7).
Δ x = x g t x a d a , Δ y = y g t y a d a , Δ z = z g t z a d a Δ w = log w g t w a , Δ l = log l g t l a , Δ h = log h g t h a Δ θ = sin ( θ g t θ a ) , d a = ( w a ) 2 + ( l a ) 2
The regression loss function is expressed as (8).
L l o c = b ( x , y , z , w , l , h , θ ) S m o o t h L 1 ( Δ b )
Classification loss function—Similar to the existing methods, we also use the focal loss to handle the class imbalance problem. The classification loss is formulated as (9):
L c l s = α a ( 1 p a ) γ log p a
where P α is the class probability of an anchor. We use the original paper’s settings α = 0.25 and γ = 2.
Directional loss function—To enable the network to learn the direction information of the bounding box, SoftMax is used as the direction loss function L d i r .
By combining all the loss functions, the total loss is defined as (10).
L = 1 N p o s ( β l o c L l o c + β c l s L c l s + β d i r L d i r )
Here, N p o s represents the number of positive matching boxes, and the loss weights β l o c , β c l s , and β d i r are empirically set to 2.0, 1.0, and 0.2, respectively.

4. Implementation Details

4.1. Dataset

All experiments in this study are tested with the KITTI [39] dataset as a benchmark. It contains real image data collected from city scenes, rural scenes, road scenes, and so on. There are up to 15 vehicles and 30 pedestrians in each image, as well as various degrees of occlusion and truncation (Figure 9).
KITTI has three categories: cars, pedestrians, and bicycles. Our dataset also contains object labels in the form of 3D tracklets, and we provide online benchmarks for stereo, optical flow, object detection, and other tasks.
The KITTI dataset comprises 7481 training samples and 7518 test samples. Each training sample contains cloud data and the corresponding RGB image data. Additionally, for each target category, there are three difficulty levels: easy, medium, and difficult. In this work, we focus only on the most challenging car categories.

4.2. Setting

We set the range of (D, H, W) to ((0, 70.4), (−40, 40), (−3, 1)), and the number of voxel divisions, in the three directions, to N = 496, M = 432, and P = 4 to generate voxels of size 90.16, 0.16, 10. The car anchor has a (width, length, height) of (1.6, 3.9, 1.5) m. additionally, matching uses positive and negative thresholds of 0.6 and 0.45.

4.3. Data Augmentation

Data augmentation is essential to obtain good performance on the KITTI dataset. Inspired by PointPillars, our expansion strategy is as follows.
First, a database is generated from the training dataset containing the complete ground truth and their corresponding point cloud data.
Then, 15 ground truth samples are selected randomly from this database during the training process and introduced into the current training point cloud.
Moreover, to consider the influence of noise, we need to transform each ground truth and point cloud datum independently and randomly. Each ground truth is rotated randomly. The degree of rotation obeys a uniform distribution (−π/2, π/2).
Finally, the entire point cloud performs global scaling and a rotation operation. The degree of scaling obeys the uniform distribution (0.95, 1.05), whereas the degree of rotation obeys the uniform distribution (−π/4, π/4).

5. Experiments

5.1. Evaluation Criteria

This section describes the Intersection-over-Union (IoU), precision rate (Precision), recall rate (Recall), and average accuracy (AP). The IoU measures the degree of coincidence between the network prediction target box and the original real target box, which can be expressed in (11), which is the ratio between the intersection between the prediction box and the prediction box and the real box.
I o U = Detectionresult GroundTruth Detectionresult GroundTruth
The role of the IoU is to help resolve the correctness of the detection results, and the general algorithm sets a threshold of T that is considered correct when the IoU > T is otherwise judged incorrect. On this basis, the prediction results can be divided into the following four categories:
(1)
TP (True positives): Positive samples were identified as positive samples;
(2)
TN (True negatives): Negative samples were identified as negative samples;
(3)
FP (False positives): Negative samples were misidentified as positive samples;
(4)
FN (False negatives): Positive samples were misidentified as negative samples.
On the basis of the IoU, the definitions of Precision and Recall can be given. The calculation process of Precision is shown in (12), which is actually among all the predicted targets, predicting the correct ratio.
P recision = TP T P + F P
The computational procedure of Recall, as shown in (13), is the proportion of samples correctly identified as positive samples out of all positive samples in the test set.
Recall = TP T P + F N
AP, which can be calculated using Precision and Recall, is actually the area under the Precision–Recall curve. Generally speaking, the better the target detection model is, the higher its AP value will be. To contrast this with existing algorithms, AP is also used as the main detection index.

5.2. Evaluation Details

We evaluated our HCNet on the KITTI validation set and the test set, to verify the superiority of our proposed model that only uses point clouds for 3D target detection, in Table 1.
We compare our HCNet with other existing algorithms performed on the KITTI validation set (Figure 10).
The algorithms to be compared include not only classic single-stage methods such as VoxelNet, SECOND, and PointPillars, but also two-stage methods with superior performance, such as the PointRCNN. Moreover, to compare our HCNet with other methods, we also submitted the results to the official KITTI server. The result of BEV and 3D are given in Table 2.
It is worth noting that our method achieved very good results on the Hard difficulty level, and achieved good performance on both BEV and 3D. For BEV, we achieved comparable performance to SAPRNET on the Moderate (Mod) difficulty level, and achieved the best performance on the Easy and Hard difficulty levels. For 3D, we demonstrate good performance on the Hard level, but the results are poor for the Easy and Mod levels. The possible reasons are as follows:
(1)
Our HC module focuses on more challenging tasks;
(2)
Our adaptive detection head overemphasizes the features of difficult objects in the distance.
In addition, our method has also achieved excellent results on AOS, which shows that our algorithm also has strong performance in the prediction of orientation (Figure 11), which is very critical for autonomous driving tasks.
We conducted multiple ablation experiments on the KITTI validation set to analyze the performance of different components of the HCNet. The training set used is the KITTI training set, the test set is the KITTI validation set, and the experimental platform is a computer with a GTX 2080Ti.

5.2.1. Different Attention Mechanisms in the HC Module

We compared the performance of multiple attention branches in the HC block on 3D target detection tasks. Channel attention (CA), height attention (HA), and height–channel attention (HCA) are shown in the Table 3.
It can be seen from Figure 12 that when fusing multiple feature maps, channel attention achieved better performance than high attention. Additionally, the HCA, which combines the two attention mechanisms in parallel, achieved the best detection results. This observation shows that the HC module we proposed can focus on the key features through weighted adjustments to the two dimensions of channel and height. By improving the information-expression ability of the feature map, the object-detection performance of the network is also improved.

5.2.2. Different Parts of the Detection Head

We have the original information-fusion module (OIFM) and the adaptive-adjustment module (AAM) in the adaptive detection head. Ablation experiments were carried out to test their effect on the detection results.
From the results in the Table 4, it can be seen that both the OIFM and AAM help improve the detection effect, and the effect of OIFM is even more prominent. Additionally, the best detection performance can be achieved by connecting two modules serially. These facts show that our adaptive detection head overcomes the sparseness of the point cloud through the fusion of original information and makes up for the uneven distribution of the point cloud through the adaptive adjustment of the feature map in two directions.

5.3. Visualization

In order to show the detection results more visually, this section uses the MayaVi library to visualize the detection results of the network. MayaVi library is a general purpose, cross-platform tool for 3D scientific data visualization.
We selected two examples (Figure 13 and Figure 14). They were obtained from the easy and hard datasets, respectively. The image on the left are the original point cloud, and the image on the right are the detection results. The white dots are formed by the LiDAR emission on the surface of the object, the green box is the target, and the number beside it is the confidence (Figure 15). Confidence means the probability that the current predicted box is the correct box.
As seen in the RGB diagram of Figure 13, this is a basic pavement scene. A good spacing remained between vehicles, with no occlusion and truncation to facilitate the measurement of the underlying performance of the algorithm.
The RGB of Figure 14 shows that various vehicles with different directions are parked on both sides of the road. High occlusion was highly truncated between each vehicle. This is very challenging for the target-detection task.

6. Conclusions

We propose a single-stage point cloud target-detection algorithm HCNet, and design the HC module through the attention mechanism to complete the feature fusion of multi-layer pseudo images and retain the original point cloud information to a greater extent. Using the parallel attention-mechanism module, the feature adjustment of the height dimension and the channel dimension is completed simultaneously, and the information-expression ability of the feature map is enhanced.
Moreover, we also designed an adaptively adjusted detection-head network for the sparsity of the point cloud and the uneven distributions of near-dense and far-field. On the one hand, the original information-fusion module is used to complete the information supplement of the feature map, thereby overcoming the sparsity of the point cloud; whereas, on the other hand, through the adaptive-adjustment module, the information in the key areas of the feature map are emphasized.
Thus, the network can pay attention to the target point cloud information that is scarce at a distance. The application of HCNet on KITTI shows that, compared to many existing algorithms, the HCNet proposed in this paper has achieved better detection results.

Author Contributions

Conceptualization, J.Z. and J.W.; methodology, J.Z. and J.W.; software, J.Z., J.W., D.X. and Y.L.; validation, J.W. and D.X.; formal analysis, J.Z.; investigation, J.W. and D.X.; resources, J.Z.; data curation, J.W. and D.X.; writing—original draft preparation, J.W.; writing—review and editing, J.Z., J.W. and D.X.; visualization, J.W. and D.X.; supervision, Y.L.; project administration, J.Z. and Y.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Natural Science Foundation of China under Grant 61801359.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Badue, C.; Guidolini, R.; Carneiro, R.V.; Azevedo, P.; Cardoso, V.B.; Forechi, A.; Jesus, L.; Berriel, R.; Paixao, T.M.; Mutz, F.; et al. Self-driving cars: A survey. Expert Syst. Appl. 2020, 165, 113816. [Google Scholar] [CrossRef]
  2. Yurtsever, E.; Lambert, J.; Carballo, A.; Takeda, K. A survey of autonomous driving: Common practices and emerging technologies. IEEE Access 2020, 8, 58443–58469. [Google Scholar] [CrossRef]
  3. Xiang, Y.; Choi, W.; Lin, Y.; Savarese, S. Data-driven 3d voxel patterns for object category recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1903–1911. [Google Scholar]
  4. Qi, C.R.; Su, H.; Mo, K.; Guibas, L.J. PointNet: Deep Learning on Point Sets for 3D Classification and Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  5. Qi, C.R.; Yi, L.; Su, H.; Guibas, L.J. Pointnet++: Deep hierarchical feature learning on point sets in a metric space. arXiv 2017, arXiv:1706.02413. [Google Scholar]
  6. Yin, Z.; Tuzel, O. VoxelNet: End-to-End Learning for Point Cloud Based 3D Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  7. Yan, Y.; Mao, Y.; Li, B. Second: Sparsely embedded convolutional detection. Sensors 2018, 18, 3337. [Google Scholar] [CrossRef] [Green Version]
  8. Lang, A.H.; Vora, S.; Caesar, H.; Zhou, L.; Yang, J.; Beijbom, O. PointPillars: Fast Encoders for Object Detection from Point Clouds. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  9. Zhao, Z.Q.; Zheng, P.; Xu, S.T.; Wu, X. Object detection with deep learning: A review. IEEE Trans. Neural Netw. Learn. Syst. 2019, 30, 3212–3232. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  10. Li, X.; Guivant, J.E.; Kwok, N.; Xu, Y. 3D backbone network for 3D object detection. arXiv 2019, arXiv:1901.08373. [Google Scholar]
  11. Li, B. 3d fully convolutional network for vehicle detection in point cloud. In Proceedings of the 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Vancouver, BC, Canada, 24–28 September 2017; pp. 1513–1518. [Google Scholar]
  12. Rusu, R.B. Semantic 3d object maps for everyday manipulation in human living environments. KI-Kunstl. Intell. 2010, 24, 345–348. [Google Scholar] [CrossRef] [Green Version]
  13. Wang, W.; Sakurada, K.; Kawaguchi, N. Incremental and enhanced scanline-based segmentation method for surface reconstruction of sparse LiDAR data. Remote Sens. 2016, 8, 967. [Google Scholar] [CrossRef] [Green Version]
  14. Narksri, P.; Takeuchi, E.; Ninomiya, Y.; Morales, Y.; Akai, N.; Kawaguchi, N. A slope-robust cascaded ground segmentation in 3D point cloud for autonomous vehicles. In Proceedings of the 2018 21st International Conference on Intelligent Transportation Systems (ITSC), Maui, HI, USA, 4–7 November 2018; pp. 497–504. [Google Scholar]
  15. Lambert, J.; Liang, L.; Morales, L.Y.; Akai, N.; Carballo, A.; Takeuchi, E.; Narksri, P.; Seiya, S.; Takeda, K. Tsukuba challenge 2017 dynamic object tracks dataset for pedestrian behavior analysis. J. Robot. Mech. 2018, 30, 598–612. [Google Scholar] [CrossRef]
  16. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar]
  17. Irshick, R. Fast r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  18. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. arXiv 2015, arXiv:1506.01497. [Google Scholar] [CrossRef] [Green Version]
  19. Chen, X.; Ma, H.; Wan, J.; Li, B.; Xia, T. Multi-view 3d object detection network for autonomous driving. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1907–1915. [Google Scholar]
  20. Shi, S.; Wang, X.; Li, H.P. 3d object proposal generation and detection from point cloud. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 16–20. [Google Scholar]
  21. Qi, C.R.; Liu, W.; Wu, C.; Su, H.; Guibas, L.J. Frustum PointNets for 3D Object Detection from RGB-D Data. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  22. Ali, W.; Abdelkarim, S.; Zidan, M.; Zahran, M.; El Sallab, A. YOLO3D: End-to-end real-time 3D Oriented Object Bounding Box Detection from LiDAR Point Cloud. In Proceedings of the ECCV 2018: “3D Reconstruction meets Semantics” Workshop, Munich, Germany, 8–14 September 2018. [Google Scholar]
  23. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  24. Yang, Z.; Sun, Y.; Liu, S.; Jia, J. 3dssd: Point-based 3d single stage object detector. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 11040–11048. [Google Scholar]
  25. Li, B.; Zhang, T.; Xia, T. Vehicle detection from 3d lidar using fully convolutional network. arXiv 2016, arXiv:1608.07916. [Google Scholar]
  26. Oda, M.; Shimizu, N.; Roth, H.R.; Karasawa, K.I.; Kitasaka, T.; Misawa, K.; Fujiwara, M.; Rueckert, D.; Mori, K. 3D FCN Feature Driven Regression Forest-Based Pancreas Localization and Segmentation. In Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support; Springer: Cham, Switzerland, 2017. [Google Scholar]
  27. Engelcke, M.; Rao, D.; Wang, D.Z.; Tong, C.H.; Posner, I. Vote3Deep: Fast Object Detection in 3D Point Clouds Using Efficient Convolutional Neural Networks. In Proceedings of the International Conference on Robotics and Automation (ICRA), Singapore, 29 May–3 June 2017. [Google Scholar]
  28. Graham, B.; Maaten, L. Submanifold Sparse Convolutional Networks. arXiv 2017, arXiv:1706.01307. [Google Scholar]
  29. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. Ssd: Single shot multibox detector. In European Conference on Computer Vision; Springer: Cham, Switzerland, 2016; pp. 21–37. [Google Scholar]
  30. Chaudhari, S.; Mithal, V.; Polatkan, G.; Ramanath, R. An Attentive Survey of Attention Models. ACM Trans. Intell. Syst. Technol. TIST 2021, 12, 1–32. [Google Scholar] [CrossRef]
  31. Venugopalan, S.; Rohrbach, M.; Donahue, J.; Mooney, R.; Darrell, T.; Saenko, K. Sequence to sequence-video to text. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 4534–4542. [Google Scholar]
  32. Jaderberg, M.; Simonyan, K.; Zisserman, A. Spatial transformer networks. arXiv 2015, arXiv:1506.02025. [Google Scholar]
  33. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  34. Wang, F.; Jiang, M.; Qian, C.; Yang, S.; Li, C.; Zhang, H.; Wang, X.; Tang, X. Residual Attention Network for Image Classification. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  35. Zhao, X.; Liu, Z.; Hu, R.; Huang, K. 3D object detection using scale invariant and feature reweighting networks. Proc. AAAI Conf. Artif. Intell. 2019, 33, 9267–9274. [Google Scholar] [CrossRef]
  36. Liu, X.; Han, Z.; Liu, Y.-S.; Zwicker, M. Point2sequence: Learning the shape representation of 3d point clouds with an attention-based sequence to sequence network. Proc. AAAI Conf. Artif. Intell. 2019, 33, 8778–8785. [Google Scholar] [CrossRef] [Green Version]
  37. Redmon, J.; Farhadi, A. YOLOv3: An Incremental Improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  38. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  39. Geiger, A.; Lenz, P.; Urtasun, R. Are we ready for autonomous driving? The kitti vision benchmark suite. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 3354–3361. [Google Scholar]
  40. Ye, Y.; Chen, H.; Zhang, C.; Hao, X.; Zhang, Z. SARPNET: Shape attention regional proposal network for liDAR-based 3D object detection. Neurocomputing 2020, 379, 53–63. [Google Scholar] [CrossRef]
Figure 1. Human visual attention.
Figure 1. Human visual attention.
Remotesensing 13 05071 g001
Figure 2. The structure of HCNet. The original point cloud is used as the input, and multiple pseudo images are fused by the HC module; in addition, the adaptive detector head is used to overcome the sparsity and uneven distribution of the point cloud.
Figure 2. The structure of HCNet. The original point cloud is used as the input, and multiple pseudo images are fused by the HC module; in addition, the adaptive detector head is used to overcome the sparsity and uneven distribution of the point cloud.
Remotesensing 13 05071 g002
Figure 3. Pseudo images generator (P = 3).
Figure 3. Pseudo images generator (P = 3).
Remotesensing 13 05071 g003
Figure 4. The structure of the HC module. It includes a height attention branch and a channel attention branch, which adjust features from two dimensions, respectively.
Figure 4. The structure of the HC module. It includes a height attention branch and a channel attention branch, which adjust features from two dimensions, respectively.
Remotesensing 13 05071 g004
Figure 5. Diagram of the height attention branch.
Figure 5. Diagram of the height attention branch.
Remotesensing 13 05071 g005
Figure 6. Diagram of the channel attention branch.
Figure 6. Diagram of the channel attention branch.
Remotesensing 13 05071 g006
Figure 7. Diagram of the backbone network.
Figure 7. Diagram of the backbone network.
Remotesensing 13 05071 g007
Figure 8. The structure of the adaptively adjusted detection head. The structure of the OIFM (original information fusion module) is shown in the lower-left part of the figure, and the structure of the AAM (adaptive adjustment module) is shown in the lower-right part of the figure.
Figure 8. The structure of the adaptively adjusted detection head. The structure of the OIFM (original information fusion module) is shown in the lower-left part of the figure, and the structure of the AAM (adaptive adjustment module) is shown in the lower-right part of the figure.
Remotesensing 13 05071 g008
Figure 9. Examples from the KITTI dataset [39].
Figure 9. Examples from the KITTI dataset [39].
Remotesensing 13 05071 g009
Figure 10. Performance of different algorithms (3D).
Figure 10. Performance of different algorithms (3D).
Remotesensing 13 05071 g010
Figure 11. Performance of different algorithms (AOS).
Figure 11. Performance of different algorithms (AOS).
Remotesensing 13 05071 g011
Figure 12. Performance of different attention mechanisms.
Figure 12. Performance of different attention mechanisms.
Remotesensing 13 05071 g012
Figure 13. Example 1 (low occlusion and truncation).
Figure 13. Example 1 (low occlusion and truncation).
Remotesensing 13 05071 g013
Figure 14. Example 2 (high occlusion and truncation).
Figure 14. Example 2 (high occlusion and truncation).
Remotesensing 13 05071 g014
Figure 15. Confidence.
Figure 15. Confidence.
Remotesensing 13 05071 g015
Table 1. The comparison of performances in 3D object detection AP (%) on the KITTI validation set for the class of car.
Table 1. The comparison of performances in 3D object detection AP (%) on the KITTI validation set for the class of car.
MethodTimes(s)3D
EasyModerateHard
VoxelNet [6]0.2381.9765.4662.85
SECOND [7]0.0585.5075.0468.78
PointPillars [8]0.01687.5077.0174.77
PointRCNN [20]0.1088.2677.7376.67
HCNET (ours)0.03288.4578.0176.72
Table 2. The comparison of performances in bird’s eye view (BEV) detection AP (%), 3D object detection AP (%), and object orientation estimation AOS (%) on the KITTI test set for the class of car.
Table 2. The comparison of performances in bird’s eye view (BEV) detection AP (%), 3D object detection AP (%), and object orientation estimation AOS (%) on the KITTI test set for the class of car.
MethodTimes(s)BEV3DOrientation
EasyModerateHardEasyModerateHardEasyModerateHard
VoxelNet [6]0.2389.3579.2677.3977.4765.1157.73N/AN/AN/A
Second [7]0.0588.0779.3777.9583.1373.6666.2087.8481.3171.85
PointPillars [8]0.01688.3586.1079.8379.0574.9968.3090.1988.7686.38
PointRCNN [20]0.1089.2886.0479.0284.3275.4267.8690.7689.5580.76
SARPNET [40]0.0588.9387.2678.6884.9275.6467.7090.1688.8680.05
HCNET (ours)0.03290.2786.9181.9081.3173.5668.4295.0591.4488.02
Table 3. The comparison of the results of different attention mechanisms in HC module on the KITTI validation set for the class of car.
Table 3. The comparison of the results of different attention mechanisms in HC module on the KITTI validation set for the class of car.
Method3D
EasyModerateHard
CA87.2277.3375.78
HA87.1977.2675.82
HCA88.4578.0176.72
Table 4. The comparison of the results of different parts in the detection head on the KITTI validation set for the class of car.
Table 4. The comparison of the results of different parts in the detection head on the KITTI validation set for the class of car.
Method3D
EasyModerateHard
AAM87.2377.3375.70
OIFM88.3278.0876.52
(AAM + OIFM)88.4578.0176.72
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhang, J.; Wang, J.; Xu, D.; Li, Y. HCNET: A Point Cloud Object Detection Network Based on Height and Channel Attention. Remote Sens. 2021, 13, 5071. https://doi.org/10.3390/rs13245071

AMA Style

Zhang J, Wang J, Xu D, Li Y. HCNET: A Point Cloud Object Detection Network Based on Height and Channel Attention. Remote Sensing. 2021; 13(24):5071. https://doi.org/10.3390/rs13245071

Chicago/Turabian Style

Zhang, Jing, Jiajun Wang, Da Xu, and Yunsong Li. 2021. "HCNET: A Point Cloud Object Detection Network Based on Height and Channel Attention" Remote Sensing 13, no. 24: 5071. https://doi.org/10.3390/rs13245071

APA Style

Zhang, J., Wang, J., Xu, D., & Li, Y. (2021). HCNET: A Point Cloud Object Detection Network Based on Height and Channel Attention. Remote Sensing, 13(24), 5071. https://doi.org/10.3390/rs13245071

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop