Next Article in Journal
Domain Usability Evaluation
Previous Article in Journal
RMU-Net: A Novel Residual Mobile U-Net Model for Brain Tumor Segmentation from MR Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Real-Time Semantic Segmentation of 3D Point Cloud for Autonomous Driving

1
Graduate School of Automotive Engineering, Kookmin University, 77 Jeongneung-ro, Seongbuk-gu, Seoul 02707, Korea
2
#01-10a, Block 44, 535 Clementi Rd, Singapore 599489, Singapore
*
Author to whom correspondence should be addressed.
Electronics 2021, 10(16), 1960; https://doi.org/10.3390/electronics10161960
Submission received: 2 July 2021 / Revised: 9 August 2021 / Accepted: 12 August 2021 / Published: 14 August 2021
(This article belongs to the Section Electrical and Autonomous Vehicles)

Abstract

:
Autonomous vehicles perceive objects through various sensors. Cameras, radar, and LiDAR are generally used as vehicle sensors, each of which has its own characteristics. As examples, cameras are used for a high-level understanding of a scene, radar is applied to weather-resistant distance perception, and LiDAR is used for accurate distance recognition. The ability of a camera to understand a scene has overwhelmingly increased with the recent development of deep learning. In addition, technologies that emulate other sensors using a single sensor are being developed. Therefore, in this study, a LiDAR data-based scene understanding method was developed through deep learning. The approaches to accessing LiDAR data through deep learning are mainly divided into point, projection, and voxel methods. The purpose of this study is to apply a projection method to secure a real-time performance. The convolutional neural network method used by a conventional camera can be easily applied to the projection method. In addition, an adaptive break point detector method used for conventional 2D LiDAR information is utilized to solve the misclassification caused by the conversion from 2D into 3D. The results of this study are evaluated through a comparison with other technologies.

1. Introduction

Scene understanding through semantic segmentation is one of the components of the perception system used in autonomous vehicles. Autonomous vehicles understand the overall situation by applying multiple attached sensors. Typically, information is acquired through radar, cameras, and LiDAR. Each sensor has its own advantages and disadvantages. The perception system configures each sensor to have a complementary relationship used to solve the shortcomings of the sensors. In addition, more accurate results are sought through redundant information of each sensor. Greater stability can be secured when there is more redundant information with high reliability. To acquire redundant information, technologies that emulate the functions of different sensors with a single sensor are being developed [1,2,3].
Semantic segmentation is a field of computer vision. Scene understanding can be divided mainly into classification, detection, and segmentation. Classification is a method of predicting a label for an image. Detection is a method of predicting the position of an image while predicting the label. Segmentation is a task dividing objects of an image into meaningful units or a method of predicting a label for every pixel.
The use of semantic information is increasing in areas such as localization, object detection, and tracking, which are the roles of LiDAR in autonomous vehicles. It is used to increase the effect of such algorithms as loop closure [1] or to increase the performance of object tracking in simultaneous localization and mapping. Methods using deep learning have been proposed for semantic segmentation of 3D LiDAR data. Such deep learning methods can be mainly divided into three types, point methods that use original raw data and do not apply preprocessing; voxel grid methods that standardize and reduce the number of data; and 2D projection methods that use a 2D projection, similar to an image [4]. Although a point method is robust against data distortion by using the original raw data, it has difficulty guaranteeing a real-time performance. For a voxel grid method, the distortion rate and computation speed vary depending on the size of the grid. Finally, in a 2D projection, data are simplified by converting the coordinate system from 3D into 2D data.
A real-time LiDAR semantic segmentation method was used in this study. Moreover, a method of projecting LiDAR data from a 3D coordinate system into a 2D image coordinate system was used. The segmentation of each pixel of a 2D-projected LiDAR image was inferred using a convolutional neural network. The LiDAR data of the 3D coordinate system were segmented by applying the results inferred from the 2D image to the 3D coordinate system. In this paper, a modified version of an existing image semantic segmentation network is proposed by considering the characteristics of the point cloud. A filter using an adaptive break point detector (ABD) was used to reduce the misclassification when applying data inferred from a 2D coordinate system to a 3D coordinate system data. Based on the above description, it operates faster than the measurement speed of the LiDAR sensor (approximately 10 Hz) and performs Semantic Segmentation of LiDAR data through reliable level of inference.

2. Related Work

Scene perception in autonomous vehicles has made rapid progress with the advent of deep learning. In particular, techniques such as semantic segmentation have been developed. However, semantic segmentation requires a large amount of computing power. This problem has been significantly resolved through parallel processing using a graphics processing unit. In addition, research on the weight pruning of deep neural networks (DNNs), such as MobileNets V2, has been conducted [5].
Studies on semantic segmentation are also being conducted for LiDAR data, including the semantic segmentation of images. An indirect method for obtaining semantic segmentation information of an image through calibration was developed. A method for directly applying LiDAR to a DNN and achieve the semantic segmentation of LiDAR data has also been applied [1]. In addition, methods for directly applying LiDAR data to DNNs are being studied, of which there are three main types: a method for applying a 3D convolution by splitting a 3D space into voxels of a given size to apply a point cloud to a DNN, a method for applying a 2D convolution by using a multi-view image as an input, and a method for directly applying a point cloud to the network [6].
PointNet, proposed by Qi et al., uses a transform network, which is an end-to-end DNN that learns the characteristics directly from a point cloud [6]. It was applied to 3D object perception and 3D semantic segmentation. Subsequently, an improved PointNet++ was proposed to learn the local characteristics [7]. VoxelNet, proposed by Zhou et al., was the first to employ an end-to-end DNN in the 3D domain. The data were simplified and standardized for application to a DNN by splitting the space into voxels and expressing only certain points as voxels [8]. SqueezeSeg, proposed by Wu et al., projects the point cloud onto the image coordinate system for use in a 2D convolution [9]. In addition, a conditional random field, used in image semantic segmentation, was applied. The results showed a faster performance than the measurement speed of the sensor by projecting onto the image coordinate system. An improved DNN was proposed through V2 and V3 for SqueezeSeg [2]. As with SqueezeSeg, a method for projecting a point cloud on the image coordinate system was used.
A DNN requires a large number of data to extract the features. The Cityscape and Mapillary datasets are mainly used in semantic segmentation of a video. In this study, the SemanticKITTI dataset was used to obtain 3D semantic segmentation training data [3,10].

3. Method

The purpose of this study was to conduct a semantic segmentation of LiDAR data that can be used in perception systems of autonomous vehicles. The method for projecting LiDAR data into a 2D image coordinate system, and the configuration and characteristics of a DNN for a semantic segmentation of the projected image, are described in this section. Postprocessing using ABD was used to limit the misclassification during the process of recovering from the inferred image using LiDAR data. After projecting from (A) the input to the 2D image coordinate system, the semantic segmented LiDAR data were output by passing through (B) the proposed DNN and (C) the ABD filter, as shown in Figure 1.

3.1. LiDAR Point Cloud Representation

The method for projecting 3D LiDAR data ( x ,   y ,   z ,   i ) onto the image coordinate system ( u ,   v ) for a 2D convolution application is described in this section. The LiDAR coordinate system can be projected onto the image coordinate system using a spherical coordinate system. The projection according to the sensor measurement method generates data with a large amount of noise, such as the overlapping of objects, when LiDAR data are projected onto the image coordinate system because 3D data are projected onto the 2D coordinate system. The data with the shortest path from the sensor are represented in the image to prevent noise. The 360 ° data acquired from the sensor are projected [11].
( u v ) = ( 1 2 [ 1 arctan ( y , x ) π 1 ]   W [ 1 ( arcsin ( z r 1 ) + f u p ) f 1 ]   H ) ,
where ( W ,   H ) denote the width and height of the image, and x ,   y ,   z denote LiDAR data. Here, f = f u p + f d o w n is the field of view of the sensor, and r = x 2 + y 2 + z 2 [12]. The [ x ,   y ,   z ,   r ,   i ] image is generated by projecting the LiDAR data and r onto the converted coordinate system using Equation (1). The created image is input into the network in the form of [ W × H × 5 ] . A spherical projection is shown in Figure 2.

3.2. Network Structure

The proposed network configuration uses DeepLabV3 as the main network, and partial convolution and semilocal convolution are the main convolution layer.

3.2.1. Partial Convolution

A partial convolution is a padding method proposed by Nvidia. The size of the input data decreases as the convolution and pooling proceed. The data may be excessively reduced, and information may be lost as the depth of the network increases. Padding is applied to extend the surroundings of the input data by filling them with a specific value to prevent data loss. Generally, zero padding is used. However, data including errors are obtained at the border of the image if it is filled with a specific value or zero. Partial convolution is a method of conditioning the output for input data by defining 0 as a hole and 1 as a non-hole by adding a binary mask. A partial convolution is helpful for data loss through the abovementioned method and has been applied to holes generated when the LiDAR data and the error at the border are projected onto a 2D coordinate system [13]. The partial convolution is shown in Figure 3.
Partial   conv ( x ) = w x 0 * r a t i o r a t i o = s u m ( p 1 ) s u m ( p 0 )

3.2.2. Semilocal Convolution

A semilocal convolution uses the fact that data in a fixed space are measured when LiDAR data are projected into a 2D coordinate system, unlike the image data of a camera. This convolution is applied by dividing the input data by α . Different kernels can be applied to the segmented convolution, and the convolution weight divided by the region is shared. Moreover, it can be learned by using the characteristics of LiDAR data because it is learned by dividing the input data by region and applying weights [14]. The semilocal convolution is shown in Figure 4.

3.2.3. Atrous Convolution

Atrous convolution creates and uses an empty space inside the kernel, unlike a conventional convolution. For segmentation, it is better for a DNN when a pixel has a wider field of view. A conventional method constructs a deeper DNN for the pixel to have a wider field of view. However, more original information is lost when the DNN is deeper. An atrous convolution expands the field of view of a pixel by creating an empty space. This is advantageous for segmentation, and a light DNN can be configured because the field of view of a pixel is expanded. Here, r represents the size of the empty space. Different ratios of r can be used to obtain multiscale features simultaneously [15]. An atrous convolution is shown in Figure 5. This convolution is used in a network, as shown in Figure 6.

3.2.4. DeepLabV3+

DeepLabV3+, used in image semantic segmentation, was applied as the backbone. DeepLabV3+ is designed for image data semantic segmentation and in encoder-decoder structures. Four methods were proposed for DeepLabV3+ from version 1 to 3+. An atrous convolution was proposed in V1, atrous spatial pyramid pooling was proposed in V2, and the ResNet structure was proposed by applying an atrous convolution in V3. The V3+ used in this study uses an atrous separable convolution [15].

3.2.5. Network Details

Data are received in the form of 64 × 1024 × 5   ( H × W × C ) as input. An encoder–decoder structure is used with DeepLabV3+ as the backbone. Xception-41 is used as the backbone network, and the entry part is replaced with a partial convolution and semilocal convolution considering that the inputs are LiDAR data. Cross entropy is used as the loss function [14].
CE ( y ^ , y ) = y ^ c i l o g y c i
This loss function is applied most frequently. Here, y ^ c i denotes the ground-truth for class c at the one-hot encoded pixel position i, and y c i denotes the predicted data of softmax. Generally, semantic segmentation is evaluated using the mean intersection over union (mIoU). The purpose is to minimize the cross entropy in the learning process to reach a high mIoU. The modified 3D model of Xception is shown in Figure 7. The network structure is shown in Figure 8.

3.3. Postprocessing

Conversion from a 3D coordinate system into a 2D coordinate system causes errors because only data with shortest path from the sensor are shown as a representative point in the 2D image coordinate system. A misclassification occurs when the data that were classified in a 2D image coordinate system are applied to 3D data. In this study, a filter using an ABD was applied to reduce the misclassification. An ABD was used for the clustering method of 2D LiDAR data.
When the distance between || p n p n 1 || is greater than the threshold circle ( D m a x ), ABD is a method that designates this as a break point [16]. If the threshold of the circle is small, the prediction will not be reached, and if it is large, an overflow will occur.
The pseudocode of the ABD, as shown in Table 1, adaptively depends on Δ θ and r, as shown in Figure 9. Here, Δ θ = θ n θ n 1 , λ denotes a user-definable constant, and σ r is the sensor noise associated with r, in which the range of influence of the circle is wider when λ is small or σ r is large. The distance r and height h of data projected on a pixel [u, v] were substituted into the filter by using the ABD as a filter in order of distance, and the classification result was applied by designating it as one object up to the break point. The parameter values were calculated empirically. In the system, λ is 10, and σ r is 2. The ABD is shown in Figure 9. The ABD postprocessing is shown in Figure 10.

4. Experiments

The network was trained and evaluated using SemanticKitti data. SemanticKitti provides LiDAR data by labeling them from Kitti data. The dataset consists of more than 43,000 scans. The data are organized in sequences within the range of 00 to 21, and 21,000 scans from 00 to 10 can be used for training because they provide the ground-truth. Sequences 11 to 21 are used as test data. The dataset provides 28 classes including moving objects, which were used in the experiment by being merged into 19 classes [10].
For the scratch training of the network, the base learning rate is 0.03, the weight decay is 0.000015, and the batch size is 38. The hardware and software configurations are presented in Table 2 and Table 3, respectively.
An mIoU evaluation, which is mainly used in a semantic segmentation evaluation, was employed to evaluate the inference results [12]. The evaluation is shown in Table 4.
mIoU = 1 C C = 1 C T P c T P c + F P c + F N c
Here, T P c ,   F P c , and F N c are the true positive, false positive, and false negative predictions of class c, respectively, and C is the number of classes. For the evaluation network, the network was improved to apply DeepLab V3+, which is a 2D semantic segmentation network, to 3D LiDAR data as a backbone. An evaluation of the network inference results is shown in Table 4. It contains the results according to the image size and the size of the decoder stride of DeepLabV3+. The network results are shown in Figure 11.
It was found that only the data projected on the 2D image were classified through the proposed filter, as shown in Figure 12. The white data are unclassified data.

5. Conclusions

A 2D network was designed for a semantic segmentation of 3D LiDAR, and a semantic segmentation algorithm using the network was proposed. The error propagation, which is a disadvantage of 2D classification, was reduced by using a postprocessed ABD filter to reduce the classification error of a 2D network. Finally, semantic segmentation was conducted for 3D LiDAR data. The practicality of this was demonstrated through its considerably faster speed than the sensor measurement speed of 10 Hz with a computing speed of 13 Hz.
A further study on the development of a network to complement classes with low classification results and for weight pruning is planned.

Author Contributions

Conceptualization, D.K.; methodology, D.K.; software, D.K.; validation, D.K.; formal analysis, D.K.; investigation, D.K.; data curation, D.K.; writing—review and editing, D.K. and J.K.; supervision, A.W. and J.K.; project administration, B.L. and J.K.; All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Ministry of Trade, Industry, and Energy (MOTIE) in Korea, under the Fostering Global Talents for Innovative Growth Program (P0008751) supervised by the Korea Institute for Advancement of Technology (KIAT).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

All training data used in this paper are available from the references.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Milioto, A.; Vizzo, I.; Behley, J.; Stachniss, C. RangeNet++: Fast and accurate LiDAR semantic segmentation. In Proceedings of the 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Macau, China, 3–8 November 2019; pp. 4213–4220. [Google Scholar]
  2. Wu, B.; Wan, A.; Yue, X.; Keutzer, K. Squeezeseg: Convolutional neural nets with recurrent CRF for real-time road-object segmentation from 3D LiDAR point cloud. In Proceedings of the 2018 IEEE International Conference on Robotics and Automation (ICRA), Brisbane, Australia, 21–26 May 2018; pp. 1887–1893. [Google Scholar]
  3. Wu, B.; Zhou, X.; Yue, X.; Keutzer, K. Squeezesegv2: Improved model structure and unsupervised domain adaptation for road-object segmentation from a LiDAR point cloud. In Proceedings of the 2019 International Conference on Robotics and Automation (ICRA), Montreal, QC, Canada, 20–24 May 2019; pp. 4376–4382. [Google Scholar]
  4. Sualeh, M.; Kim, G.-W. Simultaneous localization and mapping in the epoch of semantics: A survey. Int. J. Control Autom. Syst. 2018, 17, 729–742. [Google Scholar] [CrossRef]
  5. Kang, D.W.; Kim, D.J.; Sun, H.D.; Kim, J.H. Object classification and optimize calibration using the network in map. J. Inst. Control Robot. Syst. 2020, 26, 443–451. [Google Scholar] [CrossRef]
  6. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L. MobileNetV2: Inverted residuals and linear bottlenecks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 4510–4520. [Google Scholar]
  7. Qi, C.R.; Su, H.; Mo, K.; Guibas, L.J. PointNet: Deep learning on point sets for 3D classification and segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 652–660. [Google Scholar]
  8. Qi, C.R.; Yi, K.; Su, H.; Guibas, L.J. PointNet++: Deep hierarchical feature learning on point sets in a metric space. In Proceedings of the Advances in Neural Information Processing Systems (NIPS), Long Beach, CA, USA, 4–9 December 2017; pp. 5105–5114. [Google Scholar]
  9. Zhou, Y.; Tuzel, O. Voxelnet: End-to-end learning for point cloud based 3D object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 4490–4499. [Google Scholar]
  10. Neuhold, G.; Ollmann, T.; Bulo, S.R.; Kontschieder, P. The Mapillary vistas dataset for semantic understanding of street scenes. In Proceedings of the IEEE Intl. Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 4990–4999. [Google Scholar]
  11. Behley, J.; Garbade, M.; Milioto, A.; Quenzel, J.; Behnke, S.; Stachniss, C.; Gall, J. SemanticKITTI: A dataset for semantic scene understanding of LiDAR sequences. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, South Korea, 27 October–3 November 2019; pp. 9297–9307. [Google Scholar]
  12. Park, S.Y.; Choi, S.I.; Moon, J.; Kim, J.; Park, Y.W. Localization of an unmanned ground vehicle based on hybrid 3D registration of 360degree range data and DSM. Int. J. Control Autom. Syst. 2011, 9, 875–887. [Google Scholar] [CrossRef]
  13. Liu, G.; Shih, K.J.; Wang, T.C.; Reda, F.A.; Sapra, K.; Yu, Z.; Catanzaro, B. Partial convolution based padding. arXiv 2018, arXiv:1811.11718. [Google Scholar]
  14. Triess, L.T.; Peter, D.; Rist, C.B.; Zöllner, J.M. Scan-based semantic segmentation of LiDAR point clouds: An experimental study. In Proceedings of the 2020 IEEE Intelligent Vehicles Symposium (IV), Las Vegas, NV, USA, 19 October–13 November 2020; pp. 1116–1121. [Google Scholar]
  15. Chen, L.C.; Zhu, Y.; Papndreou, G.; Schroff, F.; Adam, H. Encoder-decoder with atrous separable convolution for semantic image segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 801–818. [Google Scholar]
  16. Kang, D.W.; Yang, J.H.; Kim, J.H. Extended Kalman filter based localization to autonomous vehicles using a 2D laser sensor. In Proceedings of the Korean Society of Automotive Engineers, Jeju, Korea, 18–20 May 2017; pp. 493–497. [Google Scholar]
  17. Landrieu, L.; Simonovsky, M. Large-scale point cloud semantic segmentation with superpoint graphs. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 4558–4567. [Google Scholar]
  18. Tatarchenko, M.; Park, J.; Koltun, V.; Zhou, Q.-Y. Tangent convolutions for dense prediction in 3D. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 3887–3896. [Google Scholar]
Figure 1. Point cloud spherical projection.
Figure 1. Point cloud spherical projection.
Electronics 10 01960 g001
Figure 2. Point cloud spherical projection.
Figure 2. Point cloud spherical projection.
Electronics 10 01960 g002
Figure 3. Partial convolution.
Figure 3. Partial convolution.
Electronics 10 01960 g003
Figure 4. Semilocal convolution.
Figure 4. Semilocal convolution.
Electronics 10 01960 g004
Figure 5. Atrous convolution.
Figure 5. Atrous convolution.
Electronics 10 01960 g005
Figure 6. Atrous convolution in a network.
Figure 6. Atrous convolution in a network.
Electronics 10 01960 g006
Figure 7. Modified 3D model of Xception.
Figure 7. Modified 3D model of Xception.
Electronics 10 01960 g007
Figure 8. Network structure.
Figure 8. Network structure.
Electronics 10 01960 g008
Figure 9. Adaptive break point detector.
Figure 9. Adaptive break point detector.
Electronics 10 01960 g009
Figure 10. ABD postprocessing.
Figure 10. ABD postprocessing.
Electronics 10 01960 g010
Figure 11. Network result.
Figure 11. Network result.
Electronics 10 01960 g011
Figure 12. Results before (left) and after (right) applying ABD filter.
Figure 12. Results before (left) and after (right) applying ABD filter.
Electronics 10 01960 g012
Table 1. Adaptive break point detector.
Table 1. Adaptive break point detector.
Pseudocode: Adaptive Break Point Detector
for n = 2 to N do
D m a x = r n 1 · sin ( Δ Φ ) sin ( λ Δ Φ ) + 3 σ r
If   || p n p n 1 || > D m a x   t h e n
  *breakpoint detected*
   f l a g n b T r u e
   f l a g n 1 b T r u e
 else
   f l a g n b F a l s e
 end
end
Table 2. Hardware configuration.
Table 2. Hardware configuration.
ItemSpec
Train DesktopIntel i9-9940X CPU 3.30 GHz
Nvidia RTX-2080Ti
RAM 128 GB
Ubuntu 16.04
Test DesktopIntel i7-9700E
Nvidia GTX 1660Ti
RAM 16 GB
Ubuntu 16.04
Table 3. Software configuration.
Table 3. Software configuration.
ItemSpec
LanguagePython 3.5
FrameworkTensorFlow 1.13.2
Table 4. IoU [%] on test set (Sequences 11 to 21).
Table 4. IoU [%] on test set (Sequences 11 to 21).
ApproachPointnet++ [7]SPGraph [17]TrangentConv [18]SqueezeSegV2 [2]RangeNet53++ [12]Proposed
Decoder1Decoder 2Decoder 1
Size50,000 pts64 × 2048 px64 × 1024 px
car53.768.386.881.886.492.5774.382.8
bicycle1.90.91.318.524.542.681527
motorcycle0.24.512.717.932.760.3314.426.8
truck0.90.911.613.425.537.310.310.8
other-vehicle0.20.810.21422.667.3123.425.8
person0.9117.120.136.275.7803.5
bicyclist1620.225.133.6025.70
motorcyclist000.53.94.751.250.20.1
road7249.582.988.691.896.2486.690.8
parking18.71.715.245.864.870.9956.266
sidewalk41.824.261.767.674.691.7167.673.4
other-ground5.60.3917.727.912.1421.628.5
building62.368.282.873.781.192.279.483
fence16.922.544.241.15575.6542.153.9
vegetation46.559.275.571.878.391.2773.877.4
trunk13.827.242.535.850.161.2344.148.2
terrain301755.560.26482.3859.164.5
pole618.330.220.238.956.4623.932.4
traffic sign8.910.522.236.352.245.830.537.3
mean IOU20.12035.939.749.963.3339.443.8
scan/s0.10.20.35013101513
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Kang, D.; Wong, A.; Lee, B.; Kim, J. Real-Time Semantic Segmentation of 3D Point Cloud for Autonomous Driving. Electronics 2021, 10, 1960. https://doi.org/10.3390/electronics10161960

AMA Style

Kang D, Wong A, Lee B, Kim J. Real-Time Semantic Segmentation of 3D Point Cloud for Autonomous Driving. Electronics. 2021; 10(16):1960. https://doi.org/10.3390/electronics10161960

Chicago/Turabian Style

Kang, Dongwan, Anthony Wong, Banghyon Lee, and Jungha Kim. 2021. "Real-Time Semantic Segmentation of 3D Point Cloud for Autonomous Driving" Electronics 10, no. 16: 1960. https://doi.org/10.3390/electronics10161960

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop