Next Article in Journal
Enhancing Multi-tissue and Multi-scale Cell Nuclei Segmentation with Deep Metric Learning
Next Article in Special Issue
A Response-Adaptive Method for Design of Validation Experiments in Computational Mechanics
Previous Article in Journal
The Brewing Industry and the Opportunities for Real-Time Quality Analysis Using Infrared Spectroscopy
Previous Article in Special Issue
Simplified Vibration PSD Synthesis Method for MIL-STD-810
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Learning Collision Situation to Convolutional Neural Network Using Collision Grid Map Based on Probability Scheme

School of Mechanical Engineering, Chonnam National University, 77 Yongbong-ro, Buk-gu, Gwangju 61186, Korea
*
Author to whom correspondence should be addressed.
Appl. Sci. 2020, 10(2), 617; https://doi.org/10.3390/app10020617
Submission received: 19 November 2019 / Revised: 10 January 2020 / Accepted: 10 January 2020 / Published: 15 January 2020
(This article belongs to the Special Issue Selected Papers from the ICMR 2019)

Abstract

:
In this paper, a Collision Grid Map (CGM) is proposed by using 3d point cloud data to predict the collision between the cattle and the end effector of the manipulator in the barn environment. The Generated Collision Grid Map using x-y plane and depth z data in 3D point cloud data is applied to a Convolutional Neural Network to predict a collision situation. There is an invariant of the permutation problem, which is not efficiently learned in occurring matter of different orders when 3d point cloud data is applied to Convolutional Neural Network. The Collision Grid Map is generated by point cloud data based on the probability method. The Collision Grid Map scheme is composed of a 2-channel. The first channel is constructed by location data in the x-y plane. The second channel is composed of depth data in the z-direction. 3D point cloud is measured in a barn environment and created a Collision Grid Map. Then the generated Collision Grid Map is applied to the Convolutional Neural Network to predict the collision with cattle. The experimental results show that the proposed scheme is reliable and robust in a barn environment.

1. Introduction

Many studies [1,2,3,4] are proposed where the robot is suitable for feeding and monitoring livestock in the barn environment. When a mobile robot feeds the cattle, there is a problem with the cow sticking their head out of the fence for eating and colliding with the end effector of the manipulator. In order to solve the collision problem, the K-Nearest Neighbor method [5] was used to cluster point cloud data to detect obstacles. The depth space method [6,7] was defined to find the obstacle in depth space and cartesian space. Obstacles were constructed with octree structures [8] to find a safe space. The ANN method [9,10] was modeled as an adaptive neural network for predicting penetration rate and rock tensile strength.
Recently, a Convolutional Neural Network (CNN) algorithm [11,12,13,14,15,16] using image data for object classification was proposed. In the proposed method [11], the neural network shows the feature of image data using gradient descent for optimizing. As one of the deep neural network models, Alex Net [12] is a network that is deeply constructed and extracts not only simple features but also high-level features to enhance the performance of the object classification. In the VGG Net [13], it uses a filter size that is smaller than Alex Net and an increased depth layer to enhance the performance. When deeper networks are optimized, there is a degradation problem that the neural network is not efficiently learned by training data. ResNet [14] addresses the degradation problem used for the residual learning method. A pheromone convolutional method [15] is used as discrete gaussian distribution to generate the efficient network path in a topological network. In PSI-CNN [16], it extracts an untrained feature map from various image resolutions. The untrained feature map is used to fuse the original feature map. PSI-CNN shows stable performance when tested on low-resolution images acquired from CCTV cameras.
Unlike images that are regularized pixel, the 3D point cloud is unordered data. These characteristics occur as invariant permutation problems when point cloud data is applied to the Convolutional Neural Network. Some references [17,18,19,20] were proposed for the Convolutional Neural Network (CNN) using a 3d point cloud. In PointNet [21], an invariant problem is solved by a symmetric function formulated multi-layer perceptron (MLP) and a max pooling function. As the end-to-end deep neural network, PointNet is directly learned by point cloud, which segmented the features of 3d objects and demonstrated high performance in classifications. PointNet++ [22] is able to learn hierarchical features with increasing scales of contexts. The network of PointNet++ deals with non-uniform densities by using density adaptive strategy. In Reference [23], sparsity of point cloud was utilized by using sliding window detection with linear classifiers and a feature-centric voting method. Vote3deep [24] is able to efficiently learn a 3d feature by a sparse convolutional layer, which is based on the voting scheme. VoxNet [25] is learned by three types of the occupancy grid maps [26,27]. Learned network shows robust performance by a grid map, which is modeled by binary, hit, and density grid maps used for 3d point cloud from lidar or the RGBD camera. As the end-to-end learning network, proposed VoxelNet [28] extracts the 3D bounding box using 4-dimensional tensor and RGB data. It is based on a Region Proposal Network, a Convolutional Middle Layer, a Feature Learning Network, and a Voxel Feature Encoding (VFE) scheme.
In this paper, the Collision Grid Map (CGM) method is proposed by using 3D data to predict collision between the cow and the end effector of the manipulator in a barn environment. In a conventional method [25], the network is learned by three types of voxel mentioned above. However, in this paper, the Collision Grid Map is used and composed of a two-channel without any voxel. Each cell in the first channel is computed by distribution of location data in the x-y plane. The second channel consists of the average of depth z data. This proposed method is not a transformed locational feature of the point cloud in the x-y plane data. The neural network is learned about location features and safety distance without an invariant permutation problem. The Generated Collision Grid Map is applied to a Convolutional Neural Network for prediction of collision. In addition, learned network classifies the collision situation in the barn environment.
The remaining sections of this paper are organized as follows. In Section 2, the Collision Grid Map scheme and Convolutional Neural Network algorithm using the Collision Grid Map are explained. The experimental setup and results are described in Section 3. The concluding remarks are presented in Section 4.

2. Collision Grid Map Scheme and Convolutional Neural Network

2.1. Collision Grid Map Scheme

There is an invariant of the permutation problem, which is not efficiently learned in an occurring matter of a different order, as 3D point cloud data is applied to a Convolutional Neural Network. The proposed method [25] modelled voxel of the binary occupancy grid, the density grid, and the hit grid. Then the network is learned by voxels to solve an invariant of the permutation problem. However, the Collision Grid Map scheme with a two-channel is proposed to predict collision.
Equation (1) shows measurement data s and grid map M. Grid map M is constructed of the two-channel and m 1 , i , m 2 , i is each of the grid channel.
s = { s i } = { x i , y i , z i } ,   M = { m 1 , i , m 2 , i }
The Posterior of the Collision Grid Map is obtained as shown in Equation (2). In order to calculate the posterior efficiently, log odds notation can be defined, as shown in Equation (3).
p ( m |   s ) = i p ( m i |   s )
l ( m i |   s ) = log ( p ( m i |   s ) 1 p ( m i |   s ) )
Equation (4) is shown as applying log odds notation to Equation (2). Log odds notation is calculated using present point cloud data and previous point cloud data. When an initial grid map is free state, the initial Collision Grid Map can be omitted.
l ( m i |   s 1 : n ) = l ( m i |   s n ) + l ( m i |   s 1 : n 1 ) l ( m 0 )
Equations (5) and (6) show l ( m i |   s n ) that is modeled to predict the collision situation.
l ( m 1 ,   i |   s n ) = G ( x n , y n )
l ( m 2 ,   i |   s n ) = z 1 :   n m 1 ,   i
The first channel of the Collision Grid Map uses the discretization function G to grid the x and y data in the point cloud. When discretization data corresponds to the grid cell, grid cell values are added up in counts one by one such as in Equation (4). The second channel is consisted of the average of depth z data. The value of the cell in the first channel is constructed from x-y data based on the probability method. The value of the cell in the second channel is shown as a safe distance between the end effector and the cow.
The example of generating the first channel in the Collision Grid Map and Collision Grid Map is shown in Figure 1. When the Collision Grid Map is composed of a two-channel, an invariant of the permutation problem that occurs to apply point cloud to a Convolutional Neural Network is solved and raw point cloud data size becomes reduced. By modeling the Collision Grid Map, networks are able to learn not only distribution of the x-y plane data but also collision safety distance that is depth z data.

2.2. Convolutional Neural Network Algorithm Using the Collision Grid Map

The Convolutional Neural Network algorithm can be shown robustly in the office and road environment. However, there is required big data due to the different appearance. In an automatic farm system, typically fences and livestock have same appearance. Unlike the office and road environment, networks can be learned with less data that is defined as a collision situation. In order to predict the collision situation, the non-collision situation is defined when there is a fence only and the collision situation is defined when cows stick their head out of the fence. In these defined situations, performance of segmentation is guaranteed even without deep networks due to normalized data of the appearance.
The convolutional layer is learned features of local data in a Convolutional Neural Network algorithm. The proposed Collision Grid Map transforms 3D point cloud data to grid data without destroying spatial characteristics of point cloud data. When networks are learned, a generated Collision Grid Map can be effectively applied to the Convolutional Neural Network to predict the collision situation.
When weights of convolutional layers are w in N by N size, Equation (7) is shown as the learned convolutional layer for a generated Collision Grid Map. c is the convolutional layer value and m * means the generated Collision Grid Map.
c i j = a = 0 N 1 b = 0 N 1 ( w a b m ( i + a ) ( j + b ) * )
Figure 2 applies the Collision Grid Map to Convolutional Neural Network. The network is a learned spatial feature about collision and a non-collision situation in the convolutional layer. In order to reduce computation, data size is reduced in the pooling layer. In a practical barn environment, measured data about the defined situation is filtered by the learned network to detect the local feature of the x-y plane and safety distance. Then, the SoftMax function transforms the filtered data to a probabilistic value. A high probability score is selected as an appropriate situation. When the Collision Grid Map is applied to a Convolutional Neural Network, there are some advantages that the network is learned not only in local features in the x-y plane data but also on a safety distance in the z direction to distinguish between the collision and non-collision situation.

3. Experimental Results

3.1. Experimental Setup

3D point cloud data is measured in the barn environment using D435 from intel as shown in Figure 3a. D435 has a maximum resolution 1920 × 1080 and a maximum frame rate of 30 FPS. It can be measured up to 10 m and 90° about the depth diagonal field of view. Our robot has intel core i5 4300 U processor and 8 GB RAM. This experiment has been performed using the robot equipped with a D435 shown in Figure 3b. When networks are learned, the CPU Intel core i9-9900 K and GPU Nvidia GeForce RTX 2080 Ti are used.
The experimented barn is 54.9 m in width and 6.05 m in height. Each section is 4 m wide and consists of a total of 10 sections. We experimented by setting the robot in a barn environment, as shown in Figure 4.
The data in the case of non-collision situation, when there is a fence only, has been measured using 2000 data. Similarly, the data in the case of the collision situation, when cows stick their head out of the fence, has been measured using 2000 data. The data is divided into 1800 training data and 200 test data each. Figure 5a shows the example of 3D point cloud data about the collision situation. Furthermore, Figure 5b shows the example of 3D point cloud data about the non-collision situation.

3.2. Result of the Collision Grid Map

In order to generate the Collision Grid Map, the map is set 150 * 100 in size and 0.01 m in resolution. Based on the sensor coordinate, x, y with the width of 1.5 m, the length of 1.0 m and depth data are used up to 2 m. When data is measured in the collision situation, Figure 5 shows the result of the Collision Grid Map. The example image was captured by a depth camera sensor shown in Figure 6a. Figure 6b shows raw 3D point cloud data. Figure 6c,d show the result of Collision Grid Map that consists of the first channel and second channel. In the second channel about depth data, the result shows that the nose of the cow becomes green due to the nearest distance based on a sensor coordinate. Similarly, Figure 7 shows results of the Collision Grid Map about a non-collision situation. Figure 7a,b show the image and 3D point cloud data each. Figure 7c,d show the result of the Collision Grid Map.
Table 1 shows a comparison of data size by types. The data format unifies pickle to compare data sizes. Raw point cloud data size averages 4508.9 KB per data and the Collision Grid Map data size is 5.7 KB per data. The Collision Grid Map data size shows a significant reduction relative to raw point cloud data due to a transformation more than 100,000 raw data to 150 × 100 size of pixel data. Vox grid, which is set 200 * 200 * 200 size, 0.01 m resolution and hit map type for VoxNet [25] data size shows 8.0 KB. The size of voxnet data is relatively larger than the Collision Grid Map data size since map size is different.

3.3. Convolutional Neural Network Algorithm Using the Collision Grid Map

A non-collision situation is defined when there is a fence only without any cow and in the case of the collision situation when cows stick their head out of the fence. A Convolutional Neural Network architecture using a Collision Grid Map consists of a convolutional layer, a pooling layer, and a fully connected layer. Figure 8 shows a block diagram of the network architecture. The convolutional layer is 3 * 3 in size and uses a Rectified Linear Unit (ReLu) activation function. The pooling layer is 2 * 2 in size and uses a max function. A dully connected layer has 1024 neurons. Then the Adam optimizer [29] is used for optimization.
Figure 9 shows visualization of the convolutional layer that is learned as a Collision Grid Map. The first convolutional layer is 3 * 3 in size and has 20 filters. In addition, the second convolutional layer is 3 * 3 in size and has 20 * 40 filters.
The learned filter is able to detect the local features, when the Collision Grid Map is put in the neural network. The appearance of the filtered Collision Grid Map is shown in Figure 10 and Figure 11. Figure 10 is an example of the collision situation. Similarly, Figure 11 is an example of the non-collision situation. The Collision Grid Map that is filtered by the first convolutional layer is shown in a filtered Collision Grid Map1. In addition, the Collision Grid Map that is filtered by a second convolutional layer is shown in a filtered Collision Grid Map2. The first convolutional layer is applied to the Collision Grid Map and the reacted edges are extracted from filtered collision grid map1. In the second convolutional layer, the network reacts as the distance, according to depth data as well as edges.
The graph for comparison of accuracy for VoxNet and Convolutional Neural Network using Collision Grid Map is shown in Figure 12 and Table 2. The accuracy is measured with a test set of collision and non-collision situations. Accuracy for Convolutional Neural Network using the Collision Grid Map is measured 19.2% higher than VoxNet.
Figure 13 shows the accuracy of detecting each situation according to the method. The accuracy of each situation is the failure that the end effector may collide with a cow when the cow sticks its head out and collides with the robot. When the cow actually sticks its head out, the Convolutional Neural Network using the Collision Grid Map is measured 92.7% whether or not the robot can classify the situation that the cow really sticks its head out. Voxnet is measured 5.4% less accurately than the Convolutional Neural Network using a Collision Grid Map method.

4. Conclusions

In this study, the Collision Grid Map scheme is proposed from point cloud data to predict collision between the end effector and the cow. Generated Collision Grid Map is applied to a Convolutional Neural Network. The Collision Grid Map is converted from 3D point cloud based on a probability scheme and considered an invariant permutation problem when a 3D point cloud is applied to a Convolutional Neural Network. Furthermore, the convolutional layer is learned as the safety distance in the collision situation as well as local features from the x-y plane when the Collision Grid Map is used. We show that the Collision Grid Map can be effectively managed with an aspect to the size of data. The result of the field tests and experiments performed in this study shows that the Convolutional Neural Network using the Collision Grid Map method can obtain robust performances and better prediction in a collision situation. The Collision Grid Map using point cloud data can be applied to 3D lidar, solid state lidar, and a 3D depth camera.
In the future, we would like to investigate the detection of the position method into our proposed scheme. Then, it is expected to be used for path planning of the manipulator to prevent collision. This is able to reduce the duration of time when the mobile robot feed the cows.

Author Contributions

Conceptualization, J.H.J. and C.-B.M.; methodology, J.H.J.; validation J.H.J. and C.-B.M.; writing—original draft preparation, J.H.J.; writing—review and editing, C.-B.M.; visualization, J.H.J.; All authors have read and agreed to the published version of the manuscript.

Funding

This work was carried out with the support of “Cooperative Research Program for Agriculture Science and Technology Development (Project No. PJ01386005)” Rural Development Administration, Republic of Korea.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Banhazi, T.M.; Babinszky, L.; Halas, V.; Tscharke, M. Precision livestock farming: Precision feeding technologies and sustainable livestock production. Int. J. Agric. Biol. Eng. 2012, 5, 54–61. [Google Scholar]
  2. Lehmann, R.J.; Reiche, R.; Schiefer, G. Future internet and the agri-food sector: State-of-the-art in literature and research. Comput. Electron. Agric. 2012, 89, 158–174. [Google Scholar] [CrossRef]
  3. Barbaros, F.; Gül, A.; Harmancİoğlu, N. An Earth observation assisted tool for irrigation water management. Fresenius Environ. Bull. 2013, 22, 705–713. [Google Scholar]
  4. Mutchek, M.A.; Williams, E.D. Design space characterization for meeting cost and carbon reduction goals: Smart irrigation controllers in the southwestern United States. J. Ind. Ecol. 2010, 14, 727–739. [Google Scholar] [CrossRef]
  5. Chen, J.H.; Song, K.T. Collision-Free Motion Planning for Human-Robot Collaborative Safety under Cartesian Constraint. In Proceedings of the 2018 IEEE International Conference on Robotics and Automation, Brisbane, QLD, Australia, 21–25 May 2018. [Google Scholar]
  6. Flacco, F.; Kröger, T.; De Luca, A.; Khatib, O. A depth space approach to human-robot collision avoidance. In Proceedings of the 2012 IEEE International Conference on Robotics and Automation, Saint Paul, MN, USA, 14–18 May 2012. [Google Scholar]
  7. Flacco, F.; Kroeger, T.; De Luca, A.; Khatib, O. A depth space approach for evaluating distance to objects. J. Intell. Robot. Syst. 2015, 80, 7–22. [Google Scholar] [CrossRef]
  8. Fetzner, A.; Frese, C.; Frey, C. A 3D representation of obstacles in the robots reachable area considering occlusions. In Proceedings of the ISR/Robotik 2014 41st International Symposium on Robotics, Munich, Germany, 2–3 June 2014; pp. 1–8. [Google Scholar]
  9. Xu, H.; Zhou, J.; G Asteris, P.; Jahed Armaghani, D.; Tahir, M.M. Supervised machine learning techniques to the prediction of tunnel boring machine penetration rate. Appl. Sci. 2019, 9, 3715. [Google Scholar] [CrossRef] [Green Version]
  10. Huang, L.; Asteris, P.G.; Koopialipoor, M.; Armaghani, D.J.; Tahir, M.M. Invasive weed optimization technique-based ANN to the prediction of rock tensile strength. Appl. Sci. 2019, 9, 5372. [Google Scholar] [CrossRef] [Green Version]
  11. LeCun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef] [Green Version]
  12. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. In Advances in Neural Information Processing Systems; MIT Press: Cambridge, MA, USA, 2012; pp. 1097–1105. [Google Scholar]
  13. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  14. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  15. Oh, E.; Lee, H. Development of a Convolution-Based Multi-Directional and Parallel Ant Colony Algorithm Considering a Network with Dynamic Topology Changes. Appl. Sci. 2019, 9, 3646. [Google Scholar] [CrossRef] [Green Version]
  16. Nam, G.; Choi, H.; Cho, J.; Kim, I.J. PSI-CNN: A pyramid-based scale-invariant CNN architecture for face recognition robust to various image resolutions. Appl. Sci. 2018, 8, 1561. [Google Scholar] [CrossRef] [Green Version]
  17. Qi, C.R.; Su, H.; Nießner, M.; Dai, A.; Yan, M.; Guibas, L.J. Volumetric and multi-view cnns for object classification on 3d data. In Proceedings of the IEEE conference on computer vision and pattern recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 5648–5656. [Google Scholar]
  18. Wu, Z.; Song, S.; Khosla, A.; Yu, F.; Zhang, L.; Tang, X.; Xiao, J. 3d shapenets: A deep representation for volumetric shapes. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1912–1920. [Google Scholar]
  19. Chen, X.; Ma, H.; Wan, J.; Li, B.; Xia, T. Multi-view 3D object detection network for autonomous driving. In Proceedings of the 30th IEEE Conference on Computer Vision and Pattern Recognition, CVPR, Honolulu, HI, USA, 22–25 July 2017. [Google Scholar]
  20. Masci, J.; Boscaini, D.; Bronstein, M.; Vandergheynst, P. Geodesic convolutional neural networks on riemannian manifolds. In Proceedings of the IEEE International Conference on Computer Vision Workshops, Santiago, Chile, 7–13 December 2015; pp. 37–45. [Google Scholar]
  21. Qi, C.R.; Su, H.; Mo, K.; Guibas, L.J. Pointnet: Deep learning on point sets for 3d classification and segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 652–660. [Google Scholar]
  22. Qi, C.R.; Yi, L.; Su, H.; Guibas, L.J. Pointnet++: Deep hierarchical feature learning on point sets in a metric space. In Advances in Neural Information Processing Systems; MIT Press: Cambridge, MA, USA, 2017; pp. 5099–5108. [Google Scholar]
  23. Wang, D.Z.; Posner, I. Voting for Voting in Online Point Cloud Object Detection. Robot. Sci. Syst. 2015, 1, 10–15607. [Google Scholar]
  24. Engelcke, M.; Rao, D.; Wang, D.Z.; Tong, C.H.; Posner, I. Vote3deep: Fast object detection in 3d point clouds using efficient convolutional neural networks. In Proceedings of the 2017 IEEE International Conference on Robotics and Automation (ICRA), Singapore, 29 May–3 June 2017; pp. 1355–1361. [Google Scholar]
  25. Maturana, D.; Scherer, S. Voxnet: A 3d convolutional neural network for real-time object recognition. In Proceedings of the 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Hamburg, Germany, 28 September–2 October 2015; pp. 922–928. [Google Scholar]
  26. Moravec, H.; Elfes, A. High resolution maps from wide angle sonar. In Proceedings of the 1985 IEEE International Conference on Robotics and Automation, St. Louis, MO, USA, 25–28 March 1985; Volume 2, pp. 116–121. [Google Scholar]
  27. Thrun, S. Learning occupancy grid maps with forward sensor models. Auton. Robot. 2003, 15, 111–127. [Google Scholar] [CrossRef]
  28. Zhou, Y.; Tuzel, O. VoxelNet: End-to-End Learning for Point Cloud Based 3D Object Detection. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018. [Google Scholar]
  29. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
Figure 1. (a) The example of generating the first channel in a Collision Grid Map using raw point cloud and (b) an overall Collision Grid Map.
Figure 1. (a) The example of generating the first channel in a Collision Grid Map using raw point cloud and (b) an overall Collision Grid Map.
Applsci 10 00617 g001
Figure 2. Applying the collision grid map to the convolutional neural network.
Figure 2. Applying the collision grid map to the convolutional neural network.
Applsci 10 00617 g002
Figure 3. (a) D435 form Intel. (b) Robot equipped with a D435.
Figure 3. (a) D435 form Intel. (b) Robot equipped with a D435.
Applsci 10 00617 g003
Figure 4. Appearance of the experimented barn environment.
Figure 4. Appearance of the experimented barn environment.
Applsci 10 00617 g004
Figure 5. (a) Example of 3D point cloud data about the collision situation. (b) Example of 3D point cloud data about the non-collision situation.
Figure 5. (a) Example of 3D point cloud data about the collision situation. (b) Example of 3D point cloud data about the non-collision situation.
Applsci 10 00617 g005
Figure 6. Result of the Collision Grid Map about the collision situation. (a) The example image that was captured by a depth camera sensor. (b) Raw 3d point cloud data. (c) The result of CGM in first channel. (d) The result of CGM in second channel.
Figure 6. Result of the Collision Grid Map about the collision situation. (a) The example image that was captured by a depth camera sensor. (b) Raw 3d point cloud data. (c) The result of CGM in first channel. (d) The result of CGM in second channel.
Applsci 10 00617 g006
Figure 7. Result of the Collision Grid Map about the non-collision situation. (a) The example image that was captured by a depth camera sensor. (b) Raw 3d point cloud data. (c) The result of CGM in first channel. (d) The result of CGM in second channel.
Figure 7. Result of the Collision Grid Map about the non-collision situation. (a) The example image that was captured by a depth camera sensor. (b) Raw 3d point cloud data. (c) The result of CGM in first channel. (d) The result of CGM in second channel.
Applsci 10 00617 g007
Figure 8. Block diagram of the network architecture.
Figure 8. Block diagram of the network architecture.
Applsci 10 00617 g008
Figure 9. Visualization of the convolutional layer that is a learned Collision Grid Map.
Figure 9. Visualization of the convolutional layer that is a learned Collision Grid Map.
Applsci 10 00617 g009
Figure 10. Appearance of the Filtered Collision Grid Map about the collision situation.
Figure 10. Appearance of the Filtered Collision Grid Map about the collision situation.
Applsci 10 00617 g010
Figure 11. Appearance of the Filtered Collision Grid Map about the non-collision situation.
Figure 11. Appearance of the Filtered Collision Grid Map about the non-collision situation.
Applsci 10 00617 g011
Figure 12. Comparison of accuracy for VoxNet and the Convolutional Neural Network using the Collision Grid Map.
Figure 12. Comparison of accuracy for VoxNet and the Convolutional Neural Network using the Collision Grid Map.
Applsci 10 00617 g012
Figure 13. Accuracy of the detection of each situation according to the method.
Figure 13. Accuracy of the detection of each situation according to the method.
Applsci 10 00617 g013
Table 1. Comparison of data size by types.
Table 1. Comparison of data size by types.
DataSize (KB)
Raw point cloud4508.9
Vox grid data8.0
Collision Grid Map5.7
Table 2. Comparison of accuracy results shown in Figure 11.
Table 2. Comparison of accuracy results shown in Figure 11.
MethodAccuracy (%)
Vox grid data73.9
Collision Grid Map93.1

Share and Cite

MDPI and ACS Style

Jo, J.H.; Moon, C.-b. Learning Collision Situation to Convolutional Neural Network Using Collision Grid Map Based on Probability Scheme. Appl. Sci. 2020, 10, 617. https://doi.org/10.3390/app10020617

AMA Style

Jo JH, Moon C-b. Learning Collision Situation to Convolutional Neural Network Using Collision Grid Map Based on Probability Scheme. Applied Sciences. 2020; 10(2):617. https://doi.org/10.3390/app10020617

Chicago/Turabian Style

Jo, Jun Hyeong, and Chang-bae Moon. 2020. "Learning Collision Situation to Convolutional Neural Network Using Collision Grid Map Based on Probability Scheme" Applied Sciences 10, no. 2: 617. https://doi.org/10.3390/app10020617

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop