Next Article in Journal
Laser-Generated Scholte Waves in Floating Microparticles
Next Article in Special Issue
Convolutional Model with a Time Series Feature Based on RSSI Analysis with the Markov Transition Field for Enhancement of Location Recognition
Previous Article in Journal
Fluxgate Sensor with Bifactor Excitation Mode
Previous Article in Special Issue
An Adaptive Calibration Algorithm Based on RSSI and LDPLM for Indoor Ranging and Positioning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Water Surface Targets Detection Based on the Fusion of Vision and LiDAR

1
School of Information Engineering, Southwest University of Science and Technology, Mianyang 621010, China
2
Laboratory of Science and Technology on Marine Navigation and Control, China State Shipbuilding Corporation, Tianjin 300131, China
3
Engineering Product Development Pillar, Singapore University of Technology and Design, Singapore 487372, Singapore
4
Tianjin Navigation Instrument Research Institute, Tianjin 300131, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(4), 1768; https://doi.org/10.3390/s23041768
Submission received: 9 January 2023 / Revised: 28 January 2023 / Accepted: 3 February 2023 / Published: 4 February 2023
(This article belongs to the Special Issue Multi‐Sensors for Indoor Localization and Tracking)

Abstract

:
The use of vision for the recognition of water targets is easily influenced by reflections and ripples, resulting in misidentification. This paper proposed a detection method based on the fusion of 3D point clouds and visual information to detect and locate water surface targets. The point clouds help to reduce the impact of ripples and reflections, and the recognition accuracy is enhanced by visual information. This method consists of three steps: Firstly, the water surface target is detected using the CornerNet-Lite network, and then the candidate target box and camera detection confidence are determined. Secondly, the 3D point cloud is projected onto the two-dimensional pixel plane, and the confidence of LiDAR detection is calculated based on the ratio between the projected area of the point clouds and the pixel area of the bounding box. The target confidence is calculated with the camera detection and LiDAR detection confidence, and the water surface target is determined by combining the detection thresholds. Finally, the bounding box is used to determine the 3D point clouds of the target and estimate its 3D coordinates. The experiment results showed this method reduced the misidentification rate and had 15.5% higher accuracy compared with traditional CornerNet-Lite network. By combining the depth information from LiDAR, the position of the target relative to the detection coordinate system origin could be accurately estimated.

1. Introduction

With the increase of military and civil affairs applications of USV (Unmanned Surface Vessel), it is becoming important to enhance the USV’s ability to identify and locate water surface targets. In order to ensure maritime safety, the ability to identify and locate warship targets is very important to protect maritime territorial rights and interests [1]. Additionally, in modern marine operations, precise weapons guidance has high requirements for identifying and localizing the target. In the civil sector, the USV for water surface cleaning and environment monitoring [2] also requires accurate target identification and localization.
In the present day, vision detection [3], LiDAR detection [4], and multi-sensor fusion detection [5] are the main methods used for target detection on the water surface. The mainstay of vision detection is the convolutional neural network that is used to train the target sample to obtain the model and feature information that can be used for subsequent detection. However, it is easily affected by factors such as light, foam, reflection, and others present in a water environment. A LiDAR sensor learns through multi-view input, which can estimate the three-dimensional geometric structure of a target. While 3D LiDAR has been widely used in target detection, there are still many difficulties associated with its practical application. The first is that target segmentation is not easy in the background environment, especially for dynamic targets. Secondly, most current methods are based on surface texture or structural characteristics, which depend on the amount of data. Finally, the computation efficiency becomes very low when it comes to meeting the accuracy requirement.
The fusion of multi-sensor data provides multi-level information from different sensors combination, resulting in a consistent interpretation of the observation environment. The actual scene presents two challenges. The view angle of the sensor is the first one. It is important to understand that the camera images are obtained from the cone of vision, while the LiDAR point clouds are acquired from the real world. The second difficulty is the different data representations. Since the image information is dense and regular, while the point cloud information is sparse and disordered, the fusion in the feature layer will have an impact on the computation effectiveness due to different representations. CornerNet-Lite network is a lightweight network of target detection. Compared with the YOLO series algorithms, the methods based on this network is improved in terms of accuracy and speed. In such structure, different sizes targets are scaled, and small objects on water surface are easier to be found. Next, CornerNet-Lite network is adopted in this paper.
We propose a detection method based on the fusion of 3D LiDAR point clouds and visual information to identify and locate water surface targets. We achieve the target detection with the following three steps. Firstly, the model is trained using the CornerNet-Lite [6] lightweight network, and the confidence and pixel coordinates of the target are calculated. Simultaneous determination of visual inspection bounding boxes. Secondly, the processed point clouds are projected onto the two-dimensional pixel plane according to the extrinsic calibration matrix of the camera and LiDAR, and the pixel area of the point clouds is determined with the projection point cloud amount in the bounding box from the previous step. The LiDAR detection confidence (LDC) is determined by the bounding box area and the point cloud pixel area, the target confidence (TC) is determined by combining the camera detection confidence (CDC) with the LiDAR detection confidence (LDC), and the water surface target is determined by combining the detection threshold with the TC. Finally, the bounding box of the detection result is used to obtain the corresponding 3D point clouds. The relative position between the target and the detection coordinate system origin (DCSO) is obtained by calculating the 3D minimum bounding box of the point clouds. Experiments demonstrate that this method eliminates the influence of ripples and reflections in the water and reduces the false rate from water target recognition. With the LiDAR depth information, the target position in the detection coordinate system can be accurately computed.
The remainder of the paper is organized as follows: a discussion of some research explanations related to this topic is provided in Section 2. Detailed explanations of the proposed approach are provided in Section 3. The results and analysis of the experiment are presented in Section 4. Conclusions are drawn in Section 5 with a discussion of future works.

2. Related Work

2.1. Water Surface Targets Detection Based on the Vision

For the traditional water surface target detection method, researchers made use of a large number of sliding windows with different sizes to traverse each image from the camera, and then use the manual feature [7,8] and support vector machine [9] classifier to recognize the target object. As the sliding process of windows is redundant and the expression capability of manual features is limited, it is very difficult to adapt traditional methods to detect targets on water surfaces. Recent advances in convolutional neural networks have made this detection possible. The methods from such advances are typically divided into two types: one-stage methods and two-stage methods. The two-stage method first extracts the candidate areas from the image and then classifies the targets within the candidate areas and regresses the bounding box, such as Fast R-CNN [10] and Faster R-CNN [11]. The one-stage approach estimates the target directly based on a pre-defined anchor frame, such as YOLO [12,13,14] or SSD [15]. The two-stage method has a greater advantage in terms of detection accuracy, while the one-stage method has higher efficiency.
Lin [16] et al. introduced the channel attention mechanism into Faster R-CNN, which could improve the target detection accuracy by suppressing redundant features. Cheng [17] et al. used blank label training and optimized YOLOv4 to augment the target network. Ma [18] et al. improved the YOLO v3 and KCF algorithms to obtain accurate identification and real-time tracking of multiple targets on the water surface. For the particularity and complexity of the water surface environment, it is generally recommended to improve the accuracy of the detection accuracy by improving the training network model, but the impact of reflection and illumination has not been adequately addressed.

2.2. Water Surface Targets Detection Based on LiDAR

In addition to restore the target’ three-dimensional dynamic information, such as the shape, size, and space, the 3D LiDAR can capture the target point cloud in real-time. According to the input format, these methods fall into three general categories: point clouds, images, and voxels. For the method of directly operating point cloud data, Charles [19] and others put forward an improved network called PointNet++. It is capable of learning data with different scales. Concerning the method of inputting the image format after point cloud projection, Chen et al. [20] employed 2D point cloud detection networks as the framework to fuse the vision information, thereby enriching the features. As a method for voxel-based input, Muratura et al. [21] proposed the VoxNet network in 2015. In this study, LiDAR point cloud data are voxelized and combined with 3D convolution to use point cloud for network training, thereby presenting a new method for target recognition. Zhou et al. [22] used a voxel network to predict the 3D bounding box of LiDAR points.
It has been verified and analyzed by Stateczny et al. [23] that water targets can be detected by a mobile 3D LiDAR, and several small targets can be detected except inflatable targets (such as fenders or air toys). Based on 3D LiDAR, Zhou et al. [24] presented a joint detection algorithm for water surface targets called DBSCAN-VoxelNet. This algorithm has an excellent performance in suppressing clutter on the water surface. Ye et al. [25] clustered the LiDAR point clouds by improving the DBSCAN algorithm, which both recognizes close obstacles and targets at tens of meters away. Zhang et al. [26] proposed a method to detect mobile targets in water based on LiDAR point clouds and image features. The combination of camera image and LiDAR point were used to determine the types of ground targets or obstacles, the detection confidence, the distance, and the azimuth information of unmanned surface vehicles. In order to perceive, detect, and avoid obstacles, Chen et al. [27] incorporated multidimensional environmental information from camera and LiDAR. The sensor network adopted in literature [28] is also an alternative to replace our communication system. Although the point cloud information can be enriched by clustering, it is difficult to identify objects accurately, for the insufficient features from the point cloud.

3. Proposed Approach for Target Detection Based on LiDAR and Vision Fusion

3.1. Overview of the Proposed System

Using the camera solely to detect targets on the water surface may result in misidentification, identification loss, etc, due to factors such as target reflection, ripples on the water surface, and reflections on the water surface. Thus, this paper proposes a method for detecting targets on the water surface by combining LiDAR and a camera. The entire process involves three stages. In the first stage, the LiDAR and camera are used to collect the original data, and the image information is then trained by CornerNet-Lite in order to achieve target classification and confidence estimation. In the second stage, a three-dimensional point cloud is projected onto a two-dimensional pixel plane through the joint calibration of the camera and LiDAR, and a point cloud’s confidence level is determined by the ratio between the pixel area occupied by the point clouds and the one of the bounding boxes. The experiments were performed to obtain an optimal weight ratio, according to the weight ratio fusion camera and LiDAR information, and then to achieve accurate classification and detection of water targets, to obtain the fusion detection of the bounding box. In the third stage, the target position is determined using the bounding box, which is framed with the minimum bounding box of the AABB(Axis-aligned bounding box). The positioning result is defined as the central coordinate of the bounding box. As shown in Figure 1, the system structure is depicted. These three steps correspond to the following three subsections.

3.2. Detection of the Water Surface Target Based on CornerNet-Lite

The vision detection part employs the CornerNet-Lite network. The structure can be found in Figure 2. Initially, the raw image collected by the camera is input into the detection network, and the raw image data are compressed into 255 × 255 and 192 × 192 data, and the latter is expanded to 255 × 255. According to the pixel size, the objects are defined as the small target, the medium one, and the large one. With the attention maps corresponding to the possible locations, sizes, and scores of targets’ presence, these targets are predicted based on different features. In order to find the most possible target areas, the first k areas with scores exceeding a specified threshold are selected as possible target areas, and these locations are input into the lightweight Hourglass network. Based on the target frame size in the results, the detection area is enlarged once to prevent the small targets from being missed. Finally, a non-maximum suppression algorithm is used to optimize the detection of edges.
As a one-stage detection algorithm, a large number of anchor frames are not used in the CornerNet-Lite network. Among the trained image samples, only the position recognized as ground-truth is considered a positive sample, whereas the other positions are considered a negative sample. During the training process, only the negative sample loss computation in a certain circular domain near the positive sample position is reduced. The reason is that when the position close to the positive sample is incorrectly detected as the key corner, the false detection result will still generate the target bounding box and overlap the bounding box that has been correctly detected. Consequently, the circle radius in this area should satisfy the requirement that the intersection ratio of positive samples within this radius range is greater than 0.7, and the reduction in loss calculation satisfies 2D Gaussian distribution e x 2 + y 2 2 δ 2   ( x 2 + y 2 δ 2 ) , where x and y represent the distances from the pixel point to the active sample, while 2 δ 2 represents the variance, i.e.,
y c i j = { e x 2 + y 2 2 δ 2 ( x 2 + y 2 + δ 2 ) 0
If our intersection ratio is less than 0.7, there will be a large case of false recognition and reduce the accuracy of the system. Based on Equation (1), the optimized focus loss function is as follows:
L det = 1 n c = 1 C x = 1 H y = 1 W { ( 1 P c i j ) α log ( P c i j ) ( 1 y c i j ) β ( P c i j ) α log ( 1 P c i j )
where n represents the total number of detected targets in the image. When y c i j = 1, the formula in the upper bracket is satisfied, and the super-parameters are controlled by the variables α and β .
In the CornerNet-Lite backbone network, Hourglass performs up-and-down sampling from the input image sequence. Simply up-sampling and down-sampling will cause a reduction in image size, which will cause a loss of detection accuracy. Therefore, smoothing loss is used to resolve this issue. The key idea is to perform a rounding operation on the reduced coordinates during down-sampling, the reduced coordinates are rounded. For one point ( x , y ) in the image, down-sampling i times maps correspond to the coordinate ( [ x i ] , [ y i ] ) on the thermal map, and to restore the image size, there will be a coordinate offset after up-sampling. The offsets are expressed as follows:
Δ k = { x k i [ y k i ] , x k i [ y k i ] }
where k represents the number of target targets in the image, and ( x k , y k ) represents the coordinate of the k-th target. A smooth L1 loss is adopted at the corner points of the positive sample when the network makes a prediction as follows.
F o f f = 1 n k = 1 n s m o o t h L 1 L o s s ( O k , O k ^ )
Although the CornerNet-Lite network can identify most targets, it can be influenced by spray, reflection, underwater disturbances, and ripple, leading to target misidentifications, because of the instability of the water surface.

3.3. Camera and LiDAR Data Fusion for Target Detection

For camera and LiDAR, external parameter calibration is required prior to sensor data fusion. Each small square on the calibration board has a side with 108 mm length, and it is arranged as a chessboard with 8 rows and 6 columns. During calibration, LiDAR and camera data from the calibration plane are collected. An illustration of the joint calibration model is shown in Figure 3.
A T c l matrix represents the conversion matrix from a camera coordinate system to a LiDAR coordinate system, and a T c m matrix represents a conversion matrix from a calibration board coordinate system to a camera coordinate system. During the calibration process, the Calibration Toolkit module of Autoware is used. As a result of the calibration, the calibration result is automatically calculated, and the extrinsic calibration matrix of the camera and LiDAR is as follows:
T c 1 = [ R c 1 t c 1 0 T 1 ] = [ 0.0501 0.152 0.987 0.119 0.999 0.00815 0.0494 0.0193 0.000548 0.988 0.152 0.0492 0 0 0 1 ]
where R c l matrix is a 3 × 3 matrix representing the rotation of the camera with respect to the LIDAR, and t c 1 is a 3 × 1 translation matrix representing the translation of the camera with respect to the LiDAR. The matrix values are obtained using Autoware’s Calibration Toolkit module, and the result represents the rotation translation of the camera in relation to the LiDAR. The LiDAR data can be transformed into the camera coordinate system using this transformation matrix. To achieve the data fusion from camera and LiDAR, the 3D point cloud data are projected to the pixel plane with the internal reference matrix obtained from the camera calibration.
The 3D point cloud data are projected into the image using the Formula (5), which ( X t , Y t , Z t ) represents the position of the target in the camera coordinate system, ( u , v ) represents the pixel coordinates, and Z c represents the camera internal reference matrix. We collect data with the camera and LiDAR simultaneously and project the 3D point clouds onto the camera plane with the T c l obtained earlier. The projection effect is shown in Figure 4.
Z c [ u v 1 ] = K = [ R c 1 t c 1 ] [ X t Y t Z t 1 ]
Considering that the LiDAR scans 360 degrees in all directions, while the camera has a relatively fixed view field, accurate projection results can be obtained when the target lies within the co-view area of the camera and the LiDAR. Figure 4 gives an example of the LiDAR projection. Figure 4a represents the raw image of the target as seen by the camera view field. Figure 4b shows the raw LiDAR point clouds of a target in camera view, where it can be seen that water is transparent to laser rays. Figure 4c is the projection of the 3D point cloud data of the target into a 2D pixel coordinate system.
When performing water surface target detection, it is prone to misidentification due to the influence of light, ripples, and reflections. After the introduction of LiDAR point cloud data, the number of point clouds in each recognition box is counted to determine whether the target box is a real target, thus playing a role in eliminating false recognition. The algorithm for fusion camera and 3D LiDAR detection is shown in Algorithm 1.
Algorithm 1: Fusion detection of LiDAR and camera
Input:
  Input a picture and the corresponding PCD
Output:
  Fusion detection results
  Target box diagonal coordinates
1: Get the result of CornerNet-Lite
2: Have S b o x , N b o x , C o n f c a m e r a
3: For all i in S b o x do
4:   For all point 3d in point 2d, do
5:     Get I (x, y)
6:   End for
7:   Statistical points quantity S l a s e r in S b o x
8:   C o n f l a s e r = λ ( S l a s e r / S b o x )
9:   Fusion confidence: α 1 C o n f l a s e r + α 2 C o n f c a m e r a > Y
10: End for
With CornerNet-Lite, most floating targets on the water surface can be detected and located, with the target coordinates and the detection confidence degree of each target, which we define as the CDC degree C o n f camera . After the 3D point clouds data have been obtained from the LiDAR, they are converted into a PCD format with a system time stamp, while making sure that each PCD frame corresponds to its image frame. According to the conversion matrix obtained from the camera calibration, the 3D point cloud is projected onto the 2D image plane. Considering a large number of 3D point clouds, it is necessary to filter some with their coordinate values before projection, thus reducing the amount of computation required. On the 2D image plane, the set of projected points I have the coordinate ( x , y )   , and the principle of the screening is shown in Equation (6).
I = { I ( x , y ) i f ( 0 x 640 )   a n d   ( 0 y 480 ) I o t h e r w i s e
For the sample image of the water environment, after projecting the point clouds onto the 2D image plane, the water environment does not have any projected point clouds, while the projected point clouds will be distributed on the floating target. We define the ratio of the projected area of the point clouds to the area of the bounding box pixels as the LDC C o n f l a s e r , as illustrated in Equation (7).
C o n f l a s e r = { λ S l a s e r S b b o x λ S l a s e r S b b o x < 1 1 otherwise
As shown in the equation above, S b b o x represents the pixel area of the prediction box corresponding to the candidate target, S l a s e r represents the pixel area occupied by the point clouds in the prediction box in the fused image, and λ represents the pixel area adjustment factor, that is, the proportion of the pixel area that corresponds to the adjusted point clouds data in the pixel area of the prediction box. According to 400 statistics test sets, the number of point clouds reflected back from a little object was small, with a ratio of point cloud area to prediction boxes of approximately 0.015 to 0.020. For small to medium-sized floating targets, the ratio between the area of the point clouds and the area of the prediction box is 0.020 to 0.025. In addition, the larger floats have an area ratio at least 0.025. This ratio occurs most frequently around 0.025, as shown in Figure 5. Therefore, the pixel area adjustment factor λ in this paper is taken as 40. All results are taken as 1 when λ ( S l a s e r / S b b o x ) are greater than or equal to 1. In the case of area ratio for small target detection, C o n f l a s e r takes the middle value of 0.7.
Following data fusion, the final confidence determination is shown in Equation (8). C o n f c a m m e r a represents the confidence level obtained from the detection of the target by the CornerNet-Lite network. C o n f l a s e r represents the confidence level of LiDAR detection as defined in Equation (7). α 1 , α 2 represent the weighting coefficients and the sum of the two values is 1. The target is considered to be detected when the final obtained target confidence level C o n f is greater than a set threshold Y. Otherwise, it is determined that no target is detected or that reflections or ripples cause interference. In visual inspection, we use a non-maximum suppression method to mark targets with C o n f c a m e r a greater than 0.7. At the same time, we only keep the detection results with C o n f laser greater than 0.7. With linear combination, our fusion detection confidence level is greater than 0.7.
C o n f = { α 1 C o n f c a m e r a + α 2 C o n f l a s e r α 1 + α 2 = 1
The statistical results shown in Figure 6 came from 200 frames of image data and LiDAR point cloud. In these data, there are over 1000 floating targets. The number of validly detected targets is determined by varying the image confidence weight α1 and the final decision threshold Y. According to the statistical results, with a determination threshold Y greater than 0.6, more floating targets are sifted out and an excessively high confidence level threshold is not appropriate. Even though the targets are essentially detected when the determination threshold is lower than 0.4, there will occur multiple frames of the same target, or even misidentification. Therefore, a too-low determination weight is also inappropriate. With a Y value of 0.5 and α1 value of 0.6, the best detection is achieved.

3.4. Creation of the 3D Minimum Wraparound Box

With a simple shape and slightly larger volume, a wraparound box involves wrapping a complex shape model in a bounding box. It is widely used in collision detection and ray tracing. After replacing the modeled target with a wraparound box, the light first makes a quick intersection test with the box. If the light does not cross the box, it does not intersect the modeled target. Since elements that do not intersect are eliminated, the intersection tests number is significantly reduced, and geometric computations become more efficient. Considering the relationship among the wraparound box tightness, the intersection test complexity and the boxes number, this paper adopts the AABB wraparound box with a simple intersection and poor tightness. According to the AABB definition, it is the smallest 6 hexahedron that wraps a target, and its edges are parallel to the axes. Therefore, an AABB can be described with only six parameters.
Make the flush coordinates of camera image pixel as p c = ( u c , v c , 1 ) , the coordinates under the camera coordinate system { C } as P c = ( X c , Y c , Z c ) T , the coordinates under the LiDAR coordinate system { L } as P l = ( X l , Y l , Z l ) T , corresponding to the coordinates in the world coordinate system as P w = ( X w , Y w , Z w ) T . As a final localization result, the target position needs to be converted to the coordinates under the world coordinate system. In Figure 7, the conversion model is shown. From the pinhole camera model, we can obtain:
Z c p c = K P c
Suppose R c and t c are the rotation and translation matrices of { C } relative to { W } and point P c are related to point P l as follows:
P c = R c P w + t c
Suppose R l and t l are the rotation and translation matrices of { L } relative to { W } and point P c are related to point P w as follows:
P l = R l P w + t l
Combining the two equations, we can obtain:
P l = R l R c 1 P c + t l R l R c 1 t c
Therefore, the rotation and translation matrices of the camera and the LIDAR are R l R c 1 and t l R l R c 1 t c , respectively, which are noted as R c l and t c l . With joint calibration, these two parameters can be calculated. Combining Formula (9), we can obtain:
P l = Z c R c l K c 1 p c + t c d
Therefore, the coordinates P w of the point P l expressed under { W } are expressed as:
P w = T l w P l = T l w ( Z c R c l K c 1 p c + t c d )
Inferring the 3D minimum wraparound box from the anchored box obtained from target detection, it is necessary to convert the 3D boundedness to a constraint on the 2D images by inferring the 3D minimum wraparound box from the anchored box obtained from target detection. Suppose the CornerNet-Lite detection result in image of the i-th image as S i , and iterate over S i to obtain the top-left, top-right, bottom-left, and bottom-right vertex coordinates of the anchored frame as T ( u l , v l ) , T ( u r , v r ) , D ( u l , v l ) , D ( u r , v r ) , respectively. Based on Equation (14), the vertices coordinates can be determined. In this case, the four vertices are connected to the optical center of the monocular camera, resulting a quadrilateral cone with the camera’s optical center as its apex. Next, the target’s anchoring frame locates in the quadrilateral. Due to projecting and extending the quadrilateral from the optical center with several different viewpoints, the target 3D region can be obtained, and the maximum and minimum values on the three axes of the region are taken as the minimum bounding box in 3D, as illustrated in Figure 8.
Define a single 3D plane as P i = ( p x i , p y i , p z i , p w i ) , the corresponding linear harness relationship can be listed as,
P i X 0
AABB wraparound box can be expressed as:
R = { ( x , y , z ) | min x max x , min y max y , min z max z }
In the formula, min x , max x , min y , max y , min z , max z represent the minimum and maximum values of the hotspot minimum box, which are projected on three rectangular axis. Each perspective corresponds to a quadrilateral which can be viewed as four planes intersecting. The six parameters of the AABB wraparound box is transformed into solving the intersection of 4n different planes. Therefore, in order to determine box parameter min x , the linear constraint relationship will be as follows:
p x i x + p y i y + p z i z + p w i 0 , i = { 1 , 2 , , 4 n }
There is a similar calculation method for the other five parameters as well. The problem falls into the category of linear programming. By using the simple method, the optimal solution is obtained through finite-step iterations. An AABB wraparound box plot of the fused detection target in the camera view is shown in Figure 9.
In order to evaluate localization error, the Root Mean Square Error (RMSE) in [29] is introduced in this paper.

4. Experimental Results

We validated the proposed method with an USV as shown in Figure 10. The computer is equipped with an Intel i7-10750H processor running at 2.6 GHz, 16 GB of RAM and a GTX1650Ti GPU card, the operation system is Ubuntu 18.04 LTS, the LiDAR is a Velodyne VLP-16 3D, and the camera model is Spedal 920PRO. The relative position between the camera and LiDAR is shown in Figure 10.

4.1. Surface Target Detection

If only vision is used to detect targets on the water surface, excessive light and reflections can cause loss of recognition, and waves can cause misidentification. Since the laser ray emitted from LiDAR can pass through the transparent medium without returning the point clouds, this characteristic can be used to solve the above problems. All data for experiments were collected from one lake. This dataset contains 4800 640*480 pixels images, along with 2400 frames of LiDAR point cloud information. These 3D point cloud information are processed into PCD format aligned with the timestamp of the image. One image in 0.5 s is selected for detection with the PCD data.
As shown in Figure 11, CornerNet-Lite detected the following results. Figure 11a demonstrates the significant misidentification of vision detection as a result of ripple effects. Due to the reflection of the target in Figure 11b, a target appears to be misidentified as well. Similarly, Figure 11c illustrates misidentification due to the effects of light and perspective. Based on the results of this experiment, it is evident that tar-get identification in the water environment is affected by factors such as the reflection of the target and the ripples on the surface of the water. In order to address the above problem, we make use of 3D LiDAR to acquire the point clouds of the surface object on the water.
When the LiDAR data are merged with the camera data to examine the samples in Figure 12, it is clear that the misidentifications caused by ripples, reflections, lighting, and other factors are well eliminated, and this is because the LiDAR point clouds do not return data in the transparent medium. By recalculating the prediction frame confidence levels based on the fused confidence Equation (8), it is possible to significantly decrease the false identification rate. Due to its distance far from the LiDAR and the lack of LiDAR data on the target, the target in the top right corner of Figure 12b is not identified. This is due to the fact that our detection results by fusion are influenced by the LiDAR detection confidence parameter. After a large number of experiments, it is seemed that the detection results can achieve an optimal solution when the visual detection weight is 0.6 and the LiDAR weight results are at 0.4. These two weights can be modified accordingly for different experimental situations. From Equation (8), can be seen that the confidence level can reach above 0.7.
To verify the accuracy, our detection algorithm is compared with three algorithms: CornerNet-Lite, YOLOv3, and YOLOv5.For each algorithm, we count the number of correct recognitions and the number of false recognitions. The accuracy rate is calculated as the ratio of the number of correct recognitions to the total number of targets in the field of view, as shown in Table 1 and Figure 13.
Table 1 indicates that the recognition accuracy of YOLO v5 is relatively low because the water surface targets are small targets with small sizes, and the downsampling times of YOLO v5 is large, which makes it difficult to learn feature information about small targets. Although YOLO v3 has better detection results than YOLO v5, the detection accuracy of small targets remains relatively low, and some false recognitions occur. There was an accuracy improvement for CornerNet-Lite algorithm over YOLO v3, but the number of false positives increased to four. Based on CornerNet-Lite, the proposed algorithm fused LiDAR point clouds projection data. As a result, the false recognition is decreased and the accuracy is improved.

4.2. Positioning of the Target AABB Bounding Box

In Section 3.4, it is described about how the projected LiDAR point cloud is converted from camera pixel coordinates to world coordinates. The AABB bounding box for the target in Figure 9 corresponds to the fusion detection result in Figure 11c. With the standard distance measurement method, the location relative to the detection coordinate system origin is acquired, and the 3 coordinates values are defined as the “true position”. For the proposed algorithm, the center of the bounding box is defined as the “estimated position“. To validate the accuracy, the distance deviation between “true position” and “estimated position” is defined as an “error”. According to the “error”, it is obvious that the detection target can be localized relatively accurately. The positioning results are shown in Table 2.
Above objects’ position are based on the fused detection. To achieve three-dimensional positions of the targets, the two-dimensional pixel coordinates of the target are converted to the coordinates in detection system, with the same confidence level in Section 4.1. We calculated the RSME for x, y, z in this table: RSME(x) = 0.073, RSME(y) = 0.079, RSME(z) = 0.065. The RMSE are all less than 0.1, indicating that the positioning results are reliable.

5. Conclusions

This paper proposed a method for recognizing and localizing targets by fusing vision and LiDAR information for unmanned surface vessel. There are two important works implemented. Firstly, the boundary box of target is obtained by vision detection, and the influence of reflection and ripple is solved by fusing LiDAR data. Our algorithm has 15.5% improvement in accuracy compared with traditional CornerNet-Lite. Secondly, the pixel coordinates of the target are determined by fusing the detection results, and the target spatial coordinates in the detection coordinate system are calculated with AABB bounding box. Several experimental tests have demonstrated that the method is more accurate than CornerNet-Lite algorithm, and it can eliminate the misidentification caused by reflections and ripples on the water surface. With the 3D minimum bounding box from target image bounding box, the detection target position is calculated.
Limitation and future work. The obvious limitation of this paper is the resolution of a cheap mechanical LiDAR. The farther the target is from the LiDAR, the fewer point cloud data are acquired from the target surface. In the target localization, the main error reason is the time synchronization of the camera and LiDAR data. In future work, the solid-state LiDAR will be integrated with small target detection algorithms to enhance accuracy, and the inter-frame interpolation is used to achieve time synchronization between the camera and LiDAR to reduce localization errors. In addition, regarding the relation between visual detection weight and LiDAR weight, some principles deduction work will be arranged.

Author Contributions

Data curation, L.W.; methodology, L.W. and B.Z. (Baorui Zhangand); writing—original draft, L.W.; writing—review and editing, Y.X., B.Z. (Bin Zhao) and R.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the Laboratory of Science and Technology on Marine Navigation and Control, China State Shipbuilding Corporation (2022010105), Natural Science Foundation of China (12175187), Natural Science Foundation of Sichuan Province (2023NSFSC0505).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Lee, J.; Nam, D.W.; Lee, J.; Moon, S.; Oh, A.; Yoo, W. A Study on the Composition of Image-Based Ship-type/class Identification System. In Proceedings of the 2020 22nd International Conference on Advanced Communication Technology (ICACT), Pyeongchang, South Korea, 16–19 February 2020; pp. 203–206. [Google Scholar] [CrossRef]
  2. Song, X.; Jiang, P.; Zhu, H. Research on Unmanned Vessel Surface Object Detection Based on Fusion of SSD and Faster-RCNN. In Proceedings of the 2019 Chinese Automation Congress (CAC), Hangzhou, China, 22–24 November 2019; pp. 3784–3788. [Google Scholar] [CrossRef]
  3. Yao, W.; Ming, H. An integration method for detection and classification of sea surface targets. In Proceedings of the IET International Radar Conference (IET IRC 2020), Online, 4–6 November 2020; pp. 990–993. [Google Scholar] [CrossRef]
  4. Zhou, Z.; Li, Y.; Cao, J.; Di, S.; Zhao, W.; Ailaterini, M. Research on Surface Target Detection Algorithm Based on 3D Lidar. In Proceedings of the 2021 International Conference on Security, Pattern Analysis, and Cybernetics (SPAC), Chengdu, China, 18–20 June 2021; pp. 489–494. [Google Scholar] [CrossRef]
  5. Wang, P.; Liu, C.; Wang, Y.; Yu, H. Advanced Pedestrian State Sensing Method for Automated Patrol Vehicle Based on Multi-Sensor Fusion. Sensors 2022, 22, 4807. [Google Scholar] [CrossRef] [PubMed]
  6. Law, H.; Teng, Y.; Russakovsky, O.; Deng, J. Cornernet-lite: Efficient keypoint based object detection. arXiv 2019, arXiv:1904.08900. [Google Scholar]
  7. Dalal, N.; Triggs, B. Histograms of oriented gradients for human detection. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05), San Diego, CA, USA, 20–25 June 2005; Volume 1, pp. 886–893. [Google Scholar]
  8. Lowe, D.G. Distinctive image features from scale-invariant keypoints. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar] [CrossRef]
  9. Suykens JA, K.; Vandewalle, J. Least squares support vector machine classifiers. Neural Process. Lett. 1999, 9, 293–300. [Google Scholar] [CrossRef]
  10. Girshick, R. Fast r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  11. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. Adv. Neural Inf. Process. Syst. 2015, 28. [Google Scholar] [CrossRef]
  12. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  13. Redmon, J.; Farhadi, A. YOLO9000: Better, faster, stronger. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7263–7271. [Google Scholar]
  14. Redmon, J.; Farhadi, A. Yolov3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  15. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.-Y.; Berg, A.C. Ssd: Single shot multibox detector. In European Conference on Computer Vision; Springer: Cham, Germany, 2016; pp. 21–37. [Google Scholar]
  16. Lin, Z.; Ji, K.; Leng, X.; Kuang, G. Squeeze and excitation rank faster R-CNN for ship detection in SAR images. IEEE Geosci. Remote Sens. Lett. 2018, 16, 751–755. [Google Scholar] [CrossRef]
  17. Cheng, L.; Deng, B.; Yang, Y.; Lyu, J.; Zhao, J.; Zhou, K.; Yang, C.; Wang, L.; Yang, S.; He, Y. Water Target Recognition Method and Application for Unmanned Surface Vessels. IEEE Access 2022, 10, 421–434. [Google Scholar] [CrossRef]
  18. Ma, Z.; Zeng, Y.; Wu, L.; Zhang, L.; Li, J.; Li, H. Water Surface Targets Recognition and Tracking Based on Improved YOLO and KCF Algorithms. In Proceedings of the 2021 IEEE International Conference on Mechatronics and Automation (ICMA), Takamatsu, Japan, 8–11 August 2021; pp. 1460–1465. [Google Scholar] [CrossRef]
  19. Qi, C.R.; Yi, L.; Su, H.; Leonidas, J. Guibas Pointnet++: Deep hierarchical feature learning on point sets in a metric space. Adv. Neural Inf. Process. Syst. 2017, 30, 5105–5114. [Google Scholar]
  20. Chen, X.; Ma, H.; Wan, J.; Li, B.; Xia, T. Multi-view 3d object detection network for autonomous driving. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1907–1915. [Google Scholar]
  21. Maturana, D.; Scherer, S. Voxnet: A 3d convolutional neural network for real-time object recognition. In Proceedings of the 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Hamburg, Germany, 28 September–2 October 2015; pp. 922–928. [Google Scholar]
  22. Zhou, Y.; Tuzel, O. Voxelnet: End-to-end learning for point cloud based 3d object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 4490–4499. [Google Scholar]
  23. Stateczny, A.; Kazimierski, W.; Gronska-Sledz, D.; Motyl, W. The empirical application of automotive 3D radar sensor for target detection for an autonomous surface vehicle’s navigation. Remote Sens. 2019, 11, 1156. [Google Scholar] [CrossRef]
  24. Zhiguo, Z.; Yiyao, L.; Jiangwei, C.; Shunfan, D. Research on algorithm of surface target detection based on 3D lidar. Prog. Laser Optoelectron. 2022, 59, 278–287. [Google Scholar]
  25. Sheng, Y.; Haixiang, X.; Hui, F. Laser Radar Surface Target Detection Based on Improved DBSCAN Algorithm. J. Wuhan Univ. Technol. 2022, 46, 89–93. [Google Scholar]
  26. Zhang, W.; Yang, C.-F.; Jiang, F.; Gao, X.-Z.; Yang, K. A water surface moving target detection based on information fusion using deep learning. J. Phys.: Conf. Ser. 2020, 1606, 012020. [Google Scholar] [CrossRef]
  27. Chen, Z.; Huang, T.; Xue, Z.; Zhu, Z.; Xu, J.; Liu, Y. A Novel Unmanned Surface Vehicle with 2D3D Fused Perception and Obstacle Avoidance Module. In Proceedings of the 2021 IEEE International Conference on Robotics and Biomimetics (ROBIO), Sanya, China, 27–31 December 2021; p. 18041809. [Google Scholar] [CrossRef]
  28. Wu, R.; Dong, J.; Wang, M. Wearable Polarization Conversion Metasurface MIMO Antenna for Biomedical Applications in 5 GHz WBAN. Biosensors 2023, 13, 73. [Google Scholar] [CrossRef] [PubMed]
  29. Pan, Y.; Dong, J. Design and Optimization of an Ultrathin and Broadband Polarization-Insensitive Fractal FSS Using the Improved Bacteria Foraging Optimization Algorithm and Curve Fitting. Nanomaterials 2023, 13, 191. [Google Scholar] [CrossRef] [PubMed]
Figure 1. System Overview.
Figure 1. System Overview.
Sensors 23 01768 g001
Figure 2. CornerNet-Lite Structure Diagram.
Figure 2. CornerNet-Lite Structure Diagram.
Sensors 23 01768 g002
Figure 3. Joint Calibration Model.
Figure 3. Joint Calibration Model.
Sensors 23 01768 g003
Figure 4. LiDAR point clouds projection. (a) Original picture of the camera. (b) Target point clouds projection. (c) Projection results.
Figure 4. LiDAR point clouds projection. (a) Original picture of the camera. (b) Target point clouds projection. (c) Projection results.
Sensors 23 01768 g004
Figure 5. Histogram of point clouds area proportion statistics.
Figure 5. Histogram of point clouds area proportion statistics.
Sensors 23 01768 g005
Figure 6. Threshold and weight values versus detection effect statistics.
Figure 6. Threshold and weight values versus detection effect statistics.
Sensors 23 01768 g006
Figure 7. Conversion model.
Figure 7. Conversion model.
Sensors 23 01768 g007
Figure 8. 3D back-projection model of the anchored frame.
Figure 8. 3D back-projection model of the anchored frame.
Sensors 23 01768 g008
Figure 9. AABB wraparound box for fusion detection results.
Figure 9. AABB wraparound box for fusion detection results.
Sensors 23 01768 g009
Figure 10. Hardware architecture diagram.
Figure 10. Hardware architecture diagram.
Sensors 23 01768 g010
Figure 11. CornerNet-Lite detection results. (a) Light influence. (b) Water reflection. (c) Ripple effect.
Figure 11. CornerNet-Lite detection results. (a) Light influence. (b) Water reflection. (c) Ripple effect.
Sensors 23 01768 g011
Figure 12. Fusion detection results. (a) Light influence. (b) Water reflection. (c) Ripple effect.
Figure 12. Fusion detection results. (a) Light influence. (b) Water reflection. (c) Ripple effect.
Sensors 23 01768 g012
Figure 13. Comparison of the results of the four algorithms. (a) Original image. (b) CornerNet-Lite. (c) YOLO v3. (d) YOLO v5. (e) The algorithm in this paper.
Figure 13. Comparison of the results of the four algorithms. (a) Original image. (b) CornerNet-Lite. (c) YOLO v3. (d) YOLO v5. (e) The algorithm in this paper.
Sensors 23 01768 g013
Table 1. Recognition of the four algorithms in the three scenarios.
Table 1. Recognition of the four algorithms in the three scenarios.
AlgorithmNumber of Correct IdentificationNumber of MisidentificationsAccuracy/%
CornerNet-Lite17473.9
YOLO v313165.0
YOLO v511057.9
Our algorithm17089.4
Table 2. Target position estimates.
Table 2. Target position estimates.
Target Serial No.True 3D Position (m)Estimated 3D Position (m)Error (m)
Object 1(2.01, −0.53, 0.02)(1.964, −0.462, −0.04)0.084
Object 2(1.92, −0.11, 0.20)(1.948, −0.037, 0.137)0.100
Object 3(1.52, 0.35, −0.31)(1.395, 0.290, −0.230)0.160
Object 4(1.31, 0.88, −0.24)(1.345, 0.761, −0.143)0.157
Object 5(2.22, 0.94, 0.63)(2.196, 0.845, 0.585)0.097
Object 6(1.85, −0.12, 0.09)(1.959, −0.147, 0.070)0.198
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, L.; Xiao, Y.; Zhang, B.; Liu, R.; Zhao, B. Water Surface Targets Detection Based on the Fusion of Vision and LiDAR. Sensors 2023, 23, 1768. https://doi.org/10.3390/s23041768

AMA Style

Wang L, Xiao Y, Zhang B, Liu R, Zhao B. Water Surface Targets Detection Based on the Fusion of Vision and LiDAR. Sensors. 2023; 23(4):1768. https://doi.org/10.3390/s23041768

Chicago/Turabian Style

Wang, Lin, Yufeng Xiao, Baorui Zhang, Ran Liu, and Bin Zhao. 2023. "Water Surface Targets Detection Based on the Fusion of Vision and LiDAR" Sensors 23, no. 4: 1768. https://doi.org/10.3390/s23041768

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop