Next Article in Journal
Long-Range Displacement Meters Based on Chipped Circular Patch Antenna
Previous Article in Journal
ClothFace: A Batteryless RFID-Based Textile Platform for Handwriting Recognition
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Target Detection Method of the Unmanned Surface Vehicle under All-Weather Conditions with an Improved YOLOV3

1
The State Key Laboratory of Robotics, Shenyang Institute of Automation, Chinese Academy of Sciences, Shenyang 110016, China
2
Institutes of Robotics and Intelligent Manufacturing, Chinese Academy of Sciences, Shenyang 110169, China
3
University of Chinese Academy of Sciences, Beijing 100049, China
4
School of Automation and Electrical Engineering, Shenyang Ligong University, Shenyang 110159, China
5
Shenyang Institute of Automation, Guangzhou, Chinese Academy of Sciences, Guangzhou 511458, China
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(17), 4885; https://doi.org/10.3390/s20174885
Submission received: 28 July 2020 / Revised: 21 August 2020 / Accepted: 25 August 2020 / Published: 28 August 2020
(This article belongs to the Section Remote Sensors)

Abstract

:
The USV (unmanned surface vehicle) is playing an important role in many tasks such as marine environmental observation and maritime security, for the advantages of high autonomy and mobility. Detecting the targets on the surface of the water with high precision ensures the subsequent task implementation. However, the changes from the lights and the surface environment influence the performance of the target detecting method in a long-term task with USV. Therefore, this paper proposed a novel target detection method by fusing DenseNet in YOLOV3 to improve the stability of detection to decrease the feature loss, while the target feature is transmitted in the layers of a deep neural network. All the image data used to train and test the proposed method were obtained in the real ocean environment with a USV in the South China Sea during a one month sea trial in November 2019. The experiment results demonstrate the performance of the proposed method is more suitable for the changed weather conditions though comparing with the existing methods, and the real-time performance is available in practical ocean tasks for USV.

1. Introduction

In recent years, the unmanned surface vehicle (USV) as a typical automatic unmanned system has made considerable and rapid development. It is playing an important role in both military and civilian missions to reduce human casualties as well as to create mission efficiencies, covering submarine tracking, environmental monitoring, patrol, reconnaissance, and so on [1]. Furthermore, USVs are also implemented in hydrographic measurement or bathymetric survey in shallow water regions because of some of their special advantages [2,3,4,5]. The autonomous and reliable navigation without obstacle collision is one of the important preconditions to ensure the completion of these tasks. To achieve superior perception performance, the USV generally requires employing heterogeneous sensors covering radar, lidar, camera, and infrared sensors [6]. They provide advantages of computer vision in terms of power consumption, size, weight, cost, and the readability of data, unlike radar or LIDAR, which may require heavy equipment placed on the vehicle [7,8,9]. Therefore, vision-based target detection on the sea for USVs has received much attention. Meanwhile, the camera is becoming one of the necessary equipments for USV to perform environment perception and object detection, especially for small and low-end vehicles, and the development of highly efficient computer vision algorithms for object detection is a huge challenge in the real complex environment [10].
In previous literatures, many researchers have contributed and provided different methods to detect objects on the water surface. The traditional methods usually detect the horizon line to distinguish the water region and the sky/land region in advance, and then several ad hoc image processing algorithms are implemented to segment the potential objects in the water region below the horizon [11,12,13,14]. Although these methods are suited for the undisturbed water surface, it is difficult to search the horizon line under a complex dynamic environment (haze, fog, sn glitter, and so on) or in the areas closed to the shoreline or in a marina. These limit the applications of USVs in open water fields. In these works, manual set low-level features such as the color [14] are used to recognize the objects, which could tend to cause false detection influenced by the real environment [15].
With the development of machine learning, deep learning algorithms have been widely used in object detection. In particular, as the typical representative of deep learning methods, the convolutional neural network (CNN) has been successfully employed in many applications. Owing to the advantages in high-level feature extraction and representation from raw data automatically, CNN has achieved more significant performance in image classification [16,17,18] and speech recognition [19,20]. Faster R-CNN, proposed in 2015 [21] as an improved form of CNN with region proposal networks, has been implemented in object detection for USVs in recent years, and the identification accuracy is improved compared with other methods [22,23,24].
However, the complex realistic external environment brings large challenges to the performance of object detection. It is regrettable that, in these works, no environmental factors are taken into account in object detection. While in the long-term task, for instance, changes of light, water droplets adhere to the lens of the camera, changes in vehicle attitude caused by current, even sea fog, and water reflection take place, all environmental changes leading to the instability of the original detection method [6,25].
YOLO (You only look once) [26,27,28] is different from the region-based deep learning methods such as R-CNN (region-based convolutional neural network) [29]. A YOLO network directly performs regression to detect targets in images without requiring the region proposal network (RPN) to detect regions of interest in advance, which speeds up the detection of the target processes [30]. As the state-of-the-art version, YOLOV3 could detect the targets with high accuracy and speed, and also performs well in the detection of small-size targets. These advantages ensure that the USV could detect the targets in real time even if the targets are still far from the USV. Thus, we proposed the object detection method for USV based on YOLOV3 in this paper. Simultaneously, DenseNet is employed to improve the original YOLOV3, which could increase the stability of the detection method to the changes of the dynamic environment by reducing the feature loss in the object detection process.
The main contributions of this paper are summarized as follows. (1) Although the YOLOV3 model has an excellent performance in small-size target detection, the ever-changing dynamic ocean environments suppress the detection performance for the reason that partial features of the target are lost during the features transmitting in the deep neural network owing to both convolution and down-sampling operation. The proposed YOLOV3-dense model ensures all the features transmit with no loss between layers of deep neural network and improves the reusability of features. (2) The raw images are fed to the deep neural network as a training dataset, and the features of the raw images cannot characterize all-weather conditions and sea condition changes owing to the constraints of data acquisition platform and sea trial period, which also limist the capability of detection under the all-weather conditions. In this paper, the training dataset is augmented to rich features of weather conditions by adjusting the brightness and rotation, which improves the robustness of the model to changes in environmental conditions.
The rest of this paper is organized as follows. In Section 2, we introduce the image data pre-processing, including data acquisition and augmentation. Next, the target detection method with an improved YOLOV3 is proposed in Section 3. The evaluation of target detection in some factors is addressed in Section 4, while the experimental results are discussed in Section 5. Finally, the conclusions are provided in Section 6.

2. Image Data Pre-Processing

2.1. Image Data Acquistion

In this study, image acquisition was conducted using a forward-looking camera with 1280 × 720 pixel resolution. The camera was installed horizontally on the top of the USV developed by Shenyang Institute of Automation, Chinese Academy of Sciences to monitor the surface of the water. The USV platform and the visual system installation are shown in Figure 1. The camera model is the iDS-2DF8837I5X of Hikvision with 8 megapixels.
The image data used in this paper were collected in the South China Sea during a one month sea trial in November 2019 under sunny weather and cloudy conditions. The collection periods were throughout the day from 07:00 to 17:00. The ship was selected as the target in this paper and 3000 images of ships were collected. The collected image data covered as many different environmental conditions as possible during the day. The environmental conditions and basic parameters of the collected images are shown in Table 1. A total of 1000 ship images were randomly selected for use as the training dataset.
To make the collected image data reflect more different environmental conditions, the 1000 images were then expanded to 4000 images by implementing data augmentation methods to yield the training dataset.

2.2. Image Data Augmentation

Considering that the intensity of light illumination and USV attitude caused by the waves vary greatly during the day, this could influence the performance in both the model training and method verifying steps. Therefore, the dataset used to train the model was augmented to enrich the experimental dataset by adjusting the brightness and rotation. This step of training dataset augmentation not only should enrich the deep feature maps of targets, but also could be considered to improve the robustness of the target detection method in the realistic environment condition. The framework of data augmentation is shown in Figure 2.

2.2.1. Data Augmentation: Brightness

To improve the robustness to the luminance varies of the natural light, the original images were augmented by adjusting the brightness, and the pre-processed images were added to the training dataset. The threshold values were randomly set in the range between l m i n and l m a x . However, if the brightness images are set too high or too low, the image annotation will become difficult because the edge of the target is unclear for manual annotation. On the other hand, these images will also influence the performance of the model training. Therefore, there will be some constraints in the threshold values selection; in this work, the threshold values of the brightness were set to [0.3, 0.7] for l m i n and l m a x , respectively.

2.2.2. Data Augmentation: Rotation

Considering the influence on camera attitude from the sea waves, especially USV sailing at high speed, the training dataset was also manual augmented by rotating the image data with different angular degrees. Here, 15° and −15° were utilized in our work, thus the training data were enhanced three times after rotation. The rotated images can also improve the detection performance of the proposed method.

2.3. Image Annotation

To compare the performance with other algorithms, the images for training the model weights were converted to PASCAL VOC format. The targets in training images were labeled manually by drawing bounding boxes with a software called LabelImg, which is a graphical annotation tool designed for use in deep learning algorithms. The completed dataset is shown in Table 2.

3. Methodologies

3.1. YOLOV3

YOLO was proposed by Redmon et al. in 2016 [26], and the core structure of YOLO is a convolutional neural network that can predict multi-class targets at one time. It can realize the end-to-end target detection in a real sense, with advantages in terms of high detection accurate rate and fast speed. YOLOV3 was released in 2018, and is the state-of-the-art version of YOLO [27].
YOLO divides the input image into a grid. If the center point of the object’s ground truth falls within a certain grid, the grid is responsible for detecting the object. Each grid outputs prediction bounding boxes, and the information for each bounding box contains five values ( x , y , w i d t h , h e i g h t ,   a n d   p r e d i c t i o n   c o n f i d e n c e ). The prediction confidence is defined as follows:
C o n f i d e n c e = p r ( O b j e c t ) × I o U p r e d t r u t h , p r ( O b j e c t ) { 0 , 1 } ,
where IoU as a standard indicator in target detection defining the detection accuracy by calculating the overlap ration between the true bounding box and the bounding box predicted using detection methods.
If the target falls in the grid, p r ( O b j e c t ) = 1 , and 0 for otherwise. Then, a tensor of dimensions is predicted by building a single CNN network:
S × S × ( B × 5 + C ) ,
where S × S is the number of grids and each grid can predict B bounding boxes, and C is the number of the object classes in the model.
YOLOV3 includes a CNN feature extractor named Darknet-53 as the backbone, which is a 53 layered CNN network. Compared with the previous versions, YOLOV3 predicts boxes at three different scales and the tensor dimensions are correspondingly changed as follows:
S × S × ( 3 × ( 4 + 1 + C ) ) ,
The loss function of the YOLOV3 constitutes three parts, including coordinate prediction error, IoU error, and classification error, shown as follows:
L o s s = i = 1 S 2 E r r c o o r d + E r r I o U + E r r c l s ,
where S 2 means the number of grids covered in the input image.
The coordinate prediction error is defined as follows:
E r r c o o r d = λ c o o r d i = 1 S 2 j = 1 B I i j o b j [ ( x i x ^ i ) 2 + ( y i y ^ i ) 2 ]   + λ c o o r d i = 1 S 2 j = 1 B I i j o b j [ ( w i w ^ i ) 2 + ( h i h ^ i ) 2 ]
where λ c o o r d is the weight of the coordinate prediction error. I i j o b j = 1 if the target falls into the jth bounding box in grid i, otherwise I i j o b j = 0 . ( x i , y i , w i , h i ) are true values of a target, and ( x ^ i , y ^ i , w ^ i , h ^ i ) are information of predicted bounding box in terms of the center coordinate, height, and width.
The IoU error is defined as follows:
E r r I o U = i = 1 S 2 j = 1 B I i j o b j ( C i C ^ i ) 2 + λ n o o b j i = 1 S 2 j = 1 B I i j o b j ( C i C ^ i ) 2 ,
where λ n o o b j is the weight of the IoU error, C i is the true confidence, and C ^ i is the predicted confidence.
The classification error is defined as follows:
E r r c l s = i = 1 S 2 j = 1 B I i j o b j c c l a s s e s ( p i ( c ) p ^ i ( c ) ) 2 ,
where c means the class to which the detected target belongs, p i ( c ) is the true probability that the target belonging to class c is in grid i, and p ^ i ( c ) is the predicted probability.

3.2. DenseNet

The targets should be detected as early as possible, especially while the USV sailing at high speed, which could save sufficient time to plan subsequent operations. This requires the detection method to be sensitive to the targets with long distance, even though the features of targets are not enough at this time. Furthermore, for the complex ocean environment, the targets are usually blurred in cases of the foggy weather or the water droplets adhere to the camera lens. These bring a huge challenge in target detection. Although the YOLOV3 is sensitive to the small-scale objects, the features information of the targets is lost in the transmission of the neural network owing to convolution and down-sampling. Therefore, in this paper, DenseNet is proposed to improve the original YOLOV3, which could make more effective use of feature information [31].
Between the transition layers of the YOLOV3, a structure referred to as the Dense Block is added to ensure all the feature information has no loss in the transition. The structure of DenseNet is demonstrated in Figure 3. The Dense Blocks facilitate feature reuse and mitigate gradient vanishing.

3.3. Proposed Method

This paper takes Darknet-53 of YOLO-V3 as the basic network structure for feature extraction. Considering that DenseNet has the characteristics of feature reuse and enhances the feature propagation, the down-sampling layers in Darknet-53 are replaced with DenseNet, which is likely to cause the feature loss.
The network structure diagram of YOLOV3-dense is shown in Figure 4. Considering both the calculating cost and the network structure, the size of the input images is changed from 1280 × 720 to 416 × 416. In the improved YOLOV3 network, the DenseNet structure replaces the 26 × 26 and 13 × 13 down-sampling layers; it contains the dense-block and transition-layer. The transfer function of dense-block is made up of batch normalization (BN), rectified linear units (ReLU), and convolution (Conv), which is used for nonlinear transformation between x 0 , x 1 , , x l 1 layers. The specific operation is as follows. In the layers with 26 × 26 resolution, the input layer x 0 first applies BN-ReLU-Conv(1 × 1) operation, and then applies BN-ReLU-Conv(3 × 3) operation and outputs x 1 . x 0 is spliced with x 1 as the new input [ x 0 , x 1 ] , and the above operation is repeated to output x 2 . [ x 0 , x 1 ] is spliced with x 2 as the new input [ x 0 , x 1 , x 2 ] , and so on. Finally, the feature layer is spliced into 26 × 26 × 512 and propagates forward. In the layers with 13 × 13 resolution, the feature layer finally is spliced into 13 × 13 × 1024 and propagates forward. The transition-layer is used to connect dense-block and the feature map applies BN-ReLU-Conv(1 × 1)-average pooling in this layer to reduce the size.
In the prediction process, the YOLOV3-dense model proposed in this paper predicts the bounding boxes at three different scales: 52 × 52, 26 × 26 and 13 × 13, and improved the detection accuracy of small targets.

4. Performance Metrics

4.1. Precision, Recall, and F-Measure

To evaluate the detection performance of the proposed YOLOV3-dense model, the original YOLOV3 and the Faster-RCNN-resnet101 are also applied to the detection of targets from realistic images obtained in the sea trial via USV, as well as the YOLOV3-dense model. The precision and recall analysis is utilized as an evaluation method after the detection of targets [32]. Precision refers to the percentage of correctly identified targets from the total extracted results. A high precision value indicates that the detection results contain a high percentage of useful information (true positive, TP) and a low percentage of false alarms (false positive, FP). The false-positive rate discussed in this study refers to the percentage of the false alarms in the total results, which has a value equal to 1 – precision:
P r e c i s i o n = T P T P   +   F P .
The term recall indicates the accuracy of detecting the target objects (i.e., ships) and refers to the true-positive rate. A high recall value indicates that most of the targets have been detected. The sum of true positive and false negative (FN) equals the actual number of targets in the total images:
R e c a l l = T P T P   +   F N .
The average precision (AP) can be calculated by the precision–recall curve as follows:
A P = 0 1 p r e c i s i o n ( r e c a l l ) d r e c a l l .
In the F-measure, both precision and recall are taken into account to evaluate the overall performance of object detection. A high F-measure score indicates that the detection results contain fewer false alarms and more correct detections. The F-measure is calculated as follows:
F - m e a s u r e = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n   +   R e c a l l .

4.2. Average Detection Time Cost

Otherwise, the average detection time cost is also evaluated in the experiments for the reason that the time cost is related to the feasibility of real-time application in practice.

5. Experimental Results and Discussions

Some experiments were implemented to evaluate the performance of the proposed model. A total of 3000 original image datasets were carried out in these experiments, which were randomly subdivided into three groups including the training dataset, validation dataset, and testing dataset. The completed dataset distribution in experiments is shown in Table 3, and all experiments are run on a server equipped with Intel XEON Gold 5217 CPU and NVIDIA RTX TITAN GPU cards.

5.1. Detection Performance Evaluation

The loss curves of the proposed YOLOV3-dense and the YOLOV3 during 45 thousand iterations are shown as Figure 5. The loss of both the two models decreases gradually with the increase of the iteration, and eventually converges to a low constant. After the 45 thousand iterations, the final loss of the proposed YOLOV3-dense is 0.67, while the final loss of the original YOLOV3 is 0.68. It is notable that the proposed YOLOV3-dense has a slightly higher convergence speed compared with YOLOV3 in the early stages of training, which means the weights of the proposed method could be trained with a lower time cost.
The evaluation index covering TP, FP, AP, and F-measure of the proposed YOLOV3-dense model and the other comparison models are listed in Table 4, and the precision–recall curves for these models during testing are shown in Figure 6.
On the basis of the above results, the F-measure of the proposed YOLOV3-dense model is 0.962, which is higher than the other two models. This indicates that the comprehensive performance of the YOLOV3-dense balancing the performance of both precision and recall is superior to the other two models. The AP of YOLOV3-dense is higher than YOLOV3 and basically equal to Faster R-CNN. The YOLOV3-dense predicted 1317 targets in the testing image data with 1406 ground truth. The performance of YOLOV3-dense in the index TP and FP is better than YOLOV3. The Faster R-CNN predicted 1363 targets, which is more than both YOLOV3-dense and YOLOV3. The TP of Faster RCNN increases only two times compared with YOLOV3-dense, but the FP of Faster R-CNN reaches 51, more than 7 times higher than YOLOV3-dense. This indicates that the Faster R-CNN has a higher false alarm compared with YOLOV3-dense; in other words, more noises are falsely identified as targets when Faster R-CNN is implemented. The experimental results demonstrated that the overall detection performance of the proposed YOLOV3-dense is superior to the other two models.

5.2. Real-Time Performance Evaluation

The average detection time cost of the proposed YOLOV3-dense model is 67.5 ms for one testing image data, which is 10 ms slower than YOLOV3 because more features were processed in the YOLOV3-dense model. However, this detection speed of YOLOV3-dense is enough for the applications with USV in real time. It notable that the average detection time cost of Faster R-CNN is 963.8 ms, more than 14 times slower than the YOLOV3-dense model, though the detection performance of Faster R-CNN is slightly better than YOLOV3-dense. However, this is difficult to apply on the USV to detect the targets, especially for the fast-moving targets.

5.3. Performance of Data Augmentation

Brightness and rotation transform were used to augmented the training data to simulate the changes of the light and the ocean environment. For the purpose of evaluating the influence of the data augmentation to target detection performance, 1000 original image data and 4000 augmented image data are utilized as input to train the proposed YOLOV3-dense model, respectively. The components of the augmented data are the same as those shown in Table 2, and the same 800 testing image data were utilized to evaluate the performance. The results are shown in Table 5 and the precision–recall curves for these two models are shown in Figure 7.
The AP and F-measure of the model trained without the data augmentation are 92.44% and 0.957, respectively, while these evaluation indexes of the model trained with the data augmentation are 93.13% and 0.962, respectively. The AP and F-measure are increased by 0.69% and 0.005, respectively, through the operation of data augmentation, which verified that the data augmentation is to some extent effective to improve the detection.

5.4. Performance under Different Environment Conditions

In the realistic environment, the changes of the light and weather, as well as the water droplets adhering to the lens of the camera, would influence the target detection performance. All the detection results were reviewed manually, and the typical detection results under different environmental conditions are illustrated in Figure 8.
The upper, middle, and lower row of Figure 8 listed the detection results achieved under different environmental conditions by implementing the model YOLOV3, Faster R-CNN, and proposed YOLOV3-dense, respectively. In the case of the scattering of light caused by water droplets adhering to the lens of the camera (Figure 8a,e,i) the YOLOV3-dense and Faster R-CNN can properly identify the target falling into the region of water droplets. For the case of light reflection (Figure 8c,g,k) and the case of cloudy weather (Figure 8d,h), and Figure 8l), the YOLOV3-dense and Faster R-CNN can also properly identify more targets than YOLOV3. These validate that DenseNet is conducive to improving the detection performance in different environmental conditions. Otherwise, the proposed YOLOV3-dense has the effect of suppressing the false alarm (marked with a red dotted circle in Figure 8f) compared with the Faster R-CNN. These experimental results further demonstrate that the proposed YOLOV3-dense in this paper is robust against the changes in the environmental conditions.

6. Conclusions

This study proposed an improved YOLOV3 model by fusing DenseNet to detect sea surface targets under different environmental conditions, which is expected to enhance the environmental adaptability of the USV during a long-term task. The YOLOV3-dense model proposed in this paper takes advantage of DenseNet’s feature reuse character to optimize the sample layer of the feature extraction part in the YOLOV3 model, and to promote feature propagation. The realistic images obtained in the sea trial via USV are used to train models and evaluate the performances of the proposed YOLOV3-dense model compared with the model YOLOV3 and Faster R-CNN with ResNet-101. The F-measure of the proposed YOLOV3-dense model is 0.962, which is higher than YOLOV3 (0.958) and Faster R-CNN (0.954). Simultaneously, the AP of YOLOV3-dense achieves 93.13%, which is higher than YOLOV3 (92.47%) and basically equal to Faster R-CNN (93.21%). However, the Faster R-CNN has a higher false alarm compared with YOLOV3-dense, as the TP of Faster R-CNN is much higher. These experimental results show that the YOLOV3-dense model proposed in this paper is superior to the YOLOV3 model and has better overall performance compared with the Faster R-CNN with ResNet-101. Besides, the YOLOV3-dense model is robust to the weather changes of the realistic ocean environment and meets the requirement of the real-time prediction (67.5 ms/frame) for USVs.
The focus of future work will be on deploying the proposed model as a hardware module on USVs and implementing it to detect sea-surface targets in actual tasks. Moreover, the detection model will be optimized to accelerate the training process and further improve the detection performance.

Author Contributions

Conceptualization, Y.L. (Yan Li), J.G., X.G., and K.L.; methodology, Y.L. (Yan Li) and J.G.; software, J.G.; validation, Y.L. (Yan Li) and J.G.; investigation, Z.W. and Y.L. (Yeteng Luo); resources, W.Z.; writing—original draft preparation, Y.L. (Yan Li) and J.G.; writing—review and editing, Y.L. (Yan Li); funding acquisition, K.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the Liaoning Provincial Natural Science Foundation of China under Grant 2020-MS-031; in part by the National Natural Science Foundation of China under Grant 61821005,51809256; in part by the National Key Research and Development Program of China under Grant No. 2016YFC0300801, 2016YFC0301601, 2016YFC0300604, 2017YFC1405401; in part by the Strategic Priority Research Program of the Chinese Academy of Sciences under Grant No. XDA13030203; in part by the Instrument Developing Project of the Chinese Academy of Sciences under Grant No. YZ201441; in part by the LiaoNing Revitalization Talents Program under Grant No. XLYC1902032; in part by the China Postdoctoral Science Foundation under Grant No. 2019M662874; and in part by the State Key Laboratory of Robotics at Shenyang Institute of Automation under Grant 2017-Z13.

Acknowledgments

The authors would like to express their gratitude to all their colleagues in the Center for Innovative Marine Robotics, Shenyang Institute of Automation, Chinese Academy of Sciences. They would also like to thank all the participants of the 2019 unmanned surface vehicle experiment in the South China Sea.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Liu, Y.; Bucknall, R. Path planning algorithm for unmanned surface vehicle formations in a practical maritime environment. Ocean Eng. 2015, 97, 126–144. [Google Scholar]
  2. Naus, K.; Marchel, Ł.; Szymak, P.; Nowak, A. Assessment of the Accuracy of Determining the Angular Position of the Unmanned Bathymetric Surveying Vehicle Based on the Sea Horizon Image. Sensors 2019, 19, 4644. [Google Scholar]
  3. Giordano, F.; Mattei, G.; Parente, C.; Peluso, F.; Santamaria, R. Integrating sensors into a marine drone for bathymetric 3D surveys in shallow waters. Sensors 2016, 16, 41. [Google Scholar]
  4. Specht, M.; Specht, C.; Lasota, H.; Cywiński, P. Assessment of the steering precision of a hydrographic Unmanned Surface Vessel (USV) along sounding profiles using a low-cost multi-Global Navigation Satellite System (GNSS) receiver supported autopilot. Sensors 2019, 19, 3939. [Google Scholar]
  5. Stateczny, A.; Burdziakowski, P.; Najdecka, K.; Domagalska-Stateczna, B. Accuracy of trajectory tracking based on nonlinear guidance logic for hydrographic unmanned surface vessels. Sensors 2020, 20, 832. [Google Scholar]
  6. Liu, Z.; Zhang, Y.; Yu, X.; Yuan, C. Unmanned surface vehicles: An overview of developments and challenges. Annu. Rev. Control 2016, 41, 71–93. [Google Scholar]
  7. Almeida, C.; Franco, T.; Ferreira, H.; Martins, A.; Santos, R.; Almeida, J.M.; Silva, E. Radar based collision detection developments on USV ROAZ II. In Proceedings of the Oceans 2009-Europe, Bremen, Germany, 11–14 May 2009; pp. 1–6. [Google Scholar]
  8. Halterman, R.; Bruch, M. Velodyne HDL-64E lidar for unmanned surface vehicle obstacle detection. In Proceedings of the Unmanned Systems Technology XII, Orlando, FL, USA, 6–9 April 2010; Volume 7692, p. 76920D. [Google Scholar]
  9. Muhovič, J.; Bovcon, B.; Kristan, M.; Perš, J. Obstacle Tracking for Unmanned Surface Vessels Using 3-D Point Cloud. IEEE J. Ocean. Eng. 2019, 45, 786–798. [Google Scholar]
  10. Kristan, M.; Kenk, V.S.; Kovačič, S.; Perš, J. Fast image-based obstacle detection from unmanned surface vehicles. IEEE T. Cybern. 2015, 46, 641–654. [Google Scholar]
  11. Rankin, A.; Matthies, L. Daytime water detection based on color variation. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, Taipei, Taiwan, 18–22 October 2010; pp. 215–221. [Google Scholar]
  12. Scherer, S.; Rehder, J.; Achar, S.; Cover, H.; Chambers, A.; Nuske, S.; Singh, S. River mapping from a flying robot: State estimation, river detection, and obstacle mapping. Auton. Rob. 2012, 33, 189–214. [Google Scholar]
  13. Fefilatyev, S.; Goldgof, D. Detection and tracking of marine vehicles in video. In Proceedings of the 19th International Conference on Pattern Recognition, Tampa, FL, USA, 8–11 December 2008; pp. 1–4. [Google Scholar]
  14. Wang, H.; Wei, Z.; Wang, S.; Ow, C.S.; Ho, K.T.; Feng, B. A vision-based obstacle detection system for unmanned surface vehicle. In Proceedings of the IEEE 5th International Conference on Robotics, Automation and Mechatronics (RAM), Qingdao, China, 17–19 September 2011; pp. 364–369. [Google Scholar]
  15. Wang, H.; Wei, Z. Stereovision based obstacle detection system for unmanned surface vehicle. In Proceedings of the 2013 IEEE International Conference on Robotics and Biomimetics (ROBIO), Shenzhen, China, 12–14 December 2013; pp. 917–921. [Google Scholar]
  16. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. In Proceedings of the Advances in Neural Information Processing Systems, Lak Taheo, NV, USA, 3–6 December 2012; pp. 1097–1105. [Google Scholar]
  17. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  18. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 770–778. [Google Scholar]
  19. Hu, B.; Lu, Z.; Li, H.; Chen, Q. Convolutional neural network architectures for matching natural language sentences. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 8–13 December 2014; pp. 2042–2050. [Google Scholar]
  20. Hinton, G.; Deng, L.; Yu, D.; Dahl, G.E.; Mohamed, A.R.; Jaitly, N.; Kingsbury, B. Deep neural networks for acoustic modeling in speech recognition: The shared views of four research groups. IEEE Signal Proc. Mag. 2012, 29, 82–97. [Google Scholar]
  21. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 7–12 December 2015; pp. 91–99. [Google Scholar]
  22. Kim, H.; Boulougouris, E.; Kim, S.H. Object detection algorithm for unmanned surface vehicle using faster R-CNN. In Proceedings of the World Maritime Technology Conference, Shanghai, China, 4–7 December 2018. [Google Scholar]
  23. Yang, J.; Xiao, Y.; Fang, Z.; Zhang, N.; Wang, L.; Li, T. An object detection and tracking system for unmanned surface vehicles. In Proceedings of the Target and Background Signatures III, Warsaw, Poland, 5 October 2017; Volume 10432, p. 104320R. [Google Scholar]
  24. Yang, J.; Li, Y.; Zhang, Q.; Ren, Y. Surface vehicle detection and tracking with deep learning and appearance feature. In Proceedings of the 2019 5th International Conference on Control, Automation and Robotics, Beijing, China, 19–22 April 2019; pp. 276–280. [Google Scholar]
  25. Shi, Q.; Li, W.; Zhang, F.; Hu, W.; Sun, X.; Gao, L. Deep CNN with multi-scale rotation invariance features for ship classification. IEEE Access 2018, 6, 38656–38668. [Google Scholar]
  26. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 779–788. [Google Scholar]
  27. Redmon, J.; Farhadi, A. Yolov3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  28. Redmon, J.; Farhadi, A. YOLO9000: Better, faster, stronger. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7263–7271. [Google Scholar]
  29. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 24–27 June 2014; pp. 580–587. [Google Scholar]
  30. Tian, Y.; Yang, G.; Wang, Z.; Wang, H.; Li, E.; Liang, Z. Apple detection during different growth stages in orchards using the improved YOLO-V3 model. Comput. Electron. Agr. 2019, 157, 417–426. [Google Scholar]
  31. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  32. Li, Y.; Xia, C.; Lee, J. Detection of small-sized insect pest in greenhouses based on multifractal analysis. Optik 2015, 126, 2138–2143. [Google Scholar]
Figure 1. Unmanned surface vehicle (USV) platform and visual system installation.
Figure 1. Unmanned surface vehicle (USV) platform and visual system installation.
Sensors 20 04885 g001
Figure 2. Framework of data augmentation.
Figure 2. Framework of data augmentation.
Sensors 20 04885 g002
Figure 3. Demonstration of DenseNet structure.
Figure 3. Demonstration of DenseNet structure.
Sensors 20 04885 g003
Figure 4. Network structure diagram of YOLOV3-dense.
Figure 4. Network structure diagram of YOLOV3-dense.
Sensors 20 04885 g004
Figure 5. Loss curves of YOLOV3-dense, YOLOV3, and faster region-based convolutional neural network (R-CNN).
Figure 5. Loss curves of YOLOV3-dense, YOLOV3, and faster region-based convolutional neural network (R-CNN).
Sensors 20 04885 g005
Figure 6. Precision–recall curves of YOLOV3-dense, YOLOV3, and Faster R-CNN.
Figure 6. Precision–recall curves of YOLOV3-dense, YOLOV3, and Faster R-CNN.
Sensors 20 04885 g006
Figure 7. Precision–recall curves of YOLOV3-dense models to evaluate the performance of data augmentation.
Figure 7. Precision–recall curves of YOLOV3-dense models to evaluate the performance of data augmentation.
Sensors 20 04885 g007
Figure 8. Detection results under different envrionment conditions. YOLOV3: (ad); Faster R-CNN: (eh); YOLOV3-dense: (il).
Figure 8. Detection results under different envrionment conditions. YOLOV3: (ad); Faster R-CNN: (eh); YOLOV3-dense: (il).
Sensors 20 04885 g008
Table 1. Environmental conditions and basic parameters of the collected images.
Table 1. Environmental conditions and basic parameters of the collected images.
ConditionSizeNumberTotal Number
Sunny1280 × 72013563000
Cloudy696
Sea flog861
Water droplets adhere to lens87
Table 2. Components of training dataset.
Table 2. Components of training dataset.
Original DataBrightnessRotationTotal
Training Dataset1000100020004000
Table 3. Components of experiment dataset.
Table 3. Components of experiment dataset.
Original DataAugmented DataTotal
Training Dataset100030004000
Validation Dataset1200NA1200
Testing Dataset800800
Table 4. Detection performance for these models. TP, true positive; FP, false positive; AP, average precision; R-CNN, region-based convolutional neural network.
Table 4. Detection performance for these models. TP, true positive; FP, false positive; AP, average precision; R-CNN, region-based convolutional neural network.
ModelGround-TruthPredictedTPFPAPF-Measure
YOLOV31406131113011092.47%0.958
Faster RCNN136313125193.21%0.954
YOLOV3-dense13171310793.13%0.962
Table 5. Performance of data augmentation.
Table 5. Performance of data augmentation.
ModelIterationAPF-Measure
Training data without data augmentation45,00092.44%0.957
Training data with data augmentation93.13%0.962

Share and Cite

MDPI and ACS Style

Li, Y.; Guo, J.; Guo, X.; Liu, K.; Zhao, W.; Luo, Y.; Wang, Z. A Novel Target Detection Method of the Unmanned Surface Vehicle under All-Weather Conditions with an Improved YOLOV3. Sensors 2020, 20, 4885. https://doi.org/10.3390/s20174885

AMA Style

Li Y, Guo J, Guo X, Liu K, Zhao W, Luo Y, Wang Z. A Novel Target Detection Method of the Unmanned Surface Vehicle under All-Weather Conditions with an Improved YOLOV3. Sensors. 2020; 20(17):4885. https://doi.org/10.3390/s20174885

Chicago/Turabian Style

Li, Yan, Jiahong Guo, Xiaomin Guo, Kaizhou Liu, Wentao Zhao, Yeteng Luo, and Zhenyu Wang. 2020. "A Novel Target Detection Method of the Unmanned Surface Vehicle under All-Weather Conditions with an Improved YOLOV3" Sensors 20, no. 17: 4885. https://doi.org/10.3390/s20174885

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop