Next Article in Journal
Unsustainable Anthropogenic Activities: A Paired Watershed Approach of Lake Urmia (Iran) and Lake Van (Turkey)
Next Article in Special Issue
Real Micro-Doppler Parameters Extraction of Spinning Targets Based on Rotating Interference Antenna
Previous Article in Journal
Exploring Spatial Network Structure of the Metropolitan Circle Based on Multi-Source Big Data: A Case Study of Hangzhou Metropolitan Circle
Previous Article in Special Issue
SEAN: A Simple and Efficient Attention Network for Aircraft Detection in SAR Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

YOLO-SD: Small Ship Detection in SAR Images by Multi-Scale Convolution and Feature Transformer Module

1
Key Laboratory of Earth Exploration and Infomation Techniques (Chengdu University of Technology), Ministry of Education, Chengdu 610059, China
2
The College of Mechanical and Electrical Engineering, Chengdu University of Technology, Chengdu 610059, China
3
The School of Resources and Environment, University of Electronic Science and Technology of China, Chengdu 611731, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2022, 14(20), 5268; https://doi.org/10.3390/rs14205268
Submission received: 21 September 2022 / Revised: 17 October 2022 / Accepted: 18 October 2022 / Published: 21 October 2022
(This article belongs to the Special Issue Microwave Remote Sensing for Object Detection)

Abstract

:
As an outstanding method for ocean monitoring, synthetic aperture radar (SAR) has received much attention from scholars in recent years. With the rapid advances in the field of SAR technology and image processing, significant progress has also been made in ship detection in SAR images. When dealing with large-scale ships on a wide sea surface, most existing algorithms can achieve great detection results. However, small ships in SAR images contain little feature information. It is difficult to differentiate them from the background clutter, and there is the problem of a low detection rate and high false alarms. To improve the detection accuracy for small ships, we propose an efficient ship detection model based on YOLOX, named YOLO-Ship Detection (YOLO-SD). First, Multi-Scale Convolution (MSC) is proposed to fuse feature information at different scales so as to resolve the problem of unbalanced semantic information in the lower layer and improve the ability of feature extraction. Further, the Feature Transformer Module (FTM) is designed to capture global features and link them to the context for the purpose of optimizing high-layer semantic information and ultimately achieving excellent detection performance. A large number of experiments on the HRSID and LS-SSDD-v1.0 datasets show that YOLO-SD achieves a better detection performance than the baseline YOLOX. Compared with other excellent object detection models, YOLO-SD still has an edge in terms of overall performance.

Graphical Abstract

1. Introduction

The imaging effect of traditional optical sensing is always affected by many factors, such as clouds and illumination, while Synthetic Aperture Radar (SAR) is not. SAR has the characteristics of strong penetration and a durable working period, so it is more suitable for ever-changing marine scenes [1,2]. With the increasingly serious challenge of maritime rights, SAR technology has become one of the important tools for marine monitoring.
When processing SAR images, the detection of ships is mainly achieved by comparing changes in pixel grey values and extracting feature information, such as the structure and shape of ship objects [3]. Poor visual effects lead to unsatisfactory detection in cluttered scenes, such as nearshore and harbors. Figure 1 shows two SAR image examples in the HRSID, a simple off-shore image (a) and a complicated in-shore image (b). In Figure 1a, the ships are large, sparsely distributed, and have a clear trail. The ships in Figure 1b have the characteristics of a small area and dense distribution, which makes ship detection difficult and challenging.
For ship detection in SAR images, traditional methods are mainly based on the feature differences between sea clutter and ship objects. The traditional methods can be divided into two types depending on the object of concern, based on auxiliary features and on statistical features. Analyzing the trail and leaking oil traces behind the ships, the algorithm based on auxiliary features achieves the indirect detection of ship targets and eliminates false detection in the results [4]. However, tail detection can only be performed if there are trails in the image, which makes the employment of this algorithm limited. The algorithm based on statistical features compares the marine background and ship objects in SAR images, analyzes the contrast information, and ultimately achieves ship detection. The Constant False Alarm Rate (CFAR) detection algorithm is the first and one of the most widely used [5]. By statistical inference and probabilistic modeling, the CFAR algorithm estimates the probability density function of ocean clutter and sets an appropriate threshold to separate the objects from clutter. However, the accuracy of pixel-based CFAR detection is poor as it is easily influenced by serious clutter and other factors [6]. According to different image characteristics, CFAR needs to select appropriate sea clutter distribution models. When the image resolution is low, a Gaussian distribution or negative exponential distribution is used to describe the sea clutter [7]. Dealing with high-resolution SAR images, Qin et al. [8] used the log-cumulants method to gain the parameters and proposed a CFAR detection algorithm based on a generalized gamma distribution, which showed better results. Linking the co-polarized channels to the burst time offset between the channels, Nunziata et al. [9] proposed an innovative dual-polarization model and a CFAR method to process the full-resolution CSK PingPong SAR data to observe ships and oil platforms. Based on the scattering characteristics of ships at sea, Ferrara et al. [10] proposed a physical model that processes full-resolution Single-Look Complex (SLC) SAR information combined with an efficient filtering technique designed to achieve the high-quality identification of targets and backgrounds at sea. The above traditional algorithms are suitable for processing single as well as simple SAR images, but are not quite effective when dealing with complex maritime situations.
Since the introduction of deep learning, it has continued to evolve and received widespread attention from scholars from all walks of life. By applying this technique to image processing, the detection accuracy and speed of tasks such as target detection and instance segmentation have been significantly improved [11]. Depending on their structure, there are two types of deep learning-based detection algorithms: one-stage algorithms and two-stage algorithms [12]. The principle of the two-stage algorithm is to generate candidate frames first and then classify them on the basis of whether they contain objects or not. As a pioneer in object detection using deep learning, the R-CNN [13] algorithm has substantially improved detection accuracy compared with traditional detection algorithms. Based on R-CNN, scholars have made improvements, resulting in excellent algorithms such as Faster R-CNN [14], Mask R-CNN [15], Dynamic R-CNN [16], Sparse R-CNN [17], and Libra R-CNN [18]. Meanwhile, the one-stage model samples the image uniformly at all locations and transforms the detection mission into a regression classification task; examples are YOLO series [19,20], RetinaNet [21], YOLOF [22], etc. Generally, the one-stage algorithm adopts the end-to-end training mode, which is usually faster but has low accuracy. The two-stage algorithm can achieve great accuracy, but its computational overhead remains large.
To make further breakthroughs in the field of machine vision, Dosovitskiy et al. [23] redesigned the transformer, which was proposed by Vaswani et al. [24,25], to encode images as sequences and proposed the first visual transformer (ViT) for image classification. The best recognition results can be achieved by applying ViT in optically natural scenes. Carson et al. [26] introduced the transformer to object recognition and proposed the Detection Transformer (DETR). To reduce the computational consumption and the false drop rate involved in SAR complex backgrounds, Li et al. [27] added a transformer encoder after the backbone ResNet101 [28] and fused semantic information and location information. Srinivas et al. [29] combined the transformer with the backbone network and achieved the best results in areas of image generation and instance segmentation. These studies have all employed hybrid structures, combining CNN-based models with transformers to achieve excellent results on computer vision tasks. Inspired by this, this paper proposes the design of a kind of transformer structure for YOLOX [30] to achieve the high-efficiency detection of small ships in SAR images.
The accurate detection of small ships has always been a challenging research topic due to the characteristics of SAR images [31]. When the area of the ship is small (typically less than 482 pixels [32,33]), it is only shown as a bright spot. Lacking feature information during detection, these are easily confused with other interference, resulting in missed detection and affecting the final results. In order to improve the detection accuracy of small ships, scholars have performed a lot of research in this field.
For the poor detection of small objects caused by class imbalances, data augmentation is employed to expand the data of small ships, improve the model’s attention to them, and thus enhance the contribution of small objects to the loss function calculation during training. For the partial or even full feature loss of small objects due to down-sampling, model optimization strategies such as feature fusion are often used to increase feature information. Specifically, on the basis of the traditional SSD detector [34], Juan et al. [35] introduced the data augmentation of rotation and expansion, added dilated convolution in the backbone, and finally improved the adaptability of the model to small objects. Chen et al. [36] inserted feature pyramids into the Region Proposal Network to compensate for the loss of small ships’ location information at the bottom of the network. Yang et al. [37] designed a perceptual field enhancement module to integrate different convolutions and pooling, which enhanced the transfer of feature information and ultimately reduced the false alarms of small objects. To achieve small ship detection in PoISAR images, Jin et al. [38] replaced all normal convolutions in the network with extended convolutions when expanding the perceptual field. In the FBR-Net network proposed by Fu et al. [39], the designed ABP structure utilizes a layer-based attention method and spatial attention method to balance the semantic information of the features in each layer, which made the network more focused on small-scale ships. To improve the detection of small ships in SAR images with complex backgrounds, Guo et al. [40] combined feature refinement, feature fusion, and head enhancement approaches to design a highly accurate detector, called CenterNet++. Chang et al. [41] proposed a GPU-based deep learning detection method, called YOLOv2, which offers superior detection speed and accuracy and greatly improves the efficiency of ship detection in SAR images. To further enhance the detection effect of small-scale ships, Su et al. [42] proposed a Spatial Information Integration Network (SII-Net). In SII-Net, a Channel-Location Attention Mechanism (CLAM) block and a multi-scale pooling layer were applied to obtain richer ship position information, and interpolations and poolings were employed after the PANet to enhance the model’s attention to targets. While the network does achieve a high overall detection accuracy, it is less effective for densely distributed ship groups. Considering that contextual information is crucial for the detection of small and dense ships, Zhao et al. [43] proposed a novel CNN-based method. In this method, as many small ship proposals as possible are first made and then combined with contextual information to exclude spurious ships from the predictions.
Based on the above analysis, this paper uses the latest YOLOX algorithm as the baseline, focuses on enhancing the feature extraction capability, captures richer contextual information, and ultimately strengthens the detection ability of ships in high-resolution SAR image datasets, especially small ships. Our contributions can be summarized as follows:
  • To improve the sensitivity of minor object detection, we designed an MSC block. It combines several parallel convolutions with the residual network, which can obtain feature information of different sizes and perform multi-scale fusion, further enhancing the representation of semantic information;
  • In addition, we proposed the FTM block. It divides the high-layer feature information into two parts, processes them using the transformer encoder, and finally merges them via a cross-stage structure. With this FTM module, our model can capture global features effectively and achieve higher detection accuracy;
  • Taking YOLOX as the baseline, we incorporated MSC and FTM, and proposed an efficient detection model YOLO-SD for ship detection in high-resolution SAR images;
  • YOLO-SD was tested on the HRSID [32] and LS-SSDD-v1.0 datasets [33]. According to the experimental results, our detection accuracy was improved dramatically compared with YOLOX, which indicates the effectiveness of our model. Besides this, we compared our design with some existing excellent networks when applied to the same dataset, and the results showed that ours still excels in overall performance.

2. Material and Methods

Firstly, our method is derived by analyzing the shortcomings of the existing method, YOLOX. Next, each key point is described in detail, including the specific architecture and working principles, and the overall structure of YOLO-SD is shown. Last, we introduce the environment’s setup, the datasets used, and the evaluation metrics.

2.1. Proposed Method Based on YOLOX

In the detection network, the bottom feature map has high resolution and rich detailed features, which makes it suitable for small ship detection. In the higher feature map, the image resolution seems low, but its semantic information becomes rich, which is appropriate for detecting large-scale ships [44]. Figure 2a shows the partial structure of YOLOX. YOLOX applies CSPDarkNet53 as the backbone and mainly uses C3, C4, and C5 for feature fusion and classification, but its actual detection of small ships is poor. Firstly, the C3 layer is located in the shallower region of the network, with rich detailed information and high resolution. YOLOX mainly utilizes it for feature extraction to achieve small ship detection. However, due to the small reception field of the C3 layer, the semantic information [45] obtained is weak. Thus, YOLOX will consider some real ships (especially small ships) as background, resulting in poor final detection. Next, the C5 layer has low resolution and is highly abstract, so lots of small ships have lost some or all of their detailed features, at which point it is no longer meaningful to process the C5 layer for small ship detection.
To solve these problems of YOLOX and improve the accuracy of small ship detection in SAR images, we have used YOLOX_L as our baseline to design a new detection model for small ships, called YOLO-SD. The partial structure of YOLO-SD is shown in Figure 2b. The MSC and FTM in the backbone CSPDarkNet53 are our designs, while the red lines indicate the new connection between the backbone and the neck. Firstly, several parallel multi-scale convolutions were inserted after the C3 layer. These convolutions help the network to obtain feature information from different reception fields and enhance semantic information. Secondly, after layer C5, in which most of the detailed features were lost, we added a newly designed FTM block, mainly consisting of a multi-headed attention layer and a fully connected layer, to optimize the feature information. Lastly, to improve the effect of feature fusion, we modified the connection between the backbone and the neck. In Figure 2b, P3, P4, and P5 are connected with MSC, C4, and FTM. In this method, we apply the newly improved feature maps to enhance the network’s attention to small ships, which in turn improves the network’s detection performance. The experiments demonstrate that YOLO-SD increases the computational overhead to a lesser extent than baseline YOLOX, but improves the accuracy significantly.

2.2. Specific Architecture of YOLO-SD

In order to optimize model performance, we considered deepening the backbone network as well as expanding its width. However, stacking structures directly not only increase the computational cost significantly, but also make the network prone to scattering [46]. Therefore, we applied the basic structure of CSPDarkNet53 [47] and improved on it by introducing the following design.
We propose the MSC module to improve the effect, as shown in the following Formulas (1) and (2):
y M S C = R e l u ( x + C o n c a t ( c 0 ( x ) , c 1 ( x ) , c 2 ( x ) ) )
{ c 0 ( x ) = w 01 3 w 00 1 x c 1 ( x ) = w 11 5 w 10 1 x c 2 ( x ) = w 2 1 x
where x is the input feature map and y is the output feature map. In Formula (2), ci (i = 0, 1, 2) represents the i-th convolution branch, and * i (i = 1, 3, 5) represents the convolution with kernel size i × i. Wij (j = 0, 1) means the weight parameters of convolution, and the lower corner indicates the j-th convolution of the i-th branch.
The diagram of the MSC feature enhancement structure is shown in Figure 3. MSC is mainly a parallel filter structure, which connects the outputs of convolutions with different kernel sizes into a single output. These parallel convolutions are performed at different scales and can extract features from different receptive fields at the same time, which has two benefits. On the one hand, it improves the feature extraction effect for ships. On the other hand, deeper features can enhance the semantic information of the feature map and improve the model’s ability to detect small ships. These parallel convolution operations occupy a lot of computer resources, so we add 1 × 1 convolution before processing to alleviate the problem. The 1 × 1 convolution can both further increase the network depth and reduce the dimension (changing the number of channels to 0.5, 0.25, and 0.25 times the number of input channels), as well as reduce the computational consumption. However, network widening and deepening induces training difficulties and gradient disappearance problems while improving performance. For this reason, we introduce the ResNet structure, which directly connects the input of MSC with the output of the concatenation operation. Through this design, we process and aggregate the information while limiting the amount of computation, deepen the network while enhancing the expressive ability, and improve the sensitivity of the model to small ships.
During detection, the model mainly relies on the backbone network to extract local feature information from SAR images. However, the large down-sampling factors involved in extraction may mean the model misses small-scale ships. In addition, the model is unable to capture sufficient global information due to the small actual receptive field of the convolutional neural network. To improve the capability of small ship detection in SAR images and minimize the leakage of small targets, we propose a Feature Transformer Module (FTM) that can capture rich global and contextual information, as shown in Figure 4.
The core of the FTM is a transformer encoder, consisting of a Multi-Headed Attention (MHA) block and a Multi-Layer Perception (MLP) block overlay. As the input to the encoder is a sequence with certain orders, we partition feature maps into sequences of specific length and width patches before the encoder. Inside the encoder, MHA enables the network to obtain the location information of surrounding ships by acquiring the relationships between ships under a global receptive field. Due to the higher learning capability of the nonlinear transform, it consists of two fully connected layers with a large number of intermediate hidden units to form an MLP block, which analyzes contextual information and enhances the characterization of ship features. In addition, a residual structure is added to keep the FTM well trained even when the layers are deepened. Layer normalization is employed to normalize the feature sequence so that the ReLu activation function can play a better role afterward. To cope with the structural gradient disappearance problem, the FTM transforms the input features by two 1 × 1 convolutions, one retaining the original features and the other using the transformer encoder. Compared to dividing the channels directly, such a division allows all the input features to be transformed, improves the reusability of features effectively, and keeps the overall computing effort lower. Using the FTM at the top layer of the backbone network before inputting to the neck, through continuous learning, contextual information is linked to enhance the correlation between ships, and thus reduce the omission of small-scale ships and improve the network’s detection ability.

2.3. Overall Structure of YOLO-SD

The overall framework of YOLO-SD is shown in Figure 5. After the SAR images are input into our model, feature extraction is first implemented by the backbone network, a modified CSPDarknet. The MSC and FTM of the design are introduced to it, which means that the final feature map contains more valid small-scale ship features. In the neck network, the first fusion was performed from top to bottom to obtain P3, P4, and P5. To retain the shallow edge, shape, and other features, a bottom-up path enhancement structure was added later and achieved the second fusion to obtain N3, N4, and N5 feature maps. Finally, decoupling heads separate the classification and regression tasks to obtain more accurate detection results for small ships in SAR images.

2.4. Dataset

In order to test the practical effect of YOLO-SD, we employed the high-resolution SAR image dataset (HRSID) and LS-SSDD-v1.0 dataset. The specific information of these two datasets is shown in Table 1. The HRSID, including a total of 16,951 ship targets, cuts 136 panoramic SAR images into 5604 images with 800 × 800 pixels. In the LS-SSDD-v1.0 dataset, 15 images with 24,000 × 16,000 pixels are cut into 9000 sub-images, also with 800 × 800 pixels. The LS-SSDD-v1.0 dataset retains the pure background image, so the detection model can learn pure background features more effectively and reduce false alarms. The SAR images in two datasets were collected from Sentinel-1 and TerraSAR-X satellites with mixed HH, HV, VV and VH polarizations. With the help of Google Earth and the Automatic Identification System (AIS), all the ships in a SAR image can be completely labeled. When experimenting, the ratio of dataset division (training dataset:validation set:test set) was set to 13:7:7 and 2:1:1, respectively. Figure 6 shows the comparison of the number of ships of three sizes in the two datasets. When the datasets processed the ship targets, they were divided into three types according to the area size: small ships (area less than 482 pixels), medium ships (area between 482 and 1452 pixels), and large ships (area greater than 1452 pixels). According to Figure 6, the typical size of small ships in the two datasets is 482 pixels.

2.5. Evaluation Metrics

For the accurate evaluation of the detection performance of each model, the indexes, including the MS COCO evaluation indexes [48], FPS, Parameters, and so on, were used in this work.
Intersection Over Union (IoU) is an important and standard index to measure the accuracy of object detection in the dataset. Its calculation is defined as follows, where A represents a real object box in the dataset, and B represents the corresponding prediction box obtained by detection models:
I o U = A B A B
Recall refers to the proportion of correctly predicted samples in all real objects, while Precision means the proportion of correctly predicted samples in the objects targeted predicted by the model. Their calculation methods are shown in Formulas (2) and (3), where TP refers to True Positive and FN refers to False Negative.
R e c a l l = T P T P + F N
P r e c i s i o n = T P T P + F P
The widely employed Mean Average Precision (mAP) is the average of the accuracy of all categories. Since there is only one type in the HRSID and LS-SSDD-v1.0 (ship), the result obtained by calculating AP is the mAP. The derivation formula of AP is shown in Formula (6), where R represents Recall and P represents Precision. Generally speaking, the higher the AP corresponding to the model, the better the detection performance of this model. Since AP is obtained by integrating P(R) with R, the Precision-Recall curve can display the overall performance of algorithms.
A P = 0 1 P ( R ) d R
This work used MS COCO evaluation indexes to reliably compare the experimental results between different detection models. According to different IoU thresholds and various object characteristics, indexes can be divided into six different types, including AP, AP50, AP75, APS, APM, and APL. Once the mentioned IoU threshold is set to 0.5 and 0.75, the results obtained by Formula (6) are AP50 and AP75. If IoU gradually increases between 0.5 and 0.95 (by 0.05), the average of the ten values obtained is AP. When only objects of a specific size are calculated, such as small (the area of the detection object is less than 322 pixels), medium (322 < area < 962 pixels), and large (the area is greater than 962 pixels) objects, the averages obtained are APS, APM, and APL.
In addition to these indicators, we also introduced some other indexes, such as Frame Per Second (FPS) to evaluate detection speed and Parameters to describe model complexity. FPS represents the number of images that can be processed per second. The time required to detect each image can be obtained by taking the inverse of the FPS, as shown in the following Formula (7). In the CNN network, the parameter can describe the complexity of the model, and its calculation formula is shown in Formula (8). In Formula (8), Kh and Kw represent the size of the convolution kernel, Cin means the number of channels of the input feature map, and Cout means the number of channels of the output feature map. Therefore, the parameter of a convolutional layer can be obtained by Formula (8), and the parameter of the entire model can be obtained by adding the parameters of all layers.
t = 1 F P S
P a r a m s = ( K h · K w · C i n ) · C o u t

3. Results

3.1. Experimental Environment

All experiments were completed on a server equipped with NVIDIA GeForce RTX 3090 and 24G video memory. Besides this, we used Python 3.7 compilation language, python 1.8.1 to realize our training, and CUDA 11.1 to speed up the calculation. Furthermore, because of the limitations of the hardware and network, the batch size was set to for all experiments, which means that our server for training needed to process four SAR images at a time.

3.2. Training and Testing

In all experiments, our model and all other models were based on the mmdetection platform [49] and applied the same settings. On the HRSID and LS-SSDD-v1.0 datasets, the learning rate was set to 0.001, the number of iterations in the training epochs was 24, the momentum was 0.9, and the weight attenuation decay was 0.0001. We employed the CSPDarkNet53 backbone network parameters pretrained on the ImageNet dataset, and set the input image resolution to 512 × 512, 640 × 640, and 800 × 800 pixels. Some image processing operations were included in the training pipeline, including Mosaic, RandomAffine, MixUp, RandomFlip, Resize, and MixUp processing. For a thorough assessment, MS COCO evaluation metrics were applied to compare the experimental results of each model. In the field of deep learning, the model is considered successful in detecting the target once the IoU between its predicted box and the real object box is higher than the threshold value of 0.5.

3.2.1. Ablation Experiments for YOLO-SD

First, to demonstrate the advance of YOLO-SD more objectively, we conducted ablation experiments on the HRSID dataset and compared it with YOLOX_L. In the experiments, the input image resolution was set to 800 × 800 pixels, and MSC and FTM were excluded from YOLO-SD separately and then trained; the results are shown in Table 2. Using MSC and FTM alone resulted in a 1.2% and 3.2% increase in final ground average accuracy, respectively, and combining them resulted in a 3.8% increase in AP. In addition, the learnable parameters for MSC and FTM are 0.20 M and 5.25 M, respectively. Compared to the baseline YOLOX_L, YOLO-SD has increased the parameters by 5.45 M, about 10%. From the experimental results, it is clear that both MSC alone and FTM can optimize the ship target information in the feature map and effectively improve the detection accuracy for small ships.

3.2.2. Comparison with YOLOX at Different Scales

Based on the concept of network splitting, width refers to the number of output channels of the network, while depth is the number of layers of the network. Depending on the depth and width, YOLOX can be divided into YOLOX_S, YOLOX_M, YOLOX_L, and YOLOX_X. Specifically, the ratio of the number of layers between them is 1:2:3:4 and the ratio of the number of output channels is 2:3:4:5. To obtain better experimental results, we tested these four complexities of YOLOX and trained our model based on the best YOLOX_L. Table 3 shows the results derived from YOLOX and YOLO-SD with different input scales. When the input was 512 × 512 pixels, the AP of YOLO-SD reached 51.9%, which was 2.3% higher than the baseline. When we used 640 × 640 and 800 × 800 pixel images, the increase was 1.2% and 3.8%, respectively, compared with YOLOX_L. When the resolution of the input SAR image increases, the AP obtained by YOLO-SD shows an upward trend. When dealing with low-resolution (512 × 512) SAR images, YOLO-SD has a higher AP than the baseline YOLOX_L, which means it can detect more ships and obtain more accurate results.
According to the results, the Precision-Recall curves of each model have been drawn as shown in Figure 7. The area between the curve and the two coordinate axes is AP. From Figure 7, we can see that the AP of our YOLO-SD includes all the results of YOLOX. All these results suggest that the AP of YOLO-SD is higher, and the overall performance of YOLO-SD is significantly better than all kinds of YOLOX.

3.2.3. Comparison with Some Existing Models

In addition, we also tested some other excellent algorithms on the HRSID and LS-SSDD-v1.0 datasets to compare their performance with ours, as shown in Table 4. In contrast to Faster R-CNN, YOLO-SD achieved an AP improvement of 1.8% and took almost equal time. Compared to Libra Faster R-CNN, Mask R-CNN, Dynamic R-CNN, and Grid R-CNN, YOLO-SD was not only faster, but also had better accuracy. When facing one-stage algorithms, such as YOLOF, YOLOv3, RetinaNet, and YOLOX_L, although our computational speed was not superior, the accuracy improvement was larger, by 13.9%, 9%, 1.2%, and 3.8%, respectively. Furthermore, in terms of the average precision APS obtained for small-scale object detection, YOLO-SD achieved 62.3%, a 3.8% improvement compared to the baseline and 2.3% higher than Dynamic R-CNN, and performs best in this respect.
The results of the experiments on the LS-SSDD-v1.0 dataset are shown in Table 5. This dataset contains a high number of small ship annotations, accounting for 99.8% of all annotations. Therefore, it allows a more direct comparison of the ship detection of each model for small targets. As there is no target box larger than 962 pixels, the APL results are all indicated using “-”. As can be seen from Table 5, our remaining five accuracy metrics increased by 1.4%, 2.4%, 2.8%, 1%, and 3.3%, respectively, compared to the baseline. YOLO-SD achieved the highest accuracy at a moderate computational speed compared with the other models. The experimental results prove that YOLO-SD has the best detection capability for small ships.
Demonstrating the advantages of YOLO-SD more visually, we plotted the P-R curves of some of the single-stage models and two-stage models separately based on the experimental results on HRSID. Compared with two-stage models (Figure 8a), the Precision of YOLO-SD is significantly higher in the Recall region of 0.8~0.9. Compared with one-stage models (Figure 8b), our Precision is slightly high in the Recall region of 0.5~0.8. The above results demonstrate that the MSC block and the FTM block can effectively improve the accuracy of small ship detection in SAR images.

3.3. Detection Results and Analysis

3.3.1. Comparison with YOLOX_L

To demonstrate the superiority of YOLO-SD over the baseline YOLOX_L, some of the visual ship detection results on HRSID are displayed later. We applied two SAR images taken at a canal to detect and compare during the test. In addition, the most representative part in the original size image was selected to be magnified, which is thought to enhance the comparison between YOLO-SD, the baseline, and the true annotated boxes. In detail, the first column of the figures shows the results obtained using the baseline detection, the second column shows the real annotated boxes in the dataset, and the third column shows the results detected by our method. The first row displays all the original images, while the second row shows the results after zooming in on the intercepted part. To illustrate the superiority of YOLO-SD, some of the visual ship detection results are displayed later.
Figure 9 shows the first SAR image of the canal. Where the boats are small and densely distributed on both sides of this river (the enlarged part in Figure 9), our model (Figure 9c) detects more accurately. However, YOLO-SD produces partial false detections when dealing with some larger disturbances.
In Figure 10, there is a large number of small ships sailing in the canal. They are distributed sparsely in the upper part of the river, and both YOLOX_L and YOLO-SD perform well. However, in the lower part of the river, YOLOX_L (Figure 10a) produced a huge number of missed detections due to the extremely narrow spacing between ships, and YOLO-SD (Figure 10c) maintained its excellent performance in detecting most of the ships. Table 6 records the detailed data of YOLOX_L and YOLO-SD detection in these two scenarios, where correct refers to the number of correct boxes and wrong refers to the number of boxes detected in error. The greater accuracy of YOLO-SD suggests that our model has an advantage over YOLOX_L in detecting densely distributed small ships.

3.3.2. Comparison with Other models

To compare the performances of each model, we then took one image from each dataset, and the visual results are shown in Figure 11 and Figure 12. Figure 11 contains fewer distractions and a large number of clear small-scale ships. As the small islands and reefs occupy few pixels, many models incorrectly recognized them as ships (Figure 11d–i). Large and small-scale ships are both included in Figure 12, and there are also some land structures in the upper right section, which may interfere with detection. As the test results show, all the models found large ships, but missed some of the small ships and incorrectly identified land structures as ships.
Table 7 records the detection results obtained for YOLO-SD with other models. Only our algorithm succeeded in identifying not only the larger objects but also the smaller ships, without any false detection. In all the results, our model performs the best in terms of detection, even when compared to Mask R-CNN and Grid R-CNN (two models with high AP in Table 4 and Table 5).

4. Discussion

As can be seen from the ablation experiments described in Table 3, both MSC and the FTM contribute to the improved accuracy of ship detection. The FTM increases the AP metric by 3.2%, which is 2% higher than MSC’s 1.2%. The introduction of MSC improved the representation of feature maps in the backbone network, while also optimizing the fusion effect of the neck. With a large global field of perception, the FTM focuses on the feature information of the ship, while enhancing the correlation between ships, ultimately reducing the number of missed small objects. With a large global receptive field, the FTM focuses on the feature information of the ship while enhancing the correlation between ships, ultimately reducing the number of missed small objects. Due to the small area of ships in SAR images, feature loss becomes severe as the network deepens. Therefore, the FTM, which focuses on optimizing the depth of the network, can obtain higher detection accuracy. The two act at different network locations and there is no conflict between them, so YOLO-SD can achieve an accuracy improvement of up to 3.8%.
For YOLOX, different complexities have different advantages. YOLOX_S has the smallest number of parameters and the fastest calculation speed, while YOLOX_X has the highest detection accuracy. By comparing the PR curves, we objectively and equitably arrive at the best ground scale for YOLOX-L, and use it as the baseline for improvement. As the size of the input image was altered (from 800 to 640 and 512), the detection accuracy of all models decreased by varying degrees. In fact, when the SAR image was reduced, the ship object area became smaller and contained less feature information, making detection more difficult. Whereas MSC enables the model to detect targets from indistinguishable complex backgrounds by exploiting the rich semantic information, the FTM improves the correlation between all ships and enables the model to detect a larger number of ship objects. As a result, compared to the baseline, YOLO-SD is more capable of detecting ships and consistently obtains the highest AP.
In practical experiments on two different datasets, our model maintained accurate detection results. YOLO-SD also has the highest detection accuracy compared to other superior models, which proves its advantages.

5. Conclusions

Given the poor effect of existing models on small ship detection in complex SAR images, we propose an improved detection model based on YOLOX in this paper, named YOLO-SD. It is combined with MSC to extract different scale features and enrich semantic information, and the FTM block to optimize features. On the HRSID and LS-SSDD-v1.0, several sets of experiments were conducted to compare with the highly representative detection methods, including Faster R-CNN, Mask R-CNN, RetinaNet, etc. The experimental results show that our network performs better and achieves the highest accuracy when dealing with small ships. However, we still found some missed detection when dealing with unclear ships. In the future, we will conduct further research on small-scale ship detection methods with lower leakage rates. It is hoped that this article can help scholars find better ideas when analyzing and dealing with scenes containing dense, small targets. Both our code and the datasets used (HRSID and LS-SSDD-v1.0) are available at the link: https://doi.org/10.6084/m9.figshare.21316290.v3 (accessed on 17 October 2022).

Author Contributions

Conceptualization, S.W.; methodology, S.W. and S.G.; software, S.W. and J.L.; validation, S.W., R.L., L.Z., H.Z. and J.L.; formal analysis, S.W.; investigation, S.W. and L.Z.; resources, S.W. and R.L.; data curation, S.W.; writing—original draft, S.W. and S.G.; writing—review and editing, S.W., S.G., Y.J. and J.Q.; visualization, S.W.; supervision, S.G. and J.L.; project administration, S.G.; funding acquisition, S.G. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (NSFC) under Grant No. 41930112.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

All abbreviations mentioned in this article are listed below:
SARSynthetic Aperture Radar
MSCMulti-Scale Convolution
FTMFeature Transformer Module
HRSIDHigh-Resolution SAR Images Dataset
CFARConstant False Alarm Rate
CNNConvolutional Neural Network
R-CNNRegion-CNN
YOLOYou Only Look Once
YOLOFYou Only Look One-level Feature
NLPNatural Language Processing
RNNRecurrent Neural Network
CVComputer Vision
ViTVision Transformer
DETRDetection Transformer
SETRSegmentation Transformer
SSDSingle Shot Detector
CBAMConvolutional Block Attention Module
RFBReceptive Fields Block
FPNFeature Pyramid Network
PAFPNPath Aggregation Feature Pyramid Network
MLPMultilayer Perceptron
FPSFrames Per Second
IoUIntersection over Union
TPTrue Positive
TNTrue Negative
FNFalse Negative
FPFalse Positive
MS COCOMicrosoft Common Objects in Context
P-R CurvePrecision-Recall Curve

References

  1. Moreira, A.; Prats-Iraola, P.; Younis, M.; Krieger, G.; Hajnsek, I.; Papathanassiou, K.P. A tutorial on synthetic aperture radar. IEEE Trans. Geosci. Remote Sens. 2013, 1, 6–43. [Google Scholar] [CrossRef] [Green Version]
  2. Guo, Q.; Wang, H.P.; Xu, F. Research progress on aircraft detection and recognition in SAR imagery. J. Radars 2020, 9, 497–513. [Google Scholar]
  3. Ao, W.; Xu, F.; Li, Y.; Wang, H. Detection and discrimination of ship targets in complex background from spaceborne ALOS-2 SAR images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 536–550. [Google Scholar] [CrossRef]
  4. Copeland, A.C.; Ravichandran, G.; Trivedi, M.M. Localized Radon transform-based detection of ship wakes in SAR images. IEEE Trans. Geosci. Remote Sens. 1995, 33, 35–45. [Google Scholar] [CrossRef]
  5. Novak, L.M.; Owirka, G.J.; Brower, W.S.; Weaver, A.L. The automatic target-recognition system in SAIP. Linc. Lab. J. 1997, 10. Available online: http://www.geo.uzh.ch/microsite/rsl-documents/research/SARlab/GMTILiterature/PDF/NOBW97.pdf (accessed on 17 October 2022).
  6. Xu, P.; Li, Q.; Zhang, B.; Wu, F.; Zhao, K.; Du, X.; Yang, C.; Zhong, R. On-Board Real-Time Ship Detection in HISEA-1 SAR Images Based on CFAR and Lightweight Deep Learning. Remote Sens. 2021, 13, 1995. [Google Scholar] [CrossRef]
  7. Novak, L.M.; Owirka, G.J.; Netishen, C.M. Performance of a high-resolution polarimetric SAR automatic target recognition system. Linc. Lab. J. 1993, 6, 11–24. [Google Scholar]
  8. Qin, X.; Zhou, S.; Zou, H.; Gao, G. A CFAR detection algorithm for generalized gamma distributed background in high-resolution SAR images. IEEE Geosci. Remote Sens. Lett. 2012, 10, 806–810. [Google Scholar]
  9. Nunziata, F.; Migliaccio, M. On the COSMO-SkyMed PingPong mode to observe metallic targets at sea. IEEE J. Ocean. Eng. 2012, 38, 71–79. [Google Scholar] [CrossRef]
  10. Ferrara, G.; Migliaccio, M.; Nunziata, F.; Sorrentino, A. Generalized-K (GK)-based observation of metallic objects at sea in full-resolution synthetic aperture radar (SAR) data: A multipolarization study. IEEE J. Ocean. Eng. 2011, 36, 195–204. [Google Scholar] [CrossRef]
  11. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  12. Wu, X.; Sahoo, D.; Hoi, S.C. Recent advances in deep learning for object detection. Neurocomputing 2020, 396, 39–64. [Google Scholar] [CrossRef] [Green Version]
  13. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Columbus, OH, USA, 24–27 June 2014; pp. 580–587. [Google Scholar]
  14. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. Adv. Neural Inf. Process. Syst. 2015, 39, 1137–1149. [Google Scholar] [CrossRef] [Green Version]
  15. Nie, X.; Duan, M.; Ding, H.; Hu, B.; Wong, E.K. Attention mask R-CNN for ship detection and segmentation from remote sensing images. IEEE Access 2020, 8, 9325–9334. [Google Scholar] [CrossRef]
  16. Zhang, H.; Chang, H.; Ma, B.; Wang, N.; Chen, X. Dynamic R-CNN: Towards high quality object detection via dynamic training. In Proceedings of the European Conference on Computer Vision (ECCV), Glasgow, UK, 23–28 August 2020; pp. 260–275. [Google Scholar]
  17. Sun, P.; Zhang, R.; Jiang, Y.; Kong, T.; Xu, C.; Zhan, W.; Tomizuka, M.; Li, L.; Yuan, Z.; Wang, C.; et al. Sparse r-cnn: End-to-end object detection with learnable proposals. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 14454–14463. [Google Scholar]
  18. Pang, J.; Chen, K.; Shi, J.; Feng, H.; Ouyang, W.; Lin, D. Libra r-cnn: Towards balanced learning for object detection. In Proceedings of the 2019 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 16–20 June 2019; pp. 821–830. [Google Scholar]
  19. Redmon, J.; Farhadi, A. Yolov3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  20. Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y.M. Yolov4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  21. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
  22. Chen, Q.; Wang, Y.; Yang, T.; Zhang, X.; Cheng, J.; Sun, J. You only look one-level feature. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 19–25 June 2021; pp. 13039–13048. [Google Scholar]
  23. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16 × 16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  24. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30. Available online: https://proceedings.neurips.cc/paper/2017/file/3f5ee243547dee91fbd053c1c4a845aa-Paper.pdf (accessed on 17 October 2022).
  25. Yin, W.; Kann, K.; Yu, M.; Schütze, H. Comparative study of CNN and RNN for natural language processing. arXiv 2017, arXiv:1702.01923. [Google Scholar]
  26. Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; Zagoruyko, S. End-to-end object detection with transformers. In Proceedings of the European Conference on Computer Vision (ECCV), Glasgow, UK, 23–28 August 2020; pp. 213–229. [Google Scholar]
  27. Li, J.; Qu, C.; Shao, J. Ship detection in SAR images based on an improved faster R-CNN. In Proceedings of the 2017 SAR in Big Data Era: Models, Methods and Applications (BIGSARDATA), Beijing, China, 13–14 November 2017; pp. 1–6. [Google Scholar]
  28. Xu, Z.; Sun, K.; Mao, J. Research on ResNet101 network chemical reagent label image classification based on transfer learning. In Proceedings of the 2020 IEEE 2nd International Conference on Civil Aviation Safety and Information Technology (ICCASIT), Weihai, China, 14–16 October 2020; pp. 354–358. [Google Scholar]
  29. Srinivas, A.; Lin, T.Y.; Parmar, N.; Shlens, J.; Abbeel, P.; Vaswani, A. Bottleneck transformers for visual recognition. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 19–25 June 2021; pp. 16519–16529. [Google Scholar]
  30. Ge, Z.; Liu, S.; Wang, F.; Li, Z.; Sun, J. Yolox: Exceeding yolo series in 2021. arXiv 2021, arXiv:2107.08430. [Google Scholar]
  31. Wang, J.; Zheng, T.; Lei, P.; Bai, X. Ground target classification in noisy SAR images using convolutional neural networks. IEEE J. Sel. Top Appl. Earth Obs. Remote Sens. 2018, 11, 4180–4192. [Google Scholar] [CrossRef]
  32. Wei, S.; Zeng, X.; Qu, Q.; Wang, M.; Su, H.; Shi, J. HRSID: A high-resolution SAR images dataset for ship detection and instance segmentation. IEEE Access 2020, 8, 120234–120254. [Google Scholar] [CrossRef]
  33. Zhang, T.; Zhang, X.; Ke, X.; Zhan, X.; Shi, J.; Wei, S.; Pan, D.; Li, J.; Su, H.; Zhou, Y.; et al. LS-SSDD-v1.0: A deep learning dataset dedicated to small ship detection from large-scale Sentinel-1 SAR images. Remote Sens. 2020, 12, 2997. [Google Scholar] [CrossRef]
  34. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. Ssd: Single shot multibox detector. In Proceedings of the European Conference on Computer Vision (ECCV), Amsterdam, The Netherlands, 8–16 October 2016; pp. 21–37. [Google Scholar]
  35. Juan, S.; Long, Y.; Hua, H. Improved SSD algorithm for small-sized SAR ship detection. J. Syst. Eng. Electron. 2020, 42, 1026–1034. [Google Scholar]
  36. Chen, P.; Li, Y.; Zhou, H.; Liu, B.; Liu, P. Detection of small ship objects using anchor boxes cluster and feature pyramid network model for SAR imagery. J. Mar. Sci. Eng. 2020, 8, 112. [Google Scholar] [CrossRef] [Green Version]
  37. Yang, X.; Zhang, X.; Wang, N.; Gao, X. A Robust One-Stage Detector for Multiscale Ship Detection with Complex Background in Massive SAR Images. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5217712. [Google Scholar] [CrossRef]
  38. Jin, K.; Chen, Y.; Xu, B.; Yin, J.; Wang, X.; Yang, J. A patch-topixel convolutional neural network for small ship detection with PolSAR images. IEEE Trans. Geosci. Remote Sens. 2020, 58, 6623–6638. [Google Scholar] [CrossRef]
  39. Fu, J.; Sun, X.; Wang, Z.; Fu, K. An anchor-free method based on feature balancing and refinement network for multiscale ship detection in SAR images. IEEE Trans. Geosci. Remote Sens. 2020, 59, 1331–1344. [Google Scholar] [CrossRef]
  40. Guo, H.; Yang, X.; Wang, N.; Gao, X. A CenterNet++ model for ship detection in SAR images. Pattern Recognit. 2021, 112, 107787. [Google Scholar] [CrossRef]
  41. Chang, Y.L.; Anagaw, A.; Chang, L.; Wang, Y.C.; Hsiao, C.Y.; Lee, W.H. Ship detection based on YOLOv2 for SAR imagery. Remote Sens. 2019, 11, 786. [Google Scholar] [CrossRef] [Green Version]
  42. Su, N.; He, J.; Yan, Y.; Zhao, C.; Xing, X. SII-Net: Spatial Information Integration Network for Small Target Detection in SAR Images. Remote Sens. 2022, 14, 442. [Google Scholar] [CrossRef]
  43. Zhao, J.; Guo, W.; Zhang, Z.; Yu, W. A coupled convolutional neural network for small and densely clustered ship detection in SAR images. Sci. China Technol. Sci. 2019, 62, 42301. [Google Scholar] [CrossRef] [Green Version]
  44. Cui, Z.; Li, Q.; Cao, Z.; Liu, N. Dense attention pyramid networks for multi-scale ship detection in SAR images. IEEE Trans. Geosci. Remote Sens. 2019, 57, 8983–8997. [Google Scholar] [CrossRef]
  45. Zhang, Z.; Zhang, X.; Peng, C.; Xue, X.; Sun, J. Exfuse: Enhancing feature fusion for semantic segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 269–284. [Google Scholar]
  46. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–15 June 2015; pp. 1–9. [Google Scholar]
  47. Wang, C.Y.; Liao, H.Y.M.; Wu, Y.H.; Chen, P.Y.; Hsieh, J.W.; Yeh, I.H. Cspnet: A new backbone that can enhance learning capability of cnn. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 390–391. [Google Scholar]
  48. Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D. Microsoft coco: Common objects in context. In Proceedings of the European Conference on Computer Vision (ECCV), Zurich, Switzerland, 6–12 September 2014; pp. 740–755. [Google Scholar]
  49. Chen, K.; Wang, J.; Pang, J.; Cao, Y.; Xiong, Y.; Li, X.; Lin, D. MMDetection: Open mmlab detection toolbox and benchmark. arXiv 2019, arXiv:1906.07155. [Google Scholar]
Figure 1. Two SAR image examples in different scenarios, with green boxes referring to ship objects labeled in the HRSID. (a) A simple off-shore image without interference; (b) a complicated image with inland interference and dense ships.
Figure 1. Two SAR image examples in different scenarios, with green boxes referring to ship objects labeled in the HRSID. (a) A simple off-shore image without interference; (b) a complicated image with inland interference and dense ships.
Remotesensing 14 05268 g001
Figure 2. Comparison of partial model structure: (a) YOLOX. (b) YOLO-SD.
Figure 2. Comparison of partial model structure: (a) YOLOX. (b) YOLO-SD.
Remotesensing 14 05268 g002
Figure 3. The structure of MSC, which adopts convolution operation with different kernel sizes and fuses after residual connection to extract and refine features in receptive fields of various sizes.
Figure 3. The structure of MSC, which adopts convolution operation with different kernel sizes and fuses after residual connection to extract and refine features in receptive fields of various sizes.
Remotesensing 14 05268 g003
Figure 4. The structure of FTM mainly consists of a Multi-Head Attention block and a Multi-Layer Perception block in a Transformer Encoder.
Figure 4. The structure of FTM mainly consists of a Multi-Head Attention block and a Multi-Layer Perception block in a Transformer Encoder.
Remotesensing 14 05268 g004
Figure 5. The overall structure of YOLO-SD, where MSC and FTM are added to the backbone, and the connections between the backbone and neck are modified.
Figure 5. The overall structure of YOLO-SD, where MSC and FTM are added to the backbone, and the connections between the backbone and neck are modified.
Remotesensing 14 05268 g005
Figure 6. Comparison of the number of ships of three sizes in the two datasets.
Figure 6. Comparison of the number of ships of three sizes in the two datasets.
Remotesensing 14 05268 g006
Figure 7. Comparison of the P-R curves of YOLOX with different complexity.
Figure 7. Comparison of the P-R curves of YOLOX with different complexity.
Remotesensing 14 05268 g007
Figure 8. Comparison of Precision-Recall curves of different models. (a) Two-stage models. (b) One-stage models.
Figure 8. Comparison of Precision-Recall curves of different models. (a) Two-stage models. (b) One-stage models.
Remotesensing 14 05268 g008
Figure 9. Results in the first river scenario, where the green boxes represent the ship positions marked in the dataset and the yellow boxes represent the detection results inferred by the detection model. (a) inference by YOLOX_L, (b) the real target frame marked in the dataset, and (c) inference by YOLO-SD.
Figure 9. Results in the first river scenario, where the green boxes represent the ship positions marked in the dataset and the yellow boxes represent the detection results inferred by the detection model. (a) inference by YOLOX_L, (b) the real target frame marked in the dataset, and (c) inference by YOLO-SD.
Remotesensing 14 05268 g009
Figure 10. Results in the second river scenario, where the green boxes represent the ship positions marked in the dataset and the yellow boxes represent the detection results inferred by the detection model. (a) inference by YOLOX_L, (b) the real target frame marked in the dataset, and (c) inference by YOLO-SD.
Figure 10. Results in the second river scenario, where the green boxes represent the ship positions marked in the dataset and the yellow boxes represent the detection results inferred by the detection model. (a) inference by YOLOX_L, (b) the real target frame marked in the dataset, and (c) inference by YOLO-SD.
Remotesensing 14 05268 g010
Figure 11. Partial detection results in LS-SSDD-v1.0, where the green boxes represent the ship positions marked in the dataset, the red boxes show the results detected by YOLO-SD and the yellow boxes represent the results inferred by other detection models. (a) ground truth; (b) YOLO-SD; (c) Dynamic R-CNN; (d) Faster R-CNN; (e) Grid R-CNN; (f) Mask R-CNN; (g) RetinaNet; (h) YOLOF; (i) YOLOv3; (j) Libra Faster R-CNN.
Figure 11. Partial detection results in LS-SSDD-v1.0, where the green boxes represent the ship positions marked in the dataset, the red boxes show the results detected by YOLO-SD and the yellow boxes represent the results inferred by other detection models. (a) ground truth; (b) YOLO-SD; (c) Dynamic R-CNN; (d) Faster R-CNN; (e) Grid R-CNN; (f) Mask R-CNN; (g) RetinaNet; (h) YOLOF; (i) YOLOv3; (j) Libra Faster R-CNN.
Remotesensing 14 05268 g011
Figure 12. Partial detection results in HRSID, where the green boxes represent the ship positions marked in the dataset, the red boxes show the results detected by YOLO-SD and the yellow boxes represent the results inferred by other detection models. (a) ground truth; (b) YOLO-SD; (c) Dynamic R-CNN; (d) Faster R-CNN; (e) Grid R-CNN; (f) Mask R-CNN; (g) RetinaNet; (h) YOLOF; (i) YOLOv3; (j) Libra Faster R-CNN.
Figure 12. Partial detection results in HRSID, where the green boxes represent the ship positions marked in the dataset, the red boxes show the results detected by YOLO-SD and the yellow boxes represent the results inferred by other detection models. (a) ground truth; (b) YOLO-SD; (c) Dynamic R-CNN; (d) Faster R-CNN; (e) Grid R-CNN; (f) Mask R-CNN; (g) RetinaNet; (h) YOLOF; (i) YOLOv3; (j) Libra Faster R-CNN.
Remotesensing 14 05268 g012aRemotesensing 14 05268 g012b
Table 1. Description of the two datasets used.
Table 1. Description of the two datasets used.
ParameterHRSIDLS-SSDD-v1.0
Resolution (m)0.5, 1, 35 × 20
Total Images56049000
Total Ships16,9516015
Image Size800 × 800 pixels800 × 800 pixels
Dataset Division
(Training Set:Validation Set:Test Set)
13:7:72:1:1
Size of ShipsSmall92426003
Medium738812
Large3210
Table 2. Results of ablation experiments.
Table 2. Results of ablation experiments.
BaselineMSCFTMAPAP50AP75APSAPMAPLParams
55.779.863.558.547.31.254.15
56.980.565.159.749.70.0254.35
58.982.867.561.552.8159.4
59.583.767.662.351.91.359.6
Results were evaluated by MS COCO evaluation indexes, "√" represents the model structure used in this test, three "√" represent YOLO-SD we proposed, and bold data is the best result.
Table 3. Comparison of experimental results of YOLOX_L on HRSID.
Table 3. Comparison of experimental results of YOLOX_L on HRSID.
SizeModeAPAP50AP75APSAPMAPLParamsSpeed/s
512YOLOX_L49.373.555.351.050.13.954.150.031
YOLO-SD49.773.956.150.953.84.759.60.034
YOLO-SD *51.977.158.353.455.08.159.60.034
640YOLOX_L53.677.960.656.052.11.954.150.034
YOLO-SD54.778.361.656.754.51.659.60.038
YOLO-SD *54.880.462.056.954.05.359.60.038
800YOLOX_L55.779.863.558.547.31.254.150.035
YOLO-SD56.280.264.159.148.10.659.60.039
YOLO-SD *59.583.767.662.351.91.359.60.039
Results were evaluated by MS COCO evaluation indexes, * indicates it loaded the parameters pretrained on HRSID, and bold data is the best result.
Table 4. Results of various models on HRSID.
Table 4. Results of various models on HRSID.
ModeAPAP50AP75APSAPMAPLParamsSpeed/s
Faster R-CNN57.779.465.957.864.224.741.120.037
Libra Faster R-CNN58.378.866.658.864.323.041.390.040
Mask R-CNN59.381.167.759.963.713.943.750.041
Dynamic R-CNN59.180.368.360.063.822.741.120.034
Grid R-CNN59.078.667.359.465.824.064.240.045
YOLOF45.668.152.245.255.08.942.060.023
YOLOv350.580.254.653.745.80.01761.520.029
RetinaNet58.382.264.559.362.621.436.330.033
YOLOX_L55.779.863.558.547.31.254.150.035
YOLO-SD *59.583.767.662.351.91.359.60.039
Results were evaluated by MS COCO evaluation indexes, * indicates it loaded the parameters pretrained on HRSID, and bold data is the best result.
Table 5. Results of various models on LS-SSDD-v1.0.
Table 5. Results of various models on LS-SSDD-v1.0.
ModeAPAP50AP75APSAPMAPLParamsSpeed/s
Faster R-CNN25.666.610.824.637.4-41.120.037
Libra Faster R-CNN25.165.910.024.038.5-41.390.039
Mask R-CNN27.170.113.426.040-43.750.041
Dynamic R-CNN26.769.612.125.338.8-41.120.034
Grid R-CNN27.270.311.325.640.1-64.240.045
YOLOF16.650.73.915.031.8-42.060.023
YOLOv320.858.07.820.231.5-61.520.029
RetinaNet21.960.77.220.236.7-36.330.033
YOLOX_L28.372.012.227.139.9-54.150.035
YOLO-SD *29.774.415.028.143.2-59.600.038
Results were evaluated by MS COCO evaluation indexes, * indicates it loaded the parameters pretrained on HRSID, and bold data is the best result.
Table 6. Specific detection results of Figure 9 and Figure 10.
Table 6. Specific detection results of Figure 9 and Figure 10.
ImageModelCorrectWrong
Ground Truth31-
Figure 9YOLOX_L72
YOLO-SD145
Ground Truth92-
Figure 10YOLOX_L340
YOLO-SD652
Correct indicates the number of ship targets successfully detected, Wrong means the number of false detections, and bold data is the best result.
Table 7. Specific detection results of Figure 11 and Figure 12.
Table 7. Specific detection results of Figure 11 and Figure 12.
ModelFigure 11Figure 12
CorrectWrongCorrectWrong
Ground Truth15-6-
YOLO-SD15060
Dynamic R-CNN13053
Faster R-CNN8153
Grid R-CNN11363
Mask R-CNN10353
RetinaNet121453
YOLOF91040
YOLOv313351
Libra Faster R-CNN8352
Correct indicates the number of ship targets successfully detected, Wrong means the number of false detections, and bold data is the best result.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Wang, S.; Gao, S.; Zhou, L.; Liu, R.; Zhang, H.; Liu, J.; Jia, Y.; Qian, J. YOLO-SD: Small Ship Detection in SAR Images by Multi-Scale Convolution and Feature Transformer Module. Remote Sens. 2022, 14, 5268. https://doi.org/10.3390/rs14205268

AMA Style

Wang S, Gao S, Zhou L, Liu R, Zhang H, Liu J, Jia Y, Qian J. YOLO-SD: Small Ship Detection in SAR Images by Multi-Scale Convolution and Feature Transformer Module. Remote Sensing. 2022; 14(20):5268. https://doi.org/10.3390/rs14205268

Chicago/Turabian Style

Wang, Simin, Song Gao, Lun Zhou, Ruochen Liu, Hengsheng Zhang, Jiaming Liu, Yong Jia, and Jiang Qian. 2022. "YOLO-SD: Small Ship Detection in SAR Images by Multi-Scale Convolution and Feature Transformer Module" Remote Sensing 14, no. 20: 5268. https://doi.org/10.3390/rs14205268

APA Style

Wang, S., Gao, S., Zhou, L., Liu, R., Zhang, H., Liu, J., Jia, Y., & Qian, J. (2022). YOLO-SD: Small Ship Detection in SAR Images by Multi-Scale Convolution and Feature Transformer Module. Remote Sensing, 14(20), 5268. https://doi.org/10.3390/rs14205268

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop