Next Article in Journal
Spectral-Based Classification of Plant Species Groups and Functional Plant Parts in Managed Permanent Grassland
Previous Article in Journal
Analysis of Ecological Blockage Pattern in Beijing Important Ecological Function Area, China
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Ship Detection Method via Redesigned FCOS in Large-Scale SAR Images

1
Graduate College, Air Force Engineering University, Xi’an 710051, China
2
Air and Missile Defense College, Air Force Engineering University, Xi’an 710051, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2022, 14(5), 1153; https://doi.org/10.3390/rs14051153
Submission received: 10 January 2022 / Revised: 19 February 2022 / Accepted: 22 February 2022 / Published: 25 February 2022

Abstract

:
Ship detection in large-scale synthetic aperture radar (SAR) images has achieved breakthroughs as a result of the improvement of SAR imaging technology. However, there still exist some issues due to the scattering interference, sparsity of ships, and dim and small ships. To address these issues, an anchor-free method is proposed for dim and small ship detection in large-scale SAR images. First, fully convolutional one-stage object detection (FCOS) as the baseline is applied to detecting ships pixel by pixel, which can eliminate the effect of anchors and avoid the missing detection of small ships. Then, considering the particularity of SAR ships, the sample definition is redesigned based on the statistical characteristics of ships. Next, the feature extraction is redesigned to improve the feature representation for dim and small ships. Finally, the classification and regression are redesigned by introducing an improved focal loss and regression refinement with complete intersection over union (CIoU) loss. Experimental simulation results show that the proposed R-FCOS method can detect dim and small ships in large-scale SAR images with higher accuracy compared with other methods.

1. Introduction

Due to the continuous improvements in the quantity and quality of synthetic aperture radar (SAR) images [1,2,3], ship detection has been widely applied for maritime management and surveillance. Ship detection has become a subject task with important theoretical research and practical applications, so that it has attracted more and more scholars’ attention [4,5,6,7,8]. For example, Li et al. [9] introduced the superpixels into constant false alarm rate (CFAR) ship detection method. Salembier et al. [10] applied graph signal processing based on Maxtree representation for ship detection. Lin et al. [11] proposed a ship detection method via superpixels and fisher vectors. Wang et al. [12] proposed the local contrast of fisher vectors (LCFVs) for detecting ships. However, these traditional methods still have limitations. On the one hand, it is difficult to design appropriate manual features for ships, resulting in weak generalization ability. On the other hand, the time cost is high due to the complex detection process.
Benefiting from the excellent performance of convolutional neural networks (CNN) in object classification, research on CNN-based object detection is booming. Object detection methods proposed in the past ten years can be roughly divided into two categories. The first are the two-stage detectors, such as faster region-based CNN (Faster R-CNN) [13]. The second are the one-stage detectors, such as you only look once (YOLO) [14], single shot multi-box detector (SSD) [15], and RetinaNet [16]. The two-stage detectors can achieve relatively good detection performance but have high time costs. However, the one-stage detectors have lower computational costs but have a weaker detection performance.
As a result of the emergence of large-scale SAR datasets [17,18,19,20,21], CNN-based detectors have been introduced for ship detection [22,23,24,25,26]. For example, Deng et al. [27] proposed a method to detect small and densely clustered ships, and it is trained from scratch. Chen et al. [28] proposed a ship detection method via an adaptive recalibration mechanism. Jin et al. [29] proposed P2P-CNN via using the ship and its surroundings to deal with the small ship problem. Gao et al. [30] proposed SAR-Net for achieving a balance between speed and accuracy. These methods are anchor-based detectors, whose detection accuracies are related to the quality of the pre-defined anchors. Although the anchor-based methods work well in detecting ships, they still have certain shortcomings. Firstly, the anchor box will introduce additional hyperparameters, which are set based on prior knowledge. When the detection object changes, the hyperparameters need to be reset. Setting inappropriate anchor boxes will result in poor detection performance, thereby reducing the generalization ability of the network. Secondly, considering the sparsity of ships, most anchor boxes tend to contain empty backgrounds, and only a few anchor boxes contain ships. This will result in far more negative samples than positive samples, resulting in an unbalanced design. Finally, a large number of anchor frames are redundant because of the sparsity of ships, thereby bringing additional computational costs.
Anchor-free methods directly predict class and location instead of using the anchor box. Anchor-free methods consist of key-point-based methods [31,32,33] and center-based methods [34,35,36,37,38]. The key-point-based methods, such as CenterNet [33] and representative points (RepPoints) [31], first detect the key points and then combine the key points for object detection. Center-based methods, such as fully convolutional one-stage object detection (FCOS) [35], FoveaBox [36], and feature selective anchor-free (FSAF) [37], directly detect objects by the center point and bounding box. Although the anchor-free methods [39,40,41] for ship detection are still under development, they have shown a better performance potential and a trade-off between speed and accuracy. However, the performance of the anchor-free method still needs to be improved due to the following issues. First, there is a lot of scattering interference from islands or the sea in SAR images. Secondly, the sparsity of ships cannot be ignored. Finally, some ships are small and dim, that is, the proportion of ships is small, and the object scattering is weak.
Therefore, a novel method called redesigned FCOS (R-FCOS) is proposed for dim and small ship detection. Specifically, we redesign the anchor-free detector FCOS to address the issues described above, and the architecture of R-FCOS is shown in Figure 1. The contributions are summarized as follows:
  • It is shown that R-FCOS can eliminate the effect of anchors and avoid missing detection of small ships.
  • Considering the particularity of SAR ships, the sample definition was redesigned based on the statistical characteristics of these ships.
  • The feature extraction was redesigned to improve the feature representation for dim and small ships.
  • The classification and regression stages were redesigned by introducing an improved focal loss and bounding box refinement with complete intersection over union (CIoU) loss.
The rest of the paper is organized as follows. Section 2 introduces the proposed method based on FCOS. In Section 3, the details of experiments and the analysis of the results are exhibited. Section 4 describes the discussion. Finally, the conclusion is presented in Section 5.

2. Materials and Methods

In this section, the FOCS network, the baseline of the proposed method, is introduced. Then, the R-FCOS method, including redesigned sample definition, feature extraction, classification, and regression are described in detail. Finally, the loss function is given.

2.1. Baseline

The core idea of FCOS is to detect objects pixel by pixel, that is, to directly predict the distance ( l , t , r , b ) from the center point to the four sides of the bounding box, as shown in Figure 2. Figure 3 shows the architecture of the FCOS network, which includes a backbone network, a feature pyramid network (FPN), a classification branch, a regression branch, and a center-ness branch. The backbone network is used to extract feature maps of the input images. In FPN, the different feature pyramid levels are used to detect multi-scale objects. The classification branch and regression branch are to achieve object classification and location, respectively. The center-ness branch is to evaluate the “center-ness” of a location. The “center-ness” represents the degree of coincidence of the position with the center of the ground truth bounding box, which is defined as:
center-ness = min ( l , r ) max ( l , r ) × min ( t , b ) max ( t , b )
The total training loss as follows:
L ( { p x , y } , { t x , y } ) = 1 N pos x , y L cls ( p x , y , c x , y ) + 1 N pos x , y [ c x , y 1 ] L reg ( t x , y , t x , y ) + 1 N pos x , y L CE ( center-ness x , y , center-ness x , y )
where p is the classification score, t = ( l , t , r , b ) is the regression prediction, and N pos is the number of positive samples. L cls , L reg , and L CE denote focal loss [16], generalized intersection over union (GIoU) loss [42], and binary cross entropy loss, respectively, and [ c x , y 1 ] denotes the indicator function as follows:
[ c x , y 1 ] = { 1 c x , y 1 0 o t h e r w i s e

2.2. Sample Definition Redesign

SAR images and natural images have obvious differences in imaging conditions and object categories. Therefore, the previous sample definition methods based on the intersection over union (IoU) threshold, or the scale range will not be suitable for SAR ship detection. Therefore, we introduced a new sample definition method, which defines the sample threshold according to the statistical characteristics of SAR ships. The specific process is as follows:
Step 1: For a ground-truth box g G on each pyramid level, the L2 distances between the predicted boxes D i D on the i th pyramid level and the ground-truth g are computed.
Step 2: The L2 distances are sorted from small to large, and the first k corresponding predicted boxes A i are selected.
Step 3: For the different pyramid level i , the total candidate positive samples are computed as follows:
A g = i A i
Step 4: The IoU between A g and g is computed as follows:
O g = IoU ( A g , g )
Step 5: The mean and standard deviation are computed as follows:
μ g = Mean ( O g )
v g = Std ( O g )
Step 6: The IoU threshold for g is as follows:
T g = μ g + v g
Step 7: For each candidate a A g , if IoU ( a , g ) T g and the center of a in g , then
Positive g = Positive g a
Step 8: Finally, all positive samples are as follows:
Positive = g G Positive g
Step 9: The rest are negative samples as follows:
Negative = D Positive

2.3. Feature Extraction Redesign

As the feature level increases, the resolution of the feature map decreases while the semantic information contained increases. For dim and small ship detection, high-resolution feature representation is essential. This is because the low-resolution feature maps contain very little object information due to the small ships with less pixels and weak scattering. Previous methods usually restore the obtained low-resolution features to high-resolution, such as Hourglass [43], SegNet [44], and the deconvolution network (DeconvNet) [45]. Different from the previous methods, we no longer add additional operations of feature resolution recovery. In other words, the high-resolution feature representation persists throughout the entire computing stage. In addition, the multi-resolution features are repeatedly fused to obtain rich semantic information and high-resolution feature representation. In summary, we redesign feature extraction via the same-resolution feature convolution (SFC) module, multi-resolution feature fusion (MFF) module, and feature pyramid (FP) module, as shown in Figure 1.
At first, the input image is down-sampled via two 3 × 3 stride-2 convolutions.
F 0 = Conv 3 × 3 , s = 2 ( Conv 3 × 3 , s = 2 ( Input ) )
where F 0 denotes the feature map, and Conv 3 × 3 , s = 2 denotes the operation of 3 × 3 stride-2 convolution.
Two types of SFC modules are used to extract feature maps, of which two residual blocks, i.e., Basicblock and Bottleneck, are the main components, as shown in Figure 4. The computation processes of Basicblock and Bottleneck are summarized as follows:
f ( F ) = Conv 1 × 1 ( F ) + Conv 1 × 1 ( Conv 3 × 3 ( Conv 1 × 1 ( F ) ) )
h ( F ) = F + Conv 3 × 3 ( Conv 3 × 3 ( F ) )
where Conv 1 × 1 and Conv 3 × 3 denote the operations of 1 × 1 stride-1 convolution and 3 × 3 stride-1 convolution, respectively; f ( · ) and h ( · ) denote the operations of Bottleneck and Basicblock, respectively.
The MFF module achieves information fusion between features through convolution and up-sample. Figure 5 shows the fusion process of three-resolution features. Given three-resolution features { F 31 ,   F 32 ,   F 33 } , the output features { F 41 , F 42 , F 43 , F 44 } are as follows:
F 41 = F 31 + Conv 1 × 1 ( Upsample s = 2 ( F 32 ) ) + Conv 1 × 1 ( Upsample s = 4 ( F 33 ) )
F 42 = Conv 1 × 1 ( Conv 3 × 3 , s = 2 ( F 31 ) ) + F 32 + Conv 1 × 1 ( Upsample s = 2 ( F 33 ) )
F 43 = Conv 3 × 3 , s = 2 ( Conv 3 × 3 , s = 2 ( F 31 ) ) + Conv 3 × 3 , s = 2 ( F 32 ) + F 33
F 44 = Conv 3 × 3 , s = 2 ( Conv 3 × 3 , s = 2 ( Conv 3 × 3 , s = 2 ( F 31 ) ) ) + Conv 3 × 3 , s = 2 ( Conv 3 × 3 , s = 2 ( F 32 ) ) + Conv 3 × 3 , s = 2 ( F 33 )
where Upsample s = 2 and Upsample s = 4 denote the operations of bilinear up-sample with stride-2 and stride-4, respectively.
The final output features are up-sampled and concatenated for the subsequent FP module.
F out = Concat ( F 41 + Upsample s = 2 ( F 42 ) + Upsample s = 4 ( F 43 ) + Upsample s = 8 ( F 44 ) )
where Upsample s = 8 denotes the operation of bilinear up-sample with stride-8, and Concat denotes the concatenation operation.
The dimension of Pin is set to 256 by a 1 × 1 convolution.
P in = Conv 1 × 1 ( F out )
The FP module is constructed through a set of average pooling operations.
P φ out = AvgPool s = 2 φ ( P in ) , φ = 1 , 2 , 3 , 4
where AvgPool s = 2 φ denotes the operation of average pooling with stride - 2 φ ; φ is the pyramid level.
Finally, a 3 × 3 convolution is appended on each level to obtain the final feature map called { P 0 , P 1 , P 2 , P 3 , P 4 } .
P φ = Conv 3 × 3 ( P φ out ) , φ = 0 , 1 , 2 , 3 , 4

2.4. Classification and Regression Redesign

To utilize the geometric features and contextual information of dim and small ships, we implemented a new feature representation for the bounding box via the features of nine fixed points, as shown in Figure 6. The deformable convolution (Dconv) is introduced to achieve the above feature representation. Specifically, these offsets between the other eight points and ( x , y ) are used as the offsets of the Dconv.
The purpose of object detection is to identify the categories of all objects and give their localizations. Object detectors generally use bounding boxes with category labels and classification scores to represent the detection results. Some bounding boxes with accurate localization may be eliminated due to low classification scores. This indicates that the classification score is not suitable for estimating detection accuracy. Therefore, we use an IoU score (IS) to simultaneously express classification and localization accuracy. Due to the sparsity of ships, most of the regions in SAR images are negative samples, and the sample imbalance is unavoidable. To solve this issue, an improved focal loss function is proposed to predict IS. Specifically, an asymmetric training sample weighting method is applied to focal loss. Focal loss is defined as:
L FL ( p , z ) = { α ( 1 p ) γ log ( p ) i f z = 1 ( 1 α ) p γ log ( 1 p ) o t h e r w i s e
where z { 1 , + 1 } is class label, p { 0 , 1 } is predicted probability, and α , γ are the weighting factor and tunable focusing parameter, respectively.
Next, we used an asymmetric sample weighting to improve focal loss function.
L IFL ( I S , q ) = { q ( q l o g ( I S ) + ( 1 q ) log ( 1 I S ) ) q > 0 α I S γ log ( 1 I S ) q = 0
where I S and q denote the predicted and target I S , respectively.
There is a lot of scattering interference from islands and the sea in SAR images. Therefore, we introduced a regression refinement branch to improve localization accuracy by a refinement factor ( Δ l , Δ t , Δ r , Δ b ) . The final output of the regression branch is ( Δ l l , Δ t t , Δ r r , Δ b b ) . To improve the speed and accuracy of loss convergence, a CIoU loss is introduced into the regression branch:
L CIoU ( d , g ) = 1 IoU + ρ 2 ( d , g ) Δ 2 + β
where d and g denote the center point of predicted and ground truth box, respectively, ρ is the Euclidean distance, Δ denotes the diagonal distance of the smallest rectangle containing the predicted and ground-truth box, and β is an impact factor.

2.5. Loss Function

The entire network is trained with a multi-task loss function as follows:
L = 1 N pos j L IFL ( I S j , q j ) + 1.5 N pos j q j L CIoU ( d j , g j ) + 2 N pos j q j L CIoU ( d j , g j )
where d j denotes the initial box at location j on feature map, and d and g are the refined and ground-truth box, respectively.

3. Results

All experiments in this paper are performed on a computer equipped with RTX 2080Ti GPU and Inter i9-9820X CPU. The operating system used is Ubuntu 16.04, and the basic framework is pytorch. The average precision (AP) and frames per second (FPS) are used as the evaluation metrics.

3.1. Dataset

The Large-Scale SAR Ship Detection Dataset-v1.0 (LS-SSDD-v1.0) [20] was constructed from Sentinel-1 satellites and included 15 images with 24,000 × 16,000 pixels, a sample image is shown in Figure 7. To facilitate network training and testing, the original image is split into 9000 sub-images with 800 × 800 pixels for training (6000) and testing (3000). According to the scale division for bounding box in Microsoft Common Objects in COntext (MS COCO), the area below 1024 pixels is a small ship, the area from 1024 pixels to 9216 pixels is a medium ship, and the area above 9216 pixels is a large ship. The distributions of ship shape and number are shown in Figure 8 and Figure 9, respectively.

3.2. Ablation Study

Since the proposed method is similar in structure to FCOS, FCOS is used as the baseline. Figure 10 shows the detection results of FCOS, FCOS with sample definition redesign (FCOS+SDR), FCOS with sample definition redesign and feature extraction redesign (FCOS+SDR+FER), and FCOS with sample definition redesign, feature extraction redesign, and classification and regression redesign (FCOS+SDR+FER+CRR).

3.2.1. Analysis of Sample Definition Redesign

As shown in Figure 10, the AP of FCOS+SDR is 1.3% higher than that of FCOS. This suggests that sample definition redesign is effective for improving detection accuracy by redefining the sample threshold according to the statistical characteristics of SAR ships.

3.2.2. Analysis of Feature Extraction Redesign

As shown in Figure 10, the APs of FCOS+SDR+FER are 2.3% and 1.0% higher than those of FCOS and FCOS+SDR, respectively, suggesting that feature extraction redesign is effective for improving detection accuracy. The main reason is that feature extraction redesign can obtain rich semantic information and high-resolution feature representation for dim and small ships.

3.2.3. Analysis of Classification and Regression Redesign

As shown in Figure 10, the APs of R-FCOS are 3.2% and 0.9% higher than those of FCOS and FCOS+SDR+FER, respectively. The main reason is that the classification and regression redesign is used to deal with these issues, i.e., complex surroundings, sparsity of ships, and dim and small ships. Specifically, the nine-point feature representation is used to extract geometric features and contextual information for dim and small ships. The improved focal loss function is used to predict IS. Finally, the regression refinement branch with CIoU loss is used to improve localization accuracy.

3.3. Comparison with Other Methods

Based on the same experimental environment, the ship detection performance of R-FCOS was compared with those of other methods such as Faster RCNN [13], SSD [15], RetinaNet [16], YOLOv3 [46], RepPoints [31], FSAF [37], and FoveaBox [36]. As illustrated in Figure 11, we can draw the following conclusions:
  • The AP of R-FCOS is better than those of other methods. Specifically, the AP of R-FCOS is 75.5%, which is 9.2%, 17.7%, 4.7%, 8.9%, 3.2%, 4.9%, and 6.1% higher than Faster RCNN, SSD, RetinaNet, YOLOv3, RepPoints, FSAF, and FoveaBox, respectively.
  • The AP of SSD is the worst and is 17.7% lower than our method. Although SSD uses high-resolution features to detect small objects, it contains less semantic information, resulting in unsatisfactory detection results. In addition, SSD reduces the input image size to 300×300, which destroys the object information in the image.
  • The APs of anchor-free methods such as RepPoints, FSAF, FoveaBox, and R-FCOS are generally better than those of anchor-based methods except for RetinaNet. This shows that the anchor-free method is more suitable for ship detection.
  • The FPS of SSD is the highest, and that of Faster RCNN is the lowest. Although the FPS of our method is only 52.0, it already meets the real-time requirements.
To visually demonstrate the detection performance of R-FCOS, Figure 12 shows the comparative results for different methods on LS-SSDD-v1.0. In the first column of Figure 11, there exist some missing ships that occur in all methods. FSAF and R-FCOS have the least number of missed ships, but our method has higher classification scores. In the second column of Figure 11, most of the ships were missed by SSD, YOLOv3, and FoveaBox, and several false alarms were obtained by Faster RCNN. In addition, false alarms and missing ships existed in RetinaNet, RepPoints, and FSAF. However, our method had only one missing ship. In the third column of Figure 11, false alarms or missing ships exist in all methods, except for YOLOv3 and R-FCOS. However, the classification scores of R-FCOS were higher than those of YOLOv3. Overall, R-FCOS is better than the other methods.

4. Discussion

Simulation experiments were carried out on the SAR ship detection dataset (SSDD) [21] to verify the model migration ability. Figure 13 shows the detection results on SSDD. The AP of R-FCOS is 97.8%, which is 3.9%, 5.8%, 1.5%, 2.8%, 1.3%, 1.4%, and 2.2% higher than Faster RCNN, SSD, RetinaNet, YOLOv3, RepPoints, FSAF, and FoveaBox, respectively. Although the FPS of R-FCOS is only 17.5, it is acceptable. Figure 14 shows the visual results of R-FCOS on SSDD.

5. Conclusions

In this paper, an anchor-free detector R-FCOS was proposed for dim and small ship detection. We have redesigned FCOS with the aim to address the issues of complex surroundings, sparsity of ships, and dim and small ships. The sample definition redesign deals with the particularity of SAR ships. The feature extraction redesign improved the feature representation for dim and small ships. The classification and regression redesign introduced an improved focal loss and regression refinement with CIoU loss. In the experimental part, we verified the effectiveness of the sample definition redesign, feature extraction redesign, and classification and regression redesign. Experimental results on LS-SSDD-v1.0 showed that, the proposed method achieves competitive detection performance in comparison with Faster RCNN, SSD, RetinaNet, YOLOv3, RepPoints, FSAF, and FoveaBox. In addition, we verified the model migration ability of the proposed method on SSDD. However, it should be noted that, although the proposed method has better detection performance, it cannot completely eliminate all false alarms and missed detection results, which requires further analysis and research.

Author Contributions

Conceptualization, M.Z. and G.H.; methodology, M.Z.; software, M.Z.; validation, M.Z., H.Z., S.W., Z.F. and S.Y.; formal analysis, M.Z.; investigation, M.Z.; resources, M.Z.; data curation, M.Z.; writing—original draft preparation, M.Z.; writing—review and editing, M.Z.; visualization, M.Z.; supervision, M.Z.; project administration, G.H.; funding acquisition, S.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Natural Science Foundation Research Project of Shaanxi Province, China, under Grant 2020JM-345.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
SARSynthetic Aperture Radar
FCOSFully Convolutional One-Stage Object Detection
CIoUComplete Intersection over Union
CFARConstant False Alarm Rate
LCFVsLocal Contrast of Fisher Vectors
CNNConvolutional Neural Network
FasterR-CNN Faster Region-based CNN
YOLOYou Only Look Once
SSDSingle Shot Multi-Box Detector
P2P-CNNPatch-to-Pixel CNN
RepPointsRepresentative Points
FSAFFeature Selective Anchor-Free
R-FCOSRedesigned FCOS
FPNFeature Pyramid Network
GIoUGeneralized Intersection over Union
IoUIntersection over Union
DeconvNetDeconvolution Network
SFCSame-Resolution Feature Convolution
MFFMulti-Resolution Feature Fusion
FPFeature Pyramid
DconvDeformable Convolution
ISIoU Score
APAverage Precision
FPSFrames Per Second
LS-SSDD-v1.0Large-Scale SAR Ship Detection Dataset-v1.0
MS COCOMicrosoft Common Objects in COntext
FCOS+SDRFCOS with Sample Definition Redesign
FCOS+SDR+FERFCOS with SDR and Feature Extraction Redesign
FCOS+SDR+FER+CRRFCOS with SDR, FER, and Classification and Regression Redesign
SSDDSAR Ship Detection Dataset

References

  1. Li, M.; Wen, G.; Huang, X.; Li, K.; Lin, S. A Lightweight Detection Model for SAR Aircraft in a Complex Environment. Remote Sens. 2021, 13, 5020. [Google Scholar] [CrossRef]
  2. Ghaderpour, E.; Pagiatakis, S.D.; Hassan, Q.K. A Survey on Change Detection and Time Series Analysis with Applications. Appl. Sci. 2021, 11, 6141. [Google Scholar] [CrossRef]
  3. Cui, Z.; Qin, Y.; Zhong, Y.; Cao, Z.; Yang, H. Target Detection in High-Resolution SAR Image via Iterating Outliers and Recursing Saliency Depth. Remote Sens. 2021, 13, 4315. [Google Scholar] [CrossRef]
  4. Pappas, O.; Achim, A.; Bull, D. Superpixel-Level CFAR Detectors for Ship Detection in SAR Imagery. IEEE Geosci. Remote Sens. Lett. 2018, 15, 1397–1401. [Google Scholar] [CrossRef] [Green Version]
  5. He, J.; Wang, Y.; Liu, H.; Wang, N.; Wang, J. A Novel Automatic PolSAR Ship Detection Method Based on Superpixel-Level Local Information Measurement. IEEE Geosci. Remote Sens. Lett. 2018, 15, 384–388. [Google Scholar] [CrossRef]
  6. Yang, F.; Xu, Q.; Li, B. Ship Detection From Optical Satellite Images Based on Saliency Segmentation and Structure-LBP Feature. IEEE Geosci. Remote Sens. Lett. 2017, 14, 602–606. [Google Scholar] [CrossRef]
  7. Wang, S.; Wang, M.; Yang, S.; Jiao, L. New Hierarchical Saliency Filtering for Fast Ship Detection in High-Resolution SAR Images. IEEE Trans. Geosci. Remote Sens. 2017, 55, 351–362. [Google Scholar] [CrossRef]
  8. Song, S.; Xu, B.; Yang, J. SAR Target Recognition via Supervised Discriminative Dictionary Learning and Sparse Representation of the SAR-HOG Feature. Remote Sens. 2016, 8, 683. [Google Scholar] [CrossRef] [Green Version]
  9. Li, T.; Liu, Z.; Xie, R.; Ran, L. An Improved Superpixel-Level CFAR Detection Method for Ship Targets in High-Resolution SAR Images. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2018, 11, 184–194. [Google Scholar] [CrossRef]
  10. Salembier, P.; Liesegang, S.; Lopez-Martinez, C. Ship Detection in SAR Images Based on Maxtree Representation and Graph Signal Processing. IEEE Trans. Geosci. Remote Sens. 2019, 57, 2709–2724. [Google Scholar] [CrossRef] [Green Version]
  11. Lin, H.; Chen, H.; Jin, K.; Zeng, L.; Yang, J. Ship Detection with Superpixel-Level Fisher Vector in High-Resolution SAR Images. IEEE Geosci. Remote Sens. Lett. 2020, 17, 247–251. [Google Scholar] [CrossRef]
  12. Wang, X.; Li, G.; Zhang, X.-P.; He, Y. Ship Detection in SAR Images via Local Contrast of Fisher Vectors. IEEE Trans. Geosci. Remote Sens. 2020, 58, 6467–6479. [Google Scholar] [CrossRef]
  13. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  15. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.; Berg, A.C. SSD: Single Shot MultiBox Detector. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; pp. 21–37. [Google Scholar]
  16. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollar, P. Focal Loss for Dense Object Detection. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 318–327. [Google Scholar] [CrossRef] [Green Version]
  17. Wang, Y.; Wang, C.; Zhang, H.; Dong, Y.; Wei, S. A SAR Dataset of Ship Detection for Deep Learning under Complex Backgrounds. Remote Sens. 2019, 11, 765. [Google Scholar] [CrossRef] [Green Version]
  18. Song, J.; Kim, D.-J.; Kang, K.-M. Automated Procurement of Training Data for Machine Learning Algorithm on Ship Detection Using AIS Information. Remote Sens. 2020, 12, 1443. [Google Scholar] [CrossRef]
  19. Wei, S.; Zeng, X.; Qu, Q.; Wang, M.; Su, H.; Shi, J. HRSID: A High-Resolution SAR Images Dataset for Ship Detection and Instance Segmentation. IEEE Access 2020, 8, 120234–120254. [Google Scholar] [CrossRef]
  20. Zhang, T.; Zhang, X.; Ke, X.; Zhan, X.; Shi, J.; Wei, S.; Pan, D.; Li, J.; Su, H.; Zhou, Y.; et al. LS-SSDD-v1.0: A Deep Learning Dataset Dedicated to Small Ship Detection from Large-Scale Sentinel-1 SAR Images. Remote Sens. 2020, 12, 2997. [Google Scholar] [CrossRef]
  21. Li, J.; Qu, C.; Shao, J. Ship detection in SAR images based on an improved faster R-CNN. In Proceedings of the 2017 SAR in Big Data Era: Models, Methods and Applications (BIGSARDATA), Beijing, China, 13–14 November 2017; pp. 1–6. [Google Scholar]
  22. Cui, Z.; Li, Q.; Cao, Z.; Liu, N. Dense Attention Pyramid Networks for Multi-Scale Ship Detection in SAR Images. IEEE Trans. Geosci. Remote Sens. 2019, 57, 8983–8997. [Google Scholar] [CrossRef]
  23. Zhao, Y.; Zhao, L.; Xiong, B.; Kuang, G. Attention Receptive Pyramid Network for Ship Detection in SAR Images. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2020, 13, 2738–2756. [Google Scholar] [CrossRef]
  24. Li, Y.; Zhang, S.; Wang, W.-Q. A Lightweight Faster R-CNN for Ship Detection in SAR Images. IEEE Geosci. Remote Sens. Lett. 2020, 1–5. [Google Scholar] [CrossRef]
  25. Hong, Z.; Yang, T.; Tong, X.; Zhang, Y.; Jiang, S.; Zhou, R.; Han, Y.; Wang, J.; Yang, S.; Liu, S. Multi-Scale Ship Detection From SAR and Optical Imagery Via A More Accurate YOLOv3. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2021, 14, 6083–6101. [Google Scholar] [CrossRef]
  26. Zhang, K.; Wu, Y.; Wang, J.; Wang, Y.; Wang, Q. Semantic Context-Aware Network for Multiscale Object Detection in Remote Sensing Images. IEEE Geosci. Remote Sens. Lett. 2021, 19, 1–5. [Google Scholar] [CrossRef]
  27. Deng, Z.; Sun, H.; Zhou, S.; Zhao, J. Learning Deep Ship Detector in SAR Images from Scratch. IEEE Trans. Geosci. Remote Sens. 2019, 57, 4021–4039. [Google Scholar] [CrossRef]
  28. Chen, C.; He, C.; Hu, C.; Pei, H.; Jiao, L. MSARN: A Deep Neural Network Based on an Adaptive Recalibration Mechanism for Multiscale and Arbitrary-Oriented SAR Ship Detection. IEEE Access 2019, 7, 159262–159283. [Google Scholar] [CrossRef]
  29. Jin, K.; Chen, Y.; Xu, B.; Yin, J.; Wang, X.; Yang, J. A Patch-to-Pixel Convolutional Neural Network for Small Ship Detection with PolSAR Images. IEEE Trans. Geosci. Remote Sens. 2020, 58, 6623–6638. [Google Scholar] [CrossRef]
  30. Gao, S.; Liu, J.M.; Miao, Y.H.; He, Z.J. A High-Effective Implementation of Ship Detector for SAR Images. IEEE Geosci. Remote Sens. Lett. 2021, 19, 1–5. [Google Scholar] [CrossRef]
  31. Yang, Z.; Liu, S.; Hu, H.; Wang, L.; Lin, S. RepPoints: Point Set Representation for Object Detection. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Korea, 27 October–2 November 2019; pp. 9656–9665. [Google Scholar]
  32. Law, H.; Deng, J. Cornernet: Detecting objects as paired keypoints. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 734–750. [Google Scholar]
  33. Zhou, X.; Wang, D.; Krahenbuhl, P. Objects as Points. arXiv 2019, arXiv:1904.07850. [Google Scholar]
  34. Zhang, S.; Chi, C.; Yao, Y.; Lei, Z.; Li, S.Z. Bridging the Gap between Anchor-Based and Anchor-Free Detection via Adaptive Training Sample Selection. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), ATSS, Seattle, WA, USA, 13–19 June 2020; pp. 9756–9765. [Google Scholar]
  35. Tian, Z.; Shen, C.; Chen, H.; He, T. FCOS: A Simple and Strong Anchor-free Object Detector. IEEE Trans. Pattern Anal. Mach. Intell. 2020. [Google Scholar] [CrossRef]
  36. Kong, T.; Sun, F.; Liu, H.; Jiang, Y.; Li, L.; Shi, J. FoveaBox: Beyound Anchor-Based Object Detection. IEEE Trans. Image Process. 2020, 29, 7389–7398. [Google Scholar] [CrossRef]
  37. Zhu, C.; He, Y.; Savvides, M. Feature Selective Anchor-Free Module for Single-Shot Object Detection. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 16–20 June 2019; pp. 840–849. [Google Scholar]
  38. Zhang, H.; Wang, Y.; Dayoub, F.; Sunderhauf, N. VarifocalNet: An IoU-aware Dense Object Detector. arXiv 2021, arXiv:2008.13367. [Google Scholar]
  39. Cui, Z.; Wang, X.; Liu, N.; Cao, Z.; Yang, J. Ship Detection in Large-Scale SAR Images Via Spatial Shuffle-Group Enhance Attention. IEEE Trans. Geosci. Remote Sens. 2021, 59, 379–391. [Google Scholar] [CrossRef]
  40. Fu, J.; Sun, X.; Wang, Z.; Fu, K. An Anchor-Free Method Based on Feature Balancing and Refinement Network for Multiscale Ship Detection in SAR Images. IEEE Trans. Geosci. Remote Sens. 2021, 59, 1331–1344. [Google Scholar] [CrossRef]
  41. Sun, Z.; Dai, M.; Leng, X.; Lei, Y.; Xiong, B.; Ji, K.; Kuang, G. An Anchor-Free Detection Method for Ship Targets in High-Resolution SAR Images. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2021, 14, 7799–7816. [Google Scholar] [CrossRef]
  42. Rezatofighi, H.; Tsoi, N.; Gwak, J.; Sadeghian, A.; Reid, I.; Savarese, S. Generalized Intersection Over Union: A Metric and a Loss for Bounding Box Regression. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 16–20 June 2019; pp. 658–666. [Google Scholar]
  43. Bulat, A.; Tzimiropoulos, G. Binarized Convolutional Landmark Localizers for Human Pose Estimation and Face Alignment with Limited Resources. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 3726–3734. [Google Scholar]
  44. Badrinarayanan, V.; Kendall, A.; Cipolla, R. SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [CrossRef]
  45. Noh, H.; Hong, S.; Han, B. Learning Deconvolution Network for Semantic Segmentation. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 1520–1528. [Google Scholar]
  46. Redmon, J.; Farhadi, A. YOLOv3: An Incremental Improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
Figure 1. The overall architecture of R-FCOS. W, H denote the width and height of the feature map.
Figure 1. The overall architecture of R-FCOS. W, H denote the width and height of the feature map.
Remotesensing 14 01153 g001
Figure 2. The predicted 4D vector of FCOS.
Figure 2. The predicted 4D vector of FCOS.
Remotesensing 14 01153 g002
Figure 3. The architecture of FCOS. C denotes the number of classes.
Figure 3. The architecture of FCOS. C denotes the number of classes.
Remotesensing 14 01153 g003
Figure 4. The architectures of two SFC modules. X denotes the channel dimension. Y denotes the number of Basicblock.
Figure 4. The architectures of two SFC modules. X denotes the channel dimension. Y denotes the number of Basicblock.
Remotesensing 14 01153 g004
Figure 5. An example of MFF module.
Figure 5. An example of MFF module.
Remotesensing 14 01153 g005
Figure 6. A new feature representation for the bounding box.
Figure 6. A new feature representation for the bounding box.
Remotesensing 14 01153 g006
Figure 7. A sample image of LS-SSDD-v1.0.
Figure 7. A sample image of LS-SSDD-v1.0.
Remotesensing 14 01153 g007
Figure 8. The distribution of the ship’s shape.
Figure 8. The distribution of the ship’s shape.
Remotesensing 14 01153 g008
Figure 9. The distribution of the ship’s number.
Figure 9. The distribution of the ship’s number.
Remotesensing 14 01153 g009
Figure 10. The bar graph of ablation studies.
Figure 10. The bar graph of ablation studies.
Remotesensing 14 01153 g010
Figure 11. The bar graph of different methods on LS-SSDD-v1.0.
Figure 11. The bar graph of different methods on LS-SSDD-v1.0.
Remotesensing 14 01153 g011
Figure 12. Comparison results of different methods on LS-SSDD-v1.0.
Figure 12. Comparison results of different methods on LS-SSDD-v1.0.
Remotesensing 14 01153 g012aRemotesensing 14 01153 g012bRemotesensing 14 01153 g012c
Figure 13. The bar graph of different methods on SSDD.
Figure 13. The bar graph of different methods on SSDD.
Remotesensing 14 01153 g013
Figure 14. Visual results of R-FCOS on SSDD. The green rectangles and red rectangles are the ground truth and detection results, respectively.
Figure 14. Visual results of R-FCOS on SSDD. The green rectangles and red rectangles are the ground truth and detection results, respectively.
Remotesensing 14 01153 g014
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhu, M.; Hu, G.; Zhou, H.; Wang, S.; Feng, Z.; Yue, S. A Ship Detection Method via Redesigned FCOS in Large-Scale SAR Images. Remote Sens. 2022, 14, 1153. https://doi.org/10.3390/rs14051153

AMA Style

Zhu M, Hu G, Zhou H, Wang S, Feng Z, Yue S. A Ship Detection Method via Redesigned FCOS in Large-Scale SAR Images. Remote Sensing. 2022; 14(5):1153. https://doi.org/10.3390/rs14051153

Chicago/Turabian Style

Zhu, Mingming, Guoping Hu, Hao Zhou, Shiqiang Wang, Ziang Feng, and Shijie Yue. 2022. "A Ship Detection Method via Redesigned FCOS in Large-Scale SAR Images" Remote Sensing 14, no. 5: 1153. https://doi.org/10.3390/rs14051153

APA Style

Zhu, M., Hu, G., Zhou, H., Wang, S., Feng, Z., & Yue, S. (2022). A Ship Detection Method via Redesigned FCOS in Large-Scale SAR Images. Remote Sensing, 14(5), 1153. https://doi.org/10.3390/rs14051153

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop