Next Article in Journal
A Hybrid Method for Vibration-Based Bridge Damage Detection
Next Article in Special Issue
Unsupervised SAR Image Change Type Recognition Using Regionally Restricted PCA-Kmean and Lightweight MobileNet
Previous Article in Journal
WiFi Access Points Line-of-Sight Detection for Indoor Positioning Using the Signal Round Trip Time
Previous Article in Special Issue
MSSDet: Multi-Scale Ship-Detection Framework in Optical Remote-Sensing Images and New Benchmark
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Lightweight Model for Ship Detection and Recognition in Complex-Scene SAR Images

College of Electronic Science and Technology, National University of Defense Technology, Changsha 410073, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2022, 14(23), 6053; https://doi.org/10.3390/rs14236053
Submission received: 29 October 2022 / Revised: 23 November 2022 / Accepted: 26 November 2022 / Published: 29 November 2022

Abstract

:
SAR ship detection and recognition are important components of the application of SAR data interpretation, allowing for the continuous, reliable, and efficient monitoring of maritime ship targets, in view of the present situation of SAR interpretation applications. On the one hand, because of the lack of high-quality datasets, most existing research on SAR ships is focused on target detection. Additionally, there have been few studies on integrated ship detection and recognition in complex SAR images. On the other hand, the development of deep learning technology promotes research on the SAR image intelligent interpretation algorithm to some extent. However, most existing algorithms only focus on target recognition performance and ignore the model’s size and computational efficiency. Aiming to solve the above problems, a lightweight model for ship detection and recognition in complex-scene SAR images is proposed in this paper. Firstly, in order to comprehensively improve the detection performance and deployment capability, this paper applies the YOLOv5-n lightweight model as the baseline algorithm. Secondly, we redesign and optimize the pyramid pooling structure to effectively enhance the target feature extraction efficiency and improve the algorithm’s operation speed. Meanwhile, to suppress the influence of complex background interference and ships’ distribution, we integrate different attention mechanism into the target feature extraction layer. In addition, to improve the detection and recognition performance of densely parallel ships, we optimize the structure of the model’s prediction layer by adding an angular classification module. Finally, we conducted extensive experiments on the newly released complex-scene SAR image ship detection and recognition dataset, named the SRSDDv1.0 dataset. The experimental results show that the minimum size of the model proposed in this paper is only 1.92 M parameters and 4.52 MB of model memory, which can achieve an excellent F1-Score performance of 61.26 and an FPS performance of 68.02 on the SRSDDv1.0 dataset.

1. Introduction

Radar is a kind of sensor that uses microwaves for active sensing. Synthetic Aperture Radar (SAR) extends the original concept of radar, using pulse compression technology and the principle of synthetic aperture, which enables the radar system to perform the two-dimensional high-resolution imaging of the target. Compared with optical and infrared sensors, SAR is not affected by weather, light, and other conditions. Therefore, SAR has the characteristics of all-day, all-weather, large-width, and high-resolution imaging, which has become a significant tool for Earth observation [1,2,3].
With the continuous development of SAR image resolution and image quality, the research on the automatic detection and identification of ship targets in SAR images has become a significant research direction in the domain of SAR image interpretation [4]. Among them, the SAR ship identification technology has involved the courtesies of a large number of researchers due to its task characteristics. In contrast, the research of ship recognition technology started relatively late, and there have been little research results. In recent years, deep learning technology has accomplished encouraging results in various fields, including target detection [5,6,7], image classification [8,9], autonomous driving [10], saliency detection [11], semantic understanding [12], and so on [13,14,15,16]. Additionally, this new technique also provides new ideas for the development of SAR target detection and recognition technology. Additionally, it is possible to detect and recognize ship targets in SAR images automatically based on deep learning technology [17,18,19,20]. For example, Zhao et al. [21] proposed an automatic identification method for SAR ships based on feature decomposition across different satellites. This method improves the performance of target location and recognition by optimizing the backbone network to extract features. Yoshida et al. [22] proposed a method to automatically detect ships in motion based on the “You Only Look Once (YOLO) v5 model”, which can effectively capture ship targets in ALOS-2 spotlight images. Zheng et al. [23] proposed an ensemble automated method (MetaBoost) for heterogeneous DCNN models based on two-stage filtering, which effectively achieved robustness and high-accuracy recognition for SAR ships. Extensive experiments on OpenSARShip and FuSARShip datasets show that the MetaBoost can significantly outperform individual classifiers and traditional ship recognition models.
As we all know, target detection and recognition have been developed and improved in natural scenes. However, SAR images have major differences from those in natural scenes, one of the most obvious points being that the change in the ship target’s morphology and background interference, such as sidelobe, tailing, and so on. Although there is significant research on SAR image target detection, target classification, and recognition, they still have the following obvious problems:
  • Due to the limitation of data quality and target labeling requirements, most of existing research methods are based on ship detection datasets such as SSDD [24], AirSARship [25], HRSID [26], and LS-SSDD-v1.0 [27]. Research on the integration of SAR ship detection and identification is lacking.
  • The scenes of the SAR image are complex and changeable, which have obvious influence on target imaging and morphological changes. A large number of ships are missed and false alarms can easily occur nearshore or in areas where targets are densely distributed.
  • Compared with traditional recognition models, most existing deep learning models show strong robustness and adaptability in target recognition performance. However, these methods have obvious shortcomings such as low model training efficiency, high deployment cost of embedded devices, and low real-time performance.
In response to the above problems, we propose a lightweight model for ship detection and recognition in complex-scene SAR images. The work of this paper mainly includes the following aspects:
  • In order to improve the model’s operating efficiency and reduce the cost of algorithm deployment, this paper improves and optimizes the algorithm based on the YOLOv5-n lightweight model. Combined with the fast pyramidal pooling structure, the target feature extraction efficiency of the neural network model is effectively improved.
  • Aiming to improve the detection and recognition performance of ship targets in high-resolution complex-scene SAR images, this paper integrates an attention mechanism into the target feature extraction layer. The proposed attention module can improve the model’s attention to target features in complex scenes and suppress the influence of background noise.
  • To optimize the performance of ship positioning and recognition in complex scenes such as nearshore or the dense distribution of ship targets, this paper introduces an angle classification module in the prediction layer of the network model to achieve the rotation detection and recognition of ship targets.
  • We conducted extensive experiments on the newly released SAR ship detection and recognition dataset named SRSDD [28] to validate the proposed improvements. The experimental results show that the proposed method in this paper not only outperforms several other deep learning methods in terms of detection and recognition performance, but also has significant advantages in terms of algorithm parameters, model size, and operation efficiency.
The rest of this paper is organized as follows. In Section 2, the related research work is briefly introduced. In Section 3, we propose the overall framework of the model in this paper and describe several optimization modules. We analyze the experimental performance of the proposed lightweight model in Section 4. Meanwhile, some visual recognition results are presented in Section 4. Finally, the conclusions are drawn in Section 5.

2. Related Work

This section mainly introduces some of the latest deep learning methods applied in the field of SAR image ship detection and recognition. According to the improvement and innovation proposed by the methods in this study, the related methods are divided into three categories: lightweight models, embedded attention mechanism models, and rotation detection models.

2.1. Lightweight Models

The majority of convolutional neural network-based ship target detection algorithms improve the identification score at the expense of technical complexity, but they are challenging to directly apply and deploy. To address this problem, Zhou et al. [29] proposed a lightweight anchorless ship detection network (LASDNet) for SAR images. The method is improved based on the single-stage anchor-free network FCOS and can achieve the lowest parameter size of 1.15 MB, the lowest computational complexity (1.01 GFLOPs), and average precision (59.25) on the HRSID dataset. A lightweight framework based on a threshold neural network (TNN) was suggested by Cui et al. [30] to quickly locate ship targets in large-scene SAR imagery. The neural network, which has a higher detection accuracy and more resilience, is used in the procedure to obtain the target detection threshold. It is important to note that the techniques’ size is considerably less than that of other deep learning models. An improved lightweight RetinaNet was recommended by Miao et al. [31] for ship detection in SAR images. The ghost module, which replaces the backbone’s shallow convolutional layers and reduces the number of deep convolutional layers, is the network’s main selling point because it allows for a significant reduction in the number of floating-point operations and parameters without sacrificing detection precision and recall. In their proposed learned detector, Light-YOLOv4, Ma et al. [32] eliminated the network’s uninteresting channels and layers, considerably lowering the network’s breadth and depth. OSCAR-RT, the first end-to-end algorithm/hardware codesign framework for SAR ship identification based on CNN on real-time satellites, was suggested by Yang et al. [33]. The model may concurrently produce highly effective field-programmable gate array (FPGA)-based hardware accelerators that can be installed on satellites, as well as precise, hardware CNN models. Yang et al. [34] proposed an efficient and lightweight object detection network combined with soft quantization. The network employs a soft quantization algorithm to reduce the impact of quantization errors on model accuracy. Yu et al. [35] suggested a fast and lightweight detection network named FASC-Net, and the model contains four kinds of network lightweight optimization modules, which greatly reduce the number of parameters of the target feature extraction network. Chang et al. [36] develop a new architecture with less layers called YOLOv2-reduced to reduce the computational time. Yu et al. [37] suggested a light ship identification network based on YOLOX-s. The network outputs a pyramid structure with a large amount of computation and builds a streamlined network on the first-level features, which effectively improves the detection efficiency. Feng et al. [38] proposed a new lightweight position-enhanced anchorless SAR ship detection algorithm LPEDet. The network uses YOLOX as the benchmark framework and redesigned a lightweight multi-scale backbone to balance the detection speed and accuracy. Li et al. [39] proposed an ultralight and high detection accuracy SAR ship detection method based on YOLOX. The network contains an ultralight and high-performance detection backbone based on ghost cross stage partial (GhostCSP) and lightweight spatially dilated convolutional pyramid (LSDP). Xu et al. [40] designed a lightweight cross-level local (L-CSP) module to decrease the computation and prune the network for a more compact detector. Guo et al. [41] proposed a lightweight single-stage SAR ship target detection model, called Yold-based Lightweight multi-scale ship detector (LMSD-YOLO), in which the DSASFF module is designed, to attain the adaptive fusion of multi-scale features with a few parameters. Liu et al. [42] proposed a lightweight network based on the YOLOv4-LITE model, which uses MobileNetv2 to extract the features and designed an improved receptive field block (RFB) structure.

2.2. Embedding Attention Mechanism Models

Ship target recognition in SAR images frequently creates more false alarms and may result in the missing detection of tiny objects because of the complex backdrop and coherent speckle noise interference. Gao et al. [43] proposed an improved attention-based yolo4 (imyolo4) model. The adverse effects of the complex background and noise are suppressed by introducing a threshold attention module (TAM), and an attention module is added to the backbone network to enhance the discriminative ability of the multi-scale target features. Peng et al. [44] suggested a new approach based on enhanced YOLOX as an anchor-free target detection method, which achieved high-speed and high-precision ship target detection by combining a coordinated attention mechanism and improving the loss function. Zha et al. [45] proposed a novel ship detection model based on multi-feature transformation and fusion (MFTF-Net) to suppress false alarms and improve small target detection performance. To reduce the interference of noisy information, the method applies a modified convolutional block attention module (CBAM) and a squeeze-excited attention (SE) mechanism to the lower and upper two layers of the network output, respectively. Jiang et al. [46] proposed an effective lightweight anchor-free detector called R-Centernet+ and introduced the CBAM module to the backbone network. Zhang et al. [47] proposed a cross-scale region predictive perception network (CSRP-Net), by designing a cross-scale self-attention (CSSA) module, which suppressed the influence of noise and complex background and enhanced the detection ability of multi-scale targets. A multi-scale detection network for ships in SAR pictures was suggested by Zhang et al. [48] and is based on attention and weighted fusion. This technique, which is based on the YOLOv5 architecture, introduces a coordinate attention block to sharpen the position characteristics of ship targets and muffle the interference from complicated backdrops. Guo et al. [49] added the CBAM module and the BiFPN module to the YOLOv5 network, which enables it to fully learn the feature information of the spatial and channel dimensions, and enhance the information fusion transfer between multi-scale targets, and the problem of missed detection of multi-scale targets is well resolved. A novel anchor-free SAR target recognition technique, AFSar, was put forth by Wan et al. [50] based on multi-scale improved representation learning. This approach emphasizes the distinct strong scattering properties of SAR targets by combining channel and spatial attenuation processes. Li et al. [51] proposed a new deep learning network, the attention-guided balanced feature pyramid network (A-BFPN), and designed a channel attention-guided fusion network (CAFN) model to obtain the optimized multi-scale features, reducing severe aliasing effects in the mixed feature maps. Zhou et al. [52] suggested a multi-scale ship detection network (MSSDNet) method based on a small model size YOLOv5 (YOLOv5s) by introducing an improved backbone network Res2Net (MRes2) with a coordinate attention module (CAM)) for the multi-scale feature extraction in scale dimension to better represent the features. Su et al. [53] suggested a new spatial information integration network (SII-Net) detection method, and designed a channel location attention mechanism (CLAM) module to extract location information along two spatial directions to enhance the detection capability of the backbone network.

2.3. Rotation Detection Models

Ship targets are often densely arranged and docked in nearshore areas. The traditional positive frame detection method will cause a large number of missed detections and false detections of targets. In order to improve the detection performance of ship targets in the berth area in SAR images, Zhao et al. [54] proposed a ship rotation detection method based on an orientation-aware feature fusion network (OFF-Net) in SAR images. This method proposes a decoupled orientation perception head that is more robust to ship targets with arbitrary orientations. At the same time, the authors also provide a high-resolution SAR ship detection dataset (OBB-HRSDD) with rotatable bounding boxes. Man et al. [55] proposed a new center-to-corner vector navigation network (CCVNet) for SAR ship rotation detection, which uses an anchor-free method to directly predict the vector from the center to the corner, which can reduce the number of separate predictions’ accumulation of errors caused by angles and scales. Cheng et al. [56] proposed an arbitrary orientation-oriented SAR ship detection network, global context-guided feature balance network (GFB-Net). The method obtains the output feature map by rotating ROIAlign, and uses it for classification and regression in the second stage of the network, and finally realizes ship detection in any direction. Zhao et al. [57] designed a head network with stepwise regression from coarse-grained to fine-grained to accurately detect ships in arbitrary orientations. At the same time, the method also combines the attention module to calibrate the multi-scale fusion features to highlight the ship information while suppressing the surrounding background interference. Li et al. [58] proposed a new directional SAR ship detector and embedded a hybrid convolutional channel attention (MCCA) module in the backbone network to enhance the ship by reweighting all channels of the feature map and boat characterization features. Zhao et al. [59] improved the original deep subdomain adaptation network (DSAN) and designed a dual branch network (DBN) embedded with an attention module to extract more discriminative deep transferable features, thereby improving the subdomain adaptation performance. Sun et al. [60] proposed an arbitrary orientation SAR ship detector (BiFA-YOLO) based on bidirectional feature fusion and angle classification. In order to effectively detect arbitrary orientation and densely distributed ships in HR SAR images, BiFA-YOLO adds an angle classification structure to the head network. He et al. [61] proposed a directional ship detector based on dual-branch probes and adaptive SAR feature enhancement, and an oriented region proposal network (ORPN) was designed to study the conversion of horizontal regions of interest to rotational regions of interest. Shao et al. [62] proposed a rotation-balanced feature alignment network (RBFA-Net) and designed an anchor-guided feature alignment network (AFAN) to adaptively align convolutional features according to the rotated anchor boxes. Xu et al. [63] proposed an arbitrary orientation-oriented SAR ship detection model combining triangle distance IoU loss (TDIoU loss) and attention-weighted feature pyramid network (AW-FPN), and proposed TDIoU loss as a rotating bounding box regression efficient solution for inconsistencies in loss metrics and discontinuities in boundaries.
Based on the above survey, it is not difficult to find that the current work of ship detection on SAR images mainly focuses on how to improve the performance of target detection in complex scenes while reducing the complexity of the model and improving the efficiency of the algorithm. This is also the starting point of our proposed method in this paper. However, different from the above research, we focus more on the comprehensive detection and recognition performance of ship targets.

3. Proposed Method

In this part, a thorough description of the suggested lightweight model will be given. First, the general structure of our model is shown. After that, each important module’s mechanism will be described.

3.1. Overall Framework

Figure 1 depicts the process and general structure of the strategy suggested in this article. We choose the YOLOv5n model as the basic framework, which has a more concise network structure and faster running and reasoning speed, so it is easier to achieve lightweight deployment. Our improved method is mainly composed of three parts: the first part is the improved backbone network in the dotted frame in Figure 1. First of all, in order to effectively suppress the background noise and enhance the prominent target features, we embed a lightweight attention module in each C3 feature extraction module on the basis of the original YOLOv5-n backbone network framework. Second, based on the improvement of the original framework, the attention structure and the simplified spatial pyramid pooling fast structure (SimSPPF) were combined to further improve the generation speed of the target candidate box and effectively save the computational cost. In the part of prediction, we added the angle classification module to the prediction layer, and improved the constraint through the loss function to achieve the directed prediction of ship targets.

3.2. C3_Attention Block

In contrast with the optical image, the ship target contains less information in the SAR image, while the background occupies most of the pixels. Therefore, in complex scenes such as inshore areas and high sea conditions, the ship features extracted by the network are less dominant. For this reason, we choose to add an attention module in the network feature extraction layer to further enhance the highlighting of target features and suppress the interference of background noise. Among many attention modules, we choose to embed the strong attention module squeeze excitation (SE) attention model and the convolutional block attention module (CBAM) based on the original C3 structure and construct new feature extraction structures C3SE and C3CBAM. Figure 2a,b show the structures of the C3SE and C3CBAM. Unlike transformer, which consumes a lot of computing resources to obtain the relationship between features, SE and CBAM modules directly calculate the feature weights in channel dimension and spatial dimension, and have significant advantages in computational efficiency and reasoning speed.
As seen in Figure 2a, C3SE initially decreases the dimension of spatial features by global average pooling based on the width and height of the feature map in order to draw emphasis to the channel dimension. Then, as seen in Equation (1), the link between the channels is established utilizing two complete connection layers and a nonlinear activation function.
F c = T ( R e L U ( T ( 1 H × W i = 1 H j = 1 W F i n ( i , j ) ) ) )
where Fin represents input features, Fc represents middle layer features, H and W represent the width and height of the channel, respectively, ReLU represents the nonlinear activation function, and T represents the fully connected layer.
Following the Sigmoid activation function, the normalized weight is acquired, and it is then multiplied by each channel of the original feature graph to complete the channel attention and recalibrate the original feature, as given in Equation (2).
F o u t = F i n · σ ( F c )
where Fout represents the output features and σ represents the Sigmoid function. The global receptive field may be produced by global average pooling. By lowering the dimension of the feature graph, the parameters and computation in the first complete connection are significantly lowered. It is restored to its original number of channels by a complete connection after a nonlinear activation function, and the correlation between channels is formed. After that, the network can pay more attention to the channel characteristics that contain more information, and reduce the attention to the features with less information.
The C3CBAM attention mechanism module is separated into two sections: spatial attention and channel attention, as seen in Figure 2b. The featured graph is input, followed by the channel attention, GAP, and GMP based on the width and height of the feature graph, the attention weight for the channel through MLP, the normalized attention weight through the Sigmoid function, and finally the weighting of the normalized attention weight to the original input feature graph channel by the channel through multiplication. Complete Equation (3)’s formula for the original feature’s channel attention recalibration:
F c = σ ( M L P ( A v g P o o l ( F i n ) ) + M L P ( M a x P o o l ( F i n ) ) )
where AvgPool represents the average pooling, MaxPool represents the maximum pooling, MLP represents a two-layer neural network.
The feature graph output through the channel also performs global maximum pooling and global average pooling based on the width and height of the feature graph, transforming the feature dimension from H × W to 1 × 1, then reducing the dimension of the feature graph after convolution kernel 7 × 7 and Relu activation function, and finally lifting it to the original dimension after a convolution. This process obtains the attention feature in spatial dimension. Finally, the feature map standardized by Sigmoid activation function and the feature map of channel attention outputs merged to complete the recalibration of the feature graph in the space and channel dimensions, as shown in Equation (4).
F o u t = σ ( f 7 × 7 ( [ A v g P o o l ( F c ) ; M a x P o o l ( F c ) ] ) )
where f 7 × 7 represents the convolution operation with 7 × 7 convolution kernel. In the spatial attention module, the global average pooling and maximum pooling obtain spatial attention features, and the correlation between spatial features is established through two convolutions, while keeping the input and output dimensions unchanged. Through the convolution operation with 7 × 7 convolution kernel, the parameters and the amount of calculation are greatly reduced, which is beneficial to the establishment of high-dimensional spatial feature correlation. The association between each feature in the channel and space is substantially improved after CBAM, which makes it easier to extract the target’s useful characteristics. The new feature graph will also obtain attention weights in the channel and spatial dimensions.

3.3. SimSPPF Block

Spatial pyramid pooling (SPP) was originally proposed by He et al. [64] to solve the problem that the size of the input image cannot always meet the requirements of the input, which leads to the distortion of the network input image. When applied to the target detection network, SPP can effectively solve the problem of graph correlation repetitive feature extraction by convolution neural network, which greatly improves the speed of generating candidate boxes and saves computational costs. Specifically, as shown in Figure 3a, SPP first carries out convolution and normalization activation processing on the input feature, and then carries on the maximum value pool processing to the activation feature on three scales, and then splices the three scale features together with convolution normalization activation to obtain the features, as shown in Equation (5).
F o u t = C B L [ M a x P o o l 5 × 5 C B L ( F i n ) ; M a x P o o l 9 × 9 C B L ( F i n ) ; M a x P o o l 13 × 13 C B L ( F i n ) ; C B L ( F i n ) ]
where CBL represents the combination of “Conv”, “BN”, and “ReLU”.
The SimSPPF module used in this paper is further improved and optimized based on the SPP structure. As shown in Figure 3b, SimSPPF also carries out the convolution and normalized activation of the input features at first, but the difference is that the activated features are not pooled at the same time at three scales, but progressive pooling. In this way, the efficiency of target feature extraction and candidate box selection can be further accelerated. The experimental results show that the processing speed of SimSPPF is significantly higher than that of SPP.

3.4. OBB Prediction Block

The original YOLOv5 network is based on rectangle regression. However, for the SAR images of complex scenes, there are many dense distributions of ship targets, especially in the nearshore scene, often docking many ships. As such, the network will suppress some high-quality prediction boxes when calculating the IOU between the target prediction frames. Therefore, the SAR ship detection method based on the original YOLOv5 network is not effective for ships close to the shore. In order to improve the detection performance, the regression mode of the network detection frame can be changed to the directed frame form. Due to the limitation of the angle, Sun et al. [60] classified and predicted ships in any direction by adding angle components to the structure of the network prediction layer. The experimental results show that the angle classification method can effectively improve the performance of ship target prediction without significantly increasing the complexity of the model. Figure 4 shows the improved rotation prediction head structure adopted in this paper.
Due to the addition of an angle classification module, the training loss function of the model proposed in this paper consists of four parts: classes loss, objectness loss, location loss, and angular loss [60], as shown in Equation (6):
L o s s = L c l s + L o b j + L l o c + L a n g
where L c l s , L o b j , L l o c , and L a n g denote the classes loss, objectness loss, location loss, and angular loss, respectively. Among them, the BCE loss is adopted for L c l s , L o b j , and L a n g , and the CIOU loss is adopted for L l o c . L c l s represents the classification loss of the positive sample [16]. L l o c represents the location loss of the positive sample. L o b j represents the BCE loss between the prediction bounding box with the ground truth box. In order to balance the losses of different scales, the L o b j of three prediction layers are dynamically weighted as Equation (7).
L o b j = λ 1 L o b j s m a l l + λ 2 L o b j m e d i u m + λ 3 L o b j l arg e
where L o b j s m a l l , L o b j m e d i u m , and L o b j l arg e , denote the objectness loss of three scales in the prediction layers, λ 1 , λ 2 , and λ 3 are the equilibrium coefficient.

4. Experiments

We conducted several tests on the recently published ship detection and identification dataset, SRSDD-v1.0, to assess the efficacy of the approaches suggested in this work. The experimental dataset, setup, and assessment indicators will initially be covered in this section. We conducted ablation tests to confirm the efficiency of the suggested module on this foundation. Finally, the outcomes of our visualization experiment using SRSDD datasets are presented and statistically evaluated against those of alternative deep-learning techniques. The results of the experiments demonstrate that the suggested technique has the best detection performance and the shortest model size.

4.1. Dataset

The SRSDD-v1.0 dataset proposed by Lei et al. [28] is built using 30 large-scene images of the Chinese GF-3 spotlight (SL) model, all with an image resolution of 1 m. These large-scene images are cropped into 1024 × 1024-pixel patches. After cropping, the dataset contains 666 SAR image slices and a total of 2884 ships. It is worth noted that this dataset is the first publicly released dataset containing different types of SAR ships. Table 1 shows the information on several datasets. In all published datasets, SRSDD is the only SAR ship dataset with different resolutions and different categories. Therefore, we conduct ship detection and recognition experiments based on this dataset. Figure 5 shows the quantity distribution of six categories of ships in the SRSDD-v1.0 dataset. Figure 6 and Figure 7, respectively, show the distribution of the length, width, and direction angle of the ships in the training and testing datasets. The overall distribution of the training and testing datasets is consistent. Figure 8 shows some examples of ships in the SRSDD data, including the six categories of Ore-oil, Bulk cargo, Finishing, LawEnforce, Dredger, and Container.

4.2. Experimental Setup

All the tests in this work were carried out in PyTorch on a computer with an Intel Core (TM) i7-10875H CPU clocked at 2.30 GHz and an NVIDIA RTX 2070 GPU, as shown in Table 2. The computer was running in Windows 10. Table 2 shows how the computer and deep learning environment were set up for our experiments. In order to solve the problem of insufficient performance caused by the small sample size of the dataset, we used some traditional methods such as rotation and flipping to expand the original dataset to a certain extent. In addition, through the inspection of the data, we found that there were some problems with the original release data label, so we revised the data’s label before training.
The stochastic gradient descent approach is used by the optimizer, and 200 training epochs are used. The learning rate is set to 0.01 and the momentum and weight decay rates are 0.0005 and 0.8, respectively. In every trial, the detection threshold IOU is set to 0.5.

4.3. Evaluation Metrics

We primarily use the accuracy, recall, and F1-score as assessment markers to compare various approaches in order to statistically measure the detection performance. The following are the definitions of recall and precision:
Pr e c i s i o n = T P T P + F P
Re c a l l = T P T P + F N
where the terms “true positives”, “false positives”, and “false negatives” stand for “true positives”, “false positives”, and “missing ships”, respectively. Precision and recall are combined in the F1 score as follows:
F 1 - s c o r e = 2 × Pr e c i s i o n × Re c a l l Pr e c i s i o n + Re c a l l
Furthermore, we applied the frames-per second (FPS) to evaluate the detection efficiency of different methods as follows:
F P S = 1 T p e r - i m g
where Tper-img represents the inference time per image. As for the algorithm complexity, we use the number of parameters, model size, and floating-point operations per second (FLOPS) to evaluate the different methods.

4.4. Ablation Studies

We performed a series of ablation tests on the SRSDD-v1.0 to evaluate the efficacy of each improvement suggested in this study.

4.4.1. Effect of C3_Attention Block

The suggested C3_Attention block module’s effects on the effectiveness of model detection are examined through the tests in this section. Table 3 presents the experimental findings based on the test dataset. It appears that the suggested C3 Attention module can significantly enhance the detection performance, particularly the recall index. Finally, the C3SE module and the C3CBAM increase precision and recall by 2.69%, 2.88%, 3.77%, and 4.12%, respectively, in comparison to the baseline. Additionally, the F1 scores are higher by 3.24 and 3.51%, indicating that the suggested improved model may perform better at comprehensive identification. In terms of model complexity, the two proposed attention modules are lightweight models, so there is no increase in the final model size and FLOPS value.

4.4.2. Effect of SimSPPF Block

Table 4 shows the comparison of experimental results with or without the addition of SimSPPF modules. We can find that replacing SPP with SimSPPF in the backbone network structure will not affect the detection performance of the model, and the comprehensive F1 value of the model detection is still 54.48. In addition, the model complexity, including the total number of parameters, model size and computational complexity, remain unchanged. In terms of the operational efficiency of the model, the addition to SimSPPF structure, can significantly improve the reasoning speed, and FPS increases by 10.28 compared with the baseline.

4.4.3. Effect of OBB Prediction Block

In this section, the OBB prediction block is verified. The same parameter settings are used for training the network in these experiments. The results in Table 5 show that the OBB prediction method yields better overall performance. Specifically, prediction with OBB finally increases the precision and recall by 0.05% and 3.64%, respectively. Furthermore, the F1 scores are increased by 1.84, implying that the effectiveness of angle classification. To demonstrate the advantages of OBB prediction more visually, we visualized the HBB and OBB prediction results, as shown in Figure 9. Among them, different color callout boxes represent different categories. We can find that there are a large number of densely arranged ships in the SRSDD dataset, and there are obvious missing ships in the detection results of HBB prediction. On the contrary, the OBB prediction results are better able to capture a large number of docked or side-by-side ships.

4.5. Comparison with Other Methods

To verify the comprehensive performance of our method, we compare the proposed model with other deep-learning methods in terms of detection performance, inference efficiency, and model complexity (obtained from the corresponding literature), as shown in Table 6. It can be seen that, in the two-stage detection network, O-RCNN [65] has the best comprehensive detection performance in the SRSDD dataset, with the highest detection precision of 64.01%, a detection recall rate of 57.61%, and a comprehensive F1 value of 60.64. In the single-stage detection model, the comprehensive performance of BBAVectors [66] is the best, but the recall rate of this method on SRSDD data is only 34.56%. However, the F1 score of the proposed models in this paper is 61.26, which is 0.86% and 20.36% higher than that of O-RCNN and BBAVectors, respectively. More importantly, our method processes the SRSDD dataset almost seven times faster than other methods, and the model size is only 4.52 M, which is significantly better than the 244 M of the original lightest model R-FCOS [67]. Although the precision index of the proposed method is 59.70%, which is lower than the 64.01% of O-RCNN and 60.56% of R-FCOS. However, the recall index of our model is 62.90%, which is significantly higher than the 57.61% of the O-RCNN and all other methods. The results show that our model can more effectively detect ship targets in complex-scene SAR images. In terms of the running efficiency of different models, the maximum FPS of two-stage detection models for processing images on the SRSDD dataset is only 8.38. Among the other one-stage detection models, R-RetinaNet [68] has the fastest processing speed, but the FPS of R-RetinaNet model is only 10.53. In contrast, through Table 6, we can clearly find that the FPS value of our model proposed in this paper is 68.02, and the processing efficiency has a significant advantage over other one-stage or two-stages detection models.
Table 6. Detection results of different CNN-based methods on SRSDD-v1.0.
Table 6. Detection results of different CNN-based methods on SRSDD-v1.0.
ModelCategoryPrecision (%)Recall (%)F1FPSModel (M)
FR-O [28]Two-stage57.1249.6653.138.09315
ROI [28,69]Two-stage59.3151.2254.977.75421
Gliding Vertex [28,70]Two-stage57.7553.9555.797.58315
O-RCNN [28,65]Two-stage64.0157.6160.648.38315
R-RetinaNet [28,68]One-stage53.5212.5520.3310.53277
R3Det [28,71]One-stage58.0615.4124.367.69468
BBAVectors [28,66]One-stage50.0834.5640.903.26829
R-FCOS [28,67]One-stage60.5618.4228.2510.15244
OursOne-stage59.7062.9061.2668.024.52

4.6. Detection and Recognition Results on SRSDD

Finally, we visualize the detection and recognition results of the proposed model in the nearshore and offshore scenarios of the SRSDD dataset, as shown in Figure 10 and Figure 11. From Figure 10, we can see that the proposed method can basically detect the inshore ships and assign the correct labels. This is mainly because the embedded attention module can strengthen the target features while suppressing the background interference information. For the nearshore scene, the background interference is serious, and the use of the attention enhancement feature can effectively remove the false alarms. Through Figure 11, it can be found that almost all the offshore ship targets are detected by the proposed method, which further verifies the effectiveness of our designed model.
In order to quantitatively analyze the detection and recognition performance of our method on SRSDD data, we calculate the detection and recognition results of six categories of ships, as shown in Table 7 and Table 8. As can be seen from Table 7, the detection rates vary greatly among different categories, in which the comprehensive F1 value of Container can reach 0.71, but that of Fishing is only 0.39. This phenomenon demonstrates that our method still has obvious space for improvement in detection performance. In terms of target recognition performance, as shown in Table 8, the recognition rate of large ships such as Container, Dredger, and Ore-oil is significantly higher than that of small ships such as Fishing and Bulk cargo. Among them, the Background FN in the meter indicates that the false alarms in the detection and recognition results are non-ship targets, which mainly includes land interferers and sea islands. In other words, there is still some room for improvement in the detection performance on all categories of ships. In addition, the results of Table 8 show that the recognition performance of Bulk Cargo, Fishing and LawEnforce is relatively low. This is because Bulk Cargo accounts for the vast majority of the SRSDD dataset, while Fishing and LawEnforce are relatively few. Although this paper adopts some data augmentation methods to improve the recognition performance, the serious data imbalance results in the low recognition rate of these three categories of ships. On the other hand, Bulk Cargo, Fishing, and LawEnforce are mostly distributed along the shore, and the target size is small and often densely distributed, which brings some challenges to network detection and recognition. In any case, the method proposed in this paper can effectively detect and identify six categories of ships to a certain extent, and the current gap is also the focus of our future work.

5. Conclusions

This research suggests a lightweight model for ship detection and recognition in complex-scene SAR images. In order to enhance the model’s detection and recognition performance in complicated scenarios and increase the model’s deployment possibilities, we optimized the module based on three aspects: the designed C3_attention module, the improved SimSPPF module, and the OBB prediction module. The results of many ablation experiments performed on the latest SRSDD-v1.0 dataset demonstrate the potency of each module created in this paper. Experimental findings demonstrate that the model suggested in this paper can perform well, with an F1-Score of 61.26 and an FPS of 68.02 on SRSDD, and using just 1.92 M parameters and 4.52 MB of model memory. It is worth noting that our approach is more suited for practical equipment than previous approaches and can meet the real-time requirements of SAR ship detection and recognition in the future. However, through many comparative experiments, we find that our method still has obvious deficiencies in the performance of ship target recognition due to the influence of image quality and resolution. Therefore, we will consider carrying out more in-depth research on the following aspects: (1) We will build a higher-quality SAR ship detection and recognition dataset including more categories of ships; (2) We will conduct more research on the improvement of SAR image quality to improve the performance of ship target recognition; (3) We will continue to investigate the study on the quick detection and recognition of ship targets in SAR images.

Author Contributions

Conceptualization, B.X.; methodology, B.X.; validation, B.X. and Z.S.; formal analysis, B.X. and Z.S.; investigation, B.X. and Z.S.; resources, Z.S.; data curation, Z.S. and J.W.; writing—original draft preparation, B.X. and Z.S.; writing—review and editing, J.W. and X.L.; visualization, Z.S.; supervision, K.J.; project administration, K.J.; funding acquisition, X.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was jointly supported by National Natural Science Foundation of China (62001480) and the Hunan Provincial Natural Science Foundation of China (2021JJ40684).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors would like to thank the pioneer researchers in SAR ship detection and other related fields.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Xiong, B. Study of Registration and Change Detection in SAR Images; National University of Defense Technology: Changsha, China, 2012. [Google Scholar]
  2. Leng, X.; Ji, K.; Xiong, B.; Kuang, G. Complex Signal Kurtosis—Indicator of Ship Target Signature in SAR Images. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–12. [Google Scholar] [CrossRef]
  3. Sun, Z.; Dai, M.; Leng, X.; Lei, Y.; Xiong, B.; Ji, K.; Kuang, G. An Anchor-Free Detection Method for Ship Targets in High-Resolution SAR Images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 7788–7816. [Google Scholar] [CrossRef]
  4. Kuang, G.; Gao, G.; Jiang, Y.; Lu, J.; Jia, C. Theory, Algorithm and Application for Target Detection in Synthetic Aperture Radar; Press of National University of Defense Technology: Changsha, China, 2007. [Google Scholar]
  5. Zhang, S.; Wu, R.; Xu, K.; Wang, J.; Sun, W. R-CNN-Based Ship Detection from High Resolution Remote Sensing Imagery. Remote Sens. 2019, 11, 631. [Google Scholar] [CrossRef] [Green Version]
  6. Zhang, Y.; Guo, L.; Wang, Z.; Yu, Y.; Liu, X.; Xu, F. Intelligent Ship Detection in Remote Sensing Images Based on Multi-Layer Convolutional Feature Fusion. Remote Sens. 2020, 12, 3316. [Google Scholar] [CrossRef]
  7. Chen, L.; Shi, W.; Deng, D. Improved YOLOv3 Based on Attention Mechanism for Fast and Accurate Ship Detection in Optical Remote Sensing Images. Remote Sens. 2021, 13, 660. [Google Scholar] [CrossRef]
  8. Zheng, W.; Liu, X.; Yin, L. Research on image classification method based on improved multi-scale relational network. PeerJ Comput. Sci. 2021, 7, e613. [Google Scholar] [CrossRef]
  9. Rostami, M.; Kolouri, S.; Eaton, E.; Kim, K. Deep Transfer Learning for Few-Shot SAR Image Classification. Remote. Sens. 2019, 11, 1374. [Google Scholar] [CrossRef] [Green Version]
  10. Zhang, J.; Su, Q.; Wang, C.; Gu, H. Monocular 3D vehicle detection with multi-instance depth and geometry reasoning for autonomous driving. Neurocomputing 2020, 403, 182–192. [Google Scholar] [CrossRef]
  11. Zhou, W.; Guo, Q.; Lei, J.; Yu, L.; Hwang, J.-N. IRFR-Net: Interactive Recursive Feature-Reshaping Network for Detecting Salient Objects in RGB-D Images. IEEE Trans. Neural Netw. Learn. Syst. 2021, 1–13. [Google Scholar] [CrossRef]
  12. Zhang, J.; Su, Q.; Tang, B.; Wang, C.; Li, Y. DPSNet: Multitask Learning Using Geometry Reasoning for Scene Depth and Semantics. IEEE Trans. Neural Netw. Learn. Syst. 2021, 1–12. [Google Scholar] [CrossRef]
  13. Zhou, W.; Yu, L.; Zhou, Y.; Qiu, W.; Wu, M.-W.; Luo, T. Local and Global Feature Learning for Blind Quality Evaluation of Screen Content and Natural Scene Images. IEEE Trans. Image Process. 2018, 27, 2086–2095. [Google Scholar] [CrossRef]
  14. Wang, Y.; Wang, H.; Zhou, B.; Fu, H. Multi-dimensional prediction method based on Bi-LSTMC for ship roll. Ocean Eng. 2021, 242, 110106. [Google Scholar] [CrossRef]
  15. Ban, Y.; Liu, M.; Wu, P.; Yang, B.; Liu, S.; Yin, L.; Zheng, W. Depth Estimation Method for Monocular Camera Defocus Images in Microscopic Scenes. Electronics 2022, 11, 2012. [Google Scholar] [CrossRef]
  16. Su, X.; Ullah, I.; Liu, X.; Choi, D. A Review of Underwater Localization Techniques, Algorithms, and Challenges. J. Sens. 2020, 2020, 6403161. [Google Scholar] [CrossRef] [Green Version]
  17. Li, J.; Qu, C.; Shao, J. Ship detection in SAR images based on an improved faster R-CNN. In Proceedings of the BIGSARDATA, Beijing, China, 13–14 November 2017; pp. 1–6. [Google Scholar]
  18. Ai, J.; Tian, R.; Luo, Q.; Jin, J.; Tang, B. Multi-Scale Rotation-Invariant Haar-Like Feature Integrated CNN-Based Ship Detection Algorithm of Multiple-Target Environment in SAR Imagery. IEEE Trans. Geosci. Remote. Sens. 2019, 57, 10070–10087. [Google Scholar] [CrossRef]
  19. Lin, Z.; Ji, K.; Leng, X.; Kuang, G. Squeeze and excitation rank fasterR-CNN for ship detection in SAR images. IEEE Geosci. Remote Sens. Lett. 2018, 16, 751–755. [Google Scholar] [CrossRef]
  20. Yasir, M.; Jianhua, W.; Mingming, X.; Hui, S.; Zhe, Z.; Shanwei, L.; Colak, A.T.I.; Hossain, S. Ship detection based on deep learning using SAR imagery: A systematic literature review. Soft Comput. 2022, 1–22. [Google Scholar] [CrossRef]
  21. Zhao, S.; Luo, Y.; Zhang, T.; Guo, W.; Zhang, Z. A Feature Decomposition-based Method for Automatic Ship Detection Crossing Different Satellite SAR Images. IEEE Trans. Geosci. Remote Sens. 2022, 60, 3201628. [Google Scholar] [CrossRef]
  22. Yoshida, T.; Ouchi, K. Detection of Ships Cruising in the Azimuth Direction Using Spotlight SAR Images with a Deep Learning Method. Remote Sens. 2022, 14, 4691. [Google Scholar] [CrossRef]
  23. Zheng, H.; Hu, Z.; Liu, J.; Huang, Y.; Zheng, M. MetaBoost: A Novel Heterogeneous DCNNs Ensemble Network With Two-Stage Filtration for SAR Ship Classification. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
  24. Zhang, T.; Zhang, X.; Li, J.; Xu, X.; Wang, B.; Zhan, X.; Xu, Y.; Ke, X.; Zeng, T.; Su, H.; et al. SAR Ship Detection Dataset (SSDD): Official Release and Comprehensive Data Analysis. Remote. Sens. 2021, 13, 3690. [Google Scholar] [CrossRef]
  25. Sun, X.; Wang, Z.; Sun, Y.; Diao, W.; Zhang, Y.; Kun, F. AIR-SARShip-1.0: High-resolution SAR Ship Detection Dataset. J. Radars 2019, 8, 852. [Google Scholar]
  26. Wei, S.; Zeng, X.; Qu, Q.; Wang, M.; Su, H.; Shi, J. HRSID: A high-resolution SAR images dataset for ship detection and instance segmentation. IEEE Access 2020, 8, 120234–120254. [Google Scholar] [CrossRef]
  27. Zhang, T.; Zhang, X.; Ke, X.; Zhan, X.; Shi, J.; Wei, S.; Pan, D.; Li, J.; Su, H.; Zhou, Y.; et al. LS-SSDD-v1.0: A deep learning dataset dedicated to small ship detection from large-scale sentinel-1 SAR images. Remote Sens. 2020, 12, 2997. [Google Scholar] [CrossRef]
  28. Lei, S.; Lu, D.; Qiu, X.; Ding, C. SRSDD-v1.0: A High-Resolution SAR Rotation Ship Detection Dataset. Remote Sens. 2021, 13, 5104. [Google Scholar] [CrossRef]
  29. Zhou, L.; Yu, H.; Wang, Y.; Xu, S.; Gong, S.; Xing, M. LASDNet: A Lightweight Anchor-Free Ship Detection Network for SAR Images. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium, Kuala Lumpur, Malaysia, 17–22 July 2022; pp. 2630–2633. [Google Scholar] [CrossRef]
  30. Cui, J.; Jia, H.; Wang, H.; Xu, F. A Fast Threshold Neural Network for Ship Detection in Large-Scene SAR Images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 6016–6032. [Google Scholar] [CrossRef]
  31. Miao, T.; Zeng, H.; Yang, W.; Chu, B.; Zou, F.; Ren, W.; Chen, J. An Improved Lightweight RetinaNet for Ship Detection in SAR Images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 4667–4679. [Google Scholar] [CrossRef]
  32. Ma, X.; Ji, K.; Xiong, B.; Zhang, L.; Feng, S.; Kuang, G. Light-YOLOv4: An Edge-Device Oriented Target Detection Method for Remote Sensing Images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 10808–10820. [Google Scholar] [CrossRef]
  33. Yang, G.; Lei, J.; Xie, W.; Fang, Z.; Li, Y.; Wang, J.; Zhang, X. Algorithm/Hardware Codesign for Real-Time On-Satellite CNN-Based Ship Detection in SAR Imagery. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–18. [Google Scholar] [CrossRef]
  34. Yang, X.; Zhang, J.; Chen, C.; Yang, D. An Efficient and Lightweight CNN Model With Soft Quantification for Ship Detection in SAR Images. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–13. [Google Scholar] [CrossRef]
  35. Yu, J.; Zhou, G.; Zhou, S.; Qin, M. A Fast and Lightweight Detection Network for Multi-Scale SAR Ship Detection under Complex Backgrounds. Remote Sens. 2021, 14, 31. [Google Scholar] [CrossRef]
  36. Chang, Y.-L.; Anagaw, A.; Chang, L.; Wang, Y.C.; Hsiao, C.-Y.; Lee, W.-H. Ship Detection Based on YOLOv2 for SAR Imagery. Remote Sens. 2019, 11, 786. [Google Scholar] [CrossRef] [Green Version]
  37. Yu, W.; Wang, Z.; Li, J.; Luo, Y.; Yu, Z. A Lightweight Network Based on One-Level Feature for Ship Detection in SAR Images. Remote Sens. 2022, 14, 3321. [Google Scholar] [CrossRef]
  38. Feng, Y.; Chen, J.; Huang, Z.; Wan, H.; Xia, R.; Wu, B.; Sun, L.; Xing, M. A Lightweight Position-Enhanced Anchor-Free Algorithm for SAR Ship Detection. Remote Sens. 2022, 14, 1908. [Google Scholar] [CrossRef]
  39. Li, S.; Fu, X.; Dong, J. Improved Ship Detection Algorithm Based on YOLOX for SAR Outline Enhancement Image. Remote Sens. 2022, 14, 4070. [Google Scholar] [CrossRef]
  40. Xu, X.; Zhang, X.; Zhang, T. Lite-YOLOv5: A Lightweight Deep Learning Detector for On-Board Ship Detection in Large-Scene Sentinel-1 SAR Images. Remote. Sens. 2022, 14, 1018. [Google Scholar] [CrossRef]
  41. Guo, Y.; Chen, S.; Zhan, R.; Wang, W.; Zhang, J. LMSD-YOLO: A Lightweight YOLO Algorithm for Multi-Scale SAR Ship Detection. Remote Sens. 2022, 14, 4801. [Google Scholar] [CrossRef]
  42. Liu, S.; Kong, W.; Chen, X.; Xu, M.; Yasir, M.; Zhao, L.; Li, J. Multi-Scale Ship Detection Algorithm Based on a Lightweight Neural Network for Spaceborne SAR Images. Remote Sens. 2022, 14, 1149. [Google Scholar] [CrossRef]
  43. Gao, Y.; Wu, Z.; Ren, M.; Wu, C. Improved YOLOv4 Based on Attention Mechanism for Ship Detection in SAR Images. IEEE Access 2022, 10, 23785–23797. [Google Scholar] [CrossRef]
  44. Peng, H.; Tan, X. Improved YOLOX’s Anchor-Free SAR Image Ship Target Detection. IEEE Access 2022, 10, 70001–70015. [Google Scholar] [CrossRef]
  45. Zha, M.; Qian, W.; Yang, W.; Xu, Y. Multi-Feature Transformation and Fusion-Based Ship Detection with Small Targets and Complex Backgrounds. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
  46. Jiang, Y.; Li, W.; Liu, L. R-CenterNet+: Anchor-Free Detector for Ship Detection in SAR Images. Sensors 2021, 21, 5693. [Google Scholar] [CrossRef] [PubMed]
  47. Zhang, L.; Liu, Y.; Huang, Y.; Qu, L. Regional Prediction-Aware Network with Cross-Scale Self-Attention for Ship Detection in SAR Images. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
  48. Zhang, L.; Chu, Z.; Zou, B. Multi Scale Ship Detection Based on Attention and Weighted Fusion Model for High Resolution SAR Images. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium, Kuala Lumpur, Malaysia, 17–22 July 2022; pp. 631–634. [Google Scholar] [CrossRef]
  49. Guo, Y.; Chen, S.; Zhan, R.; Wang, W.; Zhang, J. SAR Ship Detection Based on YOLOv5 Using CBAM and BiFPN. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium, Kuala Lumpur, Malaysia, 17–22 July 2022; pp. 2147–2150. [Google Scholar] [CrossRef]
  50. Wan, H.; Chen, J.; Huang, Z.; Xia, R.; Wu, B.; Sun, L.; Yao, B.; Liu, X.; Xing, M. AFSar: An Anchor-Free SAR Target Detection Algorithm Based on Multiscale Enhancement Representation Learning. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–14. [Google Scholar] [CrossRef]
  51. Li, X.; Li, D.; Liu, H.; Wan, J.; Chen, Z.; Liu, Q. A-BFPN: An Attention-Guided Balanced Feature Pyramid Network for SAR Ship Detection. Remote Sens. 2022, 14, 3829. [Google Scholar] [CrossRef]
  52. Zhou, K.; Zhang, M.; Wang, H.; Tan, J. Ship Detection in SAR Images Based on Multi-Scale Feature Extraction and Adaptive Feature Fusion. Remote Sens. 2022, 14, 755. [Google Scholar] [CrossRef]
  53. Su, N.; He, J.; Yan, Y.; Zhao, C.; Xing, X. SII-Net: Spatial Information Integration Network for Small Target Detection in SAR Images. Remote Sens. 2022, 14, 442. [Google Scholar] [CrossRef]
  54. Zhao, M.; Shi, J.; Wang, Y. Orientation-Aware Feature Fusion Network for Ship Detection in SAR Images. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
  55. Xiao, M.; He, Z.; Lou, A.; Li, X. Center-to-Corner Vector Guided Network for Arbitrary-Oriented Ship Detection in Synthetic Aperture Radar Images. In Proceedings of the 2022 3rd International Conference on Geology, Mapping and Remote Sensing (ICGMRS), Zhoushan, China, 22–24 April 2022; pp. 293–297. [Google Scholar] [CrossRef]
  56. Yao, C.; Bai, L.; Xue, D.; Lin, X.; Ye, Z.; Wang, Y.; Yin, K. GFB-Net: A Global Context-Guided Feature Balance Network for Arbitrary-Oriented SAR Ship Detection. In Proceedings of the 2022 7th International Conference on Image, Vision and Computing (ICIVC), Xi’an, China, 26–28 July 2022; pp. 166–171. [Google Scholar] [CrossRef]
  57. Zhao, S.; Liu, Q.; Yu, W.; Lv, J. A Single-Stage Arbitrary-Oriented Detector Based on Multi-Scale Feature Fusion and Calibration for SAR Ship Detection. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 8179–8198. [Google Scholar] [CrossRef]
  58. Li, C.; Lang, H. An Oriented SAR Ship Detector with Mixed Convolution Channel Attention Module and Geometric Non-Maximum Suppression. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 8074–8084. [Google Scholar] [CrossRef]
  59. Zhao, S.; Lang, H. Improving Deep Subdomain Adaptation by Dual-Branch Network Embedding Attention Module for SAR Ship Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 8038–8048. [Google Scholar] [CrossRef]
  60. Sun, Z.; Leng, X.; Lei, Y.; Xiong, B.; Ji, K.; Kuang, G. BiFA-YOLO: A Novel YOLO-Based Method for Arbitrary-Oriented Ship Detection in High-Resolution SAR Images. Remote Sens. 2021, 13, 4209. [Google Scholar] [CrossRef]
  61. He, B.; Zhang, Q.; Tong, M.; He, C. Oriented Ship Detector for Remote Sensing Imagery Based on Pairwise Branch Detection Head and SAR Feature Enhancement. Remote Sens. 2022, 14, 2177. [Google Scholar] [CrossRef]
  62. Shao, Z.; Zhang, X.; Zhang, T.; Xu, X.; Zeng, T. RBFA-Net: A Rotated Balanced Feature-Aligned Network for Rotated SAR Ship Detection and Classification. Remote Sens. 2022, 14, 3345. [Google Scholar] [CrossRef]
  63. Xu, Z.; Gao, R.; Huang, K.; Xu, Q. Triangle Distance IoU Loss, Attention-Weighted Feature Pyramid Network, and Rotated-SARShip Dataset for Arbitrary-Oriented SAR Ship Detection. Remote Sens. 2022, 14, 4676. [Google Scholar] [CrossRef]
  64. He, K.; Zhang, X.; Ren, S.; Sun, J. Spatial Pyramid Pooling in Deep Convolutional Networks for Visual Recognition. European Conference on Computer Vision. arXiv 2014, arXiv:1406.4729. [Google Scholar]
  65. Xie, X.; Cheng, G.; Wang, J. Oriented R-CNN for Object Detection. arXiv 2021, arXiv:2108.05699. [Google Scholar]
  66. Xu, Y.; Fu, M.; Wang, Q.; Wang, Y.; Chen, K.; Xia, G.; Bai, X. Gliding Vertex on the Horizontal Bounding Box for Multi-Oriented Object Detection. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 1452–1459. [Google Scholar] [CrossRef]
  67. Tian, Z.; Shen, C.; Chen, H.; He, T. FCOS: Fully Convolutional One-Stage Object Detection. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Korea, 27 October–2 November 2019; pp. 9626–9635. [Google Scholar]
  68. Rotated-RetinaNet. Available online: https://github.com/ming71/Rotated-RetinaNet (accessed on 27 October 2020).
  69. Law, H.; Deng, J. Cornernet: Detecting objects as paired keypoints. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 765–781. [Google Scholar]
  70. Yi, J.; Wu, P.; Liu, B.; Huang, Q.; Qu, H.; Metaxas, D. Oriented Object Detection in Aerial Images with Box Boundary-Aware Vectors. In Proceedings of the 2021 IEEE Winter Conference on Applications of Computer Vision (WACV), 5–9 January 2021; pp. 2149–2158. [Google Scholar]
  71. Yang, X.; Yan, J.; Feng, Z.; He, T. R3Det: Refined Single-Stage Detector with Feature Refinement for Rotating Object. arXiv 2019, arXiv:1908.05612. [Google Scholar] [CrossRef]
Figure 1. The overall framework of the proposed lightweight model.
Figure 1. The overall framework of the proposed lightweight model.
Remotesensing 14 06053 g001
Figure 2. The structures of two proposed different attention blocks. (a) The C3SE block; (b) The C3CBAM block. “Ci” represents the number of feature channels, “Conv” means the ordinary convolution-2D layer, “BN” means the batch normalization layer, “SiLU” means swish function, “BottleNeck” is formed by stacking multiple convolution layers, “GAvgPool” means the global average pooling-2D layer; “GMaxPool” means the global max pooling-2D layer; “FC” means the fully connection layer, and the “r” is a reduction factor, “Max-C“ represents the maximum-channel-pixel, “Mean-C” means the average-channel-pixel.
Figure 2. The structures of two proposed different attention blocks. (a) The C3SE block; (b) The C3CBAM block. “Ci” represents the number of feature channels, “Conv” means the ordinary convolution-2D layer, “BN” means the batch normalization layer, “SiLU” means swish function, “BottleNeck” is formed by stacking multiple convolution layers, “GAvgPool” means the global average pooling-2D layer; “GMaxPool” means the global max pooling-2D layer; “FC” means the fully connection layer, and the “r” is a reduction factor, “Max-C“ represents the maximum-channel-pixel, “Mean-C” means the average-channel-pixel.
Remotesensing 14 06053 g002
Figure 3. The structures of the SPP and SimSPPF block. “RF” represents the receptive field.
Figure 3. The structures of the SPP and SimSPPF block. “RF” represents the receptive field.
Remotesensing 14 06053 g003
Figure 4. The structure of the oriented bounding box (OBB) prediction block.
Figure 4. The structure of the oriented bounding box (OBB) prediction block.
Remotesensing 14 06053 g004
Figure 5. The quantity distributions of six categories of ships in SRSDD-v1.0 dataset.
Figure 5. The quantity distributions of six categories of ships in SRSDD-v1.0 dataset.
Remotesensing 14 06053 g005
Figure 6. The distributions of the ships’ width, length, and angle on the training dataset.
Figure 6. The distributions of the ships’ width, length, and angle on the training dataset.
Remotesensing 14 06053 g006
Figure 7. The distributions of the ships’ width, length, and angle on the test dataset.
Figure 7. The distributions of the ships’ width, length, and angle on the test dataset.
Remotesensing 14 06053 g007
Figure 8. Instances of SRSDD-v1.0 dataset. (a) The original SAR images. (b) The ground truth. (c) The corresponding optical images. From left to right are the Ore-oil, Bulk cargo, Finishing, LawEnforce, Dredger, and Container.
Figure 8. Instances of SRSDD-v1.0 dataset. (a) The original SAR images. (b) The ground truth. (c) The corresponding optical images. From left to right are the Ore-oil, Bulk cargo, Finishing, LawEnforce, Dredger, and Container.
Remotesensing 14 06053 g008
Figure 9. Detection results of different prediction methods on SRSDD. (a) The original SAR images. (b) The ground truth. (c) HBB prediction results. (d) OBB prediction results. Yellow ellipse indicates a false prediction.
Figure 9. Detection results of different prediction methods on SRSDD. (a) The original SAR images. (b) The ground truth. (c) HBB prediction results. (d) OBB prediction results. Yellow ellipse indicates a false prediction.
Remotesensing 14 06053 g009
Figure 10. Detection and recognition results of inshore ships. The first column is the original SAR images, the second column is the ground truth, and the third column is our results.
Figure 10. Detection and recognition results of inshore ships. The first column is the original SAR images, the second column is the ground truth, and the third column is our results.
Remotesensing 14 06053 g010
Figure 11. Detection and recognition results of offshore ships. The first column is the original SAR images, the second column is the ground truth, and the third column is our results.
Figure 11. Detection and recognition results of offshore ships. The first column is the original SAR images, the second column is the ground truth, and the third column is our results.
Remotesensing 14 06053 g011
Table 1. Comparison of different SAR ship datasets.
Table 1. Comparison of different SAR ship datasets.
DatasetSize (Pixel)Image (Num)Ship (Num)AnnotationsResolution (m)Categories
SSDD190–66811602586HBB1–151
SSDD+190–66811602586OBB1–151
Official-SSDD190–66811602586Polygon1–151
SAR-Ship-Dataset256 × 25643,81959,535HBB3–251
Air-SARship-1.03000 × 300031461HBB1, 31
Air-SARship-2.01000 × 10003002040HBB1, 31
HRSID800 × 800560416,951Polygon0.5, 1, 31
LS-SSDD-v1.024,000 × 16,000156015HBB5 × 201
RSDD-SAR512 × 512700010,263OBB2–201
SRSDD-v1.01024 × 10246662884OBB16
Table 2. Experimental setup and environment.
Table 2. Experimental setup and environment.
ProjectModel/Parameter
Systemwindows 10
RAM32 GB
CPUIntel i7-10875H
GPUNVIDIA RTX 2070
PlatformPyTorch
Codepython3.8
FrameworkCUDA10.1/cudnn7.6.5
Epochs200
Learning rate0.01
Momentum0.0005
Table 3. Performance comparison of different attention modules.
Table 3. Performance comparison of different attention modules.
MethodsPrecision (%)Recall (%)F1FPSModel (MB)FLOPS
YOLOv5n (Base)55.4253.7754.58 75.194.064.2G
Base + C3SE58.1157.5457.82 73.004.064.2G
Base + C3CBAM58.3057.8958.0972.464.064.2G
Table 4. Performance comparison of with and without the SimSPPF Block.
Table 4. Performance comparison of with and without the SimSPPF Block.
MethodsF1FPSParam (M)Model (MB)FLOPS
YOLOv5n (Base)54.58.75.191.684.064.2G
Base + SimSPPF54.5885.471.684.064.2G
Table 5. Performance comparison of different attention modules with OBB prediction method.
Table 5. Performance comparison of different attention modules with OBB prediction method.
MethodsPrecision (%)Recall (%)F1FPSModel (MB)FLOPS
YOLOv5n (Base)55.4253.7754.58 75.194.064.2G
Base + OBB55.4757.4156.4269.404.525.0G
Table 7. Detection results of six categories of ships.
Table 7. Detection results of six categories of ships.
ClassPrecision (%)Recall (%)F1
Ore-oil53.546.70.50
Bulk cargo52.659.30.56
Fishing64.328.00.39
LawEnforce44.2100.00.61
Dredger77.467.00.72
Container66.476.20.71
Table 8. The confusion matrix of proposed method in the six categories of ships.
Table 8. The confusion matrix of proposed method in the six categories of ships.
Ore-OilBulk CargoFishingLawEnforceDredgerContainerBackground FN
Ore-oil0.470.010.000.000.020.000.02
Bulk cargo0.000.530.130.000.100.100.78
Fishing0.000.000.220.000.000.000.08
LawEnforce0.000.000.000.250.000.000.04
Dredger0.000.000.000.000.630.000.03
Container0.000.000.000.000.000.710.04
Background FN0.530.450.640.750.240.190.00
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Xiong, B.; Sun, Z.; Wang, J.; Leng, X.; Ji, K. A Lightweight Model for Ship Detection and Recognition in Complex-Scene SAR Images. Remote Sens. 2022, 14, 6053. https://doi.org/10.3390/rs14236053

AMA Style

Xiong B, Sun Z, Wang J, Leng X, Ji K. A Lightweight Model for Ship Detection and Recognition in Complex-Scene SAR Images. Remote Sensing. 2022; 14(23):6053. https://doi.org/10.3390/rs14236053

Chicago/Turabian Style

Xiong, Boli, Zhongzhen Sun, Jin Wang, Xiangguang Leng, and Kefeng Ji. 2022. "A Lightweight Model for Ship Detection and Recognition in Complex-Scene SAR Images" Remote Sensing 14, no. 23: 6053. https://doi.org/10.3390/rs14236053

APA Style

Xiong, B., Sun, Z., Wang, J., Leng, X., & Ji, K. (2022). A Lightweight Model for Ship Detection and Recognition in Complex-Scene SAR Images. Remote Sensing, 14(23), 6053. https://doi.org/10.3390/rs14236053

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop