Next Article in Journal
Automated Mapping of the Freshwater Ecosystem Functional Groups of the International Union for Conservation of Nature Global Ecosystem Typology in a Large Region of Arid Australia
Previous Article in Journal
Post-Little Ice Age Equilibrium-Line Altitude and Temperature Changes in the Greater Caucasus Based on Small Glaciers
Previous Article in Special Issue
An Approach for SAR Feature Reconfiguring Based on Periodic Phase Modulation with Inter-Pulse Time Bias
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

TPNet: A High-Performance and Lightweight Detector for Ship Detection in SAR Imagery

School of Remote Sensing and Information Engineering, Wuhan University, Wuhan 430079, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2025, 17(9), 1487; https://doi.org/10.3390/rs17091487
Submission received: 21 March 2025 / Revised: 14 April 2025 / Accepted: 16 April 2025 / Published: 22 April 2025

Abstract

:
The advancement of SAR satellites enables continuous and real-time ship monitoring on water surfaces regardless of time and weather. Traditional ship detection algorithms in SAR imagery using manually designed operators lack accuracy, while many existing deep learning-based detection algorithms are computationally intensive and have room for accuracy improvement. Inspired by CenterNet, we propose the Three Points Network (TPNet). It locates the ship’s center point and estimates distances to the top-left and bottom-right corners for precise positioning. We introduce several innovative mechanisms to enhance TPNet’s performance, improving both accuracy and computational efficiency. Evaluated on the open-source SAR-Ship-Dataset, TPNet outperforms 14 other deep learning-based detection algorithms in accuracy and efficiency. Its strong generalization ability is further verified on SSDD and HRSID datasets. These results show TPNet’s potential in real-time maritime surveillance and monitoring systems.

1. Introduction

Synthetic Aperture Radar (SAR) technology, with its active microwave sensor, has emerged as a valuable tool in marine science due to its ability to operate effectively in all weather and lighting conditions. Since the launch of the SEASAT1 satellite by the United States in 1978, SAR technology has been widely utilized, and many countries have launched their own SAR satellites, leading to a surge of interest in the processing of SAR images. In particular, ship detection algorithms for SAR images have emerged as a significant research focus in remote sensing, attracting growing attention from researchers and practitioners dedicated to developing advanced detection techniques. Traditional ship detection algorithms generally employ a sliding window approach to partition an image into smaller regions. Within each region, hand-crafted features such as HOG [1], LBP [2], DoG [3] or SIFT [4] are extracted. Subsequently, a classification algorithm, typically a Support Vector Machine (SVM), is applied to ascertain the presence of a ship within each region. For example, Gan et al. [5] utilized a continuous interval rotating sliding window and HOG features to locate ships. Lin et al. [6] put forward a compact and effective feature named MSHOG, which can be computed by four steps to classify ships in SAR images. Wang et al. [7] proposed a detector based on DoG to detect ships in Radarsat-2 ScanSAR data. Guo et al. [8] utilized SIFT operator and Spatial Pyramid Matching to detect ships in optical images. In addition, the Constant False Alarm Rate (CFAR) is also frequently used in SAR ship detection. For instance, Leng et al. [9] put forward a bilateral CFAR algorithm, which combined the intensity distribution and spatial distribution of SAR images to reduce the impact of SAR ambiguity and sea clutter; Wang et al. [10] proposed an Intensity-Space Domain CFAR algorithm to detect ships in HR SAR images. However, traditional ship detection methods cannot extract advanced semantic information from images, resulting in unsatisfactory detection accuracy.
Compared to hand-designed feature extraction operators, convolutional neural networks (CNNs) can autonomously learn to extract useful features, providing strong robustness and the ability to handle more complex computer vision tasks. The pioneer of deep learning in object detection is RCNN [11], and since then, researchers have proposed more and more object detection models based on deep learning, which can be mainly divided into two categories: two-stage object detection models represented by Faster R-CNN [12] and Mask RCNN [13], and one-stage object detection models represented by the YOLO series [14,15,16,17,18,19,20,21,22].
As deep learning has been widely used in generalized target detection, many researchers are studying the application of deep learning in ship detection for SAR images. Li et al. [23] proposed an SAR image ship detection algorithm based on improved Faster RCNN. Chang et al. [24] put forward a detector named YOLOv2-reduced for SAR imagery, which is based on YOLOv2. Zhang et al. [25] presented a one-stage detector relying on a grid convolutional network. Zhang et al. [26] offered a high-speed detector constructed on a depthwise separable convolutional neural network and a pointwise convolutional neural network. Gui et al. [27] introduced a scale transfer module to facilitate the detection of small ships in SAR images. Jiao et al. [28] proposed a densely connected multi-scale network within the Faster-RCNN framework to detect multi-scale ships. Li et al. [29] utilized a deep ResNet network and transfer learning to precisely locate the target. Lin et al. [30] combined the Squeeze-and-Excitation Mechanism with Faster RCNN for ship detection in SAR imagery. Zhao et al. [31] proposed a two-stage detector called the attention receptive pyramid network, which can strengthen the relationships among non-local features and refine information at different feature maps. Bao et al. [32] put forward a technology known as the optical-SAR matching pretraining technique, which helps the SAR detector learn information from an optical detector. Zhang et al. [33] proposed a hyper-lightweight detector named HyperLi-Net, which achieved excellent performance on some public datasets. Zhang et al. [34] introduced a quad-feature pyramid network composed of four unique FPNs to enhance the detection performance. Xu et al. [35] proposed a lightweight YOLOv5 that can be deployed on an NVIDIA Jetson TX2 platform. Gao et al. [36] introduced an improved YOLOv4 based on an attention mechanism. Jiang et al. [37] proposed a tiny YOLOv4 capable of achieving high speed and good detection accuracy. Guo et al. [38] proposed an improved version of YOLOv5 based on CBAM and BiFPN, which shows better performance compared to the original YOLOv5. They also improved YOLOv5 by incorporating CBAM and BiFPN, leading to enhanced performance. Tang et al. [39] introduced DBW-YOLO, an optimized version of YOLOv7 that excels in detecting small and near-shore ships. Liu et al. [40] proposed CLFR-Det, a novel detector that makes use of the Swin transformer and two advanced modules, achieving outstanding performance in detecting tiny ships.
The ship detectors mentioned above rely on the anchor mechanism, a concept introduced by Faster R-CNN [12]. This approach involves predefining anchor boxes with fixed sizes and shapes on the feature map, which assists in locating and classifying objects. However, the anchor mechanism tends to underperform when handling objects of diverse sizes and shapes, as the anchor boxes are predetermined and fixed. These anchor boxes are typically derived from clustering the bounding boxes in the target dataset, which can constrain the model’s ability to generalize to new or unseen datasets. In contrast, an anchor-free mechanism dispenses with the need for predefined anchor boxes, enabling direct object detection based on the features extracted from the input image. Prominent examples of anchor-free detection models include CenterNet [41], FCOS [42], and PP-YOLOE [22]. Nowadays, an increasing number of SAR ship detection algorithms based on the anchor-free mechanism have been proposed. Gao et al. [43] introduced an anchor-free detector that integrates dense attention feature aggregation to enhance ship detection capabilities; Feng et al. [44] proposed a light-weight, position-enhanced anchor-free detector built upon the YOLOX framework; Yao et al. [45] developed a two-stage anchor-free detector, utilizing two innovative mechanisms to improve detection accuracy; He et al. [46] presented a Gaussian-guided detection head aimed at enhancing detector performance for small ships; Zhu et al. [47] introduced R-FCOS, an extension of FCOS, which incorporates CIoU loss to further refine detection accuracy; Sun et al. [48]. proposed an anchor-free detector named CP-FCOS, which showed excellent ship detection for high-resolution SAR imagery.
Previous studies have achieved significant advancements in enhancing the accuracy of SAR-based ship detection and improving the efficiency of detection algorithms. However, ship detection in SAR imagery using deep learning techniques continues to face substantial challenges: (1) Current research on small ship detection and complex scene detection is inadequate, and related small target detection algorithms rarely consider reducing computational requirements. (2) Many algorithms either prioritize accuracy at the expense of excessive computational resources or emphasize efficiency at the cost of reduced accuracy. Such trade-offs are not ideal for real-world applications, where both high-level accuracy and real-time performance are essential. (3) A common shortcoming of many detection algorithms is the insufficient focus on the quality of detection boxes. Even when detection is successful, the precision of these boxes directly impacts the reliability and usability of the results. High-precision detection boxes are critical for practical applications, as they determine the accuracy of subsequent analyses and decision-making processes. (4) Few studies provide a quantitative analysis of their algorithms’ generalization capabilities. This lack of a scientific quantitative evaluation makes it challenging to assess how well detection algorithms perform across different datasets.
To address these challenges, this paper draws inspiration from CenterNet [41] and proposes the Three Points Network (TPNet). TPNet enhances the detection of small ships by leveraging high-resolution feature layers. It also introduces a lightweight MBlock module to reduce computational complexity while improving detection accuracy. Furthermore, TPNet incorporates the Refine Bounding Box Head (RBH) and other innovative mechanisms to improve the quality of detection boxes. Extensive experiments across multiple datasets demonstrate TPNet’s superior generalization performance, confirming its effectiveness and robustness in various scenarios. The main contributions of this paper are:
  • We introduce TPNet, a novel SAR ship detector inspired by CenterNet. TPNet significantly enhances the detection of small ships by leveraging high-resolution feature layers for prediction. This approach addresses the limitations of existing methods in detecting small ships while maintaining computational efficiency.
  • TPNet achieves lightweight design through the introduction of MBlock, reducing computational cost to only 0.485 G FLOPs, a 92.5% reduction compared to CenterNet. Additionally, Dynamic Feature Refinement Module (DFRM), Refine Bounding-Box Head (RBH), Refine Scoring Branch (RSB), Weighted GIoU(WGIoU) Loss, and Weighted Squeeze-and-Excitation (WSE) Attention Mechanism are integrated to further boost performance.
  • Extensive experiments on the open-source SAR-Ship-Dataset demonstrate that TPNet achieves state-of-the-art performance with an average precision of 95.7% at an IoU threshold of 0.5 ( AP 50 ). Experiments on additional datasets (SSDD and HRSID) validate TPNet’s strong generalization ability. Comprehensive ablation studies also highlight the individual and combined contributions of each proposed mechanism.
The remaining part of this article is structured as follows: Section 2 provides a comprehensive overview of the TPNet algorithm. Section 3 elaborates on the experimental setup in detail, covering the datasets utilized, evaluation metrics, and experimental particulars. Section 4 presents the experimental results, including comparative experiments with other algorithms and ablation experiments. Finally, Section 5 concludes this paper with a comprehensive summary.

2. Methodology

2.1. The Basic Structure of TPNet

As shown in Figure 1, TPNet is primarily composed of three components: Backbone, Neck, and Head. The Backbone is a lightweight ResNet18 [49] called Light Weight MBlock Network(MNet), which is composed of MBlock, and responsible for extracting the features of the input image. The Neck consists of a lightweight FPN [50], designed to further refine the features extracted by MNet, thereby enhancing their representational capacity. The Head is responsible for predicting the locations and detected ships.
In TPNet, a ship is represented using three critical points: the center point, the left-top corner, and the right-bottom corner, as depicted in Figure 2. The input image has dimensions h × w × 3 , and the Lightweight MBlock Network (MNet) is utilized to extract high-level features from this image. The feature maps produced by MNet are subsequently processed by the Neck module, which performs fusion operations to generate a final feature map of size h D R × w D R × 32 . Here, D R denotes the downsample ratio, which is set to 4 and will be elaborated upon in Section 4.2.1. This feature map is then passed to the Head, which outputs 3 maps: a center map and two corner maps. The center map is a heatmap used to locate the center of the target ship, while the corner maps compute the distances from the center point to the upper-left and lower-right corners of the bounding box. Additionally, the RSB module generates a rescoring map to further refine the center map’s output.

2.2. MNet: An Efficient Backbone Architecture for Feature Extraction

In the pursuit of improving object detection performance, researchers have developed deep networks like ResNet101 [49] and Swin Transformer [51] that exhibit remarkable detection abilities. However, these networks come with drawbacks such as a large number of parameters, high computational costs, and slow running speeds, making them unsuitable for real-time ship detection. To address these limitations, many researchers have focused on designing lightweight networks. Currently, the predominant lightweight structures are MobileNet V1 [52] and MobileNet V2 [53] modules. These modules utilize depth-wise separable convolution, which includes two key components: Depthwise (DW) Convolution and Pointwise (PW) Convolution. DW Convolution operates by convolving each input channel independently, while PW Convolution applies a 1 × 1 convolution kernel to the output of the DW Convolution layer, facilitating channel transformation and nonlinear mapping. Inspired by the inverted bottleneck concept of MobileNet V2 and RepVGG [54], we have devised a high-performance lightweight module named MBlock.
As illustrated in Figure 3, MBlock consists of five key components: Conv1, Conv2, Conv3, Conv4, and the attention block. Conv1 is composed of a PW convolutional layer, a batch normalization (BN) layer, and an activation function layer, with the hard-swish function utilized as the activation function. The Conv2 layer is constructed using a sequence of DW convolutions with different kernel sizes, each followed by a BN layer. The output features from these DW-BN layers are aggregated through summation and subsequently passed through an activation function for refinement. Conv3 includes a PW convolutional layer and a BN layer, which are employed to adjust the channel dimensions of the output features. Conv4 functions as a downsampling layer, consisting of a convolutional layer and a BN layer. The Attention Block is a WSE Layer, which will be further detailed in Section 2.6.
When a feature map of size h × w × c 1 is input to MBlock, Conv1 functions to elevate the number of channels from c 1 to m i d _c, thereby preserving the feature map’s information. Here, m i d _ c = c 1 × E R , where E R is a hyperparameter representing the expansion ratio. The impact of E R on model accuracy and complexity will be examined in Section 4.2.3. To balance computational cost and accuracy, E R is set to 3 in this work. The output of Conv2 has dimensions h s × w s × m i d _c (s denotes the stride of MBlock). This output is then compressed by Conv3, producing the final feature map with dimensions h s × w s × c 2 . The role of Conv4 is to reshape the input feature map to align its size with the output feature map, facilitating their subsequent addition.
The output of MBlock is:
y = conv3 ( conv2 ( conv1 ( x ) ) ) + conv4 ( x )
Conv2 exhibits two operational modes, as depicted in Figure 4: training mode and testing mode. During training, a multi-branch network employing kernels of varying sizes, possessing an extensive set of parameters and high precision, is employed. To enhance the performance further, we incorporate a learnable parameter at the culmination of each branch, serving as a gate unit. In the testing mode, a single-branch network with reduced parameters and augmented speed is utilized, and the weights of the multi-branch network are transposed to the single-branch network via structural reparameterization. The gating unit can also be assimilated into the weights of the single-branch network via parameter reconstruction. The outcome of Conv2 is demonstrated as Equation (2).
Conv2 ( x ) = f BN DW ( 5 ) ( x ) + α 1 × BN DW ( 3 ) ( x ) + α 2 × BN DW ( 1 ) ( x ) trainig mode and stride = 1 f BN DW ( 5 ) ( x ) + α 1 × BN DW ( 3 ) ( x ) trainig mode and stride = 2 f D W ( 5 ) ( x ) testing mode
By utilizing MBlock, we have constructed MNet, as detailed in Table 1. Furthermore, a comparative evaluation of MBlock against traditional convolutional methods will be presented in Section 4.2.2.

2.3. MFPN: An Enhanced Feature Extraction Neck for Robust Feature Fusion

The features obtained by the MNet include high-level features that have high semantic information but low-resolution and low-level features that have high-resolution but low semantic information. The role of the neck is to fuse these features to improve the ability of the model to extract information. For TPNet, we selected the FPN [50] architecture as the neck component and named it MFPN (MBlock FPN). The MFPN architecture integrates MBlock and PW convolution layers specifically designed to reduce parameter count and computational complexity. Additionally, we developed a new lightweight upsampling method. In this approach, the input features are first adjusted by MBlock to match the required number of output channels. Then, the spatial dimensions are expanded using bilinear interpolation, as illustrated in Figure 5.
To optimally harness the multi-scale features extracted from the backbone— encompassing rich semantic information in high-level features and abundant detailed information in low-level features—we introduced the DFRM, as depicted in Figure 6. The computational process is structured as follows:
  • Global information extraction: adaptive average pooling is applied to each feature map to capture global information. These global feature descriptors are subsequently concatenated and processed through lightweight convolutional layers to learn a set of weights.
  • Weight calculation: the weights are derived via convolutional operations followed by a sigmoid activation function. These weights reflect the importance of each feature map in the current scene.
  • Feature refinement: the refined features are obtained by channel-wise multiplication of the weights with the corresponding feature maps.
The calculation process is outlined below. Given multi-scale input features { x i R C × H i × W i } i = 1 N , we generate channel descriptors through adaptive average pooling:
z i = F gap ( x i ) R C × 1 × 1
The compressed features are concatenated and processed through lightweight convolutional layers to generate weights:
w = σ F conv z 1 , z 2 , , z N R N × 1 × 1
Finally, the refined features are obtained through channel-wise multiplication:
x ^ i = w i x i
Here, σ denotes the sigmoid function, and  F conv is implemented via lightweight convolutional layers.
This design enhances the model’s attention to critical features and improves the flexibility of the fusion process. By learning the weights, the model can automatically adjust its reliance on different feature maps, leading to better performance in complex and varied detection tasks. The ablation study of the DFRM is presented in Section 4.2.4.

2.4. Detection Head Architecture and Output Components

In TPNet, the head component is composed of two distinct branches: the classification branch and the regression branch, as depicted in Figure 7. The classification branch produces the center map, which estimates the confidence score for each point, thereby determining the score assigned to the predicted bounding box at that location. The regression branch generates three outputs: corner map 1, corner map 2, and the rescoring map. The corner maps provide positional information for the predicted bounding boxes, while the rescoring map refines the center map’s predictions. Unlike CenterNet, TPNet determines the bounding box by predicting the horizontal and vertical distances from the center point to the upper-left and lower-right corners of the predicted box, as illustrated in Figure 2.

2.4.1. Center Map for Classification

The YOLO series employs the Intersection over Union (IoU) metric for label assignment. Specifically, an anchor is assigned a confidence level of 1 if its IoU with the true bounding box exceeds a predefined threshold, typically set at 0.7; otherwise, the confidence level is set to 0. In contrast, TPNet adopts a soft label strategy similar to that used in CenterNet. In this approach, the center point of the ship is assigned a label of 1, while the labels for surrounding points are determined according to Equation (6).
T x y = e x C x 2 2 ( σ w ) 2 + y C y 2 2 ( σ h ) 2
Here, T x y refers to the assigned value at ( x , y ) , where ( x C x ) , ( y C y ) denotes the distance from the center point, and  σ h = 0.5 × h and σ w = 0.5 × w represent the standard deviations in height and width, respectively. Here, h and w denote the height and width of the target ship.
Figure 8 depicts a ship along with its corresponding ground truth heatmap on the center map.
Our classification branch employs the focal loss, as introduced by Zhou et al. [41], which is an enhancement of cross-entropy loss. The primary objective of focal loss is to address the problem of imbalanced sample ratios. In ship detection, the majority of image pixels or objects are classified as background, while only a small fraction corresponds to the ship class. This imbalance between positive and negative samples is thus pronounced. The formula for calculating the loss in the classification branch is given by:
L c l s = 1 N x y 1 P x y α log P x y T x y = 1 1 T x y β P x y α log 1 P x y otherwise
where T x y represents the ground truth and P x y means the output of the center map, α = 2, β = 6, and N denotes the number of ships in the image.

2.4.2. Corner Map1 for Localization

As illustrated in Figure 2, in TPNet, the bounding box of a ship is determined by four distances X l e f t Y t o p X r i g h t Y b o t t o m , as well as the coordinate of the center point, which is determined by the center map. Specifically, if a point ( x 0 , y 0 ) on the center map has a value exceeding a predefined threshold (0.1), it is identified as the center point of a ship. The bounding box for the ship can then be determined using the following formula:
x 1 = D R × ( x 0 l ) = D R × x 0 e c 1 x 0 , y 0 y 1 = D R × ( y 0 t ) = D R × y 0 e c 2 x 0 , y 0 x 2 = D R × ( x 0 + r ) = D R × x 0 + e c 3 x 0 , y 0 y 2 = D R × ( x 0 + b ) = D R × y 0 + e c 4 x 0 , y 0
here DR represents the downsample ratio, and c k ( x 0 , y 0 ) denotes the output of the k-th channel of the corner map1 at the coordinates ( x 0 , y 0 ) .

2.4.3. Corner Map2 for Refining Bounding Box

Bounding box regression is a critical component of object detection, aimed at leveraging center point information for accurate regression. However, due to the effects of downsampling, the projection of the center point onto the feature map is typically non-integer, necessitating the use of information from the nearest integer coordinates rather than the center point itself. This limitation highlights the need for a more effective method to better utilize center point information to enhance the precision of bounding box regression, leading to the development of deformable convolution [55].
Unlike traditional convolution operations, deformable convolutions enhance object detection accuracy by learning deformation information to adapt to variations in object shapes. The core of deformable convolution is the deformable module, which comprises an offset prediction module and a sampling module. The offset prediction module learns the displacement of each pixel, while the sampling module uses these offsets to sample the input feature map. The formula for deformable convolution is as follows:
Y ( p 0 ) = p i R w ( p i ) · X ( p 0 + p i + Δ p i )
In this equation, Y ( p 0 ) represents the output feature map at position p 0 , X denotes the input feature map, w ( p i ) is the weight at position p i in the convolution kernel, and  Δ p i refers to the learned offsets that enable the convolution to adapt to the shape of the object. These offsets are derived through Table 2 in this article.
Given a sampling location ( x 0 , y 0 ) on the feature map, TPNet regresses an initial bounding box from this location using Equation (8). Utilizing the distance vector ( l , t , r , b ) , we can obtain nine sampling points: Center, Left, Right, Top, Bottom, Top-Left, Top-Right, Bottom-Left, Bottom-Right by Table 2 and map these points onto the feature map. The relative offsets of these points to the projection point ( x 0 , y 0 ) serve as the offsets for the deformable convolution. Since these points are manually calculated, there is no additional prediction burden, making this operation computationally more efficient compared to traditional deformable convolutions.
To enhance the accuracy of bounding box regression, the RBH module is introduced, as illustrated in Figure 9. The RBH module operates as a 3 × 3 deformable convolution, with its offsets derived from the output of the corner map1, as specified by Equation (2), distinguishing it from conventional deformable convolutions. The input to the RBH module is identical to that of corner map1, meaning both receive the feature maps produced by the regression branch, which contains extensive localization information. After applying the RBH module, the bounding box can be adjusted using Equation (10):
x 1 = D R × x 0 e c 1 x 0 , y 0 + r 1 x 0 , y 0 y 1 = D R × y 0 e c 2 x 0 , y 0 + r 2 x 0 , y 0 x 2 = D R × x 0 + e c 3 x 0 , y 0 + r 3 x 0 , y 0 y 2 = D R × y 0 + e c 4 x 0 , y 0 + r 4 x 0 , y 0
r k ( x 0 , y 0 ) represents the output of the k-th channel of the RBH module at the coordinates ( x 0 , y 0 ) . The ablation studies presented in Section 4.2.5 will illustrate that the RBH module can effectively enhance the detection accuracy of TPNet.

2.4.4. RSB for Refining Center Map

In TPNet, the center map and corner map are responsible for predicting the confidence scores and bounding boxes, respectively, and they are trained using distinct loss functions without clear mutual awareness. Consequently, while the center map predicts the confidence of each bounding box, it lacks information about the corresponding localization quality. The standard measure for evaluating the localization quality of a bounding box is its Intersection over Union (IoU) with the ground truth box. Therefore, the scores obtained from the center map can easily mismatch with the localization accuracy. For instance, some high-quality predicted boxes may be deleted because of their low scores in the center map, or some predicted boxes with poor localization accuracy may be retained because of their high scores in the center map. This phenomenon can negatively impact the detector’s performance: during Non-Maximum Suppression (NMS), predicted boxes are ranked according to their center map scores, leading to high-scoring boxes suppressing other overlapping boxes. This can lead to the removal of high-quality predictions and the retention of low-quality ones, which is particularly detrimental given TPNet’s goal of accurately predicting target ships.
To tackle the aforementioned problem of localization accuracy in object detection, inspired by the IoU-aware detector [56], a novel approach named Refining Score Branch (RSB) is proposed. RSB is incorporated into the final layer of the regression branch to predict the IoU of each regressed anchor. During the training phase, RSB is jointly trained with the classification and regression branches, the cost function for RSB is Equation (11). During inference, the center map is calibrated by multiplying the rescoring map, which is the output of the RSB. To maintain computational efficiency, RSB consists of a single 1 × 1 convolution layer followed by a sigmoid activation function to ensure the predicted IoU falls within the range [0, 1]. RSB adds minimal computational burden to the entire model while still substantially enhancing its performance.
L R S B = 1 N i = 1 N y i · log ( p i ) + ( 1 y i ) · log ( 1 p i )
here N is the number of positive points, y i is the IoU between the predicted bounding box and the ground truth box at point i, and  p i is the corresponding predicted score in the rescoring map at that point i. The ablation study of the RSB will be presented in Section 4.2.6.

2.5. WGIoU Loss Function

The CenterNet object detection framework utilizes a regression loss based on the Smooth L1 loss; however, Smooth L1 loss has several drawbacks that can limit its effectiveness in certain situations. For instance, it may result in slower convergence and is highly sensitive to hyperparameter tuning. Furthermore, it treats the four distances used for bounding box regression as independent variables, although they are inherently correlated.
To address these limitations, TPNet utilizes the Generalized Intersection over Union (GIoU) to measure the similarity between predicted and ground truth bounding boxes. GIoU considers the overlap between the predicted and ground truth boxes, making it more effective than Smooth L1 loss, particularly in scenarios where the bounding boxes are either very small or very large.
After investigating the regression loss employed by FCOS and CenterNet, we observed that both methods consider only positive samples in their regression loss, i.e.,
l o s s r e g = 1 N pos x , y l r e g ( b o x 1 , b o x 2 )
where the points are positive samples in the regression branch.
However, we are of the belief that “negative” samples in close proximity to the center of the object also play a role in enhancing the accuracy of bounding box predictions. Hence, a weighting parameter is indispensable to mirror the significance of each predicted value. To resolve this problem, we put forward the WGIoU loss. This loss function utilizes the values from the center map as weights, taking into consideration the varying importance of each prediction point. The WGIoU loss takes into account the “centerness” of each sample and modifies the GIoU loss correspondingly to reflect its importance in the localization process. It represents an improvement over the GIoU loss. The  computation process of the wGIoU loss is as in Algorithm 1.
By combining Equations (7) and (11) with Algorithm 1, we can obtain the total loss function of TPNet, which is represented as
l o s s _ t o t a l = l o s s _ c l s + l o s s _ b o x + l o s s _ R S B
The l o s s _ c l s term is responsible for classification, the  l o s s _ b o x term handles bounding box regression, and the l o s s _ R S B term optimizes the RSB. The purpose of this loss function is to train the network to accurately classify and localize ships within the input images.
Algorithm 1 The calculation process of WGIoU loss
Input: Center map: C 1 R H × W × 1 , Corner map1: C 1 R H × W × 4 , Corner map2: C 2 R H × W × 4
Output:  loss _ box
     1:
l o s s 1 = 0 , l o s s 2 = 0 , avg _ factor = 0
     2:
for all every point ( i , j )  do
     3:
    if  C ( i , j ) > 0  then
     4:
        get box1 predicted by corner map1 by Equation (8)
     5:
        get box2 predicted by corner map1 and corner map2 by Equation (10)
     6:
        get the ground truth bounding box from the annotation file denoted as true _ box
     7:
        compute the GIoU between box1 and the true _ box ( g i o u 1 ), as well as the GIoU between box2 and the true _ box ( g i o u 2 )
     8:
         l o s s 1 + = C ( i , j ) × ( 1 g i o u 1 ) , l o s s 2 + = C ( i , j ) × ( 1 g i o u 2 )
     9:
         avg _ factor + = C ( i , j )
    10:
    end if
    11:
end for
    12:
loss _ box = ( l o s s 1 + l o s s 2 ) / avg _ factor
    13:
return  loss _ box

2.6. WSE Attention Module

In complex environmental conditions, accurately detecting ships can be challenging due to the similarity in backscattering characteristics between ships and the background. Attention mechanisms offer a viable solution to this problem by enhancing the feature extraction capabilities of detection algorithms. These mechanisms enable the detector to focus on key features, thereby improving the accuracy and robustness of ship detection. Moreover, attention modules are generally lightweight, allowing for effective ship detection in complex backgrounds without compromising detection speed.
The introduction of attention modules to computer vision was pioneered by SENet [57]. This layer employs a straightforward yet effective mechanism to adaptively recalibrate channel-wise feature responses by explicitly modeling interdependencies between channels. Since its inception, the SE layer has become a cornerstone in modern deep neural network design, achieving state-of-the-art performance across various computer vision tasks, including SAR ship detection [30]. The SE layer has inspired the development of numerous attention models, including the Convolutional Block Attention Module (CBAM) [58], Coord Attention (CA) [59], Efficient Channel Attention (ECA) module [60] and effective Squeeze-and-Excitation (eSE) [61].
Despite these advancements, we found that channel attention mechanisms, such as SE layers, do not incorporate positional information. While CBAM considers positional information, it processes it separately from channel information. On the other hand, self-attention modules like the Non-Local module [62] address both channel and positional information but involve a complex computational process, which does not align with our objective of achieving fast and lightweight ship detection.
To develop a lightweight and efficient attention module, we first examine the channel attention mechanism used in SENet. As detailed in Equations (14) and (15), SENet applies a global average pooling operation on a per-channel basis, followed by two fully connected (FC) layers with non-linear activation functions. A Sigmoid function is then used to generate channel weights. The purpose of these FC layers is to capture non-linear cross-channel interactions while reducing model complexity through dimensionality reduction. However, we identified a key limitation of the SE layer: the use of average pooling does not account for positional information and instead averages across all positions within a channel.
To address this limitation, we propose an enhancement to the original Equation (15), referred to as Equation (16). This enhancement allows for the representation of the importance of each position within a channel. WSE enhances the SE module by adaptively assigning weights to different spatial locations. The modified Equation (16) is integrated with Equation (14) to define the WSE layer, which is an improvement over the traditional SE module. WSE incorporates both spatial and channel-wise attention mechanisms. The computation process for the WSE layer is detailed in Algorithm 2.
Algorithm 2 The calculation process of the WSE layer
Input: Tensor X : R H × W × C
Output: Tensor Y : R H × W × C
Require: a 1x1 convolutional layer conv whose input channels is C and output channels is 1, and two FC layers fc1 and fc2
      1:
w ( i , j ) = softmax ( conv ( X ) )
      2:
scale = i = 1 H j = 1 W w ( i , j ) · x ( i , j , k )
      3:
scale = fc 2 ( fc 1 ( scale ) )
      4:
scale = sigmoid ( scale )
      5:
Y = scale · X
      6:
return  Y
f s = σ FC ReLU FC f g
while f g can be calculated by:
f g = 1 H × W i = 1 H j = 1 W x ( i , j )
in WSE, f g can be rewritten as follows:
f g = i = 1 H j = 1 W w ( i , j ) · x ( i , j )
where w ( i , j ) measures the importance of each point, with the constraint that i = 1 H j = 1 W w ( i , j ) = 1 . Notably, the SE mechanism is a special case of the WSE layer, where w ( i , j ) = 1 H × W .

2.7. The Workflow of TPNet

The entire pipeline for TPNet to process an image is as follows:
  • Step 1: Pre-process the detected image and feed the pre-processed image to MNet. The features generated by MNet are further processed by the neck to obtain a feature map, then this feature map is fed to the Head.
  • Step 2: As shown in Figure 7, the Head generates a center map, a corner map1, a corner map2, and a rescoring map. The final center map is obtained by multiplying the center map and the rescoring map. Points on the center map with values greater than the set threshold (0.1) are identified as positive and recorded as ( x 0 , y 0 ), ( x 1 , y 1 ), etc.
  • Step 3: Using Equation (10) and the corner map1 and corner map2, bounding boxes corresponding to these points are obtained and recorded as ( x 1 0 , y 1 0 , x 2 0 , y 2 0 ) , ( x 1 1 , y 1 1 , x 2 1 , y 2 1 ) , etc.
  • Step 4: The bounding boxes from Corner map2 and their corresponding scores on the refined center map are integrated and processed through NMS to yield the final predicted ships.

3. Experiment Settings

To assess TPNet’s performance, we designed a comprehensive experimental program, comparing it with 14 state-of-the-art CNN-based detection algorithms. Our experimental framework includes dataset construction, evaluation metric design, comparative algorithm selection, and ablation experiments.

3.1. Datasets

We perform experiments on three publicly available datasets, namely SAR-Ship-Dataset [63], SSDD [23] and HRSID [64]; the characteristics of these three datasets are presented in Table 3.
The SAR-Ship Dataset comprises 43,819 images featuring a total of 59,535 ships. Each image has dimensions of 256 × 256 pixels and is cropped from GF-3 and Sentinel-1 satellite images. The dataset includes images with resolutions ranging from 3 m to 22 m and covers various polarization modes such as HH, HV, VH, and VV. It encompasses a broad range of maritime scenes, including ports, inshore areas, islands, and offshore environments. The ships in the dataset include various types, such as oil tankers, bulk carriers, large container ships, and fishing vessels. The dataset is divided into training and testing subsets with an 80:20 ratio, resulting in 35,055 samples for training and 8764 samples for testing. Figure 10 describes the ship size distribution in the SAR-Ship Dataset.
The SSDD is the first publicly available SAR ship detection dataset; it contains a total of 1160 SAR images, which have an average size of 400 × 400 pixels and are obtained from Sentinel-1, TerraSAR-X, and RadarSat-2. The dataset provides SAR images of ships at various resolutions, ranging from 1m to 10m, and from different polarizations, including HH, HV, VV, and VH. Furthermore, the dataset includes a wide variety of ship sizes, ranging from the smallest size of 7 × 7 to the largest size of 211 × 298 pixels, which can be used to evaluate the multi-scale detection performance of different models.
The HRSID dataset consists of 5,604 SAR, each with a resolution of 800 × 800 pixels, acquired from the Sentinel-1 and TerraSAR-X satellites. The dataset provides SAR ship targets with resolutions ranging from 0.1 m to 3 m, and includes three different polarization modes: HH, HV, and VV.
It is noteworthy that unlike other researchers, we did not partition the SSDD and HRSID datasets into separate training and test sets. Instead, we directly used them as the test set to evaluate the performance of the detector trained on the SAR-Ship-Dataset training set. This approach was adopted to demonstrate the superior generalization ability of TPNet.

3.2. Evaluation Metrics

We utilized six evaluation metrics: AP 50 , AP 75 , AP, AP small , AP medium , and AP large —to comprehensively assess the detection accuracy of our detector. In addition, we evaluated the computational complexity of various algorithms using the FLOPs metric and measured the detection speed of the algorithms using FPS (Frames Per Second).

3.3. Experimental Environment, and Implementation Details

The experiments were conducted on a computer equipped with an i9-9900K CPU and Tesla V100 GPU, and PaddlePaddle was utilized for the implementation. To ensure a fair comparison, TPNet and the other 14 advanced SAR ship detectors were implemented using the PaddleDetection 2.6 toolbox [65]. It is worth noting that in our work, we did not utilize any pretrained weights for TPNet, and instead trained it from scratch.
During training, we utilized the AdamW optimizer with a regularizer factor of 0.001. The initial learning rate was set to 0.01 with a warm-up period of 2000 iterations. After the warm-up, a cosine learning rate schedule was employed. All models were trained on the training set of the SAR-Ship-Dataset for 36,000 iterations with a batch size of 64. To improve the training process, we also adopted the exponential moving average (EMA) strategy with a decay rate of 0.9998. To further augment our training data, we employed various techniques such as random flipping, random expansion, and random cropping.

4. Experimental Results

4.1. Comparative Experiments

To evaluate the performance of TPNet, we selected 14 advanced CNN-based detectors for comparison: YOLOv3 [16], YOLOv4 [17], YOLOv5 [18], YOLOv7 [66], YOLOv8 [67], YOLOv10 [68], YOLOv11 [67], PP-YOLO [20], PP-YOLOv2 [21], PP-YOLOE [22], RetinaNet [69], CenterNet [41], FCOS [42], and TTFNet [70]. All detectors, including TPNet, were trained using the same methodology to ensure a fair comparison.

4.1.1. Visualization of Detection Results

Figure 11 presents the detection results of six selected detectors on five SAR images, where blue, green, and red bounding boxes represent correct, missed, and false detections, respectively. The chosen detectors include our proposed TPNet, PP-YOLO-v2 (second-best performing algorithm in terms of AP), YOLOv3 (a classic YOLO algorithm), YOLOv10 and YOLOv11 (state-of-the-art YOLO algorithms), and CenterNet (the foundational algorithm that inspired our work). We selected five typical scenarios: the first column illustrates offshore ship detection in simple backgrounds, while columns two to five depict inshore ship detection in complex scenarios. The complete detection results for all 15 algorithms are provided in Figure A1.
From Figure 11, the following conclusions can be drawn: most detection algorithms can easily identify offshore ships in simple scenes but struggle with inshore ships in complex environments, leading to missed and false detections. Compared to other algorithms, TPNet excels in handling complex-background ship detection. For instance, in the last column, although PP-YOLOv2 match TPNet’s detection accuracy, PP-YOLOv2 underperforms TPNet in the second and third columns. This highlights TPNet’s superior accuracy in complex scenarios.

4.1.2. Comparison of Experimental Results

In the previous subsection, we conducted a qualitative analysis of the detection effects of various algorithms, revealing their performance differences in practical applications. However, visual assessment alone cannot comprehensively and accurately evaluate algorithm superiority. To precisely assess TPNet’s advantages and limitations compared to other advanced algorithms, this section delves into quantitative analysis. We quantitatively compare algorithm detection performance using widely-recognized metrics: AP 50 , AP 75 , AP, AP small , AP medium , and AP large . Moreover, considering computational resource constraints in real-world applications, we include computational complexity (measured by FLOPs) to evaluate the balance between performance and efficiency. Experimental results are shown in Table 4. Through systematic analysis of these quantitative indicators, we aim to present a clear and objective overview of algorithm performance, thereby more strongly supporting our conclusions.
Based on the results presented in Table 4, we can draw the following conclusions:
TPNet is an exceptionally lightweight network, requiring only 7.54% of the computational cost of CenterNet. In terms of speed, TPNet is also the fastest among the 15 algorithms. Despite its low computational cost, TPNet achieves high detection accuracy. Specifically, TPNet outperforms 14 other detection algorithms across five out of six metrics: AP 50 , AP 75 , AP, AP small , and AP middle . The only exception is AP large , where TPNet demonstrates strong performance but does not lead. In terms of AP 50 , TPNet exceeds the second-best PP-YOLOv2 by 1% and the third-best YOLOv3 by 1.1%. For AP 75 , TPNet surpasses PP-YOLOv2 by 1.3% and leads the third-best algorithm YOLO v10 by 3.9%. This performance demonstrates TPNet’s superior localization accuracy in addition to its ship detection capabilities. The AP metric, which measures average detection performance across different IoUs, is widely regarded as a crucial indicator of detector performance. TPNet outperforms PP-YOLOv2 by 1.8% and PP-YOLOE/YOLO v10 by 2.9% in this metric. In the detection of small ships ( AP small ), TPNet leads PP-YOLOv2 by 0.4% and YOLO v10 by 1.8%, demonstrating its effectiveness in detecting smaller targets. Moreover, TPNet excels in the detection of medium-sized ships, with its AP middle exceeding PP-YOLOv2 by 2.8% and YOLOv5 by 2.9%. Although TPNet ranks behind CenterNet in AP large , it still achieves a strong second-place ranking among the 15 algorithms tested, indicating solid performance in detecting large ships.
Many researchers train and test their detection algorithms on the same dataset. While this approach provides a useful benchmark for comparison, we argue that robust object detection algorithms should demonstrate strong generalization ability. An effective algorithm should learn fundamental features from the training dataset, enabling it to perform well on other datasets. To assess this capability, we evaluated the performance of detectors trained on the SAR-Ship-Dataset’s training set using two additional widely used open-source SAR ship detection datasets: HRSID and SSDD. The results of these evaluations are presented in Table 5, Table 6 and Table 7.
Based on the results presented in Table 5, Table 6 and Table 7, we can draw the following conclusions:
When assessing performance via the AP 50 metric, algorithm performances varied across datasets. TTFNet shone on the HRSID dataset but had average results on the SSDD dataset. In contrast, TPNet excelled on the SSDD dataset and performed strongly on HRSID. Overall, for the AP 50 metric considering both datasets, TPNet was the top-performer. Regarding the AP 75 metric, YOLO v10 was outstanding on the HRSID dataset yet underperformed on the SSDD dataset. Conversely, TPNet not only led in performance on the SSDD dataset but also did quite well on HRSID. Across both datasets, TPNet had overall superiority in the AP 75 metric, making it the most effective algorithm. Evaluated by the AP metric, YOLO v11 achieved top performance on the HRSID dataset. However, on the SSDD dataset, its performance paled in comparison to TPNet. In contrast, TPNet consistently outperformed other algorithms on the SSDD dataset and was highly competitive on HRSID. Considering both datasets, TPNet emerged as the leading algorithm in terms of the AP metric, showing superior and more stable performance across different scenarios. For the AP small metric, YOLOv11 had superior performance on the HRSID dataset but underperformed on the SSDD dataset. TPNet, meanwhile, performed strongly on both datasets. Considering both, TPNet was the leading algorithm for the AP small metric. When evaluated using the AP middle metric, TPNet achieved the highest performance on both datasets, firmly establishing itself as the optimal algorithm for this detection metric. With the AP large metric, the PP-YOLOE algorithm outperformed all others on both datasets. TPNet’s performance on this metric was less impressive, suggesting areas for improvement in detecting large ships. Overall, considering performance across both datasets and all six metrics, TPNet demonstrated the best generalization ability among the 15 detectors evaluated.

4.2. Ablation Experiments

To evaluate the effectiveness of the six proposed mechanisms and understand their impact on TPNet’s performance, we conducted a series of ablation experiments, as shown in the following subsections.

4.2.1. Ablation Experiments on Downsample Ratio

Figure 10 shows the size statistics of all ships in the training set of the SAR-Ship dataset. It can be seen that the majority of the ships have length and width smaller than 50, which means they are small or medium-sized targets. To improve the detection performance on these ships, we chose the feature map with a downsampling ratio of 4 for ship detection. Meanwhile, we also conducted a comparative experiment using the feature map with a downsampling ratio of 8, and the experimental results are shown in Table 8.
Table 8 demonstrates that utilizing high-resolution feature maps can significantly enhance detection performance, particularly in the AP 75 metric. This indicates that high-resolution feature maps provide richer positional information, leading to more accurate bounding boxes. Additionally, the results on HRSID and SSDD, as shown in Table 8, indicate that TPNet with a D R value of 4 exhibits superior generalization performance compared to TPNet with a D R value of 8. Therefore, in this study, we selected D R = 4.

4.2.2. Ablation Experiments on MBlock

In order to reduce the computational cost of the model, we designed a lightweight module, MBlock, to replace the traditional convolution module. We conducted ablation experiments to evaluate the performance of MBlock, and the results are shown in Table 9.
Table 9 presents strong evidence that MBlock significantly reduces the model’s computational complexity. Specifically, the FLOPs for TPNet using MBlock are merely 48.5% of those for TPNet using traditional convolution. Moreover, MBlock significantly enhances detection performance compared to traditional convolution, particularly in three key metrics: AP 75 , AP small , and AP large , all of which show marked improvements. Additionally, the results on HRSID and SSDD, as shown in Table 9, indicate that MBlock effectively boosts the model’s generalization capability.

4.2.3. Ablation Experiments on Expansion Ratio

In MBlock, we employed an inverted bottleneck structure to retain more information. The expansion ratio plays a crucial role in determining the model’s detection performance, with a higher E R leading to improved performance at the cost of increased computational complexity. To strike a balance between accuracy and computational cost, we selected an E R value of 3. We conducted experiments with E R values ranging from 1 to 6 to examine the trade-offs between accuracy and computational cost. The results are presented in Table 10.
As shown in Table 10, TPNet’s performance generally improves as the E R value increases from 1 to 4. However, further increases in E R from 4 to 6 result in plateauing or even declines for some metrics. On the SAR-Ship-Dataset, while E R = 3 does not achieve the absolute best performance across all metrics, it achieves a favorable balance between detection accuracy and computational overhead. Specifically, although E R = 3 lags behind E R = 5 in AP75 and E R = 6 in overall AP, it consistently ranks in the upper-middle tier across all six evaluation metrics. This indicates a stable and well-rounded performance. Moreover, the results in Table 10 demonstrate that TPNet achieves optimal generalization at E R = 3, as evidenced by its superior performance on the HRSID and SSDD datasets. This suggests that E R = 3 not only enhances detection precision but also ensures robust adaptability across diverse data distributions and object scales. After careful consideration of these factors, including detection accuracy, computational efficiency, and generalization capability, we conclude that E R = 3 is the most appropriate choice for achieving the best overall performance.

4.2.4. Ablation Study of DFRM Module

To further enhance the feature extraction capability of MFPN from the backbone network, we designed the DFRM module. In this section, we evaluate the effectiveness of the DFRM module, and the experimental results are presented in Table 11.
Table 11 demonstrates the efficacy of the DFRM module. It is evident that the DFRM module introduces only a marginal increase in computational cost while bringing about a significant enhancement in detection performance. This is consistent with our design objective of improving the detector’s overall performance. The module effectively refines detection accuracy and efficiency, making it a valuable addition to the model. Additionally, the results on HRSID and SSDD, as presented in Table 11, indicate that the DFRM module enhances the model’s generalization ability to a notable degree.

4.2.5. Ablation Experiments on RBH

To further improve the localization accuracy of the model, we designed the RBH module for bounding box refinement. In this section, we evaluate the effectiveness of the RBH module, and the experimental results are presented in Table 12.
Table 12 presents the experimental results for our RBH module in bounding box refinement. The RBH module introduces a marginal increase in computational cost but brings about substantial enhancements in detection performance. Specifically, the AP 75 metric shows an improvement of 0.9%, consistent with our design objective of boosting the quality of detection bounding boxes. Moreover, the RBH module also contributes to significant improvements in five other indicators, particularly in the detector’s ability to identify large ships. Additionally, the results on HRSID and SSDD, as shown in Table 12, demonstrate that the RBH module not only elevates the model’s detection accuracy but also enhances its generalization ability.

4.2.6. Ablation Experiments on RSB

During training, the classification and regression branches operate independently. Using the center map as the confidence score fails to account for the quality of the bounding boxes generated by the regression branch. Consequently, some high-quality bounding boxes may be discarded due to low scores in the center map. To address this issue, we proposed RSB. Table 13 demonstrates the effectiveness of RSB in enhancing detection performance, particularly for high-quality bounding boxes.
Table 13 illustrates the effectiveness of the RSB module in enhancing TPNet’s detection performance. Although the RSB module causes a slight increase in computational complexity, it achieves remarkable improvements across all six metrics, with especially notable increases of 1.3% in AP 75 and 1.1% in AP large . This outcome confirms the RSB’s design goal of increasing confidence in high-quality, low-score detection boxes relative to low-quality, high-score ones. Additionally, the results on HRSID and SSDD, as shown in Table 13, further demonstrate that the RSB module can also significantly boost the generalization ability of TPNet.

4.2.7. Ablation Experiments on WGIoU Loss

In both Centernet and FCOS, the regression loss functions are limited to considering only positive samples. However, "negative" points close to the center points should also be accounted for in the regression loss function. To address this limitation, we developed the WGIoU loss and compared its performance against the standard smooth L1 loss and GIoU loss, as detailed in the following results.
Table 14 demonstrates that training the parameters X l e f t , Y t o p , X r i g h t , and Y b o t t o m collectively can significantly enhance the detection performance of the model compared to training them independently. Additionally, our proposed WGIoU loss surpasses the conventional GIoU loss, resulting in substantial improvements in detector performance, particularly in the AP 75 metric, which increases by 2.3%. Furthermore, the results on HRSID and SSDD, as shown in Table 14, indicate that the WGIoU loss also contributes to an improvement in the generalization capability of TPNet.

4.2.8. Ablation Experiments on WSE Layer

An attention mechanism can help the detector focus on important information. However, current mainstream attention modules either only focus on channel information while ignoring position information, such as the SE layer and eSE layer, or handle position and channel information separately, such as CBAM. Based on the SE layer, we propose a WSE layer that considers positional information when generating channel attention maps. We compare the proposed WSE with commonly used attention modules: SE [57], eSE [61], CBAM [58], CA [59], ECA-Net [60], and the results are shown in Table 15 (baseline means we do not use attention modules).
Table 15 offers a thorough assessment of the WSE module’s impact on TPNet’s detection performance. The WSE module consistently outperforms other mechanisms, securing the top position in five out of six evaluation metrics on the SAR-Ship dataset. Furthermore, across both the HRSID and SSDD datasets, the WSE module demonstrates superior effectiveness, consistently outperforming other attention mechanisms. With only a minimal increase in computational cost, the WSE layer significantly boosts TPNet’s performance. Among the six evaluated attention mechanisms, WSE attains the highest levels of accuracy and generalization. This highlights the WSE layer’s pivotal role in optimizing TPNet and validates its selection as the preferred attention mechanism in this study.

5. Conclusions

In this article, we present TPNet, a lightweight, anchor-free detector meticulously designed to improve the accuracy of ship detection in SAR imagery while concurrently reducing computational complexity. TPNet leverages higher-resolution feature maps to enhance detection accuracy and incorporates the novel MBlock module to diminish computational overhead while further improving accuracy. To elevate TPNet’s performance, we also introduce the DFRM, RBH, and RSB modules, an enhanced WGIoU loss function, and the WSE layer, a novel attention mechanism that effectively captures both positional and channel-wise information. Experimental evaluations across three publicly available datasets reveal that these advancements not only enhance TPNet’s performance but also substantially boost its generalization capability. A comparative analysis with 14 state-of-the-art object detection algorithms highlights TPNet’s superior performance in terms of computational efficiency, detection accuracy, and generalization capability. Future research will focus on refining TPNet to improve its accuracy and generalization performance in detecting large vessels. Additionally, future work will take into account the significant research value of arbitrary orientation target detection in remote sensing imagery.

Author Contributions

Writing—original draft, W.Z.; Writing—review & editing, S.F. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The data presented in this study are openly available at: https://aistudio.baidu.com/datasetdetail/100924 [23], accessed on 15 April 2025; https://github.com/CAESAR-Radi/SAR-Ship-Dataset.git, accessed on 15 April 2025; [62], https://github.com/chaozhong2010/HRSID.git, accessed on 15 April 2025 [63].

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

To facilitate readers’ understanding, Figure A1 provides a comprehensive overview of the detection results for all 15 ship detection algorithms evaluated in this study.
Figure A1. Ship Detection Performance Comparison of Different Algorithms on SAR Imagery. Rows: Different Detection Algorithms ((YOLOv3, YOLOv4, YOLOv5,YOLOv7,YOLOv8, YOLOv10, YOLOv11, RetinaNet, CenterNet, PP-YOLO, PP-YOLOv2, PP-YOLO-E, TTFNet, TPNet). Columns: Different Scenes (1-Simple Offshore, 2–5 Complex Inshore). Legend: Blue Boxes = Correct Detections, Green Boxes = Missed Detections, Red Boxes = False Alarm Detections.
Figure A1. Ship Detection Performance Comparison of Different Algorithms on SAR Imagery. Rows: Different Detection Algorithms ((YOLOv3, YOLOv4, YOLOv5,YOLOv7,YOLOv8, YOLOv10, YOLOv11, RetinaNet, CenterNet, PP-YOLO, PP-YOLOv2, PP-YOLO-E, TTFNet, TPNet). Columns: Different Scenes (1-Simple Offshore, 2–5 Complex Inshore). Legend: Blue Boxes = Correct Detections, Green Boxes = Missed Detections, Red Boxes = False Alarm Detections.
Remotesensing 17 01487 g0a1aRemotesensing 17 01487 g0a1b
To facilitate readers’ reading, Table A1 provides a comprehensive list of abbreviations and their corresponding full names used throughout this article. The abbreviations are presented in alphabetical order for clarity.
Table A1. The table shows abbreviations (left) and corresponding full names.
Table A1. The table shows abbreviations (left) and corresponding full names.
AbbreviationFull Name
TPNetThree Points Network
MNetMBlock Network
ERExpansion Ratio
DRDownsample Ratio
DWDepthwise
PWPointwise
BNBatch Normalization
FLOPsFloating Point Operations
MFPNMBlock FPN
DFRMDynamic Feature Refinement Module
RBHRefining Bbox Head
RSBRefining Score Branch
IoUIntersection over Union
NMSNon-Maximum Suppression
GIoUGeneralized Intersection over Union
WGIoUWeighted GIoU
SESqueeze-and-Excitation
WSEWeighted Squeeze-and-Excitation
CBAMConvolutional Block Attention Module
CACoordinate Attention
ECAEfficient Channel Attention
eSEeffective Squeeze-and-Excitation
FCFully-Connected

References

  1. Dalal, N.; Triggs, B. Histograms of Oriented Gradients for Human Detection. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), San Diego, CA, USA, 20–25 June 2005; IEEE: Piscataway, NJ, USA, 2005; Volume 1, pp. 886–893. [Google Scholar]
  2. Ojala, T.; Pietikainen, M.; Maenpaa, T. Multiresolution Gray-Scale and Rotation Invariant Texture Classification with Local Binary Patterns. IEEE Trans. Pattern Anal. Mach. Intell. 2002, 24, 971–987. [Google Scholar] [CrossRef]
  3. Marr, D.; Hildreth, E. Theory of Edge Detection. Proc. R. Soc. Lond. Ser. B. Biol. Sci. 1980, 207, 187–217. [Google Scholar]
  4. Lowe, D.G. Object Recognition from Local Scale-Invariant Features. In Proceedings of the Seventh IEEE International Conference on Computer Vision, Kerkyra, Greece, 20–27 September 1999; IEEE: Piscataway, NJ, USA, 1999; Volume 2, pp. 1150–1157. [Google Scholar]
  5. Gan, L.; Liu, P.; Wang, L. Rotation Sliding Window of the HOG Feature in Remote Sensing Images for Ship Detection. In Proceedings of the 2015 8th International Symposium on Computational Intelligence and Design (ISCID), Hangzhou, China, 12–13 December 2015; IEEE: Piscataway, NJ, USA, 2015; Volume 1, pp. 401–404. [Google Scholar]
  6. Lin, H.; Song, S.; Yang, J. Ship Classification Based on MSHOG Feature and Task-Driven Dictionary Learning with Structured Incoherent Constraints in SAR Images. Remote Sens. 2018, 10, 190. [Google Scholar] [CrossRef]
  7. Wang, Z.; Wang, C.; Wu, F.; Zhang, B.; Zhang, H.; Tang, Y. Ship Detection for Radarsat-2 ScanSAR Data Using DoG Scale-Space. In Proceedings of the 2013 IEEE International Geoscience and Remote Sensing Symposium—IGARSS, Melbourne, VIC, Australia, 21–26 July 2013; pp. 1881–1884. [Google Scholar] [CrossRef]
  8. Guo, J.; Zhu, C.R. A Novel Method of Ship Detection from Spaceborne Optical Image Based on Spatial Pyramid Matching. Appl. Mech. Mater. 2012, 190, 1099–1103. [Google Scholar] [CrossRef]
  9. Leng, X.; Ji, K.; Yang, K.; Zou, H. A Bilateral CFAR Algorithm for Ship Detection in SAR Images. IEEE Geosci. Remote Sens. Lett. 2015, 12, 1536–1540. [Google Scholar] [CrossRef]
  10. Wang, C.; Bi, F.; Zhang, W.; Chen, L. An Intensity-Space Domain CFAR Method for Ship Detection in HR SAR Images. IEEE Geosci. Remote Sens. Lett. 2017, 14, 529–533. [Google Scholar] [CrossRef]
  11. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar]
  12. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef]
  13. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask R-CNN. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2961–2969. [Google Scholar]
  14. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  15. Redmon, J.; Farhadi, A. YOLO9000: Better, Faster, Stronger. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7263–7271. [Google Scholar]
  16. Redmon, J.; Farhadi, A. YOLOv3: An Incremental Improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  17. Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y.M. YOLOv4: Optimal Speed and Accuracy of Object Detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  18. Ultralytics. YOLOv5. Available online: https://github.com/ultralytics/yolov5 (accessed on 1 February 2023).
  19. Li, C.; Li, L.; Jiang, H.; Weng, K.; Geng, Y.; Li, L.; Ke, Z.; Li, Q.; Cheng, M.; Nie, W.; et al. YOLOv6: A Single-Stage Object Detection Framework for Industrial Applications. arXiv 2022, arXiv:2209.02976. [Google Scholar]
  20. Long, X.; Deng, K.; Wang, G.; Zhang, Y.; Dang, Q.; Gao, Y.; Shen, H.; Ren, J.; Han, S.; Ding, E.; et al. PP-YOLO: An Effective and Efficient Implementation of Object Detector. arXiv 2020, arXiv:2007.12099. [Google Scholar]
  21. Huang, X.; Wang, X.; Lv, W.; Bai, X.; Long, X.; Deng, K.; Dang, Q.; Han, S.; Liu, Q.; Hu, X.; et al. PP-YOLOv2: A Practical Object Detector. arXiv 2021, arXiv:2104.10419. [Google Scholar]
  22. Xu, S.; Wang, X.; Lv, W.; Chang, Q.; Cui, C.; Deng, K.; Wang, G.; Dang, Q.; Wei, S.; Du, Y.; et al. PP-YOLOE: An Evolved Version of YOLO. arXiv 2022, arXiv:2203.16250. [Google Scholar]
  23. Li, J.; Qu, C.; Shao, J. Ship Detection in SAR Images Based on an Improved Faster R-CNN. In Proceedings of the 2017 SAR in Big Data Era: Models, Methods and Applications (BIGSARDATA), Beijing, China, 13–14 November 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 1–6. [Google Scholar]
  24. Chang, Y.L.; Anagaw, A.; Chang, L.; Wang, Y.C.; Hsiao, C.Y.; Lee, W.H. Ship Detection Based on YOLOv2 for SAR Imagery. Remote Sens. 2019, 11, 786. [Google Scholar] [CrossRef]
  25. Zhang, T.; Zhang, X. High-Speed Ship Detection in SAR Images Based on a Grid Convolutional Neural Network. Remote Sens. 2019, 11, 1206. [Google Scholar] [CrossRef]
  26. Zhang, T.; Zhang, X.; Shi, J.; Wei, S. Depthwise Separable Convolution Neural Network for High-Speed SAR Ship Detection. Remote Sens. 2019, 11, 2483. [Google Scholar] [CrossRef]
  27. Gui, Y.; Li, X.; Xue, L.; Lv, J. A Scale Transfer Convolution Network for Small Ship Detection in SAR Images. In Proceedings of the 2019 IEEE 8th Joint International Information Technology and Artificial Intelligence Conference (ITAIC), Chongqing, China, 24–26 May 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 1845–1849. [Google Scholar]
  28. Jiao, J.; Zhang, Y.; Sun, H.; Yang, X.; Gao, X.; Hong, W.; Fu, K.; Sun, X. A Densely Connected End-to-End Neural Network for Multiscale and Multiscene SAR Ship Detection. IEEE Access 2018, 6, 20881–20892. [Google Scholar] [CrossRef]
  29. Li, Y.; Ding, Z.; Zhang, C.; Wang, Y.; Chen, J. SAR Ship Detection Based on ResNet and Transfer Learning. In Proceedings of the IGARSS 2019-2019 IEEE International Geoscience and Remote Sensing Symposium, Yokohama, Japan, 28 July–2 August 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 1188–1191. [Google Scholar]
  30. Lin, Z.; Ji, K.; Leng, X.; Kuang, G. Squeeze and Excitation Rank Faster R-CNN for Ship Detection in SAR Images. IEEE Geosci. Remote Sens. Lett. 2018, 16, 751–755. [Google Scholar] [CrossRef]
  31. Zhao, Y.; Zhao, L.; Xiong, B.; Kuang, G. Attention Receptive Pyramid Network for Ship Detection in SAR Images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 2738–2756. [Google Scholar] [CrossRef]
  32. Bao, W.; Huang, M.; Zhang, Y.; Xu, Y.; Liu, X.; Xiang, X. Boosting Ship Detection in SAR Images with Complementary Pretraining Techniques. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 8941–8954. [Google Scholar] [CrossRef]
  33. Zhang, T.; Zhang, X.; Shi, J.; Wei, S. HyperLi-Net: A Hyper-Light Deep Learning Network for High-Accurate and High-Speed Ship Detection from Synthetic Aperture Radar Imagery. ISPRS J. Photogramm. Remote Sens. 2020, 167, 123–153. [Google Scholar] [CrossRef]
  34. Zhang, T.; Zhang, X.; Ke, X. Quad-FPN: A Novel Quad Feature Pyramid Network for SAR Ship Detection. Remote Sens. 2021, 13, 2771. [Google Scholar] [CrossRef]
  35. Xu, X.; Zhang, X.; Zhang, T. Lite-YOLOv5: A Lightweight Deep Learning Detector for On-Board Ship Detection in Large-Scene Sentinel-1 SAR Images. Remote Sens. 2022, 14, 1018. [Google Scholar] [CrossRef]
  36. Gao, Y.; Wu, Z.; Ren, M.; Wu, C. Improved YOLOv4 Based on Attention Mechanism for Ship Detection in SAR Images. IEEE Access 2022, 10, 23785–23797. [Google Scholar] [CrossRef]
  37. Jiang, J.; Fu, X.; Qin, R.; Wang, X.; Ma, Z. High-Speed Lightweight Ship Detection Algorithm Based on YOLO-v4 for Three-Channels RGB SAR Image. Remote Sens. 2021, 13, 1909. [Google Scholar] [CrossRef]
  38. Guo, Y.; Chen, S.; Zhan, R.; Wang, W.; Zhang, J. SAR Ship Detection Based on YOLOv5 Using CBAM and BiFPN. In Proceedings of the IGARSS 2022—2022 IEEE International Geoscience and Remote Sensing Symposium, Kuala Lumpur, Malaysia, 17–22 July 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 2147–2150. [Google Scholar]
  39. Tang, X.; Zhang, J.; Xia, Y.; Xiao, H. DBW-YOLO: A High-Precision SAR Ship Detection Method for Complex Environments. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2024, 17, 7029–7039. [Google Scholar] [CrossRef]
  40. Liu, L.; Fu, L.; Zhang, Y.; Ni, W.; Wu, B.; Li, Y.; Shang, C.; Shen, Q. CLFR-Det: Cross-Level Feature Refinement Detector for Tiny-Ship Detection in SAR Images. Knowl.-Based Syst. 2024, 284, 111284. [Google Scholar] [CrossRef]
  41. Zhou, X.; Wang, D.; Krähenbühl, P. Objects as Points. arXiv 2019, arXiv:1904.07850. [Google Scholar]
  42. Tian, Z.; Shen, C.; Chen, H.; He, T. FCOS: Fully Convolutional One-Stage Object Detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 9627–9636. [Google Scholar]
  43. Gao, F.; He, Y.; Wang, J.; Hussain, A.; Zhou, H. Anchor-Free Convolutional Network with Dense Attention Feature Aggregation for Ship Detection in SAR Images. Remote Sens. 2020, 12, 2619. [Google Scholar] [CrossRef]
  44. Feng, Y.; Chen, J.; Huang, Z.; Wan, H.; Xia, R.; Wu, B.; Sun, L.; Xing, M. A Lightweight Position-Enhanced Anchor-Free Algorithm for SAR Ship Detection. Remote Sens. 2022, 14, 1908. [Google Scholar] [CrossRef]
  45. Yao, C.; Xie, P.; Zhang, L.; Fang, Y. ATSD: Anchor-Free Two-Stage Ship Detection Based on Feature Enhancement in SAR Images. Remote Sens. 2022, 14, 6058. [Google Scholar] [CrossRef]
  46. He, B.; Zhang, Q.; Tong, M.; He, C. An Anchor-Free Method Based on Adaptive Feature Encoding and Gaussian-Guided Sampling Optimization for Ship Detection in SAR Imagery. Remote Sens. 2022, 14, 1738. [Google Scholar] [CrossRef]
  47. Zhu, M.; Hu, G.; Zhou, H.; Wang, S.; Feng, Z.; Yue, S. A Ship Detection Method via Redesigned FCOS in Large-Scale SAR Images. Remote Sens. 2022, 14, 1153. [Google Scholar] [CrossRef]
  48. Sun, Z.; Dai, M.; Leng, X.; Lei, Y.; Xiong, B.; Ji, K.; Kuang, G. An Anchor-Free Detection Method for Ship Targets in High-Resolution SAR Images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 7799–7816. [Google Scholar] [CrossRef]
  49. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  50. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature Pyramid Networks for Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  51. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin Transformer: Hierarchical Vision Transformer Using Shifted Windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 10012–10022. [Google Scholar]
  52. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications. arXiv 2017, arXiv:1704.04861. [Google Scholar]
  53. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.C. MobileNetV2: Inverted Residuals and Linear Bottlenecks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 4510–4520. [Google Scholar]
  54. Ding, X.; Zhang, X.; Ma, N.; Han, J.; Ding, G.; Sun, J. RepVGG: Making VGG-Style ConvNets Great Again. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 13733–13742. [Google Scholar]
  55. Dai, J.; Qi, H.; Xiong, Y.; Li, Y.; Zhang, G.; Hu, H.; Wei, Y. Deformable Convolutional Networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 764–773. [Google Scholar]
  56. Wu, S.; Li, X.; Wang, X. IoU-aware single-stage object detector for accurate localization. Image Vis. Comput. 2020, 97, 103911. [Google Scholar] [CrossRef]
  57. Hu, J.; Shen, L.; Sun, G. Squeeze-and-Excitation Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 7132–7141. [Google Scholar]
  58. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. CBAM: Convolutional Block Attention Module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  59. Hou, Q.; Zhou, D.; Feng, J. Coordinate Attention for Efficient Mobile Network Design. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 13713–13722. [Google Scholar]
  60. Wang, Q.; Wu, B.; Zhu, P.; Li, P.; Zuo, W.; Hu, Q. ECA-Net: Efficient Channel Attention for Deep Convolutional Neural Networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 11534–11542. [Google Scholar]
  61. Lee, Y.; Park, J. Centermask: Real-Time Anchor-Free Instance Segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 13906–13915. [Google Scholar]
  62. Wang, X.; Girshick, R.; Gupta, A.; He, K. Non-Local Neural Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 7794–7803. [Google Scholar]
  63. Wang, Y.; Wang, C.; Zhang, H.; Dong, Y.; Wei, S. A SAR Dataset of Ship Detection for Deep Learning under Complex Backgrounds. Remote Sens. 2019, 11, 765. [Google Scholar] [CrossRef]
  64. Wei, S.; Zeng, X.; Qu, Q.; Wang, M.; Su, H.; Shi, J. HRSID: A High-Resolution SAR Images Dataset for Ship Detection and Instance Segmentation. IEEE Access 2020, 8, 120234–120254. [Google Scholar] [CrossRef]
  65. PaddlePaddle. PaddleDetection. Available online: https://github.com/PaddlePaddle/PaddleDetection.git (accessed on 1 February 2023).
  66. Wang, C.Y.; Bochkovskiy, A.; Liao, H.Y.M. YOLOv7: Trainable Bag-of-Freebies Sets New State-of-the-Art for Real-Time Object Detectors. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023. [Google Scholar]
  67. Jocher, G.; Qiu, J. Ultralytics YOLO11. 2024. Version 11.0.0, Licensed Under AGPL-3.0. Available online: https://github.com/ultralytics/ultralytics (accessed on 1 February 2023).
  68. Wang, A.; Chen, H.; Liu, L.; Chen, K.; Lin, Z.; Han, J.; Ding, G. YOLOv10: Real-Time End-to-End Object Detection. arXiv 2024, arXiv:2405.14458. [Google Scholar]
  69. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal Loss for Dense Object Detection. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
  70. Liu, Z.; Zheng, T.; Xu, G.; Yang, Z.; Liu, H.; Cai, D. Training-Time-Friendly Network for Real-Time Object Detection. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 11685–11692. [Google Scholar]
Figure 1. The basic structure of TPNet.
Figure 1. The basic structure of TPNet.
Remotesensing 17 01487 g001
Figure 2. The triple points of a target ship.
Figure 2. The triple points of a target ship.
Remotesensing 17 01487 g002
Figure 3. Schematic diagram of MBlock.
Figure 3. Schematic diagram of MBlock.
Remotesensing 17 01487 g003
Figure 4. (a) Two modes of Conv2 layer; (b), structural re-parameterization of Conv2, the letter ’b’ represents the bias term in the convolutional layer.
Figure 4. (a) Two modes of Conv2 layer; (b), structural re-parameterization of Conv2, the letter ’b’ represents the bias term in the convolutional layer.
Remotesensing 17 01487 g004
Figure 5. The architecture of the neck of TPNet.
Figure 5. The architecture of the neck of TPNet.
Remotesensing 17 01487 g005
Figure 6. DFRM schematic diagram. In this diagram, © represents the concat operation, FC represents the fully connected layer, Ⓢ represents the split operation, ⊙ represents the broadcast element product, and σ represents the Sigmoid activation function.
Figure 6. DFRM schematic diagram. In this diagram, © represents the concat operation, FC represents the fully connected layer, Ⓢ represents the split operation, ⊙ represents the broadcast element product, and σ represents the Sigmoid activation function.
Remotesensing 17 01487 g006
Figure 7. The structure of the head of TPNet.
Figure 7. The structure of the head of TPNet.
Remotesensing 17 01487 g007
Figure 8. A target ship and its corresponding heatmap in the center map (the table parameters are derived from the true labels downsampled at a rate of 4, as calculated using Equation (6). The heatmap is generated based on these table parameters).
Figure 8. A target ship and its corresponding heatmap in the center map (the table parameters are derived from the true labels downsampled at a rate of 4, as calculated using Equation (6). The heatmap is generated based on these table parameters).
Remotesensing 17 01487 g008
Figure 9. The schematic diagram of RBH module.
Figure 9. The schematic diagram of RBH module.
Remotesensing 17 01487 g009
Figure 10. Size statistics of ships in the SAR-Ship dataset. (a) The width and height distribution of target ships in the dataset; (b) The height distribution of target ships in the dataset; (c) The width distribution of target ships in the dataset.
Figure 10. Size statistics of ships in the SAR-Ship dataset. (a) The width and height distribution of target ships in the dataset; (b) The height distribution of target ships in the dataset; (c) The width distribution of target ships in the dataset.
Remotesensing 17 01487 g010
Figure 11. Ship Detection Performance Comparison of Different Algorithms on SAR Imagery. Rows: Different Detection Algorithms (YOLOv3, YOLOv10, YOLOv11, CenterNet, PP-YOLOv2, TPNet). Columns: Different Scenes (1-Simple Offshore, 2–5 Complex Inshore). Legend: Blue Boxes = Correct Detections, Green Boxes = Missed Detections, Red Boxes = False Alarm Detections.
Figure 11. Ship Detection Performance Comparison of Different Algorithms on SAR Imagery. Rows: Different Detection Algorithms (YOLOv3, YOLOv10, YOLOv11, CenterNet, PP-YOLOv2, TPNet). Columns: Different Scenes (1-Simple Offshore, 2–5 Complex Inshore). Legend: Blue Boxes = Correct Detections, Green Boxes = Missed Detections, Red Boxes = False Alarm Detections.
Remotesensing 17 01487 g011
Table 1. The details of MNet (attn means attention mechanism).
Table 1. The details of MNet (attn means attention mechanism).
OperatorKernel SizeStrideattnInput SizeOutput Size
MBlock72False256 × 256 × 3128 × 128 × 24
Maxpool32False128 × 128 × 2464 × 64 × 24
MBlock52False64 × 64 × 2432 × 32 × 56
MBlock51True32 × 32 × 5632 × 32 × 56
MBlock52False32 × 32 × 5616 × 16 × 120
MBlock51True16 × 16 × 12016 × 16 × 120
MBlock52False16 × 16 × 1208 × 8 × 272
MBlock51True8 × 8 × 2728 × 8 × 272
Table 2. Offsets employed in RBH.
Table 2. Offsets employed in RBH.
PositionCoordinates
Center ( 0 , 0 )
Left ( l , 0 )
Right ( r , 0 )
Top ( 0 , t )
Bottom ( 0 , b )
Top-Left ( l , t )
Top-Right ( r , t )
Bottom-Left ( l , b )
Bottom-Right ( r , b )
l , t , r , b are concluded from Equation (8).
Table 3. Basic information of SAR-ship-dataset, SSDD and HRSID.
Table 3. Basic information of SAR-ship-dataset, SSDD and HRSID.
DatasetNum of ImagesNum of ShipsSatellitesResolution (m)
SAR-ship-dataset43,81959,535GF3 Sentinel-15–20
SSDD560416,951TerraSAR-X Sentinel-1 RadarSat-21–10
HRSID11602540TerraSAR-X Sentinel-1B TanDEM0.5–3
Table 4. Test results of different algorithms on SAR-Ship-Dataset.
Table 4. Test results of different algorithms on SAR-Ship-Dataset.
AlgorithmAP50 (%)AP75 (%)AP (%)APsmall (%)APmiddle (%)APlarge (%)FLOPs (G)FPS
YOLOv3 [16]94.660.755.851.961.455.912.377302.2
YOLOv4 [17]91.653.451.646.958.054.39.766300.5
YOLOv5 [18]94.166.158.254.066.453.08.637298.7
YOLOv7 [66]84.348.247.242.554.225.68.391293.2
YOLOv8 [67]93.264.257.753.264.059.615.378305.7
YOLOv10 [68]93.767.559.254.665.963.89.632233.7
YOLOv11 [67]93.465.558.153.664.559.114.664297.1
PP-YOLO [20]92.662.355.951.262.563.69.153300.0
PP-YOLOv2 [21]94.770.160.356.066.563.79.153303.7
PP-YOLOE [22]93.866.859.254.566.065.08.879298.9
RetinaNet [69]80.234.139.531.451.643.812.988103.6
CenterNet [41]91.855.152.846.561.473.66.43338.7
TTFNet [70]92.764.557.651.965.269.411.931119.4
FCOS [42]90.056.952.944.963.752.612.865298.0
TPNet95.771.462.156.469.373.30.485513316.9
Table 5. Test results of different algorithms on HRSID.
Table 5. Test results of different algorithms on HRSID.
AlgorithmAP50 (%)AP75 (%)AP (%)APsmall (%)APmiddle (%)APlarge (%)
YOLOv3 [16]69.141.940.143.424.71.1
YOLOv4 [17]71.141.740.243.428.11.1
YOLOv5 [18]65.940.538.642.419.50.0
YOLOv7 [66]64.141.937.941.122.00.0
YOLOv8 [67]71.247.043.147.225.40.1
YOLOv10 [68]71.549.444.348.028.70.8
YOLOv11 [67]72.449.344.548.528.60.3
PP-YOLO [20]59.738.535.438.720.80.2
PP-YOLOv2 [21]70.745.542.245.826.10.4
PP-YOLOE [22]69.647.943.246.334.82.5
RetinaNet [69]59.425.329.131.724.00.2
CenterNet [41]70.444.941.944.929.40.9
TTFNet [70]73.246.343.346.827.10.1
FCOS [42]55.815.924.331.014.80.1
TPNet71.047.743.346.635.50.3
Table 6. Test results of different algorithms on SSDD.
Table 6. Test results of different algorithms on SSDD.
AlgorithmAP50 (%)AP75 (%)AP (%)APsmall (%)APmiddle (%)APlarge (%)
YOLOv3 [16]77.728.937.235.940.726.6
YOLOv4 [17]78.929.537.436.840.320.3
YOLOv5 [18]45.910.618.618.819.63.5
YOLOv7 [66]72.533.136.633.137.538.2
YOLOv8 [67]78.430.638.138.040.512.2
YOLOv10 [68]82.534.340.339.942.921.1
YOLOv11 [67]78.332.438.538.340.916.3
PP-YOLO [20]82.936.841.640.545.023.2
PP-YOLOv2 [21]79.634.739.738.043.529.2
PP-YOLOE [22]71.930.835.629.545.838.8
RetinaNet [69]74.423.833.633.836.220.9
CenterNet [41]75.531.037.135.541.716.8
TTFNet [70]74.625.334.134.735.011.5
FCOS [42]69.217.328.729.430.719.4
TPNet86.144.245.944.150.821.6
Table 7. Combined test results of different algorithms on HRSID and SSDD.
Table 7. Combined test results of different algorithms on HRSID and SSDD.
AlgorithmAP50 (%)AP75 (%)AP (%)APsmall (%)APmiddle (%)APlarge (%)
YOLOv3 [16]146.870.877.379.365.427.7
YOLOv4 [17]150.071.277.680.268.421.4
YOLOv5 [18]111.851.157.261.239.13.5
YOLOv7 [66]136.675.074.574.259.538.2
YOLOv8 [67]149.677.681.285.265.912.3
YOLOv10 [68]154.083.784.687.971.621.9
YOLOv11 [67]150.781.783.086.869.516.6
PP-YOLO [20]142.675.377.079.265.823.4
PP-YOLOv2 [21]150.380.281.983.869.629.6
PP-YOLOE [22]141.578.778.875.880.641.3
RetinaNet [69]133.849.162.765.560.221.1
CenterNet [41]145.975.979.080.471.117.7
TTFNet [70]147.871.677.481.562.111.6
FCOS [42]125.033.253.060.472.519.5
TPNet157.191.989.290.786.321.9
Table 8. Test results on different datasets under different D R values.
Table 8. Test results on different datasets under different D R values.
DatasetDRAP50 (%)AP75 (%)AP (%)APsmall (%)APmiddle (%)APlarge (%)FLOPs (G)
SAR-Ship
Dataset
495.771.462.156.469.373.30.485513
894.967.659.554.067.861.10.378620
HRSID471.047.743.346.635.50.3-
863.732.533.737.625.20.0-
SSDD486.144.245.944.150.821.6-
884.434.841.440.444.622.7-
HRSID
and SSDD
4157.191.989.290.786.321.9-
8148.167.375.178.069.822.7-
Table 9. Test results on different datasets under different operators.
Table 9. Test results on different datasets under different operators.
DatasetOperatorAP50 (%)AP75 (%)AP (%)APsmall (%)APmiddle (%)APlarge (%)FLOPs (G)
SAR-Ship
Dataset
Traditional Convolution95.270.061.054.968.763.91.000
MBlock95.771.462.156.469.373.30.485513
HRSIDTraditional Convolution65.039.236.940.928.20.00
MBlock71.047.743.346.635.50.30
SSDDTraditional Convolution81.630.438.639.339.317.00
MBlock86.144.245.944.150.821.60
HRSID
and
SSDD  
Traditional Convolution146.669.675.580.267.517.00
MBlock157.191.989.290.786.321.90
Table 10. Test results on different datasets under different E R values.
Table 10. Test results on different datasets under different E R values.
DatasetERAP50 (%)AP75 (%)AP (%)APsmall (%)APmiddle (%)APlarge (%)FLOPs (G)
SAR-Ship
Dataset
195.168.860.655.167.968.90.234794
295.370.361.756.169.172.30.360131
395.771.462.156.469.373.30.485513
495.772.662.556.769.870.40.610942
595.672.162.256.469.568.40.736417
695.872.562.756.970.072.70.861937
HRSID171.346.743.245.640.10.20
271.046.442.946.035.80.10
371.047.743.346.635.50.30
470.450.244.647.538.60.00
571.850.643.347.840.80.70
671.649.244.647.737.00.50
SSDD184.939.243.441.647.919.80
283.739.943.141.247.623.50
386.144.245.944.150.821.60
485.839.843.943.646.420.80
584.240.944.040.850.225.70
683.138.642.841.147.520.50
HRSID
and
SSDD
1156.285.986.687.288.020.00
2154.786.386.087.283.423.60
3157.191.989.290.786.321.90
4156.290.088.591.185.020.80
5156.091.587.388.691.026.40
6154.787.887.488.884.521.00
Table 11. Effectiveness of DFRM on different datasets.
Table 11. Effectiveness of DFRM on different datasets.
DatasetWhether to Use DFRMAP50 (%)AP75 (%)AP (%)APsmall (%)APmiddle (%)APlarge (%)FLOPs (G)
SAR-Ship
Dataset
No95.369.961.355.668.771.90.485452
Yes95.771.462.156.469.373.30.485513
HRSIDNo71.649.444.948.036.70.2-
Yes71.047.743.346.635.50.3-
SSDDNo84.540.243.542.447.122.0-
Yes86.144.245.944.150.821.6-
HRSID
and SSDD
No156.189.688.490.483.822.2-
Yes157.191.989.290.786.321.9-
Table 12. Effectiveness of RBH on different datasets.
Table 12. Effectiveness of RBH on different datasets.
DatasetWith RBHAP50 (%)AP75 (%)AP (%)APsmall (%)APmiddle (%)APlarge (%)FLOPs (G)
SAR-Ship
Dataset
NO95.670.561.455.768.766.90.474843
YES95.771.462.156.469.373.30.485513
HRSIDNO68.445.941.044.631.10.3-
YES71.047.743.346.635.50.3-
SSDDNO84.340.344.142.648.123.3-
YES86.144.245.944.150.821.6-
HRSID
and SSDD
NO152.786.285.187.279.223.6-
YES157.191.989.290.786.321.9-
Table 13. Effectiveness of RSB on different datasets.
Table 13. Effectiveness of RSB on different datasets.
DatasetWith RSBAP50 (%)AP75 (%)AP (%)APsmall (%)APmiddle (%)APlarge (%)FLOPs (G)
SAR-Ship
Dataset
NO95.470.161.555.568.572.20.485378
YES95.771.462.156.469.373.30.485513
HRSIDNO69.445.641.644.933.20.0-
YES71.047.743.346.635.50.3-
SSDDNO84.340.344.142.648.122.3-
YES86.144.245.944.150.821.6-
HRSID
and SSDD
NO153.785.985.787.581.322.3-
YES157.191.989.290.786.321.9-
Table 14. Test results of TPNet with different loss functions on various datasets.
Table 14. Test results of TPNet with different loss functions on various datasets.
DatasetLoss FunctionAP50 (%)AP75 (%)AP (%)APsmall (%)APmiddle (%)APlarge (%)
SAR-Ship
Dataset
Smooth L192.959.555.250.262.459.4
GIoU95.369.160.555.167.471.5
WGIoU95.771.462.156.469.373.3
HRSIDSmooth L166.141.438.842.829.40.0
GIoU65.739.937.642.627.10.0
WGIoU71.047.743.346.635.50.3
SSDDSmooth L181.633.340.341.042.614.6
GIoU83.734.641.242.142.720.4
WGIoU86.144.245.944.150.821.6
HRSID
and
SSDD
Smooth L1147.774.779.183.872.014.6
GIoU149.474.578.884.769.820.4
WGIoU157.191.989.290.786.321.9
Table 15. Test results of TPNet with different attention modules on various datasets.
Table 15. Test results of TPNet with different attention modules on various datasets.
DatasetAttention ModuleAP50 (%)AP75 (%)AP (%)APsmall (%)APmiddle (%)APlarge (%)FLOPs (G)
SAR-Ship
Dataset
Baseline95.470.561.455.368.361.70.483484
SE95.571.061.956.368.965.80.483787
eSE95.371.361.755.969.167.10.484338
CBAM95.070.961.856.169.064.50.485767
CA95.671.361.956.269.769.50.517191
ECA95.369.961.055.368.562.90.485767
WSE95.771.462.156.469.373.30.485513
HRSIDBaseline68.745.441.945.231.60.0-
SE66.344.540.643.631.50.0-
eSE68.243.540.343.830.70.0-
CBAM67.244.140.844.330.10.0-
CA68.344.140.944.529.40.0-
ECA67.643.239.943.930.20.0-
WSE71.047.743.346.635.50.3-
SSDDBaseline81.736.240.941.941.620.4-
SE82.936.041.142.641.317.0-
eSE80.732.539.640.940.415.4-
CBAM84.236.641.942.044.018.7-
CA83.334.741.341.343.415.8-
ECA81.930.839.340.341.210.1-
WSE86.144.245.944.150.821.6-
HRSID
and
SSDD
Baseline150.481.682.887.173.220.4-
SE149.280.581.786.272.817.0-
eSE148.976.079.984.771.115.4-
CBAM151.480.782.786.374.118.7-
CA151.678.882.285.872.815.8-
ECA149.574.079.284.271.410.1-
WSE157.191.989.290.786.321.9-
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zuo, W.; Fang, S. TPNet: A High-Performance and Lightweight Detector for Ship Detection in SAR Imagery. Remote Sens. 2025, 17, 1487. https://doi.org/10.3390/rs17091487

AMA Style

Zuo W, Fang S. TPNet: A High-Performance and Lightweight Detector for Ship Detection in SAR Imagery. Remote Sensing. 2025; 17(9):1487. https://doi.org/10.3390/rs17091487

Chicago/Turabian Style

Zuo, Weikang, and Shenghui Fang. 2025. "TPNet: A High-Performance and Lightweight Detector for Ship Detection in SAR Imagery" Remote Sensing 17, no. 9: 1487. https://doi.org/10.3390/rs17091487

APA Style

Zuo, W., & Fang, S. (2025). TPNet: A High-Performance and Lightweight Detector for Ship Detection in SAR Imagery. Remote Sensing, 17(9), 1487. https://doi.org/10.3390/rs17091487

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop