Next Article in Journal
STAR-RIS-Enabled AOA Positioning Algorithm
Previous Article in Journal
Centralized Multi-Sensor GLMB Smoother for Multi-Target Tracking
Previous Article in Special Issue
Clothing-Agnostic Pre-Inpainting Virtual Try-On
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Ship Target Detection in SAR Imagery Based on Band Recombination and Multi-Scale Feature Enhancement

Institute of Geospatial Information, Information Engineering University, Zhengzhou 450001, China
*
Author to whom correspondence should be addressed.
Electronics 2025, 14(23), 4728; https://doi.org/10.3390/electronics14234728
Submission received: 2 November 2025 / Revised: 24 November 2025 / Accepted: 25 November 2025 / Published: 30 November 2025

Abstract

Synthetic aperture radar images have all-weather and all-time capabilities and are widely used in the field of ship target surveillance at sea. However, its detection accuracy is often limited by factors such as complex sea conditions, diverse ship scales, and image noise. Aiming at the problems such as inconsistent scale of ship target detection in SAR images, difficulty in detecting small targets, and interference from complex backgrounds, this paper proposes a ship detection method for SAR images based on band recombination and multi-scale feature enhancement. Firstly, aiming at the problem that the single-channel replication mode adopted by the deep neural network cannot fully extract the ship target information in SAR images, a band recombination method was designed to enhance the ship information in the images. Furthermore, the coordinate channel attention and bottleneck Transformer attention mechanisms are introduced in the backbone part of the network to enhance the network’s representation ability of the target spatial distribution and maintain the global feature modeling ability. Finally, a multi-scale feature enhancement and multi-scale effective feature aggregation module was designed to improve the detection accuracy of multi-scale ships in wide-format images. The experimental results on the LS-SSDD and HRSID datasets show that the average accuracies of the method proposed in this paper reach 78.1% and 94.5% respectively, which are improved by 6.9% and 0.8% compared with the baseline model, and are superior to other advanced algorithms, verifying the effectiveness of the method proposed in this paper. Meanwhile, the algorithm proposed in this paper has also demonstrated good performance in wide-format SAR images of actual large scenes. The method proposed in this paper effectively improves the problems of missed detection and false detection of small-target ships in SAR images of large scenes. At the same time, it enhances the efficiency of rapid and accurate detection in large scenes and can provide reliable technical support for the field of maritime target surveillance.

1. Introduction

Synthetic Aperture Radar (SAR) has become increasingly vital in the field of intelligent interpretation of remote sensing images due to its all-weather and day-and-night monitoring capabilities [1,2]. Its sensitivity to metallic objects makes ships easily distinguishable from the background in SAR images. Furthermore, the ability to scan large swathes—often hundreds of kilometers wide—provides significant advantages for efficient wide-area maritime monitoring and dynamic tracking. Consequently, ship target detection based on SAR images is widely applied in maritime traffic management, maritime target surveillance, and disaster emergency response [3]. Traditional SAR ship detection methods often rely on statistical models, such as Constant False Alarm Rate (CFAR) algorithms. These methods determine a threshold to distinguish ship targets from the background in SAR images by performing statistical analysis and modeling of the background clutter [4]. To address the characteristics of different SAR images and meet diverse application requirements, researchers have proposed various statistical models, including the Gaussian, Gamma, Normal, and K-distributions [5,6]. These methods form the foundation of model-based target detection. Advancements in statistical theory have further led to sophisticated multi-channel adaptive detectors, such as Generalized Likelihood Ratio Test (GLRT)-based approaches, which are designed to enhance selectivity and performance in multi-channel data [7]. While these algorithms benefit from simple principles and computational efficiency, their detection performance is highly dependent on the accuracy of the clutter statistical model. While these model-based methods are mathematically rigorous and offer well-defined performance guarantees, they face challenges posed by the complex and non-stationary characteristics of real-world SAR backgrounds and the requirement for manual parameter tuning. In scenarios involving complex sea states, varying imaging conditions, or small, blurred targets, they often suffer from insufficient robustness and declining detection accuracy, making it difficult to meet the demands of high-precision marine monitoring.
Recent research has focused on addressing specific challenges, such as complex maritime conditions and multi-scale targets. For example, Bakirci [8] provided a comprehensive overview of advanced ship detection and ocean monitoring using satellite imagery and deep learning, highlighting the integration of multi-source data for marine science applications. Furthermore, the role of polarimetric information is critical; Adil et al. [9] analyzed the polarimetric scattering behavior of vessels at different incidence angles, underscoring the importance of polarization features for improving target-clutter discrimination. To tackle the issue of scale variation and complex backgrounds, models incorporating dynamic convolution and adaptive fusion mechanisms have been proposed, Li [10] demonstrating enhanced robustness. Our work builds upon these advancements by proposing a unified framework that systematically addresses the interrelated challenges of input data representation, multi-scale feature extraction, and global context modeling for SAR ship detection.
In recent years, advances in deep learning have enabled algorithms to autonomously learn key features from massive image datasets without complex modeling processes, demonstrating superior accuracy and robustness in object detection tasks. Two major categories of mature algorithms have emerged: the first is two-stage detection algorithms, which initially extract region proposals and then apply image classification to these candidate regions. Representative models include R-CNN [11], Faster R-CNN [12], and Cascade R-CNN [13]. In contrast, single-stage detection algorithms eliminate the region proposal step, directly predicting target locations from input images based on localization tasks. Typical examples include YOLO [14], SSD [15], and RetinaNet [16]. Furthermore, with the rise of Transformer models, numerous approaches have begun incorporating them into object detection. The Facebook team pioneered DETR (Detection Transformer) [17], an end-to-end object detection framework based on Transformer architecture. Although these classical detection algorithms have achieved remarkable success in natural image processing, their direct application to SAR ship target detection remains challenging due to the unique noise characteristics inherent to SAR images.
To address challenges in SAR ship target detection—including varying target scales, poor small-target detectability, and complex background interference—numerous researchers have refined the aforementioned classical algorithms. From a technical perspective, existing SAR ship detection methods primarily focus on improving the following model categories: CNN-based single-stage detection frameworks are widely adopted for their high efficiency. He et al. [18] reduced missed and false detections of multi-scale targets in complex scenes by optimizing feature extraction and detection head design. Tang et al. [19] enhanced YOLOv7’s ability to learn target scale diversity through their AMMRF module, improving detection accuracy while maintaining a lightweight structure. Liu et al. [20] enabled the FESAR method to better handle multi-scale variations by integrating convolutional enhancement and spatial relationship analysis modules. To address poor adaptability to slender ship shapes and sensitivity in small-target localization, Tang et al. [21] designed adaptive coordinate generation and multi-frequency perception guidance mechanisms, boosting detection accuracy for specially shaped targets. STSAR-YOLOv5 [22], LWM-YOLO [23], and Gao [24] were specifically optimized for small-target detection, complex backgrounds, and large-scale scenarios, respectively. While these methods balance detection efficiency and accuracy, they face limitations such as the loss of small-target features during repeated downsampling in deep networks. Transformer-based detection frameworks have emerged as another major direction due to their powerful global modeling capabilities. Xia et al. [25] proposed CRTransSar, which combines the global context awareness of Transformers with the local feature representation of CNNs, effectively addressing SAR targets’ scattering properties and background interference. Qu et al. [26] enhanced inter-target correlations through mask-guided Transformer encoders. RDB-DINO [27], SFRT-DETR [28], and small-target-enhanced RT-DETR [29] all built upon the Transformer architecture with improvements targeting small object sensitivity and multi-scale focus. These approaches mitigate insufficient global context modeling in complex backgrounds but face challenges such as high computational complexity and limited local detail perception. For wide-swath, low-resolution SAR images, Kang et al. [30] modified the region proposal strategy based on Faster R-CNN to achieve rapid ship detection. Li et al. [31] leveraged multi-polarization information and an improved VGG16 network to enhance classification performance for low-resolution targets. He et al. [32] improved recognition in medium-to-low resolution images through densely connected CNNs. SSD-YOLO [33] effectively suppressed background noise via small-target feature enhancement and self-attention path aggregation. These methods address specific issues of sparse features and low resolution in wide-swath images, but still have room for improvement in balancing small-target sensitivity and robustness to complex backgrounds.
Although the aforementioned methods have optimized SAR ship detection performance to some extent, inherent characteristics of SAR data—such as large-scale coverage, complex background interference, and sparse target distribution—continue to result in relatively low detection accuracy and frequent missed detections of small targets. Specifically, two technical issues persist: First, when processing single-channel SAR images, existing deep learning-based object detection models commonly employ a channel replication strategy to convert inputs into three channels. While this approach formally adapts the data to network architectures, it fails to effectively capture the transform-domain characteristics of SAR images, leading to significant deficiencies in target representation learning due to suboptimal information utilization. This highlights a gap between simple data adaptation and principled, theoretically-grounded input representation learning for SAR data. Second, ship detection in SAR images often suffers from small-target sizes and high missed detection rates. During target extraction, multiple downsampling operations in the network inevitably cause partial pixel information loss, subsequently giving rise to both missed detections and false alarms.
In summary, to address the challenges of single-channel information limitation, background clutter interference, and difficult feature extraction for small targets in SAR images, this study proposes a ship detection method integrating band reorganization and multi-scale enhancement. The proposed approach aims to improve the accuracy of ship target detection in SAR images, thereby providing more effective technical support for dynamic maritime vessel surveillance.

2. Methodology

The overall network architecture, illustrated in Figure 1, is a cohesive framework designed to systematically overcome key limitations of existing deep learning models in SAR ship detection. Rather than a simple aggregation of techniques, each component is purposefully integrated to address a specific shortcoming: (1) The Band Reorganization (BR) module is proposed to overcome the inefficiency of simple channel replication, which fails to extract discriminative ship information and amplifies noise. (2) The Coordinate Attention (CA) mechanism is introduced to enhance the network’s perceptual capability for locating ship targets amidst vast and cluttered backgrounds, a task where standard convolutions are deficient. (3) The Bottleneck Transformer with Triple Attention (BOT3) is incorporated to address the limited global feature modeling capability of traditional CNNs, which is crucial for understanding context in wide-swath SAR imagery. (4) To tackle the fundamental challenge of significant scale variation, two dedicated modules are designed: the Multi-Scale Feature Enhancement (MSFE) module focuses on extracting robust edge and detail features for small targets, while the Multi-Scale Effective Feature Aggregation (MEFA) module ensures effective integration of features across different levels of the network. This interconnected design ensures that each part complements the others, leading to a robust and efficient detection system.

2.1. SAR Image Band Reorganization

Deep learning networks were originally developed for natural images, which are typically three-channel RGB data, whereas SAR image consists of single-channel data [34]. Conventional methods replicate the single channel to construct pseudo-three-channel inputs. While this meets network input requirements, it introduces three issues: repeated noise superposition, feature homogenization, and loss of critical information. To address these limitations, this paper proposes a BR method based on multi-scale decomposition and dynamic fusion, which optimizes input representation through Laplacian pyramid reconstruction (as shown in Figure 2). The mathematical foundation of our band reorganization approach is built upon multi-scale pyramid decomposition with three hierarchical levels (m = 3), where each level captures distinct frequency components of the SAR image. The core procedure comprises the following four steps:

2.1.1. Gaussian Pyramid Hierarchical Noise Suppression

First, iterative Gaussian convolution and downsampling are performed on the original SAR image to generate multi-level low-resolution images (Gaussian pyramid). Higher pyramid levels exhibit progressively reduced resolution, thereby suppressing the influence of background noise. We employ a 3-level pyramid structure (k = 0, 1, 2) with a fixed downsampling ratio of 2:1 between consecutive levels.
G k = downsample ( G k 1 g 5 × 5 ) , g 5 × 5 = 1 256 1 4 6 4 1 4 16 24 16 4 6 24 36 24 6 4 16 24 16 4 1 4 6 4 1
where G k denotes the Gaussian pyramid image at the k-th level. g 5 × 5 is a 5 × 5 Gaussian convolution kernel matrix with standard deviation, and ⊗ denotes the 2D convolution operation between the image matrix and the kernel matrix. G k 1 is the image matrix at the ( k 1 ) -th level (higher resolution). The higher the layer, the lower the resolution. High-frequency noise decreases progressively with convolution and downsampling, while the bottom layer image retains global radiation features. The frequency domain response of the Gaussian kernel matches the high-frequency distribution characteristics of the noise in SAR images. By utilizing its low-pass filtering properties, it can effectively attenuate speckle noise, and the downsampling operation further reduces the spatial correlation of the noise.

2.1.2. High-Frequency Detail Extraction via Laplacian Pyramid

Multi-scale high-frequency components are extracted from the Gaussian pyramid via inverse difference reconstruction: The Laplacian pyramid is constructed across three levels (k = 0, 1, 2) using bilinear interpolation for upsampling operations.
L k = G k ( upsample ( G k + 1 ) g 5 × 5 )
In the above formula, L k is a matrix representing the high-frequency detail information of the k-th layer. G k + 1 are the Gaussian pyramid image matrices at the ( k + 1 ) -th levels, respectively. upsample ( · ) is the upsampling function that enlarges the image via interpolation. Since the Gaussian pyramid images have already smoothed the original data, low-frequency background noise is effectively suppressed in the high-frequency detail components. The Laplacian pyramid essentially captures multi-scale gradient information of the image. Through the “difference of Gaussians” operation applied across pyramid levels, it precisely isolates salient high-frequency details strongly associated with ship targets while filtering out the low-frequency background interference remaining after Gaussian smoothing.

2.1.3. Goal-Oriented Dynamic Weight Fusion

An adaptive weight allocation mechanism is designed to perform weighted fusion of multi-level high-frequency features: The fusion weights are dynamically computed based on the energy concentration within ship regions, where the ship binary mask M s h i p is generated using Otsu’s thresholding method.
D = k = 0 m ω k L k , ω k = L k M s h i p F j = 0 m L j M s h i p F
where D is the fused detail feature map matrix obtained through weighted summation. ω k is a scalar weight for the k-th layer Laplacian detail L k . M s h i p is the ship binary mask matrix generated based on scattering intensity thresholding, with ship regions as 1 and background as 0. • represents element-wise multiplication (Hadamard product). · F denotes the Frobenius norm, which calculates the square root of the sum of squared matrix elements, measuring the “energy” of detail information within ship regions. This weighting mechanism assigns higher weights to high-level details (e.g., L 0 ) in target-dense regions, while enhancing the contribution of low-level details (e.g., L 2 ) in areas with complex backgrounds.
The original image I r a w contains rich low-frequency information, providing the overall structure and background context of the scene, which is essential for accurate scene understanding. In contrast, the fused detail features D primarily consist of high-frequency information, emphasizing target edges, textures, and other fine details that aid in distinguishing different targets.
A three-channel input is constructed by selecting the original image G 0 , the first-level details L 1 , and the second-level details L 2 : This specific channel selection ( G 0 , L 2 , D ) was empirically determined to achieve an optimal balance among global context, intermediate details, and enhanced high-frequency information.
I R G B = [ G 0 , L 2 , D ]
where I R G B is the constructed three-channel input image tensor. [ · , · , · ] denotes the channel concatenation operation that stacks three single-channel matrices along the channel dimension. G 0 is the original image matrix preserving global information. L 2 is the second-level detail matrix.

2.2. Coordinate Attention Mechanism

Due to the absence of spectral information in SAR images and interference from coherent speckle and cross-shaped artifacts, the geometric representation of targets is highly susceptible to background noise. To enhance the network’s capability to characterize the spatial distribution features of targets, this paper introduces a CA mechanism [35], as illustrated in Figure 3. The CA mechanism is configured with a reduction ratio of r = 16 for channel reduction in the intermediate transformation.
Ship targets often exhibit significant directional elongation (e.g., along the major axis). The Coordinate Attention mechanism inherently captures long-range dependencies by performing 1D global pooling on the feature maps during the information aggregation stage. This operation separately captures long-range dependencies along the vertical and horizontal directions, effectively addressing the limitation of traditional attention mechanisms in modeling features along the elongated structure of ships. By computing features for the c-th channel separately along the vertical and horizontal directions, it locates targets while preserving precise positional information in the other direction, thereby effectively enhancing the accuracy of ship detection in the SAR image.
z c h = 1 W P a v g ( x )
z c w = 1 H P a v g ( x )
where z c h and z c w represent the outputs of the c-th channel feature map with vertical dimension H and horizontal span W, respectively, and P a v g ( ) denotes the average pooling layer. The separate pooling operations along the vertical and horizontal directions decouple the spatial encoding of targets. This approach preserves precise positional information in one dimension while compensating for the positional ambiguity caused by standard 2D pooling through cross-directional feature fusion. Consequently, the resulting attention weights can focus more precisely on ship target regions. This design aligns well with the pronounced directional specificity of ships in SAR imagery, enabling more efficient capture of spatial correlations along their major axis.
During the CA computation phase, the feature maps generated along the vertical and horizontal directions are concatenated to fuse multi-dimensional spatial information. This is followed by h-swish activation, convolutional operations, and batch normalization to enhance spatial perception capability for ships, resulting in a new intermediate feature map X H W . The intermediate feature transformation uses a 1 × 1 convolution that reduces channels to C/r, followed by batch normalization and h-swish activation.
X H W = C B h 1 × 1 c o n c a t z c h , z c w
where C B h 1 × 1 denotes a module comprising a 1 × 1 convolutional layer, batch normalization, and h-swish activation, while C o n c a t represents the concatenation of feature maps along the channel dimension.
In the attention weight generation phase, the intermediate feature map is decomposed into vertical and horizontal vectors. These vectors then pass through convolutional layers for channel adjustment and are activated by Sigmoid, yielding attention vectors X h and X w for the two respective directions. The final attention feature map Y , which highlights ship target characteristics, is generated through weighted computation. The final attention application uses element-wise multiplication between the input feature map and the generated attention maps.
Y = C B S 1 × 1 X h C B S 1 × 1 X w x
where C B S 1 × 1 denotes a module consisting of a 1 × 1 convolutional layer, batch normalization, and Sigmoid activation; ⊗ represents the outer product operator; and x refers to the input feature map.

2.3. Bottleneck Transformer

SAR image is characterized by complex textures and non-uniform noise distribution. When processing such high-dimensional nonlinear data, traditional convolutional neural networks often suffer from low computational efficiency and inadequate global feature modeling capabilities. To address these limitations, this paper introduces a Transformer-based BOT3 module (Bottleneck Transformer with Triple Attention) [36], whose schematic is shown in Figure 4. Designed with SAR imaging characteristics in mind, BOT3 incorporates lightweight mechanisms that maintain powerful global feature modeling capabilities while significantly reducing computational costs. The BOT3 module employs 8 attention heads with head dimension d = 64, and uses relative positional encoding with learnable parameters for both horizontal and vertical directions.
The BOT3 module adopts a hybrid CSP-Transformer architecture, whose core integrates a Multi-Head Self-Attention (MHSA) mechanism to enhance feature representation for multi-scale targets. This enables the capture of long-range dependencies between ships and marine backgrounds from multiple subspaces. The workflow is designed to model long-range dependencies between ships and the background, thereby overcoming the limited receptive field of traditional CNNs. It proceeds as follows: input features pass through three 1 × 1 convolutional layers to generate the Query ( Q ), Key ( K ), and Value ( V ). Simultaneously, a two-dimensional positional embedding, which adapts to the spatial characteristics of SAR imagery and prevents the self-attention mechanism from losing positional information, producing vertical ( R h ) and horizontal ( R w ) encoding vectors. These vectors are then summed to form a position encoding matrix R adapted to the 2D features. Matrix operations generate a fused feature map and a content attention map, which are summed and normalized, then multiplied by weights to obtain the output feature map Z . This process can be expressed as follows:
Q = C o n v l × l Q ( X )
K = C o n v 1 × 1 K X
V = C o n v 1 × 1 V X
R = R h + R w
where R h R H × d m o d e l and R w R W × d m o d e l are learnable positional encoding matrices for height and width dimensions, respectively.
E = Q R T + Q K T
A = S o f t m a x ( E )
Z = V A T
where X R N × C × W × H denotes the input feature map, with N representing the batch size, C the number of channels, and W and H the width and height of the feature map, respectively. E is the logit matrix that combines attention to position and content. A is the attention weight matrix.
Given the relatively limited data volume and computational resources, the Bottleneck Transformer module reduces parameter count by optimizing convolutional and self-attention operations. It also incorporates residual connections to mitigate gradient vanishing in deep networks, thereby enhancing training stability and efficiency. The module adds only one convolutional layer before and after the MHSA for dimension adjustment, enabling efficient processing of SAR images. We employ n = 3 consecutive Bottleneck Transformer blocks in the BOT3 module for an optimal performance–complexity trade-off.
X out = C o n ν l × 1 x + M H S A C o n ν l × 1 X
where M H S A ( ) denotes the multi-head self-attention operation. X o u t is the output tensor of the Bottleneck Transformer block, incorporating residual connection (+).
The BOT3 module follows the CSP architecture, enhancing expressive capability through feature branching and fusion. The input features are processed by two 1 × 1 convolutional layers to generate feature maps X 1 and X 2 . After extracting enhanced features by stacking n Bottleneck Transformers, the results are concatenated along the channel dimension, followed by a 1 × 1 convolutional layer to map the features to the output channels X BOT 3 . The CSP split ratio is set to 0.5, meaning the feature channels are equally divided between the transformer path and identity path.
C o n v 1 × 1 ( X ) [ X 1 , X 2 ]
X m = B o t t l e n e c k T r a n s f o r m e r n X 1
X B O T 3 = C o n v 1 × 1 c o n c a t X m , X 2
The BOT3 module effectively integrates low-level spatial features X 2 with high-level global contextual representations X m . The resulting output feature map concurrently captures both spatial details and contextual information, thereby providing more robust feature inputs for the detection head.

2.4. Multi-Scale Feature Enhancement Module

An SAR image inherently contains complexities such as coherent speckle noise and complex marine backgrounds. Coupled with the common characteristics of remote sensing images, these factors often lead to misjudgments due to feature similarity in small ship target detection. Furthermore, traditional backbone networks exhibit limited capability when processing SAR images, often yielding features with insufficient semantic information and restricted receptive fields, making it difficult to effectively distinguish small ship targets from the background. To address these issues, this paper draws inspiration from the Feature Enhancement Module (FEM) [37] and constructs an MSFE module, as illustrated in Figure 5. This module comprises several key components—a unique multi-branch dilated convolution structure, an edge enhancer, and feature fusion mechanisms—that work collaboratively to enhance the feature representation of small ship targets. The MSFE module processes input features of dimension C × H × W and maintains the same output dimension through careful dimension preservation in all operations.
The multi-branch dilated convolution structure divides the input feature map into four branches, each performing a 1 × 1 convolutional operation on the input features. The first branch employs a residual structure, effectively preserving critical feature information of small targets by constructing an equivalent mapping. This design is particularly important for detecting small ship targets in SAR images: their features are typically weak and prone to loss during convolutional operations, but the residual structure ensures these key characteristics are retained, providing a solid foundation for subsequent detection and recognition. The other three branches perform cascaded standard convolution operations with kernel sizes of 1 × 3, 3 × 1, and 3 × 3, respectively. This diverse kernel configuration helps process the input feature map from different perspectives and scales, thereby extracting richer feature information. Notably, the two middle branches incorporate additional dilated convolution layers, enabling the extracted feature maps to capture broader contextual information. In complex marine backgrounds of SAR images, environmental cues around small ship targets are crucial for accurate identification. The dilated convolution layers expand the receptive field, effectively preserving these contextual details and thereby improving the model’s detection accuracy for small targets and its adaptability to complex backgrounds. All convolutional layers use padding=’same’ to maintain spatial dimensions, and ReLU activation follows each convolution except the final output layer.
The output feature maps of the latter three branches can be expressed as follows:
C 1 = C B R 3 × 3 [ C B R 1 × 1 ( x ) ]
C 2 = D C B R 3 × 3 5 C B R 3 × 1 C B R 1 × 3 C B R 1 × 1 x
C 3 = D C B R 3 × 3 5 C B R 1 × 3 C B R 3 × 1 C B R 1 × 1 x
where C B R m × n ( ) denotes a module comprising a convolutional layer (kernel size m × n), batch normalization, and a ReLU activation layer. Here, D C B R 3 × 3 5 specifically represents a convolutional layer with a 3 × 3 kernel and a dilation rate of 5. x is the input original feature map, while C 1 , C 2 , and C 3 are the output feature maps of the three branches after standard and dilated convolutions, respectively. The dilation rate of 5 was empirically determined to provide optimal receptive field expansion without grid artifacts, and all 1 × 1 convolutions use 256 output channels.
Subsequently, the multi-scale feature maps from the multiple branches are fused along the channel dimension, followed by channel dimensionality reduction using a 1 × 1 convolutional kernel, yielding the fused feature map C 4 . This process can be expressed as follows:
C 4 = C B 1 × 1 C o n c a t ( C 1 , C 2 , C 3 ) + C B 1 × 1 x
where C o n c a t ( ) denotes the concatenation of feature maps along the channel dimension, and C B 1 × 1 ( ) represents a module consisting of a convolutional layer with kernel size 1 × 1 and a batch normalization layer. The final 1 × 1 convolution reduces the channel dimension from 768 (3 × 256) back to 256 to maintain dimension consistency.
The Edge Enhancer (EE) emphasizes crucial boundaries of ships in the image through detailed processing and information extraction of the input feature map. Specifically, the EE first applies an average pooling layer P a v g ( ) to suppress high-frequency noise and detailed information in the feature map, resulting in a preliminarily smoothed feature map A v g P o o l . This operation can be mathematically expressed as
A v g P o o l = P a v g ( C 5 )
where the average pooling uses a kernel size of 3 × 3 and a stride of 1 to preserve spatial resolution while providing sufficient smoothing.
Subsequently, the edge information feature map C 5 is obtained by subtracting the average-pooled feature map A v g P o o l from the edge information E , expressed as:
E = C 5 A v g P o o l
This process successfully separates potential edge information by leveraging the differential characteristics between edge regions and smooth areas in the image. To further enhance the extracted edge information and make it more prominent in subsequent feature representations, the obtained edge information E is sequentially processed through a convolutional layer and an activation function. The convolutional layer performs feature extraction and transformation on the edge information, while the activation function introduces nonlinearity to strengthen its representational capacity. The enhanced edge information C 6 is obtained through this series of operations, expressed as follows:
C 6 = C B S ( E )
where C B S ( ) denotes a module comprising a convolutional layer with kernel size 1 × 1, a batch normalization layer, and a ReLU activation layer.
Subsequently, the enhanced edge information C 6 is added to the original edge information feature map C 5 to obtain the final enhanced feature map X , expressed as
X = C 6 + C 5
Through this approach, the Edge Enhancer effectively highlights key object boundaries, enabling the enhanced feature map to strengthen the representation of object edges while preserving original feature information. This provides richer and more accurate feature representations for subsequent target detection and recognition tasks.
Finally, the fused feature map C 4 from the four branches and the edge-enhanced feature map obtained through multiple Edge Enhancers are concatenated along the channel dimension. The concatenated result is then passed through a 1 × 1 convolutional layer for dimensionality reduction, yielding the final feature map X end . The MSFE module employs 3 Edge Enhancer blocks in series, with each enhancer processing the output of the previous one. This process can be expressed as
X end = C B R 1 × 1 C o n c a t C 4 , X 1 + X 2 + X 3
where X 1 , X 2 , X 3 represent the outputs from the three sequential Edge Enhancer blocks.

2.5. Multi-Scale Effective Feature Aggregation Module

In large-scene wide-swath SAR ship target detection, significant variations in ship scales—from small fishing boats to large cargo vessels—result in considerable differences in their pixel sizes within images. Additionally, complex and variable marine backgrounds interfere with detection, while substantial image noise further exacerbates the difficulty, posing severe challenges to ship target detection. Specifically, small fishing boats occupy few pixels and exhibit weak features in SAR images, making them highly susceptible to being flooded by background noise. Although large cargo ships have relatively larger sizes, they still occupy limited regions in wide-swath images, and the structural and contour differences among various ship types further increase detection difficulty. To address these issues, this paper designs a Multi-scale Effective Feature Aggregation (MEFA) module to significantly enhance the feature representation capability of convolutional neural networks when processing a large-scene wide-swath SAR image. The MEFA module takes two input feature maps, X 1 and X 2 , from different network levels and fuses them through concatenation before multi-scale processing in Figure 6.
The MEFA module integrates feature map X 1 from different levels of the backbone network with intermediate feature map X 2 from the detection head, combining multi-scale information to enable the model to simultaneously capture local details from low-level features and global semantics from high-level features. This allows the model to better adapt to the significant scale variations of ship targets in large-scene wide-swath SAR images and improve detection capability for ships of different sizes. The fused feature map is split into four branches: X 1 , X 2 , X 3 , X 4 . Three of these branches undergo convolutional operations with kernel sizes of 3 × 3, 5 × 5, and 7 × 7, respectively, extracting multi-scale feature information from the image. For small ship targets, smaller kernels (e.g., 3 × 3) capture fine-grained details, while for large ship targets, larger kernels (e.g., 7 × 7) capture overall structural information. These three branches are then concatenated with the original fused feature map along the channel dimension, preserving both multi-scale target information and original image content. The kernel size progression (3, 5, 7) was designed to capture increasingly larger contextual information, with corresponding padding values (1, 2, 3) selected to maintain spatial dimensions. This process can be mathematically expressed as follows:
X end = C B R 1 × 1 C o n c a t C 4 , X 1 + X 2 + X 3
x 1 = C o n v 1 × 1 1 4 X c o n c a t R C × H × W
x 2 = C o n v 3 × 3 P = 1 1 4 X c o n c a t R C × H × W
x 3 = C o n v 5 × 5 P = 2 1 4 X c o n c a t R C × H × W
x 4 = C o n v 7 × 7 P = 3 1 4 X c o n c a t R C × H × W
X f u s e d = C o n v 1 × 1 c o n c a t x 1 , x 2 , x 3 , x 4 , dim = 1 R C × H × W
where C o n v n × n P = m denotes a grouped convolution with kernel size n×n and padding m. All convolutional layers in the multi-scale branches use 64 output channels, resulting in 256 total channels after concatenation, which are then reduced back to C channels by the final 1 × 1 convolution.
Incorporating the CBAM attention mechanism, dual visual focus is achieved through channel attention and spatial attention, dynamically suppressing interference from regions such as waves and islands by weighting the input features. The CBAM component uses a reduction ratio r = 16 in the channel attention MLP, and the spatial attention employs 7 × 7 convolution for spatial feature aggregation.
W c = S i g m o i d M L P P a v g X + M L P P m a x X
W s = S i g m o i d C o n v 1 × 1 M e a n ( X ) , M a x ( X )
Y 1 = X 1 W c W s , Y 2 = X 2 W c W s
Y M E F A = S i g m o i d X f u s e d Y 1 + S i g m o i d X f u s e d Y 2 + Y 1 + Y 2
where P a v g ( ) denotes the average pooling layer, P m a x ( ) denotes the max pooling layer, M e a n ( ) represents average pooling along the channel dimension, and M a x ( ) represents max pooling along the channel dimension. W c and W s are channel and spatial attention weight tensors. Y 1 and Y 2 are attention-weighted feature map tensors. Y M E F A is the final output tensor of the MEFA module. The final fusion strategy employs element-wise multiplication between the multi-scale fused features and attention-weighted inputs, followed by residual connections to preserve original information.

3. Experimental Results and Analysis

3.1. Dataset and Experimental Configuration

To evaluate and compare the performance of different algorithms in SAR image target detection tasks, experiments were conducted using commonly adopted SAR ship target detection datasets, including the LS-SSDD dataset [38] and the HRSID dataset [39]. Key parameters of these datasets are summarized in Table 1. Additionally, large-scene wide-swath SAR images from real-world scenarios, all sourced from Sentinel-1, were also utilized.
LS-SSDD, introduced in 2020, is the first public benchmark dataset designed for large-scene small-target SAR ship detection. It is constructed from Sentinel-1 Interferometric Wide Swath (IW) mode data, with a ground range resolution of 5 m and an azimuth resolution of 20 m. The dataset covers multiple typical maritime regions worldwide and comprises 15 original SAR images, each with dimensions of 24,000 × 16,000 pixels. The training set and test set were split in an 8:2 ratio.
HRSID consists of 5604 SAR images captured by Sentinel-1 and TerraSAR-X, each with a size of 800 × 800 pixels and resolutions of 0.5 m, 1 m, and 3 m. The dataset contains 16,965 ship instances, including 15,510 small ships, 1344 medium ships, and 111 large ships. The data were divided into training, test, and validation sets in a 6:2:2 ratio.
Prior to conducting experiments, each dataset was divided into training and test sets. The training set was used to train the model, and when the training loss and accuracy stabilized, predictions were made on the test set images to validate the model’s generalization capability. The model was built using the PyTorch1.10 deep learning framework with the following configuration: a batch size of 16, an initial learning rate of 0.00001, and a maximum of 150 training epochs. The model was optimized using the Stochastic Gradient Descent (SGD) algorithm and trained on an NVIDIA 4090D GPU with 24 GB of memory provided by the AutoDL platform.

3.2. Evaluation Metrics

This paper employs precision (P), recall (R), and average precision (AP) to evaluate the performance of the detection model. Precision is defined as the proportion of correctly detected ships among all detected ships:
P = T P T P + F P
Recall represents the proportion of successfully detected targets among all actual targets:
R = T P T P + F N
where F N denotes false negatives (missed detections), T P represents true positives (correct detections satisfying two conditions: correct category—predicted class matches the ground truth class, and correct localization—IoU between predicted and ground truth boxes is greater than or equal to a preset threshold). F P indicates false positives (false detections).
A P is determined by the precision–recall curve and is defined as follows:
A P = 0 1 P ( R ) d R
P ( R ) is represented by the area under the precision–recall curve. It reflects the accuracy of the target detection model. Following the COCO dataset A P calculation method, m A P 50 and m A P 50 : 95 are computed separately. m A P 50 is the average precision when the Intersection over Union (IoU) between the detection box and the ground truth box is greater than 0.5, while m A P 50 : 95 denotes the average precision with IoU thresholds ranging from 0.5 to 0.95 in increments of 0.05. Computational Load (GFLOPs) and Parameter Count (Para, M) represent the model’s complexity and computational cost. FPS (Frames Per Second) indicates the model’s inference speed.

3.3. Ablation Study on Network Architecture Effectiveness

To validate the effectiveness of the proposed method and provide theoretical motivation for each module design, a progressive ablation study was conducted to investigate the contribution of each improved module to the SAR ship detection task.
Using the LS-SSDD dataset and YOLOv8 as the baseline model, the experiments sequentially incorporated BR, CA, BOT3, MSFE, and MEFA modules to evaluate their individual improvements. As shown in Table 2, the BR module improved m A P 50 from 0.712 to 0.743 and increased recall by 4.3% without increasing computational cost (maintaining 3.0 GFLOPs), confirming its effectiveness as an efficient preprocessing strategy that suppresses background noise in SAR images via channel restructuring. This confirms that frequency-domain processing provides a computationally efficient foundation for subsequent modules.
Further incorporation of the CA mechanism enhanced the model’s feature extraction capability for ships, raising m A P 50 to 0.760 and P to 0.864, validating its role in capturing directional spatial dependencies crucial for ship target localization. The addition of the BOT3 module improved m A P 50 : 95 to 0.311, highlighting the importance of global context modeling for robust ship detection in complex marine environments.
Introducing the MEFA module further increased m A P 50 to 0.78 through dynamic weighted fusion of multi-scale features. Finally, integrating the MSFE branch expanded the receptive field via dilated convolutions and balanced detection performance across targets of different scales. Although precision slightly decreased to 0.867, m A P 50 : 95 reached a peak of 0.316. This precision–recall trade-off illustrates the complementary nature of these modules: while MEFA enhances feature aggregation, MSFE improves multi-scale representation at the cost of some false positives.
A critical finding emerges from analyzing module combinations: omitting the BR and CA using only BOT3, MEFA, and MSFE caused a sharp drop in m A P 50 : 95 to 0.744. This strongly confirms that the modules exhibit synergistic effects: the foundational preprocessing (BR) and spatial attention (CA) modules are essential for the effective operation of the more complex multi-scale modules.
When all modules were combined, the proposed method achieved the best performance across all metrics except P, where it was slightly lower than the configuration without MSFE. The MSFE module’s multi-scale receptive fields and feature enhancement capabilities improve detection sensitivity, especially for small targets, but simultaneously increase susceptibility to background artifacts in complex SAR scenes. This is quantitatively evidenced by the 1.5% increase in recall accompanied by a 1.5% decrease in precision.
The relationship between the precision decrease and multi-scale feature fusion can be visualized in Figure 7, which presents representative false detection cases. We attribute the MSFE module’s performance characteristics to its ability to capture multi-scale target information through feature fusion and enhancement, improving feature complementarity and expressive power, strengthening key features while suppressing noise, and adapting to complex scenes by learning multi-scenario feature patterns—all of which benefit target detection. P, defined as the ratio of correctly detected targets to all detections, decreased slightly after multi-scale feature fusion. Although the detection of small targets increased, some irrelevant false detections were also introduced, as illustrated in Figure 7. R, defined as the ratio of correctly detected targets to all annotated targets, improved, indicating that the model successfully detected more true targets. This suggests that multi-scale feature fusion enhances small-target detection capability but also introduces a certain degree of false detections.
In the LS-SSDD dataset, the tendency to over-learn background pseudo-features and feature-level conflicts, combined with the complexity of SAR scenes and sparse target distribution, led to increased false detections and a noticeable drop in P. However, since some true positives were still correctly identified, R and mAP were less affected. This demonstrates that our module combination achieves an optimal balance for SAR ship detection, where the gains in recall and overall mAP outweigh the slight precision reduction.
As the proposed MSFE module is an improved version of the FEM module, an ablation experiment was also conducted by replacing MSFE with FEM. The results show that the FEM module performs worse than the MSFE module in all indicators, providing concrete evidence that our architectural modifications in MSFE are theoretically justified and empirically superior for small ship target feature representation. Meanwhile, the more complex structure of MSFE also led to a slight increase in Para and GFLOPs.
To further validate the module complementarity, we conducted replacement experiments with alternative approaches. The proposed BR module surpassed wavelet transform (WT) and principal component analysis (PCA) methods, confirming the superiority of our multi-scale decomposition and dynamic fusion approach for SAR-specific channel reorganization.
These results collectively validate that each module addresses distinct challenges in SAR ship detection while exhibiting complementary synergistic effects: BR provides foundational feature optimization through frequency-domain processing, CA enhances directional spatial perception, BOT3 improves global contextual representation, while MEFA and MSFE strengthen multi-scale feature fusion—together forming an efficient and robust framework that delivers measurable improvements beyond mere architectural complexity.
To further investigate the generalizability and theoretical robustness of the proposed method, ablation experiments were also conducted on the HRSID dataset. The results demonstrate consistent performance patterns that reinforce our theoretical motivations. The BR module effectively enhances model performance across both datasets, with improvements observed in P, R, and m A P 50 , confirming its universal effectiveness in SAR image preprocessing. The CA module significantly increases inference speed while maintaining accuracy, validating its efficiency in spatial attention modeling. The BOT3 and MSFE modules achieve a favorable balance between accuracy and computational efficiency across both datasets, demonstrating the generalizability of our global context and multi-scale feature learning approaches.
A particularly revealing finding emerges from the HRSID experiments: when both the BR and CA modules were omitted, R dropped significantly to 0.848. In contrast, when all modules were integrated, the model achieved relatively superior results across P, R, and mAP, providing cross-dataset evidence for the necessity of the complete architecture and the synergistic effects between modules.
The computational complexity analysis across both datasets reveals important theoretical insights into our architectural design. As shown in Table 2 and Table 3, the gradual incorporation of modules leads to increases in both parameter count and computational load. The introduction of the CA module causes a noticeable jump in these metrics, which can be attributed to its use of global pooling, fully connected matrix operations, and activation functions—all of which require substantial parameters and serial operations. These components provide theoretical benefits by enabling the model to focus on ship-related channel features, suppress background noise, and improve the detection accuracy of small targets.
Subsequent modules, such as BOT3 and MEFA, also increase parameters and computations but are designed with multi-branch parallel structures (e.g., multi-scale branches in BOT3, cross-layer feature fusion in MEFA). Theoretical analysis indicates that these operations—primarily based on parallel convolution and feature concatenation—are better aligned with GPU parallel acceleration capabilities, leading to higher hardware utilization. In addition, the newly added parameters are mostly weights of branch convolutions, which are gradually stacked and do not exceed the original complexity threshold of the model. Therefore, the impact on computation and FPS is more gradual.
This computational pattern analysis provides theoretical justification for our architectural choices: while individual modules increase complexity, their careful integration maintains practical deployability while delivering measurable performance improvements that justify the added computational cost.
The SAR images in the LS-SSDD and HRSID datasets differ in noise level, contrast, and target scale, resulting in variations in computational resource consumption and the model’s parallel adaptability after module stacking. Moreover, the magnitude and rate of increase and decrease in indicators in the two datasets are different, but the overall trend remains consistent. This consistency across diverse SAR datasets strongly supports the theoretical robustness and generalizability of our proposed architecture.

3.4. Comparative Experiments with Other Methods

To validate the effectiveness of the proposed method in SAR image ship target detection, comparative experiments were conducted on both the LS-SSDD and HRSID datasets. The models included in the comparison were categorized into three groups to ensure thorough and insightful analysis: Classical detectors (Faster R-CNN [12], Cascade R-CNN [13], EfficientDet [40], SSD [15], FCOS [41], and GL-DETR [42]), Specialized SAR detection models (HR-SDNet [43], SARNet [44], YOLO-SARSI [19], MSIF [45], MSFA [46], FESAR [20], SARNAS [47], SSD-YOLO [33], FDI-YOLO [48], ROSD [49], SIF-Net [50], DRGD-YOLO [51], and PSG [52]), YOLO series baseline models (YOLOv5s [53], YOLOv8 [54], YOLOv11 [55], and YOLOv12 [56]).
On the LS-SSDD dataset, which focuses on small-target detection, the proposed method achieved an m A P 50 of 0.781, outperforming all compared classical and newer models (as shown in Table 4). In particular, our method achieves 0.316 on m A P 50 : 95 , which is significantly higher than other methods, indicating that the generated prediction boxes have higher boundary accuracy. The performance superiority stems from our holistic approach: the Band Reorganization module provides cleaner inputs by suppressing background noise, while the multi-scale enhancement and attention mechanisms significantly improve target-clutter discrimination. The most notable improvements are observed in m A P 50 : 95 (13.3% higher than YOLOv8) and recall, demonstrating enhanced boundary precision and reduced missed detections.
Compared to YOLO series baseline models, the introduction of BR, CA, BOT3, and MSFE structures led to a significant improvement in detection performance while maintaining the high efficiency inherent to single-stage detectors. In terms of model complexity, the proposed method achieves a balance with moderate parameter counts and computational load, while reaching an inference speed of 175.44 FPS. This frame rate substantially exceeds that of most two-stage models and Transformer-based detectors, demonstrating strong potential for real-time processing.
To further validate the generalization capability of the proposed method in complex scenarios, experimental results on the HRSID dataset demonstrate that our approach achieves an m A P 50 of 0.945—the best among all compared methods. Compared with specialized SAR ship detection models, e.g., FESAR and SSD-YOLO, the proposed method achieves a significant breakthrough in detection accuracy. Notably, while maintaining high precision (0.923), the proposed method also achieves the highest recall (0.889) among all models. This indicates its ability to more comprehensively identify targets in complex marine environments and dense scenes while effectively controlling false detections. The balanced improvement across both precision and recall metrics underscores our method’s capability to handle the precision–recall trade-off that often challenges SAR ship detection. These results confirm the effectiveness of the proposed multi-scale enhancement and attention mechanisms in improving model generalization.
As summarized in Table 4 and Table 5, the proposed method achieves optimal performance on both the LS-SSDD and HRSID datasets in terms of core detection metrics ( m A P 50 and m A P 50 : 95 ), demonstrating its high accuracy and strong generalization capability for ship detection in wide-swath SAR images.
To more intuitively demonstrate the superiority of the proposed algorithm in small-target detection, representative scenes from both the LS-SSDD and HRSID datasets were selected for comparative analysis against other algorithms, as illustrated in Figure 8. The experimental results show that the proposed algorithm achieves optimal performance in small-target detection tasks (highlighted by the blue bounding boxes in the figure), significantly reducing both missed detections and false alarms while generally achieving higher confidence scores. These visual results corroborate the quantitative analysis, particularly demonstrating our method’s effectiveness in detecting challenging small targets that are frequently missed by other approaches. These outcomes fully demonstrate that the proposed algorithm exhibits stronger robustness in wide, large-scene, small-target detection and holds broader application potential.

3.5. Generalization Performance Evaluation

Building upon the benchmark validation using public datasets (LS-SSDD and HRSID), this section further evaluates the adaptability and generalization capability of the proposed method for real-world large-scene wide-swath SAR images by incorporating the SAHI slicing inference strategy. The experimental data were selected from three geographically and oceanographically distinct regions: the U.S. West Coast, East Coast, and the northwestern Gulf of Mexico. These areas were chosen to validate the method’s robustness across different marine environments—featuring steep coastlines on the West Coast, and gentle continental shelves with busy shipping lanes on the East Coast and in the Gulf of Mexico.
The data were sourced from Sentinel-1 Interferometric Wide Swath (IW) mode Single Look Complex (SLC) products provided by the Alaska Satellite Facility (ASF), with VV + VH polarization and a standard coverage range of approximately 250 km. The raw data downloaded from these websites were processed through a series of operations—including orbit correction, radiometric calibration, multi-looking, filtering, and geocoding—to generate images consistent with the format of the LS-SSDD dataset. For incidence angles ranging from 30° to 45°, which are characteristic of the Sentinel-1 Interferometric Wide (IW) swath mode, ships typically exhibit higher backscattering values in the co-polarization channel (VV), while lower backscattering is observed in the cross-polarization channel (VH) [57,58]. In cross-polarization, ships present a smaller radar cross-section (RCS), and the clutter often falls below the noise floor, making detection noise-limited [58]. The use of cross-polarization offers notable advantages, particularly for acquisitions with small incidence angle [58]. Moreover, since the noise floor is largely independent of incidence angle, cross-polarization enables more uniform ship detection performance across the image swath. Therefore, for the purpose of this study, which focuses on developing a robust detection network, VH-polarized SAR imagery was selected. Large sub-regions were extracted from the full-scene data for testing. During processing, the SAHI framework was employed for sliding window inference. Given the image resolution of approximately 10 meters per pixel, the base slice size was set to 800 × 800 pixels with an overlap ratio of 20% to effectively mitigate missed detections caused by ship targets being truncated at slice boundaries.
Finally, a weighted non-maximum suppression algorithm was applied to merge the initial detection results from all slices. An IoU threshold of 0.4 was set, and confidence-weighted averaging was performed on bounding boxes in overlapping regions to generate seamless, large-scene detection results.
To intuitively demonstrate the detection performance of the proposed method in complex large-scene real-world data, the detection results for the three scenarios are visualized in Figure 9, Figure 10, Figure 11 and Figure 12. For each image, a representative sub-region on the right side is selected and displayed as a zoomed-in view to clearly illustrate the actual detection effects. The annotation rules for the detection results are as follows: red rectangular boxes indicate ship targets correctly detected by the proposed algorithm, blue elliptical boxes mark missed detections, and green boxes indicate false detections. Analysis of the zoomed-in regions shows that the proposed method performs well in detecting small-scale targets. However, under interference from complex background clutter, instances remain where island or reef features are misclassified as ship targets.
Given the current lack of publicly available large-scale annotated datasets for SAR ship detection, this study employs expert visual interpretation as the quasi-ground truth for quantitative evaluation. The researchers performed systematic visual analysis of the full-scene image without reference to any algorithmic outputs, identifying targets based on morphological characteristics, scattering intensity, and land-sea contextual relationships. The detection results of the proposed algorithm were subsequently compared against this ground truth set, as summarized in Table 6. The proposed method demonstrates excellent generalization performance across wide-swath SAR images from different geographical regions. In the West Coast scenario, where the background is relatively simple and sea clutter interference is weak (providing an optimal testing environment), the method achieved the highest detection rate and precision. Although precision decreases in the more complex backgrounds of the East Coast and Gulf of Mexico scenarios, the detection rate remains consistently above 80%. This confirms the method’s stable and reliable ship detection capability across diverse wide-area scenes. The East Coast and Gulf of Mexico scenarios present greater complexity, with dense shipping traffic, gentle coastlines, and offshore facilities generating numerous false targets. These conditions place higher demands on the algorithm’s ability to distinguish foreground targets from background clutter, leading to a reduction in precision. Nevertheless, even under these challenging conditions, the proposed method maintains a detection rate above 80%, highlighting its strong robustness in capturing true targets within complex backgrounds.

4. Conclusions

SAR image-based methods for ship detection were often challenged by intense background clutter and the difficulty of extracting discriminative features from small targets in wide-swath images, resulting in limited detection performance. To mitigate these issues, this paper proposed a novel ship detection algorithm that integrated multi-channel feature extraction with a multi-scale enhancement scheme for small targets. The proposed algorithm enhanced input feature representation through channel recombination. It further incorporated a CA mechanism and a BOT3 module to bolster spatial feature discrimination and global context modeling. Additionally, a dedicated multi-scale module was constructed to facilitate feature refinement and information aggregation, collectively improving the detection accuracy for multi-scale ship targets. Extensive experiments on the LS-SSDD and HRSID datasets demonstrated that our algorithm significantly enhanced the detection of small ship targets in wide-swath SAR images, offering reliable technical support for maritime surveillance in complex, large-scale scenarios. Nevertheless, the proposed approach still exhibited certain limitations, including false positives for large ships, high computational complexity, and insufficient robustness to physical distortions inherent in wide-swath images. Future research will prioritize developing false positive suppression strategies, pursuing lightweight model design, and incorporating AIS data association alongside physical distortion modeling to further enhance the algorithm’s robustness and practical utility.

Author Contributions

Conceptualization, Y.Z., K.Z., H.G., J.L., Z.G. and X.L.; methodology, Y.Z.; validation, Y.Z. and K.Z.; formal analysis, Y.Z. and K.Z.; investigation, Y.Z. and K.Z.; resources, Y.Z. and K.Z.; data curation, Y.Z. and K.Z.; writing—original draft preparation, Y.Z., K.Z., H.G., J.L., Z.G. and X.L.; writing—review and editing, Y.Z. and K.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China, No. 42301464; China Postdoctoral Science Foundation, No. 2023M744302; The Key Project of Science and Technology of Henan Province, No. 252102211061; and Natural Science Foundation of Henan Province, No. 252300420301; Key Laboratory Project, No. KX25203.

Data Availability Statement

The experimental data for the generalization test were obtained from https://search.asf.alaska.edu/(accessed on 10 January 2025).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Wu, L.B. SAR Image Processing and Target Recognition; Aviation Industry Press: Beijing, China, 2013. [Google Scholar]
  2. Wei, S.J.; Jiang, P.F.; Yuan, Q.Z. Detection and Recognition of SAR Small Ship Objects Using Deep Neural Network. J. Northwestern Polytech. Univ. 2019, 37, 546–552. [Google Scholar] [CrossRef]
  3. Zheng, Z.Y.; Zhang, T.; Liu, Z.Y.; Li, Y.J.; Sun, C.M. Ship target detection in SAR images based on RCIoU loss function. J. Shandong Univ. (Eng. Sci.) 2022, 52, 15–22. [Google Scholar]
  4. Leng, X.G. Study on Ship Detection Technology Based on Complex-Valued Information in Single-Channel SAR Imagery. Master’s Thesis, National University of Defense Technology, Changsha, China, 2019. [Google Scholar]
  5. El-Darymli, K.; Gill, E.W.; Mcguire, P.; Power, D.; Moloney, C. Automatic target recognition in synthetic aperture radar imagery: A state-of-the-art review. IEEE Access 2016, 4, 6014–6058. [Google Scholar] [CrossRef]
  6. Gao, G. Statistical modeling of SAR images: A survey. Sensors 2010, 10, 775–795. [Google Scholar] [CrossRef]
  7. Liu, W.; Cao, H.; Xiong, G.; Liu, J.; Qi, C. GLRT-based detectors with enhanced selectivity for mismatched signals through a random-signal approach. Signal Process. 2026, 240, 110375. [Google Scholar] [CrossRef]
  8. Bakirci, M. Advanced ship detection and ocean monitoring with satellite imagery and deep learning for marine science applications. Reg. Stud. Mar. Sci. 2025, 81, 103975. [Google Scholar] [CrossRef]
  9. Adil, M.; Nunziata, F.; Buono, A.; Velotto, D.; Migliaccio, M. Polarimetric Scattering by a Vessel at Different Incidence Angles. IEEE Geosci. Remote Sens. Lett. 2023, 20, 4008605. [Google Scholar] [CrossRef]
  10. Li, Z.; Deng, Z.; Hao, K.; Zhao, X.; Jin, Z. A Ship Detection Model Based on Dynamic Convolution and an Adaptive Fusion Network for Complex Maritime Conditions. Sensors 2024, 24, 859. [Google Scholar] [CrossRef] [PubMed]
  11. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar]
  12. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef]
  13. Cai, Z.; Vasconcelos, N. Cascade r-cnn: Delving into high quality object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 6154–6162. [Google Scholar]
  14. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  15. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. SSD: Single Shot MultiBox Detector. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; pp. 21–37. [Google Scholar]
  16. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal Loss for Dense Object Detection. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2999–3007. [Google Scholar]
  17. Zhao, Y.; Lv, W.Y.; Xu, S.L.; Wei, J.M.; Wang, G.Z.; Dang, Q.Q.; Liu, Y.; Chen, J. DETRs Beat YOLOs on Real-time Object Detection. In Proceedings of the 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 17–21 June 2024; pp. 16965–16974. [Google Scholar]
  18. He, S.; Wang, Y.Z.; Yang, Z.W. Multi-scale ship detection algorithm in SAR images in complex scenes. Appl. Electron. Tech. 2025, 51, 59–64. [Google Scholar]
  19. Tang, H.; Gao, S.; Li, S.; Wang, P.; Liu, J.; Wang, S.; Qian, J. A Lightweight SAR Image Ship Detection Method Based on Improved Convolution and YOLOv7. Remote Sens. 2024, 16, 486. [Google Scholar] [CrossRef]
  20. Liu, C.; Yan, C. FESAR: SAR ship detection model based on local spatial relationship capture and fused convolutional enhancement. Mach. Vis. Appl. 2024, 35, 34. [Google Scholar] [CrossRef]
  21. Tang, F.L.; Li, W.S.; Qiao, G.J. Ship Detection in SAR Images Based on Adaptive Coordinate Generation and Multi-Frequency Perception. Laser Optoelectron. Prog. 2025, 62, 2028006. [Google Scholar]
  22. Xu, J.T.; Wang, Z.W.; Zhao, G.P. Small target detection algorithm based on SAR images. Comput. Eng. Des. 2025, 46, 570–577. [Google Scholar]
  23. Zhang, Y.; Zhang, Y.; Wang, T.; Wang, B.Y. Lightweight research on ship target detection in large-scale SAR images. J. Geo-Inf. Sci. 2025, 27, 256–270. [Google Scholar]
  24. Gao, D.; Li, M.; Fan, D.Z.; Dong, Y.; Li, Z.X.; Wang, R. A ship detection method from lightweight SAR images under complex backgrounds. J. Geo-Inf. Sci. 2024, 26, 2612–2625. [Google Scholar]
  25. Xia, R.; Chen, J.; Huang, Z.; Wan, H.; Wu, B.; Sun, L.; Yao, B.; Xiang, H.; Xing, M. CRTransSar: A Visual Transformer Based on Contextual Joint Representation Learning for SAR Ship Detection. Remote Sens. 2022, 14, 1488. [Google Scholar] [CrossRef]
  26. Qu, H.C.; Shen, L.; Guo, W.; Wang, J.K. Ships Detection in SAR Images Based on Anchor-Free Model With Mask Guidance Features. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 666–675. [Google Scholar] [CrossRef]
  27. Qin, C.; Zhang, L.P.; Wang, X.Q.; Li, G.; He, Y.; Liu, Y.H. RDB-DINO: An Improved End-to-End Transformer With Refined De-Noising and Boxes for Small-Scale Ship Detection in SAR Images. IEEE Trans. Geosci. Remote Sens. 2025, 63, 1–17. [Google Scholar] [CrossRef]
  28. Cao, J.; Han, P.H.; Liang, H.P.; Niu, Y. SFRT-DETR: A SAR ship detection algorithm based on feature selection and multi-scale feature focus. Signal Image Video Process. 2025, 19, 115. [Google Scholar]
  29. Zhang, H.S.; Wu, W.; Xu, J.; Wu, F.; Ji, Y.M. A Ship Detection Method for SAR Images Based on Small Target Feature Enhanced RT-DETR. Comput. Sci. 2025, 1–11. [Google Scholar] [CrossRef]
  30. Kang, M. Study on the Methods of SAR Ship Detection and Recognition Based on Deep Learning. Master’s Thesis, National University of Defense Technology, Changsha, China, 2017. [Google Scholar]
  31. Li, J.Q.; Jiang, Z.J.; Yao, L.B.; Jian, T. A Ship Target Fusion Recognition Algorithm Based on Deep Learning. Ship Electron. Eng. 2020, 40, 31–35. [Google Scholar]
  32. He, J.L. Research on Detection and Classification of Ship Targets in SAR Images. Master’s Thesis, Xidian University, Xi’an, China, 2019. [Google Scholar]
  33. Fu, X.Y.; Zhou, Z.C.; Meng, H.; Li, S.T. A synthetic aperture radar small ship detector based on transformers and multi-dimensional parallel feature extraction. Eng. Appl. Artif. Intell. 2024, 137, 109049. [Google Scholar] [CrossRef]
  34. Jiao, L.C.; Zhao, J. A Survey on the New Generation of Deep Learning in Image Processing. IEEE Access 2019, 7, 172231–172263. [Google Scholar] [CrossRef]
  35. Wang, Y.; Wang, W.; Li, Y.; Jia, Y.D.; Xu, Y.; Ling, Y.; Ma, J.Q. An attention mechanism module with spatial perception and channel information interaction. Complex Intell. Syst. 2024, 10, 5427–5444. [Google Scholar] [CrossRef]
  36. Srinivas, A.; Lin, T.Y.; Parmar, N.; Shlens, J.; Abbeel, P.; Vaswani, A. Bottleneck Transformers for Visual Recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 16519–16529. [Google Scholar]
  37. Zhang, Y.; Ye, M.; Zhu, G.Y.; Liu, Y.; Guo, P.Y.; Yan, J.H. FFCA-YOLO for Small Object Detection in Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2024, 62, 1–15. [Google Scholar] [CrossRef]
  38. Zhang, T.W.; Zhang, X.L.; Ke, X.; Zhan, X.; Shi, J.; Wei, S.J.; Pan, D.C.; Li, J.W.; Su, H.; Zhou, Y.; et al. LS-SSDD-v1.0: A Deep Learning Dataset Dedicated to Small Ship Detection from Large-Scale Sentinel-1 SAR Images. Remote Sens. 2020, 12, 2997. [Google Scholar] [CrossRef]
  39. Wei, S.J.; Zeng, X.; Qu, Q.; Wang, M.; Su, H.; Shi, J. HRSID: A High-Resolution SAR Images Dataset for Ship Detection and Instance Segmentation. IEEE Access 2020, 8, 120234–120254. [Google Scholar] [CrossRef]
  40. Tan, M.; Pang, R.; Le, Q.V. EfficientDet: Scalable and efficient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 10781–10790. [Google Scholar]
  41. Tian, Z.; Shen, C.; Chen, H.; He, T. Fcos: Fully convolutional one-stage object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Long Beach, CA, USA, 15–20 June 2019; pp. 9627–9636. [Google Scholar]
  42. Li, C.; Hei, Y.Q.; Xi, L.H.; Li, W.T.; Xiao, Z. GL-DETR: Global-to-Local Transformers for Small Ship Detection in SAR Images. IEEE Geosci. Remote Sens. Lett. 2024, 21, 1–5. [Google Scholar] [CrossRef]
  43. Wei, S.J.; Su, H.; Ming, J.; Wang, C.; Yan, M.; Kumar, D.; Shi, J.; Zhang, X.L. Precise and robust ship detection for high-resolution SAR imagery based on HR-SDNet. Remote Sens. 2020, 12, 167. [Google Scholar] [CrossRef]
  44. Gao, S.; Liu, J.M.; Miao, Y.H.; He, Z.J. A high-effective implementation of ship detector for SAR images. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
  45. Zhang, S.F.; Chi, C.; Yao, Y.Q.; Lei, Z.; Li, S.Z. Bridging the gap between anchor-based and anchor-free detection via adaptive training sample selection. arXiv 2020, arXiv:1912.02424. [Google Scholar]
  46. Li, Y.X.; Li, X.; Li, W.J.; Hou, Q.B.; Liu, L.; Cheng, M.M.; Yang, J. SARDet-100K: Towards Open-Source Benchmark and ToolKit for Large-Scale SAR Object Detection. arXiv 2025, arXiv:2403.06534. [Google Scholar]
  47. Du, W.T.; Chen, J.; Zhang, C.C.; Zhao, P.; Wan, H.Y.; Zhou, Z.; Cao, Y.C.; Huang, Z.X.; Li, Y.S.; Wu, B.C. SARNas: A hardware-aware SAR target detection algorithm via multi-objective neural architecture search. IEEE Trans. Geosci. Remote Sens. 2023, 61, 1–23. [Google Scholar]
  48. Wang, P.; Luo, Y.; Zhu, Z.L. FDI-YOLO: Feature disentanglement and interaction network based on YOLO for SAR object detection. Expert Syst. Appl. 2025, 260, 125442. [Google Scholar] [CrossRef]
  49. Yang, X.; Zhang, X.; Wang, N.N.; Gao, X.B. A robust one-stage detector for multiscale ship detection with complex background in massive SAR images. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–12. [Google Scholar] [CrossRef]
  50. Wang, H.; Liu, S.L.; Lv, Y.X.; Li, S.Y. Scattering information fusion network for oriented ship detection in SAR images. IEEE Geosci. Remote Sens. Lett. 2023, 20, 1–5. [Google Scholar] [CrossRef]
  51. Cao, R.; Sui, J. A Dynamic Multi-Scale Feature Fusion Network for Enhanced SAR Ship Detection. Sensors 2025, 25, 5194. [Google Scholar] [CrossRef]
  52. Zhu, H.; Li, D.; Wang, H.; Yang, R.; Liang, J.; Liu, S.; Wan, J. A Progressive Saliency-Guided Small Ship Detection Method for Large Scene SAR Images. Remote Sens. 2025, 17, 3085. [Google Scholar] [CrossRef]
  53. Khanam, R.; Hussain, M. What is YOLOv5: A deep look into the internal features of the popular object detector. arXiv 2024, arXiv:2407.20892. [Google Scholar] [CrossRef]
  54. Varghese, R.; Sambath, M. Yolov8: A novel object detection algorithm with enhanced performance and robustness. In Proceedings of the 2024 International Conference on Advances in Data Engineering and Intelligent Computing Systems (ADICS), Chennai, India, 18–19 April 2024; pp. 1–6. [Google Scholar]
  55. Khanam, R.; Hussain, M. Yolov11: An overview of the key architectural enhancements. arXiv 2024, arXiv:2410.17725. [Google Scholar] [CrossRef]
  56. Tian, Y.; Ye, Q.; Doermann, D. Yolov12: Attention-centric real-time object detectors. arXiv 2025, arXiv:2502.12524. [Google Scholar]
  57. Torres, R.; Snoeij, P.; Geudtner, D.; Bibby, D.; Davidson, M.; Attema, E.; Potin, P.; Rommen, B.; Floury, N.; Brown, M.; et al. GMES Sentinel-1 mission. Remote Sens. Environ. 2012, 120, 9–24. [Google Scholar] [CrossRef]
  58. Pelich, R.; Chini, M.; Hostache, R.; Matgen, P.; Lopez-Martinez, C.; Nuevo, M.; Ries, P.; Eiden, G. Large-scale automatic vessel monitoring based on dual-polarization Sentinel-1 and AIS data. Remote Sens. 2019, 11, 1078. [Google Scholar] [CrossRef]
Figure 1. The overall frame structure of this paper.
Figure 1. The overall frame structure of this paper.
Electronics 14 04728 g001
Figure 2. SAR image band reorganization. R, G, and B are the three channels of an image.
Figure 2. SAR image band reorganization. R, G, and B are the three channels of an image.
Electronics 14 04728 g002
Figure 3. Coordinate channel attention mechanism network structure.
Figure 3. Coordinate channel attention mechanism network structure.
Electronics 14 04728 g003
Figure 4. BOT3 network structure.
Figure 4. BOT3 network structure.
Electronics 14 04728 g004
Figure 5. Multi-scale feature enhancement module.
Figure 5. Multi-scale feature enhancement module.
Electronics 14 04728 g005
Figure 6. Multi-Scale Effective Feature Aggregation module.
Figure 6. Multi-Scale Effective Feature Aggregation module.
Electronics 14 04728 g006
Figure 7. Visualization of False Detection Cases in the Detection Model with the MSFE Module. The red box indicates the actual detected target, while the yellow box indicates the ship target falsely detected by the method in this paper.
Figure 7. Visualization of False Detection Cases in the Detection Model with the MSFE Module. The red box indicates the actual detected target, while the yellow box indicates the ship target falsely detected by the method in this paper.
Electronics 14 04728 g007
Figure 8. The detection results of the comparative experiments on the HRSID and LS-SSDD datasets. The red boxes indicate the detection boxes of each algorithm, while the blue boxes mark small-scale ship targets that are easily missed by most comparison algorithms.
Figure 8. The detection results of the comparative experiments on the HRSID and LS-SSDD datasets. The red boxes indicate the detection boxes of each algorithm, while the blue boxes mark small-scale ship targets that are easily missed by most comparison algorithms.
Electronics 14 04728 g008
Figure 9. Accurate detection results of vessel targets in the Gulf of Mexico, USA.
Figure 9. Accurate detection results of vessel targets in the Gulf of Mexico, USA.
Electronics 14 04728 g009
Figure 10. Detection results of ship targets in the Gulf of Mexico region of the United States. The blue circles represent missed detections, and the green ones represent false detections.
Figure 10. Detection results of ship targets in the Gulf of Mexico region of the United States. The blue circles represent missed detections, and the green ones represent false detections.
Electronics 14 04728 g010
Figure 11. Detection results of ship targets in the West Coast area of the United States. The blue circles represent missed detections.
Figure 11. Detection results of ship targets in the West Coast area of the United States. The blue circles represent missed detections.
Electronics 14 04728 g011
Figure 12. Detection results of ship targets in the East Coast area of the United States.
Figure 12. Detection results of ship targets in the East Coast area of the United States.
Electronics 14 04728 g012
Table 1. Some parameters of the LS-SSDD dataset and the HRSID dataset.
Table 1. Some parameters of the LS-SSDD dataset and the HRSID dataset.
ParameterLS-SSDDHRSID
SatelliteSentinel-1Sentinel-1, TerraSAR-X
Sensor ModeIWSM, ST, HS
Resolution 10 × 10 m0.5 m, 1 m, 3 m
PolarizationVV, VHHH, HV, VV
Image Size 800 × 800 pixels 800 × 800 pixels
Table 2. Ablation experiment results on the LS-SSDD dataset. ✓ indicates that the component is included in the model variant. Bold indicates the optimal result.
Table 2. Ablation experiment results on the LS-SSDD dataset. ✓ indicates that the component is included in the model variant. Bold indicates the optimal result.
Yolov8BRCABOT3MEFAMSFEPR mAP 50 mAP 50 : 95 GFLOPsParaFPS
0.8280.6260.7120.2793.08.1625.00
0.8520.6530.7430.2993.08.1669.23
0.8630.6640.7660.30928.511.1370.37
0.8620.6650.7650.30829.112.0639.23
0.8640.6650.7650.30636.215.3556.33
0.8670.6640.7650.30637.214.3400.00
0.8640.6690.760.30711.128.5294.12
0.8710.6740.7720.31112.029.1270.27
0.8680.6730.7580.29744.918.5200.00
0.8720.6760.7500.29645.018.5200.00
0.8800.6810.7800.31315.237.8208.33
0.8670.6820.7810.31619.445.6175.44
0.8510.6440.7440.29345.619.4333.33
+FEM0.8770.6790.7720.31113.632.7250.00
+FEM0.8600.6800.7790.31317.840.5204.08
+WT0.8600.6680.7550.2963.08.1183.56
+PCA0.8570.6270.7310.2913.08.1196.45
Table 3. Ablation experiment results on the HRSID dataset. ✓ indicates that the component is included in the model variant. Bold indicates the optimal result.
Table 3. Ablation experiment results on the HRSID dataset. ✓ indicates that the component is included in the model variant. Bold indicates the optimal result.
Yolov8BRCABOT3MEFAMSFEPR mAP 50 mAP 50 : 95 GFLOPsParaFPS
0.9250.8620.9370.6933.08.1190.31
0.9130.8810.9410.7213.08.1192.31
0.9290.8640.9400.70628.511.1357.14
0.9220.8800.9420.72229.111.9238.10
0.9220.8650.9370.69536.215.3153.85
0.9220.8840.9430.72137.214.3243.90
0.9230.8820.9420.72411.128.5181.82
0.9080.8790.9430.73012.029.1208.33
0.9130.8730.9430.72744.918.5203.34
0.9150.8620.9430.72945.018.5192.31
0.9200.8900.9440.73015.237.8196.08
0.9350.8890.9450.72419.445.6125.00
0.9340.8480.9420.72245.619.4270.27
+FEM0.9170.8790.9430.72713.632.7322.58
+FEM0.9160.8890.9440.72417.840.5132.36
+WT0.9310.8690.9390.69819.445.6134.45
+PCA0.9220.8710.9390.70819.445.6145.55
Table 4. Performance comparison of the proposed method with other methods on the LS-SSDD dataset. Bold indicates the optimal result.
Table 4. Performance comparison of the proposed method with other methods on the LS-SSDD dataset. Bold indicates the optimal result.
MethodPR mAP 50 mAP 50 : 95 GFLOPsParaFPS
Faster R-CNN [12]0.8280.6260.7120.279134.3841.35
Cascade R-CNN [13]0.8520.6530.7430.299162.1869.15
EfficientDet [40]0.8640.6690.760.307107.5239.40
HR-SDNet [43]0.8710.6740.7720.311260.3990.92
SARNet [44]0.5830.8190.762104.2042.60
GL-DETR [42]0.7810.7920.7510.14918.5
MSIF [45]0.8650.7810.751
MSFA [46]0.8580.7120.7540.286
DRGD-YOLO [51]0.7560.30110.42.95
PSG [52]0.73321.5
YOLO-SARSI [19]0.7370.28518.43
YOLOv5s [53]0.8060.7230.7617.12.5454.55
YOLOv8s [54]0.8280.6260.7120.2793.08.1625.00
YOLOv11 [55]0.8210.6220.7030.2566.325.8476.19
YOLOv12 [56]0.8050.6360.7220.2706.325.3434.60
Our Model0.8670.6820.7810.31619.445.6175.44
Table 5. Performance comparison of the proposed method with other methods on the HRSID dataset.
Table 5. Performance comparison of the proposed method with other methods on the HRSID dataset.
MethodPR mAP 50 mAP 50 : 95 GFLOPsParaFPS
Faster R-CNN [12]0.8880.7250.7790.589134.3841.358.9
Cascade R-CNN [13]0.8610.8230.8110.551162.1869.158.9
SARNas [47]0.8920.8830.9081.285.46
GL-DETR [42]0.9140.8750.92116.1
FESAR [20]0.9350.8580.9403.546.8
FCOS [41]0.7760.9020.8770.373170.6050.807.8
SSD [15]0.9410.5840.8200.58287.8024.4013.1
ROSD [49]0.9270.8810.927123.5065.80
SIF-Net [50]0.7990.8270.797
DRGD-YOLO [51]0.9240.8560.9310.69810.42.95
PSG [52]0.8700.83118.1
SSD-YOLO [33]0.9280.8720.93021.307.37
FDI-YOLO [48]0.9620.8190.9090.7107.92.27200
YOLOv5s [53]0.8630.8210.9040.6027.12.5184.45
YOLOv8s [54]0.9250.8620.9370.6213.08.1190.31
YOLOv11 [55]0.9210.8390.9350.6186.325.8202.12
YOLOv12 [56]0.9100.8210.9110.6026.325.3178.34
Our Model0.9230.8890.9450.63119.445.6125.00
Table 6. Quantitative index table of image detection in different regions.
Table 6. Quantitative index table of image detection in different regions.
Image LocationFalse Alarm RateDetection RatePrecisionTPFNFD
West Coast9.6%90.4%90.4%75108
Mexico Gulf12.5%87.5%87.5%28134
East Coast20%80%80%1293
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhou, Y.; Zhu, K.; Guo, H.; Lu, J.; Gong, Z.; Liu, X. Ship Target Detection in SAR Imagery Based on Band Recombination and Multi-Scale Feature Enhancement. Electronics 2025, 14, 4728. https://doi.org/10.3390/electronics14234728

AMA Style

Zhou Y, Zhu K, Guo H, Lu J, Gong Z, Liu X. Ship Target Detection in SAR Imagery Based on Band Recombination and Multi-Scale Feature Enhancement. Electronics. 2025; 14(23):4728. https://doi.org/10.3390/electronics14234728

Chicago/Turabian Style

Zhou, Yi, Kun Zhu, Haitao Guo, Jun Lu, Zhihui Gong, and Xiangyun Liu. 2025. "Ship Target Detection in SAR Imagery Based on Band Recombination and Multi-Scale Feature Enhancement" Electronics 14, no. 23: 4728. https://doi.org/10.3390/electronics14234728

APA Style

Zhou, Y., Zhu, K., Guo, H., Lu, J., Gong, Z., & Liu, X. (2025). Ship Target Detection in SAR Imagery Based on Band Recombination and Multi-Scale Feature Enhancement. Electronics, 14(23), 4728. https://doi.org/10.3390/electronics14234728

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop