Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (122)

Search Parameters:
Keywords = rotated bounding box

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
24 pages, 26435 KB  
Article
Oil and Gas Facility Detection in High-Resolution Remote Sensing Images Based on Oriented R-CNN
by Yuwen Qian, Song Liu, Nannan Zhang, Yuhua Chen, Zhanpeng Chen and Mu Li
Remote Sens. 2026, 18(2), 229; https://doi.org/10.3390/rs18020229 - 10 Jan 2026
Viewed by 175
Abstract
Accurate detection of oil and gas (O&G) facilities in high-resolution remote sensing imagery is critical for infrastructure surveillance and sustainable resource management, yet conventional detectors struggle with severe class imbalance, extreme scale variation, and arbitrary orientation. In this work, we propose OGF Oriented [...] Read more.
Accurate detection of oil and gas (O&G) facilities in high-resolution remote sensing imagery is critical for infrastructure surveillance and sustainable resource management, yet conventional detectors struggle with severe class imbalance, extreme scale variation, and arbitrary orientation. In this work, we propose OGF Oriented R-CNN (Oil and Gas Facility Detection Oriented Region-based Convolutional Neural Network), an enhanced oriented detection model derived from Oriented R-CNN that integrates three improvements: (1) O&G Loss Function, (2) Class-Aware Hard Example Mining (CAHEM) module, and (3) Feature Pyramid Network with Feature Enhancement Attention (FPNFEA). Working in synergy, they resolve the coupled challenges more effectively than any standalone fix and do so without relying on rigid one-to-one matching between modules and individual issues. Evaluated on the O&G facility dataset comprising 3039 high-resolution images annotated with rotated bounding boxes across three classes (well sites: 3006, industrial and mining lands: 692, drilling: 244), OGF Oriented R-CNN achieves a mean average precision (mAP) of 82.9%, outperforming seven state-of-the-art (SOTA) models by margins of up to 27.6 percentage points (pp) and delivering a cumulative gain of +10.5 pp over Oriented R-CNN. Full article
Show Figures

Figure 1

22 pages, 1777 KB  
Article
DP2PNet: Diffusion-Based Point-to-Polygon Conversion for Single-Point Supervised Oriented Object Detection
by Peng Li, Limin Zhang and Tao Qu
Sensors 2026, 26(1), 329; https://doi.org/10.3390/s26010329 - 4 Jan 2026
Viewed by 285
Abstract
Rotated Bounding Boxes (RBBs) for oriented object detection are labor-intensive and time-consuming to annotate. Single-point supervision offers a cost-effective alternative but suffers from insufficient size and orientation information, leading existing methods to rely heavily on complex priors and fixed refinement stages. In this [...] Read more.
Rotated Bounding Boxes (RBBs) for oriented object detection are labor-intensive and time-consuming to annotate. Single-point supervision offers a cost-effective alternative but suffers from insufficient size and orientation information, leading existing methods to rely heavily on complex priors and fixed refinement stages. In this paper, we propose DP2PNet (Diffusion-Point-to-Polygon Network), the first diffusion model-based framework for single-point supervised oriented object detection. DP2PNet features three key innovations: (1) A multi-scale consistent noise generator that replaces manual or external model priors with Gaussian noise, reducing dependency on domain-specific information; (2) A Noise Cross-Constraint module based on multi-instance learning, which selects optimal noise point bags by fusing receptive field matching and object coverage; (3) A Semantic Key Point Aggregator that aggregates noise points via graph convolution to form semantic key points, from which pseudo-RBBs are generated using convex hulls. DP2PNet supports dynamic adjustment of refinement stages without retraining, enabling flexible accuracy optimization. Extensive experiments on DOTA-v1.0 and DIOR-R datasets demonstrate that DP2PNet achieves 53.82% and 53.61% mAP50, respectively, comparable to methods relying on complex priors. It also exhibits strong noise robustness and cross-dataset generalization. Full article
Show Figures

Figure 1

26 pages, 102536 KB  
Article
SPOD-YOLO: A Modular Approach for Small and Oriented Aircraft Detection in Satellite Remote Sensing Imagery
by Jiajian Chen, Pengyu Guo, Yong Liu, Lu Cao, Dechao Ran, Kai Wang, Wei Hu and Liyang Wan
Remote Sens. 2025, 17(24), 3963; https://doi.org/10.3390/rs17243963 - 8 Dec 2025
Viewed by 497
Abstract
The accurate detection of small, densely packed and arbitrarily oriented aircraft in high-resolution remote sensing imagery remains highly challenging due to significant variations in object scale, orientation and background complexity. Existing detection frameworks often struggle with insufficient representation of small objects, instability of [...] Read more.
The accurate detection of small, densely packed and arbitrarily oriented aircraft in high-resolution remote sensing imagery remains highly challenging due to significant variations in object scale, orientation and background complexity. Existing detection frameworks often struggle with insufficient representation of small objects, instability of rotated bounding box regression and inability to adapt to complex background. To address these limitations, we propose SPOD-YOLO, a novel detection framework specifically designed for small aircraft in remote sensing images. This method is based on YOLOv11, combined with the feature attention mechanism of swintransformer, through targeted improvements on cross-scale feature modelling, dynamic convolutional adaptation, and rotational geometry optimization to achieve effective detection. Additionally, we have constructed a new dataset based on satellite remote sensing images, which has high density of small aircraft with rotated bounding box annotations to provide more realistic and challenging evaluation settings. Extensive experiments on MAR20, UCAS-AOD and the constructed dataset demonstrate that our method achieves consistent performance gains over state-of-the-art approaches. SPOD-YOLO achieves an 4.54% increase in mAP50 and a 11.78% gain in mAP50:95 with only 3.77 million parameters on the constructed dataset. These results validate the effectiveness and robustness of our approach in complex remote sensing scenarios, offering a practical advancement for the detection of small objects in aerospace imagery. Full article
Show Figures

Figure 1

20 pages, 8646 KB  
Article
Fine-Grained Multispectral Fusion for Oriented Object Detection in Remote Sensing
by Xin Lan, Shaolin Zhang, Yuhao Bai and Xiaolin Qin
Remote Sens. 2025, 17(22), 3769; https://doi.org/10.3390/rs17223769 - 20 Nov 2025
Cited by 1 | Viewed by 903
Abstract
Infrared–visible-oriented object detection aims to combine the strengths of both infrared and visible images, overcoming the limitations of a single imaging modality to achieve more robust detection with oriented bounding boxes under diverse environmental conditions. However, current methods often suffer from two issues: [...] Read more.
Infrared–visible-oriented object detection aims to combine the strengths of both infrared and visible images, overcoming the limitations of a single imaging modality to achieve more robust detection with oriented bounding boxes under diverse environmental conditions. However, current methods often suffer from two issues: (1) modality misalignment caused by hardware and annotation errors, leading to inaccurate feature fusion that degrades downstream task performance; and (2) insufficient directional priors in square convolutional kernels, impeding robust object detection with diverse directions, especially in densely packed scenes. To tackle these challenges, in this paper, we propose a novel method, Fine-Grained Multispectral Fusion (FGMF), for oriented object detection in the paired aerial images. Specifically, we design a dual-enhancement and fusion module (DEFM) to obtain the calibrated and complementary features through weighted addition and subtraction-based attention mechanisms. Furthermore, we propose an orientation aggregation module (OAM) that employs large rotated strip convolutions to capture directional context and long-range dependencies. Extensive experiments on the DroneVehicle and VEDAI datasets demonstrate the effectiveness of our proposed method, yielding impressive results with accuracies of 80.2% and 66.3%, respectively. These results highlight the effectiveness of FGMF in oriented object detection within complex remote sensing scenarios. Full article
Show Figures

Figure 1

18 pages, 2645 KB  
Article
Advancing YOLOv8-Based Wafer Notch-Angle Detection Using Oriented Bounding Boxes, Hyperparameter Tuning, Architecture Refinement, and Transfer Learning
by Eun Seok Jun, Hyo Jun Sim and Seung Jae Moon
Appl. Sci. 2025, 15(21), 11507; https://doi.org/10.3390/app152111507 - 28 Oct 2025
Viewed by 973
Abstract
Accurate angular alignment of wafers is essential in ion implantation to prevent channeling effects that degrade device performance. This study proposes a real-time notch-angle-detection system based on you only look once version 8 with oriented bounding boxes (YOLOv8-OBB). The proposed method compares YOLOv8 [...] Read more.
Accurate angular alignment of wafers is essential in ion implantation to prevent channeling effects that degrade device performance. This study proposes a real-time notch-angle-detection system based on you only look once version 8 with oriented bounding boxes (YOLOv8-OBB). The proposed method compares YOLOv8 and YOLOv8-OBB, demonstrating the superiority of the latter in accurately capturing rotational features. To enhance detection performance, hyperparameters—including initial learning rate (Lr0), weight decay, and optimizer—are optimized using an one factor at a time (OFAT) approach followed by grid search. Architectural improvements, including spatial pyramid pooling fast with large selective kernel attention (SPPF_LSKA), a bidirectional feature pyramid network (BiFPN), and a high-resolution detection head (P2 head), are incorporated to improve small-object detection. Furthermore, a gradual unfreezing strategy is employed to support more effective and stable transfer learning. The final system is evaluated over 100 training epochs and tracked up to 5000 epochs to verify long-term stability. Compared to baseline models, it achieves higher accuracy and robustness in angle-sensitive scenarios, offering a reliable and scalable solution for high-precision wafer-notch detection in semiconductor manufacturing. Full article
Show Figures

Figure 1

20 pages, 5472 KB  
Article
Research on Indoor 3D Semantic Mapping Based on ORB-SLAM2 and Multi-Object Tracking
by Wei Wang, Ruoxi Wu, Yan Dong and Huilin Jiang
Appl. Sci. 2025, 15(20), 10881; https://doi.org/10.3390/app152010881 - 10 Oct 2025
Viewed by 1296
Abstract
The integration of semantic simultaneous localization and mapping (SLAM) with 3D object detection in indoor scenes is a significant challenge in the field of robot perception. Existing methods typically rely on expensive sensors and lack robustness and accuracy in complex environments. To address [...] Read more.
The integration of semantic simultaneous localization and mapping (SLAM) with 3D object detection in indoor scenes is a significant challenge in the field of robot perception. Existing methods typically rely on expensive sensors and lack robustness and accuracy in complex environments. To address this, this paper proposes a novel 3D semantic SLAM framework that integrates Oriented FAST and Rotated BRIEF-SLAM2 (ORB-SLAM2), 3D object detection, and multi-object tracking (MOT) techniques to achieve efficient and robust semantic environment modeling. Specifically, we employ an improved 3D object detection network to extract semantic information and enhance detection accuracy through category balancing strategies and optimized loss functions. Additionally, we introduce MOT algorithms to filter and track 3D bounding boxes, enhancing stability in dynamic scenes. Finally, we deeply integrate 3D semantic information into the SLAM system, achieving high-precision 3D semantic map construction. Experiments were conducted on the public dataset SUNRGBD and two self-collected datasets (robot navigation and XR glasses scenes). The results show that, compared with the current state-of-the-art methods, our method demonstrates significant advantages in detection accuracy, localization accuracy, and system robustness, providing an effective solution for low-cost, high-precision indoor semantic SLAM. Full article
Show Figures

Figure 1

27 pages, 26151 KB  
Article
EfficientRDet: An EfficientDet-Based Framework for Precise Ship Detection in Remote Sensing Imagery
by Weikang Zuo and Shenghui Fang
Remote Sens. 2025, 17(18), 3160; https://doi.org/10.3390/rs17183160 - 11 Sep 2025
Viewed by 1063
Abstract
Detecting arbitrarily oriented ships in remote sensing images remains challenging due to diverse orientations, complex backgrounds, and scale variations, leading to a struggle in balancing detector accuracy with efficiency. We propose EfficientRDet, an enhanced rotated-ship detection algorithm built upon the EfficientDet framework. EfficientRDet [...] Read more.
Detecting arbitrarily oriented ships in remote sensing images remains challenging due to diverse orientations, complex backgrounds, and scale variations, leading to a struggle in balancing detector accuracy with efficiency. We propose EfficientRDet, an enhanced rotated-ship detection algorithm built upon the EfficientDet framework. EfficientRDet adapts to rotated objects via an angle prediction branch and then significantly boosts accuracy with a novel pseudo-two-stage paradigm comprising a Rotated-Bounding-Box Refinement Branch (RRB) and a Class-Score Refinement Branch (CRB). Further precision is gained through an optimized Enhanced BiFPN (E-BiFPN), an Attention Head, and Distribution Focal (DF) angle representation. Extensive experiments on the HRSC2016 (optical) and RSDD-SAR datasets show that EfficientRDet consistently outperforms state-of-the-art methods, achieving 97.60% AP50 on HRSC2016 and 93.58% AP50 on RSDD-SAR. Comprehensive ablation studies confirm the effectiveness of all proposed mechanisms. EfficientRDet thus offers a promising and practical solution for precise, efficient ship detection across diverse remote sensing imagery. Full article
Show Figures

Figure 1

21 pages, 14861 KB  
Article
Feature Equalization and Hierarchical Decoupling Network for Rotated and High-Aspect-Ratio Object Detection
by Wenbin Gao, Jinda Ji and Donglin Jing
Symmetry 2025, 17(9), 1491; https://doi.org/10.3390/sym17091491 - 9 Sep 2025
Viewed by 805
Abstract
Current mainstream remote sensing target detection algorithms mostly estimate the rotation angle of targets by designing different bounding box descriptions and loss functions. However, they fail to consider the symmetry–asymmetry duality anisotropy in the distribution of key features required for target localization. Moreover, [...] Read more.
Current mainstream remote sensing target detection algorithms mostly estimate the rotation angle of targets by designing different bounding box descriptions and loss functions. However, they fail to consider the symmetry–asymmetry duality anisotropy in the distribution of key features required for target localization. Moreover, the equivalent feature extraction mode of shared convolutional kernels may lead to difficulties in accurately predicting parameters with different attributes, thereby reducing the performance of the detector. In this paper, we propose the Feature Equalization and Hierarchical Decoupling Network (FEHD-Net), which comprises three core components: a Symmetry-Enhanced Parallel Interleaved Convolution Module (PICM), a Parameter Decoupling Module (PDM), and a Critical Feature Matching Loss Function (CFM-Loss). PICM captures diverse spatial features over long distances by integrating square convolution and multi-branch continuous orthogonal large kernel strip convolution sequences, thereby enhancing the network’s capability in processing long-distance spatial information. PDM decomposes feature maps with different properties and assigns them to different regression branches to estimate the parameters of the target’s rotating bounding box. Finally, to stabilize the training of anchors with different qualities that have captured the key features required for detection, CFM-Loss utilizes the intersection ratio between anchors and true value labels, as well as the uncertainty of convolutional regression during training, and designs an alignment criterion (symmetry-aware alignment) to evaluate the regression ability of different anchors. This enables the network to fine-tune the processing of templates with different qualities, achieving stable training of the network. A large number of experiments demonstrate that compared with existing methods, FEHD-Net can achieve state-of-the-art performance on DOTA, HRSC2016, and UCAS-AOD datasets. Full article
(This article belongs to the Special Issue Symmetry and Asymmetry Study in Object Detection)
Show Figures

Figure 1

28 pages, 5678 KB  
Article
Enhanced YOLOv8 with DWR-DRB and SPD-Conv for Mechanical Wear Fault Diagnosis in Aero-Engines
by Qifan Zhou, Bosong Chai, Chenchao Tang, Yingqing Guo, Kun Wang, Xuan Nie and Yun Ye
Sensors 2025, 25(17), 5294; https://doi.org/10.3390/s25175294 - 26 Aug 2025
Cited by 5 | Viewed by 1514
Abstract
Aero-engines, as complex systems integrating numerous rotating components and accessory equipment, operate under harsh and demanding conditions. Prolonged use often leads to frequent mechanical wear and surface defects on accessory parts, which significantly compromise the engine’s normal and stable performance. Therefore, accurately and [...] Read more.
Aero-engines, as complex systems integrating numerous rotating components and accessory equipment, operate under harsh and demanding conditions. Prolonged use often leads to frequent mechanical wear and surface defects on accessory parts, which significantly compromise the engine’s normal and stable performance. Therefore, accurately and rigorously identifying failure modes is of critical importance. In this study, failure modes are categorized into notches, scuffs, and scratches based on original bearing structure images. The YOLOv8 architecture is adopted as the base framework, and a Dilated Reparameterization Block (DRB) is introduced to enhance the Dilation-Wise Residual (DWR) module. This structure uses a large convolutional kernel to capture fragmented and sparse features in wear images, ensuring a wide receptive field. The concept of structural reparameterization is incorporated into DWR to improve its ability to capture detailed target information. Additionally, the standard convolutional layer in the head of the improved DWR-DRB structure is replaced by Spatial-Depth Convolution (SPD-Conv) to reduce the loss of wear morphology and enhance the accuracy of fault feature extraction. Finally, a fusion structure combining Focaler and MPDIoU is integrated into the loss function to leverage their strengths in handling imbalanced classification and bounding box geometric regression. The proposed method achieves effective recognition and diagnosis of mechanical wear fault patterns. The experimental results demonstrate that, compared to the baseline YOLOv8, the proposed method improves the mAP50 for fault diagnosis and recognition from 85.4% to 91%. Full article
(This article belongs to the Section Fault Diagnosis & Sensors)
Show Figures

Figure 1

17 pages, 3827 KB  
Article
A Deep Learning Approach to Teeth Segmentation and Orientation from Panoramic X-Rays
by Mou Deb, Madhab Deb and Mrinal Kanti Dhar
Signals 2025, 6(3), 40; https://doi.org/10.3390/signals6030040 - 8 Aug 2025
Cited by 3 | Viewed by 3716
Abstract
Accurate teeth segmentation and orientation are fundamental in modern oral healthcare, enabling precise diagnosis, treatment planning, and dental implant design. In this study, we present a comprehensive approach to teeth segmentation and orientation from panoramic X-ray images, leveraging deep-learning techniques. We built an [...] Read more.
Accurate teeth segmentation and orientation are fundamental in modern oral healthcare, enabling precise diagnosis, treatment planning, and dental implant design. In this study, we present a comprehensive approach to teeth segmentation and orientation from panoramic X-ray images, leveraging deep-learning techniques. We built an end-to-end instance segmentation network that uses an encoder–decoder architecture reinforced with grid-aware attention gates along the skip connections. We introduce oriented bounding box (OBB) generation through principal component analysis (PCA) for precise tooth orientation estimation. Evaluating our approach on the publicly available DNS dataset, comprising 543 panoramic X-ray images, we achieve the highest Intersection-over-Union (IoU) score of 82.43% and a Dice Similarity Coefficient (DSC) score of 90.37% among compared models in teeth instance segmentation. In OBB analysis, we obtain the Rotated IoU (RIoU) score of 82.82%. We also conduct detailed analyses of individual tooth labels and categorical performance, shedding light on strengths and weaknesses. The proposed model’s accuracy and versatility offer promising prospects for improving dental diagnoses, treatment planning, and personalized healthcare in the oral domain. Full article
Show Figures

Figure 1

21 pages, 15647 KB  
Article
Research on Oriented Object Detection in Aerial Images Based on Architecture Search with Decoupled Detection Heads
by Yuzhe Kang, Bohao Zheng and Wei Shen
Appl. Sci. 2025, 15(15), 8370; https://doi.org/10.3390/app15158370 - 28 Jul 2025
Cited by 2 | Viewed by 2135
Abstract
Object detection in aerial images can provide great support in traffic planning, national defense reconnaissance, hydrographic surveys, infrastructure construction, and other fields. Objects in aerial images are characterized by small pixel–area ratios, dense arrangements between objects, and arbitrary inclination angles. In response to [...] Read more.
Object detection in aerial images can provide great support in traffic planning, national defense reconnaissance, hydrographic surveys, infrastructure construction, and other fields. Objects in aerial images are characterized by small pixel–area ratios, dense arrangements between objects, and arbitrary inclination angles. In response to these characteristics and problems, we improved the feature extraction network Inception-ResNet using the Fast Architecture Search (FAS) module and proposed a one-stage anchor-free rotation object detector. The structure of the object detector is simple and only consists of convolution layers, which reduces the number of model parameters. At the same time, the label sampling strategy in the training process is optimized to resolve the problem of insufficient sampling. Finally, a decoupled object detection head is used to separate the bounding box regression task from the object classification task. The experimental results show that the proposed method achieves mean average precision (mAP) of 82.6%, 79.5%, and 89.1% on the DOTA1.0, DOTA1.5, and HRSC2016 datasets, respectively, and the detection speed reaches 24.4 FPS, which can meet the needs of real-time detection. Full article
(This article belongs to the Special Issue Innovative Applications of Artificial Intelligence in Engineering)
Show Figures

Figure 1

20 pages, 4423 KB  
Article
Pointer Meter Reading Recognition Based on YOLOv11-OBB Rotated Object Detection
by Xing Xu, Liming Wang, Chunhua Deng and Bi He
Appl. Sci. 2025, 15(13), 7460; https://doi.org/10.3390/app15137460 - 3 Jul 2025
Viewed by 1638
Abstract
In the domain of intelligent inspection, the precise recognition of pointer meter readings is of paramount importance for monitoring equipment conditions. To address the challenges of insufficient robustness and diminished detection accuracy encountered in practical applications of existing methods for recognizing pointer meter [...] Read more.
In the domain of intelligent inspection, the precise recognition of pointer meter readings is of paramount importance for monitoring equipment conditions. To address the challenges of insufficient robustness and diminished detection accuracy encountered in practical applications of existing methods for recognizing pointer meter readings based on object detection, we propose a novel approach that integrates YOLOv11-OBB rotating object detection with adaptive template matching techniques. Firstly, the YOLOv11 object detection algorithm is employed, incorporating a rotational bounding box (OBB) detection mechanism; This effectively enhances the feature extraction capabilities related to pointer rotation direction and dial center, thereby boosting detection robustness. Subsequently, an enhanced angle resolution algorithm is leveraged to develop a mapping model that establishes a relationship between pointer the deflection angle and the instrument range, facilitating precise reading calculation. Experimental findings demonstrate that the proposed method achieves a mean Average Precision (mAP) of 99.1% in a self-compiled pointer instrument dataset. The average relative error of readings is 0.41568%, with a maximum relative error of less than 1.1468%. Furthermore, the method exhibits robustness and reliability when handling low-quality meter images characterized by blur, darkness, overexposure, and tilt. The proposed approach provides a highly adaptable and reliable solution for pointer meter reading recognition in the intelligent industrial field, with significant practical value. Full article
Show Figures

Figure 1

22 pages, 11308 KB  
Article
TIAR-SAR: An Oriented SAR Ship Detector Combining a Task Interaction Head Architecture with Composite Angle Regression
by Yu Gu, Minding Fang and Dongliang Peng
Remote Sens. 2025, 17(12), 2049; https://doi.org/10.3390/rs17122049 - 13 Jun 2025
Cited by 3 | Viewed by 1127
Abstract
Oriented ship detection in Synthetic Aperture Radar (SAR) images has broad applications in maritime surveillance and other fields. While deep learning advancements have significantly improved ship detection performance, persistent challenges remain for existing methods. These include the inherent misalignment between regression and classification [...] Read more.
Oriented ship detection in Synthetic Aperture Radar (SAR) images has broad applications in maritime surveillance and other fields. While deep learning advancements have significantly improved ship detection performance, persistent challenges remain for existing methods. These include the inherent misalignment between regression and classification tasks and the boundary discontinuity problem in oriented object detection. These issues hinder efficient and accurate ship detection in complex scenarios. To address these challenges, we propose TIAR-SAR, a novel oriented SAR ship detector featuring a task interaction head and composite angle regression. First, we propose a task interaction detection head (Tihead) capable of predicting both oriented bounding boxes (OBBs) and horizontal bounding boxes (HBBs) simultaneously. Within the Tihead, a “decompose-then-interact” structure is designed. This structure not only mitigates feature misalignment but also promotes feature interaction between regression and classification tasks, thereby enhancing prediction consistency. Second, we propose a joint angle refinement mechanism (JARM). The JARM addresses the non-differentiability problem of the traditional rotated Intersection over Union (IoU) loss through the design of a composite angle regression loss (CARL) function, which strategically combines direct and indirect angle regression methods. A boundary angle correction mechanism (BACM) is then designed to enhance angle estimation accuracy. During inference, BACM dynamically replaces an object’s OBB prediction with its corresponding HBB if the OBB exhibits excessive angle deviation when the angle of the object is near the predefined boundary. Finally, the performance and applicability of the proposed methods are evaluated through extensive experiments on multiple public datasets, including SRSDD, HRSID, and DOTAv1. Experimental results derived from the use of the SRSDD dataset demonstrate that the mAP50 of the proposed method reaches 63.91%, an improvement of 4.17% compared with baseline methods. The detector achieves 17.42 FPS on 1024 × 1024 images using an RTX 2080 Ti GPU, with a model size of only 21.92 MB. Comparative experiments with other state-of-the-art methods on the HRSID dataset demonstrate the proposed method’s superior detection performance in complex nearshore scenarios. Furthermore, when further tested on the DOTAv1 dataset, the mAP50 can reach 79.1%. Full article
Show Figures

Figure 1

21 pages, 6270 KB  
Article
Cross-Level Adaptive Feature Aggregation Network for Arbitrary-Oriented SAR Ship Detection
by Lu Qian, Junyi Hu, Haohao Ren, Jie Lin, Xu Luo, Lin Zou and Yun Zhou
Remote Sens. 2025, 17(10), 1770; https://doi.org/10.3390/rs17101770 - 19 May 2025
Cited by 4 | Viewed by 922
Abstract
The rapid progress of deep learning has significantly enhanced the development of ship detection using synthetic aperture radar (SAR). However, the diversity of ship sizes, arbitrary orientations, densely arranged ships, etc., have been hindering the improvement of SAR ship detection accuracy. In response [...] Read more.
The rapid progress of deep learning has significantly enhanced the development of ship detection using synthetic aperture radar (SAR). However, the diversity of ship sizes, arbitrary orientations, densely arranged ships, etc., have been hindering the improvement of SAR ship detection accuracy. In response to these challenges, this study introduces a new detection approach called a cross-level adaptive feature aggregation network (CLAFANet) to achieve arbitrary-oriented multi-scale SAR ship detection. Specifically, we first construct a hierarchical backbone network based on a residual architecture to extract multi-scale features of ship objects from large-scale SAR imagery. Considering the multi-scale nature of ship objects, we then resort to the idea of self-attention to develop a cross-level adaptive feature aggregation (CLAFA) mechanism, which can not only alleviate the semantic gap between cross-level features but also improve the feature representation capabilities of multi-scale ships. To better adapt to the arbitrary orientation of ship objects in real application scenarios, we put forward a frequency-selective phase-shifting coder (FSPSC) module for arbitrary-oriented SAR ship detection tasks, which is dedicated to mapping the rotation angle of the object bounding box to different phases and exploits frequency-selective phase-shifting to solve the periodic ambiguity problem of the rotated bounding box. Qualitative and quantitative experiments conducted on two public datasets demonstrate that the proposed CLAFANet achieves competitive performance compared to some state-of-the-art methods in arbitrary-oriented SAR ship detection. Full article
Show Figures

Figure 1

19 pages, 8533 KB  
Article
Rotation-Invariant Feature Enhancement with Dual-Aspect Loss for Arbitrary-Oriented Object Detection in Remote Sensing
by Zhao Hu, Xiangfu Meng, Xinsong Liu and Zhuxiang Sun
Appl. Sci. 2025, 15(10), 5240; https://doi.org/10.3390/app15105240 - 8 May 2025
Viewed by 1451
Abstract
Object detection in remote sensing imagery plays a pivotal role in various applications, including aerial surveillance and urban planning. Despite its significance, the task remains challenging due to cluttered backgrounds, the arbitrary orientations of objects, and substantial scale variations across targets. To address [...] Read more.
Object detection in remote sensing imagery plays a pivotal role in various applications, including aerial surveillance and urban planning. Despite its significance, the task remains challenging due to cluttered backgrounds, the arbitrary orientations of objects, and substantial scale variations across targets. To address these issues, we proposed RFE-FCOS, a novel framework that synergizes rotation-invariant feature extraction with adaptive multi-scale fusion. Specifically, we introduce a rotation-invariant learning (RIL) module, which employs adaptive rotation transformations to enhance shallow feature representations, thereby effectively mitigating interference from complex backgrounds and boosting geometric robustness. Furthermore, a rotation feature fusion (RFF) module propagates these rotation-aware features across hierarchical levels through an attention-guided fusion strategy, resulting in richer, more discriminative representations at multiple scales. Finally, we propose a novel dual-aspect RIoU loss (DARIoU) that simultaneously optimizes horizontal and angular regression tasks, facilitating stable training and the precise alignment of arbitrarily oriented bounding boxes. Evaluated on the DIOR-R and HRSC2016 benchmarks, our method demonstrates robust detection capabilities for arbitrarily oriented objects, achieving competitive performance in both accuracy and efficiency. This work provides a versatile solution for advancing object detection in real-world remote sensing scenarios. Full article
Show Figures

Figure 1

Back to TopTop