Next Article in Journal
Spatial Downscaling of Soil Moisture Product to Generate High-Resolution Data: A Multi-Source Approach over Heterogeneous Landscapes in Kenya
Previous Article in Journal
Comparison of a Continuous Forest Inventory to an ALS-Derived Digital Inventory in Washington State
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

TMBO-AOD: Transparent Mask Background Optimization for Accurate Object Detection in Large-Scale Remote-Sensing Images

1
College of Computer Science and Technology, Harbin Engineering University, Nantong Street, Harbin 150001, China
2
National Engineering Laboratory for E-Government Modeling and Simulation, Harbin Engineering University, Nantong Street, Harbin 150001, China
3
Defense Innovation Institute, Chinese Academy of Military Science, Beijing 100071, China
4
Intelligent Game and Decision Laboratory, Chinese Academy of Military Science, Beijing 100071, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2025, 17(10), 1762; https://doi.org/10.3390/rs17101762
Submission received: 12 April 2025 / Revised: 13 May 2025 / Accepted: 15 May 2025 / Published: 18 May 2025
(This article belongs to the Section Remote Sensing Image Processing)

Abstract

Recent advancements in deep-learning and computer vision technologies, coupled with the availability of large-scale remote-sensing image datasets, have accelerated the progress of remote-sensing object detection. However, large-scale remote-sensing images typically feature extensive and complex backgrounds with small and sparsely distributed objects, which pose significant challenges to detection performance. To address this, we propose a novel framework for accurate object detection, termed transparent mask background optimization for accurate object detection (TMBO-AOD), which incorporates a clear focus module and an adaptive filtering framework. The clear focus module constructs an empirical background pool using a Gaussian distribution and introduces transparent masks to prepare for subsequent optimization stages. The adaptive filtering framework can be applied to anchor-based or anchor-free models. It dynamically adjusts the number of candidates generated based on background flags, thereby optimizing the label assignment process. This approach not only alleviates the imbalance between positive and negative samples but also enhances the efficiency of candidate generation. Furthermore, we introduce a novel separated loss function that strengthens both foreground and background consistencies. Specifically, it focuses the model’s attention on foreground objects while enabling it to learn the consistency of background features, thus improving its ability to distinguish objects from the background. We employ YOLOv8 combined with our proposed optimizations to evaluate our model in many datasets, demonstrating improvements in both accuracy and efficiency. Additionally, we validate the effectiveness of our adaptive filtering framework in both anchor-based and anchor-free methods. When implemented with YOLOv5 (anchor based), the framework reduces the candidate generation time by 48.36%, while the YOLOv8 (anchor-free) implementation achieves a 46.81% reduction, both with maintained detection accuracy.

1. Introduction

Remote-sensing object detection plays critical roles in a wide range of real-world applications, including disaster response, environmental monitoring, and urban development. However, detecting objects accurately and efficiently in large-scale remote-sensing imagery remains a formidable challenge. These challenges can be broadly encapsulated in three defining characteristics—large and complex backgrounds, limited target sizes, and low target densities—collectively referred to as the “3L” problem. As illustrated in Figure 1, this phenomenon is pervasive across prominent benchmark datasets, such as AI-TODv2.0 [1], DIOR [2], DOTA-v1.5 [3], and NWPU VHR-10 [4]. Typically, foreground objects occupy less than 2% of the total image area, resulting in extreme class imbalance and making models highly susceptible to background interference. This imbalance propagates two core problems: (1) Background regions often mimic target features, thereby confusing the detector, and (2) detection pipelines waste computational resources on these dominant background regions, particularly under tile-based processing strategies required for gigapixel-scale imagery.
To alleviate such inefficiencies, prior work has employed techniques such as attention mechanisms and class-balancing strategies—including online hard example mining (OHEM) [5], spatial OHEM (S-OHEM) [6], focal loss [7], the gradient-harmonizing mechanism (GHM) [8], and PAA-RPN [9]. Although partially effective, these methods are fundamentally constrained by their downstream position in the detection pipeline, only acting after a large pool of often irrelevant candidates has already been generated. This reactive nature severely limits their scalability and effectiveness, especially in high-resolution imagery, where background content dominates.
In this paper, we propose a novel detection framework, termed TMBO-AOD (transparent mask background optimization for accurate object detection), which introduces three synergistic modules designed to optimize the detection process at both the structural and algorithmic levels. First, the clear focus module leverages an empirically derived background knowledge pool to segment input images into foreground and background regions using flag indicators. Background regions are suppressed via transparent masking, thereby reducing feature interference while preserving critical foreground information. Second, the adaptive filtering framework improves the candidate generation efficiency by dynamically adjusting the number of proposals based on the background density, leading to significant reductions in the candidate volume—48.36% for YOLOv5 and 46.81% for YOLOv8. Third, we introduce the separation loss function, which jointly optimizes foreground enhancement and background consistency, encouraging the model to emphasize discriminative object features while maintaining invariance to redundant background patterns.
Unlike most methods that rely solely on preprocessing strategies, such as image-partitioning and -overlapping segmentation [10,11,12,13,14], our clear focus module provides an orthogonal solution that eliminates the risk of object truncation while enhancing small-object feature learning. Conventional segmentation-based approaches are limited in their ability to disentangle useful features from vast background noise—a limitation we directly address through background-aware masking.
TMBO-AOD is compatible with both anchor-based and anchor-free detection paradigms. Anchor-based methods, such as YOLO [15,16,17], SSD [18], and Faster R-CNN [19], benefit from predefined anchor priors but often require extensive parameter tuning and are computationally expensive [4,12,13,14,20,21,22,23,24,25,26,27,28,29]. In contrast, anchor-free methods—including CornerNet [30], CenterNet [31,32,33], FoveaBox [34], ExtremeNet [35], RepPoints [36], CSP [37], FCOS [38], and others applied to remote sensing [39,40,41,42,43]—simplify the pipeline by predicting object attributes directly. However, they often struggle with background interference and class imbalance. Our adaptive filtering framework bridges this gap, improving the candidate efficiency and optimizing the label assignment for both detection paradigms.
Moreover, the loss function’s design has played a pivotal role in advancing the object detection performance in remote sensing. Traditional IoU-based losses, such as GIoU [44], DIoU [45], CIoU [46], EIoU [47], WIoU [48], and SIoU [49], have been widely adopted to refine bounding box regression. Specialized loss functions for remote sensing, including the gradient calibration loss (GCL) [50], circular smooth label (CSL) [51], and discriminative distribution loss [52], have further improved orientation and localization accuracies. Other domain-specific approaches, like Redet [53], RFLA [54], ClusDet [55], and CDMNet [56], tackle challenges related to object size and density. In contrast to these methods, our separation loss function introduces a novel dual-component formulation that not only enhances the learning of foreground features but also enforces consistency across background regions, effectively filtering out distractive signals during the optimization.
In summary, our contributions are as follows:
(1) We propose the clear focus module, which constructs an empirical background pool and applies transparent masking guided by flag indicators, significantly reducing the effect of the background interference on object detection;
(2) We introduce the adaptive filtering framework, which adaptively selects candidate numbers and optimizes the label assignment based on background flags, improving the candidate efficiency in both anchor-based and anchor-free models;
(3) We present the separation loss function, which enhances model’s attention to foreground objects while learning consistent background features, yielding an over 48% reduction in candidate proposals and consistent accuracy improvements;
(4) Extensive experiments using AI-TODv2.0, DIOR, DOTA-v1.5, and NWPU VHR-10 demonstrate the generalizability and efficiency of TMBO-AOD across multiple detector backbones, achieving mAP gains of 0.3–7.1% while reducing the computational overhead.

2. Methodology

TMBO-AOD is composed of two primary components: the clear focus module and the adaptive filtering framework. As depicted in Figure 2, the clear focus module (Figure 2a) segments the input image into a predefined number of sub-blocks. Each sub-block is then compared to an empirical background pool. Sub-blocks identified as the background are assigned transparent masks to mitigate the influence of the background features. The adaptive filtering framework (Figure 2b) seamlessly integrates with both anchor-free and anchor-based methods, facilitating feature extraction and fusion. The adaptive filtering framework dynamically creates a varying number of candidate frames by analyzing the background and foreground sub-blocks that have been processed by the transparent masks. This process ensures a balance between positive and negative samples, thereby enhancing the optimization of the label assignment process. It is worth noting that the concept of an anchor does not exist in the anchorless model, in which case, an anchor is analogous to a proposal. For the sake of the uniformity of the expression, we unify anchors or proposals as candidates. Ultimately, TMBO-AOD yields accurate detection results.

2.1. Clear Focus Module

The clear focus module (CFM) comprises two main components: an empirical background pool and a transparent mask generation module.

2.1.1. Empirical Background Pool

As illustrated in Figure 3, our study selected a total of one hundred representative images, with twenty-five images from each of the four predominant remote-sensing background categories: ocean, land, forest, and rock. This carefully curated selection achieves an optimal balance between diversity coverage and computational practicality. Although expanding the dataset could marginally improve the statistical precision, our empirical analysis confirms that the one hundred images in these four basic categories effectively capture the essential background characteristics in most remote-sensing scenarios.
The RGB color space directly encodes the intensities of the red, green, and blue channels but is highly sensitive to lighting variations. In contrast, the HSV color space provides better color consistency under changing illumination conditions. To mitigate illumination effects, we converted the collected RGB background images to HSV background images and computed the means and standard deviations of the hue (H) and saturation (S) channels, excluding the brightness channel to further reduce the sensitivity to lighting changes. These statistical values were then used to construct Gaussian distributions, which form the empirical background pool (EBP).
Because the background pool is highly data dependent, it must be recalculated when applied to a different dataset to ensure accuracy. For remote-sensing imagery, the existing four background types are typically sufficient. However, for datasets with different scene compositions, resampling is necessary to capture dataset-specific background characteristics. Fortunately, although resampling is required, the computation of statistical values remains straightforward. Moreover, leveraging prior background classification can significantly improve the efficiency of remote-sensing image interpretation. By first identifying the dominant background type in a given image, the number of unnecessary computations in irrelevant regions can be reduced, allowing detection algorithms to focus on areas with higher probabilities of containing targets. This background-aware processing strategy not only accelerates inferences but also enhances detection accuracy by mitigating false positives in uniform background regions. The formulae for computing the mean and variance from the H and S channels are as follows:
μ ¯ h = 1 M j = 1 M ( 1 N i = 1 N h i )
σ ¯ h = 1 M j = 1 M ( 1 N i = 1 N ( h i μ ¯ h ) )
μ ¯ s = 1 M j = 1 M ( 1 N i = 1 N s i )
σ ¯ s = 1 M j = 1 M ( 1 N i = 1 N ( s i μ ¯ s ) )
where M represents the number of images, N represents the total number of pixels in the channel, h i represents the value of the ith pixel in the Hth channel, s i represents the value of the ith pixel in the Sth channel, μ ¯ h represents the mean value of the Hth channel, σ ¯ h represents the standard deviation of the H-th channel, μ ¯ s represents the mean value of the Sth channel, and σ ¯ s represents the standard deviation of the S-th channel.
The formulae for establishing Gaussian distributions for H and V for the four backgrounds are as follows:
X N ( μ ¯ X , σ ¯ X 2 ) , X { H ocean , H land , H forest , H rock }
Y N ( μ ¯ Y , σ ¯ Y 2 ) , Y { S ocean , S land , S forest , S rock }
where X and Y are random variables representing the hue and saturation, respectively. H ocean , H land , H forest , H rock , S ocean , S land , S forest , and S rock represent the Gaussian distributions of the H characteristics and S characteristics of the four types of backgrounds: ocean, land, forest, and reef, respectively.
The Gaussian distributions of the H and S channels for the four types of backgrounds are stored in the EBP. The input image will be divided into several sub-blocks based on a uniform rule. Specifically, the shorter side of the image is rounded up to 100 (with any excess not processed). This dimension is then divided by two and four to derive the side lengths of the square sub-blocks. The image will subsequently be partitioned into these square sub-blocks, while regions that cannot be fully covered by the sub-blocks will not be processed. In addition, the coordinates of the four vertices of each sub-block in the original image will be recorded. When the target image is divided into sub-blocks, the H and S channel means and variances of each sub-block are calculated to construct their respective Gaussian distributions (similar to those in (1)–(6)) as I H and I S . These distributions are then compared with those in the EBP by computing the Wasserstein distance. The shortest average Wasserstein distance across the H and S channels is denoted as δ and is calculated using the following formulae:
W ( I H , X ) = | μ ¯ X μ I H | + σ ¯ X 2 + σ I H 2
W ( I S , Y ) = | μ ¯ Y μ I S | + σ ¯ Y 2 + σ I S 2
δ = m i n ( W ( I H , X ) + W ( I S , Y ) 2 ) x
where I H and I S are random variables based on the Gaussian distribution representing the characteristics of the sub-block in the H and S channels, respectively; W is the Wasserstein distance; and δ is the average distance obtained by comparing the Gaussian distribution of the image’s sub-block with those of the four backgrounds in the EBP, the shorter the distance, the more similar the sub-block is to a particular background. Sub-blocks with δ < 1 (an empirically determined threshold) are classified as background and assigned a corresponding flag, where 0 represents the background and 1 represents the foreground.

2.1.2. Transparent Mask

While comparing the image’s sub-blocks with those of the empirical background pool, a transparent mask is applied to the sub-blocks to minimize the interference of the background features in the target detection. The traditional approach directly adds a transparent channel based on semantic understanding, which can hinder compatibility with various target detection methods. However, we simulate transparency by superimposing a grayscale image onto the original image and adjusting its brightness using gamma correction. The gamma correction formula is as follows:
O = I γ
where I is the pixel value in the input image, o is the pixel value after the gamma correction, and γ is the gamma value, the higher the value, the brighter the image. Specifically, when δ < 0.1 , indicating a high probability of the sub-block being a background block, the luminance value corresponding to γ = 3 is applied. High-brightness grayscale maps can attenuate background features. For 0.1 < δ 0.2 , γ is set at 1.8. There is a medium likelihood that the δ value in this range corresponds to the background, which will be superimposed onto a medium-brightness grayscale map. Sub-blocks of δ outside these ranges have a high probability of being the foreground, so no transparency mask is applied. Through this process, the resulting image is effectively refined to focus on the target regions, enhancing the detection performance.

2.2. Adaptive Filtering Framework

The adaptive filtering framework (AFF) and the CFM (Section 2.1) together form the core of the TMBO-AOD architecture. This two-stage filtering system first reduces the background interference at the pixel level through the CFM and then optimizes candidate generation at the detection level using the AFF, significantly enhancing the precision and speed of the target localization.
The CFM is primarily responsible for background-aware filtering, suppressing background noise and enhancing the foreground signal. Building upon this, the AFF broadens the focus by optimizing candidate generation at the detection level. This involves filtering out irrelevant background proposals and refining the selection of high-quality candidates, ensuring that only the best candidates are passed to the subsequent classification and regression stages.
Anchor-based and anchor-free detection paradigms fundamentally differ in their approaches to candidate generation and label assignment. Anchor-based methods rely on a predefined set of anchors with fixed sizes and aspect ratios, which provide a structured and straightforward approach for localizing objects. However, this rigidity can limit flexibility, as these fixed anchors may struggle to adapt to objects of varying scales and shapes. In contrast, anchor-free methods avoid using fixed anchors entirely, instead generating dynamic “virtual anchors” (commonly referred to as proposals) directly from the feature maps. This allows for more adaptive and context-aware candidate generation, which can be especially beneficial when dealing with complex or highly variable objects.
Given these fundamental differences, our AFF framework accommodates both paradigms, each with its unique candidate generation strategy. To standardize the terminology in this paper, we refer to the candidate boxes used for predicting the positions and categories of objects as “candidates”, which encompass both the fixed anchors used in anchor-based methods and the dynamically generated proposals of anchor-free approaches. Below, we describe the implementation of the AFF for both detection paradigms, highlighting how background flags are leveraged to optimize candidate generation and improve the detection performance.

2.2.1. Anchor-Based Methods

The most representative anchor-based methods include the single-stage YOLO series [15] and the two-stage faster R-CNN series [19]. Both utilize anchors of varying sizes and ratios to guide detection models in learning true object bounding boxes, although the anchor generation strategies differ. As shown in Figure 4a, for the last feature map layer in the detection, YOLO uses preset anchor sizes, and if these sizes deviate significantly from the true values, clustering is applied to regenerate fixed-sized anchors, which are then used across all the grid cells in the feature map. Faster R-CNN employs a regional proposal network (RPN) to generate anchors using a sliding window approach with three scales and three aspect ratios. As shown in Figure 4c, in our adaptive filtering detection framework, the background flag of the sub-block in the region where the center of the grid cell or the center of the window is located is determined before generating anchor points. Using this information, the number of anchors is dynamically adjusted: If the background flag is 0 (indicating a high probability of being the background), the number of anchors is reduced (e.g., from nine to six). If the background flag is 1, the anchor count remains unchanged. Both anchor-based methods classify positive and negative samples based on IoU thresholds. Most anchors generated in background regions are assigned as negative samples. However, with adaptive filtering, the number of negative samples is significantly reduced, addressing the issue of imbalanced positive and negative samples and effectively improving the velocity of the model’s operation.

2.2.2. Anchor-Free Methods

Our framework also offers a potential label assignment strategy for anchor-free methods. Taking YOLOv8 as an example, its label assignment is based on the task-aligned one-stage object detection (TOOD) [57] strategy. As illustrated in Figure 4b, YOLOv8 treats each grid in the feature map as a proposal and predicts the offsets of the four bounding box edges relative to the grid center. Positive samples are defined as those meeting the following criteria: The predicted box center lies within the ground truth (GT), and the IoU exceeds a predefined threshold. A ranking score, t = S a U b , is used, where S a represents the category prediction score, U b is the IoU value, and a and b are hyperparameters. The top-K positive samples (default K = 13) are selected, while others are treated as negative samples. If multiple GTs correspond to a single proposal, the GT with the highest IoU is selected. Prior to this process, our AFF enhances the allocation of positive and negative samples using background flags, as shown in Figure 4d. Specifically, if a proposal originates from a grid with a background flag of 0, a method similar to “dropout” is applied to probabilistically discard the proposal. This approach effectively reduces the number of negative samples without causing imbalances because of excessive sample elimination.
Together with the CFM, the AFF forms the backbone of the TMBO-AOD architecture, providing a two-layer filtering mechanism that first removes irrelevant background information and then refines candidate selection. This dual-level optimization directly contributes to the high detection accuracy observed in our experiments, especially in complex background scenarios, where conventional methods struggle.

2.3. Separable Loss Function

The separable loss function (SLF) is a classification loss specifically designed to enhance the performance of object detection models by explicitly distinguishing between background and foreground regions, thereby enabling the model to learn to differentiate between these regions more effectively. This distinction contributes to both improving feature consistency and enhancing detection accuracy for both foreground objects and background features. The loss function integrates the focal loss ( F L ) for foreground classification with a custom background loss ( L o s s B G ) to more effectively address background regions. The formulae for the classification loss ( L o s s c l s ) and background loss are presented below:
L o s s c l s = i ( y i · λ 0 · F L ( p i , y i ) ) + ( 1 y i ) · ( λ 1 · L o s s B G ( f i ) )
where p i is the predicted probability for the target class, y i indicates whether the detection box is in a foreground region ( y i = 1 ) or a background region ( y i = 0 ), and λ 0 and λ 1 are weights controlling the balance between foreground and background learning.

2.3.1. Foreground Loss ( y i = 1 )

In foreground regions, the SLF uses the focal loss to focus on challenging samples and reduce the influence of easily classified background elements as follows:
F L ( p i , y i ) = ( 1 p i ) θ log ( p i ) if y i = 1
The Focal Loss ( F L ( p i , y i ) ) is effective for addressing class imbalance by assigning higher importance to hard-to-classify samples, which is particularly useful in complex remote-sensing imagery.

2.3.2. Background Loss ( y i = 0 )

In background regions, the SLF incorporates a background loss that measures the similarity between the feature vector ( f i ) of each region and a reference background vector ( f ¯ B G ) from the empirical background pool. This loss is defined as follows:
L o s s B G = α · f i f ¯ B G 2 + ( 1 α ) · 1 f i , f ¯ B G f i · f ¯ B G
where f i is the feature vector of the ith region (typically the output from the last convolutional layer), f ¯ B G is the mean feature vector representing typical background features, and α ( 0 , 1 ) balances the contributions of the two terms. The first term captures the overall distance between f i and f ¯ B G , using the Euclidean distance, helping the model to identify regions that deviate significantly from the background distribution. This is particularly useful for separating object-like structures from typical background noise. The second term uses cosine similarity to measure the alignment of feature vectors, focusing on their direction rather than just their magnitude. This helps to distinguish features that share similar spatial structures but differ in the semantic content, reducing the risk of false positives.
Together, these two components guide the model to learn a more precise background representation, improving its ability to differentiate foreground objects from complex background regions.
At this point, the important components of the TMBO-AOD have been introduced. It is important to highlight that the improvements offered by the TMBO-AOD are not confined to remote-sensing data; TMBO-AOD can be broadly applied to various types of datasets. In numerous computer vision tasks, such as autonomous driving, video surveillance, and medical image analysis, challenges like background interference and sample imbalance are equally significant. Consequently, dynamically generating candidate frames in the AFF and incorporating the background flags obtained from the CFM can effectively improve the detection accuracy and efficiency in these fields. By adjusting the candidate frame generation strategy and combining it with the SFL, the TMBO-AOD can be better aligned with the target features in different scenes, which significantly improves its detection performance in different datasets.

3. Global Experiments

3.1. Datasets

We evaluate the TMBO-AOD in four datasets: AI-TODv2.0, DIOR, DOTAv1.5, and NWPU VHR-10. AI-TOD contains 700,621 object instances across 28,036 aerial images, spanning 8 distinct categories. Compared to other aerial image object detection datasets, the average object size in AI-TOD is approximately 12.8 pixels, significantly smaller than the objects in other datasets. DIOR is a standard dataset for object detection in remote-sensing imagery, designed to offer a rich set of labeled data for detecting various targets. The dataset includes both high-resolution and low-resolution images, with target categories such as buildings, roads, vehicles, pedestrians, and more. It is widely used in fields like remote-sensing image analysis, target detection, and automated monitoring. DOTA is a large-scale, high-resolution aerial image dataset, comprising 2806 aerial images from 15 cities and regions, with a total of 188,282 annotated objects. These objects span multiple categories, including airplanes, ships, vehicles, basketball courts, and others. The NWPU VHR-10 dataset is a challenging ten-class geospatial object detection dataset. The dataset contains a total of 800 VHR optical remote-sensing images, of which 715 color images were obtained from Google Earth, with spatial resolutions ranging from 0.5 to 2 m. Eighty-five sharpened color–infrared images were acquired from the Vaihingen data, with a spatial resolution of 0.08 m. This dataset contains a total of 1000 VHR optical remote-sensing images.

3.2. Evaluation Metrics

The evaluation metrics used in the experiment are the average precision (AP), mean average precision (mAP), candidate generation time (CGT), and number of floating-point operations per second (FLOPS).
AP and mAP are commonly used metrics in the field of object detection. The formulae for calculating AP and mAP are as follows:
AP = 0 1 P ( R ) d R
mAP = 1 N i = 1 N 0 1 A P i
where P ( R ) represents the precision–recall curve, R is the recall, and N denotes the total number of object categories in the dataset. AP is computed by integrating the precision–recall curve over recall values from 0 to 1, reflecting the model’s ability to balance precision and recall across different confidence thresholds. The mAP is obtained by averaging the AP values across all the object categories, providing an overall measure of the detection accuracy. Specifically, AP and mAP are categorized as A P 50 , m A P 50 , A P 75 , m A P 75 , A P 50 95 , m A P 50 95 , A P v t , A P t , A P s , and A P m . A P 50 , A P 75 , and A P 50 95 are the average precision scores calculated over the IoU ranges 0.5, 0.75, and from 0.5 to 0.95, respectively. Furthermore, m A P 50 , m A P 75 , and m A P 50 95 are the mean values of the APs for each category, calculated over the IoU ranges 0.5, 0.75, and from 0.5 to 0.95, respectively. Additionally, A P v t , A P t , A P s , and A P m are the APs of very tiny (2–8 pixels), tiny (8–16 pixels), small (16–32 pixels), and medium-sized (32–64 pixels) objects defined in the AI-TOD dataset. The calculation methods of the AP and mAP are similar but differ in their specific applications and interpretations. Therefore, for different comparison experiments, we choose either AP or mAP as an evaluation index according to the real situation.
In addition, CGT is the time (in seconds) required to generate candidate boxes in the target detection task. CGT depends on the complexity of the generation algorithm and affects the velocity and accuracy of the target detection system. A shorter candidate generation time indicates a more efficient detection framework. Furthermore, FLOPS values are included as a metric for the computational complexity, reflecting the number of floating-point operations required for inferences. Lower FLOPS values indicate reduced computational costs, which are critical for real-time and resource-constrained applications.

3.3. Implementation Details

We propose two versions of the TMBO-AOD for comparative and ablation studies: TMBO-AOD5, based on YOLOV5, and TMBO-AOD8, based on YOLOV8. The former serves as an anchor-based model, while the latter functions as an anchor-free model, with notable differences in their feature extraction and feature fusion components. The generalization of the TMBO-AOD can be demonstrated through comparisons of both versions of the model with classical object detection models and the current outstanding model. The TMBO-AOD framework is an empirical-background-pool-based screening detection framework, allowing for the replacement of both the backbone network and the feature fusion network according to specific requirements. The TMBO-AOD is implemented in PyTorch 2.0 and operates on a system equipped with two NVIDIA 4090 GPUs. The model is trained using distributed data parallel (DDP) to enhance the training efficiency, achieving twice the velocity relative to that for training on a single GPU. Our designed model did not utilize any pretrained weights during the training process. The training process is optimized using SGD, with momentum parameters β 1 = 0.9 and β 2 = 0.95 , and cosine decay was employed, with a base learning rate of 3.75 × 10−5. Random scaling, random cropping, and random horizontal flipping are also employed. Additionally, the input image’s size is standardized to 1280 × 1280 pixels. Unless otherwise stated, all the comparison methods in the experiments utilized the official code, and all the settings are consistent with those provided in the official documentation.

3.4. Comparisons with the SOTA Methods

3.4.1. Experiment with AI-TODv2.0

In Table 1, we compare the performances of the two versions of the TMBO-AOD (TMBO-AOD5 and TMBO-AOD8) with those of other baseline models in the AI-TODv2.0 test set. These baseline models include the mainstream methods in computer vision: faster R-CNN, DETR [58], IQDet [59], DetectoRS [60], a Swin Transformer [61], YOLOv8 [62], and YOLOv9 [63], as well as the remote-sensing image-oriented SOTA models, iRMB [64], MSC [65], RFLA [48], NWD [66], LSK [67], BRSTD [68], and HANet [69].
Overall, TMBO-AOD8 outperforms the current leading methods across multiple metrics, demonstrating substantial performance advantages. Specifically, TMBO-AOD8 achieves 24.4%, 56.1%, and 23.8% increases in detection accuracy for AP50–90, AP50, and AP75, respectively, representing improvements of 0.3% for AP50–90, 0.6% for AP50, and 5.1% for AP75 compared to the detection performances of the second highest performing model, RFLA [64]. In terms of the detection performances for small APs = 35.1) and medium-sized (APm = 54.9) targets, TMBO-AOD8 demonstrates a significant advantage over the third highest performing model, HANet [69], with an improvement of 7.8% for APs. Although the detection performance for APm is slightly lower than that for HANet, it ranks second among all the models. Furthermore, compared to the computational complexity of the most parameter-efficient model, the Swin Transformer [61] (11.8 GFLOPS), that of TMBO-AOD8 is in the medium range (123.2 GFLOPS), yet TMBO-AOD8 significantly enhances the detection performance, particularly in detecting both very small targets (APvt = 10.9) and tiny targets (APt = 38.1), outperforming all the other models. Conversely, the lightweight TMBO-AOD5 model exhibits a more advantageous computational complexity of only 55.9 GFLOPS, although its detection performance is slightly lower than that of TMBO-AOD8. The TMBO-AOD5 model achieves 22.4% and 49.9% increases in detection accuracy for AP50–90 and AP50, respectively, outperforming those of recent high-performing models, such as NWD [58] and BRSTD [61], as well as all the other models in small-target detection. It also achieves the second highest accuracy (26.5%) in tiny-target detection, demonstrating exceptionally high parametric efficiency. Compared to the lightweight model, YOLOv9 [63] (67.7 GFLOPS), the TMBO-AOD5 model exhibits lower computational complexity while demonstrating higher detection accuracy across nearly all the metrics.
In summary, our models achieve an exceptional balance between computational efficiency and detection accuracy. TMBO-AOD5 delivers a 22.4% improvement in detection accuracy for AP50–90 at just 55.9 GFLOPS, demonstrating remarkable computational efficiency. Although TMBO-AOD8 increases the computational load to 123.2 GFLOPS (2.2× higher than that of TMBO-AOD), it provides only a 9% relative improvement in detection accuracy for AP50–90 (24.4%), revealing clearly diminishing returns from simply scaling up the model size. In addition, TMBO-AOD5’s 55.9 GFLOPS computational demand is fully in line with the real-time processing power of mainstream edge processors (in the 50–100 GFLOPS range), which enables TMBO-AOD5 to be effectively deployed in edge devices. In the future, we will further validate this capability to ensure its feasibility in resource-constrained environments.

3.4.2. Experiment with DOTAv1.5

In this experiment, we use the DOTAv1.5 dataset with the “train” set for training and the “val” set for testing. In addition to the models listed in Table 1, we compare TMBO-AOD8 with several other baseline models, including the classic visual domain model MobileNet [70], FasterNet [71], and the remote-sensing object detection domain models TOSO [72], GWD [73], BDR-Net [74], and O2DFFE [75].
As shown in Table 2, TMBO-AOD8 consistently outperforms all the other mainstream object detection models in the DOTAv1.5 dataset, achieving the highest overall mAP50 of 78.1%, significantly ahead of the second-ranked BDR-Net (71.0%) and third-ranked O2DFFE (70.8%). Among the 16 object categories, TMBO-AOD8 achieves the best AP50 in 12 categories, including complex scenarios, like basketball court (BC), where it leads by 6.1% over the second best model. This superior performance demonstrates the robustness of TMBO-AOD8 in handling dense targets and complex backgrounds, a key advantage provided by its CFM, which effectively reduces background interference by constructing an empirical background pool based on Gaussian distribution, allowing the model to concentrate more accurately on foreground targets.
Moreover, TMBO-AOD8 maintains strong performances in categories with high background complexity, such as swimming pools (SPs) and harbors (HAs), surpassing the second-place model’s performances by 3.2% and 1.1%, respectively. In challenging small-target detection, such as a small vehicle (SV) and a large vehicle (LV), TMBO-AOD8 also outperforms the other models by 4.3% and 0.6%, respectively. This is mostly attributed to the SLF, which enhances the model’s ability to distinguish foreground objects from complex backgrounds, further optimizing the overall detection performance.
Importantly, compared to O2DFFE, which also focuses on mitigating background interference, TMBO-AOD8 demonstrates a more comprehensive approach to background suppression. Although O2DFFE relies on a keypoint attention mechanism and prototype contrastive learning to enhance the foreground, TMBO-AOD8 incorporates a more nuanced strategy through its CFM and SLF, effectively balancing positive and negative sample distributions and dynamically optimizing candidate generation, resulting in a more accurate and stable detection performance across various object scales and scene complexities. The partial inference results of TMBO-AOD8 in the DOTAv1.5 validation set are shown in Figure 5.
Additionally, we performed qualitative analyses on four models, with the visualization results presented in Figure 6. As shown, our TMBO-AOD8 model consistently achieved the highest detection accuracy across three randomly selected images. However, it is noteworthy that in the third image, the performance of TMBO-AOD8 did not exhibit a significant improvement compared to those of the other models. Upon further analysis, this may be attributed to the relatively small proportion of background regions in the image, limiting the effectiveness of the CLM and AFF modules, which are designed to optimize background suppression. Nevertheless, TMBO-AOD8 still demonstrated robust overall performance.

3.4.3. Experiment with DIOR

In this experiment, we compare the performance of TMBO-AOD8 with those of the other baseline models in the DIOR test set. Figure 7 shows the classification accuracies and average accuracies of TMBO-AOD8 and the other mainstream methods in the DIOR dataset, and it can be seen that TMBO-AOD8 significantly outperforms the other mainstream methods in the DIOR dataset and achieves the best overall performance, with a mAP of 65.07%, outperforming models such as Faster R-CNN (60.77%), RetinaNet (59.32%), and the other classical models. TMBO-AOD8 improves upon the sub-optimal model (YWCSL) by 0.45%. The radar plot reveals that TMBO-AOD8 performs well across most categories (e.g., “storagetank” (ST), “trainstation” (TS), and “airport” (APO)). These categories usually have complex backgrounds or contain small targets, showing the model’s adaptability in multi-scene and multi-target detection tasks. Meanwhile, in terms of the coverage area, TMBO-AOD8 is able to maintain a balanced high performance across categories, rather than excelling in only a few categories, demonstrating its comprehensiveness in the overall detection task. The model’s wide range of advantages can be attributed to its transparent mask and background optimization strategies, which effectively separate the target from the background by constructing empirical background pools with Gaussian distributions, especially in scenarios with complex backgrounds and confusing targets (e.g., “airport” and “harbor”), and greatly reducing the number of false positives. The adaptive filtering framework dynamically adjusts the number of candidates, which not only optimizes the detection efficiency but also significantly alleviates the positive and negative sample imbalance problem, thus improving the detection of small-target categories (e.g., “basketballcourt” (BC) and “windmill” (WM)). In addition, the separation loss function enhances the feature consistency between the foreground and background, which enables the model to maintain stable detection accuracy in diverse scenes.

3.4.4. Experiment with NUPUVHR-10

We used the NUPUVHR-10 dataset to compare several different loss functions, namely, the generalized loss functions DIoU [45], CIoU [46], EIoU [47], WIoU [48] and SIoU [49] for visual detection; focal loss [7] for small-target detection; and GWD [66] and KLD [77] for remote-sensing rotating-target detection. Because horizontal detection boxes can be viewed as a special case of rotating detection boxes (when the angle is 0°, 90°, or any multiple thereof), the fundamental principles of GWD and KLD remain applicable. Therefore, we also conduct a comparative analysis with GWD and KLD. Figure 8 illustrates the m A P 0.5 : 0.95 performance of the various loss functions over training epochs, demonstrating the superiority of our proposed SLF. Compared to traditional IoU-based losses such as SIoU, WIoU, and CIoU, which plateau at lower accuracy levels, the SLF continues improving, even in later training stages, achieving the highest final detection accuracy. Distribution-based losses, like GWD and KLD, show moderate improvements over IoU-based methods but still fall short of the SLF, while focal loss, despite its strength in handling class imbalances, remains slightly below the SLF, suggesting that explicitly modeling background information provides additional performance gains. The zoomed-in section further highlights how the SLF maintains a consistent advantage, demonstrating its ability to refine object detection through both foreground and background learning. Unlike conventional losses, which treat the background as a generic negative sample, the SLF integrates a focal-loss-based foreground optimization with a background loss component that explicitly models background features’ consistency, reducing the number of false positives and enhancing the detection’s robustness. Moreover, by leveraging an empirical background pool, the SLF refines the distinction between the foreground objects and background clutter, making it more effective in complex remote-sensing imagery. These combined factors enable the SLF to achieve superior detection accuracy, improved convergence stability, and better adaptability to challenging detection environments, making it a compelling alternative to conventional loss functions.

4. Discussion

4.1. Ablation Experiments Based on the Number of Sub-Blocks

To evaluate the impact of the TMBO-AOD framework on the detection accuracy for different sub-block configurations, we conducted ablation experiments on four datasets (DOTAv1.5, AI-TODv2.0, NUPUVHR-10, and DIOR) using anchor-based YOLOv5 and anchor-free YOLOv8. The results, as shown in Table 3, indicate significant performance improvements across all the datasets when incorporating the TMBO-AOD framework. Notably, the configuration with four sub-blocks achieved the highest performance gains for both models.
Specifically, YOLOv5 exhibits the most substantial improvement in the DIOR dataset, with an AP50 increase of 9.36% (from 48.58% to 53.13%) and an AP50–90 increase of 13.58% (from 19.87% to 22.57%). For YOLOv8, the most remarkable improvement is observed in the DIOR dataset, achieving an AP50 increase of 10.45% (from 49.77% to 54.97%) and an AP50–90 increase of 19.63% (from 20.35% to 24.34%). Across all the datasets, the configuration with four sub-blocks consistently outperformed both the baseline (no sub-blocks) and the two-sub-block configuration, highlighting the robustness of the TMBO-AOD framework.
The heatmaps in Figure 9 visually illustrate the advantages of incorporating sub-blocks into the TMBO-AOD framework. Compared to the baseline YOLO models, the enhanced models with four sub-blocks demonstrate a more concentrated focus on target regions while effectively suppressing the background noise. The addition of transparent masks to sub-blocks, as a part of TMBO-AOD’s design, enables the models to isolate foreground targets and minimize interference from background regions. This effect is particularly evident in high-density datasets, like AI-TODv2.0, where dense and small targets are better localized with sharper boundary definitions.
The effectiveness of the four-sub-block configuration can be attributed to several key mechanisms within the TMBO-AOD framework. First, the adaptive filtering mechanism optimizes the balance between positive and negative samples, improving the detection accuracy for small and ambiguous boundary targets. Second, the separation loss function enhances the distinction between the foreground and background, as reflected in the heatmaps, where background regions consistently exhibit low responses, while target regions display clear and precise high-heat areas.
When the number of sub-blocks increases, the TMBO-AOD framework expands its masking area, further reducing the background interference and sharpening its focus on target regions. This trend is particularly pronounced in the DIOR dataset, where the baseline YOLOv8 often struggles to concentrate on targets, resulting in scattered high-response regions. In contrast, the TMBO-AOD-enhanced YOLOv8 with four sub-blocks achieves significantly improved target localization, as seen in Figure 9.
However, certain regions in the heatmaps reveal limitations in the TMBO-AOD framework’s performance. When targets are more dispersed, the framework can struggle to completely isolate all the background areas. In some cases, portions of the background sub-blocks may inadvertently capture parts of the targets, resulting in over-detection, where target objects are incorrectly classified as background elements. This effect is primarily because of the relatively large size of the sub-blocks, which can fail to effectively differentiate small, target-like background features from actual targets.
As demonstrated in Figure 9, increasing the number of sub-blocks from two to four significantly improves the separation of the foreground and background, allowing for finer-grained masking and more precise focus on target regions. However, this improvement comes at the cost of reduced model efficiency, as a greater number of sub-blocks requires more computational resources to process the finer distinctions between background and foreground regions. Additionally, this approach can sometimes lead to over-detection if the model starts to interpret fine background textures or isolated background objects as potential targets, thus reducing the overall precision.
Therefore, finding an optimal balance between the sub-block size and overall model performance is critical. This balance helps to avoid both efficiency losses and over-detection, ensuring that the model effectively isolates true foreground targets without being misled by background noise. We explore the effect of the sub-block size on the model’s performance in Section 4.1.
Further validation of the four-sub-block configuration is provided in Table 4, which presents the performances of YOLOv5 and YOLOv8 in the DIOR dataset after 20 training epochs. The TMBO-AOD framework yields substantial accuracy improvements, with YOLOv5 and YOLOv8 achieving respective increases of 9.42% and 8.76%. Moreover, the framework significantly reduces the candidate generation time, with YOLOv5 achieving a 48.36% reduction and YOLOv8 a 46.81% reduction. These results highlight the dual benefits of the TMBO-AOD: enhanced detection accuracy and improved computational efficiency, making it highly suitable for real-world applications.

4.2. Ablation Experiments with Hyperparameters in the Loss Function

We tuned the untrained TMBO-AOD8 model in the AI-TODv2.0 dataset, as shown in Table 5, to identify the optimal hyperparameter settings for the SLF. This process involved two stages as follows:
First, we focused on the weighting parameters, λ 0 and λ 1 , which balance the relative importance of the foreground and background components in the SLF. We fixed the values of α and θ and tested different combinations of λ 0 and λ 1 . The results indicate that the model generally achieved better performance when λ 0 < λ 1 , with the optimal combination being λ 0 = 1.0 and λ 1 = 2.0 . This configuration resulted in the highest overall performance, achieving AP50–90 = 19.83%, AP50 = 49.15%, and AP75 = 20.26%. This outcome suggests that assigning a higher weight to the background loss helps the model to better capture the background features, effectively filtering out irrelevant regions and focusing on a smaller set of potential target regions.
Next, we fixed the optimal λ 0 and λ 1 values and explored the impact of the background-modeling parameter, α , which controls the balance between the Euclidean distance and cosine similarity in the background loss function, on the detection accuracy. This balance is critical, as it influences how the model distinguishes subtle background variations. Our experiments revealed that setting α = 0.7 resulted in the best performance, likely because this configuration provides a balanced emphasis on both the feature magnitude and direction, enhancing the model’s ability to capture complex background patterns.
Finally, we refined the focal modulation parameter, θ , and found that θ = 2.5 provided the best results. This value effectively balances the impacts of hard and easy samples, contributing to the overall detection performance.
After a series of ablation experiments, the optimal hyperparameter configuration was identified as α = 0.7 , θ = 2.5 , λ 0 = 1.0 , and λ 1 = 2.0 , achieving the best overall performance. This combination effectively balances the tradeoff between precise foreground classification and robust background filtering, resulting in a significant boost in detection accuracy across varying IoU thresholds.

4.3. Ablation Experiments with Contributions from Each Component

The ablation results in Table 6 highlight the impact of each core component within the TMBO-AOD5 framework on the detection accuracy. Given that the CFM serves as the foundational component for both the AFF and the SLF, the latter two cannot be independently evaluated. With only the CFM enabled, the model shows significant gains, boosting mAP50-90 from 18.1% to 19.7% and mAP50 from 37.2% to 38.4%, reflecting the effectiveness of empirical background pooling in reducing the background interference. This setup also improves mAP75 from 17.0% to 18.2%, indicating more precise boundary localization. Adding the AFF further enhances medium-scale target detection, raising mAPm from 41.9% to 45.2%, though it slightly compromises the accuracy of the very small-target detection (mAPvt drops from 8.3% to 8.0%), suggesting the need for consistent background learning. In contrast, the combination of the CFM and SLF provides a more balanced boost, improving the accuracies of both small- and medium-sized-target detections (mAPs and mAPm, respectively) while maintaining the overall detection precision. The full TMBO-AOD5 configuration, integrating the CFM, AFF, and SLF, achieves the highest performance across all the metrics, with mAP50-90 reaching 22.4% and mAP50 peaking at 49.9%. This comprehensive improvement underscores the synergistic effect of precise background optimization, adaptive candidate filtering, and effective feature separation.

5. Conclusions

In this paper, we propose the TMBO-AOD framework to address the challenges of background interference and class imbalance in remote-sensing object detection. By introducing a clear focus module, an adaptive filtering framework, and a novel separation loss function, our method effectively mitigates background noise, improves anchor generation efficiency, and enhances detection performance. Extensive experiments in four datasets and with both anchor-based and anchor-free models demonstrate the robustness and generalizability of the TMBO-AOD. This framework not only achieves significant performance improvements but also establishes a novel perspective for addressing remote-sensing detection challenges, paving the way for more accurate and efficient object detection in complex scenarios.

Author Contributions

Conceptualization, T.F.; methodology, T.F.; software, T.F.; validation, T.F.; writing—original draft preparation, T.F.; writing—review and editing, B.Y.; visualization, T.F.; supervision, H.D., B.Y. and B.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by a study of multi-objective evolutionary clustering algorithms (KY10600200048), research on environmental characterization and trajectory-planning technology of micro unmanned aerial vehicles for dynamic uncertain environments (62303486), and research on the mobile augmented-reality-map construction method with visual inertia fusion (61902423). The authors have no competing interests to declare that are relevant to the content of this article.

Data Availability Statement

The datasets used in this study are all publicly available and have been cited in the text.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Wang, J.; Yang, W.; Guo, H.; Zhang, R.; Xia, G.S. Tiny Object Detection in Aerial Images. In Proceedings of the 25th International Conference on Pattern Recognition (ICPR), Virtual Conference, 10–15 January 2021; pp. 3791–3798. [Google Scholar]
  2. Li, K.; Wan, G.; Cheng, G.; Meng, L.; Han, J. Object Detection in Optical Remote Sensing Images: A Survey and a New Benchmark. ISPRS J. Photogramm. Remote Sens. 2020, 159, 296–307. [Google Scholar] [CrossRef]
  3. Xia, G.-S.; Bai, X.; Ding, J.; Zhu, Z.; Belongie, S.; Luo, J.; Datcu, M.; Pelillo, M.; Zhang, L. DOTA: A Large-scale Dataset for Object Detection in Aerial Images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018; pp. 3974–3983. [Google Scholar]
  4. Cheng, G.; Zhou, P.; Han, J. Learning Rotation-Invariant Convolutional Neural Networks for Object Detection in VHR Optical Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2016, 54, 7405–7415. [Google Scholar] [CrossRef]
  5. Shrivastava, A.; Gupta, A.; Girshick, R. Training Region-Based Object Detectors with Online Hard Example Mining. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 761–769. [Google Scholar]
  6. Li, M.; Zhang, Z.; Yu, H.; Chen, X.; Li, D. S-OHEM: Stratified Online Hard Example Mining for Object Detection. In Proceedings of the Second CCF Chinese Conference on Computer Vision (CCCV), Tianjin, China, 11–14 October 2017; pp. 166–177. [Google Scholar]
  7. Lin, T.-Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal Loss for Dense Object Detection. In Proceedings of the IEEE international conference on computer vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
  8. Li, B.; Liu, Y.; Wang, X. Gradient Harmonized Single-Stage Detector. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; pp. 8577–8584. [Google Scholar]
  9. Yang, K.; Zhang, H.; Zhou, D.; Dong, L. PaaRPN: Probabilistic Anchor Assignment with Region Proposal Network for Visual Tracking. Inf. Sci. 2022, 598, 19–36. [Google Scholar] [CrossRef]
  10. Li, Z.; Liu, Y.; Wang, X.; Zhang, Y.; Chen, H. Deep Learning-Based Object Detection Techniques for Remote Sensing Images: A Survey. Remote Sens. 2022, 14, 2385. [Google Scholar] [CrossRef]
  11. He, Y.; Sun, X.; Gao, L.; Zhang, B. Ship Detection without Sea-Land Segmentation for Large-Scale High-Resolution Optical Satellite Images. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Valencia, Spain, 22–27 July 2018; pp. 717–720. [Google Scholar]
  12. Li, W.; Dong, R.; Fu, H.; Yu, L. Large-Scale Oil Palm Tree Detection from High-Resolution Satellite Images Using Two-Stage Convolutional Neural Networks. Remote Sens. 2018, 11, 11. [Google Scholar] [CrossRef]
  13. Song, Z.; Sui, H.; Hua, L. A Hierarchical Object Detection Method in Large-Scale Optical Remote Sensing Satellite Imagery Using Saliency Detection and CNN. Int. J. Remote Sens. 2021, 42, 2827–2847. [Google Scholar] [CrossRef]
  14. Van Etten, A. You Only Look Twice: Rapid Multi-Scale Object Detection in Satellite Imagery. arXiv 2018, arXiv:1805.09512. [Google Scholar]
  15. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  16. Redmon, J.; Farhadi, A. YOLOv3: An Incremental Improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  17. Zhu, X.; Lyu, S.; Wang, X.; Zhao, Q. TPH-YOLOv5: Improved YOLOv5 Based on Transformer Prediction Head for Object Detection on Drone-Captured Scenarios. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 11–17 October 2021; pp. 2778–2788. [Google Scholar]
  18. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. Ssd: Single shot multibox detector. In Proceedings of the European Conference on Computer Vision (ECCV), Amsterdam, The Netherlands, 11–14 October 2017; pp. 21–37. [Google Scholar]
  19. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef]
  20. Liu, S.; Shi, H.; Guo, Z. Remote Sensing Image Object Detection Based on Improved SSD. In Proceedings of the IEEE International Conference on Geoscience and Remote Sensing, Kuala Lumpur, Malaysia, 17–22 July 2022; pp. 421–424. [Google Scholar]
  21. Lv, H.; Qian, W.; Chen, T.; Yang, H.; Zhou, X. Multiscale Feature Adaptive Fusion for Object Detection in Optical Remote Sensing Images. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
  22. Shamsolmoali, P.; Zareapoor, M.; Yang, J.; Granger, E.; Chanussot, J. Enhanced Single-Shot Detector for Small Object Detection in Remote Sensing Images. In Proceedings of the IEEE International Conference on Image Processing (ICIP), Bordeaux, France, 16–19 October 2022; pp. 1716–1719. [Google Scholar]
  23. Han, J.; Ding, J.; Li, J.; Xia, G.-S. Align Deep Features for Oriented Object Detection. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5602511. [Google Scholar] [CrossRef]
  24. Yang, X.; Yan, J.; Feng, Z.; He, T. R3Det: Refined Single-Stage Detector with Feature Refinement for Rotating Object. Proc. Aaai Conf. Artif. Intell. 2021, 35, 3163–3171. [Google Scholar] [CrossRef]
  25. Xie, X.; Cheng, G.; Wang, J.; Yao, X.; Han, J. Oriented R-CNN for Object Detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 11–17 October 2021; pp. 3520–3529. [Google Scholar]
  26. Ding, J.; Xue, N.; Long, Y.; Xia, G.-S.; Lu, Q. Learning RoI Transformer for Oriented Object Detection in Aerial Images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 2849–2858. [Google Scholar]
  27. Dong, Z.; Wang, M.; Wang, Y.; Zhu, Y.; Zhang, Z. Object Detection in High Resolution Remote Sensing Imagery Based on Convolutional Neural Networks with Suitable Object Scale Features. IEEE Trans. Geosci. Remote Sens. 2020, 58, 2104–2114. [Google Scholar] [CrossRef]
  28. Liu, Z.; Hu, J.; Weng, L.; Yang, Y. Rotated Region Based CNN for Ship Detection. In Proceedings of the IEEE International Conference on Image Processing (ICIP), Beijing, China, 17–20 September 2017; pp. 900–904. [Google Scholar]
  29. Yang, X.; Yang, J.; Yan, J.; Zhang, Y.; Zhang, T.; Guo, Z.; Sun, X.; Fu, K. SCRDet: Towards More Robust Detection for Small, Cluttered and Rotated Objects. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 8232–8241. [Google Scholar]
  30. Law, H.; Deng, J. CornerNet: Detecting Objects as Paired Keypoints. Int. J. Comput. Vis. 2020, 128, 435–446. [Google Scholar] [CrossRef]
  31. Law, H.; Teng, Y.; Russakovsky, O.; Deng, J. CornerNet-Lite: Efficient Keypoint Based Object Detection. arXiv 2019, arXiv:1904.08900. [Google Scholar]
  32. Duan, K.; Bai, S.; Xie, L.; Qi, H.; Huang, Q.; Tian, Q. CenterNet: Keypoint Triplets for Object Detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 6569–6578. [Google Scholar]
  33. Zhou, X.; Wang, D.; Krähenbühl, P. Objects as Points. arXiv 2019, arXiv:1904.07850. [Google Scholar]
  34. Kong, T.; Sun, F.; Liu, H.; Jiang, Y.; Li, L.; Shi, J. FoveaBox: Beyond Anchor-Based Object Detection. IEEE Trans. Image Process. 2020, 29, 7389–7398. [Google Scholar] [CrossRef]
  35. Zhou, X.; Zhuo, J.; Krähenbühl, P. Bottom-Up Object Detection by Grouping Extreme and Center Points. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 850–859. [Google Scholar]
  36. Yang, Z.; Liu, S.; Hu, H.; Wang, L.; Lin, S. RepPoints: Point Set Representation for Object Detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 9657–9666. [Google Scholar]
  37. Liu, W.; Hasan, I.; Liao, S. Center and Scale Prediction: Anchor-Free Approach for Pedestrian and Face Detection. Pattern Recognit. 2023, 135, 109071. [Google Scholar] [CrossRef]
  38. Tian, Z.; Shen, C.; Chen, H.; He, T. FCOS: A Simple and Strong Anchor-Free Object Detector. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 1922–1933. [Google Scholar] [CrossRef]
  39. Zhang, Y.; Sheng, W.; Jiang, J.; Jing, N.; Wang, Q.; Mao, Z. Priority Branches for Ship Detection in Optical Remote Sensing Images. Remote Sens. 2020, 12, 1196. [Google Scholar] [CrossRef]
  40. Dai, P.; Yao, S.; Li, Z.; Zhang, S.; Cao, X. ACE: Anchor-Free Corner Evolution for Real-Time Arbitrarily-Oriented Object Detection. IEEE Trans. Image Process. 2022, 31, 4076–4089. [Google Scholar] [CrossRef] [PubMed]
  41. Yi, J.; Wu, P.; Liu, B.; Huang, Q.; Qu, H.; Metaxas, D. Oriented Object Detection in Aerial Images with Box Boundary-Aware Vectors. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Virtual, 5–9 January 2021; pp. 2150–2159. [Google Scholar]
  42. Cheng, G.; Wang, J.; Li, K.; Xie, X.; Lang, C.; Yao, Y.; Han, J. Anchor-Free Oriented Proposal Generator for Object Detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–11. [Google Scholar] [CrossRef]
  43. Li, J.; Tian, Y.; Xu, Y.; Zhang, Z. Oriented Object Detection in Remote Sensing Images with Anchor-Free Oriented Region Proposal Network. Remote Sens. 2022, 14, 1246. [Google Scholar] [CrossRef]
  44. Rezatofighi, H.; Tsoi, N.; Gwak, J.; Sadeghian, A.; Reid, I.; Savarese, S. Generalized Intersection over Union: A Metric and a Loss for Bounding Box Regression. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 658–666. [Google Scholar]
  45. Zheng, Z.; Wang, P.; Liu, W.; Li, J.; Ye, R.; Ren, D. Distance-IoU Loss: Faster and Better Learning for Bounding Box Regression. Proc. Aaai Conf. Artif. Intell. 2020, 34, 12993–13000. [Google Scholar] [CrossRef]
  46. Zheng, Z.; Wang, P.; Ren, D.; Liu, W.; Ye, R.; Hu, Q.; Zuo, W. Enhancing geometric factors in model learning and inference for object detection and instance segmentation. IEEE Trans. Cybern. 2021, 52, 8574–8586. [Google Scholar] [CrossRef]
  47. Zhang, Y.F.; Ren, W.; Zhang, Z.; Jia, Z.; Wang, L.; Tan, T. Focal and efficient IOU loss for accurate bounding box regression. Neurocomputing 2022, 506, 146–157. [Google Scholar] [CrossRef]
  48. Tong, Z.; Chen, Y.; Xu, Z.; Yu, R. Wise-IoU: Bounding Box Regression Loss with Dynamic Focusing Mechanism. arXiv 2023, arXiv:2301.10051. [Google Scholar]
  49. Gevorgyan, Z. SIoU Loss: More Powerful Learning for Bounding Box Regression. arXiv 2022, arXiv:2205.12740. [Google Scholar]
  50. Ming, Q.; Miao, L.; Zhou, Z.; Song, J.; Pizurica, A. Gradient calibration loss for fast and accurate oriented bounding box regression. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5611015. [Google Scholar] [CrossRef]
  51. Yang, X.; Yan, J. Arbitrary-oriented object detection with circular smooth label. In Proceedings of the European Conference on Computer Vision (ECCV), Glasgow, UK, 23–28 August 2020; pp. 677–694. [Google Scholar]
  52. Ming, Q.; Miao, L.; Zhou, Z.; Vercheval, N.; Pizurica, A. Not all boxes are equal: Learning to optimize bounding boxes with discriminative distributions in optical remote sensing images. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5622514. [Google Scholar] [CrossRef]
  53. Han, J.; Ding, J.; Xue, N.; Xia, G.-S. ReDet: A rotation-equivariant detector for aerial object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 19–25 June 2021; pp. 2786–2795. [Google Scholar]
  54. Xu, C.; Wang, J.; Yang, W.; Yu, H.; Yu, L.; Xia, G.-S. RFLA: Gaussian Receptive Field Based Label Assignment for Tiny Object Detection. In Proceedings of the European Conference on Computer Vision (ECCV), Tel Aviv, Israel, 23–27 October 2022; pp. 526–543. [Google Scholar]
  55. Li, Z.; Sun, S.; Li, Y.; Sun, B.; Tian, K.; Qiao, L.; Lu, X. Aerial Image Object Detection Method Based on Adaptive ClusDet Network. In Proceedings of the 2021 IEEE 21st International Conference on Communication Technology (ICCT), Tianjin, China, 13–16 October 2021; pp. 1091–1096. [Google Scholar]
  56. Shi, J.; Wang, C. CDMnet: Cloud Detection in Remote Sensing Images Based on CNN. J. Phys. Conf. Ser. 2023, 2640, 012013. [Google Scholar] [CrossRef]
  57. Feng, C.; Zhong, Y.; Gao, Y.; Scott, M.R.; Huang, W. TOOD: Task-aligned one-stage object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 11–17 October 2021; pp. 3490–3499. [Google Scholar]
  58. Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; Zagoruyko, S. End-to-End Object Detection with Transformers. In European Conference on Computer Vision; Springer: Cham, Switzerland, 2020; pp. 213–229. [Google Scholar]
  59. Ma, Y.; Liu, S.; Li, Z.; Sun, J. Iqdet: Instance-Wise Quality Distribution Sampling for Object Detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 1717–1725. [Google Scholar]
  60. Qiao, S.; Chen, L.-C.; Yuille, A. DetectoRS: Detecting Objects with Recursive Feature Pyramid and Switchable Atrous Convolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 10208–10219. [Google Scholar]
  61. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin Transformer: Hierarchical Vision Transformer Using Shifted Windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 10012–10022. [Google Scholar]
  62. Ultralytics. Available online: https://github.com/ultralytics/ultralytics (accessed on 12 April 2025).
  63. Wang, C.-Y.; Yeh, I.-H.; Liao, H.-Y.M. YOLOv9: Learning what you want to learn using programmable gradient information. arXiv 2024, arXiv:2402.13616. [Google Scholar]
  64. Zhang, J.; Li, X.; Li, J.; Liu, L.; Xue, Z.; Zhang, B.; Jiang, Z.; Huang, T.; Wang, Y.; Wang, C. Rethinking mobile block for efficient attention-based models. In Proceedings of the 2023 IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 2–3 October 2023; pp. 1389–1400. [Google Scholar]
  65. Guo, M.-H.; Lu, C.-Z.; Hou, Q.; Liu, Z.; Cheng, M.-M.; Hu, S.-M. SegNeXt: Rethinking Convolutional Attention Design for Semantic Segmentation. Proc. Adv. Neural Inf. Process. Syst. (NeurIPS) 2022, 35, 1140–1156. [Google Scholar]
  66. Xu, C.; Wang, J.; Yang, W.; Yu, H.; Yu, L.; Xia, G.-S. Detecting tiny objects in aerial images: A normalized Wasserstein distance and a new benchmark. ISPRS J. Photogramm. Remote Sens. 2022, 190, 79–93. [Google Scholar] [CrossRef]
  67. Li, Y.; Hou, Q.; Zheng, Z.; Cheng, M.-M.; Yang, J.; Li, X. Large selective kernel network for remote sensing object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 18–22 June 2023; pp. 16794–16805. [Google Scholar]
  68. Huang, S.; Lin, C.; Jiang, X.; Qu, Z. BRSTD: Bio-Inspired Remote Sensing Tiny Object Detection. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5643115. [Google Scholar] [CrossRef]
  69. Guo, G.; Chen, P.; Yu, X.; Han, Z.; Ye, Q.; Gao, S. Save the tiny, save the all: Hierarchical activation network for tiny object detection. IEEE Trans. Circuits Syst. Video Technol. 2024, 34, 221–234. [Google Scholar] [CrossRef]
  70. Howard, A.; Sandler, M.; Chu, G.; Chen, L.C.; Chen, B.; Tan, M.; Wang, W.; Zhu, Y.; Pang, R.; Vasudevan, V.; et al. Searching for MobileNetV3. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 16–20 June 2019; pp. 1314–1324. [Google Scholar]
  71. Chen, J.; Kao, S.H.; He, H.; Zhuo, W.; Wen, S.; Lee, C.H.; Chan, S.H.G. Run, Don’t Walk: Chasing higher FLOPS for faster neural networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 18–22 June 2023; pp. 12021–12031. [Google Scholar]
  72. Feng, P.; Lin, Y.; Guan, J.; He, G.; Shi, H.; Chambers, J. TOSO: Student’s-T distribution aided one-stage orientation target detection in remote sensing images. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Barcelona, Spain, 4–8 May 2020; pp. 4057–4061. [Google Scholar]
  73. Yang, X.; Yan, J.; Ming, Q.; Wang, W.; Zhang, X.; Tian, Q. Rethinking rotated object detection with Gaussian Wasserstein distance loss. In Proceedings of the International Conference on Machine Learning (ICML), Virtual Event, 18–24 July 2021; pp. 11830–11840. [Google Scholar]
  74. Wang, H.; Liao, Y.; Li, Y.; Fang, Y.; Ni, S.; Luo, Y.; Jiang, B. BDR-Net: Bhattacharyya Distance-Based Distribution Metric Modeling for Rotating Object Detection in Remote Sensing. IEEE Trans. Instrum. Meas. 2023, 72, 1–12. [Google Scholar] [CrossRef]
  75. Lin, P.; Wu, X.; Wang, B. Oriented Object Detection Based on Foreground Feature Enhancement in Remote Sensing Images. Remote Sens. 2022, 14, 6226. [Google Scholar] [CrossRef]
  76. Dai, L.; Liu, H.; Tang, H.; Wu, Z.; Song, P. AO2-DETR: Arbitrary-oriented object detection transformer. IEEE Trans. Circuits Syst. Video Technol. 2022, 33, 2342–2356. [Google Scholar] [CrossRef]
  77. Yang, X.; Zhang, G.; Yang, X.; Zhou, Y.; Wang, W.; Tang, J.; He, T.; Yan, J. Detecting rotated objects as Gaussian distributions and its 3-D generalization. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 4335–4354. [Google Scholar] [CrossRef]
Figure 1. Four remote-sensing image datasets and their common features.
Figure 1. Four remote-sensing image datasets and their common features.
Remotesensing 17 01762 g001
Figure 2. The overall framework of the TMBO-AOD: (a) our proposed clear focus module and (b) our proposed adaptive filtering framework.
Figure 2. The overall framework of the TMBO-AOD: (a) our proposed clear focus module and (b) our proposed adaptive filtering framework.
Remotesensing 17 01762 g002
Figure 3. The empirical background pool and transparency masks in clear focus. The blue rectangular box represents the four random sub-blocks in the image.
Figure 3. The empirical background pool and transparency masks in clear focus. The blue rectangular box represents the four random sub-blocks in the image.
Remotesensing 17 01762 g003
Figure 4. (a) Anchor generation for anchor-based models. (b) Proposal generation and label assignment for anchor-free models. (c) Anchor generation for anchor-based models combined with the adaptive filtering framework. (d) Proposal generation and label assignment for anchor-free models combined with the adaptive filtering framework. The orange rectangle in the picture represents the sliding window, and the yellow rectangle represents the grid.
Figure 4. (a) Anchor generation for anchor-based models. (b) Proposal generation and label assignment for anchor-free models. (c) Anchor generation for anchor-based models combined with the adaptive filtering framework. (d) Proposal generation and label assignment for anchor-free models combined with the adaptive filtering framework. The orange rectangle in the picture represents the sliding window, and the yellow rectangle represents the grid.
Remotesensing 17 01762 g004
Figure 5. Partial inference results for TMBO-AOD8 in the DOTAv1.5 validation set.
Figure 5. Partial inference results for TMBO-AOD8 in the DOTAv1.5 validation set.
Remotesensing 17 01762 g005
Figure 6. Visualization of the detection results of four excellent models for the same three images. It is worth noting that the values following each model refer to the detection accuracies.
Figure 6. Visualization of the detection results of four excellent models for the same three images. It is worth noting that the values following each model refer to the detection accuracies.
Remotesensing 17 01762 g006
Figure 7. Comparison of classification detection results with those of state-of-the-art models in the DIOR dataset. Note that after the models’ names are their mAPs.
Figure 7. Comparison of classification detection results with those of state-of-the-art models in the DIOR dataset. Note that after the models’ names are their mAPs.
Remotesensing 17 01762 g007
Figure 8. The mAP curves with different loss functions in the NUPUVHR-10 dataset.
Figure 8. The mAP curves with different loss functions in the NUPUVHR-10 dataset.
Remotesensing 17 01762 g008
Figure 9. The heatmaps of inference results of YOLOv5 and YOLOv8 incorporating TMBO-AOD ideas in four datasets: DOTAv1.5, AI-TODv2.0, NUPUVHR-10, and DIOR.
Figure 9. The heatmaps of inference results of YOLOv5 and YOLOv8 incorporating TMBO-AOD ideas in four datasets: DOTAv1.5, AI-TODv2.0, NUPUVHR-10, and DIOR.
Remotesensing 17 01762 g009
Table 1. A comparison of the detection results with those of the state-of-the-art models in AI-TODv2.0. All the AP values in the table are expressed as percentages, but this is omitted for clarity. Note that the red numbers refer to the best results, the blue numbers refer to the second best results.
Table 1. A comparison of the detection results with those of the state-of-the-art models in AI-TODv2.0. All the AP values in the table are expressed as percentages, but this is omitted for clarity. Note that the red numbers refer to the best results, the blue numbers refer to the second best results.
MethodVenueAP50–90AP50AP75APvtAPtAPsAPmGFLOPS
Faster R-CNNTPAMI201712.725.77.40.011.628.333.9289.3
DETR [58]ECCV20202.911.70.50.62.63.312.1227
iRMB [64]ICCT202121.344.719.95.023.933.853.957.1
IQDet [59]CVPR20219.421.711.32.19.215.631.868.4
DetectoRS [60]CVPR202115.032.612.10.110.628.940.0244
Swin Transformer–tiny [61]ICCV20219.724.45.35.410.415.324.8105.2
Swin Transformer–sin [61]ICCV202112.932.76.26.710.120.937.611.8
MSC [65]NeurIPS202221.442.717.19.115.529.147.932.8
RFLA [54]ECCV202224.155.518.79.625.830.137.6-
NWD [66]ISPRS202223.954.616.98.924.129.236.5246
LSK [67]CVPR202321.243.417.710.015.630.353.963.4
YOLOv8s [62]arXiv202315.436.120.34.111.935.545.929.7
BRSTD [68]TGRS202423.654.216.610.422.830.337.664.3
YOLOv9s [63]arXiv202421.742.622.24.911.637.051.467.7
HANet [69]TCSVT202424.053.719.810.322.227.355.5171.3
TMBO-AOD5 (Ours)-22.449.922.89.926.532.752.855.9
TMBO-AOD8 (Ours)-24.456.123.810.938.135.154.9123.2
Note that red refers to the best results and blue refers to the second best results. Because of the limited space in the table, the percentage symbol (%) is omitted from all the AP values.
Table 2. Comparison of detection results with those of state-of-the-art models in the DOTAv1.5 dataset.
Table 2. Comparison of detection results with those of state-of-the-art models in the DOTAv1.5 dataset.
MethodVenueAP50 (%)mAP50 (%)
PLSHSTBDTCBCGTFHABRLVSVHCRASBFSPCC
Faster R-CNN [19]TPAMI201781.977.584.373.190.676.158.761.143.970.972.656.863.148.071.120.263.7
MobileNet [70]CVPR201974.832.417.313.529.454.769.679.817.745.719.314.752.053.88.89.939.4
TOSO [72]ICASSP202187.265.644.348.946.168.083.090.746.068.948.760.866.570.617.932.362.5
GWD [73]PMLR202188.177.878.576.685.980.165.263.834.961.771.354.561.258.362.338.112.9
Swin Transformer–tiny [61]ICCV202140.515.13.85.310.427.432.248.65.219.94.17.213.023.51.720.713.9
Swin Transformer–sin [61]ICCV202142.110.23.41.313.225.835.648.05.724.93.66.015.121.01.522.123.1
AO2-DETR [76]TCSVT202279.679.777.678.180.674.761.258.642.474.555.369.666.953.673.224.766.3
O2DFFE [77]RS202280.286.182.381.590.879.272.073.453.580.967.566.860.360.173.125.770.8
FasterNet [71]CVPR202347.219.68.36.414.530.541.660.78.126.59.314.719.025.82.618.969.3
YOLOV8sarXiv202389.974.550.255.450.371.685.491.452.873.459.166.175.071.443.239.365.9
BDR-Net [74]TIM202389.779.782.978.490.883.173.170.046.074.775.448.263.160.766.842.971.0
HANet [69]TCSVT202490.571.249.247.852.573.385.690.957.674.351.066.375.675.642.740.165.9
TMBO-AOD8 (Ours)-90.888.184.375.284.989.285.092.558.081.579.774.276.672.976.444.878.1
Note that red refers to the best results and blue refers to the second best results.
Table 3. The performances of YOLOv5 and YOLOv8 incorporating the TMBO-AOD without pretraining in the four datasets.
Table 3. The performances of YOLOv5 and YOLOv8 incorporating the TMBO-AOD without pretraining in the four datasets.
MethodDatasetSub-BlocksAP50–90 (%)AP50 (%)
YOLOV5DOTAv1.5039.0963.29
239.5764.21
441.9666.95
AI-TODv2.0018.1945.97
218.3646.25
419.1547.01
NUPUVHR-10058.2394.19
258.9694.88
459.2795.30
DIOR019.8748.58
221.2249.60
422.5753.13
YOLOV8DOTAv1.5039.2763.89
240.1864.56
442.1268.17
AI-TODv2.0018.5646.00
219.3746.34
420.1547.22
NUPUVHR-10059.6795.82
261.0196.33
461.7598.21
DIOR020.3549.77
221.5650.82
422.3454.97
Note that bold font represents optimal performance.
Table 4. The performances of YOLOv5 and YOLOv8 incorporating the TMBO-AOD without pretraining in the DIOR dataset.
Table 4. The performances of YOLOv5 and YOLOv8 incorporating the TMBO-AOD without pretraining in the DIOR dataset.
MethodAP50–90 (%)AP50 (%)CGT (s)
YOLOV54.1111.871.22
YOLOV5 + TMBO-AOD6.7517.68:0.63 (↑48.36%)
YOLOV84.6312.760.47
YOLOV8 + TMBO-AOD7.9219.200.25 (↑46.81%)
Note that upward arrow represents performance improvement and bold font represents optimal performance.
Table 5. Hyperparameter combinations and detection performance metrics in the AI-TODv2.0 dataset.
Table 5. Hyperparameter combinations and detection performance metrics in the AI-TODv2.0 dataset.
α θ λ 0 λ 1 AP50–90 (%)AP50 (%)AP75 (%)
0.52.01.01.019.6047.7318.25
2.01.019.2946.6119.86
1.02.019.8349.1520.26
1.03.019.7749.0919.91
0.62.01.02.019.4248.7319.02
0.719.9949.7320.45
0.819.2348.5918.88
0.73.01.02.019.2248.1318.90
4.019.1248.2519.70
2.520.3550.2221.56
Note that bold font represents optimal performance.
Table 6. Ablation results for the various components in TMBO-AOD5.
Table 6. Ablation results for the various components in TMBO-AOD5.
CFMAFFSLFmAP50–90mAP50mAP75mAPvtmAPtmAPsmAPm
18.137.217.08.315.431.441.9
19.738.418.28.315.932.343.7
20.239.417.68.016.832.545.2
21.543.521.99.319.732.451.4
22.449.922.89.926.532.752.8
Note that bold font represents optimal performance. The check mark represents the adoption of the module.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Fu, T.; Dong, H.; Yang, B.; Deng, B. TMBO-AOD: Transparent Mask Background Optimization for Accurate Object Detection in Large-Scale Remote-Sensing Images. Remote Sens. 2025, 17, 1762. https://doi.org/10.3390/rs17101762

AMA Style

Fu T, Dong H, Yang B, Deng B. TMBO-AOD: Transparent Mask Background Optimization for Accurate Object Detection in Large-Scale Remote-Sensing Images. Remote Sensing. 2025; 17(10):1762. https://doi.org/10.3390/rs17101762

Chicago/Turabian Style

Fu, Tianyi, Hongbin Dong, Benyi Yang, and Baosong Deng. 2025. "TMBO-AOD: Transparent Mask Background Optimization for Accurate Object Detection in Large-Scale Remote-Sensing Images" Remote Sensing 17, no. 10: 1762. https://doi.org/10.3390/rs17101762

APA Style

Fu, T., Dong, H., Yang, B., & Deng, B. (2025). TMBO-AOD: Transparent Mask Background Optimization for Accurate Object Detection in Large-Scale Remote-Sensing Images. Remote Sensing, 17(10), 1762. https://doi.org/10.3390/rs17101762

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop