Next Article in Journal
Real-Time Remote Monitoring of Environmental Conditions and Actuator Status in Smart Greenhouses Using a Smartphone Application
Next Article in Special Issue
VMESR: Variable Mamba-Enhanced Super-Resolution for Real-Time Road Scene Understanding with Automotive Vision Sensors
Previous Article in Journal
Broadband Impedance Matching for Immersed CMUTs: An End-to-End Design-to-Measurement Validation Framework
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

ATDIOU: Arctangent Differential Loss Function for Bounding Box Regression

1
Xi’an Institute of Optics and Precision Mechanics of CAS, Xi’an 710119, China
2
University of Chinese Academy of Sciences, Beijing 100049, China
*
Author to whom correspondence should be addressed.
Sensors 2026, 26(5), 1545; https://doi.org/10.3390/s26051545
Submission received: 6 January 2026 / Revised: 10 February 2026 / Accepted: 25 February 2026 / Published: 1 March 2026
(This article belongs to the Special Issue AI for Emerging Image-Based Sensor Applications)

Abstract

Object detection is a fundamental task in computer vision. Bounding box regression (BBR) losses are critical to detector performance. However, evaluation measures that rely on the Intersection over Union (IoU) between the predicted and ground truth boxes are highly sensitive to positional deviations, which can hinder optimization. To alleviate this issue, we propose ATDIoU, a novel arctangent-differential loss for bounding-box regression. ATDIoU computes distance similarity between a predicted and a ground truth box by modeling the distances between their corresponding vertices as a two-dimensional arctangent differential distribution (ATD). This arctangent differential-based design mitigates bounding box drift and reduces sensitivity to localization errors. As a result, it guides the model to learn target positions more effectively. We evaluate ATDIoU by integrating it into YOLOv6 and conducting experiments on PASCAL VOC and VisDrone2019. The results demonstrate that ATDIoU yields improvements of 1.4% and 0.7% in mean average precision (mAP) relative to MPDIoU.

1. Introduction

Precise bounding box regression (BBR) is fundamental to object detection. As a core task in computer vision, object detection underpins a diverse array of applications, including autonomous driving, video surveillance, medical image analysis, and remote sensing. The accuracy of bounding box localization critically influences the performance of downstream tasks, such as instance segmentation, object tracking, and activity recognition. Modern detectors, including both one-stage and two-stage frameworks, typically use Intersection over Union (IoU)–based metrics or their variants to supervise the localization process [1,2,3,4,5,6,7,8]. Due to their intrinsic scale invariance and alignment with evaluation metrics, IoU and its derivative, IoU loss, remain the predominant choice for BBR.
Despite their efficacy in high-overlap scenarios, IoU-based losses present two significant limitations. First, they exhibit high sensitivity to spatial misalignment. In instances where the predicted and ground truth boxes are non-overlapping, the IoU metric yields a zero value, resulting in vanishing gradient. This phenomenon severely impedes model optimization, particularly during the initial training phases or when processing hard examples. Second, these losses demonstrate instability regarding boundary deviations. Slight positional shifts near edges can cause large fluctuations in overlap. Such instability destabilizes gradient updates and impedes convergence [9,10,11,12,13,14,15,16]. These deficiencies are particularly pronounced in tasks involving small objects or dense scenes. In these contexts, precise corner alignment is imperative, and the tolerance for error is negligible.
To mitigate these deficiencies, extensive research has focused on extending the standard IoU metric. A prevalent strategy involves augmenting the loss function with geometric penalties derived from center distance, aspect ratio, and the area of the minimum enclosing box. For instance, DIoU [12] explicitly penalizes the normalized distance between the centroids of the predicted and ground truth boxes to accelerate convergence. CIoU [12] expands upon this by incorporating a term for aspect ratio consistency. Subsequent variants, such as SIoU [13] and EIoU [14], introduce angle-aware components and reformulate aspect ratio penalties to enhance optimization stability. Concurrently, alternative research directions substitute overlap-based measures with metrics grounded in distance or distributional similarity, such as the Wasserstein distance [17,18]. These approaches are particularly advantageous as they preserve informative gradients even in non-overlapping scenarios.
However, a crucial limitation remains: most of these formulations continue to supervise the bounding box as a single, holistic entity. Specifically, they aggregate error signals based on global properties such as area, center, or aspect ratio. This structure inherently renders precise corner alignment an implicit learning target rather than an explicit optimization objective. Consequently, the regression head must infer the correct vertex positions indirectly from these holistic cues. This limitation is particularly evident in challenging detection scenarios, characterized by dense layouts, and pronounced scale variation. Datasets such as VisDrone2019 [19] exemplify these conditions. Under such circumstances, even a well-centered box with an appropriate aspect ratio can experience substantial IoU degradation due to misaligned corners. Therefore, corner misalignment emerges as a major source of localization error and a key factor contributing to IoU instability, especially for elongated or irregularly shaped objects.
To address this gap, we introduce ATDIoU, a vertex arctangent differential distance loss that explicitly supervises the four corresponding vertices of the predicted and ground truth boxes. Our core insight is that by directly targeting the primary source of localization error, the vertices, we can deliver a more focused and effective supervisory signal.
The method maps each vertex offset to a two-dimensional arctangent differential (ATD) similarity, as shown in Figure 1. This mapping yields a response that is bounded, smooth, and monotonic. Furthermore, the response remains sensitive and approximately linear for small deviations, which is crucial for promoting precise corner alignment and achieving high IoU thresholds. Conversely, the function saturates gracefully for large deviations. This property prevents unbounded penalties and ensures stable, non-zero gradients even in non-overlapping cases, a characteristic that is particularly beneficial during the initial training stages. Finally, the four vertex similarities are aggregated to form a robust localization signal that suppresses corner drift, maintains informative gradients throughout training, and reduces sensitivity to noisy offsets.
We integrated the proposed ATDIoU into the YOLOv6 [2] framework and evaluated its performance on PASCAL VOC [20,21] and the challenging VisDrone2019 [19] datasets. The empirical results unmistakably demonstrate consistent and substantial improvements in mean average precision (mAP) over representative IoU-based losses, including the recent MPDIoU [15] and WIoU [16]. These findings indicate that explicit vertex supervision, when paired with a carefully designed arctangent differential mapping, provides an effective and complementary alternative to overlap-centric objectives, thereby offering a promising direction for enhancing BBR precision.
Our contributions are summarized as follows:
  • We introduce ATDIoU, an arctangent-differential loss for BBR that augments IoU with corner-wise distance supervision. Unlike overlap-only or center-distance objectives, ATDIoU explicitly constrains the box geometry by penalizing the offsets of the corresponding top-left and bottom-right corners through a bounded arctangent-differential mapping, improving localization stability.
  • We provide empirical analyses of gradient behavior and optimization dynamics, showing that the arctangent-differential mapping yields smoother, non-saturating gradients under localization errors, which helps reduce box drift during training.
  • Extensive experiments verify that integrating ATDIoU into a modern detector, such as YOLOv6, YOLOV10, achieves consistent mAP gains on both the PASCAL VOC and the VisDrone2019 datasets.

2. Related Work

2.1. IoU-Based BBR Losses

The evolution of BBR losses has been largely dominated by the IoU metric and its numerous extensions, owing to its intuitive geometric interpretation and direct alignment with evaluation metrics. The standard IoU loss, defined as L I o U = 1 I o U , suffers from vanishing gradients when two boxes do not overlap. To address this, GIoU [11] extends IoU by penalizing the area of the smallest enclosing box, thereby ensuring non-zero gradients when the boxes do not overlap. Nevertheless, the enclosing-box area is a global scalar that aggregates multiple geometric discrepancies into a single quantity. Consequently, different misalignment modes can yield similar enclosing areas, making the resulting gradients less informative for vertex-level correction, particularly when the dominant error is corner offset rather than pure center displacement. However, ATDIoU addresses this limitation by introducing vertex-direction supervision, which decomposes the mismatch into deviations of corresponding vertices. The proposed arctangent-differential mapping yields a bounded, continuous penalty, preventing unstable updates caused by extreme outliers while retaining a usable optimization signal in non-overlapping regimes. Importantly, because the penalty is defined at the vertex level, the gradients carry explicit directional information, encouraging each predicted corner to move toward its ground truth counterpart.
The subsequent development of DIoU [12] marked a significant step forward by incorporating the normalized central distance between the two boxes. This directly minimizes the distance between centers, leading to much faster convergence. CIoU [12] built upon DIoU by adding a consistency term for the aspect ratio, further improving the accuracy of the regressed boxes. Despite these improvements, both DIoU and CIoU can be influenced by the aspect ratio term in ways that do not always perfectly correlate with IoU improvement.
More recent variants have sought to incorporate additional geometric factors. SIoU [13] rethinks the cost function by introducing an angle cost that aims to minimize the angle between the central line and the x- or y-axis. This modification potentially leads to faster convergence and better final performance. EIoU [14] decomposes the loss into distance, overlap, and aspect ratio components, and reformulates the aspect ratio term to be more stable during training. WIoU [16] introduces a dynamic focusing mechanism that moderates the penalty on low-quality examples, thereby improving overall generalization. Another recent innovation, MPDIoU [15], leverages the minimum point distance to simplify the similarity calculation while maintaining effectiveness. These methods demonstrate the ongoing effort to refine the geometric supervision of BBR.
Nevertheless, IoU-family losses remain fundamentally overlap-centric. This design leads to two limitations. First, gradients may be brittle when boxes overlap only marginally or are disjoint, as the primary supervisory signal remains tied to the intersection area. Second, slight shifts near box boundaries can still induce unstable updates, because the IoU value is highly sensitive to edge contacts. These issues are especially pronounced for crowded scenes, where precise vertex placement is critical to disentangling adjacent objects.

2.2. Object Detection

Modern object detection frameworks are commonly categorized into three major families: two-stage detectors, one-stage detectors, and Transformer-based detectors. The evolution of these frameworks has correspondingly shaped the requirements and design of BBR losses.
Two-stage detectors, pioneered by the R-CNN series and subsequently refined into Faster R-CNN [5], first generate region proposals and then perform classification and bounding box refinement on those proposals. This paradigm typically achieves robust accuracy, benefitting from multi-scale feature representations, such as Feature Pyramid Networks (FPN) [22,23]. Region-of-interest (RoI) pooling or alignment operations facilitate refined feature extraction for each proposal, which underscores the importance of precise BBR within the second-stage regression module.
One-stage detectors streamline the pipeline by directly regressing bounding boxes from dense sampling locations across the image, thereby prioritizing computational speed and architectural simplicity. Anchor-based methods (e.g., SSD [24], RetinaNet [25], and the YOLO family [2,6,26]) rely on predefined anchor boxes. Techniques such as focal loss [25] were introduced to alleviate the extreme foreground–background class imbalance inherent in dense prediction. In parallel, advances in feature pyramid fusion modules, including PANet [27] and BiFPN [28], have substantially enhanced object detection by improving the bidirectional flow of spatial and semantic information. More recent anchor-free detectors (e.g., CornerNet [29], FCOS [30]) obviate the need for handcrafted anchor designs by predicting keypoints (e.g., corners) or center–size pairs. Further progress has been driven by dynamic label assignment strategies and quality-aware prediction heads, which improve both training stability and detection accuracy. Given that these models rely on direct regression without a secondary refinement stage, the choice of BBR loss is particularly paramount.
Transformer-based detectors such as DETR [31] have recently reconceptualized object detection as a set prediction task, thus eliminating many manual components, including anchors and non-maximum suppression. Although promising, the original DETR exhibits slow convergence. Subsequent variants like Deformable-DETR [32] and DINO [33] ameliorate this deficiency by introducing deformable attention mechanisms and enhanced denoising training, which enables faster convergence and stronger multi-scale reasoning. Real-time extensions such as RT-DETR [31] further demonstrate ongoing efforts to render Transformer-based detection viable for industrial deployment. The end-to-end nature of these models also imposes specific requirements on the regression loss, which must maintain stability and effectiveness within a complex, set-based optimization landscape.

2.3. Vertex-Aware Localization

While the corners of a bounding box represent its most definitive geometric features, their explicit and direct supervision within loss functions remains relatively underexplored, particularly in mainstream detection frameworks. Corner-based supervision is a recognized concept in keypoint-style detectors such as CornerNet [29], which directly predicts heatmaps for the top-left and bottom-right corners. However, this approach pertains to a specific detection paradigm and does not readily translate into a general BBR loss suitable for coordinate-based regression.
In standard regression heads, vertex alignment is typically an implicit target. Losses such as Smooth L1, when applied to box parameters, influence corner placement only indirectly. Although some methods encourage edge alignment through auxiliary constraints, most BBR objectives still aggregate localization errors at the holistic box level, relying on metrics such as overlap, center distance, or aspect ratio. Consequently, explicit modeling of vertex-to-vertex relationships and direct optimization of corner distances have remained relatively niche.
A parallel line of research has explored alternatives to IoU by shifting towards distance-based or distribution-based metrics. For instance, the Normalized Wasserstein Distance (NWD) [17] has been proposed for tiny object detection because it is less sensitive to minor spatial shifts than IoU. Similarly, the Gaussian Wasserstein Distance [18] and the Kullback–Leibler Divergence have been investigated for rotated object detection, where IoU computation becomes more complex and unstable. These works underscore the potential of direct distance measures in addressing the shortcomings of overlap-centric approaches.
Our proposed ATDIoU lies at the intersection of these ideas. It draws inspiration from the benefits of direct geometric supervision seen in keypoint-based methods and from the robustness of distance-based losses. Crucially, it implements these principles in a way that is compatible with mainstream coordinate-based detectors. ATDIoU computes a bounded similarity between corresponding vertices using an arctangent differential mapping. By supervising the specific geometric elements, the corners, that most directly contribute to edge and corner drift, ATDIoU provides a more targeted and effective localization signal that effectively complements existing holistic losses.

3. Methods

In this section, we detail the design philosophy and mathematical formulation of ATDIoU. First, to validate the effectiveness of the proposed regression behavior, we construct simulated experiments that examine scale, and aspect ratio relationships. We then introduce the arctangent-differential function (ATD), which serves as the core mathematical component of our approach. Finally, we formally define the ATDIoU loss, which computes distance similarity between predicted and ground truth boxes through vertex-based arctangent differential mappings, thereby enabling more precise and reliable localization supervision.

3.1. Simulation Experiment

We further evaluate the regression behavior of ATDIoU using simulation experiments [12]. The setup spans the principal relationships between bounding boxes distance, scale, and aspect ratio. Specifically, seven unit-area target boxes (areas fixed at 1) with aspect ratios of 1:4, 1:3, 1:2, 1:1, 2:1, 3:1, and 4:1 are defined. To simplify the setup, the centers of the target boxes are fixed and evenly distributed across 6000 locations, as depicted in Figure 2. (I) Distance: For each location, we sample points uniformly within a circle of radius 0.3 centered at ( 0.5 , 0.5 ) ; the seven anchor scales and seven aspect ratios are uniformly instantiated within this region, yielding both overlapping and non-overlapping cases. (II) Scale: For each location, the anchor-box area is set to one of 0.50, 0.67, 0.75, 1.00, 1.33, 1.50, 2.00. (III) Aspect ratio: For each location and scale, the seven aspect ratios listed above are used, matching the target box settings. In total, we instantiate 6000 × 7 × 7 × 7 = 2 , 058 , 000 regression cases. The total error metric (E) used for evaluation is defined as follows:
E i = n = 1 6000 t x , y , w , h B t , n ( i ) B t , n g t ,
where B t , n i denotes the t-th parameter of the current box for sample n at iteration i, and  B t , n g t is the corresponding ground truth. Training uses stochastic gradient descent (SGD) with a step learning rate scheduler (initial learning rate 0.1; step size 80) for 150 epochs.

3.2. Arctangent Differential Function

The arctangent-differential function has diverse applications in mathematics and engineering [34], as illustrated in Figure 3a. In deep learning, it can serve as a loss component to improve gradient flow and adapt to data distributions, thereby accelerating convergence. This formulation helps models capture complex patterns and enhances generalization. For object detection, using the arctangent-differential function as a BBR loss enables more accurate localization of bounding boxes, improving both accuracy and robustness of the detector.
Inspired by the properties of the arctangent, shown in Figure 3, we propose a BBR loss based on its derivative (the arctangent-differential function). The arctangent is defined as
f ( x ) = arctan ( x ) ,
with derivative
f ( x ) = 1 1 + x 2 ,
where x R is the input, f ( x ) denotes the arctangent, and  f ( x ) its derivative. The graphs of f ( x ) and f ( x ) are shown in Figure 3a and Figure 3b, respectively.

3.3. ATDIoU

ATDIoU measures the similarity between a predicted box and its ground truth by mapping vertex-wise Euclidean offsets to a bounded 2D arctangent-differential score, as shown in Figure 4. After feature extraction, the detector outputs multiple predicted boxes. These boxes are compared with the ground truth using a BBR loss. ATDIoU emphasizes near ground truth predictions by assigning greater optimization weight to boxes that are already close to the target, thereby accelerating the refinement of high-quality predictions and speeding up bounding box convergence. Specifically, after feature extraction by the neural network, a 4-D output ( x , y , w , h ) is produced. As shown in Figure 1, ( x , y ) denotes the center of the predicted box (P), and  ( w , h ) denotes its width and height. The top-left corner a and bottom-right corner c of P are
a = x 1 2 w ; y 1 2 h , c = x + 1 2 w ; y + 1 2 h .
We denote their coordinates as ( x t p , y t p ) and ( x b p , y b p ) , respectively. Let the corresponding ground truth corners be b = ( x t g t , y t g t ) for the top-left and g = ( x b g t , y b g t ) for the bottom-right.
Figure 4. IoU and arctangent differential metrics. IoU between ground truth ( G T ) and predicted box (P). Arctangent differential distance-similarity computed for point pairs ( a , b ) and ( c , g ) .
Figure 4. IoU and arctangent differential metrics. IoU between ground truth ( G T ) and predicted box (P). Arctangent differential distance-similarity computed for point pairs ( a , b ) and ( c , g ) .
Sensors 26 01545 g004
The IoU between a ground truth ( G T ) and a predicted box (P) (Algorithm 1) is defined as
I o U = | G T P | | G T P | .
Let a and c denote the top-left and bottom-right corners of the ground truth, and b and g the corresponding corners of the predicted box. Define the squared Euclidean distances
w t 2 = x t g t x t p 2 2 + y t g t y t p 2 2 ,
w b 2 = x b g t x b p 2 2 + y b g t y b p 2 2 ,
and the arctangent differential weights
D t = x , y arctan ( w t 2 ) = 1 1 + w t 2 ,
D b = x , y arctan ( w b 2 ) = 1 1 + w b 2 .
ATDIoU is then defined as
A T D I o U = I o U α D t β D b ,
where α and β are tuning parameters (both set to 0.03 in our experiments). Consequently, the ATDIoU loss is
L A T D I o U = 1 A T D I o U .
Algorithm 1 ATDIoU as BBR loss
  1:
input:   Predicted box P and ground truth G T coordinates. P = ( x 1 p , y 1 p , x 2 p , y 2 p ) , G T = x 1 g , y 1 g , x 2 g , y 2 g :
  2:
output:  L A T D I o U .
  3:
For the predicted box B, ensuring x 2 p > x 1 p and y 2 p > y 1 p .
  4:
w 1 2 = x 1 p x 1 g 2 + y 1 p y 1 g 2
  5:
w 2 2 = x 2 p x 2 g 2 + y 2 p y 2 g 2
  6:
Calculating area of S g : S g = x 2 g x 1 g y 2 g y 1 g
  7:
Calculating area of S p : S p = x 2 p x 1 p y 2 p y 1 p
  8:
Calculating intersection i between S g and S p : x 1 i = m a x x 1 p , x 1 g , x 2 i = m i n x 2 p , x 2 g y 1 i = m a x y 1 p , y 1 g , y 2 i = m i n y 2 p , y 2 g i = ( x 2 i x 1 i ) ( y 2 i y 1 i ) , i f x 2 i > x 1 i , y 2 i > y 1 i 0 , o t h e r w i s e .
  9:
I o U = i U ,   where U = S g + S p i
10:
A T D I o U = I o U α 1 + w 1 2 β 1 + w 2 2
11:
L A T D I o U = 1 A T D I o U

4. Experiments

To thoroughly evaluate the performance of ATDIoU in object detection scenarios, this section presents extensive comparative experiments. We integrated the proposed loss function into the YOLOv6L framework and conducted evaluations on two datasets: PASCAL VOC and VisDrone2019. Through detailed descriptions of the experimental setup and comprehensive quantitative analyses, we perform a side-by-side comparison of ATDIoU with several mainstream BBR loss functions. In addition, we further examine the robustness and efficacy of our method.

4.1. Datasets

The PASCAL Visual Object Classes (PASCAL VOC) [20,21] benchmark dataset holds considerable significance in the field of computer vision. In this work, we utilize a combination of the VOC 2007 and VOC 2012 versions. The dataset encompasses 20 common object categories, which can be broadly grouped into four classes: humans, animals (e.g., birds, cats, dogs), vehicles (e.g., aeroplanes, bicycles, cars), and indoor objects (e.g., bottles, chairs, sofas). It provides rich annotation types, including image-level classification labels, bounding boxes for object detection, and semantic segmentation masks. Owing to its high annotation quality, well-characterized scenes, and relatively balanced category distribution, PASCAL VOC has become a foundational benchmark for assessing the core performance of object detection algorithms. It is widely used to evaluate a model’s feature extraction capability and its generalization ability under typical ground-level visual conditions.
The VisDrone2019 [19] dataset is a large-scale benchmark designed for unmanned aerial vehicle (UAV) visual analysis, targeting the unique challenges of object detection and tracking from aerial viewpoints. It comprises tens of thousands of static images and video sequences captured by multiple UAV platforms across diverse urban environments, weather conditions (including clear, overcast, and nighttime scenes), and illumination settings, with more than 2.6 million manually annotated bounding boxes. The dataset includes ten common categories of ground and traffic targets, such as pedestrians, vehicles, lorries, and buses. Unlike ground-level datasets such as PASCAL VOC, VisDrone2019 presents a distinct bird’s-eye perspective and is characterized by extremely small object scales, high object density, complex and dynamic backgrounds, and severe occlusion. As a result, VisDrone2019 has become an essential benchmark for evaluating the performance of deep learning models in small object detection, dense scene understanding, and robustness to complex aerial environments.

4.2. Experimental Setup

YOLOv6 [2] is a single-stage object detection framework developed by Meituan’s Visual Intelligence Department. Unlike traditional YOLO variants, it abandons the conventional anchor-based paradigm in favor of an anchor-free design, thereby eliminating the need for extensive anchor-box hyperparameter tuning and reducing computational overhead. At the architectural level, YOLOv6 introduces the EfficientRep backbone, which leverages reparameterization techniques. During training, this backbone adopts a multi-branch structure to enhance feature extraction and improve gradient backpropagation efficiency. At inference time, the branches are merged into a mathematically equivalent single-path structure, which increases hardware computational density and improves memory access efficiency. In addition, YOLOv6 employs an efficient decoupled head that separates classification and regression tasks. Together with advanced label-assignment strategies, quantization-aware training, and knowledge distillation, these design choices further maximize the model’s performance potential.
All experiments were conducted on two NVIDIA RTX 3090 (NVIDIA, Santa Clara, CA, USA) servers running PyTorch 1.13 with CUDA 11.6 on Ubuntu 20.04. We fine-tuned the YOLOv6L pretrained model using a batch size of 48 for 150 epochs. The optimizer used was stochastic gradient descent (SGD) with a learning rate of 0.0032, momentum of 0.843, and weight decay of 0.00036. Data augmentation techniques included rotation, translation, flipping, image decay, and Mosaic augmentation. All experiments were conducted using identical parameter settings.

4.3. Metrics

To comprehensively evaluate the performance of the object detection models, we adopt the standard MS COCO evaluation protocol, which uses mean average precision (mAP) as its primary metric across multiple IoU thresholds and object scales. Specifically, m A P 50 and m A P 75 denote detection accuracy at IoU thresholds of 0.50 and 0.75, reflecting both lenient and strict localization requirements. The overall m A P 95 (i.e., m A P @ [ 0.50 : 0.95 ] ) is computed by averaging the results over the IoU threshold set T = 0.50 , 0.55 , , 0.95 with an interval of 0.05:
m A P 95 = m A P @ [ 0.50 : 0.95 ] = 1 | T | t T 1 C i = 1 C A P i , t , | T | = 10 .
To analyze performance at different object scales, we report the standard COCO metrics: m A P s for small objects (area < 32 2 pixels), m A P m for medium objects ( 32 2 area < 96 2 pixels), and m A P l for large objects (area 96 2 pixels).
All metrics are derived from Precision (P) and Recall (R), defined as:
P = T P T P + F P , R = T P T P + F N .
For each category, the Average Precision ( A P ) is computed as the area under its precision–recall (PR) curve:
A P = 0 1 P ( R ) , d R .
The overall category-level mAP is then obtained by averaging A P across all C categories:
m A P = 1 C i = 1 C A P i .

4.4. Experimental Analysis

To comprehensively validate the effectiveness of ATDIoU, we conduct an in-depth evaluation combining quantitative metrics with qualitative visualization results. The experimental findings demonstrate that ATDIoU not only performs strongly on general detection tasks but also exhibits notable advantages in small object detection, dense scenes, and high-precision localization scenarios.
Firstly, the quantitative results on PASCAL VOC dataset (Table 1) highlight ATDIoU’s superior localization accuracy. ATDIoU achieves an overall mAP of 0.701, outperforming the second-best DIoU by 1.1 percentage points. Notably, on the m A P 75 metric, which reflects a model’s capability under stringent localization requirements, ATDIoU reaches 0.767, substantially surpassing mainstream loss functions such as GIoU and DIoU.
This improvement stems from ATDIoU’s design philosophy: by explicitly supervising the offsets of all four bounding box vertices rather than relying solely on holistic overlap, it more effectively suppresses corner drift. Consequently, ATDIoU performs consistently better at higher IoU thresholds. Moreover, ATDIoU achieves the best performance across all scale metrics, m A P s , m A P m , and m A P l . These gains indicate that the smooth and bounded nature of the arctangent-differential function enables stable gradient flow and robust regression behavior across objects of varying scales.
Secondly, the experimental results on the VisDrone2019 dataset (Table 2) further demonstrate ATDIoU’s robustness in dense small object scenarios. On the highly challenging m A P s metric, ATDIoU obtains a score of 0.135, outperforming both EIoU and SIoU. Since small objects are extremely sensitive to pixel-level positional deviations, traditional IoU-based losses often exhibit unstable gradients. ATDIoU, however, leverages the high sensitivity of the arctangent function near zero to perform fine-grained calibration for small deviations, substantially enhancing recall for tiny objects. Additionally, ATDIoU achieves the best performance on the m A P m metric with a score of 0.377, confirming its ability to provide more discriminative and reliable supervision for small-to-medium objects in complex aerial scenes.
To further validate the generalization capability of the ATDIoU loss function across diverse detection frameworks, we integrated it into the YOLOv10m model and conducted comparative experiments on the VisDrone2019 dataset. As detailed in Table 3, ATDIoU demonstrated superior performance across all four key metrics, precision (0.439), recall (0.353), mAP@50 (0.343), and mAP@95 (0.185), surpassing mainstream loss functions such as EIoU and GIoU. These findings align with the model’s excellent performance within the YOLOv6 framework, collectively confirming the effectiveness and robustness of the proposed method.
Finally, the qualitative inference results (Figure 5, Figure 6 and Figure 7) provide intuitive confirmation of the quantitative improvements. As shown in Figure 6, in general scenes, ATDIoU not only produces bounding boxes that align more accurately with object boundaries but also increases detection confidence (e.g., aircraft confidence rising from 0.92 under WIoU to 0.94). This suggests that precise localization feedback assists the classification branch in learning more discriminative features.
In the low-contrast and occluded scenarios illustrated in Figure 5, ATDIoU effectively mitigates missed detections because the arctangent differential distribution provides stable gradient propagation even in cases of blurred boundaries or non-overlapping boxes. Furthermore, in dense scenarios such as those shown in Figure 7, ATDIoU exhibits strong robustness against mutual interference, successfully separating closely adjacent objects and avoiding the optimization ambiguities commonly encountered with overlap-centric loss functions.
As shown in the gradient norm curves in Figure 8 and the quantitative statistics in Table 4, different BBR losses exhibit distinct optimization behaviors. Overall, all losses maintain relatively large gradient L 2 norms (primarily in the range of 1.6–1.9), indicating that none of them suffers from severe gradient vanishing. However, both the gradient magnitude and its stability vary significantly across methods. ATDIoU consistently stays near the upper envelope of all curves, achieving the highest average gradient norm ( m e a n = 1.7746 ) and the smallest standard deviation and coefficient of variation ( s t d = 0.1853 , C V = 0.1044 ). This superior performance leads to the best composite score (1.6069). These findings imply that ATDIoU provides not only richer gradient information but also a more stable optimization signal throughout training. MPDIoU ranks second, with a slightly lower mean gradient and higher C V , suggesting good but somewhat more oscillatory gradients compared with ATDIoU. In contrast, conventional IoU-based losses, such as GIoU, WIoU, DIoU/EIoU, and SIoU, present lower gradient means (ranging from 1.67 1.70 ) and noticeably larger C V values (up to 0.1431 for SIoU). This corresponds to more frequent and sharper fluctuations visible in the curves. Collectively, these results demonstrate that ATDIoU achieves a superior balance between gradient magnitude and stability, thereby offering a stronger and more reliable optimization driving force for BBR on VisDrone2019.
In summary, by introducing vertex-level differential supervision, ATDIoU effectively overcomes the limitations of traditional IoU-based losses—namely, their insufficient sensitivity to positional bias and their unstable gradient behavior at extreme scales.

5. Discussion

Although ATDIoU consistently improves localization accuracy, two practical considerations should be noted when deploying it as a training objective. First, ATDIoU adds a small number of elementwise operations (arctangent and division) on top of IoU computation, which slightly increases training-time cost but does not introduce extra parameters or change the detector’s inference-time computation. Second, ATDIoU involves two weights ( α and β ). In our experiments, we fixed α = β = 0.03 across both PASCAL VOC and VisDrone2019, but transferring to other detectors or domains may benefit from modest tuning; a practical starting point is to keep α and β equal and adjust them based on the localization–recall trade-off.
Future work will focus on (i) reducing training-time overhead via efficient approximations of the arctangent-differential term, (ii) designing adaptive weighting to eliminate manual selection of α and β when transferring across datasets, and (iii) extending ATDIoU to broader localization settings (e.g., dense detection regimes and other box parameterizations) and combining it with complementary objectives for further gains.

6. Conclusions

The ATDIoU loss exhibits notable advantages in object detection. Compared with MPDIoU, it increases AP50 by 1.4% on the PASCAL VOC and by 0.7% on the VisDrone dataset. The ATDIoU can be attributed to the adopted arctangent differential formulation, which helps mitigate bounding box drift and reduce sensitivity to localization errors. Consequently, ATDIoU encourages the model to learn object locations more effectively, thereby improving overall detection performance. Nonetheless, ATDIoU may face practical challenges, including increased computational overhead and sensitivity to hyperparameters. Future work could focus on algorithmic optimizations to reduce compute cost, systematic studies of applicability across diverse vision tasks, and combinations with complementary loss functions to address remaining limitations.

Author Contributions

Q.T., H.Q. and Y.T. contributed to methodology design, algorithm development, and manuscript writing and editing. X.F., W.H. and M.X. provided financial support, laboratory resources, and experimental guidance. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Doctoral Program of the Young Elite Scientists Sponsorship Program implemented by the China Association for Science and Technology and the Xi’an Institute of Optics and Precision Mechanics, CAS.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

All datasets used in this study are publicly available and ethically compliant. There are no competing interests associated with the data. The code used and analyzed during the current study is available from the corresponding or first author upon reasonable request.

Conflicts of Interest

The authors declare that they have no known competing financial interests or personal relationships that could have influenced the work reported in this paper.

References

  1. Han, B.; Tang, Q.; Song, J.; Wang, Z.; Yang, Y. GAOC: A Gaussian Adaptive Ochiai Loss for Bounding Box Regression. Sensors 2026, 26, 368. [Google Scholar] [CrossRef] [PubMed]
  2. Li, C.; Li, L.; Jiang, H.; Weng, K.; Geng, Y.; Li, L.; Ke, Z.; Li, Q.; Cheng, M.; Nie, W.; et al. YOLOv6: A single-stage object detection framework for industrial applications. arXiv 2022, arXiv:2209.02976. [Google Scholar] [CrossRef]
  3. Tang, Q.; Su, C.; Tian, Y.; Zhao, S.; Yang, K.; Hao, W.; Feng, X.; Xie, M. YOLO-SS: Optimizing YOLO for enhanced small object detection in remote sensing imagery. J. Supercomput. 2025, 81, 303. [Google Scholar] [CrossRef]
  4. Xie, M.; Tang, Q.; Tian, Y.; Feng, X.; Shi, H.; Hao, W. DCN-YOLO: A Small-Object Detection Paradigm for Remote Sensing Imagery Leveraging Dilated Convolutional Networks. Sensors 2025, 25, 2241. [Google Scholar] [CrossRef] [PubMed]
  5. Wang, S. Effectiveness of traditional augmentation methods for rebar counting using UAV imagery with Faster R-CNN and YOLOv10-based transformer architectures. Sci. Rep. 2025, 15, 33702. [Google Scholar] [CrossRef] [PubMed]
  6. Önal, O.; Dandıl, E. Unsafe-Net: YOLO v4 and ConvLSTM based computer vision system for real-time detection of unsafe behaviours in workplace. Multimed. Tools Appl. 2025, 84, 34967–34993. [Google Scholar] [CrossRef]
  7. He, Q.; Xu, A.; Ye, Z.; Zhou, W.; Cai, T. Object detection based on lightweight YOLOX for autonomous driving. Sensors 2023, 23, 7596. [Google Scholar] [CrossRef] [PubMed]
  8. Shi, Z.; Hu, J.; Ren, J.; Ye, H.; Yuan, X.; Ouyang, Y.; He, J.; Ji, B.; Guo, J. HS-FPN: High frequency and spatial perception FPN for tiny object detection. In Proceedings of the AAAI Conference on Artificial Intelligence, Philadelphia, PA, USA, 25 February–4 March 2025; Volume 39, pp. 6896–6904. [Google Scholar]
  9. Wu, D.; Shang, M.; Luo, X.; Wang, Z. An L 1-and-L 2-norm-oriented latent factor model for recommender systems. IEEE Trans. Neural Netw. Learn. Syst. 2021, 33, 5775–5788. [Google Scholar] [CrossRef] [PubMed]
  10. Jun, J.H.; Lee, J.S. Effective mitigation of impulsive vibration in satellite electronic units using acoustic black hole structure. Appl. Acoust. 2024, 225, 110165. [Google Scholar] [CrossRef]
  11. Rezatofighi, H.; Tsoi, N.; Gwak, J.; Sadeghian, A.; Reid, I.; Savarese, S. Generalized intersection over union: A metric and a loss for bounding box regression. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–17 June 2019; pp. 658–666. [Google Scholar]
  12. Zheng, Z.; Wang, P.; Liu, W.; Li, J.; Ye, R.; Ren, D. Distance-IoU loss: Faster and better learning for bounding box regression. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 12993–13000. [Google Scholar]
  13. Zhang, Y.F.; Ren, W.; Zhang, Z.; Jia, Z.; Wang, L.; Tan, T. Focal and efficient IOU loss for accurate bounding box regression. arXiv 2021, arXiv:2101.08158. [Google Scholar] [CrossRef]
  14. Gevorgyan, Z. SIoU loss: More powerful learning for bounding box regression. arXiv 2022, arXiv:2205.12740. [Google Scholar] [CrossRef]
  15. Ma, S.; Xu, Y. Mpdiou: A loss for efficient and accurate bounding box regression. arXiv 2023, arXiv:2307.07662. [Google Scholar] [CrossRef]
  16. Tong, Z.; Chen, Y.; Xu, Z.; Yu, R. Wise-IoU: Bounding box regression loss with dynamic focusing mechanism. arXiv 2023, arXiv:2301.10051. [Google Scholar]
  17. Wang, J.; Xu, C.; Yang, W.; Yu, L. A normalized Gaussian Wasserstein distance for tiny object detection. arXiv 2021, arXiv:2110.13389. [Google Scholar]
  18. Yang, X.; Yan, J.; Ming, Q.; Wang, W.; Zhang, X.; Tian, Q. Rethinking rotated object detection with gaussian wasserstein distance loss. In International Conference on Machine Learning; PMLR: Cambridge, MA, USA, 2021; pp. 11830–11841. [Google Scholar]
  19. Du, D.; Zhu, P.; Wen, L.; Bian, X.; Lin, H.; Hu, Q.; Peng, T.; Zheng, J.; Wang, X.; Zhang, Y.; et al. VisDrone-DET2019: The vision meets drone object detection in image challenge results. In Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops, Seoul, Republic of Korea, 27–28 October 2019. [Google Scholar]
  20. Everingham, M.; Eslami, S.M.A.; Gool, L.V.; Williams, C.K.I.; Winn, J.; Zisserman, A. The pascal visual object classes challenge: A retrospective. Int. J. Comput. Vis. 2015, 111, 98–136. [Google Scholar] [CrossRef]
  21. Everingham, M.; Gool, L.V.; Williams, C.K.I.; Winn, J.; Zisserman, A. The pascal visual object classes (voc) challenge. Int. J. Comput. Vis. 2010, 88, 303–338. [Google Scholar] [CrossRef]
  22. Xu, J.; Ren, H.; Cai, S.; Zhang, X. An improved faster R-CNN algorithm for assisted detection of lung nodules. Comput. Biol. Med. 2023, 153, 106470. [Google Scholar] [CrossRef] [PubMed]
  23. Qiao, S.; Chen, L.C.; Yuille, A. Detectors: Detecting objects with recursive feature pyramid and switchable atrous convolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 10213–10224. [Google Scholar]
  24. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. Ssd: Single shot multibox detector. In Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Proceedings, Part I 14; Springer International Publishing: Berlin/Heidelberg, Germany, 2016; pp. 21–37. [Google Scholar]
  25. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollar, P. Focal loss for dense object detection. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
  26. Redmon, J.; Farhadi, A. Yolov3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar] [CrossRef]
  27. Wang, K.; Liew, J.H.; Zou, Y.; Zhou, D.; Feng, J. Panet: Few-shot image semantic segmentation with prototype alignment. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 9197–9206. [Google Scholar]
  28. Chen, J.; Mai, H.; Luo, L.; Chen, X.; Wu, K. Effective feature fusion network in BIFPN for small object detection. In Proceedings of the 2021 IEEE International Conference on Image Processing (ICIP); IEEE: Piscataway, NJ, USA, 2021; pp. 699–703. [Google Scholar]
  29. Law, H.; Deng, J. Cornernet: Detecting objects as paired keypoints. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 734–750. [Google Scholar]
  30. Tian, Z.; Shen, C.; Chen, H.; He, T. FCOS: Fully convolutional one-stage object detection. In Proceedings of the IEEE International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 9627–9636. [Google Scholar]
  31. Zhao, Y.; Lv, W.; Xu, S.; Wei, J.; Wang, G.; Dang, Q.; Liu, Y.; Chen, J. Detrs beat yolos on real-time object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 16965–16974. [Google Scholar]
  32. Zhu, X.; Su, W.; Lu, L.; Li, B.; Wang, X.; Dai, J. Deformable detr: Deformable transformers for end-to-end object detection. arXiv 2020, arXiv:2010.04159. [Google Scholar]
  33. Zhang, D.; Li, F.; Liu, S.; Zhang, L.; Su, H.; Zhu, J.; Ni, H.Y.-S.L.M. Hino: Det with improved denoising anchor boxes for end-to-end object detection. In Proceedings of the International Conference on Learning Representations, Virtual Event, 25–29 April 2023. [Google Scholar]
  34. Kong, Y.; Tian, D.; Xiu, J.; Che, X. Nonlinear tracking differentiator based on feedforward and arctangent function. In Proceedings of the 2020 39th Chinese Control Conference (CCC); IEEE: Piscataway, NJ, USA, 2020; pp. 2968–2973. [Google Scholar]
Figure 1. Illustration of the ATDIoU computation. S 1 , S 2 , and S 3 denote the regions indicated by the arrows; their areas are computed from the widths and heights of the corresponding bounding boxes. D denotes the distance between the corresponding corner points of the two bounding boxes.
Figure 1. Illustration of the ATDIoU computation. S 1 , S 2 , and S 3 denote the regions indicated by the arrows; their areas are computed from the widths and heights of the corresponding bounding boxes. D denotes the distance between the corresponding corner points of the two bounding boxes.
Sensors 26 01545 g001
Figure 2. Anchor locations and target boxes in the simulation. (a) all cases. (b) main case.
Figure 2. Anchor locations and target boxes in the simulation. (a) all cases. (b) main case.
Sensors 26 01545 g002
Figure 3. Cartesian coordinate systems for the arctangent and arctangent-differential functions. (a) arctangent function. (b) arctangent-differential function.
Figure 3. Cartesian coordinate systems for the arctangent and arctangent-differential functions. (a) arctangent function. (b) arctangent-differential function.
Sensors 26 01545 g003
Figure 5. Inference performance across BBR losses on small and low-contrast targets.
Figure 5. Inference performance across BBR losses on small and low-contrast targets.
Sensors 26 01545 g005
Figure 6. Inference performance on PASCAL VOC across BBR losses. ATDIoU achieves more accurate localization and higher detection confidence than alternative losses.
Figure 6. Inference performance on PASCAL VOC across BBR losses. ATDIoU achieves more accurate localization and higher detection confidence than alternative losses.
Sensors 26 01545 g006
Figure 7. Inference performance across BBR losses on the VisDrone2019.
Figure 7. Inference performance across BBR losses on the VisDrone2019.
Sensors 26 01545 g007
Figure 8. Gradient update performance of BBR losses.
Figure 8. Gradient update performance of BBR losses.
Sensors 26 01545 g008
Table 1. Performance of BBR losses on PASCAL VOC, including the mean, variance, and 95% confidence interval of ATDIoU.
Table 1. Performance of BBR losses on PASCAL VOC, including the mean, variance, and 95% confidence interval of ATDIoU.
mAP s mAP m mAP l mAP 95 mAP 50 mAP 75
EIoU 0.274 ± 0.002 0.572 ± 0.005 0.788 ± 0.002 0.689 ± 0.001 0.873 ± 0.003 0.747 ± 0.003
GIoU 0.267 ± 0.003 0.570 ± 0.003 0.790 ± 0.001 0.693 ± 0.002 0.873 ± 0.004 0.755 ± 0.004
SIoU 0.268 ± 0.003 0.559 ± 0.002 0.786 ± 0.003 0.685 ± 0.003 0.872 ± 0.002 0.750 ± 0.000
MPDIoU 0.265 ± 0.002 0.563 ± 0.002 0.787 ± 0.001 0.684 ± 0.003 0.871 ± 0.002 0.749 ± 0.001
DIoU 0.266 ± 0.001 0.566 ± 0.002 0.787 ± 0.002 0.690 ± 0.000 0.878 ± 0.001 0.746 ± 0.003
WIoU 0.263 ± 0.001 0.562 ± 0.003 0.787 ± 0.003 0.692 ± 0.004 0.878 ± 0.004 0.755 ± 0.002
ATDIoU 0.283 ± 0.002 0.57 ± 0.003 0.798 ± 0.001 0.70 ± 0.000 0.886 ± 0.003 0.767 ± 0.000
95 % C I 0.283 ± 0.005 0.57 ± 0.007 0.798 ± 0.002 0.70 ± 0.000 0.886 ± 0.007 0.767 ± 0.000
Table 2. Performance of BBR losses on the VisDrone2019, including the mean, variance, and 95% confidence interval of ATDIoU.
Table 2. Performance of BBR losses on the VisDrone2019, including the mean, variance, and 95% confidence interval of ATDIoU.
mAP s mAP m mAP l mAP 95 mAP 50 mAP 75
EIoU0.130 ± 0.00090.374 ± 0.00410.614 ± 0.00300.238 ± 0.00110.396 ± 0.00310.238 ± 0.0028
GIoU0.126 ± 0.00350.373 ± 0.00510.631 ± 0.00090.239 ± 0.00240.396 ± 0.00350.246 ± 0.0037
SIoU0.134 ± 0.00310.364 ± 0.00330.621 ± 0.00350.236 ± 0.00340.397 ± 0.00190.240 ± 0.0009
MPDIoU0.129 ± 0.00170.367 ± 0.00290.621 ± 0.00100.234 ± 0.00310.395 ± 0.00200.239 ± 0.0012
DIoU0.130 ± 0.00280.372 ± 0.00290.611 ± 0.00200.236 ± 0.00120.397 ± 0.00110.236 ± 0.0036
WIoU0.131 ± 0.00230.373 ± 0.00300.621 ± 0.00260.244 ± 0.00350.403 ± 0.00400.245 ± 0.0019
ATDIoU0.135 ± 0.00180.377 ± 0.00060.626 ± 0.00270.242 ± 0.00050.404 ± 0.00260.247 ± 0.0002
95 % C I 0.135 ± 0.00450.377 ± 0.00150.626 ± 0.00680.242 ± 0.00120.404 ± 0.00650.247 ± 0.0005
Table 3. Performance of YOLOv10m integrated with BBR losses on the VisDrone2019.
Table 3. Performance of YOLOv10m integrated with BBR losses on the VisDrone2019.
Precision Recall mAP 50 mAP 95
NWD0.4250.3410.3350.175
GWD0.4310.3460.3370.179
EIoU0.4300.3480.3390.182
GIoU0.4290.3490.3380.180
SIoU0.4260.3430.3330.178
DIoU0.4330.3480.3390.181
ATDIoU0.4390.3530.3430.185
Table 4. Statistical comparison of BBR losses: mean, standard deviation (Std), coefficient of variation (CV), and comprehensive score.
Table 4. Statistical comparison of BBR losses: mean, standard deviation (Std), coefficient of variation (CV), and comprehensive score.
Mean ↑Std ↓CV ↓Score ↑
MPDIOU1.74760.18900.10811.5771
GIoU1.70160.19520.11471.5265
WIoU1.68620.21720.12881.4938
DIoU1.67570.22820.13621.4748
EIoU1.67570.22820.13621.4748
SIoU1.66780.23860.14311.4590
ATDIoU1.77460.18530.10441.6069
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Tang, Q.; Qiang, H.; Tian, Y.; Feng, X.; Hao, W.; Xie, M. ATDIOU: Arctangent Differential Loss Function for Bounding Box Regression. Sensors 2026, 26, 1545. https://doi.org/10.3390/s26051545

AMA Style

Tang Q, Qiang H, Tian Y, Feng X, Hao W, Xie M. ATDIOU: Arctangent Differential Loss Function for Bounding Box Regression. Sensors. 2026; 26(5):1545. https://doi.org/10.3390/s26051545

Chicago/Turabian Style

Tang, Qiang, Hao Qiang, Yuan Tian, Xubin Feng, Wei Hao, and Meilin Xie. 2026. "ATDIOU: Arctangent Differential Loss Function for Bounding Box Regression" Sensors 26, no. 5: 1545. https://doi.org/10.3390/s26051545

APA Style

Tang, Q., Qiang, H., Tian, Y., Feng, X., Hao, W., & Xie, M. (2026). ATDIOU: Arctangent Differential Loss Function for Bounding Box Regression. Sensors, 26(5), 1545. https://doi.org/10.3390/s26051545

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop