Next Article in Journal
Optimized Lung Nodule Classification Using CLAHE-Enhanced CT Imaging and Swin Transformer-Based Deep Feature Extraction
Previous Article in Journal
No Reproducibility, No Progress: Rethinking CT Benchmarking
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

DBA-YOLO: A Dense Target Detection Model Based on Lightweight Neural Networks

1
College of Information Science and Engineering, Hunan Normal University, Changsha 410000, China
2
College of Engineering and Design, Hunan Normal University, Changsha 410000, China
*
Author to whom correspondence should be addressed.
J. Imaging 2025, 11(10), 345; https://doi.org/10.3390/jimaging11100345
Submission received: 18 July 2025 / Revised: 26 September 2025 / Accepted: 29 September 2025 / Published: 4 October 2025
(This article belongs to the Section Computer Vision and Pattern Recognition)

Abstract

Current deep learning-based dense target detection models face dual challenges in industrial scenarios: high computational complexity leading to insufficient inference efficiency on mobile devices, and missed/false detections caused by dense small targets, high inter-class similarity, and complex background interference. To address these issues, this paper proposes DBA-YOLO, a lightweight model based on YOLOv10, which significantly reduces computational complexity through model compression and algorithm optimization while maintaining high accuracy. Key improvements include the following: (1) a C2f PA module for enhanced feature extraction, (2) a parameter-refined BIMAFPN neck structure to improve small target detection, and (3) a DyDHead module integrating scale, space, and task awareness for spatial feature weighting. To validate DBA-YOLO, we constructed a real-world dataset from cigarette package images. Experiments on SKU-110K and our dataset show that DBA-YOLO achieves 91.3% detection accuracy (1.4% higher than baseline), with mAP and mAP75 improvements of 2–3%. Additionally, the model reduces parameters by 3.6%, balancing efficiency and performance for resource-constrained devices.

1. Introduction

In recent years, deep learning has markedly improved the efficiency of object detection, delivering superior performance and notable success over traditional methods. This paradigm shift has shown strength across diverse areas such as autonomous driving [1], face recognition [2], and text recognition [3], and it has become integral to many downstream applications. Despite the revolutionary impact of deep learning in detection, challenges remain; for example, state-of-the-art detectors are often computationally intensive. This demand creates a significant barrier to deployment on resource-constrained devices, a dilemma amplified by the rapid proliferation of mobile platforms such as law-enforcement recorders and fixed-wing UAVs [4]. Against this backdrop, we target dense-object detection on edge devices and emphasize a design orientation that simultaneously pursues high accuracy and high efficiency.
In real-world settings, dense-object detection is especially challenging in retail shelf displays [5], UAV imagery [6], and the cigarette-package images showcased in Figure 1. We focus on methodological advances for dense detection—multi-scale feature fusion, attention-guided aggregation under crowding/occlusion, and real-time decoding—and cite application studies to indicate cross-scene applicability across merchandise/crowd scenarios and UAV-based image analysis [7,8,9]. These scenarios typically contain many similar or identical instances packed at high density, often with occlusions. As density increases, detectors must preserve fine-grained localization cues and remain robust to severe overlaps and complex background clutter.
Convolutional Neural Networks (CNNs) have demonstrated strong capabilities in image understanding and feature representation [10]. By extracting key visual features, CNN-based models can significantly improve detection performance. Dense-object detectors are broadly categorized into two paradigms: two-stage and single-stage. Two-stage detectors follow the cascade of “region proposal—feature extraction—classification”, typified by the R-CNN family; although accurate, their reliance on proposal generation and feature resampling constrains real-time performance in dense scenes. In contrast, single-stage detectors adopt dense prediction and directly perform classification and regression without pre-generated proposals. In practice, detection systems are often deployed on edge/mobile devices with limited computing power. (Zaidi et al., 2022) [11]. Large models struggle under such constraints, so lightweight design (e.g., efficient operators and compact backbones) and model compression (pruning/quantization) have become key directions. Although these techniques reduce complexity, maintaining accuracy in dense small-object settings is still challenging because aggressive compression can attenuate early high-resolution cues and destabilize localization. To mitigate this, handcrafted lightweight networks such as MobileNet (Sandler et al., 2018) [12], ShuffleNet (Ma et al., 2018) [13], and GhostNet (Han et al., 2020) [14] optimize convolutions to build efficient architectures suitable for constrained devices. We keep the exposition method-centric and summarize application works briefly to avoid drifting off theme.
In dense retail shelves and cigarette package scenes, objects are tightly arranged with high spatial proximity and similar appearance, making boundary discrimination difficult; the coexistence of multiple scales in a single image further increases the challenge. Consequently, directly transferring lightweight models designed for low-density settings can be suboptimal. Our design adheres to three principles: (i) preserve and exploit early high-resolution information; (ii) perform direction-balanced and cost-aware multi-scale fusion; and (iii) employ attention-guided, foreground-prioritized feature selection to withstand crowding and occlusion. As a representative real-time family, the YOLO series aims to balance speed and accuracy. Given YOLOv10’s scalable architecture and favorable accuracy/latency trade-off (Wang et al., 2024) [15], we take YOLOv10 as the baseline and refine the architecture, feature-extraction strategies, and localization head to meet the stringent requirements of dense scenes on edge devices. In implementation and evaluation, we favor compact, synergistic designs over additive stacking and validate them quantitatively on mainstream benchmarks.
We start from YOLOv10 as the base model. To harvest more informative features, we propose a plug-and-play neck—the weighted bidirectional multi-branch assisted feature pyramid network (BIMAFPN). In BIMAFPN, BiSAF (shallow assisted fusion) maintains shallow backbone information via bidirectional connections to enhance small-object detection, and BiAAF (advanced assisted fusion) enriches output-layer gradients via multidirectional connections; meanwhile, BiFPN improves efficiency and accuracy through bidirectional cross-scale connections with weighted fusion. Second, because feature regions differ in information content and foreground cues should be emphasized, we introduce the C2f PA module that augments C2f with attention, enabling adaptive reweighting by importance to improve feature selection. Finally, for accurate detection in complex dense scenes, we design the attention-enhanced head DyDHead, which builds on the YOLOv10 head with a dynamic mechanism that combines Dynamic Deformable Convolution (DyDCNv3) and multiple attentions to strengthen feature-map representation. We validate our approach on SKU110K and a curated cigarette package dataset. In addition, to probe cross-scene generalization, we conduct a small-scale comparative test on the VisDrone dataset with several representative models as supplementary evidence.
Our contributions are as follows:
  • We propose an improved C2f PA module in the backbone that adaptively adjusts feature weights according to their importance, thereby improving feature extraction. Unlike simply stacking generic attention, C2f PA prioritizes foreground information and selectively enhances shallow, fine-grained cues, targeting the early-stage erasure problem common to dense small objects.
  • We present DyDHead, an improved detection head built upon YOLOv10 that integrates novel dynamic convolution, adaptive feature enhancement, and multi-scale semantic awareness for more accurate target characterization in complex scenes, yielding significant performance gains. DyDHead combines dynamic deformable sampling with hierarchical attention to alleviate localization instability under occlusion/overlap, while a lightweight path design keeps the extra overhead controlled.
  • We propose BIMAFPN, a weighted bidirectional multi-branch assisted FPN that combines BiFPN with auxiliary branches for richer interaction and fusion. BiSAF preserves shallow information for small-object sensitivity; BiAAF enriches output-layer gradients via multidirectional links; and BiFPN provides learnable, bidirectional cross-scale fusion to improve efficiency and accuracy while reducing parameters. Unlike directly concatenating a generic neck, BIMAFPN employs a “shallow-fidelity + high-level gain” dual-assist pathway explicitly tailored to dense small objects and supplies features matched to the detection head.
  • We build a practical cigarette package dataset for testing, comprising 1073 images and 50,173 instances at 960 × 1280 resolution. As an application-neutral dense benchmark, it supports reproducible evaluation for methods targeting crowding and small objects. We plan to expand it to 3000 real images and, with augmentation, to 5000 images and 200,000 instances for public release.

2. Related Work

2.1. Real-Time Dense Small-Object Detection

Research on real-time, dense small-object detection has progressed along two lines and is gradually converging. On the YOLO line, designs emphasize lightweight multi-scale fusion and very low end-to-end latency: early versions broadened scale coverage with SPP and PANet [16,17,18]. YOLOv8 [19] adopts an anchor-free, decoupled head to stabilize optimization. YOLOv10 [15] further restructures the pipeline for edge deployment under tight latency budgets, achieving fewer parameters and lower FLOPs at comparable accuracy while maintaining stable throughput. The subsequent YOLOv11/12 [20,21] introduce attention-centric refinements in the backbone/neck while preserving real-time performance.
In parallel, transformer-based detectors reformulate detection as one-to-one set prediction and use attention to model global dependencies: DETR [22] establishes the end-to-end paradigm via Hungarian matching; Deformable DETR [23] accelerates convergence and improves small-object performance through multi-scale deformable attention; and RT-DETR/RT-DETRv2 [24] advance these ideas toward real-time, deployable regimes with practical attention/sampling strategies and training “freebies”. In addition, DEIM [25] introduces dynamic early inference/early exit into end-to-end detection—using confidence/entropy-based or token-pruning adaptive inference to reduce latency and energy while largely preserving accuracy.
A complementary strand on dense/crowded scenes directly addresses overlap and tiny objects, including crowd-aware de-duplication, one-to-one assignment in fully convolutional detectors, and sparse/dense query mechanisms for high-resolution small objects (e.g., CrowdDet, DeFCN/POTO, QueryDet, DDQ) [26,27,28,29]; building on this, the dynamic inference direction (e.g., DynamicDet [30]) further employs a detection-oriented dynamic architecture with an adaptive router to select inference paths on demand across multi-scale features, and it uses a detection-loss-based exit criterion to realize variable-speed inference, thereby covering a broader accuracy–speed trade-off with a single model. Taken together, three practical principles have emerged for dense scenarios: preserve early high-resolution cues, perform cost-aware and direction-balanced multi-scale fusion, and leverage attention-guided foreground-prioritized feature selection to remain robust under crowding and occlusion.
Positioning and baseline choice. Given our target scenarios with high object density and strict latency/memory budgets, we adopt YOLOv10 as the primary baseline: it offers an excellent speed–accuracy–latency trade-off, lower parameter and FLOP budgets suitable for on-device deployment, and stable throughput under edge constraints while being structurally compatible with lightweight multi-scale fusion necks and heads.

2.2. Multi-Scale Feature Fusion

Multi-scale feature fusion aggregates features at different resolutions to enable cross-scale object detection. FPN [31] adopts a top-down pathway with lateral connections to inject high-level semantics into low-level maps, while PANet [32] complements this with a bottom-up path that strengthens localization and shortens the information flow between shallow and deep layers. Building on these ideas, BiFPN [33] introduces repeated bidirectional fusion with learnable non-negative weights and lightweight refinement, normalizing per-scale contributions and markedly improving the accuracy–efficiency trade-off.
Regarding fusion quality and selectivity, AugFPN [34] narrows semantic gaps across levels via enhancement and attention modules, whereas CE-FPN [35] mitigates channel-reduction loss and cross-scale aliasing through sub-pixel fusion and channel-attention guidance. To further alleviate semantic degradation across non-adjacent levels, AFPN [36] explicitly models long-range interactions between high- and low-level features and employs progressive aggregation, improving cross-scale fusion and small-object performance while maintaining low inference overhead.
Recent studies emphasize lightweight yet selective fusion tailored to dense scenes and edge deployment. HR-FPN [37] preserves high-resolution pathways to enhance tiny-object representation under crowding; in addition, HS-FPN [38] jointly models high-frequency components and spatial perception, further improving cross-scale discrimination in complex backgrounds. Overall, the literature has evolved from static, topology-driven pyramids toward adaptive, attention-enhanced, and computation-efficient fusion: learnable weighting, direction-balanced bidirectional links, and content/attention-aware selection are becoming common practice, providing a viable foundation for dense small-object detection under tight FLOPs and memory budgets.

2.3. Detection Head

A detection head maps fused pyramid features to class probabilities and bounding boxes across multiple scales. Early YOLO variants typically placed per-level heads (e.g., P3–P5) on top of FPN/PANet to perform classification and regression in parallel. As applications moved toward real-time and dense scenarios, research increasingly focused on design choices that directly affect optimization stability and latency, including branch organization for prediction, anchor paradigms, and label assignment strategies.
First, decoupled prediction and anchor-free designs have become mainstream. Separating classification from regression alleviates gradient interference and improves localization quality, and—combined with re-parameterized convolutions and efficient normalizations—reduces inference latency (e.g., YOLOv6, YOLOv7) [39,40]. Meanwhile, moving from handcrafted anchors to anchor-free heads, together with IoU-aware losses and improved label assignment, enhances generalization across scales and aspect ratios; recent work (e.g., YOLOv9, YOLOv10) [15,41] further strengthens consistency training and one-to-one decoding to reduce NMS dependence and suppress false positives.
Second, selective attention/dynamic aggregation and end-to-end query-based detection improve separability under dense or crowded conditions. A representative line (e.g., DyHead) [42] integrates attention into head-side aggregation to perform lightweight selection along level, spatial, and channel dimensions, enhancing small-object representation without excessive FLOPs. In parallel, end-to-end query detectors (e.g., RT-DETR and its v2) [24] couple the “head’’ with a transformer decoder, using one-to-one matching and deployable sampling/decoding strategies to simplify post-processing and improve recall and stability in complex backgrounds.

3. Method

3.1. Overview

In this section, we introduce DBA-YOLO, a lightweight model for dense target detection as shown in Figure 2.
The architecture consists of a backbone, a neck, and a head, in which a modified C2f PA module is integrated to capture normalized object sizes at 80 × 80 , 40 × 40 , and 20 × 20 resolutions, respectively. The neck uses a streamlined multi-scale feature fusion network, BIMAFPN, which enhances the Path Aggregation Network (PAN) by removing the inefficient “Concat” module and adding four feature fusion modules to focus on the object area and mitigate complex background effects. A modified detection head, DyDHead, is used for localization, classification probability, and object scoring, and it consists of three detection layers representing feature maps of different sizes.

3.2. C2f PA Module

The C2f PA module, shown in Figure 3, is an enhanced version of the CSP Bottleneck module with two convolution layers from YOLOv8. The key modification is the introduction of the ParNetAttention attention mechanism, which improves feature selection and multi-scale fusion efficiency. This module consists of three main branches: Channel Attention (SSE), 1 × 1 Convolution, and 3 × 3 Convolution. ParNetAttention enhances feature selection as follows:
Channel Attention (SSE): Global Average Pooling (GAP) is applied to extract global information from the input features. GAP computes the average of each feature map over its entire spatial dimensions (height and width), producing a single value for each feature map. This is followed by a 1 × 1 convolution to capture channel relationships, and the result is passed through a Sigmoid activation to generate attention weights. These weights are then multiplied by the input features to enhance important channels and suppress redundant ones.
x s s e = σ f 1 × 1 GAP ( x ) · x
where GAP ( x ) = 1 H × W i = 1 H j = 1 W x i , j is Global Average Pooling.
A 1 × 1 convolution: The 1 × 1 convolution is used for feature transformation, followed by Batch Normalization (BN) to stabilize the training process.
x 1 × 1 = BN f 1 × 1 ( x )
A 3 × 3 convolution: The 3 × 3 convolution extracts local features, and Batch Normalization (BN) ensures training stability.
x 3 × 3 = BN f 3 × 3 ( x )
Feature fusion and output: The features from the 1 × 1 convolution, 3 × 3 convolution, and Channel Attention (SSE) are fused by element-wise addition, and the output feature map is obtained through SiLU activation:
y = SiLU ( x 1 × 1 + x 3 × 3 + x s s e )
The feature map, after being activated by SiLU, is first divided using the split operation along the channel dimension, reducing the computation and allowing the model to process different subsets of features. Then, multiple Bottleneck modules are applied for deeper feature extraction. Finally, the features from different branches are merged using the concat operation to enhance the feature representation.
This feature fusion strategy enables the model to process information at different scales, while the attention mechanism dynamically adjusts the importance of each channel. This significantly improves feature selection capabilities, balancing accuracy and speed, making it suitable for visual tasks that require both real-time performance and high accuracy.

3.3. Multi-Scale Attention Feature Fusion Network

In object detection tasks, objects may appear at different scales due to factors like distance, angle, and occlusion. A single feature extraction method struggles to capture information across these varying scales, leading to potential information loss. The feature pyramid fusion framework addresses this challenge by processing multi-resolution data to detect objects and features at different sizes. We propose the BIMAFPN architecture (Figure 2), which combines a multi-branch auxiliary feature pyramid network (MAFPN) with a bidirectional feature pyramid network (BiFPN). This approach leverages BiFPN’s multi-scale feature fusion capability while incorporating outputs from the spine and neck to maintain optimal shallow information. The advanced auxiliary fusion module, embedded in the neck, conveys diverse gradient information, enhancing the model’s ability to suppress irrelevant background features, ultimately improving detection accuracy and efficiency.
Our lighter-weight neck design, BiFPN, enhances feature selection through weighted cross-scale connectivity. This reduces redundant computations, unifies multi-scale feature channels, and lowers computational costs by fixing the head_channel, minimizing the number of parameters and operations in the detection head. Multi-scale feature maps are analyzed, enabling the model to capture various levels of detail and emphasize the foreground targets for improved detection.
BiFPN realizes multi-layer feature fusion by bidirectional path, where P l out represents the output feature of level l, while K l denotes the set of input levels connected to level l (e.g., adjacent levels). Resize refers to upsampling or downsampling operations used to align feature resolutions. Conv represents depthwise separable convolution (including Batch Normalization, BN, and activation functions) to reduce computational complexity. Trainable fusion weights w l , k are constrained to be non-negative (e.g., ReLU) and normalized. The BiFPN formula is as follows:
P l out = Conv k K l w l , k ε + k K l w l , k · Resize ( P k in )
As shown in Figure 4, the main goal of BiSAF is to combine deep information with features at the same level and high resolution shallow layers within the backbone network to preserve rich localization details and enhance the spatial representation of the network. In addition, we utilize 1 × 1 convolution to control the number of channels in the shallow information, ensuring that the number of channels in each input is the same, and ensuring that BiFPN operations can be performed without affecting subsequent learning. Let P n 1 , P n and P n + 1 R H × W × C denote the feature mapping at different resolutions, where P n , P n and P n denote the feature layers of the backbone network and the two paths of the BIMAFPN. U denotes an upsampling operation. Down denotes a 3 × 3 downsampling convolution accompanied by a Batch Normalization layer. SiLU denotes a SiLU function, C denotes a 1 × 1 convolution controlling the number of channels. The output after applying BiSAF is as follows:
P n = BiFPN SiLU C Down ( P n 1 ) , SiLU C ( P n ) , U ( P n + 1 )
To improve the interactive use of feature layer information, we integrate the BiAAF module into the deeper layers of the BIMAFPN for multi-scale feature fusion. Specifically, Figure 5 shows the AAF connections in P n , which aggregate information from the shallow high-resolution layer P n + 1 , the shallow low-resolution layer P n 1 , the sibling shallow layer P n , and the previous layer P n 1 . This enables the final output layer to merge data from four distinct layers, significantly enhancing detection performance of medium-sized targets. According to the traditional FPN single-path architecture, we assume that the initial bootstrap information is already embedded in the shallow layers of the BIMAFPN. Therefore, we equalize the number of channels in each layer to ensure that the model obtains different outputs. The outputs after applying BiAAF are as follows:
P n = bifpn Down ( P n 1 ) , Down ( P n 1 ) , P n , U ( P n + 1 )

3.4. DyDHead Schematic

The improved detection head DyDHead, as schematically shown in Figure 6, consists of three main modules: scale-aware attention, spatial-aware attention, and task-aware attention. The input tensor to these modules has three key dimensions: L represents the number of levels in the feature pyramid, corresponding to different resolutions of the feature map; S represents the spatial dimension (height and width), i.e., the number of spatial locations in the feature map; and C represents the channel dimension, i.e., the number of feature channels at each spatial location. Each attention module operates on one of these dimensions: scale-aware attention works on L (levels), spatial-aware attention operates on S (spatial locations), and task-aware attention focuses on C (channels). This separation allows the model to focus on different aspects of the feature map, improving detection efficiency and accuracy.
Given a feature tensor F R L × S × C , the general formula for applying self-attention is
W ( F ) = π ( F ) · F
where π ( · ) is an attention function, but using a fully connected layer for self-attention is computationally too expensive, so we convert the attention function into three consecutive attentions:
W ( F ) = π C ( π S ( π L ( F ) · F ) · F ) · F
where π L ( · ) , π S ( · ) , and π C ( · ) are three different attention functions applied to dimensions L, S, and C. π L ( · ) corresponds to the scale-aware attention module, dynamically fusing features of different scales according to their semantic importance. The specific process corresponds to the Scale Attn module in the DyDHead detail design in Figure 7.
π L ( F ) · F = σ f 1 S C s , c F · F
π S ( · ) corresponds to a spatially aware attention module based on fusion of features to continuously attend to discriminative regions co-occurring in spatial locations and feature layers. Considering the high-dimensionality of the spatial dimension S, we decompose this module into two steps: first, we make the attention learning sparse by using DCNV3, and then we aggregate features across layers at the same spatial location, where DCNV3 is a third-generation version of deformable convolutional networks designed to improve the performance of Convolutional Neural Networks (CNNs) in dealing with changes in the shapes and locations of objects in images. DCNV3 is optimized on the basis of the previous two generations of DCNs by introducing grouping operations and dynamic offsets to further enhance the deformation-awareness of the model.
π S ( F ) · F = 1 L l = 1 L k = 1 K w l , k · F ( l ; p k + Δ p k ; c ) · Δ m k
where K is the number of sparsely sampled positions, p k + Δ p k are the positions offset by a self-learned spatial offset Δ p k for attending to discriminative regions, and Δ m k is a self-learned importance scalar at position p k .
To enable joint learning and generalize the representation of different objects, π C ( · ) corresponds to a task-aware attention module that dynamically turns feature channels on and off to support different tasks.
π C ( F ) · F = max α 1 ( F ) · F c + β 1 ( F ) , α 2 ( F ) · F c + β 2 ( F )
F c denotes the feature slice on the c-th channel, while [ α 1 , α 2 , β 1 , β 2 ] T = θ ( · ) is a hyperfunction for learning the control activation threshold. In the implementation of the function θ ( · ) , the L × S dimensions are first pooled globally on average to reduce the dimensionality, then two fully connected layers and a normalization layer are used, and finally an offset Sigmoid function is applied to normalize the output to the range [ 1 , 1 ] . The exact process corresponds to the Task Attn module in the DyDHead detailed design in Figure 7.

3.5. Loss Function

We follow the default YOLOv10 recipe for losses—CIoU for box regression, Distribution Focal Loss (DFL) for distributional regression, and BCEWithLogits with focal modulation for classification—without changing the default loss weights. The total loss is the weighted sum
L = λ box L box + λ cls L cls + λ dfl L dfl ,
where λ box , λ cls , λ dfl follow the baseline’s dynamic weighting strategy for joint optimization, and task-aligned assignment (TAL) [43] is adopted to reduce classification–localization misalignment.
For box regression, we use the Complete IoU loss (CIoU) [44]:
L CIoU = 1 IoU + ρ 2 ( b , b gt ) c 2 + α v ,
v = 4 π 2 arctan w gt h gt arctan w h 2 , α = v ( 1 IoU ) + v .
where ρ ( · , · ) is the Euclidean distance between box centers, c is the diagonal of the minimum enclosing box, and w , h (resp. w g t , h g t ) are the predicted (resp. ground-truth) width and height.
To obtain finer localization at fractional coordinates, we adopt Distribution Focal Loss (DFL) [45]:
L dfl = i = 0 n ( y i + 1 y g t ) log P ( y i ) + ( y g t y i ) log P ( y i + 1 ) ,
where y g t is the continuous target, y i and y i + 1 are adjacent bins, and P ( y i ) , P ( y i + 1 ) are their predicted probabilities; we follow the baseline binning strategy (finer near small-object ranges).
For classification, BCEWithLogits with focal modulation [46] is used to handle class imbalance under multi-label supervision:
L cls = 1 N k = 1 N α t ( 1 p t ) γ y k log σ ( p k ) + ( 1 y k ) log 1 σ ( p k ) ,
where p k is the logit for sample k, σ ( · ) is the Sigmoid, and p t = σ ( p k ) if y k = 1 (else p t = 1 σ ( p k ) ); α t is a class-balance factor and γ is the focusing parameter. We follow the baseline’s curriculum schedule for γ over training epochs T.
Following TAL [43], the training target confidence couples localization and classification quality via
Score = IoU β · ClsScore 1 β ,
where β [ 0 , 1 ] is a balance factor (default 0.5 ).

4. Experiments

4.1. Dataset

The experiment utilized SKU-110K [5] and a custom cigarette pack dataset. SKU-110K comprises 11,762 images with 173,678 instances, collected from diverse supermarket stores. The dataset underwent a split of 8821 training samples and 2941 test samples. The self-built cigarette package dataset includes over 1000 real images captured in various cities, weather conditions, and times of day, each sized at 960 pixels × 1280 pixels. To augment diversity and prevent overfitting, 50 natural scene images without cigarette packages were included as negative samples. Labeling was performed using labelImg, resulting in 1073 images, 50,173 marker samples, with 887 images in the training set and 186 samples in the test set. As the original cigarette package images are proprietary, mosaic processing was applied to all displayed faces for privacy reasons, as illustrated in Figure 8.

4.2. Experimental Environment and Parameter Settings

The experiment used Python 3.10.14, CUDA 12.1, RTX 4090D GPU (24 GB; NVIDIA Corporation, Santa Clara, CA, USA), and Xeon 8474C CPU (Intel Corporation, Santa Clara, CA, USA). The input image sizes for SKU-110K and cigarette packet datasets were 640 × 640 and 1280 × 1280 , respectively, and mosaic data enhancement was used to increase diversity. Training involves 100 epochs for SKU-110K and 300 epochs for the cigarette packet dataset to achieve optimal weights. Stochastic Gradient Descent (SGD) with momentum of 0.937 and weight decay coefficient of 0.0005 is used for gradient updating. The initial learning rate is 10 2 and the final learning rate is 10 4 with a batch size of 16.

4.3. Evaluation Metrics

The experiment uses evaluation indicators in the field of deep learning object detection, such as precision, recall, and average precision (AP). The formulas are as follows:
Recall = T P T P + F N
Precision = T P T P + F P
A P = 0 1 P ( R ) d R
where T P represents the number of positive samples correctly predicted as positive, F P represents the number of positive samples incorrectly predicted as positive, and F N represents the number of positive samples incorrectly predicted as negative. P represents precision, and R represents recall rate. We used the same evaluation method as COCO [47], which reported the mean average precision (mAP) when IoU = 0.5:0.05:0.95 (The IoU ranges from 0.5 to 0.95 with a step size of 0.05). In addition, we report the A P 50 (IoU = 0.5) and the A P 75 (IoU = 0.75).

4.4. Result

To further investigate the effectiveness and practical deployability of DBA-YOLO for cigarette package target detection, we conduct systematic experiments on the SKU-110K dataset. DBA-YOLO achieves consistent improvements over YOLOv10n in mAP and A P 75 ; meanwhile, compared with the transformer-based RT-DETR-R18, it attains higher accuracy with fewer parameters, reflecting a more favorable accuracy–efficiency trade-off. To comprehensively cover the baseline spectrum for shelf-style dense small-object scenarios, we also include a specialized dense detector—DDQ R-CNN R50 and perform side-by-side evaluation under a unified protocol (640 × 640, single-scale, no TTA); the results show that DBA-YOLO, while remaining lightweight, exhibits robust advantages in both accuracy and efficiency (see Table 1). From the overall comparison in Table 1, DBA-YOLO leads among lightweight models (n/tiny tier), showing more stable localization at the stricter IoU threshold ( A P 75 ); it also maintains comparatively low parameter counts and GFLOPs while matching or surpassing some larger “s” series models in accuracy. These results collectively indicate that DBA-YOLO is feasible for deployment in resource-constrained scenarios.
Figure 9 shows the trend of the results of different models on the same dataset. The experimental results demonstrate that our method consistently outperforms all compared YOLO variants across both mAP@0.5 and mAP@0.5:0.95 metrics. Specifically, our model achieves faster convergence, higher overall detection accuracy, and greater stability throughout training. Notably, in the more rigorous mAP@0.5:0.95 metric, our approach maintains a clear lead over state-of-the-art lightweight models such as YOLOv8s and YOLOv10n, highlighting its superior ability to localize and classify objects precisely. These advantages indicate the strong effectiveness and robustness of our method in practical object detection tasks.
Table 1 highlights the significant improvements of our proposed method, with a 1.9% increase in mean average accuracy m A P and a 2.6% increase in m A P 75 compared to the baseline model, along with a 3.6% reduction in the number of parameters. Additionally, our method shows a corresponding increase in m A P and m A P 75 compared to other YOLO series models. While there is a corresponding increase in the number of parameters in YOLOv8s and YOLOv10s, the results are similar. However, YOLOv8s and YOLOv10s exhibit a significant increase in the number of parameters and GFLOPs. The visual representations in Figure 10 demonstrate the improvement of bounding box fitting on the SKU-110K dataset, which significantly reduces the overlap and distinguishes the boundaries between similar objects. The edge regions are better recognized, indicating that our model performs well in accurate bounding box localization, especially for densely distributed objects.
In order to gain further insight into the effectiveness of the DBA-YOLO model in cigarette package target detection and its real possibility of practical application, this paper conducts experiments on the cigarette package dataset, and the results of the experiments are shown in Table 2. Table 2 shows that our model achieves a significant gain of 1.4% in terms of m A P . A significant gain of 1.2% is also realized for A P 75 . Moreover, compared with the transformer-based RT-DETR-R18, our approach delivers higher accuracy with far fewer parameters, m A P 81.0 vs. 79.4 ( + 1.6 ) , A P 50 99.4 vs. 99.3 ( + 0.1 ) , and A P 75 93.9 vs. 93.1 ( + 0.8 ) , while reducing parameters from 198.73 × 10 5 to 26.16 × 10 5 ( 86.8 % ) . These results further support the practicality of DBA-YOLO for deployment in resource-constrained scenarios.
Figure 11 shows the object detection results on the cigarette package dataset. It can be seen that our model can accurately localize cigarette packages in complex backgrounds. Compared with other object detection algorithms, our model outperforms YOLOv10n by 1.4% in terms of A P 75 , is 1.3% higher than YOLOv11n, and is 1.4% higher than YOLOv12n, which is a significant improvement in performance. Furthermore, comparing to YOLOv10, the improved adaptability will allow deployment of the model in a variety of environments as the model parameters are significantly reduced. It can be transferred to mobile devices such as law enforcement equipment while maintaining reasonable detection accuracy, thus enabling real-time detection.
As shown in Figure 12b,c, the reflected objects from the YOLOv10 glass mirror are recognized as detected, while the DBA-YOLO is unaffected by interference and obtains accurate recognition results. We attribute the improvement to the optimized detection head, which enables the model to focus on cigarette packages and mitigates background interference. Overall, DBA-YOLO achieves the best performance on cigarette package detection, surpassing state-of-the-art baselines in m A P and A P 75 .
Figure 13 shows the comparison of the inference heat maps of YOLOv10 and DBA-YOLO models on the cigarette packet dataset and SKU-110 dataset. In heat map (b), YOLOv10 is weak in capturing local features (e.g., edges) of the target, and the heat map reflects a less precise region, while DBA-YOLO’s heat map (c) significantly reduces the interference of the background and the character’s arm, and the model focuses more on local features (e.g., text and edges) of the target and is able to discriminate the characteristics of the target region in more detail. In heat map (e), although YOLOv10 pays better attention to the box as a whole, some of the hotspots are distributed in regions unrelated to the target (e.g., character’s arms, background edges), and there is a certain amount of distraction; in contrast, DBA-YOLO’s heat map (f) pays more attention to the target region (e.g., the box that is held), with more precise alignment with the edges of the target object, less interference from the background, and a better ability to exclude irrelevant information. In heat map (h), YOLOv10’s hotspot mainly focuses on the center part of the goods on the shelves, but some areas receive less attention, so it fails to completely cover all goods, especially some smaller or inconspicuous items, resulting in missed detection; in contrast, DBA-YOLO’s heat map (i) covers more goods, including small or edge objects that may be overlooked by YOLOv10, with more comprehensive detection, which reduces the possibility of missed detection, and the heat map is more uniform and covers a wider range of items, especially for small and dense objects. Overall, DBA-YOLO outperforms YOLOv10 in terms of target area focus, localized feature capture capability, and comprehensiveness.
In addition to the experiments conducted on the SKU-110K and our custom cigarette package dataset, we further validated the performance of DBA-YOLO on the VisDrone-2019 dataset [48]. VisDrone-2019 is a large-scale UAV imagery dataset featuring rich and densely distributed object detection scenarios, making it particularly suitable for detecting small and dense objects. This dataset includes aerial images captured in various scenes and environments, containing numerous objects of different scales, densities, and partial occlusions.
VisDrone-2019 covers aerial scenes with crowded, small, and partially occluded objects under complex backgrounds. On this benchmark, DBA-YOLO reaches 38.5% mAP@0.5 and 23.4% AP 75 , showing consistent gains over strong baselines (Table 3).
This paper also evaluates the effectiveness of different detection heads to demonstrate the feasibility of DyDHead, as shown in Table 4. All comparisons use the same backbone (YOLOv10n) and neck, the same input size (640), and identical schedule/augmentations; only the detection head differs. In the comparison across multiple heads, DyDHead achieves better detection performance. Although its parameter count may increase relative to some alternatives, the accuracy gains are larger. Compared with SEAMHead, RSCD, TADDH, and LSCD, the metrics m A P , A P 50 , and A P 75 are substantially improved, with a modest parameter increase in certain settings. In practice, DyDHead produces fewer false detections; relative to MultiSEAMHead, it uses fewer parameters while improving accuracy. These results indicate that DyDHead extracts target-region features more precisely, exploits multi-scale information more efficiently, and strengthens small-object detection—particularly under complex backgrounds.
This paper also evaluates the effectiveness of different neck networks to demonstrate the feasibility of BIMAFPN, as shown in Table 5. All comparisons use the same backbone (YOLOv10n) and head, the same input size (640), and identical schedule/augmentations; only the neck differs. In comparisons across multiple necks, BIMAFPN achieves better detection performance while reducing the parameter count relative to most alternatives. Against RCSOSA, m A P and A P 75 are comparable, and BIMAFPN uses roughly half the parameters. Compared with other necks, BIMAFPN not only requires fewer parameters but also yields higher m A P , A P 50 , and A P 75 , indicating that the BIMAFPN module more efficiently exploits multi-scale features and strengthens small-object detection. This is particularly useful for detectors that are sensitive to complex backgrounds.

4.5. Ablation Study and Discussion

By optimizing the performance of the proposed DBA-YOLO model in terms of detection accuracy and speed in the SKU-110K dataset, we select YOLOv10 as the baseline model and analyze the performance of each component, and the experimental results are shown in Table 6.

4.5.1. Effectiveness of BIMAFPN Module

After replacing the original YOLOv10 neck with the proposed BIMAFPN module, the number of parameters is significantly reduced. The SAF (Shallow-Assisted Fusion) mechanism preserves shallow features via bidirectional connections, improving the detection of small objects. Meanwhile, the AAF (Advanced-Assisted Fusion) enhances gradient flow at the output stage through multidirectional connections. Additionally, the BiFPN structure introduces cross-scale bidirectional paths and weighted feature fusion, improving detection accuracy and efficiency. As shown in Table 6, this modification reduces the number of parameters by approximately 30% while achieving a 0.5% improvement in mAP and a 0.8% gain in A P 75 .

4.5.2. Effectiveness of C2f PA Module

The proposed C2f PA feature fusion module adaptively adjusts the weights of input features based on their relative importance, enhancing feature extraction under complex background conditions. With the integration of an attention mechanism, the model’s representational capacity is further improved. As shown in Table 6, although the parameter count increases slightly, the mAP improves by 0.8%, and A P 75 improves by 1.2%, demonstrating the effectiveness of the proposed module.

4.5.3. Effectiveness of DyDHead Module

The enhanced DyDHead detection head incorporates dynamic convolution and multiple attention mechanisms to better capture semantic representations in complex scenes. This improves both the robustness and accuracy of the detection. As observed in Table 6, while the number of parameters increases marginally, the mAP increases by 0.9% and A P 75 improves by 1.4% compared to the baseline, validating the effectiveness of the proposed detection head.

5. Discussion

This work presents DBA-YOLO, a lightweight detector tailored for dense-target scenarios under resource-constrained settings. The experiments highlight several contributions that advance the state of the art. Although the absolute gains are modest, DBA-YOLO achieves 1.2–2.6% improvements on mAP/ AP 75 with fewer parameters and comparable computing resources, strengthening the accuracy–efficiency trade-off for edge deployment; qualitative evidence (Figure 9, Figure 10, Figure 11 and Figure 12) shows fewer false positives and tighter boxes under crowding/overlap, which global metrics may underestimate. First, integrating C2f PA, BIMAFPN, and DyDHead forms a robust framework for enhanced multi-scale feature extraction and fusion, addressing the long-standing difficulty of small-object detection where conventional models often struggle. In particular, the scale-, spatial-, and task-aware dynamic attentions in DyDHead enable more effective capture of salient cues at varying abstraction levels, yielding strong performance in dense detection tasks. Ablations indicate complementary roles—BIMAFPN for cross-scale aggregation, C2f PA for foreground cues, and DyDHead for separability in dense regions—without inflating depth/width, thereby preserving real-time throughput.
Despite these positive results, several limitations remain. The model performs well in controlled environments, but more dynamic settings—such as extreme illumination or heavy occlusion—can still affect accuracy, motivating further work on robustness. Moreover, while attention improves cross-scale accuracy, there is room to optimize detection in highly textured backgrounds and for very small objects. Concretely, this includes two areas: (i) ultra-small objects near image boundaries are occasionally missed on SKU-110K, and (ii) glass reflections in cigarette imagery can induce spurious responses. Going forward, we aim to strengthen early high-resolution pathways for tiny-object recall and explore reflection-robust attention/de-noising while keeping real-time budgets.
Failure case analysis: Figure 14 provides practitioner-oriented examples. In panel (a), glossy film on cigarette packages introduces specular highlights together with partial hand/edge occlusions; near-duplicate appearances at small scales make NMS more conservative, leading to false negatives (red boxes mark missed instances). In panel (b), shelf-edge blur and extreme density cause boundary objects to become truncated or undersized after resizing, again yielding missed detections. These patterns align with our error logs: most failures occur when targets are tiny, partially visible, or strongly reflective.
Improvement directions: We will proceed along three restrained paths: (1) Data and training: We will adopt augmentations closer to real scenes and moderate sample distribution to mitigate biases from reflection, occlusion, and extreme density. (2) Architecture and representation: We will moderately reinforce early high-resolution features and cross-scale interaction, improving the visibility and separability of tiny and boundary objects while preserving real-time performance. (3) Inference and evaluation: We will refine inference settings for ultra-dense shelves and conduct more systematic error decomposition to continuously locate bottlenecks. Across all paths, we adhere to a “lightweight-first, compute-controlled” principle.

6. Conclusions

This paper presents DBA-YOLO, a new approach for dense target detection in complex contexts, utilizing an improved C2f PA module as a backbone feature extraction network. DBA-YOLO is a lightweight network that reduces model parameters and maintains comparable computational complexity, making it suitable for deployment on mobile devices. The BIMAFPN, as a multi-scale attentional feature fusion network, enhances feature extraction and improves detection accuracy. Replacing the original detection head with DyDHead helps the model perform better in complex scenarios. Validation on real cigarette package datasets proves the practical feasibility of DBA-YOLO, which achieves excellent performance on SKU-110K and cigarette package datasets compared with YOLOv10n. With ∼3.6% fewer parameters, DBA-YOLO improves mAP by 1.9% and AP 75 by 2.6% on SKU-110K. On the cigarette package dataset, it achieves an mAP of 81.0% (+1.6), AP 50 of 99.4%, and AP 75 of 93.9%. These results satisfy the detection requirements of complex scenarios. The experimental results confirm that DBA-YOLO outperforms existing dense target detection models in terms of accuracy and local bounding box prediction in complex environments.

Author Contributions

Conceptualization, Z.H.; methodology, Z.H. and Q.T.; software simulation, Z.H. and Q.T.; formal analysis, Q.T. and J.Y.; data curation, Z.H., H.N., and C.L.; writing—original draft preparation, Z.H.; writing—review and editing, Q.T. and J.Y.; supervision, J.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Hunan Provincial Education Science Planning Office, grant number XJK24AJD042. The APC was funded by the authors.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data that support the findings of this study are available from the corresponding author upon reasonable request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Liang, L.; Ma, H.; Zhao, L.; Xie, X.; Hua, C.; Zhang, M.; Zhang, Y. Vehicle detection algorithms for autonomous driving: A review. Sensors 2024, 24, 3088. [Google Scholar] [CrossRef]
  2. EL Fadel, N. Facial Recognition Algorithms: A Systematic Literature Review. J. Imaging 2025, 11, 58. [Google Scholar] [CrossRef]
  3. AlKendi, W.; Gechter, F.; Heyberger, L.; Guyeux, C. Advancements and challenges in handwritten text recognition: A comprehensive survey. J. Imaging 2024, 10, 18. [Google Scholar] [CrossRef]
  4. Haq, M.A. Planetscope nanosatellites image classification using machine learning. Comput. Syst. Sci. Eng. 2022, 42, 1031–1046. [Google Scholar] [CrossRef]
  5. Goldman, E.; Herzig, R.; Eisenschtat, A.; Goldberger, J.; Hassner, T. Precise detection in densely packed scenes. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 5227–5236. [Google Scholar]
  6. Haq, M.A.; Rahaman, G.; Baral, P.; Ghosh, A. Deep learning based supervised image classification using UAV images for forest areas classification. J. Indian Soc. Remote Sens. 2021, 49, 601–606. [Google Scholar] [CrossRef]
  7. Ma, Z.; Liu, D.; Cui, Z.; Zhao, Y. AdaptCD: An adaptive target region-based commodity detection system. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 5486–5495. [Google Scholar]
  8. Alsubai, S.; Dutta, A.K.; Alghayadh, F.; Alamer, B.H.; Pattanayak, R.M.; Ramesh, J.V.N.; Mohanty, S.N. Design of Artificial Intelligence Driven Crowd Density Analysis for Sustainable Smart Cities. IEEE Access 2024, 12, 121983–121993. [Google Scholar] [CrossRef]
  9. Haq, M.A. CNN based automated weed detection system using UAV imagery. Comput. Syst. Sci. Eng. 2022, 42, 837–849. [Google Scholar] [CrossRef]
  10. Sharif Razavian, A.; Azizpour, H.; Sullivan, J.; Carlsson, S. CNN features off-the-shelf: An astounding baseline for recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Columbus, OH, USA, 24–27 June 2014; pp. 806–813. [Google Scholar]
  11. Zaidi, S.S.A.; Ansari, M.S.; Aslam, A.; Kanwal, N.; Asghar, M.; Lee, B. A survey of modern deep learning based object detection models. Digit. Signal Process. 2022, 126, 103514. [Google Scholar] [CrossRef]
  12. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.C. Mobilenetv2: Inverted residuals and linear bottlenecks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 4510–4520. [Google Scholar]
  13. Ma, N.; Zhang, X.; Zheng, H.T.; Sun, J. Shufflenet v2: Practical guidelines for efficient cnn architecture design. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–10 September 2018; pp. 116–131. [Google Scholar]
  14. Han, K.; Wang, Y.; Tian, Q.; Guo, J.; Xu, C.; Xu, C. Ghostnet: More features from cheap operations. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 1580–1589. [Google Scholar]
  15. Wang, A.; Chen, H.; Liu, L.; Chen, K.; Lin, Z.; Han, J.; Ding, G. Yolov10: Real-time end-to-end object detection. Adv. Neural Inf. Process. Syst. 2024, 37, 107984–108011. [Google Scholar]
  16. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  17. Redmon, J.; Farhadi, A. YOLO9000: Better, faster, stronger. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, USA, 21–26 July 2017; pp. 7263–7271. [Google Scholar]
  18. Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y.M. Yolov4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2004.10934. [Google Scholar] [CrossRef]
  19. Varghese, R.; Sambath, M. Yolov8: A novel object detection algorithm with enhanced performance and robustness. In Proceedings of the 2024 International Conference on Advances in Data Engineering and Intelligent Computing Systems (ADICS), Chennai, India, 18–19 April 2024; pp. 1–6. [Google Scholar]
  20. Khanam, R.; Hussain, M. Yolov11: An overview of the key architectural enhancements. arXiv 2024, arXiv:2410.17725. [Google Scholar] [CrossRef]
  21. Tian, Y.; Ye, Q.; Doermann, D. Yolov12: Attention-centric real-time object detectors. arXiv 2025, arXiv:2502.12524. [Google Scholar]
  22. Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; Zagoruyko, S. End-to-end object detection with transformers. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 213–229. [Google Scholar]
  23. Zhu, X.; Su, W.; Lu, L.; Li, B.; Wang, X.; Dai, J. Deformable detr: Deformable transformers for end-to-end object detection. arXiv 2021, arXiv:2010.04159. [Google Scholar]
  24. Lv, W.; Zhao, Y.; Chang, Q.; Huang, K.; Wang, G.; Liu, Y. Rt-detrv2: Improved baseline with bag-of-freebies for real-time detection transformer. arXiv 2024, arXiv:2407.17140. [Google Scholar]
  25. Huang, S.; Lu, Z.; Cun, X.; Yu, Y.; Zhou, X.; Shen, X. Deim: Detr with improved matching for fast convergence. In Proceedings of the Computer Vision and Pattern Recognition Conference, Nashville, TN, USA, 11–15 June 2025; pp. 15162–15171. [Google Scholar]
  26. Chu, X.; Zheng, A.; Zhang, X.; Sun, J. Detection in crowded scenes: One proposal, multiple predictions. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 12214–12223. [Google Scholar]
  27. Wang, J.; Song, L.; Li, Z.; Sun, H.; Sun, J.; Zheng, N. End-to-end object detection with fully convolutional network. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Online, 19–25 June 2021; pp. 15849–15858. [Google Scholar]
  28. Yang, C.; Huang, Z.; Wang, N. QueryDet: Cascaded sparse query for accelerating high-resolution small object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 13668–13677. [Google Scholar]
  29. Zhang, S.; Wang, X.; Wang, J.; Pang, J.; Lyu, C.; Zhang, W.; Luo, P.; Chen, K. Dense distinct query for end-to-end object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 7329–7338. [Google Scholar]
  30. Lin, Z.; Wang, Y.; Zhang, J.; Chu, X. Dynamicdet: A unified dynamic architecture for object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 6282–6291. [Google Scholar]
  31. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature Pyramid Networks for Object Detection. In Proceedings of the CVPR, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  32. Liu, S.; Qi, L.; Qin, H.; Shi, J.; Jia, J. Path Aggregation Network for Instance Segmentation. In Proceedings of the CVPR, Honolulu, HI, USA, 21–26 July 2018; pp. 8759–876. [Google Scholar]
  33. Tan, M.; Pang, R.; Le, Q.V. EfficientDet: Scalable and Efficient Object Detection. In Proceedings of the CVPR, Seattle, WA, USA, 13–19 June 2020; pp. 10781–10790. [Google Scholar]
  34. Guo, C.; Fan, B.; Zhang, Q.; Liu, X.; Zhang, M.; Lu, H. AugFPN: Improving Multi-scale Feature Learning for Object Detection. In Proceedings of the CVPR, Online, 14–19 June 2020; pp. 12559–12568. [Google Scholar]
  35. Luo, Y.; Cao, X.; Zhang, J.; Guo, J.; Shen, H.; Wang, T.; Feng, Q. CE-FPN: Enhancing Channel Information for Object Detection. Multimed. Tools Appl. 2022, 81, 30685–30704. [Google Scholar] [CrossRef]
  36. Yang, G.; Lei, J.; Zhu, Z.; Cheng, S.; Feng, Z.; Liang, R. AFPN: Asymptotic Feature Pyramid Network for Object Detection. In Proceedings of the 2023 IEEE International Conference on Systems, Man, and Cybernetics (SMC), Vienna, Austria, 5–8 October 2023; pp. 2184–2189. [Google Scholar]
  37. Chen, Z.; Ji, H.; Zhang, Y.; Zhu, Z.; Li, Y. High-Resolution Feature Pyramid Network for Small Object Detection. IEEE Trans. Circuits Syst. Video Technol. 2024, 34, 475–489. [Google Scholar] [CrossRef]
  38. Shi, Z.; Hu, J.; Ren, J.; Ye, H.; Yuan, X.; Ouyang, Y.; He, J.; Ji, B.; Guo, J. High Frequency and Spatial Perception Feature Pyramid Network for Object Detection. In Proceedings of the AAAI Conference on Artificial Intelligence, Philadelphia, PA, USA, 27 February–4 March 2025. [Google Scholar]
  39. Li, C.; Li, L.; Jiang, H.; Weng, K.; Geng, Y.; Li, L.; Ke, Z.; Li, Q.; Cheng, M.; Nie, W.; et al. YOLOv6: A Single-Stage Object Detection Framework for Industrial Applications. arXiv 2022, arXiv:2209.02976. [Google Scholar] [CrossRef]
  40. Wang, C.Y.; Bochkovskiy, A.; Liao, H.-Y.M. YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. arXiv 2023, arXiv:2207.02696. [Google Scholar]
  41. Wang, C.Y.; Yeh, I.H.; Mark Liao, H.Y. YOLOv9: Learning What You Want to Learn Using Programmable Gradient Information. arXiv 2024, arXiv:2402.13616. [Google Scholar] [CrossRef]
  42. Dai, Z.; Cai, Q.; Lin, Y.; Chen, Y.; Ding, M.; Xie, E.; Zhang, W.; Hu, H.; Dai, J. Dynamic Head: Unifying Object Detection Heads with Attentions. In Proceedings of the CVPR, Online, 19–25 June 2021. [Google Scholar]
  43. Zhang, T.; Cheng, C.; Lu, C.; Li, K.; Yang, X.; Li, G.; Zhang, L. TOOD: Task-aligned One-stage Object Detection. In Proceedings of the ICCV, Seoul, Republic of Korea, 10–17 October 2021. [Google Scholar]
  44. Zheng, Z.; Wang, P.; Liu, W.; Li, J.; Ye, R.; Ren, D. Distance-IoU loss: Faster and better learning for bounding box regression. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 12993–13000. [Google Scholar]
  45. Li, X.; Wang, W.; Wu, L.; Chen, S.; Hu, X.; Li, J.; Tang, J.; Yang, J. Generalized Focal Loss: Learning Qualified and Distributed Bounding Boxes. In Proceedings of the NeurIPS, Online, 6–12 December 2020. [Google Scholar]
  46. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal Loss for Dense Object Detection. In Proceedings of the ICCV, Venice, Italy, 22–29 October 2017. [Google Scholar]
  47. Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft coco: Common objects in context. In Proceedings of the Computer Vision–ECCV 2014: 13th European Conference, Zurich, Switzerland, 6–12 September 2014; Proceedings, Part V; Springer: Berlin/Heidelberg, Germany, 2014; pp. 740–755. [Google Scholar]
  48. Zhu, P.; Wen, L.; Du, D.; Bian, X.; Fan, H.; Hu, Q.; Ling, H. Detection and tracking meet drones challenge. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 7380–7399. [Google Scholar] [CrossRef]
  49. Yu, Z.; Huang, H.; Chen, W.; Su, Y.; Liu, Y.; Wang, X. Yolo-facev2: A scale and occlusion aware face detector. arXiv 2022, arXiv:2208.02019. [Google Scholar] [CrossRef]
  50. Chen, X.; Hu, D.; Cheng, Y.; Chen, S.; Xiang, J. EDT-YOLOv8n-Based Lightweight Detection of Kiwifruit in Complex Environments. Electronics 2025, 14, 147. [Google Scholar] [CrossRef]
  51. Yan, C.; Xu, E. ECM-YOLO: A real-time detection method of steel surface defects based on multiscale convolution. J. Opt. Soc. Am. A 2024, 41, 1905–1914. [Google Scholar] [CrossRef]
  52. Wang, H.; Liu, X.; Song, L.; Zhang, Y.; Rong, X.; Wang, Y. Research on a train safety driving method based on fusion of an incremental clustering algorithm and lightweight shared convolution. Sensors 2024, 24, 4951. [Google Scholar] [CrossRef] [PubMed]
  53. Dong, S.; Wang, L.; Du, B.; Meng, X. ChangeCLIP: Remote sensing change detection with multimodal vision-language representation learning. ISPRS J. Photogramm. Remote Sens. 2024, 208, 53–69. [Google Scholar] [CrossRef]
  54. Chen, J.; Mai, H.; Luo, L.; Chen, X.; Wu, K. Effective feature fusion network in BIFPN for small object detection. In Proceedings of the 2021 IEEE International Conference on Image Processing (ICIP), Anchorage, AK, USA, 19–22 September 2021; pp. 699–703. [Google Scholar]
  55. Li, H.; Li, J.; Wei, H.; Liu, Z.; Zhan, Z.; Ren, Q. Slim-neck by GSConv: A lightweight-design for real-time detector architectures. J. Real-Time Image Process. 2024, 21, 62. [Google Scholar] [CrossRef]
  56. Wang, C.; He, W.; Nie, Y.; Guo, J.; Liu, C.; Wang, Y.; Han, K. Gold-YOLO: Efficient object detector via gather-and-distribute mechanism. Adv. Neural Inf. Process. Syst. 2023, 36, 51094–51112. [Google Scholar]
  57. Kang, M.; Ting, C.M.; Ting, F.F.; Phan, R.C.W. ASF-YOLO: A novel YOLO model with attentional scale sequence fusion for cell instance segmentation. Image Vis. Comput. 2024, 147, 105057. [Google Scholar] [CrossRef]
  58. Du, Z.; Hu, Z.; Zhao, G.; Jin, Y.; Ma, H. Cross-layer feature pyramid transformer for small object detection in aerial images. arXiv 2024, arXiv:2407.19696. [Google Scholar] [CrossRef]
  59. Kang, M.; Ting, C.M.; Ting, F.F.; Phan, R.C.W. RCS-YOLO: A fast and high-accuracy object detector for brain tumor detection. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Vancouver, Canada, 8–12 October 2023; Springer: Berlin/Heidelberg, Germany, 2023; pp. 600–610. [Google Scholar]
  60. Xu, X.; Jiang, Y.; Chen, W.; Huang, Y.; Zhang, Y.; Sun, X. Damo-yolo: A report on real-time object detection design. arXiv 2022, arXiv:2211.15444. [Google Scholar]
  61. Ye, R.; Shao, G.; He, Y.; Gao, Q.; Li, T. YOLOv8-RMDA: Lightweight YOLOv8 network for early detection of small target diseases in tea. Sensors 2024, 24, 2896. [Google Scholar] [CrossRef] [PubMed]
  62. Chen, Y.; Zhang, C.; Chen, B.; Huang, Y.; Sun, Y.; Wang, C.; Fu, X.; Dai, Y.; Qin, F.; Peng, Y.; et al. Accurate leukocyte detection based on deformable-DETR and multi-level feature fusion for aiding diagnosis of blood diseases. Comput. Biol. Med. 2024, 170, 107917. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Examples of dense imagery: (a) retail shelf display; (b) cigarette packages.
Figure 1. Examples of dense imagery: (a) retail shelf display; (b) cigarette packages.
Jimaging 11 00345 g001
Figure 2. The architecture of the DBA-YOLO method.
Figure 2. The architecture of the DBA-YOLO method.
Jimaging 11 00345 g002
Figure 3. C2f×PA module.
Figure 3. C2f×PA module.
Jimaging 11 00345 g003
Figure 4. BiSAF structure.
Figure 4. BiSAF structure.
Jimaging 11 00345 g004
Figure 5. BiAAF structure.
Figure 5. BiAAF structure.
Jimaging 11 00345 g005
Figure 6. DyDHead schematic. The asterisk (*) marks this face only.
Figure 6. DyDHead schematic. The asterisk (*) marks this face only.
Jimaging 11 00345 g006
Figure 7. DyDHead structure. (a) Detailed structure. (b) Scale and task attention modules.
Figure 7. DyDHead structure. (a) Detailed structure. (b) Scale and task attention modules.
Jimaging 11 00345 g007
Figure 8. Cigarette package images.
Figure 8. Cigarette package images.
Jimaging 11 00345 g008
Figure 9. Result curves on SKU-110K: (a) The curve on the SKU-110K with mAP@0.5. (b) The curve on the SKU-110K dataset with mAP@0.5:0.95.
Figure 9. Result curves on SKU-110K: (a) The curve on the SKU-110K with mAP@0.5. (b) The curve on the SKU-110K dataset with mAP@0.5:0.95.
Jimaging 11 00345 g009
Figure 10. SKU-110K results. (a,d) Full images; the yellow box marks the region of interest (ROI) used for zoomed comparison. (b,e) Zoomed ROI with YOLOv10 predictions. (c,f) Zoomed ROI with DBA-YOLO predictions.
Figure 10. SKU-110K results. (a,d) Full images; the yellow box marks the region of interest (ROI) used for zoomed comparison. (b,e) Zoomed ROI with YOLOv10 predictions. (c,f) Zoomed ROI with DBA-YOLO predictions.
Jimaging 11 00345 g010
Figure 11. Result curves on cigarette package images: (a) The curve on cigarette images with mAP@0.75. (b) The curve on cigarette images with mAP@0.5:0.95.
Figure 11. Result curves on cigarette package images: (a) The curve on cigarette images with mAP@0.75. (b) The curve on cigarette images with mAP@0.5:0.95.
Jimaging 11 00345 g011
Figure 12. Cigarette package images results. (a,d) Original images; (b,e) YOLOv10 predictions; (c,f) DBA-YOLO predictions. Yellow boxes mark the region of interest (ROI) used for side-by-side comparison; green boxes indicate ground-truth bounding boxes.
Figure 12. Cigarette package images results. (a,d) Original images; (b,e) YOLOv10 predictions; (c,f) DBA-YOLO predictions. Yellow boxes mark the region of interest (ROI) used for side-by-side comparison; green boxes indicate ground-truth bounding boxes.
Jimaging 11 00345 g012
Figure 13. Thermograms (a,d,g) are the original images, (b,e,h) are the thermograms of YOLOv10, and (c,f,i) are the thermograms of DBA-YOLO.
Figure 13. Thermograms (a,d,g) are the original images, (b,e,h) are the thermograms of YOLOv10, and (c,f,i) are the thermograms of DBA-YOLO.
Jimaging 11 00345 g013
Figure 14. Failure cases. Red dashed boxes indicate missed detections (false negatives). (a) Cigarette package example images; (b) SKU-110K example images.
Figure 14. Failure cases. Red dashed boxes indicate missed detections (false negatives). (a) Cigarette package example images; (b) SKU-110K example images.
Jimaging 11 00345 g014
Table 1. SKU-110K result.
Table 1. SKU-110K result.
DatasetModel mAP (%) AP 50 (%) AP 75 (%)Params (× 10 5 )GFLOPs
SKU-110Kyolov5n56.388.964.125.097.2
yolov5s58.190.266.891.2324.0
RTMDet-Tiny40.162.546.948.738.03
yolov6n56.489.064.042.3811.9
yolov8n57.089.365.130.118.2
yolov8s58.790.567.8113.6028.6
yolov9t57.089.165.220.067.8
yolov10n56.989.665.027.077.3
yolov10s58.790.667.780.6724.8
yolov11n56.488.664.425.906.4
yolov12n56.689.164.525.686.5
RT-DETR-R1858.289.666.9198.7356.9
DDQ R-CNN38.190.357.6632.8050.2
OURS58.890.667.626.167.9
Table 2. Cigarette packet result.
Table 2. Cigarette packet result.
DatasetModel mAP / % AP 50 / % AP 75 / % Params/ 10 5
Cigarette packetyolov5n79.998.392.825.09
yolov5s80.699.193.391.23
yolov6n80.198.392.642.38
yolov8n80.298.693.130.11
yolov8s81.099.293.5113.6
yolov9t79.798.292.820.06
yolov10n79.698.792.727.07
yolov10s80.799.193.780.67
yolov11n79.798.692.925.90
yolov12n79.698.892.725.68
RT-DETR-R1879.499.393.1198.73
Ours81.099.493.926.16
Table 3. Comparison of object detection performance on VisDrone-2019 validation and test datasets.
Table 3. Comparison of object detection performance on VisDrone-2019 validation and test datasets.
DatasetModelPRmAP/%mAP50/%mAP75/%Params/ 10 5
Visdrone-valyolov10n0.4580.3520.335.220.427.07
yolov11n0.4410.3419.533.719.325.9
yolov12n0.440.33519.333.119.125.68
OURS0.5040.36922.938.523.426.16
Visdrone-testyolov10n0.3860.30214.827.114.427.07
yolov11n0.3930.29615.127.115.125.9
yolov12n0.390.29215.22715.225.68
OURS0.4360.31117.630.717.826.16
Table 4. Model comparison on SKU-110 dataset.
Table 4. Model comparison on SKU-110 dataset.
HeadmAP (%)AP50 (%)AP75 (%)Params ( × 10 5 )GFLOPS
Ours57.890.066.427.87.7
SEAMHead [49]56.689.364.425.27.3
TADDH [50]56.589.764.819.98.4
MultiSEAM [51]56.689.364.567.39.3
LSCD [52]56.989.565.119.56.2
RSCD [53]55.287.562.820.56.5
Table 5. Model comparison on SKU-110 dataset.
Table 5. Model comparison on SKU-110 dataset.
Neck NetworksmAP (%)AP50 (%)AP75 (%)Params ( × 10 5 )GFLOPS
Ours57.489.765.818.96.3
bifpn [54]57.189.365.117.26.0
slimneck [55]56.689.364.323.95.9
goldyolo [56]55.988.664.853.98.9
ASF [57]57.189.665.123.06.9
CFPT [58]56.389.663.918.96.4
RCSOSA [59]57.490.065.641.115.3
GFPN [60]57.089.565.233.27.0
EfficientRepBiPAN [61]56.889.564.827.36.8
HSFPN [62]56.088.663.519.36.7
Table 6. Ablation experiments for each module in the DBA-YOLO model.
Table 6. Ablation experiments for each module in the DBA-YOLO model.
C2f PABIMAFPNDyDHeadmAP/% AP 75 /%Params ( × 10 5 )GFLOPs
56.965.027.07.3
57.766.227.78.4
57.465.818.96.3
57.866.427.87.7
57.966.424.07.4
58.266.828.88.7
58.066.725.47.3
58.867.626.17.9
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

He, Z.; Yang, J.; Ning, H.; Li, C.; Tang, Q. DBA-YOLO: A Dense Target Detection Model Based on Lightweight Neural Networks. J. Imaging 2025, 11, 345. https://doi.org/10.3390/jimaging11100345

AMA Style

He Z, Yang J, Ning H, Li C, Tang Q. DBA-YOLO: A Dense Target Detection Model Based on Lightweight Neural Networks. Journal of Imaging. 2025; 11(10):345. https://doi.org/10.3390/jimaging11100345

Chicago/Turabian Style

He, Zhiyong, Jiahong Yang, Hongtian Ning, Chengxuan Li, and Qiang Tang. 2025. "DBA-YOLO: A Dense Target Detection Model Based on Lightweight Neural Networks" Journal of Imaging 11, no. 10: 345. https://doi.org/10.3390/jimaging11100345

APA Style

He, Z., Yang, J., Ning, H., Li, C., & Tang, Q. (2025). DBA-YOLO: A Dense Target Detection Model Based on Lightweight Neural Networks. Journal of Imaging, 11(10), 345. https://doi.org/10.3390/jimaging11100345

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop