Next Article in Journal
Implementation, Validation and Clinical Testing of Oximetry Device for Microcirculation Assessment in Oral Tissue
Previous Article in Journal
Ship Motion Attitude Prediction Model Based on FMD-IBKA-BTGN
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

The Enhance-Fuse-Align Principle: A New Architectural Blueprint for Robust Object Detection, with Application to X-Ray Security

1
Guangdong Provincial Key Laboratory of Cyber-Physical System, School of Automation, Guangdong University of Technology, Guangzhou 510006, China
2
College of Engineering, Shantou University, Shantou 515063, China
3
School of Computer, Guangdong University of Technology, Guangzhou 510006, China
4
Department of Computer Science, KU Leuven, 3000 Leuven, Belgium
5
Department of Mechanical Engineering, KU Leuven, 3000 Leuven, Belgium
*
Authors to whom correspondence should be addressed.
Sensors 2025, 25(21), 6603; https://doi.org/10.3390/s25216603
Submission received: 12 September 2025 / Revised: 20 October 2025 / Accepted: 24 October 2025 / Published: 27 October 2025
(This article belongs to the Section Sensing and Imaging)

Abstract

Object detection in challenging imaging domains like security screening, medical analysis, and satellite imaging is often hindered by signal degradation (e.g., noise, blur) and spatial ambiguity (e.g., occlusion, extreme scale variation). We argue that many standard architectures fail by fusing multi-scale features prematurely, which amplifies noise. This paper introduces the Enhance-Fuse-Align (E-F-A) principle: a new architectural blueprint positing that robust feature enhancement and explicit spatial alignment are necessary preconditions for effective feature fusion. We implement this blueprint in a model named SecureDet, which instantiates each stage: (1) an RFCBAMConv module for feature Enhancement; (2) a BiFPN for weighted Fusion; (3) ECFA and ASFA modules for contextual and spatial Alignment. To validate the E-F-A blueprint, we apply SecureDet to the highly challenging task of X-ray contraband detection. Extensive experiments and ablation studies demonstrate that the mandated E-F-A sequence is critical to performance, significantly outperforming both the baseline and incomplete or improperly ordered architectures. In practice, enhancement is applied prior to fusion to attenuate noise and blur that would otherwise be amplified by cross-scale aggregation, and final alignment corrects mis-registrations to avoid sampling extraneous signals from occluding materials.

1. Introduction

Object detection in complex imaging environments represents one of the most persistent challenges in computer vision, where traditional approaches frequently fail due to inherent domain-specific limitations. Across diverse fields, from medical diagnosis to security screening and remote sensing, practitioners encounter fundamental obstacles including noise interference, multiscale object variations, occlusion handling, and feature degradation that significantly compromise detection accuracy and reliability. Conventional object detection frameworks, originally designed for natural image scenarios characterized by clean backgrounds, consistent lighting, and well-defined object boundaries, are often ill-equipped for real-world applications. These applications require robust performance in environments that can be demonstrably affected by severe noise, atmospheric interference, overlapping structures, and variable acquisition conditions. For example, in medical imaging, radiologists face the critical task of detecting subtle tumors within noisy CT scans, where inherent contrast variations and imaging artifacts can obscure vital diagnostic features [1]. Security personnel face the daunting task of identifying concealed contraband in heavily cluttered X-ray baggage imagery with extensive object occlusion [2]; see also advances in real-world X-ray benchmarks and detection models [3,4,5]. In addition, enhanced detonator detection in X-ray baggage via image manipulation and deep CNNs has been explored [6], and robust perception under adverse visibility has leveraged thermal imaging for human and vehicle detection [7,8]. Beyond security screening, real-time object detection for assistive navigation further exemplifies the need for efficient detectors [9]. Remote sensing analysts struggle with small object detection in vast-scale variations while contending with atmospheric effects and complex terrain heterogeneity [10,11]. While recent deep learning advancements have individually demonstrated considerable potential to address specific challenges for multi-domain object detection, a significant impediment to developing truly robust systems is the absence of a unified framework that systematically integrates feature enhancement [12], data fusion, and spatial alignment. Current methodologies are often siloed, typically excelling in one area at the expense of others. For example, some methods prioritize noise reduction without effectively leveraging multiscale features [13,14], while others focus on fusion while neglecting the critical issue of spatial misalignment inherent in oriented object detection [15].
This paper introduces the Enhance-Fuse-Align (E-F-A) principle, a comprehensive framework that systematically addresses the three fundamental pillars of robust object detection in challenging domains. The E-F-A principle encompasses (1) Enhancement of degraded features through adaptive attention mechanisms, (2) Fusion of multiscale contextual information using bidirectional feature pyramids, and (3) Alignment of spatial representations to handle orientation and scale variations. We demonstrate this principle through SecureDet, a novel architecture specifically designed for X-ray contraband detection that achieves state-of-the-art performance while maintaining computational efficiency suitable for real-time security screening applications.
Foundational detectors and architectural primitives have laid the groundwork for modern object detection: YOLOv3 popularized single-stage real-time detection with multi-scale predictions, Feature Pyramid Networks (FPN) introduced top-down multi-scale feature fusion, and deformable convolutions enabled adaptive spatial sampling for geometric variations, which our E–F–A design builds upon and extends in challenging X-ray settings [16,17,18]. For security X-ray imaging, large-scale benchmarks such as SIXray and the GDXray NDT dataset have been instrumental for evaluation and algorithm development [19,20].
Consequently, the central contribution of this work is not solely the introduction of the SecureDet detector, but rather the proposal and rigorous validation of a general architectural blueprint: the Enhance-Fuse-Align (E-F-A) principle. We contend that for domains like X-ray security screening, which are marked by severe signal degradation and spatial ambiguity, the specific ordered sequence of operations—first enhancing features, then performing spatial alignment, and finally fusing multi-scale representations—is of paramount importance. SecureDet serves as our instantiation designed to rigorously test this principle. To ensure a fair evaluation that isolates the benefits of E-F-A, we utilize YOLOv8 as the baseline backbone. Extensive experimentation, including critical ablation studies that disentangle module roles and their ordering, unequivocally demonstrates that this principled, disciplined design yields superior results compared to simply assembling advanced components, thereby clearly distinguishing our contribution from routine YOLO variants.

2. Related Work

2.1. Challenges in Medical Imaging Object Detection

Medical imaging presents unique challenges where noise, artifacts, and anatomical complexity significantly impact object detection performance. Tumor detection in noisy CT scans exemplifies these fundamental challenges, where low-dose imaging protocols introduce substantial noise while preserving diagnostic accuracy remains critical. Zhang et al. [21] demonstrated how annotation quality directly impacts segmentation performance in mandible segmentation of CT images, revealing that even small amounts of label noise can severely degrade the effectiveness of the deep learning model. Karimi et al. [22] provided a comprehensive analysis of noise-resistant training methods, identifying data scarcity and annotation quality as the main bottlenecks in medical image analysis. New methods are also being developed to recover high-quality images from degraded sources by using diffusion models [23] to guide the restoration process [24]. Recent advances in noise reduction for medical imaging have focused on preserving diagnostic features while minimizing artifacts. Mileto et al. [13] evaluated iterative reconstruction algorithms combined with CNN-based denoising, showing that while traditional iterative methods modify spatial resolution non-linearly, deep learning approaches can maintain diagnostic quality. Al-Antari et al. [25] introduced the Quadratic CNN (Q-CNN) architecture that enhances sensitivity to analyze noisy radiographs without requiring training on noisy images, demonstrating superior robustness compared to conventional approaches. These medical imaging challenges parallel those found in security screening, where object detection must operate reliably despite image degradation and complex backgrounds.

2.2. Remote Sensing and Satellite Imaging Challenges

Remote sensing represents another domain where fundamental imaging challenges mirror those in security applications. Yang et al. [10] identified critical issues in oriented object detection including feature misalignment, spatial misalignment, and periodicity of angle problems that cause training instability. The challenge of small object detection across vast scale variations is particularly acute in satellite imagery, where objects occupy few pixels while maintaining detection accuracy across diverse geographical contexts [26]. Wei et al. [27] addressed multi-scale attention for small object detection in remote sensing, highlighting insufficient utilization of small object information and weak robustness against complex backgrounds. These challenges directly parallel X-ray security screening scenarios where small contraband items must be detected within cluttered baggage environments. Liu et al. [28] provided comprehensive analysis of small object detection in aerial images, emphasizing inadequate positive samples and inaccurate localization—problems that extend across multiple challenging imaging domains. Dubovik et al. [11] outlined grand challenges in satellite remote sensing, identifying atmospheric effects and sensor limitations that create fundamental trade-offs between spatial coverage and resolution. These atmospheric and environmental degradation effects mirror the image quality issues encountered in medical CT imaging and X-ray security screening, establishing common technical foundations across diverse application domains.
At the architectural level, widely used primitives include channel–spatial attention modules (CBAM) and efficient channel attention (ECA) for feature enhancement, spatial pyramid pooling (SPP) for multi-scale context aggregation, and bidirectional feature pyramid fusion as instantiated in EfficientDet for cross-scale aggregation, which we build upon in our E–F–A design [29,30,31,32]. In addition, multi-resolution post-filtering with attention for image quality enhancement and cross-modal infrared–visible image fusion further motivate our enhancement and fusion strategies [33,34]. Within the architectural primitives, FPN provides a top-down pathway with lateral connections for multi-scale feature fusion, while BiFPN extends this idea with bidirectional cross-scale connections and learnable per-edge fusion weights for iterative refinement, which we adopt as the fusion operator in SecureDet [32]. For completeness, classic baselines such as Faster R-CNN and SSD established, respectively, the two-stage RPN-based pipeline and the single-stage dense default-box formulation that underlie many subsequent detectors [35,36].

3. Method

3.1. Overall Architecture and Design Philosophy

The architecture of SecureDet is predicated on the Enhance-Fuse-Align (E-F-A) principle. As illustrated in Figure 1, the workflow begins with a backbone network that extracts multi-scale features from a 3-channel (RGB) input image at a 640 × 640 resolution. By integrating RFCBAMConv and C2f+RFCBAMConv modules, the backbone yields four feature maps: P2 (160 × 160, 128 channels), P3 (80 × 80, 256 channels), P4 (40 × 40, 512 channels), and P5 (20 × 20, 1024 channels). Among these, the high-resolution P2 map is critical for capturing the fine-grained details required for small object detection.
The core of the architecture is the “Fuse” stage, orchestrated by a BiFPN. The efficacy of BiFPN stems from its use of learnable weights for feature fusion, which allows the network to prioritize more informative features across different scales. This is realized through a bidirectional data flow: a top-down pathway propagates high-level semantic context, while a bottom-up pathway conveys precise spatial localization information. Within this fusion framework, we introduce a key “Enhance” strategy focused on the P2 feature map. Specifically, the P2 map (160 × 160, 128 channels) first undergoes a convolutional operation to match the dimensions of P3 (80 × 80, 256 channels). This transformed P2 map is then integrated with both the original backbone P3 feature and the P3 feature from the top-down pathway. This tripartite fusion injects the detailed spatial information from P2 into the P3 level, substantially improving localization accuracy. The entire bidirectional fusion process is applied iteratively to ensure a comprehensive synthesis of semantic and spatial features, ultimately producing P3, P4, and P5 feature maps unified at a 256-channel dimension.
Following fusion, the architecture proceeds to the “Align” stage. The fused P3, P4, and P5 feature maps are further refined by the Enhanced Context Fusion Attention (ECFA) and Attentional Self-Feature Augmentation (ASFA) modules. Finally, these refined multi-scale feature maps are fed into the detection head to perform the final object detection task.

3.2. Enhance Module—RFCBAMConv

Design Rationale: While standard attention modules like CBAM are effective, their reliance on a fixed receptive field limits performance, particularly on X-ray images characterized by blurred boundaries and large-scale variations. To overcome this limitation, we propose the Receptive Field Context-aware CBAM (RFCBAMConv). Unlike conventional approaches, RFCBAMConv first establishes a rich, multi-scale spatial context before the channel and spatial attention mechanisms are applied. This preemptive enrichment of features enables the attention module to operate on inputs that already encode diverse spatial patterns. Consequently, this design enhances the model’s ability to focus on discriminative information under challenging conditions, while simultaneously attenuating noise and improving the signal-to-noise ratio of features for subsequent fusion.
The RFCBAMConv is designed to enhance feature representation through the integration of multi-scale spatial context with adaptive channel and spatial attention mechanisms. The processing pipeline is structured as follows:
Initially, a channel attention mechanism, adopting principles from the Squeeze-and-Excitation (SE) block, is employed to generate channel weights M ch R B × C × 1 × 1 . Input features X R B × C × H × W are first passed through global average pooling to derive per-channel statistics as shown in Equation (1):
z c = 1 H × W i = 1 H j = 1 W X c ( i , j )
Statistics are subsequently processed within a bottleneck structure that consists of two fully connected layers featuring a reduction ratio r = 16 and a sigmoid activation function.
Concurrently, multi-scale receptive field features are generated. A k × k depth-wise separable convolution is utilized to capture local spatial patterns, followed by a spatial unfolding operation that integrates diverse contextual information into an enriched feature map F unfold . Subsequently, a spatial attention mechanism computes saliency maps M sp over F unfold through the concatenation of max and average pooling statistics across channel dimensions, followed by the passage through a 3 × 3 convolutional layer with sigmoid activation, as described in Equation (2):
M sp = σ Conv 3 × 3 ( [ max c ( F unfold ) , avg c ( F unfold ) ] )
Finally, dual-attention modulation is applied. The unfolded features F unfold are sequentially weighted by M ch ( F ch = F unfold M ch ) and subsequently by M sp ( F att = F ch M sp ). These attention-refined features are fed into the output layer of the module: a k × k convolution with stride k for spatial downsampling and channel transformation, followed by Batch Normalization and ReLU activation, resulting in the final output Y as formulated in Equation (3):
Y = ReLU ( BN ( Conv k × k , s = k ( F unfold M ch M sp ) ) )
This integrated, hierarchical design allows RFCBAMConv to first identify globally important feature types (channels) and then dynamically pinpoint their precise spatial locations using learned local descriptors. By leveraging rich multi-scale context, the module effectively focuses on discriminative features, demonstrating superior capability in handling complex imagery, such as the overlapping and semi-transparent characteristics of X-ray images, as illustrated in Figure 2.
Implementation logic and parameter selection. RFCBAMConv first builds multi-scale spatial context via spatial unfolding on local descriptors (depth-wise separable k × k ), then applies dual attention sequentially—channel attention with a bottleneck (reduction r = 16 , sigmoid) followed by spatial attention obtained from concatenated max/avg channel statistics passed through a 3 × 3 convolution with sigmoid; finally, an output Conv k × k , s = k performs spatial downsampling and channel transformation, followed by Batch Normalization and ReLU. This “enhance-then-attend” ordering places attention on context-enriched features, improving robustness under noise and blur.
As shown in Table 1, the RFCBAMConv component hyperparameters provide detailed configuration settings for optimal performance.

3.3. Fuse Module—BiFPN

The BiFPN module processes the enhanced feature pyramid P 3 , P 4 , P 5 generated by the Enhance module. Its central innovation is the conceptualization of each bidirectional pathway (top-down and bottom-up) as a distinct feature network layer. These layers can be stacked iteratively to facilitate progressively higher-level feature fusion. The fusion operation itself employs a fast normalized weighted sum. Specifically, for a node combining multiple input features I i , the output feature O is computed as:
O = Conv i w i j w j + ϵ · I i
Here, w i denotes a learnable weight for each input feature I i . These weights are constrained to be non-negative via a post-weight ReLU activation, explicitly enabling the network to learn the relative importance of each input feature map. The constant ϵ (typically 10 4 ) is added to the denominator to guarantee numerical stability. Subsequently, the aggregated weighted sum is passed through a depthwise separable convolution, followed by Batch Normalization and activation functions, yielding the final fused feature map. Within the SecureDet architecture, BiFPN functions as the principal F-module, iteratively performing weighted fusion across multiple scales. Its repeated bidirectional cross-scale connections and weighted feature aggregation yield a feature pyramid rich in both high-level semantic context and fine-grained spatial details. This comprehensive representation forms a robust foundation for the subsequent alignment and detection stages.

3.4. Align Module

3.4.1. Enhanced Contextual Feature Alignment (ECFA)

Design Rationale: Effectively resolving severe occlusions requires that a model perceives both global scene context and fine-grained local details. Conventional feature pyramids, however, often sacrifice the latter for the former, creating a fundamental trade-off that compromises localization precision. To address this challenge, we designed the ECFA module. It synergistically combines a Pyramid Spatial Pooling (PSP) backbone to capture long-range context with a Tanh-gated local attention mechanism to preserve and refine spatial boundaries. This dual-path architecture ensures the integration of global context while maintaining the high-fidelity local information critical for distinguishing overlapping objects. Consequently, the resulting contextual alignment suppresses spurious responses induced by clutter and stabilizes features against degradation, thereby significantly enhancing localization robustness.
The enhancement of neural network feature representations is facilitated by the ECFA module through the adaptive integration of multi-scale contextual information, alongside the refinement of spatial characteristics by means of attention mechanisms, as illustrated in Figure 3. Both long-range contextual dependencies and fine-grained local spatial relationships are captured, thus improving feature discriminability for complex visual perception tasks.
The module begins with initial receptive field expansion and channel attenuation. Input features F R C × H × W undergo a 3 × 3 convolution to reduce channel dimension, yielding F = Conv 3 × 3 ( F ) R C / 2 × H × W . This stabilizes feature representations and prepares them for subsequent multi-scale processing. Pyramid Spatial Pooling (PSP) then aggregates context from diverse spatial scales by applying adaptive average pooling across predefined grid sizes (e.g., 6 × 6 , 3 × 3 , 2 × 2 , 1 × 1 ) to F , generating multi-scale feature maps for the cross-scale attention stage.
During the cross-scale attention stage, spatially preserved features F are transformed into Queries ( Q ), while multi-scale pooled features are mapped to Keys ( K ) and Values ( V ) via 1 × 1 convolutions. The interaction between Q and K is regarded as essential for the learning of relevant contextual dependencies. Attention weights A are computed using Equation (5):
A = Softmax ( Q · K ) R H W × S
These weights drive the weighted aggregation of Value vectors, effectively distilling multi-scale context into a refined feature map C . Here, S denotes the number of pooled scales used in PSP. A subsequent local refinement module further processes these features by generating a spatial attention mask M using a bottleneck convolutional pathway ( 1 × 1 conv → 3 × 3 conv) and a Tanh activation. Contextual features C are then element-wise modulated by M and fused with the original C via a residual connection, yielding an enhanced feature map C .
The ECFA module’s final output Y is produced via a direct residual connection, integrating refined contextual features C with the initial channel-reduced features F as shown in Equation (6):
Y = F + C R C / 2 × H × W
Original input information is preserved while dynamically aggregated and spatially calibrated context is leveraged, enhancing feature discriminability.

3.4.2. Adaptive Spatial Feature Alignment (ASFA)

The ASFA module is designed to enhance multi-scale feature fusion by enabling adaptive spatial alignment and calibration through learnable spatial transformations and attention mechanisms. As shown in Figure 4, the Adaptive Spatial Feature Alignment module architecture demonstrates the dual-stream processing approach for contextual and semantic features. This module employs a dual-stream architecture, processing contextual features ( X cp ) and semantic features ( X sp ) separately. Prior to fusion, low-resolution semantic features undergo a 3 × 3 convolutional transformation followed by bilinear upsampling to match the spatial resolution of contextual features, as formulated in Equation (7):
F sp = I Conv 3 × 3 32 ( X sp ) R N × C hid × H × W
High-resolution contextual features are extracted concurrently via an independent 3 × 3 convolutional layer. The deformable sampling mechanism represents the core innovation of ASFA. Preprocessed features are concatenated and input into an offset prediction network, which generates spatial offsets ( Δ l , Δ h ) and attention weights ( A ) . Subsequently, these learnable offsets are utilized to deform a standard sampling grid, with adaptive sampling locations computed as illustrated in Equation (8):
G l ( i , j ) = G base ( i , j ) + Δ l ( i , j ) [ W , H ] T
The adaptive spatial transformation is employed to achieve precise realignment of features, thereby enhancing robustness to geometric variations. It also reduces the sampling of noisy or occluding pixels by steering the deformable grid toward informative target structures, thereby decreasing false positives under degraded X-ray signals. The spatially calibrated features and contextual features are subsequently integrated through an attention-weighted mechanism, resulting in the production of the final output as articulated in Equation (9):
Y = W 1 F ˜ sp + W 2 F ˜ cp
where W 1 and W 2 are defined as learned attention weights. The ‘Groups’ strategy (G = 2) has been incorporated to balance representational capacity with computational efficiency, thereby rendering ASFA effective for tasks that necessitate fine-grained multi-scale feature integration.

4. Experiments and Results

4.1. Experimental Configuration and Protocol

Experiments were conducted on an NVIDIA RTX 4090 workstation using Python/PyTorch with Ultralytics YOLOv8. As shown in Figure 5, comprehensive dataset statistics demonstrate the diversity and complexity of the evaluation benchmarks. The training protocol for X-ray contraband detection uses a private dataset with strong augmentation and regularization to enhance robustness and generalization. Cross-domain evaluation on OPIXray and HiXray is included. Key settings: 640 × 640 input resolution, batch size of 16, and 300 epochs; losses combine BCE for classification/objectness with CIoU for bounding-box regression; optimization employs SGD (initial learning rate 0.01) with a cosine schedule and warmup, EMA, weight decay, and light label smoothing. Inference uses confidence and NMS IoU thresholds of 0.25 and 0.7, respectively. Performance is reported using Precision, Recall, mAP@0.5, and mAP@0.5:0.95, together with Params, GFLOPs, and FPS. Early stopping is based on validation mAP@0.5:0.95, and the cross-domain evaluation substantiates generalization. Architectural scaling parameters of SecureDet variants (n/s/m) are summarized in Table 2 for quick reference.
As shown in Table 3, the baseline training and inference configurations provide standardized parameters for fair comparison across all evaluated models.

4.2. Comprehensive Performance Analysis and Benchmarking

This work performed extensive benchmarking of SecureDet against 13 state-of-the-art detection models, spanning anchor-based (Fast R-CNN [37]), transformer-based (RT-DETR [38]), efficiency-optimized (EfficientDet), and recent YOLO variants (v8–v13 [39,40,41,42,43,44]). Table 4 summarizes comprehensive performance metrics: precision (P), recall (R), mAP@0.5, and mAP@0.5:0.95, alongside critical computational efficiency indicators such as parameters, GFLOPs, and FPS. SecureDet demonstrates consistent superiority across all model scales; its flagship SecureDet-m variant achieves exceptional performance (92.60% P, 74.61% R, 82.26% mAP@0.5, 72.31% mAP@0.5:0.95), substantially outperforming the YOLOv8s baseline (+2.97% mAP@0.5, +5.93% mAP@0.5:0.95) and even surpassing the stronger YOLOv11m by +1.58% recall and +0.62% mAP@0.5:0.95, while maintaining real-time inference. Architectural comparisons reveal SecureDet’s advantages: against RT-DETR-R50, it yields +7.91% higher mAP@0.5 and +10.88% higher mAP@0.5:0.95 with 15.8% fewer parameters (19.2 M vs. 22.8 M) and comparable inference speed (+0.9% FPS; 59.2 vs. 58.7); against EfficientDet-d2, it achieves improved accuracy at comparable computational cost. Furthermore, SecureDet consistently delivers favorable accuracy-efficiency trade-offs across its model scales; the lightweight SecureDet-n variant outperforms YOLOv8n with reduced parameter count, and SecureDet-m offers an optimal balance of high accuracy, manageable parameters, and efficient computation. These results are obtained under consistent evaluation configurations, demonstrating both individual module contributions and their combinatorial effects. Complementing its broad performance gains, SecureDet exhibits outstanding localization precision, as demonstrably proven by its robustness across varying Intersection over Union (IoU) thresholds. The substantial improvements in mAP@0.5:0.95 (+5.93% over YOLOv8s, +0.62% over YOLOv11m) directly signify enhanced spatial accuracy that persists even under stricter localization criteria. This inherent robustness confirms that SecureDet not only increases the detection rate for contraband items but also localizes them with superior spatial fidelity compared to existing methods. Such high localization precision is critically imperative for automated security screening systems, where accurate object boundary delineation is fundamental for effective threat assessment and minimizing false positives or negatives.

4.3. Ablation Study

To systematically evaluate the independent and synergistic influences of each proposed module, comprehensive ablation experiments were conducted on the private dataset, with detailed quantitative results presented in Table 5 across eight distinct configurations. The analysis commenced by establishing the YOLOv8s baseline performance at 79.29% mAP@0.5. Individual module evaluations revealed contrasting effects: the BiFPN module, when applied in isolation, caused a marked performance degradation to 77.93% mAP@0.5 (−1.36%), indicating susceptibility to noise amplification in the absence of robust feature processing. In contrast, the RFCBAMConv module alone yielded a positive impact, improving performance to 79.85% mAP@0.5 (+0.56%), thus underscoring the inherent value of enhanced feature extraction via adaptive receptive fields. Subsequent examinations of module combinations illuminated critical architectural dependencies and limitations. The ECFA+ASFA pairing synergistically boosted performance to 80.79% mAP@0.5 (+1.50%), confirming complementary roles. However, integrating BiFPN with this combination (ECFA+ASFA+BiFPN) reduced performance to 77.88% mAP@0.5, and the RFC+BiFPN configuration achieved only 78.95% mAP@0.5, both reinforcing the reliance of BiFPN on high-quality, contextually enriched input features. Furthermore, the RFC+ECFA+ASFA configuration, despite its increased parameter count, underperformed the simpler ECFA+ASFA pair (78.82% vs. 80.79% mAP@0.5), highlighting that advanced backbone features were suboptimally utilized without an effective fusion mechanism, thereby hindering information flow. Together, these results corroborate that performing enhancement before fusion mitigates degradations (noise/blur) and prevents their amplification across the pyramid, while the final alignment stage further limits extraneous sampling from occlusions. In striking contrast to partial configurations, the full SecureDet-s model, synergistically integrating all four modules, achieved peak performance metrics: 81.63% mAP@0.5 and 69.54% mAP@0.5:0.95. These figures represent substantial absolute improvements of +2.34% and +3.16%, respectively, over the baseline. Critically, this holistic architecture demonstrated remarkable parameter efficiency, employing merely 7.55M parameters—a 45% reduction compared to the more complex RFC+ECFA+ASFA configuration (13.22 M)—while delivering superior accuracy. These empirical findings yield three overarching insights: first, the order and integration strategy of modules are more pivotal than individual capabilities; BiFPN, for instance, transforms from detrimental to significantly beneficial contingent upon receiving properly processed input features. Second, a clear hierarchical dependency governs model efficacy: RFCBAMConv lays the foundation with robust feature extraction, ECFA+ASFA refine these features through contextual enhancement and spatial alignment, and BiFPN optimally fuses this high-quality, processed information. Third, the complete architecture’s superior accuracy and efficiency validate the holistic design approach, confirming its suitability and efficacy for advanced X-ray contraband detection.

4.4. Dataset Evaluation

To further validate the generalization capability of our approach, we evaluated SecureDet on two public X-ray security screening datasets: OPIXray and HiXray. The results, summarized in Table 6, demonstrate the superior performance of our method compared to several state-of-the-art YOLO models. Additionally, for our private X-ray dataset, we employed a 70/20/10 split for training/validation/testing, strictly preventing sample leakage between partitions; for the public datasets (OPIXray, HiXray), we followed the official splits provided by the dataset maintainers.
On the OPIXray dataset, SecureDet achieved a mAP@0.5 of 91.16%, outperforming the next best method, YOLOv9s, which scored 90.49%. More importantly, our model achieved a recall of 87.11%, a significant improvement over other models. This high recall is critical in security applications, as it minimizes the risk of missing dangerous items.
On the HiXray dataset, SecureDet also demonstrated excellent performance. It achieved a recall of 81.11% and a mAP@0.5 of 82.59%, surpassing other YOLO variants. These results confirm the effectiveness and generalization capability of our proposed method across different datasets and contraband categories. The consistent improvements in both precision and recall validate that SecureDet provides a more reliable solution for real-world security inspection systems.
Our findings demonstrate the effectiveness and generalization of the proposed approach across diverse datasets and contraband categories. Notably, our model exhibits performance variations between the OPIXray (mAP@0.5: 91.16%) and HiXray (mAP@0.5: 82.59%) datasets. This discrepancy primarily arises from the distinct foundational challenges inherent to each dataset. OPIXray, specifically constructed to address contraband detection under severe occlusion (e.g., knives), centers on mitigating spatial ambiguity. Our SecureDet architecture, particularly its alignment modules (ECFA and ASFA), is engineered to tackle precisely these spatial misalignments and contextual confusions, leading to its superior performance on this benchmark. Conversely, HiXray, derived from authentic airport security scans, incorporates a wider array of items (e.g., power banks, liquids, cosmetics). This diversity introduces pronounced material ambiguity (where different organic substances can exhibit similar X-ray signatures) and significant intra-class morphological variations. While HiXray also contains occlusions, its primary challenges are more pronounced in discerning material properties and morphological characteristics than solely spatial ambiguities. Consequently, even though our model maintains strong performance exceeding baselines on HiXray, a degree of performance variation is understandable and expected, given its architectural emphasis on spatial occlusion resolution when confronted with these more complex, multifaceted challenges.
The experimental results demonstrated the effectiveness of the method proposed in this paper for security item detection tasks. An enhancement in the model’s capability for detecting safety threat items was observed, accompanied by a high recognition rate. This advancement provides a new technical path for the improvement of the reliability of security inspection systems.

4.5. Visual Performance Analysis

Comprehensive insights into SecureDet’s performance across diverse challenging scenarios are presented through a detailed visual analysis of detection results on representative X-ray security screening cases. This qualitative evaluation is complemented by quantitative metrics, illustrating the practical effectiveness of architectural innovations in real-world conditions.

4.5.1. Ablation Study Visualization Analysis

Figure 6 illustrates the detection performance of eight module configurations on a challenging X-ray luggage image, characterized by dense, overlapping, and varied-density electronics, which is representative of scenarios encountered in X-ray security screening. Moderate detection performance was exhibited by the baseline YOLOv8s, while the detection of small and occluded objects was often missed. The performance of BiFPN in isolation was noted to be inadequate, resulting in fewer detections and imprecise localization, thereby confirming its negative impact. Enhancements in feature discrimination and object boundary definition were attributed to RFCBAMConv, while improvements in spatial precision for small and boundary items were facilitated by ECFA + ASFA.
Combinations were characterized by mixed results. RFC+BiFPN yielded limited multi-scale gains; however, false positives and imprecision were introduced. ECFA+ASFA+BiFPN regressed, with confusion observed in overlapping regions, thereby highlighting the input dependence of BiFPN. RFCBAM+ECFA+ASFA exhibited inconsistencies in the presence of clutter, with a lack of effective fusion noted. The complete SecureDet-s model achieved comprehensive detection capabilities, accurate localization of objects across varying scales, occlusions, and congestion. This observation confirms the synergistic integration of all four modules, which contributed to substantially enhanced visual detection.

4.5.2. Severe Occlusion Scenario Analysis

Figure 7 presents a visual evaluation of the detection performance of eight models across three X-ray baggage scanning scenarios, progressively designed to assess robustness under increasing object complexity and occlusion levels. The first scenario, characterized by moderate object overlap and semi-transparency typical of X-ray imaging, involved the identification of a single primary contraband item. Under conditions of partial occlusion, Fast R-CNN, YOLOv8s, YOLOv10s, and YOLOv13s failed entirely to detect the target, while EfficientDet and RT-DETR exhibited minor localization deviations, which were directly attributable to the occluded nature of the object. In sharp contrast, SecureDet-s achieved precise target localization with bounding boxes highly consistent with ground truth, demonstrating the capability to effectively address this level of occlusion. As the scenarios progressed to the second stage, featuring denser object stacking and greater occlusion, detection difficulty significantly amplified. Here, SecureDet-s emerged as the sole model capable of accurately detecting the contraband, maintaining high fidelity with ground truth bounding boxes, thereby confirming superior performance under more severe occlusive challenges. The third and most complex scenario involved numerous objects and extreme occlusion, which severely degraded the localization accuracy of most models—Fast R-CNN, EfficientDet, and YOLOv10s produced false detections, and YOLOv13s missed two items. Despite these extreme conditions, SecureDet-s displayed remarkable robustness, with predicted bounding boxes remaining highly proximate to ground truth, notwithstanding one missed detection and minor localization differences. This resilience is crucial, as conventional methods often struggle to identify local outliers within clusters of varying densities, a problem analogous to detecting a small, occluded item in a cluttered bag [45]. Collectively, these results highlight the exceptional capability of SecureDet-s in occlusion analysis, demonstrating enhanced robustness for complex X-ray baggage scanning tasks where overcoming occlusion is paramount.

4.5.3. Small Detection Performance

SecureDet-s demonstrates excellent, consistent, and robust performance in small object detection within challenging X-ray security imaging environments, characterized by complex overlapping structures, low-contrast features, and sparse small objects. As illustrated in Figure 8 (middle row), the unique detection and precise localization of a minute object with subtle features in a demanding test case was achieved by SecureDet-s. This capability starkly contrasts with leading state-of-the-art detectors—including Fast R-CNN, EfficientDet, RT-DETR, YOLOv8s, YOLOv10s, and YOLOv13s—all of which failed to identify this critical target. The advanced ability of SecureDet-s to perceive and accurately locate tiny, indistinct objects is highlighted by this outcome. Across the entire test set, remarkable detection consistency is maintained by SecureDet-s, with precise detections closely aligning with Ground Truth (GT) annotations (as exemplified in Figure 7, first row) and reliable performance on difficult samples where frequent missed detections are exhibited by other models. When compared to other advanced methods, distinct advantages are offered by SecureDet-s: superior stability and higher recall rates are achieved over the YOLO series (v8s/v10s/v13s) for challenging small object detection tasks. In the context of X-ray imagery’s characteristic low-contrast and complex backgrounds, enhanced feature extraction and scene adaptability is exhibited by SecureDet-s compared to EfficientDet and RT-DETR. Robust detection consistency is provided by SecureDet-s, surpassing that of Fast R-CNN. The proven ability of SecureDet-s to deliver stable and reliable detection in these demanding environments underscores its significant potential for applications requiring exceptionally high precision in small object detection, such as security screening, threat detection, and medical imaging diagnostics.

4.5.4. Multi-Scale Detection Performance

The inherent density and overlap in X-ray security imagery, exemplified by objects such as coiled cables and stacked batteries, are associated with formidable challenges for target detection. As shown in Figure 9, multi-scale detection performance analysis demonstrates SecureDet’s superior capability in handling extreme size variations. Ground Truth (GT) accurately annotate all critical targets; however, considerable deficiencies in existing methods have been observed in these demanding scenarios. Fast R-CNN is capable of identifying principal cable bundles but frequently fails to capture fine-grained details. EfficientDet is characterized by broad detection coverage but suffers from imprecise bounding box localization. Limited efficacy has been exhibited by RT-DETR in dense overlap regions, which impedes its capability to resolve complex, interwoven structures. Baseline models, including YOLOv8s, YOLOv10s, and YOLOv13s, have demonstrated compromised accuracy. In stark contrast, SecureDet-s achieves superior performance, with results being highly consistent with GT and the successful identification of all major contraband items. This advancement has been highlighted as a potential enhancement for security screening.

4.5.5. Attention Visualization of Feature Learning and Feature Impact

To elucidate differences in feature learning across various architectures and their impact on detection performance, we employed Grad-CAM analysis. As shown in Figure 10, Grad-CAM-based attention visualization reveals the salient features driving detection decisions across different architectures. This technique generated attention maps focusing on key layers within each detector. Standardized procedures and normalized attention intensities facilitated consistent and equitable cross-model comparisons. Under conditions of severe occlusion, the baseline architecture (YOLOv8ss) demonstrated diffused, background-biased attention. Conversely, SecureDet-s directed its focus to intrinsic structural features of the target, such as edges, material boundaries, and connection points. It also effectively leveraged contextual information to pinpoint the unobscured portions of contraband.

5. Discussion

This study introduced SecureDet, a specialized YOLOv8-based architecture for X-ray contraband detection. Marked improvements in detection accuracy were demonstrated through our experimental results, particularly for occluded, small, and multi-scale objects, while real-time inference speeds were maintained. The findings are contextualized within the broader field of data-driven modeling, and the core scientific contributions of this work are discussed.

5.1. Architectural Synergy as the Core Contribution

A primary critique of application-specific deep learning models is that they often appear to be mere assemblages of pre-existing components. While the foundational concepts employed in SecureDet—such as channel-spatial attention, bidirectional feature fusion, and deformable alignment—are indeed established in the computer vision literature, our primary contribution is not the invention of these primitives. Rather, the novelty of this work lies in the discovery and validation of their non-trivial, synergistic integration tailored specifically for the challenges of X-ray contraband detection. This approach aligns with a growing trend in specialized manufacturing and inspection domains, where the intelligent combination and interpretation of machine learning models are paramount [46,47].
The most compelling evidence for this assertion is presented in the ablation study (Table 5). A critical architectural dependency was demonstrated: the isolated application of a weighted fusion neck (BiFPN) on the baseline YOLOv8 backbone was proven to be detrimental, resulting in a degradation of performance by 1.36% mAP@0.5. This finding is significant and non-obvious; it is shown that a theoretically powerful module may become counterproductive if it operates on features that are not adequately conditioned. This phenomenon, wherein advanced models fail in the absence of proper data or feature pre-processing, constitutes a recurring challenge in machine learning applications, ranging from medical imaging to industrial process control. A clear, domain-specific example of this principle has been provided by this work. The success of SecureDet is derived from a clear design principle validated by the results obtained.
  • Enhance First: The RFCBAMConv module first enriches the backbone features, making them more discriminative and robust to noise and blur. This is a crucial step, as standard feature detectors often show degraded performance when images are blurred, which is a common issue in low-lighting or rapid-scan conditions [48].
  • Align Second: The ECFA and ASFA modules then explicitly model contextual relationships and correct for spatial misalignments across feature hierarchies.
  • Fuse Last: Only after features are enhanced and aligned does the BiFPN module effectively fuse multi-scale information, leading to the peak performance of the full SecureDet model.
The contribution of SecureDet is characterized as an architectural blueprint, which establishes a necessary sequence of operations for this problem domain. It has been demonstrated that, in the context of X-ray imagery, the order and interdependence of advanced components are more critical than their individual capabilities. This principle may inform the design of future detectors.

5.2. Mapping Architectural Innovations to X-Ray Image Characteristics and Remaining Challenges

A strong correspondence between specific architectural choices and the underlying physical characteristics of X-ray imaging has been revealed through a deeper analysis of the results. The success of SecureDet is attributed to a targeted approach designed to address the inherent challenges of the domain, which further assists in illuminating the model’s remaining limitations.
The ablation study presented in Table 5 demonstrates that the combination of ECFA and ASFA yielded the most substantial performance improvement, quantified as +1.50% mAP@0.5. As shown in Table A1, comprehensive statistical analysis provides detailed quantitative metrics across all evaluation scenarios, including precision, recall, and F1 scores for different detection challenges. This finding suggests that resolving contextual ambiguity and spatial misalignment constitutes a significant challenge in X-ray contraband detection. In X-ray images, object overlap is characterized not merely as occlusion but as a semi-transparent superposition, where pixel intensities are determined by the material densities and thicknesses of multiple objects along the beam’s path.
The Enhanced Contextual Feature Alignment (ECFA) module is designed to address the issue through the implementation of self-attention for the modeling of long-range dependencies. Through this methodology, inferences regarding the presence of an object, such as a firearm, can be made despite local features being compromised by overlapping items, as the visible components are related to a comprehensive understanding of the baggage content [49].
The importance of the Adaptive Spatial Feature Alignment (ASFA) module employing deformable convolution has been established. A capability to adjust the sampling grid for the purpose of concentrating on the feature points of the target object while disregarding the extraneous signals from occluding foreground or background items is provided. This methodology is classified as a data-driven approach to achieve object separation within a superimposed signal.
The challenge of separating superimposed objects from a 2D projection shares conceptual similarities with optimal transport problems, where one seeks to find an optimal coupling between different probability distributions [50]. RFCBAMConv alone yields a modest gain of +0.56% mAP@0.5, indicating limited standalone impact. The issue of extreme scale variation and blurred boundaries is addressed through the creation of a multi-scale feature representation prior to the application of attention [51]. This approach guarantees that the subsequent alignment and fusion modules are supplied with features that are inherently robust to the low-resolution and scattering effects typically encountered in security scanners.
However, the limitations of the model are regarded as informative as its successes. In the most challenging occlusion scenario (Figure 7, third row), SecureDet-s, despite outperforming all other models, registered one missed detection. This observation indicates a fundamental boundary of what purely data-driven, 2D models can achieve. When an object is heavily occluded to the extent that its distinctive features are almost entirely obscured by those of another object, a lack of sufficient information occurs, preventing confident detection. It is suggested that a powerful, implicit understanding of X-ray physics has been acquired by the model; however, failures emerge when the 2D projection becomes mathematically non-unique or ambiguous. This limitation serves to underscore the potential value of physics-based models [52], which, while computationally intensive [53], could resolve such ambiguities effectively [54]. Future inquiries may investigate a hybrid approach that employs a data-driven model for real-time prediction, while high-uncertainty cases could be flagged for analysis by a more slowly operating, physics-informed algorithm [55], thereby integrating the strengths of both paradigms.

5.3. Overcoming Data Scarcity and the “Black Box” Problem

A significant hurdle in the development of robust deep learning models for security and manufacturing applications arises from the scarcity of large, comprehensively annotated datasets. This challenge has been extensively documented in fields including medical imaging [56] and CFRP drilling analysis [57]. The initial training phase employed a private dataset; however, strong performance on public benchmarks, such as OPIXray and HiXray (Table 6), indicates the generalization capability of the model. Such outcomes suggest that architectural enhancements confer a degree of robustness, aiding in the mitigation of dependency on extensive training sets. This finding has been corroborated in studies employing techniques such as virtual sample generation to address data scarcity.
The “black box” nature of deep learning models is implicitly addressed through the work presented herein. By systematically deconstructing the architecture and evaluating each component’s contribution via ablation studies, a degree of interpretability is provided. The success of the ECFA and ASFA modules is linked to the physical challenges associated with object occlusion and spatial distortion in X-ray imaging. This validation of components serves as a practical, qualitative form of model inspection. Integration of more formal eXplainable AI (XAI) methods, such as Gradient-weighted Class Activation Mapping (Grad-CAM), is suggested for future work to visualize the model’s focus and further diagnose its decision-making process, akin to methodologies applied in electrochemical machining and acoustic monitoring [58,59]. Graph-based machine learning models [60] can also be used to facilitate the identification of relationships between detected items, with unusual co-occurrences potentially flagged as candidates for further evaluation. Such methods would enhance trust in the system and offer insights into potential failure modes, thereby improving system reliability for critical security deployment [61].

5.4. Implications and Future Directions

The performance of SecureDet is characterized by a robust and efficient framework for the enhancement of contraband detection in practical security scenarios. The accurate localization of heavily occluded and small items is recognized as a direct contribution to public safety, possibly resulting in a reduction of cognitive load on human operators and a minimization of false alarm rates. Understanding user willingness to adopt new security technologies is key, similar to how researchers study consumer preferences and barriers to adopting new mobility services [62]. Several future research avenues are identified based on this work:
  • XAI Integration: Applying methods like SHAP and Grad-CAM to understand which input features (e.g., textures, material densities) the model prioritizes for different contraband types.
  • Semi-Supervised Learning: Exploring non-fully supervised learning paradigms to leverage the vast amounts of unlabeled X-ray data available, thereby reducing the reliance on costly manual annotation and addressing the “label scarcity” challenge head-on.
  • Multi-Modal Fusion: Incorporating data from other sensor modalities, such as dual-energy X-ray or 3D computed tomography (CT) scans, to provide richer information for detection.
  • Deployment and Optimization: Further optimizing the model for deployment on edge devices with limited computational resources, ensuring its practical applicability in a wider range of security checkpoints.

6. Conclusions

This paper presents SecureDet, a novel YOLO-based framework that establishes new performance benchmarks for X-ray contraband detection by addressing fundamental limitations through principled architectural innovations. Synergistic components are integrated in the approach, including adaptive multi-scale feature extraction, uncertainty-aware fusion, cross-scale attention, and deformable convolution-based aggregation, which contribute to substantial performance gains (e.g., +2.97% mAP over YOLOv8s baseline) while real-time inference is maintained. Methodologically, the theoretical understanding of feature representation for challenging imaging conditions is advanced by SecureDet. Practically, enhancements are observed in security screening effectiveness and passenger experience, with broad potential for wider applications in fields such as medical imaging. A robust foundation for next-generation automated detection systems is provided by this systematic framework, along with clear future research directions.

Limitations Under Extreme Occlusion and Future Directions

While SecureDet demonstrates robust performance under noise and clutter, extreme occlusion fundamentally limits its effectiveness. This limitation arises because severe occlusion removes critical discriminative evidence. Consequently, cross-scale fusion mechanisms may incorrectly weigh occluder textures, and spatial alignment strategies can lead to mis-sampling of occluded regions. Further compounding these issues, Non-Maximum Suppression (NMS) can suppress partially visible true positives when overshadowed by high-scoring false positives from clutter. Additionally, our training dataset contains insufficient examples with very high occlusion ratios.
To address these multifaceted challenges, our future research will focus on several key areas. We will implement advanced occlusion-aware data augmentation strategies, including techniques like Cutout, Hide-and-Seek, Random Erasing, and instance-level copy-pasting with realistic occluders. Architecturally, we aim to introduce a visibility-aware auxiliary head designed to judiciously gate feature fusion and re-calibrate detection confidences. Enhancements to deformable or iterative alignment mechanisms within ECFA/ASFA are also planned. Furthermore, we will explore adaptive post-processing techniques, such as Soft-NMS or class-aware NMS, to better handle competing detections. Concurrently, we intend to bolster context aggregation capabilities by leveraging larger receptive fields or more efficient lightweight attention mechanisms. To rigorously evaluate our progress in this challenging domain, we commit to reporting stratified recall and detailed error analyses categorized by occlusion level, complemented by targeted failure-case visualizations and ablation studies.

Author Contributions

Conceptualization, Y.L. (Yuduo Lin) and H.W.; methodology, Y.L. (Yuduo Lin); software, Y.L. (Yuduo Lin); validation, Y.L. (Yuduo Lin) and Y.L. (Yanfeng Lin); formal analysis, Y.L. (Yuduo Lin) and Y.L. (Yanfeng Lin); investigation, Y.L. (Yuduo Lin); resources, H.W. and M.W.; data curation, Y.L. (Yuduo Lin) and Y.L. (Yanfeng Lin); writing—original draft preparation, Y.L. (Yuduo Lin); writing—review and editing, H.W., Y.L. (Yanfeng Lin) and M.W.; visualization, Y.L. (Yuduo Lin) and Y.L. (Yanfeng Lin); supervision, H.W. and M.W.; project administration, H.W. and M.W.; funding acquisition, H.W. and M.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China grant number 62176056 and the Guangdong Provincial Key Laboratory of Cyber-Physical System grant number 2020B1212060069. The APC was funded by Guangdong University of Technology.

Institutional Review Board Statement

Not applicable for studies not involving humans or animals.

Informed Consent Statement

Not applicable for studies not involving humans.

Data Availability Statement

The datasets used in this study are publicly available: OPIXray dataset can be accessed at https://github.com/OPIXray-author/OPIXray, accessed on 15 December 2024 and HiXray dataset can be accessed at https://github.com/HiXray-author/HiXray, accessed on 15 December 2024. The source code and trained models will be made available upon reasonable request to the corresponding author.

Acknowledgments

The authors would like to thank the Guangdong Provincial Key Laboratory of Cyber-Physical System for providing computational resources and technical support. We also acknowledge the open-source datasets OPIXray and HiXray for enabling this research.

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

Abbreviations

The following abbreviations are used in this manuscript:
YOLOYou Only Look Once
mAPmean Average Precision
RFCBAMConvReceptive Field CBAM Convolution
BiFPNBidirectional Feature Pyramid Network
ECFAEnhanced Contextual Feature Alignment
ASFAAdaptive Spatial Feature Alignment
CBAMConvolutional Block Attention Module
FPNFeature Pyramid Network
SGDStochastic Gradient Descent
GPUGraphics Processing Unit

Appendix A. Supplementary Statistical Analysis

Table A1. Comprehensive statistical analysis of detection performance for Figure 5, Figure 6, Figure 7 and Figure 8 across different methods.
Table A1. Comprehensive statistical analysis of detection performance for Figure 5, Figure 6, Figure 7 and Figure 8 across different methods.
FigureMethodGTTPFNFPPrecisionRecallF1
Figure 6only-BiFPN31201.0000.3330.500
yolov8s31201.0000.3330.500
RFCBAM32110.6670.6670.667
RFC + BiFPN30320.0000.0000.000
ECFA + ASFA32110.6670.6670.667
ECFA + ASFA + BiFPN32110.6670.6670.667
RFC + ECFA + ASFA32101.0000.6670.800
SecureDet-s33001.0001.0001.000
Figure 7Fast R-CNN52410.6670.3330.444
EfficientDet53401.0000.4290.600
RT-DETR52301.0000.4000.571
YOLOv8s52301.0000.4000.571
YOLOv10s52310.6670.4000.500
YOLOv13s52301.0000.4000.571
SecureDet-s55010.8331.0000.909
Figure 8Fast R-CNN52310.6670.4000.500
EfficientDet52310.6670.4000.500
RT-DETR52301.0000.4000.571
YOLOv8s52301.0000.4000.571
YOLOv10s52301.0000.4000.571
YOLOv13s51401.0000.2000.333
SecureDet-s55001.0001.0001.000
Figure 9Fast R-CNN31201.0000.3330.500
EfficientDet32101.0000.6670.800
RT-DETR30300.0000.0000.000
YOLOv8s32101.0000.6670.800
YOLOv10s32101.0000.6670.800
YOLOv13s32101.0000.6670.800
SecureDet-s33001.0001.0001.000
Note: GT = Ground Truth objects; TP = True Positives; FN = False Negatives; FP = False Positives; Precision = TP/(TP + FP); Recall = TP/(TP+FN); F1 = 2 × (Precision × Recall)/(Precision + Recall).

Appendix B. Supplementary Retrieved Images for Zoom-In Comparison

To facilitate better visual comparison as requested for Figure 5, Figure 6, Figure 7 and Figure 8, we provide the original retrieved images in high resolution below; readers can directly zoom into the regions of interest to inspect fine details and occlusion boundaries.
Figure A1. Original retrieved image corresponding to Figure 6 and Figure 9.
Figure A1. Original retrieved image corresponding to Figure 6 and Figure 9.
Sensors 25 06603 g0a1
Figure A2. Original retrieved image corresponding to Figure 7 (severe occlusion scenario).
Figure A2. Original retrieved image corresponding to Figure 7 (severe occlusion scenario).
Sensors 25 06603 g0a2
Figure A3. Original retrieved image corresponding to Figure 8 (small-object detection).
Figure A3. Original retrieved image corresponding to Figure 8 (small-object detection).
Sensors 25 06603 g0a3
These originals are provided to enable precise zoom-in comparisons of the cropped areas shown in the main figures.

References

  1. Moodley, T.; Crush, L.; Brits, M. Lung tumor segmentation: A review of the state of the art. Front. Comput. Sci. 2024, 6, 1423693. [Google Scholar] [CrossRef]
  2. Wei, Y.; Tao, R.; Wu, Z.; Ma, Y.; Zhang, L.; Liu, X. Occluded Prohibited Items Detection: An X-ray Security Inspection Benchmark and De-occlusion Attention Module. In Proceedings of the 28th ACM International Conference on Multimedia (MM), Seattle, WA, USA, 12–16 October 2020; pp. 138–146. [Google Scholar]
  3. Tao, R.; Wei, Y.; Jiang, B.; Li, B.; You, Y.; Liu, X. Towards real-world X-ray security inspection: A high-quality benchmark and lateral inhibition module. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 10203–10212. [Google Scholar]
  4. Chen, L.; Yang, F.; Wang, H.; Zhang, Y.; Liu, X. Atomic number prior guided network for prohibited items detection from heavily cluttered X-ray imagery. Front. Phys. 2022, 10, 1026209. [Google Scholar] [CrossRef]
  5. Viriyasaranon, T.; Jung, S.W.; Hwang, S.J. MFA-net: Object detection for complex X-ray cargo and baggage security imagery. PLoS ONE 2022, 17, e0272961. [Google Scholar] [CrossRef] [PubMed]
  6. Oulhissane, H.A.; Bouchentouf, T.; Oulad-Abbou, D.; Yahyaouy, A.; Sabri, A. Enhanced detonators detection in X-ray baggage inspection by image manipulation and deep convolutional neural networks. Sci. Rep. 2023, 13, 4185. [Google Scholar] [CrossRef] [PubMed]
  7. Tsai, P.F.; Liao, C.H.; Yuan, S.M. Using Deep Learning with Thermal Imaging for Human Detection in Heavy Smoke Scenarios. Sensors 2022, 22, 5351. [Google Scholar] [CrossRef]
  8. Bhadoriya, A.S.; Vegamoor, V.; Rathinam, S. Vehicle Detection and Tracking Using Thermal Cameras in Adverse Visibility Conditions. Sensors 2022, 22, 4567. [Google Scholar] [CrossRef]
  9. Wang, W.; Jing, B.; Yu, X.; Sun, Y.; Yang, L.; Wang, C. YOLO-OD: Obstacle Detection for Visually Impaired Navigation Assistance. Sensors 2024, 24, 7621. [Google Scholar] [CrossRef]
  10. Yang, X.; Yang, J.; Yan, J.; Zhang, Y.; Zhang, T.; Guo, Z.; Sun, X.; Fu, K. Oriented Object Detection in Optical Remote Sensing Images using Deep Learning: A Survey. arXiv 2023, arXiv:2302.10473v4. [Google Scholar]
  11. Dubovik, O.; Schuster, G.L.; Xu, F.; Hu, Y.; B"osch, H.; Landgraf, J.; Li, Z. Grand challenges in satellite remote sensing. Front. Remote Sens. 2021, 2, 619818. [Google Scholar] [CrossRef]
  12. Wang, Y.; Zhang, F.L.; Dodgson, N.A. Target Scanpath-Guided 360-Degree Image Enhancement. In Proceedings of the AAAI Conference on Artificial Intelligence, Philadelphia, PA, USA, 25 February–4 March 2025; Volume 39, pp. 8169–8177. [Google Scholar]
  13. Mileto, A.; Guimaraes, L.S.; McCollough, C.H.; Fletcher, J.G.; Yu, L. CT noise-reduction methods for lower-dose scanning: Strengths and weaknesses of iterative reconstruction algorithms and new techniques. RadioGraphics 2021, 41, 1493–1508. [Google Scholar]
  14. Yao, Z.; Ge, J.; Cao, P.; Wu, M.; Qian, J.; Li, Y.; Reynaerts, D. Advancements in process monitoring and quality control for electrical discharge machining: A comprehensive review. J. Mater. Process. Technol. 2025, 345, 119081. [Google Scholar] [CrossRef]
  15. Guo, C.; Fan, B.; Zhang, Q.; Xiang, S.; Pan, C. AugFPN: Improving multi-scale feature learning for object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 12595–12604. [Google Scholar]
  16. Redmon, J.; Farhadi, A. YOLOv3: An Incremental Improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar] [CrossRef]
  17. Lin, T.Y.; Dollár, P.; Girshick, R.B.; He, K.; Hariharan, B.; Belongie, S.J. Feature Pyramid Networks for Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–16 July 2017. [Google Scholar]
  18. Dai, J.; Qi, H.; Xiong, Y.; Li, Y.; Zhang, G.; Hu, H.; Wei, Y. Deformable Convolutional Networks. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–23 October 2017. [Google Scholar]
  19. Miao, C.; Xie, L.; Wan, F.; Su, C.; Liu, H.; Jiao, J.; Ye, Q. SIXray: A Large-Scale Security Inspection X-Ray Benchmark for Prohibited Item Discovery in Overlapping Images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  20. Mery, D.; Riffo, V.; Zscherpel, U.; Mondragón, G.; Lillo, I.; Zuccar, I.; Lobel, H.; Carrasco, M. GDXray: The Database of X-ray Images for Nondestructive Testing. J. Nondestruct. Eval. 2015, 34, 42. [Google Scholar] [CrossRef]
  21. Zhang, Y.; Yang, L.; Chen, J.; Fredericksen, M.; Hughes, D.P.; Chen, D.Z. Deep adversarial networks for biomedical image segmentation utilizing unannotated images. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Lima, Peru, 4–8 October 2020; pp. 1–9. [Google Scholar]
  22. Karimi, D.; Dou, H.; Warfield, S.K.; Gholipour, A. Deep learning with noisy labels: Exploring techniques and remedies in medical image analysis. Med. Image Anal. 2020, 65, 101759. [Google Scholar] [CrossRef]
  23. Wang, Y.; Zhang, F.L.; Dodgson, N.A. Scantd: 360° scanpath prediction based on time-series diffusion. In Proceedings of the 32nd ACM International Conference on Multimedia, Melbourne, VIC, Australia, 28 October–1 November 2024; pp. 7764–7773. [Google Scholar]
  24. Wen, J.; Cui, J.; Zhao, Z.; Yan, R.; Gao, Z.; Dou, L.; Chen, B.M. SyreaNet: A Physically Guided Underwater Image Enhancement Framework Integrating Synthetic and Real Images. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), London, UK, 29 May–2 June 2023; pp. 5177–5183. [Google Scholar]
  25. Al-Antari, M.A.; Han, S.M.; Kim, T.S. See through the noise: Revolutionizing medical image diagnosis with quadratic convolutional neural network (Q-CNN). Int. J. Mach. Learn. Cybern. 2024, 15, 1435–1456. [Google Scholar]
  26. Adegun, A.A.; Viriri, S.; Ogundokun, R.O. State-of-the-Art Deep Learning Methods for Objects Detection in Remote Sensing Satellite Images. Sensors 2023, 23, 5849. [Google Scholar] [CrossRef]
  27. Wei, X.; Li, Z.; Wang, Y. SED-YOLO based multi-scale attention for small object detection in remote sensing. Sci. Rep. 2025, 15, 3125. [Google Scholar] [CrossRef] [PubMed]
  28. Liu, J.; Zhao, Y.; Chen, Y.; Zhang, B.; Zheng, W.; Yang, M. A survey of small object detection based on deep learning in aerial images. Artif. Intell. Rev. 2025, 58, 1–74. [Google Scholar] [CrossRef]
  29. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. CBAM: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  30. Wang, Q.; Wu, B.; Zhu, P.; Li, P.; Zuo, W.; Hu, Q. ECA-Net: Efficient channel attention for deep convolutional neural networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 11531–11539. [Google Scholar]
  31. He, K.; Zhang, X.; Ren, S.; Sun, J. Spatial pyramid pooling in deep convolutional networks for visual recognition. In Proceedings of the European Conference on Computer Vision, Zurich, Switzerland, 6–12 September 2014; pp. 346–361. [Google Scholar]
  32. Tan, M.; Pang, R.; Le, Q.V. Efficientdet: Scalable and efficient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 10781–10790. [Google Scholar]
  33. Schiopu, I.; Munteanu, A. Deep Learning Post-Filtering Using Multi-Head Attention and Multiresolution Feature Fusion for Image and Intra-Video Quality Enhancement. Sensors 2022, 22, 1353. [Google Scholar] [CrossRef]
  34. Xie, Y.; Fei, Z.; Deng, D.; Meng, L.; Niu, F.; Sun, J. MEEAFusion: Multi-Scale Edge Enhancement and Joint Attention Mechanism Based Infrared and Visible Image Fusion. Sensors 2024, 24, 5860. [Google Scholar] [CrossRef]
  35. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. In Proceedings of the Advances in Neural Information Processing Systems (NeurIPS), Montreal, QC, Canada, 7–12 December 2015. [Google Scholar]
  36. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. SSD: Single Shot MultiBox Detector. In Proceedings of the European Conference on Computer Vision (ECCV), Amsterdam, The Netherlands, 11–14 October 2016; pp. 21–37. [Google Scholar]
  37. Girshick, R. Fast R-CNN. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015. [Google Scholar]
  38. Zhao, Y.; Lv, W.; Xu, S.; Wei, J.; Wang, G.; Dang, Q.; Liu, Y.; Chen, J. DETRs Beat YOLOs on Real-time Object Detection. arXiv 2023, arXiv:2304.08069. [Google Scholar]
  39. Jocher, G.; Chaurasia, A.; Qiu, J. Ultralytics YOLOv8. 2023. Available online: https://github.com/ultralytics/ultralytics/blob/main/docs/en/models/yolov8.md (accessed on 15 December 2024).
  40. Wang, C.Y.; Yeh, I.H.; Liao, H.Y.M. YOLOv9: Learning What You Want to Learn Using Programmable Gradient Information. arXiv 2024, arXiv:2402.13616. [Google Scholar] [CrossRef]
  41. Wang, A.; Chen, H.; Liu, L.; Chen, K.; Lin, Z.; Han, J.; Ding, G. YOLOv10: Real-Time End-to-End Object Detection. arXiv 2024, arXiv:2405.14458. [Google Scholar]
  42. Jocher, G.; Qiu, J. Ultralytics YOLO11, Version 11.0.0. 2024. Available online: https://github.com/ultralytics/ultralytics/blob/main/docs/en/models/yolo11.md (accessed on 15 December 2024).
  43. Tian, Y.; Ye, Q.; Doermann, D. YOLOv12: Attention-Centric Real-Time Object Detectors. arXiv 2025, arXiv:2502.12524. [Google Scholar]
  44. Lei, M.; Li, S.; Wu, Y.; Hu, H.; Zhou, Y.; Zheng, X.; Ding, G.; Du, S.; Wu, Z.; Gao, Y. YOLOv13: Real-Time Object Detection with Hypergraph-Enhanced Adaptive Visual Perception. arXiv 2025, arXiv:2506.17733. [Google Scholar]
  45. Smiti, A. A critical overview of outlier detection methods. Comput. Sci. Rev. 2020, 38, 100306. [Google Scholar] [CrossRef]
  46. Wu, M.; Arshad, M.H.; Saxena, K.K.; Qian, J.; Reynaerts, D. Profile prediction in ECM using machine learning. Procedia CIRP 2022, 113, 410–416. [Google Scholar] [CrossRef]
  47. Wu, M.; Yao, Z.; Ye, L.; Verbeke, M.; Karsmakers, P.; Reynaerts, D. Geometrical Feature Classification in Electrical Discharge Machining Using In-Process Monitoring and Machine Learning. Procedia CIRP 2025, 137, 462–467. [Google Scholar] [CrossRef]
  48. Zhao, Z. BALF: Simple and Efficient Blur Aware Local Feature Detector. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 3–8 January 2024; pp. 3362–3372. [Google Scholar]
  49. Luo, J.; Wu, Q.; Wang, Y.; Zhou, Z.; Zhuo, Z.; Guo, H. MSHF-YOLO: Cotton growth detection algorithm integrated multi-semantic and high-frequency features. Digit. Signal Process. 2025, 167, 105423. [Google Scholar] [CrossRef]
  50. Yan, S.; Wang, Y.; Zhao, K.; Shi, P.; Zhao, Z.; Zhang, Y.; Li, J. HeMoRa: Unsupervised Heuristic Consensus Sampling for Robust Point Cloud Registration. In Proceedings of the Computer Vision and Pattern Recognition Conference, Nashville, TN, USA, 10–17 June 2025; pp. 1363–1373. [Google Scholar]
  51. Guo, H.; Wu, Q.; Wang, Y. AUHF-DETR: A Lightweight Transformer with Spatial Attention and Wavelet Convolution for Embedded UAV Small Object Detection. Remote Sens. 2025, 17, 1920. [Google Scholar] [CrossRef]
  52. Wu, M.; Guo, Z.; Qian, J.; Reynaerts, D. Multi-Ion-Based Modelling and Experimental Investigations on Consistent and High-Throughput Generation of a Micro Cavity Array by Mask Electrolyte Jet Machining. Micromachines 2022, 13, 2165. [Google Scholar] [CrossRef]
  53. Wu, M.; Hazak Arshad, M.; Kumar Saxena, K.; Reynaerts, D.; Guo, Z.; Liu, J. Experimental and Numerical Investigations on Fabrication of Surface Microstructures Using Mask Electrolyte Jet Machining and Duckbill Nozzle. J. Manuf. Sci. Eng. -Trans. ASME 2023, 145, 051006. [Google Scholar] [CrossRef]
  54. Wu, M.; Guo, Z.; He, J.; Chen, X. Modeling and simulation of the material removal process in electrolyte jet machining of mass transfer in convection and electric migration. Procedia CIRP 2018, 68, 488–492. [Google Scholar] [CrossRef]
  55. Chen, B.; Zha, J.; Cai, Z.; Wu, M. Predictive modelling of surface roughness in precision grinding based on hybrid algorithm. CIRP J. Manuf. Sci. Technol. 2025, 59, 1–17. [Google Scholar] [CrossRef]
  56. Zhang, X.; Wang, J.; Wei, J.; Yuan, X.; Wu, M. A Review of Non-Fully Supervised Deep Learning for Medical Image Segmentation. Information 2025, 16, 433. [Google Scholar] [CrossRef]
  57. Ge, J.; Yao, Z.; Wu, M.; Almeida, J.H.S., Jr.; Jin, Y.; Sun, D. Tackling data scarcity in machine learning-based CFRP drilling performance prediction through a broad learning system with virtual sample generation (BLS-VSG). Compos. Part B Eng. 2025, 305, 112701. [Google Scholar] [CrossRef]
  58. Wu, M.; Yao, Z.; Verbeke, M.; Karsmakers, P.; Gorissen, B.; Reynaerts, D. Data-driven models with physical interpretability for real-time cavity profile prediction in electrochemical machining processes. Eng. Appl. Artif. Intell. 2025, 160, 111807. [Google Scholar] [CrossRef]
  59. Yao, Z.; Wu, M.; Qian, J.; Reynaerts, D. Intelligent discharge state detection in micro-EDM process with cost-effective radio frequency (RF) radiation: Integrating machine learning and interpretable AI. Expert Syst. Appl. 2025, 291, 128607. [Google Scholar] [CrossRef]
  60. Ahmedt-Aristizabal, D.; Armin, M.A.; Denman, S.; Fookes, C.; Petersson, L. Graph-Based Deep Learning for Medical Diagnosis and Analysis: Past, Present and Future. Sensors 2021, 21, 4758. [Google Scholar] [CrossRef] [PubMed]
  61. Wu, M.; Shukla, S.; Vrancken, B.; Verbeke, M.; Karsmakers, P. Data-Driven Approach to Identify Acoustic Emission Source Motion and Positioning Effects in Laser Powder Bed Fusion with Frequency Analysis. Procedia CIRP 2025, 133, 531–536. [Google Scholar] [CrossRef]
  62. Alonso-González, M.J.; Hoogendoorn-Lanser, S.; van Oort, N.; Cats, O.; Hoogendoorn, S. Drivers and barriers in adopting Mobility as a Service (MaaS) – A latent class cluster analysis of attitudes. Transp. Res. Part A Policy Pract. 2020, 132, 378–401. [Google Scholar] [CrossRef]
Figure 1. The overall architecture of the SecureDet model. The SecureDet architecture is engineered for robust multi-scale feature fusion. It employs a Bidirectional Feature Pyramid Network (BiFPN) as its foundational backbone, which is synergistically augmented with an Asymmetric Connection, an Adaptive Spatial Feature Fusion (ASFA) module, and an Edge-aware Context Feature Aggregation (ECFA) module. Specifically, the ECFA module addresses contextual alignment, while the ASFA module is dedicated to spatial and geometric alignment. This integrated framework culminates in the generation of multi-level feature maps—P3, P4, and P5 (with corresponding strides of 8, 16, and 32)—thereby enabling precise detection of objects across a wide range of scales.
Figure 1. The overall architecture of the SecureDet model. The SecureDet architecture is engineered for robust multi-scale feature fusion. It employs a Bidirectional Feature Pyramid Network (BiFPN) as its foundational backbone, which is synergistically augmented with an Asymmetric Connection, an Adaptive Spatial Feature Fusion (ASFA) module, and an Edge-aware Context Feature Aggregation (ECFA) module. Specifically, the ECFA module addresses contextual alignment, while the ASFA module is dedicated to spatial and geometric alignment. This integrated framework culminates in the generation of multi-level feature maps—P3, P4, and P5 (with corresponding strides of 8, 16, and 32)—thereby enabling precise detection of objects across a wide range of scales.
Sensors 25 06603 g001
Figure 2. Receptive Field CBAM Convolution (RFCBAMConv) Module Architecture. Sigmoid activations are confined to attention maps (channel and spatial) for weight estimation; the main feature path applies Batch Normalization and ReLU, avoiding saturation and maintaining stable gradients.
Figure 2. Receptive Field CBAM Convolution (RFCBAMConv) Module Architecture. Sigmoid activations are confined to attention maps (channel and spatial) for weight estimation; the main feature path applies Batch Normalization and ReLU, avoiding saturation and maintaining stable gradients.
Sensors 25 06603 g002
Figure 3. Enhanced Contextual Feature Alignment (ECFA) Module Architecture.
Figure 3. Enhanced Contextual Feature Alignment (ECFA) Module Architecture.
Sensors 25 06603 g003
Figure 4. Adaptive Spatial Feature Alignment (ASFA) Module Architecture. The semantic branch applies a 3 × 3 convolution followed by 4 × bilinear upsampling before fusion.
Figure 4. Adaptive Spatial Feature Alignment (ASFA) Module Architecture. The semantic branch applies a 3 × 3 convolution followed by 4 × bilinear upsampling before fusion.
Sensors 25 06603 g004
Figure 5. Dataset overview and statistics: training set contains 7780 images and 18,885 instances; test set contains 1944 images and 4671 instances; total is 9724 images and 23,556 instances. Eight categories are included: knife, scissors, lighter, USBFlashDisk, pressure, plasticBottleWithaNozzle, seal, battery. Occlusion distribution (by instances): training—none 78.15%, light 12.36%, medium 6.96%, heavy 2.52%; test—none 78.35%, light 11.52%, medium 7.38%, heavy 2.76%; total—none 78.27%, light 12.19%, medium 7.05%, heavy 2.57%. Occlusion levels are defined by the proportion of object area occluded: none (0%), light (<20%), medium (20–50%), heavy (>50%).
Figure 5. Dataset overview and statistics: training set contains 7780 images and 18,885 instances; test set contains 1944 images and 4671 instances; total is 9724 images and 23,556 instances. Eight categories are included: knife, scissors, lighter, USBFlashDisk, pressure, plasticBottleWithaNozzle, seal, battery. Occlusion distribution (by instances): training—none 78.15%, light 12.36%, medium 6.96%, heavy 2.52%; test—none 78.35%, light 11.52%, medium 7.38%, heavy 2.76%; total—none 78.27%, light 12.19%, medium 7.05%, heavy 2.57%. Occlusion levels are defined by the proportion of object area occluded: none (0%), light (<20%), medium (20–50%), heavy (>50%).
Sensors 25 06603 g005
Figure 6. Ablation study visualization results showing SecureDet’s superior detection performance across different model configurations. The figure demonstrates comprehensive evaluation of eight different architectural combinations on challenging X-ray luggage images with dense, overlapping electronics. SecureDet-s achieves perfect precision and recall while baseline methods struggle with missed detections of small and occluded objects. Detailed quantitative metrics are provided in Table A1. “GT” denotes “Ground Truth”.
Figure 6. Ablation study visualization results showing SecureDet’s superior detection performance across different model configurations. The figure demonstrates comprehensive evaluation of eight different architectural combinations on challenging X-ray luggage images with dense, overlapping electronics. SecureDet-s achieves perfect precision and recall while baseline methods struggle with missed detections of small and occluded objects. Detailed quantitative metrics are provided in Table A1. “GT” denotes “Ground Truth”.
Sensors 25 06603 g006
Figure 7. Severe occlusion scenario analysis demonstrating SecureDet’s superior performance in detecting contraband items under heavy occlusion conditions. SecureDet-s achieves the highest detection accuracy across all scenarios. Detailed quantitative metrics are provided in Table A1. “GT” denotes “Ground Truth”.
Figure 7. Severe occlusion scenario analysis demonstrating SecureDet’s superior performance in detecting contraband items under heavy occlusion conditions. SecureDet-s achieves the highest detection accuracy across all scenarios. Detailed quantitative metrics are provided in Table A1. “GT” denotes “Ground Truth”.
Sensors 25 06603 g007
Figure 8. Small detection performance analysis showing SecureDet’s superior capability in detecting small objects in X-ray security screening. SecureDet-s achieves perfect detection while other methods fail. Detailed quantitative metrics are provided in Table A1. “GT” denotes “Ground Truth”. Red boxes indicate predicted bounding boxes for knives by baseline detectors; when no corresponding GT exists, they represent false positives. Specifically, in the second series scenario, Fast R-CNN’s red box marks a knife prediction where none is present, i.e., a misdetection.
Figure 8. Small detection performance analysis showing SecureDet’s superior capability in detecting small objects in X-ray security screening. SecureDet-s achieves perfect detection while other methods fail. Detailed quantitative metrics are provided in Table A1. “GT” denotes “Ground Truth”. Red boxes indicate predicted bounding boxes for knives by baseline detectors; when no corresponding GT exists, they represent false positives. Specifically, in the second series scenario, Fast R-CNN’s red box marks a knife prediction where none is present, i.e., a misdetection.
Sensors 25 06603 g008
Figure 9. Multi-scale detection performance analysis demonstrating SecureDet’s superior capability in handling extreme size variations. See Table A1 for detailed quantitative metrics. GT” denotes “Ground Truth”.
Figure 9. Multi-scale detection performance analysis demonstrating SecureDet’s superior capability in handling extreme size variations. See Table A1 for detailed quantitative metrics. GT” denotes “Ground Truth”.
Sensors 25 06603 g009
Figure 10. Grad-CAM-based attention visualization across different architectures (YOLOv8s and SecureDet-s) on representative X-ray scenarios The highlighted regions indicate salient features driving the final predictions.
Figure 10. Grad-CAM-based attention visualization across different architectures (YOLOv8s and SecureDet-s) on representative X-ray scenarios The highlighted regions indicate salient features driving the final predictions.
Sensors 25 06603 g010
Table 1. RFCBAMConv component hyperparameters (design-formula view).
Table 1. RFCBAMConv component hyperparameters (design-formula view).
ComponentSetting
Channel attention bottlenecktwo FC layers, reduction ratio r = 16 , activation: sigmoid
Spatial attention [ max c ( F unfold ) , avg c ( F unfold ) ] Conv 3 × 3 sigmoid
Context constructiondepth-wise separable Conv k × k ; spatial unfold (patch/stride: design-formula)
Attention applicationsequential weighting: M ch then M sp
Output layer Conv k × k with stride s = k ; BatchNorm + ReLU
Table 2. Architectural scaling parameters of SecureDet variants.
Table 2. Architectural scaling parameters of SecureDet variants.
VariantBackboneDepth MultiplierWidth MultiplierDesign Objective
SecureDet-nYOLOv8n0.330.25Maximize efficiency for edge devices
SecureDet-sYOLOv8s0.330.50Balance speed and accuracy for general scenarios
SecureDet-mYOLOv8m0.670.75Pursue highest accuracy for high-performance servers
Table 3. Baseline training and inference configuration (compact).
Table 3. Baseline training and inference configuration (compact).
ItemSetting
Input resolution 640 × 640
Batch size16
Epochs300
OptimizerSGD (initial learning rate 0.01 )
LossesBCE (cls, obj) + CIoU (box)
Schedulecosine with warmup; EMA; weight decay; light label smoothing (design-formula)
Augmentationstrong data augmentation (design-formula)
Early stoppingby validation mAP@0.5:0.95
Inference thresholdsconfidence 0.25 , NMS IoU 0.7
MetricsPrecision, Recall, mAP@0.5, mAP@0.5:0.95; Params, GFLOPs, FPS
Table 4. Performance comparison of different YOLO models.
Table 4. Performance comparison of different YOLO models.
MethodP (%)R (%)mAP@0.5 (%)mAP@0.5:0.95 (%)Params (M)GFLOPsFPS
YOLOv8s91.06572.72579.28766.37913.828.683.4
YOLOv8m83.5981.381.3268.325.978.945.3
YOLOv8n86.89464.54173.33956.0793.28.8122.3
Fast R-CNN84.78162.53770.84954.065N/AN/AN/A
RT-DETR-R5087.75765.61874.34661.42822.8139.858.7
EfficientDet-d289.27470.92279.08663.3988.111.041.57
YOLOv10l91.24271.12377.72364.49724.4120.346.4
YOLOv10x91.98873.63378.46669.48729.5160.435.5
YOLOv11n83.23559.89569.84352.5752.96.597.8
YOLOv11s90.77572.21879.59766.0889.421.561.1
YOLOv11m93.92973.0382.08871.68620.168.025.46
YOLOv13n84.240.750.537.53.17.5126.9
YOLOv13s84.353.362.247.09.321.492.3
YOLOv13m85.10760.11268.31153.27920.267.553.5
SecureDet-n88.00667.61875.52959.973.07.1113.1
SecureDet-s93.55873.03481.62869.5447.5521.478.5
SecureDet-m92.60274.61182.25872.31219.258.959.2
Note: Bold values indicate the best performance results among all compared methods for each metric.
Table 5. Ablation study results. SecureDet-s achieves its parsimonious 7.55 M architecture by replacing YOLOv8s’s substantial 11.13 M PANet neck with a lean BiFPN. Conversely, RFC+ECFA+ASFA augments the original neck, accumulating significantly more parameters (13.22 M).
Table 5. Ablation study results. SecureDet-s achieves its parsimonious 7.55 M architecture by replacing YOLOv8s’s substantial 11.13 M PANet neck with a lean BiFPN. Conversely, RFC+ECFA+ASFA augments the original neck, accumulating significantly more parameters (13.22 M).
ConfigurationRFCBAMBiFPNECFAASFAP (%)R (%)mAP@0.5mAP@0.5:0.95Params (M)
YOLOv8s (Baseline) 91.0772.7379.2966.3811.13
only-BiFPN 87.8770.0877.9359.997.37
RFCBAM 90.0171.7679.8564.9111.41
RFC+BiFPN 90.0572.1878.9564.6110.95
ECFA+ASFA 91.5973.5280.7965.9012.50
ECFA+ASFA+BiFPN 89.8869.1177.8861.757.07
RFC+ECFA+ASFA 88.8872.6378.8264.4313.22
SecureDet-s (Full Model)93.5673.0381.6369.547.55
Note: ✓ indicates that the corresponding component/module is included in that configuration. Bold values indicate the best performance results achieved by our full SecureDet-s model compared to all ablation configurations.
Table 6. Performance comparison on different datasets.
Table 6. Performance comparison on different datasets.
DatasetsMethodsP (%)R (%)mAP@0.5 (%)mAP@0.5:0.95 (%)
OPIXrayFast R-CNN81.06572.72579.28738.379
YOLOv8s90.46686.53189.88943.242
RT-DETR89.586.53590.49154.469
EfficientDet86.40982.52887.92542.674
YOLOv12s84.78162.53789.84942.065
SecureDet-s89.80887.10691.15654.436
HiXrayFast R-CNN82.17661.17269.4251.885
YOLOv8s87.83270.4778.09962.295
RT-DETR93.43173.48181.25968.345
EfficientDet90.1876.7380.98250.011
YOLOv12s89.27470.92279.08663.398
SecureDet-s84.79581.1183.58569.12
Note: Bold values indicate the best performance results among all compared methods for each metric in the respective dataset.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lin, Y.; Lin, Y.; Wu, H.; Wu, M. The Enhance-Fuse-Align Principle: A New Architectural Blueprint for Robust Object Detection, with Application to X-Ray Security. Sensors 2025, 25, 6603. https://doi.org/10.3390/s25216603

AMA Style

Lin Y, Lin Y, Wu H, Wu M. The Enhance-Fuse-Align Principle: A New Architectural Blueprint for Robust Object Detection, with Application to X-Ray Security. Sensors. 2025; 25(21):6603. https://doi.org/10.3390/s25216603

Chicago/Turabian Style

Lin, Yuduo, Yanfeng Lin, Heng Wu, and Ming Wu. 2025. "The Enhance-Fuse-Align Principle: A New Architectural Blueprint for Robust Object Detection, with Application to X-Ray Security" Sensors 25, no. 21: 6603. https://doi.org/10.3390/s25216603

APA Style

Lin, Y., Lin, Y., Wu, H., & Wu, M. (2025). The Enhance-Fuse-Align Principle: A New Architectural Blueprint for Robust Object Detection, with Application to X-Ray Security. Sensors, 25(21), 6603. https://doi.org/10.3390/s25216603

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop