Next Article in Journal
Dual-Band Planar Microwave Solid Complex Dielectric Constant Sensor System Based on E-Interdigital Structure
Previous Article in Journal
Integrating Inertial Sensors to Assess Physical Performance and In-Match Demands for the International Selection of Cerebral Palsy Football Players
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Dual-Path CSDETR: Cascade Stochastic Attention with Object-Centric Priors for High-Accuracy Fire Detection

1
Key Laboratory of Fire Protection Technology for Industry and Public Building, Ministry of Emergency Management, Tianjin 300381, China
2
Tianjin Fire Science and Technology Research Institute of MEM, Tianjin 300381, China
3
Sino-European Institute of Aviation Engineering, Civil Aviation University of China, Tianjin 300300, China
4
Institute of Energy, Peking University, Beijing 100871, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(18), 5788; https://doi.org/10.3390/s25185788
Submission received: 15 August 2025 / Revised: 13 September 2025 / Accepted: 15 September 2025 / Published: 17 September 2025
(This article belongs to the Section Intelligent Sensors)

Abstract

Detecting dynamic and amorphous objects like fire and smoke poses significant challenges in object detection. To address this, we propose Dual-Path Cascade Stochastic DETR (Dual-Path CSDETR). Unlike Cascade DETR, our model introduces cascade stochastic attention (CSA) to model the irregular morphologies of fire and smoke through variational inference, combined with a dual-path architecture that enables bidirectional feature interaction for enhanced learning efficiency. By integrating object-centric priors from bounding boxes into each decoder layer, the model refines attention mechanisms to focus on critical regions. Experiments show that Dual-Path CSDETR achieves 94% AP50 on fire/smoke detection, surpassing deterministic baselines.

1. Introduction

Fire incidents pose severe threats to life and property, underscoring the urgent need for reliable detection systems. Despite advancements in deep learning [1,2] and transformer-based detection [3,4,5], accurately identifying dynamic, irregular fire/smoke patterns remains challenging due to their variability in appearance [6], diverse environmental contexts, and limited training data. Traditional methods struggle with three critical limitations: sensitivity to combustion materials and fire stages [6], poor generalization across scenarios (e.g., forests vs. indoor environments), and deterministic attention mechanisms’ inability to model complex variations without extensive data.
To address these challenges, we propose Dual-Path Cascade Stochastic DETR (Dual-Path CSDETR), which introduces a cascade stochastic attention (CSA) module leveraging variational inference to probabilistically model fire/smoke morphologies. Unlike conventional approaches, CSA integrates a dual-path architecture—combining a stochastic downward path for iterative attention refinement via bounding box priors and a deterministic upward path for feature stabilization—to enhance bidirectional information flow. By embedding object-centric gamma priors from predicted bounding boxes at each decoder layer, the model adaptively focuses on critical regions while maintaining computational efficiency for real-time deployment. Experiments demonstrate that our framework significantly outperforms existing methods, achieving 94% AP50 in fire/smoke detection.
The remainder of this paper is organized as follows: Section 2 reviews related work, Section 3 details the Dual-Path CSDETR methodology, Section 4 presents experimental results, and Section 5 concludes the study.

2. Related Works

Recent advances in object detection have been driven by the integration of convolutional and transformer-based architectures. Convolutional Neural Networks (CNNs), including seminal works like Faster R-CNN [7] and YOLO [8], established robust frameworks for feature extraction and bounding box regression. The introduction of the Detection Transformer (DETR) [9] marked a paradigm shift by eliminating handcrafted components through transformer-based object queries and bipartite matching losses. While DETR achieves competitive accuracy, its limitations in small-object detection and occlusion handling prompted innovations like Deformable DETR [10], which accelerates convergence by 40% through deformable attention mechanisms, and Cascade DETR [11], which improves occluded object detection by 8.2% mAP via multi-stage refinement.
In fire and smoke detection, domain-specific adaptations of these frameworks have emerged. Decoder-free transformer models [12] enable early flame recognition through simplified architectures, while Deformable DETR-based detectors [13] achieve real-time performance in dynamic fire scenarios. Lightweight variants like YOLO-FM [14] reduce parameters by 30% while maintaining 45 FPS on edge devices and YOLOv5s hybrids [15] integrate channel attention modules to boost smoke detection F1-scores to 0.89. Despite these advancements, persistent challenges arise from dataset limitations (e.g., 12% annotation errors in public fire datasets [16]) and environmental variability, where deterministic attention mechanisms [17,18,19] falter under rapid illumination changes or dense smoke occlusion [20].
Stochastic attention mechanisms, which model attention weights as probabilistic distributions, offer a promising solution. Early hard-attention approaches [21] employed categorical distributions, but faced optimization barriers due to non-reparameterizability. Bayesian attention [22] addressed this through reparameterizable Dirichlet distributions, yet its application remains confined to shallow NLP models. Recent extensions like deep stochastic networks [23] demonstrate potential in machine translation, but lack adaptation to vision tasks requiring spatial uncertainty modeling—a critical gap for fire detection, where probabilistic attention could mitigate false negatives in occluded scenarios by 9% [22]. This underscores the need to integrate stochastic principles into transformer-based detectors, enabling adaptive feature focusing in high-variance fire environments.

3. Preliminaries

3.1. Bayesian Attention Modules

Bayesian Attention Modules (BAMs) [9] introduce stochasticity into attention mechanisms by modeling attention weights as random variables governed by probabilistic distributions. Unlike traditional deterministic attention, BAM parameterizes unnormalized attention scores W ^ using continuous distributions, such as Weibull or Lognormal, which capture uncertainty in the attention process. These parameters are non-regularized, treated as latent variables learned from the input features. To enable gradient-based optimization, reparameterization is applied, turning the sampling process deterministic and allowing efficient backpropagation. The variational objective, L B A M , combines the task-specific loss L t a s k with a KL divergence term, regularizing the posterior distribution q ( W ^ | X ) against a data-dependent prior p ( W ^ | X )
L B A M = E q ( α ˜ | X ) [ L t a s k ( y , f ( X , α ˜ ) ) ] D K L ( q ( α ˜ | X ) p ( α ˜ | X ) )
where the α ˜ denotes the set of Gamma shape parameters learned by the network and D K L denotes the KL divergence operator. This formulation allows BAMs to incorporate uncertainty in attention, enhancing flexibility and generalization, while ensuring efficient training via reparameterization.

3.2. Cascade-DETR

Cascade-DETR, introduced by Ye, M. et al. [11], enhances the Detection Transformer (DETR) by improving localization accuracy and generalization across diverse domains. Its key innovation is the Cascade Attention Layer, which refines the attention mechanism in the detection decoder. In standard DETR models, the decoder uses cross-attention over the entire image feature map, which can lead to imprecise localization, especially in complex scenes. Cascade-DETR addresses this by iteratively refining attention in each decoding layer, focusing on regions predicted by the previous layer. Specifically, the attention mechanism at layer l + 1 is confined to regions inside the predicted bounding box B l from layer l, ensuring that attention progressively narrows down to relevant areas. This refinement improves localization precision and enhances DETR’s performance across datasets like COCO and UDB10.

4. Dual-Path Cascade Stochastic DETR

4.1. Cascade Stochastic Attention Layers

In this section, we present the formulation for learning the distribution of stochastic attention weights. Note that unnormalized attention weights have been shown to outperform normalized weights in capturing distribution patterns [22]. Let the unnormalized attention weights be denoted as W ^ = { W ^ l } l = 1 : L , treated as data-dependent latent variables sampled from distribution q ϕ Following prior work [22,24], we employ amortized variational inference q ϕ ( W ^ ) to approximate the true posterior of local attention weights under a Bayesian framework, conditioned on training data ( x , y ) .
For each attention layer l, we approximate the posterior distribution of unnormalized attention weights p ( W ^ l | x , y ) with a variational distribution q ϕ ( W ^ l ) , that is, q ϕ ( W ^ l ) p ( W ^ l x , y ) . In particular, we parameterize q ϕ ( W ^ l ) as a Weibull distribution and the prior p η ( W ^ l ) as a Gamma distribution:
q ϕ ( W ^ l ) = Weibull ( k l ; λ l = f ϕ ( K l , Q l ) )
p η ( W ^ l ) = Gamma ( α l = f η ( K l ) ; β )
where the K l and Q l denote the key and query in l layer attention, where k l , λ l are the Weibull shape and scale parameters, α l , β denote the Gamma shape and rate parameters, and f ϕ and f η are neural networks parameterized by ϕ and η , respectively. The hyperparameter β is a globally fixed, while k l is layer-specific.
The rationale for distribution choices is twofold: (1) The Weibull distribution offers dynamic adaptability to scale variations, which is crucial for fire detection scenarios. In the early layers, a small k l mimics an exponential distribution to explore diffuse regions (e.g., smoke edges). As the layers deepen, increasing k l reduces variance, resembling a Gaussian-like distribution for precise fire-core localization. (2) The Gamma distribution serves as a natural prior for modeling positive-valued variables, such as attention weights. It also enables the efficient computation of the KL divergence. Thus, the combination of these distributions supports effective learning and model convergence in fire detection tasks.
The variational formulation extends naturally to multi-layer attention with L layers. Both the approximate posterior q ϕ ( W ^ 1 : L ) and prior p η ( W ^ 1 : L ) are modeled as joint distributions. Using the chain rule to preserve layer-wise dependencies,
q ϕ ( W ^ 1 : L ) = l = 1 L q ϕ ( W ^ l | W ^ l 1 ) , p η ( W ^ 1 : L ) = l = 1 L p η ( W ^ l | W ^ l 1 )
where the prior employs factorized Gamma distributions for conjugacy. The semi-analytic KL divergence between the Weibull posterior and Gamma prior ensures efficient gradient computation (Section 4.3).

4.2. Dual-Path CSDETR Architecture

In this section, we describe the architecture of Dual-Path-based CSDETR, which injects both local object-centric bias and a dual-path into the CSA module in Section 4.1. This architecture is based on the standard Cascade DETR [11], which consists of encoder layers and decoder layers. The encoder remains unchanged, with features passed to the decoder. The decoder uses learnable object queries to localize and classify fire/smoke. Self-attention remains deterministic to model query relationships, while cross-attention employs stochastic attention layers to capture fire/smoke distributions. Two modifications integrate CSA with Cascade DETR, bounding-box-based prior distribution and dual-path architecture, enhancing information utilization under limited training data.
  • Dual-Path Architecture
As shown in the work introduced by Zhang, S et al. and Ren, W. et al. [23,25], using only the downward structure (i.e., previous decoder layers) can lead to instability. A dual-path architecture is employed by combining downward (prior-driven) and upward (likelihood-driven) paths that stabilize training. Hence, λ l in Equation (2) is obtained by
λ l = σ · ln 1 + exp ( f ϕ , 1 l ( h l ) ) + exp ( Φ l ) Γ ( 1 + 1 / k l ) the stochastic downward path h l = ln 1 + exp ( f ϕ , l l ( h l + 1 ) ) the deterministic upward path
where σ [ 0 , 1 ] is the importance weight. f ϕ , 1 l , f ϕ , 2 l are linear projections, and Φ l is the scaled dot product between keys K l and queries Q l , i.e., Φ l = Q l · K l d k . The upward path propagates deterministic hidden states h L , initialized from h L + 1 = softmax ( Φ 1 ) , to refine layer-wise features (Figure 1).
  • Object-Centric Prior-Based Attention Refinement
As highlighted in Cascade-DETR, local information surrounding each object is crucial for accurate detection [11]. Building upon this insight, we integrate the object-centric spatial priors to inform the learning of parameters of p η ( W ^ l ) and q ϕ ( W ^ l ) , i.e., λ l and α l in Equations (2) and (3), ensuring that the model’s attention is focused on object-relevant regions. Mathematically, the refined key is given by K box l = M ( B l 1 ) , where M is a mapping function that extracts feature maps, refined to better align with the bounding box B l 1 from the previous layer. Next, using the object-centric prior as a constraint, we rewrite p η ( W ^ l ) in Equation (3) as
p η ( W ^ l ) = Gamma ( α l = f η ( K box l ) ; β )
A two-layer Multi-Layer Perceptron (MLP) is employed to process K box l and produce the predicted parameters α l , which are used to shape the distribution of the object’s prior [23]:
α l = softmax f η , 2 l ( ReLU ( f η , 1 l ( K box l ) ) )
where f η , 1 l , f η , 2 l are two linear layers connected by ReLU.
The keys within the bounding box are also refined to generate q ϕ ( W ^ l ) , which models the uncertainty or variability of the features within the box. The attention distribution for regions outside the box is preserved from the previous layer. Specifically, Φ l in Equation (5) is updated as Φ l = Q l · K box l d k .
By refining the keys within the bounding box and incorporating them into the CSA, the model enhances its attention on regions with a higher likelihood of fire and smoke, as shown in Figure 2. The overall structure of the Dual-Path CSDETR decoder is illustrated in Figure 3.

4.3. Training and Inference

Our Cascade Stochastic DETR is trained end-to-end with a multi-task loss L Loss :
L Loss = L Detect ELBO
where L Detect (inherited from the Cascade-DETR) supervizes bounding box regression and classification, while the Evidence Lower Bound (ELBO), in terms of the learned parameters W ^ , combines reconstruction loss and layer-wise KL divergence:
ELBO = E q ϕ ( W ^ ) log p θ ( y x , W ^ ) KL ( q ϕ ( W ^ ) p η ( W ^ ) )
The KL divergence from the prior to variational distributions can be computed in a semi-analytic way [22] as
KL ( q ϕ ( W ^ ) p η ( W ^ ) ) = l = 1 L E q ϕ ( W ^ 1 : l 1 ) KL ( q ϕ ( W ^ l W ^ 1 : l 1 ) p η ( W ^ l W ^ 1 : l 1 ) ) analytic
The semi-analytic KL term decomposes into layer-specific divergences between the Weibull posterior q ϕ ( W ^ l W ^ 1 : l 1 ) and Gamma prior p η ( W ^ l W ^ 1 : l 1 ) :
KL ( q ϕ ( W ^ l ; k , λ l ) p η ( W ^ l ; α l , β ) ) = α l k l α l log λ l + log k l + β λ l Γ 1 + 1 k l γ log β + log Γ ( α l )
Using Rao-Blackwellization, we reduce Monte Carlo variance by integrating analytic KL components. Reparameterization with uniform noise ϵ enables differentiable sampling:
ELBO = E ϵ [ log p θ ( y x , r ( ϵ ) ) l = 1 L KL q ϕ ( W ^ l r ϕ ( ϵ 1 : l 1 ) ) p η ( W ^ l r ϕ ( ϵ 1 : l 1 ) ) ]
where W ^ = r ( ϵ ) = λ ln ( 1 ϵ ) 1 / k , which is the reparameterized expression of the Weibull distribution, ϵ is a uniform random variable ϵ Uniform ( 0 , 1 ) .
During both learning and inference, we sample ϵ l to generate unnormalized attention weights using r l ( ϵ l ) , and normalize them with the previously mentioned equation. By employing this reparameterization technique, we manage the inherent randomness in the attention mechanism, enabling gradient-based optimization, which is essential for large-scale training tasks.

5. Experiments

5.1. Datasets and Evaluation Metrics

To evaluate the proposed method for detecting highly variable fire and smoke, four public datasets, ranging from 502 images to 21,000 images, were selected, representing different scenarios. The details of the datasets are shown in Table 1. We evaluated Dual-Path DETR using mAP@[0.5: 0.95], AP50, and AP75 to quantify detection accuracy across IoU thresholds, where higher values indicate stronger fire localization and classification performance.

5.2. Comparison with State-of-the-Art

We evaluated Dual-Path Cascade Stochastic DETR against Fast-RCNN, Yolo-FM, Deformable DETR, and Cascade DETR on FireNet, Smoke, D-Fire, and DFS-Fire-Smoke datasets. Using ResNet-101 (ImageNet-pretrained) with AdamW optimizer (backbone/transformer: l r : 1 × 10 5 / 1 × 10 4 ), Fast-RCNN, Deformable-DETR, Cascade DETR, and our model were trained for 50 epochs on FireNet and Smoke, while Yolo-FM was trained for 100 epochs. On D-Fire and DFS-Fire-Smoke, Fast-RCNN, Deformable-DETR, Cascade DETR, and our model were trained for 150 epochs, and Yolo-FM for 300 epochs. All of the simulations and tests were performed on a workstation with four 3090 GPUs. The total batch size was set to 16, with 4 per GPU.
The performance results are shown in Table 2, and the best result within each evaluation metric is highlighted in bold. As shown in Table 2, our model achieved state-of-the-art AP50 (0.94 on FireNet, 0.92 on DFS-Fire-Smoke) and superior AP75, demonstrating precise localization, critical for early fire detection. Learning curves in Figure 4 reveal consistent mAP@[0.5, 0.95] dominance across datasets, particularly for challenging “other” classes, validating the CSA module’s false-positive suppression. The object detection results of the proposed method are displayed in Figure 5, further illustrating its robustness and accuracy in detecting fire and smoke in various scenarios. The paired t-test is carried out to verify the improvement significance. The H0 hypothesis was set to no significant improvement. We chose only Fast-RCNN, YOLO-FM, Deformable-DETR, Cascade-DETR as the baseline model due to the time constraints. Each model was trained and tested on the same datasets five times. The Precision results are used to calculated the p-values in the t-test. Paired t-tests (Figure 6) confirmed statistically significant improvements over baselines ( p < 0.05 on FireNet/DFS-Fire-Smoke) and marginal gains on Smoke/D-Fire (p = 0.05–0.1).

5.3. Ablation Study

In this section, we conduct experiments to evaluate the proposed Dual-Path DETR. First, we compare the performance of Bayesian attention with SSA across different decoder layers, using Cascade-DETR as the baseline and varying the number of stochastic decoder layers with Gamma priors from 1 to 6, to assess the impact of stochastic attention on model performance. Additionally, we study the efficiency of the dual-path structure by comparing it to a single stochastic downward path, highlighting the advantages of integrating both stochastic and deterministic pathways in improving attention refinement and overall model robustness.
In contrast, Table 3 examines the impact of applying Bayesian attention to different decoder layers. While the results indicate that increasing the number of layers using Bayesian attention generally improves performance, the accuracy remains lower compared to SSA. This is because Bayesian attention, which focuses on individual attention layers, is less suitable for the DETR structure in object detection tasks. The comprehensive application of SSA across all layers enables the model to better capture the complex distributions necessary for fire and smoke detection.
The findings from this ablation study highlight the importance of utilizing a sufficient number of SSA layers to fully exploit the benefits of stochastic attention in the Dual-Path DETR model. By replacing all layers with SSA, the model can more effectively capture the complex variability required for accurate fire and smoke detection, leading to higher AP values.
The results of the first experiment are presented in Table 3 and Table 4. Table 4 shows that when SSA is applied only to the first three layers while the remaining layers use deterministic attention, the model fails to achieve high AP values. This underperformance is likely due to the limited number of SSA layers, which are insufficient to learn a robust attention distribution, especially when mixed with deterministic attention in the upper layers. The limited application of SSA does not capture the complex variability necessary for precise fire and smoke detection. However, when all decoder layers are replaced with SSA incorporating Gamma distributions, there is a marked improvement in performance. This improvement is attributed to the model’s ability to effectively learn the joint distribution across multiple layers, allowing for better handling of uncertainty and variability, which is crucial for accurate detection.
We also conducted another ablation study to examine the role of the deterministic upward path. In this study, we compared two variants of the Dual-Path DETR: one with the deterministic upward path and one without it. The results are shown in Figure 7.
The observations from this experiment indicate that the deterministic upward path significantly enhances model stability. For all datasets, the inclusion of the deterministic upward path results in a more stable learning process and improved convergence. In contrast, without the upward path, the learning process becomes highly unstable and difficult to converge. This instability likely arises from the increased uncertainty and variance in the attention weights, which the deterministic upward path helps to mitigate.
These findings highlight the critical role of the deterministic upward path in ensuring the stability and convergence of the Dual-Path DETR. By incorporating this path, the model can achieve more reliable and consistent performance across different datasets.

6. Conclusions

We propose Dual-Path Cascade Stochastic DETR for fire/smoke detection, integrating cascade stochastic attention and dual-path architecture to address challenges of irregular, dynamic objects. By modeling attention via Weibull/Gamma distributions with object-centric bounding box priors, the framework achieves 94% AP50, outperforming deterministic baselines. Future work will prioritize real-time deployment via model pruning [30], false alert reduction through uncertainty quantification [25], temporal modeling for video analysis, and multi-sensor fusion using Dempster-Shafer theory [31].

Author Contributions

Conceptualization, D.Y.; methodology, D.Y.; software, D.Y.; validation, D.Y. and W.R.; formal analysis, D.Y., B.H. and X.Z.; investigation, D.Y. and X.Z.; resources, B.H.; data curation, W.R.; writing—original draft preparation, D.Y.; writing—review and editing, B.H.; supervision, B.H.; funding acquisition, D.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by Key Laboratory of Fire Protection Technology for Industry and Public Building, Ministry of Emergency Management Open Project 2023KLIB01 and National Fire and Rescue Administration Science Project 2023XFCX37,Natural Science Foundation of Inner Mongolia 2025YQ008, and National Natural Science Foundation of China 52304034.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

These data were derived from the following resources available in the public domain: D-Fire Dataset: gaiasd. Dfiredataset. Available online: https://github.com/gaiasd/DFireDataset, 2023. Accessed on 12 July 2024. DFS-fire-smoke dataset: Siyuan Wu, Xinrong Zhang, Ruqi Liu, and Binhai Li. A dataset for fire and smoke object detection. Multimedia Tools and Applications, pages 1–20, 2022. DFS-dataset: Available online: https://github.com/siyuanwu/DFS-FIRE-SMOKE-Dataset/tree/main (accessed on 16 September 2025).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Wang, X.; Wang, B.; Luo, P.; Wang, L.; Wu, Y. A Metric Learning-Based Improved Oriented R-CNN for Wildfire Detection in Power Transmission Corridors. Sensors 2025, 25, 3882. [Google Scholar] [CrossRef]
  2. Desikan, J.; Singh, S.K.; Jayanthiladevi, A.; Bhushan, S.; Rishiwal, V.; Kumar, M. Hybrid Machine Learning-Based Fault-Tolerant Sensor Data Fusion and Anomaly Detection for Fire Risk Mitigation in IIoT Environment. Sensors 2025, 25, 2146. [Google Scholar] [CrossRef]
  3. Zhou, K.; Jiang, S. Forest fire detection algorithm based on improved YOLOv11n. Sensors 2025, 25, 2989. [Google Scholar] [CrossRef]
  4. Buriboev, A.S.; Abduvaitov, A.; Jeon, H.S. Integrating Color and Contour Analysis with Deep Learning for Robust Fire and Smoke Detection. Sensors 2025, 25, 2044. [Google Scholar] [CrossRef]
  5. Zhang, Z.; Tan, L.; Robert, T.L.K. An improved fire and smoke detection method based on YOLOv8n for smart factories. Sensors 2024, 24, 4786. [Google Scholar] [CrossRef]
  6. Abdessemed, F.; Bouam, S.; Arar, C. A Review on Forest Fire Detection and Monitoring Systems. In Proceedings of the 2023 International Conference on Electrical Engineering and Advanced Technology (ICEEAT), Batna, Algeria, 5–7 November 2023; Volume 1, pp. 1–7. [Google Scholar]
  7. Girshick, R. Fast R-CNN. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  8. Redmon, J.; Divvala, S.K.; Girshick, R.B.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 26 June–1 July 2016; pp. 779–788. [Google Scholar]
  9. Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; Zagoruyko, S. End-to-end object detection with transformers. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 213–229. [Google Scholar]
  10. Zhu, X.; Su, W.; Lu, L.; Li, B.; Wang, X.; Dai, J. Deformable DETR: Deformable Transformers for End-to-End Object Detection. arXiv 2021, arXiv:2010.04159. [Google Scholar] [CrossRef]
  11. Ye, M.; Ke, L.; Li, S.; Tai, Y.W.; Tang, C.K.; Danelljan, M.; Yu, F. Cascade-DETR: Delving into High-Quality Universal Object Detection. In Proceedings of the 2023 IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 2–3 October 2023; pp. 6681–6691. [Google Scholar]
  12. Wang, X.; Li, M.; Gao, M.; Liu, Q.; Li, Z.; Kou, L. Early smoke and flame detection based on transformer. J. Saf. Sci. Resil. 2023, 4, 294–304. [Google Scholar] [CrossRef]
  13. Liang, T.; Zeng, G. FSH-DETR: An Efficient End-to-End Fire Smoke and Human Detection Based on a Deformable DEtection TRansformer (DETR). Sensors 2024, 24, 4077. [Google Scholar] [CrossRef]
  14. Geng, X.; Su, Y.; Cao, X.; Li, H.; Liu, L. YOLOFM: An improved fire and smoke object detection algorithm based on YOLOv5n. Sci. Rep. 2024, 14, 4543. [Google Scholar] [CrossRef] [PubMed]
  15. Liang, D.; Bui, T.; Wang, G. Fire and Smoke Detection Method Based on Improved YOLOv5s. In Proceedings of the 2023 8th International Conference on Communication, Image and Signal Processing (CCISP), Chengdu, China, 17–19 November 2023; pp. 293–300. [Google Scholar]
  16. Yar, H.; Khan, Z.A.; Ullah, F.U.M.; Ullah, W.; Baik, S.W. A modified YOLOv5 architecture for efficient fire detection in smart cities. Expert Syst. Appl. 2023, 231, 120465. [Google Scholar] [CrossRef]
  17. Koshy, R.; Elango, S. Applying social media in emergency response: An attention-based bidirectional deep learning system for location reference recognition in disaster tweets. Appl. Intell. 2024, 54, 5768–5793. [Google Scholar] [CrossRef]
  18. Lv, G.; Dong, L.; Xu, W. Hierarchical interactive multi-granularity co-attention embedding to improve the small infrared target detection. Appl. Intell. 2023, 53, 27998–28020. [Google Scholar] [CrossRef]
  19. Wang, J.; Yu, J.; He, Z. DECA: A novel multi-scale efficient channel attention module for object detection in real-life fire images. Appl. Intell. 2022, 52, 1362–1375. [Google Scholar] [CrossRef]
  20. Zeng, K.; Sun, X.; He, H.; Tang, H.; Shen, T.; Zhang, L. Fuzzy preference matroids rough sets for approximate guided representation in transformer. Expert Syst. Appl. 2024, 255, 124592. [Google Scholar] [CrossRef]
  21. Xu, K.; Ba, J.; Kiros, R.; Cho, K.; Courville, A.C.; Salakhutdinov, R.; Zemel, R.S.; Bengio, Y. Show, Attend and Tell: Neural Image Caption Generation with Visual Attention. arXiv 2015, arXiv:1502.03044. [Google Scholar] [CrossRef]
  22. Fan, X.; Zhang, S.; Chen, B.; Zhou, M. Bayesian Attention Modules. arXiv 2020, arXiv:2010.10604. [Google Scholar] [CrossRef]
  23. Zhang, S.; Fan, X.; Chen, B.; Zhou, M. Bayesian Attention Belief Networks. arXiv 2021, arXiv:2106.05251. [Google Scholar] [CrossRef]
  24. Blei, D.M.; Kucukelbir, A.; McAuliffe, J.D. Variational inference: A review for statisticians. J. Am. Stat. Assoc. 2017, 112, 859–877. [Google Scholar] [CrossRef]
  25. Ren, W.; Jin, Z. Phase space visibility graph. Chaos Solitons Fractals 2023, 176, 114170. [Google Scholar] [CrossRef]
  26. Olafenwa, M. FireNET Dataset [Dataset]; GitHub: San Francisco, CA, USA, 2023. [Google Scholar]
  27. Pacini, M. Smoke Dataset [Dataset]; Roboflow Universe: Des Moines, IA, USA, 2022. [Google Scholar]
  28. Gaiasd. DFireDataset. 2023. Available online: https://github.com/gaiasd/DFireDataset (accessed on 12 July 2024).
  29. Wu, S.; Zhang, X.; Liu, R.; Li, B. A dataset for fire and smoke object detection. In Multimedia Tools and Applications; Springer: Berlin/Heidelberg, Germany, 2022; pp. 1–20. [Google Scholar]
  30. Ren, W.; Jin, N.; Wang, T. An Interdigital Conductance Sensor for Measuring Liquid Film Thickness in Inclined Gas–Liquid Two-Phase Flow. IEEE Trans. Instrum. Meas. 2024, 73, 9505809. [Google Scholar] [CrossRef]
  31. Li, N.; Martin, A.; Estival, R. Heterogeneous information fusion: Combination of multiple supervised and unsupervised classification methods based on belief functions. Inf. Sci. 2021, 544, 238–265. [Google Scholar] [CrossRef]
Figure 1. Dual-path architecture.
Figure 1. Dual-path architecture.
Sensors 25 05788 g001
Figure 2. The proposed stochastic attention layer.
Figure 2. The proposed stochastic attention layer.
Sensors 25 05788 g002
Figure 3. The proposed Dual-Path CSDETR.
Figure 3. The proposed Dual-Path CSDETR.
Sensors 25 05788 g003
Figure 4. Learning process comparision on FireNet, Smoke, D-Fire and DFS-Fire-Smoke datasets: (a) FireNet. (b) Smoke. (c) D-Fire. (d) DFS-Fire-Smoke.
Figure 4. Learning process comparision on FireNet, Smoke, D-Fire and DFS-Fire-Smoke datasets: (a) FireNet. (b) Smoke. (c) D-Fire. (d) DFS-Fire-Smoke.
Sensors 25 05788 g004
Figure 5. Fire detection results on different datasets: (a) FireNet. (b) Smoke. (c) D-Fire. (d) DFS-Fire-Smoke.
Figure 5. Fire detection results on different datasets: (a) FireNet. (b) Smoke. (c) D-Fire. (d) DFS-Fire-Smoke.
Sensors 25 05788 g005
Figure 6. Paired t-test p-values comparing Dual-Path DETR with baseline models.
Figure 6. Paired t-test p-values comparing Dual-Path DETR with baseline models.
Sensors 25 05788 g006
Figure 7. Comparison of Dual-Path architecture with stochastic downward path architecture: (a) FireNet. (b) Smoke. (c) D-Fire. (d) DFS-Fire-Smoke.
Figure 7. Comparison of Dual-Path architecture with stochastic downward path architecture: (a) FireNet. (b) Smoke. (c) D-Fire. (d) DFS-Fire-Smoke.
Sensors 25 05788 g007
Table 1. Overview of datasets.
Table 1. Overview of datasets.
DatasetsClassesPositive/Total Ratio
FireNet [26]Fire210/502
Smoke [27]Smoke741/746
D-Fire [28]Fire & Smoke4658/21,000
DFS-Fire-Smoke [29]Fire & Smoke & Other6308/9462
Table 2. Performance comparison of the proposed method and other classical detectors.
Table 2. Performance comparison of the proposed method and other classical detectors.
DatasetsEvaluationFast-RCNNYolo-FMDeformable
DETR
Cascade
DETR
Ours
FireNetAP500.850.890.820.920.94
AP750.790.850.810.880.91
mAP@[0.5, 0.95]0.750.730.790.810.88
SmokeAP500.920.840.890.900.91
AP750.810.790.850.870.89
mAP@[0.5, 0.95]0.830.820.790.840.86
D-FireAP500.880.780.910.890.94
AP750.760.730.790.900.88
mAP@[0.5, 0.95]0.740.710.810.820.85
DFS-Fire-SmokeAP500.820.840.870.860.92
AP750.730.790.750.810.84
mAP@[0.5, 0.95]0.710.730.750.800.84
Table 3. Performance comparison of varying decoder layer numbers without bounding box information.
Table 3. Performance comparison of varying decoder layer numbers without bounding box information.
Datasets Evaluation Decoder Layers with Bayesian Attention
1 Layer 2 Layers 3 Layers 4 Layers 5 Layers 6 Layers
FireNetAP500.380.390.420.440.470.50
AP750.310.330.360.380.410.44
mAP@[0.5, 0.95]0.290.350.410.420.450.46
SmokeAP500.360.380.400.430.460.49
AP750.300.320.350.370.400.43
mAP@[0.5, 0.95]0.290.310.330.390.380.42
D-FireAP500.300.320.350.380.410.44
AP750.240.260.290.320.350.38
mAP@[0.5, 0.95]0.260.250.280.300.330.36
DFS-Fire-SmokeAP500.250.270.300.330.360.39
AP750.190.210.240.270.300.33
mAP@[0.5, 0.95]0.210.230.220.310.330.35
Table 4. Performance comparison of Dual-Path CSDETR with varying decoder layer numbers.
Table 4. Performance comparison of Dual-Path CSDETR with varying decoder layer numbers.
Datasets Evaluation Decoder Layers with the SSA Mechanism
1 Layer 2 Layers 3 Layers 4 Layers 5 Layers 6 Layers
FireNetAP500.600.590.680.740.850.92
AP750.530.530.510.550.670.88
mAP@[0.5, 0.95]0.570.550.620.610.770.91
SmokeAP500.630.590.680.690.720.89
AP750.590.430.550.540.610.84
mAP@[0.5, 0.95]0.570.560.570.650.630.86
D-FireAP500.560.550.570.650.820.92
AP750.480.440.430.480.660.82
mAP@[0.5, 0.95]0.500.520.490.500.770.89
DFS-Fire-SmokeAP500.510.590.650.660.860.89
AP750.410.450.440.490.620.73
mAP@[0.5, 0.95]0.450.520.600.520.800.84
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yu, D.; Han, B.; Zhao, X.; Ren, W. Dual-Path CSDETR: Cascade Stochastic Attention with Object-Centric Priors for High-Accuracy Fire Detection. Sensors 2025, 25, 5788. https://doi.org/10.3390/s25185788

AMA Style

Yu D, Han B, Zhao X, Ren W. Dual-Path CSDETR: Cascade Stochastic Attention with Object-Centric Priors for High-Accuracy Fire Detection. Sensors. 2025; 25(18):5788. https://doi.org/10.3390/s25185788

Chicago/Turabian Style

Yu, Dongxing, Bing Han, Xinyi Zhao, and Weikai Ren. 2025. "Dual-Path CSDETR: Cascade Stochastic Attention with Object-Centric Priors for High-Accuracy Fire Detection" Sensors 25, no. 18: 5788. https://doi.org/10.3390/s25185788

APA Style

Yu, D., Han, B., Zhao, X., & Ren, W. (2025). Dual-Path CSDETR: Cascade Stochastic Attention with Object-Centric Priors for High-Accuracy Fire Detection. Sensors, 25(18), 5788. https://doi.org/10.3390/s25185788

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop