DBFormer: A Dual-Branch Adaptive Remote Sensing Image Resolution Fine-Grained Weed Segmentation Network
Abstract
1. Introduction
2. Materials and Data
2.1. Data Sources
2.2. Dataset Preparation
3. Methods
3.1. Overall Framework of the Model
- 1.
- Encoder MEncoder
- Feature Embedding and Hierarchical Processing: Initially, the input image is divided into overlapping patches (overlap patch embedding) to preserve local continuity while reducing resolution and computational complexity. Subsequently, four cascaded Transformer blocks are employed, each using an overlap patch-merging mechanism to efficiently transmit information while reducing dimensionality.
- Efficient Self-Attention (Efficient Self-Attn): The encoder incorporates a self-attention mechanism to effectively capture long-range dependencies, thereby enhancing the model’s global contextual understanding of complex agricultural areas.
- Mix–Feed Forward Network (Mix-FFN): Following each Transformer block, a mix-FFN is applied to perform nonlinear feature transformations, enriching the feature representation and improving the model’s ability to discriminate field characteristics.
- Feature Output: After multiple layers of Transformer processing, the encoder generates a deep feature map FEncoder, which is then passed to the decoder for further processing.
- 2.
- Decoder MDecoder
- SMLP Feature Transformation: In the decoding stage, the SMLP structure is first applied to map the features, enhancing the compatibility among multi-scale features to ensure effective fusion across different levels.
- Dynamic Context Aggregation Branch (DCA-Branch): This branch is primarily responsible for modeling global contextual information. By employing adaptive downsampling and attention mechanisms to perform weighted fusion on multi-scale features, it strengthens cross-scale information interaction, enabling a more accurate recognition of large-scale field regions.
- Local Detail Enhancement Branch (LDE-Branch): Focused on extracting local details, this branch utilizes localized convolution operations and detail enhancement strategies to strengthen the boundary features of small field regions, ensuring precise segmentation of minute weed targets.
- Feature Fusion and Edge-Aware Loss: At the end of the decoder, a feature fusion operation integrates the features extracted by the DCA-Branch and the LDE-Branch. An Edge-Aware Loss is then applied to reinforce field boundary information, thereby improving the model’s capability to segment complex terrains and mixed-crop regions.
- Final Output: After multiple layers of SMLP processing, dual-branch fusion, and upsampling operations, the decoder ultimately produces a high-precision segmentation result Iresult with clear boundaries and coherent overall regions.
3.2. Model Encoder MEncoder
3.2.1. Overlap Patch Embeddings
3.2.2. Stage 1: Initial Feature Extraction
3.2.3. Stage 2: Multi-Scale Feature Extraction
3.2.4. Stage 3: High-Level Feature Extraction
3.2.5. Stage 4: Deep Feature Representation
3.3. Model Decoder MDecoder
3.3.1. SMLP Stage
3.3.2. Dynamic Context Aggregation Branch (DCA-Branch)
- 1.
- Channel Splitting
- 2.
- Adaptive Down-sampling Factor Calculation
- 3.
- Down-sampling and Depthwise Convolution
- 4.
- Upsampling and Attention Fusion
3.3.3. Local Detail Enhancement Branch (LDE-Branch)
- The first block expands the channels using a 1 × 1 convolution, then applies a 3 × 3 depthwise convolution to extract local textures followed by an activation function;
- The second block further enhances feature representation with another 3 × 3 depthwise convolution and lastly compresses the channels back to their original dimensions. The output is added to the input as a residual connection, as formulated in Equation (13):Flde = DMLP (LDEsub) = LDEsub + Δ,
3.3.4. Feature Fusion and Edge-Aware Loss
3.3.5. Output Stage
3.4. Experimental Environment
- Hyperparameter Settings: The initial learning rate was set to 0.001 and adjusted using cosine annealing. The batch size was fixed at 4. The network was trained using the AdamW optimizer in conjunction with a cross-entropy loss to improve boundary sensitivity.
- Training Settings: All experiments were conducted on a system equipped with an NVIDIA GeForce RTX 2080 Ti GPU, an Intel Core i7-8700K CPU, and 32 GB of RAM. The training process was carried out for 100 epochs with early stopping employed to prevent overfitting.
4. Results
4.1. Model Comparison
4.2. Evaluation of Different Models
4.3. Test Results of Different Models
4.3.1. Tobacco Dataset
4.3.2. Sunflower Dataset
4.4. Ablation Study
4.4.1. Determination of Ablation Targets
4.4.2. Experimental Setup
- Training Process: All ablation experiments were performed on the same training and validation sets, the Tobacco Dataset and Sunflower Dataset, that were used for the complete model, and with identical training parameters.
- Evaluation Metrics: The performance of each ablation variant was evaluated using pixel-based error metrics, including Precision, Pixel Accuracy, mIoU, and Recall. The results were compared against those obtained by the complete model.
4.4.3. Results Analysis
- Removing the DCA-Branch results in a decrease in mIoU to 84.77% on the Tobacco Dataset and a decrease to 79.29% on the Sunflower Dataset, accompanied by noticeable reductions in Pixel Accuracy and Recall. This indicates that the DCA-Branch is critical for effectively modeling global context and capturing cross-scale information, which is particularly important for the accurate segmentation of large-scale regions.
- Removing the LDE-Branch leads to an mIoU drop to 84.33% on the Tobacco Dataset and a drop to 80.65% on the Sunflower Dataset, demonstrating that the LDE-Branch plays a significant role in extracting fine-grained details and edge information, thereby enhancing the model’s ability to recognize small weeds and crop boundaries.
- Removing the Edge-Aware Loss causes a slight reduction in mIoU, to 86.00%, on the Tobacco Dataset and a more pronounced decrease, to 80.13%, on the Sunflower Dataset. Although the impact relative to the Tobacco Dataset is relatively minor, the effect for the Sunflower Dataset is more significant, underscoring the positive role of edge constraints in improving boundary clarity and reducing segmentation errors.
- Removing both DCA and the Edge-Aware Loss (DBFormer-noDCA-noEAL) leads to a further performance drop compared to individual removals: mIoU falls by 5.21% (to 81.27%) on the Tobacco Dataset and by 6.18% (to 79.31%) on the Sunflower Dataset, indicating a strong interaction between global context modeling and edge supervision.
- Similarly, DBFormer-noLDE-noEAL sees mIoU decreases of 6.15% (to 80.33%) and 7.05% (to 78.44%) on the Tobacco Dataset and the Sunflower Dataset, respectively—greater than the sum of its single removals—highlighting the synergistic contribution of local detail enhancement and edge-aware loss.
5. Discussion
5.1. Summary of Research Contributions
- Innovative Dual-Branch Network Architecture: DBFormer integrates a dynamic context aggregation branch (DCA-Branch) and a local detail enhancement branch (LDE-Branch). Through adaptive downsampling, depthwise convolutions, and attention mechanisms, the network achieves efficient multi-scale information fusion. The DCA-Branch captures large-scale background and global semantic information, while the LDE-Branch, enhanced via residual structures, reinforces local texture and edge features. Their complementary actions enable the model to suppress over-segmentation while accurately capturing small targets.
- High Accuracy and Stable Segmentation Performance: Experiments on both the Tobacco Dataset and the Sunflower Dataset demonstrate that the complete DBFormer model achieves outstanding performance in terms of Precision, Pixel Accuracy, mIoU, and Recall, while also exhibiting low Over-segmentation and Under-segmentation scores. These results validate the model’s ability to maintain stable and precise segmentation in complex backgrounds, confirming its application value in precision agriculture remote sensing image segmentation.
- Effective Ablation Study Validation: Ablation experiments conducted using the DCA-Branch, LDE-Branch, and Edge-Aware Loss reveal that the removal of any of these modules results in a significant performance drop. This further confirms the necessity and complementary roles of these key components in information fusion and feature enhancement.
- Good Generalization Capability: The excellent performance of DBFormer across different crop datasets demonstrates its strong generalization ability and its adaptability to a variety of agricultural scenarios. This approach provides stable and reliable technical support for weed detection and crop management using remote sensing images, thereby facilitating smart monitoring and management in precision agriculture.
5.2. Computational Efficiency of the Models
5.3. Limitations and Future Work
- Dataset Diversity and Complexity: Our experiments focus on the Tobacco and Sunflower Datasets, which differ in resolution and seasonality but still represent limited crop types and environmental conditions. Future work should evaluate DBFormer on datasets that are larger and more diverse—covering different crop species, weed varieties, growth stages, and imaging conditions (e.g., varying illumination and flight altitudes)—to fully assess its generalization and robustness.
- Real-Time Processing and Resource Consumption: Although DBFormer has achieved a favorable balance in computational efficiency and memory usage, there is still room for optimization under extremely resource-constrained or real-time scenarios. Future research may explore network pruning, quantization, or design strategies that are more lightweight in order to further reduce computational complexity and enhance deployment performance on embedded systems or mobile devices.
- Multi-Modal Information Fusion: The current model primarily relies on single-source multi-spectral remote sensing images. Future work could explore the fusion of multi-modal data (e.g., LiDAR, or hyperspectral information) to utilize additional dimensions of information, thereby enhancing target segmentation performance and improving the discrimination of fine-grained targets with complex backgrounds.
- Further Exploration of Context and Edge Constraints: Although the dynamic context aggregation branch and edge-aware loss have proven effective in capturing global context and enhancing edge features, there is still a risk of misclassification in regions with blurred edges or weak textures. Future work could explore more refined context modeling and edge enhancement strategies, such as the incorporation of self-supervised learning or multi-scale cross-attention mechanisms, to further improve segmentation accuracy.
- Cross-Domain Adaptation and Transfer Learning: Domain shifts—in sensor type, geographic region, or imaging season—can degrade model performance (Figure 10 shows the failure cases). Integrating transfer learning, unsupervised domain adaptation (e.g., adversarial alignment), or meta-learning strategies would enhance DBFormer’s adaptability across varied remote sensing platforms and crop–weed systems.
6. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Chen, C.; Yuan, X.; Gan, S.; Luo, W.; Bi, R.; Li, R.; Gao, S. A new vegetation index based on UAV for extracting plateau vegetation information. Int. J. Appl. Earth Obs. Geoinf. 2024, 128, 103668. [Google Scholar] [CrossRef]
- Tang, Y.; Qiu, F.; Wang, B.; Wu, D.; Jing, L.; Sun, Z. A deep relearning method based on the recurrent neural network for land cover classification. GISci. Remote Sens. 2022, 59, 1344–1366. [Google Scholar] [CrossRef]
- Ha, N.T.T.; Vinh, P.Q.; Thao, N.T.P.; Linh, P.H.; Parsons, M.; Van Manh, N. A Method for Assessing the Lake Trophic Status Using Hyperspectral Reflectance (400–900 nm) Measured Above Water. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2024, 17, 17890–17902. [Google Scholar] [CrossRef]
- Mishra, R.K.; Agarwal, A.; Shukla, A. Predicting ground level PM2. 5 concentration over Delhi using Landsat 8 satellite data. Int. J. Remote Sens. 2021, 42, 827–838. [Google Scholar] [CrossRef]
- Xing, X.; Yu, B.; Kang, C.; Huang, B.; Gong, J.; Liu, Y. The synergy between remote sensing and social sensing in urban studies: Review and perspectives. IEEE Geosci. Remote Sens. Mag. 2024, 12, 108–137. [Google Scholar] [CrossRef]
- Zhou, D.; Xiao, J.; Bonafoni, S.; Berger, C.; Deilami, K.; Zhou, Y.; Frolking, S.; Yao, R.; Qiao, Z.; Sobrino, J.A. Satellite remote sensing of surface urban heat islands: Progress, challenges, and perspectives. Remote Sens. 2018, 11, 48. [Google Scholar] [CrossRef]
- López, A.; Jurado, J.M.; Ogayar, C.J.; Feito, F.R. A framework for registering UAV-based imagery for crop-tracking in Precision Agriculture. Int. J. Appl. Earth Obs. Geoinf. 2021, 97, 102274. [Google Scholar] [CrossRef]
- Qiu, B.; Jiang, F.; Chen, C.; Tang, Z.; Wu, W.; Berry, J. Phenology-pigment based automated peanut mapping using sentinel-2 images. GISci. Remote Sens. 2021, 58, 1335–1351. [Google Scholar] [CrossRef]
- Zhao, R.; Shi, F. A novel strategy for pest disease detection of Brassica chinensis based on UAV imagery and deep learning. Int. J. Remote Sens. 2022, 43, 7083–7103. [Google Scholar] [CrossRef]
- Cui, Y.; Chen, X.; Xiong, W.; He, L.; Lv, F.; Fan, W.; Luo, Z.; Hong, Y. A soil moisture spatial and temporal resolution improving algorithm based on multi-source remote sensing data and GRNN model. Remote Sens. 2020, 12, 455. [Google Scholar] [CrossRef]
- Gao, J.; Liao, W.; Nuyttens, D.; Lootens, P.; Vangeyte, J.; Pižurica, A.; He, Y.; Pieters, J.G. Fusion of pixel and object-based features for weed mapping using unmanned aerial vehicle imagery. Int. J. Appl. Earth Obs. Geoinf. 2018, 67, 43–53. [Google Scholar] [CrossRef]
- Du, X.; Zare, A. Multiresolution multimodal sensor fusion for remote sensing data with label uncertainty. IEEE Trans. Geosci. Remote Sens. 2019, 58, 2755–2769. [Google Scholar] [CrossRef]
- Cui, C.; Sun, X.; Fu, B.; Shang, X. SSANet-BS: Spectral–spatial cross-dimensional attention network for hyperspectral band selection. Remote Sens. 2024, 16, 2848. [Google Scholar] [CrossRef]
- Fu, B.; Sun, X.; Cui, C.; Zhang, J.; Shang, X. Structure-preserved and weakly redundant band selection for hyperspectral imagery. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2024, 17, 12490–12504. [Google Scholar] [CrossRef]
- Andrea, C.-C.; Daniel, B.B.M.; Misael, J.B.J. Precise weed and maize classification through convolutional neuronal networks. In Proceedings of the 2017 IEEE Second Ecuador Technical Chapters Meeting (ETCM), Salinas, Ecuador, 16–20 October 2017; pp. 1–6. [Google Scholar]
- Milioto, A.; Lottes, P.; Stachniss, C. Real-time blob-wise sugar beets vs. weeds classification for monitoring fields using convolutional neural networks. ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci. 2017, 4, 41–48. [Google Scholar] [CrossRef]
- Espejo-Garcia, B.; Mylonas, N.; Athanasakos, L.; Fountas, S.; Vasilakoglou, I. Towards weeds identification assistance through transfer learning. Comput. Electron. Agric. 2020, 171, 105306. [Google Scholar] [CrossRef]
- Alam, M.; Alam, M.S.; Roman, M.; Tufail, M.; Khan, M.U.; Khan, M.T. Real-time machine-learning based crop/weed detection and classification for variable-rate spraying in precision agriculture. In Proceedings of the 2020 7th international conference on electrical and electronics engineering (ICEEE), Antalya, Turkey, 14–16 April 2020; pp. 273–280. [Google Scholar]
- Karimi, Y.; Prasher, S.; Patel, R.; Kim, S. Application of support vector machine technology for weed and nitrogen stress detection in corn. Comput. Electron. Agric. 2006, 51, 99–109. [Google Scholar] [CrossRef]
- Ishak, A.J.; Mokri, S.S.; Mustafa, M.M.; Hussain, A. Weed detection utilizing quadratic polynomial and ROI techniques. In Proceedings of the 2007 5th Student Conference on Research and Development, Selangor, Malaysia, 12–11 December 2007; pp. 1–5. [Google Scholar]
- Wendel, A.; Underwood, J. Self-supervised weed detection in vegetable crops using ground based hyperspectral imaging. In Proceedings of the 2016 IEEE international conference on robotics and automation (ICRA), Stockholm, Sweden, 16–21 May 2016; pp. 5128–5135. [Google Scholar]
- Rasti, P.; Ahmad, A.; Samiei, S.; Belin, E.; Rousseau, D. Supervised image classification by scattering transform with application to weed detection in culture crops of high density. Remote Sens. 2019, 11, 249. [Google Scholar] [CrossRef]
- Nkemelu, D.K.; Omeiza, D.; Lubalo, N. Deep convolutional neural network for plant seedlings classification. arXiv 2018, arXiv:1811.08404. [Google Scholar]
- Partel, V.; Kakarla, S.C.; Ampatzidis, Y. Development and evaluation of a low-cost and smart technology for precision weed management utilizing artificial intelligence. Comput. Electron. Agric. 2019, 157, 339–350. [Google Scholar] [CrossRef]
- Sa, I.; Chen, Z.; Popović, M.; Khanna, R.; Liebisch, F.; Nieto, J.; Siegwart, R. weednet: Dense semantic weed classification using multispectral images and mav for smart farming. IEEE Robot. Autom. Lett. 2017, 3, 588–595. [Google Scholar] [CrossRef]
- Huang, H.; Lan, Y.; Yang, A.; Zhang, Y.; Wen, S.; Deng, J. Deep learning versus Object-based Image Analysis (OBIA) in weed mapping of UAV imagery. Int. J. Remote Sens. 2020, 41, 3446–3479. [Google Scholar] [CrossRef]
- Machidon, A.L.; Krašovec, A.; Pejović, V.; Machidon, O.M. SqueezeSlimU-Net: An Adaptive and Efficient Segmentation Architecture for Real-Time UAV Weed Detection. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2025, 18, 5749–5764. [Google Scholar] [CrossRef]
- Wang, D.; Du, B.; Zhang, L. Fully contextual network for hyperspectral scene parsing. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–16. [Google Scholar] [CrossRef]
- Song, D.; Dong, Y.; Li, X. Context and difference enhancement network for change detection. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 9457–9467. [Google Scholar] [CrossRef]
- Gao, Y.; Li, L.; Weiss, M.; Guo, W.; Shi, M.; Lu, H.; Jiang, R.; Ding, Y.; Nampally, T.; Rajalakshmi, P. Bridging real and simulated data for cross-spatial-resolution vegetation segmentation with application to rice crops. ISPRS J. Photogramm. Remote Sens. 2024, 218, 133–150. [Google Scholar] [CrossRef]
- Xiao, D.; Kang, Z.; Fu, Y.; Li, Z.; Ran, M. Csswin-unet: A Swin-unet network for semantic segmentation of remote sensing images by aggregating contextual information and extracting spatial information. Int. J. Remote Sens. 2023, 44, 7598–7625. [Google Scholar] [CrossRef]
- Li, W.; Liang, S.; Chen, K.; Chen, Y.; Ma, H.; Xu, J.; Ma, Y.; Guan, S.; Fang, H.; Shi, Z. AgriFM: A Multi-source Temporal Remote Sensing Foundation Model for Crop Mapping. arXiv 2025, arXiv:2505.21357. [Google Scholar]
- Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30, 6000–6010. [Google Scholar]
- Moazzam, S.I.; Khan, U.S.; Qureshi, W.S.; Nawaz, T.; Kunwar, F. Towards automated weed detection through two-stage semantic segmentation of tobacco and weed pixels in aerial Imagery. Smart Agric. Technol. 2023, 4, 100142. [Google Scholar] [CrossRef]
- Fawakherji, M.; Potena, C.; Pretto, A.; Bloisi, D.D.; Nardi, D. Multi-spectral image synthesis for crop/weed segmentation in precision farming. Robot. Auton. Syst. 2021, 146, 103861. [Google Scholar] [CrossRef]
- Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the Medical image computing and computer-assisted intervention–MICCAI 2015: 18th international conference, Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
- Zhao, H.; Shi, J.; Qi, X.; Wang, X.; Jia, J. Pyramid scene parsing network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2881–2890. [Google Scholar]
- Chen, L.-C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-decoder with atrous separable convolution for semantic image segmentation. In Proceedings of the European conference on computer vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 801–818. [Google Scholar]
- Xie, E.; Wang, W.; Yu, Z.; Anandkumar, A.; Alvarez, J.M.; Luo, P. SegFormer: Simple and efficient design for semantic segmentation with transformers. Adv. Neural Inf. Process. Syst. 2021, 34, 12077–12090. [Google Scholar]
- Xu, J.; Xiong, Z.; Bhattacharyya, S.P. PIDNet: A real-time semantic segmentation network inspired by PID controllers. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 19529–19539. [Google Scholar]
- Qurratulain, S.; Zheng, Z.; Xia, J.; Ma, Y.; Zhou, F. Deep learning instance segmentation framework for burnt area instances characterization. Int. J. Appl. Earth Obs. Geoinf. 2023, 116, 103146. [Google Scholar] [CrossRef]
- Yu, L.; Zeng, Z.; Liu, A.; Xie, X.; Wang, H.; Xu, F.; Hong, W. A lightweight complex-valued DeepLabv3+ for semantic segmentation of PolSAR image. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 930–943. [Google Scholar] [CrossRef]
- Ma, Z.; Xia, M.; Lin, H.; Qian, M.; Zhang, Y. FENet: Feature enhancement network for land cover classification. Int. J. Remote Sens. 2023, 44, 1702–1725. [Google Scholar] [CrossRef]
- Du, R.; Ma, Z.; Xie, P.; He, Y.; Cen, H. PST: Plant segmentation transformer for 3D point clouds of rapeseed plants at the podding stage. ISPRS J. Photogramm. Remote Sens. 2023, 195, 380–392. [Google Scholar] [CrossRef]
- Xiao, P.; Zhang, X.; Zhang, H.; Hu, R.; Feng, X. Multiscale optimized segmentation of urban green cover in high resolution remote sensing image. Remote Sens. 2018, 10, 1813. [Google Scholar] [CrossRef]
- Xiang, D.; Zhang, F.; Zhang, W.; Tang, T.; Guan, D.; Zhang, L.; Su, Y. Fast pixel-superpixel region merging for SAR image segmentation. IEEE Trans. Geosci. Remote Sens. 2020, 59, 9319–9335. [Google Scholar] [CrossRef]
- Lei, L.; Chai, G.; Yao, Z.; Li, Y.; Jia, X.; Zhang, X. A novel self-similarity cluster grouping approach for individual tree crown segmentation using multi-features from UAV-based LiDAR and multi-angle photogrammetry data. Remote Sens. Environ. 2025, 318, 114588. [Google Scholar] [CrossRef]
Name | Configuration |
---|---|
CPU | Intel Core(TM) i7-8700K 32 G, Intel Corporation, Santa Clara, CA, USA |
GPU | NVIDIA GeForce RTX 2080 Ti, NVIDIA, Santa Clara, CA, USA |
Operating System | Windows 10 |
Deep Learning Framework | Pytorch 1.12 |
Datasets | Model | Precision (%) | Pixel Accuracy (%) | mIoU (%) | Recall (%) |
---|---|---|---|---|---|
Tobacco Dataset | UNet [36] | 87.72 ± 0.60 | 86.80 ± 0.55 | 78.46 ± 0.50 | 85.55 ± 0.60 |
PSPNet [37] | 88.05 ± 0.65 | 79.54 ± 0.60 | 72.78 ± 0.55 | 79.23 ± 0.60 | |
DeepLab V3+ [38] | 83.36 ± 0.70 | 81.53 ± 0.65 | 72.24 ± 0.60 | 80.15 ± 0.65 | |
SegFormer [39] | 87.62 ± 0.55 | 84.42 ± 0.50 | 76.57 ± 0.45 | 82.44 ± 0.50 | |
PIDNet [40] | 89.92 ± 0.50 | 86.98 ± 0.45 | 82.65 ± 0.40 | 88.00 ± 0.50 | |
DBFormer (Ours) | 92.89 ± 0.45 | 92.08 ± 0.50 | 86.48 ± 0.40 | 91.50 ± 0.50 | |
Sunflower Dataset | UNet [36] | 85.21 ± 0.65 | 81.17 ± 0.60 | 75.46 ± 0.55 | 80.12 ± 0.60 |
PSPNet [37] | 84.12 ± 0.70 | 76.44 ± 0.65 | 70.00 ± 0.60 | 75.80 ± 0.65 | |
DeepLab V3+ [38] | 77.14 ± 0.75 | 80.82 ± 0.70 | 71.06 ± 0.65 | 79.50 ± 0.70 | |
SegFormer [39] | 80.33 ± 0.70 | 70.44 ± 0.65 | 66.97 ± 0.60 | 68.97 ± 0.65 | |
PIDNet [40] | 88.48 ± 0.55 | 83.92 ± 0.50 | 81.06 ± 0.45 | 83.30 ± 0.50 | |
DBFormer (Ours) | 94.52 ± 0.35 | 88.73 ± 0.40 | 85.49 ± 0.45 | 88.00 ± 0.40 |
Metric | DBFormer | PIDNet | p-Value |
---|---|---|---|
OS score | 0.038 ± 0.006 | 0.057 ± 0.008 | 0.035 |
US score | 0.042 ± 0.005 | 0.065 ± 0.007 | 0.008 |
Metric | DBFormer | PIDNet | p-Value |
---|---|---|---|
OS score | 0.034 ± 0.006 | 0.072 ± 0.010 | 0.012 |
US score | 0.029 ± 0.005 | 0.063 ± 0.009 | 0.028 |
Datasets | Model | Precision (%) | Pixel Accuracy (%) | mIoU (%) | Recall (%) |
---|---|---|---|---|---|
Tobacco Dataset | DBFormer | 92.89 ± 0.45 | 92.08 ± 0.50 | 86.48 ± 0.40 | 91.50 ± 0.50 |
DBFormer-noDCA | 92.45 ± 0.50 | 90.45 ± 0.55 | 84.77 ± 0.45 | 90.12 ± 0.55 | |
DBFormer-noLDE | 91.89 ± 0.55 | 90.41 ± 0.60 | 84.33 ± 0.50 | 89.80 ± 0.60 | |
DBFormer-noEAL | 92.58 ± 0.55 | 91.79 ± 0.50 | 86.00 ± 0.40 | 91.52 ± 0.50 | |
DBFormer-noDCA-noEAL | 91.32 ± 0.60 | 89.54 ± 0.65 | 81.27 ± 0.55 | 89.21 ± 0.65 | |
DBFormer-noLDE-noEAL | 90.41 ± 0.60 | 88.72 ± 0.55 | 80.33 ± 0.50 | 87.52 ± 0.60 | |
Sunflower Dataset | DBFormer | 94.52 ± 0.35 | 88.73 ± 0.40 | 85.49 ± 0.45 | 88.00 ± 0.40 |
DBFormer-noDCA | 92.59 ± 0.50 | 82.34 ± 0.60 | 79.29 ± 0.55 | 82.05 ± 0.60 | |
DBFormer-noLDE | 90.66 ± 0.60 | 85.27 ± 0.55 | 80.65 ± 0.50 | 84.79 ± 0.55 | |
DBFormer-noEAL | 92.79 ± 0.50 | 83.29 ± 0.55 | 80.13 ± 0.45 | 83.06 ± 0.55 | |
DBFormer-noDCA-noEAL | 91.15 ± 0.55 | 80.21 ± 0.60 | 79.31 ± 0.50 | 80.11 ± 0.60 | |
DBFormer-noLDE-noEAL | 90.03 ± 0.55 | 82.10 ± 0.55 | 78.44 ± 0.50 | 81.75 ± 0.55 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
She, X.; Tang, Z.; Pan, X.; Zhao, J.; Liu, W. DBFormer: A Dual-Branch Adaptive Remote Sensing Image Resolution Fine-Grained Weed Segmentation Network. Remote Sens. 2025, 17, 2203. https://doi.org/10.3390/rs17132203
She X, Tang Z, Pan X, Zhao J, Liu W. DBFormer: A Dual-Branch Adaptive Remote Sensing Image Resolution Fine-Grained Weed Segmentation Network. Remote Sensing. 2025; 17(13):2203. https://doi.org/10.3390/rs17132203
Chicago/Turabian StyleShe, Xiangfei, Zhankui Tang, Xin Pan, Jian Zhao, and Wenyu Liu. 2025. "DBFormer: A Dual-Branch Adaptive Remote Sensing Image Resolution Fine-Grained Weed Segmentation Network" Remote Sensing 17, no. 13: 2203. https://doi.org/10.3390/rs17132203
APA StyleShe, X., Tang, Z., Pan, X., Zhao, J., & Liu, W. (2025). DBFormer: A Dual-Branch Adaptive Remote Sensing Image Resolution Fine-Grained Weed Segmentation Network. Remote Sensing, 17(13), 2203. https://doi.org/10.3390/rs17132203