Forest Road Extraction via Optimized DeepLabv3+ and Multi-Temporal Remote Sensing for Wildfire Emergency Response
Abstract
1. Introduction
2. Materials and Methods
2.1. Study Area and Data Preprocessing
2.1.1. Overview of the Study Area
2.1.2. Data Acquisition and Temporal Phase Selection
2.1.3. Data Preprocessing and Dataset Construction
2.2. Methodology
2.2.1. Network Architecture Selection
2.2.2. Multi-Scale Dynamic Spatial-Channel Attention (MCSA) Module
2.2.3. Enhanced ASPP Module via Strip Pooling
2.2.4. Loss Function Design
2.2.5. Multi-Temporal Fusion Strategy
3. Results
3.1. Experimental Details and Evaluation Metrics
3.1.1. Implementation Details and Training Strategy
3.1.2. Evaluation Metrics
- (1)
- Pixel-level Accuracy Metrics: These metrics assess the fundamental performance of the model in pixel-wise classification. In the context of forest road extraction, high Precision indicates high reliability of the predicted roads with fewer false positives (over-detection), while high Recall implies the model’s ability to identify as many ground-truth roads as possible with fewer false negatives (omissions) and better connectivity. The F1-score, as the harmonic mean of Precision and Recall, provides a balanced comprehensive evaluation. Intersection over Union (IoU) is the gold standard for measuring the overlap between the predicted mask and the ground truth, serving as one of the most rigorous and commonly used core metrics in semantic segmentation tasks. The calculation formulas for these metrics are as follows:where TP, FP, FN, and TN represent the number of pixels for true positives, false positives, false negatives, and true negatives, respectively.
- (2)
- Robustness Metrics for Class Imbalance: Since standard Accuracy can be misleading in class-imbalanced scenarios, Balanced Accuracy and the Matthews Correlation Coefficient (MCC) were introduced. The former treats the sparse foreground (roads) and the dominant background equally by calculating the average recall of both classes, thereby fairly reflecting the model’s recognition capability for rare road categories. The latter, which comprehensively accounts for TP, FP, FN, and TN, is considered one of the most robust evaluation metrics for imbalanced datasets. Its value ranges from −1 (complete misclassification) to +1 (perfect prediction), with 0 representing random guessing. The relevant formulas are as follows:
3.2. Ablation Study
3.3. Loss Function Modification
3.4. Experimental Results of the Multi-Temporal Fusion Strategy
4. Discussion
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Qian, C.S.; Wu, Z.Y.; Chen, C.D. The types of vegetation in China. Acta Geogr. Sin. 1956, 37–92. (In Chinese) [Google Scholar]
- Cui, R.K.; Qian, L.H.; Wang, Q.H. Research progress on fire fighting function evaluation of forest road networks. World For. Res. 2023, 36, 32–37. (In Chinese) [Google Scholar]
- Wang, D.D. Analysis and suggestions on the response work to mountain fire disasters in Chongqing in August 2022. Disaster Reduct. China 2023, 40–43. (In Chinese) [Google Scholar]
- Qin, X.H.; Zhao, X.D. Development of forest roads in Australia and its enlightenment. World For. Res. 2021, 34, 112–117. (In Chinese) [Google Scholar]
- Wang, L.Z.; Zhang, W.; Liu, C.M.; Wang, Y.P.; Bao, J.Q.; Liu, H. Impacts of the construction, planning and maintenance of forest roads on forest fire prevention work. For. Fire Prev. 2023, 41, 16–18. (In Chinese) [Google Scholar] [CrossRef]
- Bai, X.P.; Chen, S.Z.; He, Y.J.; Li, M.; Wang, Y.F. Current situation and enlightenment of foreign forest road development. World For. Res. 2015, 28, 85–91. (In Chinese) [Google Scholar]
- Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 39, 640–651. [Google Scholar]
- Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional networks for biomedical image segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention (MICCAI), Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
- Badrinarayanan, V.; Kendall, A.; Cipolla, R. SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [CrossRef]
- Ge, X.S.; Cao, W. An improved DeepLabV3+ network method for road extraction from high-resolution remote sensing images. Remote Sens. Inf. 2022, 37, 40–46. (In Chinese) [Google Scholar]
- Tao, C.; Qi, J.; Li, Y.; Wang, H.; Li, H. Spatial information inference net: Road extraction using road-specific contextual information. ISPRS J. Photogramm. Remote Sens. 2019, 158, 155–166. [Google Scholar] [CrossRef]
- Panboonyuen, T.; Vateekul, P.; Jitkajornwanich, K.; Lawawirojwong, S. An enhanced deep convolutional encoder-decoder network for road segmentation on aerial imagery. In Proceedings of the International Conference on Industrial, Engineering and Other Applications of Applied Intelligent Systems, Arras, France, 27–30 June 2017; pp. 191–201. [Google Scholar]
- Bajcsy, R.; Tavakoli, M. Computer recognition of roads from satellite pictures. IEEE Trans. Syst. Man Cybern. 1976, 6, 623–637. [Google Scholar] [CrossRef]
- Fan, C.F.; Li, Z.M. Edge detection method and its real-time implementation in road recognition system. Signal Process. 1998, 337–345+357. (In Chinese) [Google Scholar]
- Kass, M.; Witkin, A.; Terzopoulos, D. Snakes: Active contour models. Int. J. Comput. Vis. 1988, 1, 321–331. [Google Scholar] [CrossRef]
- Mena, J.B. State of the art on automatic road extraction for GIS update: A novel classification. Pattern Recognit. Lett. 2003, 24, 3037–3058. [Google Scholar] [CrossRef]
- Zhang, Q.; Couloigner, I. Comparing different localization approaches of the radon transform for road centerline extraction from classified satellite imagery. In Proceedings of the 18th International Conference on Pattern Recognition (ICPR’06), Hong Kong, China, 20–24 August 2006; Volume 2, pp. 138–141. [Google Scholar]
- Xu, Y.; Xie, Z.; Feng, Y.; Chen, Z. Road extraction from high-resolution remote sensing imagery using deep learning. Remote Sens. 2018, 10, 1461. [Google Scholar] [CrossRef]
- Yang, P.; Xiao, H.; Lin, C.; Xie, X. UGD-DLinkNet: An enhanced network for occluded road extraction using attention mechanisms and uncertainty estimation. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2025, 18, 24144–24161. [Google Scholar] [CrossRef]
- Li, X.; Li, Y.; Ai, J.; Shu, Z.; Xia, J.; Xia, Y. Semantic segmentation of UAV remote sensing images based on edge feature fusing and multi-level upsampling integrated with Deeplabv3. PLoS ONE 2023, 18, e0279097. [Google Scholar] [CrossRef]
- Winiwarter, L.; Coops, N.C.; Bastyr, A.; Roussel, J.-R.; Zhao, D.Q.R.; Lamb, C.T.; Ford, A.T. Extraction of Forest Road Information from CubeSat Imagery Using Convolutional Neural Networks. Remote Sens. 2024, 16, 1083. [Google Scholar] [CrossRef]
- Kleinschroth, F.; Laporte, N.; Laurance, W.F.; Goetz, S.J.; Ghazoul, J. Road expansion and persistence in forests of the Congo Basin. Nat. Sustain. 2019, 2, 628–634. [Google Scholar] [CrossRef]
- Chinanews Network. The Fire Situation Tends to Be Stable During the Continuous Fighting of Kunming Jinning Forest Fire. Available online: https://www.yn.chinanews.com.cn/news/2024/0415/76251.html (accessed on 17 November 2025). (In Chinese)
- Zhu, H. Study on vegetation diversity in Yunnan Province. J. Southwest For. Univ. (Nat. Sci.) 2022, 42, 1–12. (In Chinese) [Google Scholar]
- Yang, J.; Dai, J.H.; Yao, H.R.; Tao, Z.X.; Zhu, M.Y. Changes in vegetation distribution and vegetation activity in the Hengduan Mountains from 1992 to 2020. Acta Geogr. Sin. 2022, 77, 16. (In Chinese) [Google Scholar]
- Xiong, J.N.; Peng, C.; Cheng, W.M.; Li, W.; Liu, Z.Q.; Fan, C.K.; Sun, H.Z. Analysis of vegetation coverage change in Yunnan Province based on MODIS-NDVI. J. Geo-Inf. Sci. 2018, 20, 1830–1840. (In Chinese) [Google Scholar]
- Wu, Z.Y.; Zhu, Y.C. Vegetation in Yunnan; Science Press: Beijing, China, 1987. (In Chinese) [Google Scholar]
- Zeng, J.M. A study on the geographical distribution of Pinus and Cunninghamia species in Yunnan. J. Southwest For. Univ. (Nat. Sci.) 2021, 41, 1–12. (In Chinese) [Google Scholar]
- Zeng, J.M. A study on the classification system and geographical distribution of natural forests in Yunnan. J. Southwest For. Univ. (Nat. Sci.) 2018, 38, 1–18+231. (In Chinese) [Google Scholar]
- Zhu, H. A study on the vegetation geography of evergreen broad-leaved forests in Yunnan. Chin. J. Plant Ecol. 2021, 45, 224–241. (In Chinese) [Google Scholar] [CrossRef]
- Li, D.R. China’s High-Resolution Earth Observation System (CHEOS): Advances and Perspectives. ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci. 2022, V-3-2022, 583–590. [Google Scholar] [CrossRef]
- Maurer, T. How to Pan-Sharpen Images Using the Gram-Schmidt Pan-Sharpen Method—A Recipe. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2013, XL-1/W1, 239–244. [Google Scholar] [CrossRef]
- Demir, I.; Koperski, K.; Lindenbaum, D.; Pang, G.; Huang, J.; Basu, S.; Hughes, F.; Tuia, D.; Raskar, R. DeepGlobe 2018: A challenge to parse the earth through satellite images. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Salt Lake City, UT, USA, 18–22 June 2018; pp. 172–17209. [Google Scholar]
- Chen, L.-C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. Encoder-decoder with atrous separable convolution for semantic image segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 801–818. [Google Scholar]
- Abdollahi, A.; Pradhan, B.; Shukla, N.; Chakraborty, S.; Alamri, A. Deep Learning Approaches Applied to Remote Sensing Datasets for Road Extraction: A State-Of-The-Art Review. Remote Sens. 2020, 12, 1444. [Google Scholar] [CrossRef]
- Hoeser, T.; Kuenzer, C. Object Detection and Image Segmentation with Deep Learning on Earth Observation Data: A Review-Part I: Evolution and Recent Trends. Remote Sens. 2020, 12, 1667. [Google Scholar] [CrossRef]
- Wang, J.; Sun, K.; Cheng, T.; Jiang, B.; Deng, C.; Zhao, Y.; Liu, D.; Mu, Y.; Tan, M.; Wang, X.; et al. Deep High-Resolution Representation Learning for Visual Recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 3349–3364. [Google Scholar] [CrossRef]
- Jocher, G.; Chaurasia, A.; Qiu, J. Ultralytics YOLO. 2023. Available online: https://github.com/ultralytics/ultralytics (accessed on 13 March 2026).
- Kirillov, A.; Mintun, E.; Ravi, N.; Mao, H.; Rolland, C.; Gustafson, L.; Xiao, T.; Whitehead, S.; Berg, A.C.; Lo, W.-Y.; et al. Segment Anything. In Proceedings of the 2023 IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 2–6 October 2023; pp. 3992–4003. [Google Scholar]
- Chen, K.; Liu, C.; Chen, H.; Zhang, H.; Li, W.; Zou, Z.; Shi, Z. RSPrompter: Learning to Prompt for Remote Sensing Instance Segmentation Based on Visual Foundation Model. IEEE Trans. Geosci. Remote Sens. 2024, 62, 1–17. [Google Scholar] [CrossRef]
- Bai, H.; Ren, C.; Huang, Z.; Gu, Y. A dynamic attention mechanism for road extraction from high-resolution remote sensing imagery using feature fusion. Sci. Rep. 2025, 15, 17556. [Google Scholar] [CrossRef]
- Hou, Q.; Zhang, L.; Cheng, M.-M.; Feng, J. Strip pooling: Rethinking spatial pooling for scene parsing. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 14–19 June 2020; pp. 4002–4011. [Google Scholar]
- Lin, T.-Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
- Suzuki, S. Topological structural analysis of digitized binary images by border following. Comput. Vis. Graph. Image Process. 1985, 30, 32–46. [Google Scholar] [CrossRef]
- Loshchilov, I.; Hutter, F. Sgdr: Stochastic gradient descent with warm restarts. arXiv 2016, arXiv:1608.03983. [Google Scholar]
- Yosinski, J.; Clune, J.; Bengio, Y.; Lipson, H. How transferable are features in deep neural networks? Adv. Neural Inf. Process. Syst. 2014, 27, 3320–3328. [Google Scholar]
- Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]










| Dataset Name | Number of Image Pairs | Spatial Resolution | Image Size | Training/Validation/Test Set Split | Usage Description |
|---|---|---|---|---|---|
| DeepGlobe Road Extraction Dataset | 8570 | 0.5 m | 1024 × 1024 | 6226 images (training set), 1243 images (validation set), 1101 images (test set) | Model pre-training, preliminary feature extraction |
| Multi-temporal Forest Area Remote Sensing Dataset | 1456 | 0.8 m | 1024 × 1024 | 1165 images (training set), 146 images (validation set), 145 images (test set) | Model fine-tuning, forest road training and evaluation |
| Model | Core Architecture | Spatial Detail Processing Strategy | Suitability for Elongated/Weak Targets |
|---|---|---|---|
| FCN | Encoder–Decoder | Lost first, then roughly recovered | Low |
| U-Net | Symmetric Encoder–Decoder | Sacrificed first, then compensated | Medium |
| SegNet | Encoder–Decoder | Record positions first, then recovered | Medium–Low |
| DeepLabV3+ | Encoder–Decoder + Atrous Convolution | Maintained at all times | High |
| Method | Precision (%) | Recall (%) | F1 Score (%) | IoU (%) | MCC (%) |
|---|---|---|---|---|---|
| FCN | 37.7 | 58.9 | 45.9 | 29.8 | 39.6 |
| SegNet | 38.1 | 59.3 | 46.4 | 30.2 | 40.5 |
| U-Net | 40.2 | 60.9 | 48.4 | 31.8 | 43.1 |
| DeepLabV3+ (Baseline) | 40.6 | 61.1 | 48.8 | 32.2 | 43.5 |
| +MCSA | 46.8 | 62.9 | 53.7 | 36.7 | 48.2 |
| +Improved ASPP | 45.3 | 62.4 | 52.5 | 35.6 | 47.1 |
| Combined Application | 51.6 | 64.1 | 57.1 | 39.9 | 51.5 |
| Loss Function Configuration | Weight (α) | Weight (β) | Precision (%) | Recall (%) | F1 Score (%) | IoU (%) | MCC (%) |
|---|---|---|---|---|---|---|---|
| Baseline (CE Loss) | - | - | 40.6 | 61.1 | 48.8 | 32.2 | 43.5 |
| Configuration 1 | 1.0 | 0.0 | 47.8 | 58.2 | 52.4 | 35.5 | 46.9 |
| Configuration 2 | 0.7 | 0.3 | 53.1 | 63.4 | 57.8 | 40.6 | 52.3 |
| Configuration 3 | 0.5 | 0.5 | 49.2 | 61.8 | 54.8 | 37.7 | 49.1 |
| Configuration 4 | 0.3 | 0.7 | 50.3 | 62.5 | 55.7 | 38.6 | 50.2 |
| Configuration 5 | 0.0 | 1.0 | 44.1 | 60.3 | 50.9 | 34.2 | 45.6 |
| Method | Precision (%) | Recall (%) | F1 Score (%) | IoU (%) | MCC (%) |
|---|---|---|---|---|---|
| Loss Function Improvement | 53.1 | 63.4 | 57.8 | 40.6 | 52.3 |
| Model Improvement | 51.6 | 64.1 | 57.1 | 39.9 | 51.5 |
| Combined Effect | 53.8 | 64.9 | 58.8 | 41.6 | 53.2 |
| Evaluation Metric | Precision (%) | Recall (%) | F1 Score (%) | IoU (%) | MCC (%) |
|---|---|---|---|---|---|
| Summer Dataset | 55.3 | 60.2 | 57.6 | 41.2 | 53.1 |
| Winter Dataset | 49.7 | 67.8 | 57.3 | 39.5 | 51.8 |
| After Temporal Fusion | 54.1 | 65.5 | 59.3 | 41.8 | 53.5 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2026 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license.
Share and Cite
Gao, Z.; Li, Z.; Yao, W.; Zhang, T.; Qiu, S.; Liu, Z. Forest Road Extraction via Optimized DeepLabv3+ and Multi-Temporal Remote Sensing for Wildfire Emergency Response. Appl. Sci. 2026, 16, 3228. https://doi.org/10.3390/app16073228
Gao Z, Li Z, Yao W, Zhang T, Qiu S, Liu Z. Forest Road Extraction via Optimized DeepLabv3+ and Multi-Temporal Remote Sensing for Wildfire Emergency Response. Applied Sciences. 2026; 16(7):3228. https://doi.org/10.3390/app16073228
Chicago/Turabian StyleGao, Zhuoran, Ziyang Li, Weiyuan Yao, Tingtao Zhang, Shi Qiu, and Zhaoyan Liu. 2026. "Forest Road Extraction via Optimized DeepLabv3+ and Multi-Temporal Remote Sensing for Wildfire Emergency Response" Applied Sciences 16, no. 7: 3228. https://doi.org/10.3390/app16073228
APA StyleGao, Z., Li, Z., Yao, W., Zhang, T., Qiu, S., & Liu, Z. (2026). Forest Road Extraction via Optimized DeepLabv3+ and Multi-Temporal Remote Sensing for Wildfire Emergency Response. Applied Sciences, 16(7), 3228. https://doi.org/10.3390/app16073228

