Deep Learning-Based Collapsed Building Mapping from Post-Earthquake Aerial Imagery
Abstract
1. Introduction
- Evaluating the performance of such methods under challenging scenarios, such as poor lighting and complex scene conditions;
- Mitigating the uncertainty introduced by the limited and morphologically diverse samples of collapsed buildings, thereby enhancing the model’s generalization performance and transferability to new disaster events;
- Validating the reliability of inferred results so that they can be confidently used as the basis for official damage assessments.
2. Datasets and Methods
2.1. Outline of the 2024 Noto Peninsula Earthquake
2.2. Outline of the 2016 Kumamoto Earthquakes
2.3. Selected Aerial Images
2.4. Dataset Construction
- Visible roof damage exposing internal structural fragments (see Figure 4a);
- Significant roof structural failure (see Figure 4b,d);
- Loss of roof texture continuity, such as fragmented roofing (see Figure 4c);
- Complete loss of defining structural characteristics, with the building turning into debris or ruins.
2.5. Proposed Network
2.5.1. Network Architecture
- Encoder: In the encoder, we employ a pretrained transformer-based backbone, Pyramid Vision Transformer-V2-B2 [20] (PVT-V2-B2), to extract hierarchical features across four stages () with channel dimensions of 64, 128, 320, and 512, respectively, ranging from low-level to high-level, from the input image;
- Decoder: In the decoder, the highest-level feature is first processed by two consecutive 3 × 3 convolutions to generate a coarse segmentation result . Subsequently, this lower-resolution result is transformed into an uncertainty map by the Uncertainty Guide Fusion Module (UGFM). The uncertainty map guides the feature fusion between low-level and high-level features, producing fused features () with channel dimensions of 64, 128, and 320, respectively. These fused features are then fed into two consecutive 3 × 3 convolutions, which are used to adjust channel dimensions and enhance the nonlinear representational capacity of the decoder to generate the segmentation result for the next stage.
2.5.2. Outline of the Pyramid Vision Transformer v2
2.5.3. Proposed Uncertainty Guide Fusion Module
- Significant variations in appearance (e.g., scale, shape, texture, damage pattern);
- Ambiguous boundaries with surrounding objects (e.g., shadows or vegetation);
- Inconsistent distribution, appearing either sparsely or densely across the region.
2.5.4. Loss Function
2.6. Experimental Configuration and Training Methods
2.7. Methods Comparison
3. Results
3.1. In-Domain Data Evaluation
Quantitative and Visual Comparison
3.2. Out-of-Domain Data Evaluation
Quantitative and Visual Comparison
3.3. Ablation Study
3.4. Quantitative Analysis with Field Survey Data
3.4.1. On-Site Field Survey from Mashiki
3.4.2. Evaluation of the Reliability of Predicted Collapsed Areas
- Some buildings categorized as Levels 1–4 in the field survey exhibited sufficiently visible external damage in aerial images, leading the model to classify them as collapsed. Conversely, some Level 5 buildings appeared visually intact from the aerial perspective, causing the model to classify them as non-collapsed;
- Some buildings exhibited partial damage. While the model successfully extracted the collapsed portions, these buildings were missing in the final output because their reference coordinates fell within the non-collapsed area of the building.
4. Discussion
- Application to Buildings with Different Structural Materials: Although this study validated the model’s transferability by applying it to post-earthquake imagery from a different event (i.e., the 2016 Kumamoto earthquake), which differs from the training dataset (Noto Peninsula earthquake) in terms of lighting conditions, location, and season, both datasets predominantly feature wooden buildings. Therefore, the model’s detection targets in this study were largely limited to wooden structures. Whether this method can be effectively transferred to buildings constructed with other materials—which may exhibit different collapse patterns—remains to be verified;
- Stage-Specific Uncertainty Weighting in UGFM: In this study, the amplification range of uncertainty within UGFM was kept identical across all decoding stages, implicitly assuming that uncertainty at different depths is equally important. However, in practice, uncertainty at deeper stages may play a more critical role, or conversely, earlier stages may require stronger guidance. Future work could investigate progressively increasing or decreasing stage-specific uncertainty amplification factors, which may further refine the feature fusion process and improve overall performance;
- In real-world post-earthquake scenarios, there is often no time to manually annotate new datasets, making it infeasible to create labeled data for each event. As a result, models must rely on training from previously available data and be directly applied to unseen scenarios. However, due to the scarcity of post-earthquake RS data, such models are prone to overfitting to the training domain, which limits their generalizability. Although this study introduced the UGFM to enhance model transferability, it is fundamentally still reliant on previously labeled data. The semi-supervised learning techniques could be incorporated, i.e., where the model automatically generates pseudo-labels during inference, which may furthermore improve the model’s adaptability to new disasters without requiring manual annotations.
5. Conclusions
Author Contributions
Funding
Data Availability Statement
Acknowledgments
Conflicts of Interest
Abbreviations
EDC | Earthquake Damage Certification |
DVC | Disaster Victim Certificate |
UFGM | Uncertainty-Guided Fusion Module |
PVT | Pyramid Vision Transformer |
MMS | Moment Magnitude Scale |
CNN | Convolutional Neural Network |
ViT | Vision Transformer |
SRA | Spatial Reduction Attention |
CE | Cross-Entropy |
UM | Uncertainty Map |
RS | Remote Sensing |
GSI | Geospatial Information Authority of Japan |
References
- Japan Meteorological Agency. Major Damaging Earthquakes near Japan (Since 1996). Available online: https://www.data.jma.go.jp/eqev/data/higai/higai1996-new.html (accessed on 20 October 2024). (In Japanese)
- Cabinet Office of Japan Disaster Management in Japan. Guideline for the Application of the Standards for Recognition of Damage Caused by Disasters to Houses. 2024. Available online: https://www.bousai.go.jp/taisaku/unyou.html (accessed on 24 October 2024). (In Japanese)
- The Japan Broadcasting Corporation. Issuance of Certificate Disaster Victim Certificate in Ishikawa. Available online: https://www3.nhk.or.jp/news/html/20240202/k10014345371000.html (accessed on 24 October 2024). (In Japanese).
- Matsuoka, M.; Yamazaki, F. Interferometric Characterization of Areas Damaged by the 1995 Kobe Earthquake Using Satellite SAR Images. In Proceedings of the 12th World Conference on Earthquake Engineering, Auckland, New Zealand, 30 January–4 February 2000; Volume 2. [Google Scholar]
- Naito, S.; Hao, K.X.; Senna, S.; Saeki, T.; Nakamura, H.; Fujiwara, H.; Azuma, T. Investigation of Damages in Immediate Vicinity of Co-Seismic Faults during the 2016 Kumamoto Earthquake. J. Disaster Res. 2017, 12, 899–915. [Google Scholar] [CrossRef]
- Vu, T.T.; Matsuoka, M.; Yamazaki, F. Detection and Animation of Damage Using Very High-Resolution Satellite Data Following the 2003 Bam, Iran Earthquake. Earthq. Spectra 2005, 21 (Suppl. 1), 319–327. [Google Scholar] [CrossRef]
- Miura, H.; Aridome, T.; Matsuoka, M. Deep Learning-Based Identification of Collapsed, Non-Collapsed and Blue Tarp-Covered Buildings from Post-Disaster Aerial Images. Remote Sens. 2020, 12, 1924. [Google Scholar] [CrossRef]
- Wiguna, S.; Adriano, B.; Mas, E.; Koshimura, S. Evaluation of Deep Learning Models for Building Damage Mapping in Emergency Response Settings. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2024, 17, 5651–5667. [Google Scholar] [CrossRef]
- Adriano, B.; Yokoya, N.; Xia, J.; Miura, H.; Liu, W.; Matsuoka, M.; Koshimura, S. Learning from Multimodal and Multitemporal Earth Observation Data for Building Damage Mapping. ISPRS J. Photogramm. Remote Sens. 2021, 175, 132–143. [Google Scholar] [CrossRef]
- Kuo, W.-N.; Lin, S.-Y. Multimodal Models for Assessing Earthquake-Induced Building Damage Using Metadata and Satellite Imagery. J. Build. Eng. 2025, 111, 113467. [Google Scholar] [CrossRef]
- Han, D.; Yang, G.; Xie, R.; Lu, W.; Huang, M.; Liu, S. A Multilevel Damage Assessment Framework for Mixed-Hazard Buildings with Global Spatial Feature Guidance Module and Change Feature Attention in VHR Remote Sensing Images. IEEE Geosci. Remote Sens. Lett. 2024, 21, 1–5. [Google Scholar] [CrossRef]
- Zhan, Y.; Liu, W.; Maruyama, Y. Damaged Building Extraction Using Modified Mask R-CNN Model Using Post-Event Aerial Images of the 2016 Kumamoto Earthquake. Remote Sens. 2022, 14, 1002. [Google Scholar] [CrossRef]
- Yu, K.; Wang, S.; Wang, Y.; Gu, Z.; Wang, Y. DBA-RTMDet: A High-Precision and Real-Time Instance Segmentation Method for Identification of Damaged Buildings in Postearthquake UAV Imagery. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2025, 18, 19577–19593. [Google Scholar] [CrossRef]
- Naito, S.; Tomozawa, H.; Mori, Y.; Monma, N.; Nakamura, H.; Fujiwara, H. Development of the Deep Learning-Based Damage Detection Model for Buildings Utilizing Aerial Photographs of Multiple Earthquakes. J. Jpn. Assoc. Earthq. Eng. 2021, 21, 3_72–3_118. [Google Scholar] [CrossRef]
- Xie, Y.; Feng, D.; Chen, H.; Liu, Z.; Mao, W.; Zhu, J.; Hu, Y.; Baik, S.W. Damaged Building Detection from Post-Earthquake Remote Sensing Imagery Considering Heterogeneity Characteristics. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–17. [Google Scholar] [CrossRef]
- Liu, J.; Luo, Y.; Chen, S.; Wu, J.; Wang, Y. BDHE-Net: A Novel Building Damage Heterogeneity Enhancement Network for Accurate and Efficient Post-Earthquake Assessment Using Aerial and Remote Sensing Data. Appl. Sci. 2024, 14, 3964. [Google Scholar] [CrossRef]
- Cote, M.; Saeedi, P. Automatic Rooftop Extraction in Nadir Aerial Imagery of Suburban Regions Using Corners and Variational Level Set Evolution. IEEE Trans. Geosci. Remote Sens. 2012, 51, 313–328. [Google Scholar] [CrossRef]
- Awrangjeb, M.; Zhang, C.; Fraser, C.S. Improved Building Detection Using Texture Information. Int. Arch. Photogramm. Remote Sens. Spatial Inf. Sci. 2013, 38, 143–148. [Google Scholar] [CrossRef]
- Kendall, A.; Gal, Y. What Uncertainties Do We Need in Bayesian Deep Learning for Computer Vision? In Proceedings of the 31st International Conference on Neural Information Processing Systems (NIPS’17), Long Beach, CA, USA, 4–9 December 2017; Curran Associates Inc.: Red Hook, NY, USA, 2017; pp. 5580–5590. [Google Scholar] [CrossRef]
- Wang, W.; Xie, E.; Li, X.; Fan, D.P.; Song, K.; Liang, D.; Lu, T.; Luo, P.; Shao, L. Pyramid Vision Transformer: A Versatile Backbone for Dense Prediction without Convolutions. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 11–17. [Google Scholar] [CrossRef]
- The Report of the 2024 Noto Earthquake. Available online: https://www.jma.go.jp/jma/press/2401/01c/202401012130.html (accessed on 24 October 2024). (In Japanese)
- Quiet+. Seismic Intensity Distribution Map. Available online: https://app.quietplus.kke.co.jp/quakes?mode=quick (accessed on 24 October 2024). (In Japanese).
- Cabinet Office of Japan Disaster Management in Japan. Damage from the 2024 Noto Peninsula Earthquake. Available online: https://www.bousai.go.jp/updates/r60101notojishin/r60101notojishin/index.html (accessed on 24 October 2024). (In Japanese)
- Cabinet Office of Japan Disaster Management in Japan: The Report of 2016 Kumamoto Earthquakes. Available online: https://www.bousai.go.jp/updates/h280414jishin/index.html (accessed on 24 October 2024). (In Japanese)
- Geospatial Information Authority of Japan. Available online: https://www.gsi.go.jp/top.html (accessed on 24 October 2024).
- QGIS Development Team. Available online: https://qgis.org (accessed on 24 October 2024).
- Statistics Bureau of Japan. Type of House, Construction Method and Structure. Available online: https://www.stat.go.jp/data/jyutaku/2008/nihon/2_1.html (accessed on 24 October 2024).
- Okada, S.; Takai, N. Classifications of Structural Types and Damage Patterns of Buildings for Earthquake Field Investigation. J. Struct. Constr. Eng. 1999, 64, 65–72. [Google Scholar] [CrossRef]
- Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image Is Worth 16×16 Words: Transformers for Image Recognition at Scale. arXiv 2020, arXiv:2010.11929. [Google Scholar] [CrossRef]
- Li, J.; He, W.; Cao, W.; Zhang, L.; Zhang, H. UANet: An Uncertainty-Aware Network for Building Extraction from Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2024, 62, 1–13. [Google Scholar] [CrossRef]
- Tsubame. Supercomputer of Science Tokyo. Available online: https://www.titech.ac.jp/english/news/2024/069452 (accessed on 24 October 2024).
- Ronneberger, O.; Fischer, P.; Brox, T. UNet: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; Springer: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar] [CrossRef]
- Wang, J.; Sun, K.; Cheng, T.; Jiang, B.; Deng, C.; Zhao, Y.; Liu, D.; Mu, Y.; Tan, M.; Wang, X.; et al. Deep High-Resolution Representation Learning for Visual Recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 3349–3364. [Google Scholar] [CrossRef]
- Chen, L.C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder–Decoder with Atrous Separable Convolution for Semantic Image Segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 801–818. [Google Scholar] [CrossRef]
- Li, R.; Wang, L.; Zhang, C.; Duan, C.; Zheng, S. A2-FPN for Semantic Segmentation of Fine-Resolution Remotely Sensed Images. Int. J. Remote Sens. 2022, 43, 1131–1155. [Google Scholar] [CrossRef]
- Li, R.; Zheng, S.; Zhang, C.; Duan, C.; Wang, L.; Atkinson, P.M. ABCNet: Attentive Bilateral Contextual Network for Efficient Semantic Segmentation of Fine-Resolution Remotely Sensed Imagery. ISPRS J. Photogramm. Remote Sens. 2021, 181, 84–98. [Google Scholar] [CrossRef]
- Cho, S.; Xiu, H.; Matsuoka, M. Backscattering Characteristics of SAR Images in Damaged Buildings Due to the 2016 Kumamoto Earthquake. Remote Sens. 2023, 15, 2181. [Google Scholar] [CrossRef]
Area | Patch_Num (No_Augmentation) | Non_Collapsed_Area 1 (m2) | Collapsed_Area 1 (m2) | ||||
---|---|---|---|---|---|---|---|
Wajima Machinomachi Ukai | 2604 | 1,024,038 | 244,109 | ||||
Suzu | 228 | 143,663 | 15,226 | ||||
Mashiki | 56 | 43,254 | 33,502 | ||||
Augmentation Strategies | 1.Random_Horizontal_Flip 2. Random_GaussianBlur Random probability: 0.5 |
Training Setting | |
---|---|
Image Number of Training Dataset | 2604 (Training: 2354, Validation: 250) |
Data Augmentation | Random Horizontal Flipping, Gaussian blur |
Image Number of Test Dataset | 228 (In-Domain), 56 (Out-of-Domain) |
Framework | PyTorch |
GPU | NVIDIA H100 ((NVIDIA Corp., Santa Clara, CA, USA)) |
Batchsize | 32 |
Initial Learning Rate | |
Training Strategy | AdamW Optimizer and Poly Learning Rate Schedule |
Random Seed | 2333 |
Epochs | 200 |
Networks | Category | Parameters(M) | |||
---|---|---|---|---|---|
Non-Collapsed | Collapsed | ||||
Precision | Recall | Precision | Recall | ||
UNet | 0.898 | 0.874 | 0.633 | 0.742 | 147.81 |
HRNetv2 | 0.912 | 0.883 | 0.696 | 0.740 | 65.76 |
DeepLabv3+ | 0.875 | 0.841 | 0.607 | 0.675 | 42.40 |
A2FPN | 0.834 | 0.863 | 0.619 | 0.612 | 12.16 |
ABCNet | 0.764 | 0.762 | 0.615 | 0.240 | 13.85 |
Proposed | 0.864 | 0.922 | 0.676 | 0.787 | 29.15 |
Networks | Category | |||
---|---|---|---|---|
Non-Collapsed | Collapsed | |||
Precision | Recall | Precision | Recall | |
UNet | 0.638 | 0.772 | 0.647 | 0.435 |
HRNetv2 | 0.757 | 0.668 | 0.609 | 0.537 |
DeepLabv3+ | 0.568 | 0.751 | 0.612 | 0.261 |
A2FPN | 0.606 | 0.745 | 0.593 | 0.489 |
ABCNet | 0.308 | 0.110 | 0.926 | 0.000 |
Proposed | 0.714 | 0.861 | 0.774 | 0.660 |
Test Data | Networks | Category | |||
---|---|---|---|---|---|
Non-Collapsed | Collapsed | ||||
Precision | Recall | Precision | Recall | ||
Suzu | Base | 0.906 | 0.887 | 0.696 | 0.753 |
Base + UGFM | 0.864 | 0.922 | 0.676 | 0.787 | |
Mashiki | Base | 0.742 | 0.803 | 0.786 | 0.582 |
Base + UGFM | 0.714 | 0.861 | 0.774 | 0.660 |
Damage Class | Characteristics of the Damage | Corresponding Images | Number | |
---|---|---|---|---|
Major Damage | Level 5 | Collapsed | 399 | |
Level 4 | Large interlayer deformation (not collapsed) | 132 | ||
Level 3 | Large distortion or large inclination | 192 | ||
Level 2 | Damage to roof and walls (including the foundations) | 392 | ||
Level 1 | Damage to walls (including the foundations) | 354 | ||
Under Major Damage | Level 0 | Under 50% of economic damage to the building | 4568 |
Model Prediction | Building Point Location Problem | |||
---|---|---|---|---|
Collapsed | Non-Collapsed | |||
Field Survey | Level 5 | |||
Level 4 | ||||
Level 1–3 |
Confusion Matrix | Field Survey (Case1) | Total | Precision | ||
---|---|---|---|---|---|
Level 0–4 | Level 5 | ||||
Model Prediction | Non-collapsed | 5248 | 126 | 5374 | 0.976 |
Collapsed | 153 | 255 | 408 | 0.625 | |
Total | 5401 | 381 | |||
Recall | 0.972 | 0.669 |
Confusion Matrix | Field Survey (Case2) | Total | Precision | ||
---|---|---|---|---|---|
Level 0 | Level 1–5 | ||||
Model Prediction | Non-collapsed | 4321 | 1053 | 5374 | 0.804 |
Collapsed | 52 | 356 | 408 | 0.872 | |
Total | 4374 | 1409 | |||
Recall | 0.988 | 0.253 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Lyu, H.; Oshio, H.; Matsuoka, M. Deep Learning-Based Collapsed Building Mapping from Post-Earthquake Aerial Imagery. Remote Sens. 2025, 17, 3116. https://doi.org/10.3390/rs17173116
Lyu H, Oshio H, Matsuoka M. Deep Learning-Based Collapsed Building Mapping from Post-Earthquake Aerial Imagery. Remote Sensing. 2025; 17(17):3116. https://doi.org/10.3390/rs17173116
Chicago/Turabian StyleLyu, Hongrui, Haruki Oshio, and Masashi Matsuoka. 2025. "Deep Learning-Based Collapsed Building Mapping from Post-Earthquake Aerial Imagery" Remote Sensing 17, no. 17: 3116. https://doi.org/10.3390/rs17173116
APA StyleLyu, H., Oshio, H., & Matsuoka, M. (2025). Deep Learning-Based Collapsed Building Mapping from Post-Earthquake Aerial Imagery. Remote Sensing, 17(17), 3116. https://doi.org/10.3390/rs17173116