Next Article in Journal
Geometry-Dependent Efficiency of Dean-Flow Affected Lateral Particle Focusing and Separation in Periodically Inhomogeneous Microfluidic Channels
Next Article in Special Issue
Part-Based Obstacle Detection Using a Multiple Output Neural Network
Previous Article in Journal
Printing Technologies as an Emerging Approach in Gas Sensors: Survey of Literature
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Rapid Post-Earthquake Structural Damage Assessment Using Convolutional Neural Networks and Transfer Learning

1
School of Architecture, Civil, Energy, and Environment Engineering, Kyungpook National University, 80 Daehakro, Bukgu, Daegu 41566, Korea
2
Department of Civil Engineering, Kyungpook National University, 80 Daehakro, Bukgu, Daegu 41566, Korea
3
Department of Robot and Smart System Engineering, Kyungpook National University, 80 Daehakro, Bukgu, Daegu 41566, Korea
*
Authors to whom correspondence should be addressed.
Sensors 2022, 22(9), 3471; https://doi.org/10.3390/s22093471
Submission received: 29 March 2022 / Revised: 28 April 2022 / Accepted: 28 April 2022 / Published: 3 May 2022
(This article belongs to the Special Issue Machine Learning in Robust Object Detection and Tracking)

Abstract

:
The adoption of artificial intelligence in post-earthquake inspections and reconnaissance has received considerable attention in recent years, owing to its exponential increase in computation capabilities and inherent potential in addressing disadvantages associated with manual inspections. Herein, we present the effectiveness of automated deep learning in enhancing the assessment of damage caused by the 2017 Pohang earthquake. Six classical pre-trained convolutional neural network (CNN) models are implemented through transfer learning (TL) on a small dataset, comprising 1780 manually labeled images of structural damage. Feature extraction and fine-tuning TL methods are trained on the image datasets. The performances of various CNN models are compared on a testing image dataset. Results confirm that the MobileNet fine-tuned model offers the best performance. Therefore, the model is further developed as a web-based application for classifying earthquake damage. The severity of damage is quantified by assigning damage assessment values, derived using the CNN model and gradient-weighted class activation mapping. The web-based application can effectively and automatically classify structural damage resulting from earthquakes, rendering it suitable for decision making, such as in resource allocation, policy development, and emergency response.

1. Introduction

Classification of the magnitude of damage to buildings and infrastructure attributed to seismic events is essential for enhancing post-earthquake reconnaissance and ensuring safe and effective recovery efforts. Conventionally, property damage attributed to earthquakes is documented manually using labor-intensive methods [1,2,3,4,5]. Manual damage inspections may be time consuming and involve arbitrary judgment by a novice inspector who may not be adequately trained. These disadvantages can be addressed by performing fully automated inspections using computer-vision technologies [6]. The automated deep learning (DL) method may be critical for enabling the rapid real-time detection and classification of structural damage (SD) attributed to earthquakes.
DL algorithms for image classification may be applicable for assessing SDs [6,7,8,9,10,11]. Gao and Mosalam [6] created an image database known as “Structural ImageNet,” which implements a visual geometry group (VGG) convolutional neural network (CNN) model through transfer learning (TL) to classify SD caused by earthquakes. They curated the Pacific Earthquake Engineering Research (PEER) Hub ImageNet [12] dataset, which serves as a benchmark for similar computer-vision-based classification and detection tasks [13]. Nahata et al. [7] employed the VGG16 TL model to classify post-earthquake building damage into four categories. After training the model with more than 1,200 images, they obtained training and validation accuracies of 97.85% and 89.38%, respectively. In addition, DL methods have been exploited for damage-detection tasks, in which bounding boxes are used to identify and localize SD [14,15].
Decision makers can allocate the appropriate resources to retrofit, repair, and recover facilities by locating and quantifying SDs. A numerical scale that quantifies the magnitude of SD to facilitate such efforts has been established. Li et al. [16] identified a mismatch in the damage detected using conventional approaches. They proposed a novel approach to quantify the severity of SD, using a smooth image heat map based on gradient-weighted class activation mapping (Grad-CAM). In fact, this approach has been employed in several applications, such as post-disaster damage assessments [10] and steel frame damage investigations [17], and demonstrated performances superior to or comparable with other state-of-the-art methods, while requiring low computation time. We employed the approach to quantify and locate SD caused by the 2017 Pohang earthquake using CNN based on TL strategies. Through TL, CNN models can learn complex patterns from data without needing a large amount of training data. Additionally, they can generalize well to new datasets, which is important when dealing with SD that may vary in appearance from one instance to another. Therefore, the performance of Feature Extraction (FE) and Fine Tuning (FT) TL methods on SD image datasets were compared, in order to explore the possibility of applying the knowledge from a pre-trained model (source domain) to another (target domain), by tuning some of the model parameters. Finally, the optimal CNN model used to implement the approach was deployed on an interactive webpage that automatically classifies SD caused by earthquakes. Invisible damage, which is beyond the scope of this study, is typically examined via anomaly detection in structural members using specialized sensors and signal-processing techniques. However, in the abovementioned study, damage was considered visible to either the human eye or computer vision. This novel approach can facilitate rapid responses following an earthquake. Researchers have successfully identified SD characteristics using classification [6,7], bounding box detection [8], and segmentation techniques [9]. However, most of those methods do not involve a tool with a post-disaster assessment framework that is accessible to the structural engineering community.
Moreover, only a few studies have considered the deployment of post-earthquake damage classification, rendering it less useful for industrial applications and field validation. By contrast, both object localization with Grad-CAM and model deployment for practical applications are considered in the current study. This novel approach is relevant to researchers and practitioners as it fills the research gap by providing an interactive tool for SD assessment.
The remainder of this paper is organized as follows: Section 2 presents a brief overview of related studies. Section 3 describes the data acquisition process and methodology, and Section 4 discusses the results of the CNN model training, damage localization, and quantification. Section 5 presents an interactive webpage for damage classification, and Section 6 presents the conclusions and future research directions. The workflow of the research method is shown in Figure 1.

2. Related Studies

This section provides a brief overview of existing studies related to the current study. A comprehensive review of the applications of computer-vision-based civil infrastructure inspection has recently been presented [18]. Pan and Yang [8] implemented an object detection algorithm to quantify damage to structural elements and the associated repair costs. Their proposed algorithm achieved average precisions of 98.2% and 84.5% on the training and testing image datasets, respectively. For automated post-earthquake inspection, Hoskere et al. [9] proposed a multiscale deep CNN, incorporating ResNet23 and VGG19 as damage classifiers and damage segmenters, which achieved accuracies of 88.8% and 71.4%, respectively. Liang [11] investigated an image-based approach for inspecting bridges by considering system, component, and local damage level detection. The proposed DL network comprises a pre-trained VGG-16 CNN for system-level failure classification, a faster region-based CNN for component-level bridge column detection, and a fully convolutional network for damage segmentation. Bayesian optimization enhanced the model performance and afforded an accuracy exceeding 90% for all the three-level tasks considered.
Some disadvantages in the existing multiclass damage assessment approach mentioned earlier include dataset class imbalance, which results in overfitting, lack of scalability and flexibility of the CNN architecture for solving various challenges, noisy training data, and a complex CNN architecture [19]. Therefore, recent applications of CNN-based models for SD assessments focus more on quality data preparation, the algorithmic optimization of the CNN model architecture, and damage quantification. Techniques typically adopted for quality image data preparation include image enhancement approaches, such as gray-level thresholding, histogram equalization, and adaptive histogram equalization [20]. Moreover, the algorithmic optimization of hyperparameters enhances the accuracy of CNN-based models and reduces the computational power used for execution [21]. Recently, Kim et al. [22] developed an optimized LeNet (OLeNet) model by tuning a shallow LeNet-5 CNN architecture for concrete surface crack detection. Consequently, OLeNet achieved an optimum validation accuracy of 99.8% at 19 epochs within 220 s of model training. Meanwhile, pre-trained deep CNN architectures, including ResNet, VGG16, and Inception, required at least 45 epochs to achieve the same validation accuracy within 524 s.

3. Methodology

3.1. Data Acquisition, Division, and Preprocessing

A total of 2750 images were acquired from field investigations [1,2,3,4,5] for different earthquakes. This study focuses on the Pohang earthquake. However, data obtained from other earthquakes were used to build a robust model to increase generalizability. A summary of the image datasets is presented in Table 1. Light damage indicates hairline cracks in structural elements, whereas moderate damage indicates wider cracks and spalling in concrete. By contrast, severe damage represents elemental collapse or structural failure [23].
The methodology involves a supervised learning image classification problem. Therefore, the labeled image dataset was split into two to train and evaluate the model’s performance after each epoch. The ratio of the training and validation sets was empirically set at 4:1. In addition, the validation datasets were used to test the training performance of the models after each epoch. A total of 1780 images were selected from the database, of which 1600 were used for training and validation (Table 2). To address the data imbalance during model training, each damage class was penalized by assigning class weights of 1.0, 1.5, 1.5 and 2.4 to the severe, light, moderate, and no damage classes, respectively. A total of 180 images were obtained exclusively from the damage database of the Pohang earthquake and these were used to evaluate the generalizability of the trained model. Figure 2 shows a sample of 1600 images selected to train the CNN model.

3.2. TL Using Pre-Trained CNN Models

Six pre-trained classical CNN setups were implemented via TL. TL is an efficient approach used for training a small dataset, whereby a neural network pre-trained on a large dataset in the source domain is applied to the target domain. The underlying hypothesis of TL is that common features learned from a sufficiently large dataset are transferred to different datasets [24]. For practical applications, two strategies are used while conducting TL in deep CNNs: feature extraction (FE) and FT. We used FE and FT TL methods to train the models on the datasets. In the FE method, the fully connected layers are removed from a network that has been pre-trained on the ImageNet dataset, while maintaining the convolutional base as a feature extractor. The pre-trained network serves as an arbitrary feature extractor that performs convolutional operations once on the input image during forward propagation, stops at the pre-specified layer, and uses the outputs of that layer as bottleneck features. In summary, the pre-trained CNN models serve as the backbone for FE, in which all the parameters in the convolution layers are frozen, whereas the fully connected layers are updated during backpropagation [25].
However, the FT method requires the unfreezing and retraining of the pre-trained convolutional base through backpropagation. During retraining, the convolutional layers learn mid- to high-order features, such as edges, which are more specific to the dataset in the target domain than the more generic features from the dataset in the source domain. Because the parameters in the last convolutional layer are unfrozen and updated during backpropagation, FT typically requires more computational time than FE. The procedures for TL using FE and FT are shown in Figure 3. Similar studies using the TL approach for SD assessment include real-time crack detection using unmanned aerial vehicles [24], building defect detection [26], concrete bridge surface damage detection [27], and crack segmentation on masonry surfaces [28].
Well-established versions of VGGNet are VGG16 (16 layers) and VGG19 (19 layers), which contain 138 and 144 million parameters, respectively. The VGGNet architecture comprises five convolutional blocks, with each block containing two or more convolutional layers and a max-pooling layer. ReLU activation functions are provided in all hidden layers, and the output comprises three fully connected layers with softmax functions. Applications of pre-trained VGGNets through TL include crack detection [29], bolt-loosening detection [30], steel damage condition assessment [31], building defect detection [26], and post-earthquake SD assessment [6,7].
The inception network is engineered significantly for performance improvement and has a relatively lower error rate compared with VGGNet. Different versions of the inception modules that have evolved include V1, V2, V3, and V4. Within the inception block, parallel filter operations are applied to the input from the previous layer, followed by depth-wise concatenation of the filter outputs. Previous applications of inception networks in image classification include crack detection [32] and tunnel rock structure identification [33].
Xception is an extension of inceptionV3, where the convolutional layers are replaced with depth-wise separable convolutions. It comprises blocks of convolution and separable convolution followed by batch normalization and max-pooling layers. Use cases of Xception include aerial visual geolocalization [34] and construction site safety [35].
ResNet is a deep neural network that is based on residual learning. ResNet50 comprises 50 main layers and 177 layers, whereas ResNet101 comprises 101 main layers and a total of 347 layers. ResNet has been successfully applied to bridge component extraction [36] and road crack detection [37].
MobileNet comprises a class of efficient models based on depth-wise separable convolutions, which are widely used for mobile applications. The MobileNet block typically comprises batch normalization, 3 × 3 depth-wise convolution, 1 × 1 convolution layers, and ReLU activation. Because MobileNets have fewer parameters and a higher classification accuracy, they are typically adopted to build lightweight deep neural networks. MobileNet is used for road damage detection [38] and post-hurricane aerial damage assessment [39]. The pseudocode of the algorithm for the CNN model is presented in Table 3.
Each model was trained with an SGD optimizer on a high-performance computer with an Intel (R) Core i7-8700 CPU @ 3.20 GHz, 32 GB RAM, and an NVIDIA RTX Quadro 5000 GPU in a Keras/TensorFlow environment. A preliminary experiment was performed on the dataset based on a learning rate of 0.0001, a momentum set of 0.9, and a batch size of 32 images. The number of training epochs was set to 60 for all the experiments, and the images were resized to 224 × 224 × 3 before training. The validation set was used to tune the hyperparameters and optimize the weights of the CNN model. During FT, only the final convolutional block of the pre-trained model was retrained. In addition, a dropout rate of 0.5 was used between fully connected dense layers to reduce overfitting. To avoid overfitting problems, data augmentation techniques such as image cropping, standardization, random shifts, and horizontal image flips were implemented during model training. The properties of the pre-trained CNN models considered in this study are listed in Table 4.

4. Results and Discussion

Several experiments were performed to establish the performance of the 12 CNN models on image datasets. The potential of both FE and FT TL methods for structural image classification is analyzed in this section.

4.1. FE with Bottleneck Features

Figure 4a,b show the FE results of FE using the six pre-trained models. The pre-trained MobileNet CNN model exhibited training and validation accuracies of approximately 59% and 58.4%, respectively. Thus, it outperformed all the other models.
Notably, the ResNet50 model demonstrated categorically unsatisfactory performance compared with the other models, indicating that the architecture of the ResNet50 model was deeper and more difficult to train than those of the other models. Similarly, the VGG16 and VGG19 models demonstrated unsatisfactory performance, which might be due to their shallow architectures. However, the superior accuracy of MobileNet suggests that it is the best model for mobile application development.

4.2. FT

The FT results for the six pre-trained models are shown in Figure 5a,b. Similarly, the pre-trained MobileNet CNN model outperformed the other models in terms of its training and validation accuracies of approximately 73.4% and 71.8%, respectively.

4.3. Comparison between FE and FT

The FT method performed better than the FE method for all models and datasets considered in this study. However, the FT method is computationally expensive because it involves retraining one convolutional block. Figure 6 shows the training and validation accuracies for each model implemented through TL.
The results of the testing accuracy analyses for all the models are presented as bar charts in Figure 7.

4.4. Comparative Study: Effect of Dataset Size on Fine-Tuned Model

Because DL models are generally data intensive, the effect of data size on the performance of the fine-tuned MobileNet model was examined by gradually increasing the amount of training image data (Figure 8).
An increase in the number of training images considerably affected the performance of the model (Figure 9). For example, the testing accuracies of the fine-tuned MobileNet model for datasets A, B, and C were 88.3%, 90.6%, and 95.6%, respectively. Thus, we infer that adding more training data to the model can improve its validation accuracy. Moreover, this is consistent with the findings of [6], which suggests that increasing the data and fine-tuning the convolutional blocks can improve the model performance.
The fine-tuned MobileNet CNN model, which exhibited optimal performance with a testing accuracy of 88.3%, was selected for deployment in a web-based application for earthquake-damage classification. Figure 10 shows plots of the confusion matrix used to evaluate the model performance of the testing images.
To assess the performance of the fine-tuned MobileNet CNN model, the testing accuracy was compared with those of various CNN architectures used for similar SD classification tasks. A comparison of the different models with the optimal model is presented in Table 5. Accuracy can be expressed as the ratio of the true predictions to the total predicted cases in the dataset. The precision metric measures the classifier’s ability to correctly identify positive classes. The recall metric is the ratio of positive instances that are correctly detected by the classifier to the total number of positive instances. The mathematical expressions for accuracy, precision, recall, and F1 score are shown in Equations (1a)–(1d), respectively.
A c c u r a c y   =   T P   +   T N T P   +   T N   +   F P   +   F N
P r e c i s i o n   =   T P T P   +   F P
R e c a l l   =   T P T P   +   F N
F 1   =   2   ×   p r e c i s i o n   ×   r e c a l l p r e c i s i o n   +   r e c a l l   =   T P T P   +   F N   +   F P 2
where TP = number of true positives, TN = number of true negatives, FP = number of false positives, and FN = number of false negatives.
The proposed model was trained on datasets containing images of all structural members similar to those used by Gao and Mosalam [6], which involve extremely noisy backgrounds. However, the dataset considered by Pan and Yang [8] contained only images of reinforced concrete structural columns with less background noise; hence, their approach afforded higher accuracy.
A sample of the testing images with predictions obtained from the fine-tuned MobileNet model is shown in Figure 11. Despite the varying inclination of the camera view and light intensity of the images, the model successfully predicted the SD classes, with extremely few instances of incorrect predictions. For example, it predicted light damage in two cases, as shown in Figure 9b, instead of the ground truth, which indicates moderate damage. This misclassification can be attributed to the overlapping of hairline cracks (light damage) and wide cracks (moderate damage) in the images. Similarly, moderate damage was occasionally misclassified as severe damage, which might be attributed to background noise, such as the presence of iron bars and large window voids in the images. Hence, a more robust bounding box object-detection technique or other forms of damage localization in the model should be considered to overcome this deficiency.
The accuracy of computer-vision-based SD assessment is mainly affected by the complexity of the structure and damage. The damage assessment results can be affected by the varying lighting conditions, occlusion, and insufficient known reference points on a damaged structure that can be used for comparison with pre-damage images to accurately assess damage levels. Moreover, SD caused by debris and rubble can often be difficult or impossible to detect using computer-vision algorithms alone.

4.5. Visualization and Localization of Damage Using Grad-CAM

Grad-CAM is a visualization technique that visualizes and clarifies predictions from large classes of CNNs to render them more transparent. Initially published by Selvaraju et al. [40], Grad-CAM uses the gradient of the target concept in the last convolution layer to create an approximate localization map that highlights the areas of interest to predict the concept.
Grad-CAM was used to extract gradients from the fine-tuned MobileNet CNN model in the final convolutional layer to generate localization maps that identify relevant regions in the test images. This visualization technique is advantageous over the conventional bounding-box method, which is subjective as it requires manual annotations. The heat maps generated via Grad-CAM exhibit smooth boundaries, which provide insight into the precise location of defects or damage in the SD images. Figure 12 shows representative images from different SD classes localized using the Grad-CAM and guided Grad-CAM methods.
The mislocalization of the moderate damage image (in Figure 12b) is attributed to the lower predicted probability (88.83%) for this image compared with the light (98.13 %) and severe (97.29 %) damage images.

4.6. Damage Severity Measurement

Following the approach of Li et al. [16], damage severity was quantified by assigning a damage assessment value (DAV) obtained from Grad-CAM-based damage detection map (DDM). Mathematically, for an input image x with output damage class yD of the VGG19 CNN model, the gradient-based weight parameter wk is the aggregate of gradients in y with respect to fk(i,j) for i and j, and is expressed as follows:
w k   =   1 14   ×   14   i , j y D f i , j k ,
where fk(i,j) is the k-th feature map in the last convolutional layer (which measures 14 × 14 × 512 in this study), i = 1, …, 14, j = 1, …, 14, and k = 1, …, 512.
For feature maps fk and the corresponding weights wk, a 14 × 14 matrix S can be defined as
s i , j   =   R e L U ( k w k f i , j k ) ,
where ReLU() eliminates the effects of negative values and emphasizes positive values.
In the DDM, numerical values are assigned to quantify the damage severity based on the pixel intensity. Higher pixel intensities reflect more severe damage and are represented by a heat map in the DDM. The average numerical values obtained from the heat map of an image are regarded as the overall DAV, which quantifies the damage severity of the image. Hence, a high DAV indicates severe damage and is defined as follows:
DAV   =   1 14   ×   14   k s i , j ,
where si,j represents the elements in matrix S, and the dimensions of S are 14 × 14.
The DAV ranges between 0 (no damage) and 1 (total collapse).
An annotation tool, known as LabelMe [41], accessible at http://labelme.csail.mit.edu, is used to annotate the SD images manually. The numerical values are assigned as follows: no damage = 0, light damage = 0.25, moderate damage = 0.5, severe damage = 0.75, and total collapse = 1 [42]. The annotated sample images are shown in Figure 13, along with their corresponding severity values.

5. Development of CNN Model as Interactive Web Application

Access to trained DL/machine learning models in portable and interactive formats can facilitate real-time practical damage assessments. As shown in Figure 14, the optimal earthquake damage classifier model is converted to a Tensorflow.js compatible format and deployed as a web application with an easy-to-use graphic user interface. Tensorflow.js, built on the Tensorflow framework, facilitates the conversion of machine learning models to JavaScript formats, accessible through web browsers [43]. In addition to the superior accuracy of the optimal MobileNet CNN model, its lightweight size renders it the best model among all the trained models. An interactive web application is a useful tool that allows users to upload SD images and rapidly determine the class of SD with the corresponding confidence level of prediction. The prediction probability is computed based on the softmax function, as shown in Equation (5).
P ( y i )   =   e x p y i j   =   1 n exp y j ,
where P( y i ) is the prediction probability of class i, y i is the output score for class i, and n is the number of classes.
The trained CNN model is accessible to web browsers at https://bit.ly/3hXRyyc. This allows emergency responders to rapidly assess post-earthquake damage and make informed decisions regarding resource allocation. In addition, users can upload images captured at ground level from different sources to identify the SD and further validate the performance of the proposed earthquake-damage-classifier model.

6. Conclusions and Recommendations

Artificial intelligence for post-earthquake inspections and reconnaissance has recently received significant attention, owing to its exponential increase in computational capabilities and the inherent potential of artificial intelligence to address the disadvantages associated with manual inspections, including subjectivity. In this study, we used data from the 2017 Pohang earthquake to demonstrate the potential of automated DL for rapid and accurate inspections of post-earthquake damage with insignificant human input.
Our key findings are as follows:
  • The FT method outperformed the FE method for all the CNN models evaluated. However, the FT method is more computationally complex than the FE method because it involves retraining one convolutional block.
  • The MobileNet model exhibited the best performance for both the FE and FT TL methods, exhibiting testing accuracies of 76.1% and 88.3%, respectively. The superiority of the MobileNet model in performing classification promoted its deployment as a web-based application for earthquake-damage classification.
  • The web application successfully predicted the damage class in new images of seismic damage with high certainty. In addition, interactive web pages can rapidly and automatically classify SD from earthquakes, thereby facilitating decision making in response to earthquakes.
In this study, we demonstrated the potential of automated DL to facilitate post-earthquake damage inspections and surveys. Despite the limitations of this study, including the lack of a large and sophisticated training dataset and the complexity of the four damage classes, future studies will be conducted that focus on establishing a large benchmark dataset with high-quality annotations, such as the PEER Hub ImageNet [12]. In addition, future experiments, involving unmanned aerial vehicles, will be performed to capture real-time images from SD sites that can be sent to a webpage interface for fully automated damage assessment.

Author Contributions

Conceptualization, P.D.O., S.-S.P. and B.K.; methodology, P.D.O. and S.-S.P.; software, P.D.O.; validation, P.D.O. and B.K.; data curation, P.D.O.; writing—original draft preparation, P.D.O.; writing—review and editing, P.D.O., S.-S.P., B.K. and D.-E.L.; visualization, P.D.O.; supervision, S.-S.P., B.K. and D.-E.L.; project administration, B.K. and D.-E.L.; funding acquisition, D.-E.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by a National Research Foundation of Korea (NRF) grant, funded by the Korean government (MSIT) (No. NRF-2018R1A5A1025137).

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The abbreviations used in this manuscript are as follows:
CNNConvolutional neural network
DAVDamage assessment value
DDMDamage detection map
DLDeep learning
FEFeature extraction
FTFine-tuning
GPUGraphic processing unit
Grad-CAMGradient-weighted class activation mapping
OLeNetOptimized LeNet
PEERPacific Earthquake Engineering Research
ReLURectified linear unit
SDStructural damage
TLTransfer learning
VGGVisual geometry group

References

  1. Lowes, L.; DesRoches, R.; Eberhard, M.; Parra-Montesinos, G. NSF RAPID: Collection of Damage Data Following Haiti Earthquake (NEES-2010-0872) 2017. Available online: https://datacenterhub.org/resources/14372 (accessed on 21 August 2020).
  2. NCREE; Purdue University. Performance of Reinforced Concrete Buildings in the 2016 Taiwan (Meinong) Earthquake 2016. Available online: https://datacenterhub.org/resources/14098 (accessed on 21 August 2020).
  3. Shah, P.; Pujol, S.; Puranam, A.; Laughery, L. Database on Performance of Low-Rise Reinforced Concrete Buildings in the 2015 Nepal Earthquake; Purdue University Research Repository: Lafayette, IN, USA, 2015; Available online: https://datacenterhub.org/resources/238 (accessed on 21 August 2020).
  4. Sim, C.; Laughery, L.; Chiou, T.C.; Weng, P. 2017 Pohang Earthquake-Reinforced Concrete Building Damage Survey. 2018. Available online: https://datacenterhub.org/resources/14728 (accessed on 21 August 2020).
  5. Sim, C.; Villalobos, E.; Smith, J.P.; Rojas, P.; Pujol, S.; Puranam, A.Y.; Laughery, L. Performance of Low-Rise Reinforced Concrete Buildings in the 2016 Ecuador Earthquake. 2016. Available online: https://datacenterhub.org/resources/14160 (accessed on 21 August 2020).
  6. Gao, Y.; Mosalam, K.M. Deep transfer learning for image-based structural damage recognition. Comput. Aid. Civ. Infrastruct. Eng. 2018, 33, 748–768. [Google Scholar] [CrossRef]
  7. Nahata, D.; Mulchandani, H.; Bansal, S.; Muthukumar, G. Post-earthquake assessment of buildings using deep learning. arXiv 2019, arXiv:1907.07877. [Google Scholar]
  8. Pan, X.; Yang, T.Y. Postdisaster image-based damage detection and repair cost estimation of reinforced concrete buildings using dual convolutional neural networks. Comput. Aided Civ. Infrastruct. Eng. 2020, 35, 495–510. [Google Scholar] [CrossRef]
  9. Hoskere, V.; Narazaki, Y.; Hoang, T.A.; Spencer, B.F., Jr. Vision-based structural inspection using multiscale deep convolutional neural networks. In Proceedings of the 3rd Huixian International Forum on Earthquake Engineering for Young Researchers, Urbana-Champaign, IL, USA, 11–12 August 2012. [Google Scholar]
  10. Zhai, W.; Peng, Z.R. Damage assessment using Google street view: Evidence from hurricane Michael in Mexico Beach, Florida. Appl. Geogr. 2020, 123, 102252. [Google Scholar] [CrossRef]
  11. Liang, X. Image-based post-disaster inspection of reinforced concrete bridge systems using deep learning with Bayesian optimization. Comput. Aided Civ. Infrastruct. Eng. 2019, 34, 415–430. [Google Scholar] [CrossRef]
  12. Gao, Y.; Mosalam, K.M. PEER Hub ImageNet (Ø-Net): A Large-Scale Multi-Attribute Benchmark Dataset of Structural Images; Report No. 2019/07; Earthquake Engineering Research Center Headquarters at the University of California: Berkeley, CA, USA, 2019. [Google Scholar] [CrossRef]
  13. Naser, M.Z.; Kodur, V.; Thai, H.T.; Hawileh, R.; Abdalla, J.; Degtyarev, V.V. StructuresNet and FireNet: Benchmarking databases and machine learning algorithms in structural and fire engineering domains. J. Build. Eng. 2021, 44, 102977. [Google Scholar] [CrossRef]
  14. Cha, Y.J.; Choi, W.; Büyüköztürk, O. Deep learning-based crack damage detection using convolutional neural networks. Comput. Aided Civ. Infrastruct. Eng. 2017, 32, 361–378. [Google Scholar] [CrossRef]
  15. Ghosh Mondal, T.G.; Jahanshahi, M.R.; Wu, R.T.; Wu, Z.Y. Deep learning-based multi-class damage detection for autonomous post-disaster reconnaissance. Str. Control Health Monit. 2020, 27, e2507. [Google Scholar] [CrossRef]
  16. Li, X.; Caragea, D.; Zhang, H.; Imran, M. Localizing and quantifying infrastructure damage using class activation mapping approaches. Soc. Netw. Anal. Min. 2019, 9, 44. [Google Scholar] [CrossRef]
  17. Kim, B.; Yuvaraj, N.; Park, H.W.; Preethaa, K.R.S.; Pandian, R.A.; Lee, D. Investigation of steel frame damage based on computer vision and deep learning. Autom. Constr. 2021, 132, 103941. [Google Scholar] [CrossRef]
  18. Spencer, B.F.; Hoskere, V.; Narazaki, Y. Advances in computer vision-based civil infrastructure inspection and monitoring. Engineering 2019, 5, 199–222. [Google Scholar] [CrossRef]
  19. Sony, S.; Dunphy, K.; Sadhu, A.; Capretz, M. A Systematic Review of Convolutional Neural Network-Based Structural Condition Assessment Techniques. Eng. Struct. 2021, 226, 111347. [Google Scholar] [CrossRef]
  20. Kim, B.; Serfa Juan, R.O.S.; Lee, D.-E.; Chen, Z. Importance of image enhancement and CDF for fault assessment of photovoltaic module using IR thermal image. Appl. Sci. 2021, 11, 8388. [Google Scholar] [CrossRef]
  21. Kim, B.; Yuvaraj, N.; Sri Preethaa, K.R.; Santhosh, R.; Sabari, A. Enhanced pedestrian detection using optimized deep convolution neural network for smart building surveillance. Soft Comput. 2020, 24, 17081–17092. [Google Scholar] [CrossRef]
  22. Kim, B.; Yuvaraj, N.; Sri Preethaa, K.R.; Arun Pandian, R. Surface crack detection using deep learning with shallow CNN architecture for enhanced computation. Neural Comput. Appl. 2021, 33, 9289–9305. [Google Scholar] [CrossRef]
  23. ATC-58. Development of next Generation Performance-Based Seismic Design Procedures for New and Existing Buildings. ATC 2007. Redwood City, CA, USA. Available online: https://www.atcouncil.org (accessed on 28 March 2022).
  24. Yuvaraj, N.; Kim, B.; Preethaa, K.R.S. Transfer learning based real-time crack detection using unmanned aerial system. Int. J. High-Rise Build. 2020, 9, 351–360. [Google Scholar] [CrossRef]
  25. Dhanamjayulu, C.; Nizhal, U.N.; Maddikunta, P.K.R.; Gadekallu, T.R.; Iwendi, C.; Wei, C.; Xin, Q. Identification of malnutrition and prediction of BMI from facial images using real-time image processing and machine learning. IET Image Process. 2022, 16, 647–658. [Google Scholar] [CrossRef]
  26. Perez, H.; Tah, J.H.M.; Mosavi, A. Deep learning for detecting building defects using convolutional neural networks. Sensors 2019, 19, 3556. [Google Scholar] [CrossRef] [Green Version]
  27. Zhang, C.; Chang, C.-C.; Jamshidi, M. Concrete bridge surface damage detection using a single-stage detector. Comput. Aided Civ. Infrastruct. Eng. 2020, 35, 389–409. [Google Scholar] [CrossRef]
  28. Dais, D.; Bal, İ.E.; Smyrou, E.; Sarhosis, V. Automatic crack classification and segmentation on masonry surfaces using convolutional neural networks and transfer learning. Autom. Constr. 2021, 125, 103606. [Google Scholar] [CrossRef]
  29. Dung, C.V.; Anh, L.D. Autonomous concrete crack detection using deep fully convolutional neural network. Autom. Constr. 2019, 99, 52–58. [Google Scholar] [CrossRef]
  30. Zhang, Y.; Sun, X.; Loh, K.J.; Su, W.; Xue, Z.; Zhao, X. Autonomous bolt loosening detection using deep learning. Str. Health Monit. 2019, 19, 105–122. [Google Scholar] [CrossRef]
  31. Liu, H.; Zhang, Y. Image-driven structural steel damage condition assessment method using deep learning algorithm. Measurement 2019, 133, 168–181. [Google Scholar] [CrossRef]
  32. Ali, L.; Alnajjar, F.; Jassmi, H.A.; Gocho, M.; Khan, W.; Serhani, M.A. Performance evaluation of deep CNN-based crack detection and localization techniques for concrete structures. Sensors 2021, 21, 1688. [Google Scholar] [CrossRef] [PubMed]
  33. Chen, J.; Yang, T.; Zhang, D.; Huang, H.; Tian, Y. Deep learning based classification of rock structure of tunnel face. Geosci. Front. 2021, 12, 395–404. [Google Scholar] [CrossRef]
  34. Harvey, W.; Rainwater, C.; Cothren, J. Direct aerial visual geolocalization using deep neural networks. Remote Sens. 2021, 13, 4017. [Google Scholar] [CrossRef]
  35. Nath, N.D.; Behzadan, A.H.; Paal, S.G. Deep learning for site safety: Real-time detection of personal protective equipment. Autom. Constr. 2020, 112, 103085. [Google Scholar] [CrossRef]
  36. Narazaki, Y.; Hoskere, V.; Hoang, T.A.; Fujino, Y.; Sakurai, A.; Spencer, B.F. Vision-based automated bridge component recognition with high-level scene consistency. Comput. Aided Civ. Infrastruct. Eng. 2020, 35, 465–482. [Google Scholar] [CrossRef]
  37. Bang, S.; Park, S.; Kim, H.; Kim, H. Encoder–decoder network for pixel-level road crack detection in black-box images. Comput. Aided Civ. Infrastruct. Eng. 2019, 34, 713–727. [Google Scholar] [CrossRef]
  38. Maeda, H.; Sekimoto, Y.; Seto, T.; Kashiyama, T.; Omata, H. Road damage detection and classification using deep neural networks with smartphone images. Comput. Aided Civ. Infrastruct. Eng. 2018, 33, 1127–1141. [Google Scholar] [CrossRef]
  39. Cheng, C.S.; Behzadan, A.H.; Noshadravan, A. Deep learning for post-hurricane aerial damage assessment of buildings. Comput. Aided Civ. Infrastruct. Eng. 2021, 36, 695–710. [Google Scholar] [CrossRef]
  40. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-CAM: Visual explanations from deep networks via gradient-based localization. Int. J. Comput. Vis. 2020, 128, 336–359. [Google Scholar] [CrossRef] [Green Version]
  41. Russell, B.C.; Torralba, A.; Murphy, K.P.; Freeman, W.T. LabelMe: A database and web-based tool for image annotation. Int. J. Comput. Vis. 2008, 77, 157–173. [Google Scholar] [CrossRef]
  42. Nia, K.R.; Mori, G. Building Damage Assessment Using Deep Learning and Ground-Level Image Data. In Proceedings of the 14th Conference on Computer and Robot Vision (CRV), Edmonton, AB, Canada, 16–19 May 2017; IEEE Publications: Piscataway, NJ, USA, 2017; Volume 2017, pp. 95–102. [Google Scholar] [CrossRef]
  43. TensorFlow. Js. Available online: https://www.tensorflow.org/js (accessed on 22 February 2022).
Figure 1. Workflow of the research method used in current study.
Figure 1. Workflow of the research method used in current study.
Sensors 22 03471 g001
Figure 2. Samples used for training images for each damage class: (a) no, (b) light, (c) moderate, and (d) severe damage.
Figure 2. Samples used for training images for each damage class: (a) no, (b) light, (c) moderate, and (d) severe damage.
Sensors 22 03471 g002
Figure 3. Procedure for TL using (a) FE and (b) FT.
Figure 3. Procedure for TL using (a) FE and (b) FT.
Sensors 22 03471 g003aSensors 22 03471 g003b
Figure 4. Plots of accuracy for models trained using FE TL: (a) training and (b) validation.
Figure 4. Plots of accuracy for models trained using FE TL: (a) training and (b) validation.
Sensors 22 03471 g004
Figure 5. Plots of accuracy for models trained using FT TL: (a) training and (b) validation.
Figure 5. Plots of accuracy for models trained using FT TL: (a) training and (b) validation.
Sensors 22 03471 g005
Figure 6. Training and validation accuracies of various CNN models implemented through TL: (a) VGG16, (b) VGG19, (c) Inception, (d) MobileNet, (e) ResNet50, and (f) Xception models.
Figure 6. Training and validation accuracies of various CNN models implemented through TL: (a) VGG16, (b) VGG19, (c) Inception, (d) MobileNet, (e) ResNet50, and (f) Xception models.
Sensors 22 03471 g006aSensors 22 03471 g006b
Figure 7. Bar charts showing testing accuracies for (a) FT and (b) FE of CNN models.
Figure 7. Bar charts showing testing accuracies for (a) FT and (b) FE of CNN models.
Sensors 22 03471 g007
Figure 8. Summary of training and validation datasets for comparative study.
Figure 8. Summary of training and validation datasets for comparative study.
Sensors 22 03471 g008
Figure 9. Bar charts showing testing accuracies for datasets A, B, and C using fine-tuned MobileNet CNN model.
Figure 9. Bar charts showing testing accuracies for datasets A, B, and C using fine-tuned MobileNet CNN model.
Sensors 22 03471 g009
Figure 10. Confusion matrix for the fine-tuned MobileNet CNN model.
Figure 10. Confusion matrix for the fine-tuned MobileNet CNN model.
Sensors 22 03471 g010
Figure 11. Sample testing images of structural damage with predicted probability for cases of (a) correct and (b) incorrect predictions.
Figure 11. Sample testing images of structural damage with predicted probability for cases of (a) correct and (b) incorrect predictions.
Sensors 22 03471 g011
Figure 12. Representative images illustrating damage visualization and localization analyses via gradient-weighted class activation mapping (Grad-CAM) methods for images of (a) light, (b) moderate, and (c) severe damage.
Figure 12. Representative images illustrating damage visualization and localization analyses via gradient-weighted class activation mapping (Grad-CAM) methods for images of (a) light, (b) moderate, and (c) severe damage.
Sensors 22 03471 g012
Figure 13. Sample images with annotations for severity and corresponding damage assessment value (DAV) scores for images of (a) light damage, (b) moderate damage, and (c) severe damage.
Figure 13. Sample images with annotations for severity and corresponding damage assessment value (DAV) scores for images of (a) light damage, (b) moderate damage, and (c) severe damage.
Sensors 22 03471 g013
Figure 14. Graphical user interface for web-based application that integrates optimal MobileNet damage classifier model.
Figure 14. Graphical user interface for web-based application that integrates optimal MobileNet damage classifier model.
Sensors 22 03471 g014
Table 1. Categorized summary of the image dataset.
Table 1. Categorized summary of the image dataset.
Image SourceNo DamageLight
Damage
Moderate DamageSevere
Damage
Pohang (2017) [4]49294187551
Haiti (2010) [1]5255174127
Nepal (2015) [3]152153123255
Taiwan (2016) [2]3992734
Ecuador (2016) [5]4108115188
Total2607096261155
Table 2. Categorized summary of images in training, validation, and testing datasets.
Table 2. Categorized summary of images in training, validation, and testing datasets.
ImageNo DamageLight
Damage
Moderate DamageSevere Damage
Training160320320480
Validation408080120
Testing45454545
Total245445445645
Table 3. CNN model algorithm pseudocode.
Table 3. CNN model algorithm pseudocode.
CNN Algorithm
Programming language used for implementation: Python.
Libraries for CNN model building: Tensorflow and Keras.
Libraries used for image augmentation: OpenCV and computer vision library.
Libraries used for visualizations: Matplotlib and 2D graph tool.
1.
Let X be the input image of the batch and y be the label for the image X.
2.
Extract features from the image using a CNN algorithm.
  • Freeze all the pretrained convolutional blocks to serve as a feature extractor or fine tune by unfreezing the last convolutional blocks.
  • Obtain feature maps of the first layer a0 after passing the image into the convolution layer with 7 × 7 filters and apply batch normalization function along with ReLU function.
  • Apply the global average pooling function to the output tensor a0.
  • Flatten the output to obtain a feature vector.
3.
Execute the feature classification network on the feature vector.
  • Initialize the weight w and bias b arrays of the linear network comprising 256 neural nodes. Add 50% dropout to serve as a regularizer and reduce overfitting.
  • Perform z = w. afeature + b.
  • Perform ReLU activation function al = max (z, 0).
  • Initialize weight w and bias b arrays of linear network with four neural nodes.
  • Perform z = w. al + b.
  • Perform ReLU activation function al = max (z, 0).
  • Apply softmax function on al to obtain the probability distribution of the four classes: no, light, moderate, and severe damage.
Table 4. Comparison of proposed pre-trained CNN models.
Table 4. Comparison of proposed pre-trained CNN models.
ModelNo. of ParametersDepth of LayersSize (MB)
VGG16138.4 M16528
VGG19143.7 M19549
Inception23.9 M18992
Xception22.9 M8188
ResNet25.6 M10798
MobileNet4.3 M5516
Table 5. CNN-based SD classification models compared with current study.
Table 5. CNN-based SD classification models compared with current study.
Task DescriptionAlgorithmAccuracy (%)* Precision (%)* Recall (%)References
Classification of damage in all structural membersVGG1668.8--[6]
Classification of damage in columns onlyResNet5087.47--[8]
Classification of damage in all structural membersMobileNet88.38988.2Current work
* Values are not provided in the referenced study.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Ogunjinmi, P.D.; Park, S.-S.; Kim, B.; Lee, D.-E. Rapid Post-Earthquake Structural Damage Assessment Using Convolutional Neural Networks and Transfer Learning. Sensors 2022, 22, 3471. https://doi.org/10.3390/s22093471

AMA Style

Ogunjinmi PD, Park S-S, Kim B, Lee D-E. Rapid Post-Earthquake Structural Damage Assessment Using Convolutional Neural Networks and Transfer Learning. Sensors. 2022; 22(9):3471. https://doi.org/10.3390/s22093471

Chicago/Turabian Style

Ogunjinmi, Peter Damilola, Sung-Sik Park, Bubryur Kim, and Dong-Eun Lee. 2022. "Rapid Post-Earthquake Structural Damage Assessment Using Convolutional Neural Networks and Transfer Learning" Sensors 22, no. 9: 3471. https://doi.org/10.3390/s22093471

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop