Tsunami Damage Detection with Remote Sensing: A Review

: Tsunamis are rare events compared with the other natural disasters, but once it happens, it can be extremely devastating to the coastal communities. Extensive inland penetration of tsunamis may cause the difﬁculties of understanding its impact in the aftermath of its generation. Therefore the social needs to technologies of detecting the wide impact of great tsunamis have been increased. Recent advances of remote sensing and technologies of image analysis meet the above needs and lead to more rapid and efﬁcient understanding of tsunami affected areas. This paper provides a review of how remote sensing methods have developed to contribute to post-tsunami disaster response. The evaluations in the performances of the remote sensing methods are discussed according to the needs of tsunami disaster response with future perspective.


Introduction
The word Tsunami comes from a Japanese term derived from the characters "tsu" (meaning harbor) and "nami" (meaning wave), and is now widely common in the world. Physically, tsunami is defined as a series of water waves caused by the sudden displacement of a large volume of water, usually in the sea or even in a large lake. Tremendous volumes of water displacement makes a tsunami and devastate coastal regions, sometimes across the ocean. Earthquakes, volcanic eruptions, landslides, underwater landslides, and large mass impact into the ocean (meteorite/asteroid impacts or similar impact events) all have potentials to generate tsunamis.
Tsunami events are rare compared with the other natural disasters but can be extremely devastating. In total, 16 major tsunamis killed 250,900 people in 21 countries between 1996 and 2015 [1,2], and it averages more than 15,600 deaths per single event. This rate is likely to be high compared with other natural disasters including storms (such as cyclones or typhoons), floods, and earthquakes. It is also true that this rate is significantly contributed by two catastrophic events; the 2004 Sumatra-Andaman earthquake tsunami (hereafter, Indian Ocean tsunami) and the 2011 Great East Japan earthquake tsunami (hereafter, Tohoku tsunami), but it characterizes a tsunami as a devastating natural disaster [3].
In the aftermath of catastrophic tsunami disasters, our society is confronted with significant difficulties of understanding impacts with the limited time, resources, and information. For instance, in the 2004 Indian Ocean tsunami that propagated the entire Indian Ocean and caused extensive impacts to 12 countries. As a result of the vast damage on structures, infrastructures, communication networks, and the failure of emergency response system, whole picture of the impact could not be addressed for months.
Thanks to recent advances and improvements in satellite sensors, data accessibility, applications, and services, many space agencies support data-sharing policies that facilitate access to remotely-sensed data for more efficient use in disaster management [4]. Tremendous progress has been made with sophisticated methods to analyze imageries and geospatial data in near real-time via geo-web-services and crowd-sourcing [5], and those can be used in disaster management and emergency response. Satellite earth observations achieved consistent and repeated coverage of the world, and that makes it possible to understand and share the information on disaster impacts among the countries, regardless of time and weather conditions. This paper aims to provide a review of the advances of remote sensing methods and its applications to detecting the tsunami impacts, and to evaluate their performances in post-disaster response point of view with some discussions of future perspective. The paper is organized as follows. Section 2 describes the physical characteristics of tsunamis and summarizes the features that can be acquired by remote sensing. Section 3 focuses on the consequences of tsunami's inland penetration and discusses the tsunami damage interpretation by various sensors and platforms. Sections 4 and 5 introduce recent advances of machine/deep learning to detect tsunami damage and discusses its performances. Section 6 discusses future perspectives of machine/deep learning to enhance tsunami damage detection capability. Section 7 presents our conclusions.

Tsunami Physics and Acquisition of Tsunami Features with Remote Sensing
What tsunami features can be acquired by remote sensing platforms and sensors during its generation, propagation and after inland penetration? This is an interesting question, since tsunamis are unpredictable phenomena. Table 1 summarizes the remote sensing platforms and sensors that have acquired tsunami features during and after the events.
Tsunami is categorized as long wave in water surface waves, and its wave length (L) is much longer than the water depth (h) where a tsunami propagates [3]. In general, when h/L < 1/20, a wave has the characteristics of long wave. Consider a tsunami is generated offshore at depths of several thousand meters by a sudden sea bottom deformation due to an earthquake. The wavelength of a traveling tsunami varies from tens to hundreds of kilometers according to the fault rupture scale and the sea surface height varies from several centimeters to several meters. In many cases, tsunamis in such deep water are much smaller in amplitude than those near the coasts, then are amplified enormously as approaching the coasts. This makes tsunamis difficult to detect over deep water, where ships are unable to feel their passage, and it has been believed that only some offshore sensors, e.g., ocean bottom pressure sensors [6], can detect the tsunami's passage.
However, there was probably only an exception in 2004. The tsunami passage in the Indian Ocean, generated by the 2004 Sumatra-Andaman earthquake, was measured by Jason-1 satellite. Jason-1 was a mission satellite launched by National Aeronautics and Space Administration (NASA, United States) and Centre National d'études Spatiales (CNES, France) and has measured the sea surface height of traveling tsunami along the track approximately two hours after the earthquake. This was probably the first example for satellite remote sensing to detect the traveling tsunami in mid-ocean [7]. Figure 1 illustrates the track of Jason-1 (track 109) on 26 December 2004 and the extracted tsunami height along the track [8]. In the figure, the snapshot of modeled tsunami at 2 h after the earthquake is shown to explain how Jason-1 measured the sea surface height along tsunami propagation direction. Jason-1 clearly detected the sea surface of tsunami front and its peak propagating southward from 3 • S to 5 • S in latitude, which is critical to determine the tsunami source model [9], i.e., tsunami generation mechanism.  Figure 1. (a) The track of the satellite Jason-1 mission on 26 December 2004 with the modeled tsunami propagating at 2 h after the Sumatra-Andaman earthquake [9]. Note that Jason-1 spent approximately 8.3 minutes to fly over the track from 5 • S to 20 • N. (b) The extracted tsunami height along the track [8].
As a tsunami approaches the coast and the water depth becomes shallower, reducing the tsunami's traveling speed compresses the tsunami's wave length, and its amplitude increases enormously. When a tsunami penetrates on land, its characteristic changes significantly, from that of a water wave to a strong inundation flow [3].
To describe the overland flow of a tsunami, there are some important quantities that can be measured with remote sensing, e.g., tsunami inundation extent (how far tsunami penetrates inland) and flow velocities. The aerial videos can help with understanding how the tsunami would penetrate inland. Immediately after the 2011 Tohoku tsunami occurred, NHK (Japan Broadcasting Corporation) sent a crew to Sendai coast on a helicopter to broadcast the tsunami attack. The tsunami attacked Sendai coast approximately 1 h after the earthquake and NHK crew succeeded to shoot the moment of tsunami inundation on the Sendai coast (e.g., https://www.youtube.com/watch?v=zxm050h0k2I). The video includes scientific information of the process of tsunami's inland penetration. Figure 2 shows the procedure for mapping tsunami front [10]. The continuously recorded video was divided into individual frames to be calibrated and rectified by 2-D projective transformation with ground control points which could be identified in pre-event maps and images.
As shown in Figure 2c, the time series of tsunami front on land could be identified with the timestamps of the video. When the distance between two tsunami front lines is divided by the time between the frames, the "tsunami front velocity" can be estimated. In addition, when we focus on the movement of floating objects in the video, the "tsunami flow velocity" is measured [11]. This information is significantly valuable to understand the physical process of tsunami inland penetration, e.g., tsunami flow velocity, hydrodynamic forces on structures.

Tsunami Damage Interpretation in Remote Sensing
Identifying tsunami-induced building damage have been conducted in recent years with increasing interest as a response by scientists due to the devastating events and the need for rapid damage assessment to support disaster relief efforts. This section summarizes the methodologies for detecting tsunami impacts and its effects in the built-up areas recognizable as tsunami damage. Table 2 presents a list of events and related literature focusing on the tsunami-induced damage.
While remote sensing has been applied to monitoring and understanding physical processes of tsunami hazards, the impacts to the societies and damage caused by disasters need to be addressed for disaster response [12,13]. In this sense, satellite remote sensing is advantageous when the impact is of large-scale and a wide area is needed to be grasped rapidly [14]. Satellite data acquisition and fast data processing has increased through these years due to technological advances in the constellation of satellites and the interest in archiving and sharing the information gathered by multiple sensors. Thus, remote sensing has been used to estimate possible damage after major disasters [15,16].
A comprehensive review of earthquake-induced building damage detection with remote sensing was provided by past studies [17]. For instance, two approaches were discussed; multi-temporal techniques and mono-temporal techniques. A multi-temporal technique uses a preand post-event image to evaluate the changes between images and interpret them as possible damage. A mono-temporal method applies only post-event data. In the case of building damage induced by a tsunami, its characteristics are different from the damage mechanisms due to earthquake, therefore, some techniques follow similar principles but need further treatment to adequately interpret the level of tsunami damage [18].
First, it is important to recognize that most of the methods used for damage interpretation or damage recognition need urban-based maps (i.e., building footprint data). Sometimes also building height is required to conduct damage estimation [19,20]. Official cadastral information, open data sources (e.g., OpenStreetMap), reference sources (e.g., Google Street Map, Google Earth), and built-up area extraction methods [21,22] are used to overcome this first step in the damage assessment and damage estimation process. Otherwise, land cover classification, performed from remote sensing data collected before the disaster in question, would be necessary.
In the case of tsunami damage estimation, aerial photos [23], and optical satellite images were used first to classify tsunami affected-areas and extract inundation areas [24]. Similarly, visual interpretation of pre-and post-images of high-resolution optical images are used to classify building damage in the aftermath of a disaster [25]. Currently, this method is still conducted after many types of disasters, including meteorological hazards, geophysical hazards, deliberate and accidental man-made disasters, and other humanitarian disasters as rapid mapping efforts by emergency operational frameworks (e.g., Copernicus Emergency Management Service -European Commission [26], the UN-SPIDER [27], and the Sentinel Asia initiative in the Asia-Pacific [28]).

Using Optical Images
A unique feature of tsunami inundation zone is caused by the sea water penetration on land and cascading effects. Thus, the optical satellite remote sensing for mapping tsunami affected areas focuses on extracting the changes in spectral features related to vegetation, soil, and water observed in a pair of optical satellite images.
Changes in the normalized difference vegetation index (NDVI), soil index (NDSI), and water index (NDWI) are used to extract the inundated areas. In particular case of NDWI, many uses of the spectral bands are verified when looking for land surface changes in particular water body, water level or flooding [15,24,[57][58][59][60]. The use of these indices require a careful comparison of their indices between the pre-and post-event images (or between affected and non affected areas) to decide a threshold of classification. The value of this threshold maximizes the difference between cumulative frequency distributions in order to define at least two classes [24]. Theoretically, NDWI varies from −1 to 1 according to the spectral features of ground surface. Especially, water strongly absorbs near-infrared (NIR) band of light and it indicates higher value of NDWI. Therefore, the analysis for extracting tsunami inundation is performed on the assumption that the NDWI values in the inundated areas are higher because of the existence of water, and they suddenly decrease on the dry land across the inundation limit. This is the basic idea to determine the threshold value of NDWI to extract the inundated zones. Visual inspection of the optical images in affected areas has also been studied with Digital Elevation Model (DEM) to determine tsunami run-up heights [16,62], however it might overestimate the heights due to limited resolution and accuracy of DEM used. In addition, in other areas, tsunami run-up might have been higher than interpreted through visual inspection, because the water limit was not visible due to the lack of debris or other hints of tsunami inundation. An example of inundation limit extraction using Advanced Spaceborne Thermal Emission and Reflection Radiometer (ASTER) data and Shuttle Radar Topography Mission (SRTM) data is shown in Figure 4.

Using Synthetic Aperture Radar (SAR) Data
Since microwaves can penetrate through clouds and can operate in day or night conditions, Synthetic Aperture Radar (SAR) is widely used for monitoring the earth surfaces. The advantage of using SAR data is well-known for detecting surface changes and displacement [63], and has been recognized to be useful for post-disaster damage assessment [30]. Since the first launch of high-resolution SAR satellites in 2007, i.e., TerraSAR-X and CosmoSkyMed, archives of high-resolution SAR data have been constructed, and these data are promising for improving the methods in damage detection.
Lower sigma naught values (strength of the reflected signal, i.e., backscattering coefficient) in SAR data indicate the areas are covered by water due to the specular reflection mechanism of the microwaves on smooth surface. Other backscattering mechanisms, such as double bounce in urban areas and volumetric scattering in forest areas, might be present as well. However, the effect of the acquisition conditions (i.e., sensor depression angle and ascending/descending satellite path) must be taken into account. Radar shadowing, layover and foreshortening are such examples. Semi-automated methods to extract tsunami inundation zone were developed using change detection in SAR data similar to optical image approaches [44,64]. Backscattering coefficients from water body and land cover areas are sampled from pre-event data. Next, a threshold value of back scattering intensity histograms is chosen so water bodies and other features can be classified reliably. Figure 5 shows an example of tsunami inundation areas extracted from TerraSAR-X images in the case of the 2011 Tohoku tsunami [64].

Interpretation of Tsunami-Induced Building Damage
Building damage interpretation is a part of the damage assessment and emergency mapping processes conducted immediately after a disaster. Several satellite-based emergency mapping (SEM) systems have been deployed by local, national and international agencies to carry out this task [65,66]. Mapping the damage helps improving disaster relief effectiveness and reduce suffering and fatalities. The Copernicus Programme [26], the UN-SPIDER [27], and the Sentinel Asia initiative in the Asia-Pacific [28] are some of the earth observation systems supporting disaster response using remote sensing technology.
To identify tsunami-induced building damage one can use optical pre-and post-event images to visually inspect the affected area and observe changes in buildings. It is also used in earthquake-induced [17] or hurricane-induced damage assessments [25,67]. Unfortunately optical images require free cloud conditions to clearly observe the earth surface. To overcome the limitation of these passive sensors (optical image), active sensors (radar satellite image) are used to identify built-up-areas [68,69] and building damage [43,70,71].

Using Optical Images
High-resolution satellite images and aerial photos can be used to detect individual building damage. For instance, post-tsunami satellite imagery of IKONOS was used by the Japan International Cooperation Agency (JICA) after the 2004 Indian Ocean tsunami to visually inspect the damage in Banda Aceh of Sumatra island, Indonesia [72]. This information was then used to develop tsunami fragility functions [9] as a measure to estimate the structural damage due to tsunami attack.
In case of the 2011 Tohoku tsunami, visual inspection of building damage was conducted [23] using aerial photos acquired on the following days (12,13,19 March and 1, 5 April of 2011) by Geospatial Information Authority of Japan (GSI). In this case, the missing roof was the proxy of classification as "washed-away" or "surviving". While the accuracy of this method depends on quality of the image and the expert/user having a keen eye to spot changes, the time-consuming task becomes a drawback in disaster response stages (see Figure 6). Optical images would also be applied to monitor the recovery and reconstruction process after a disaster [32,33].

Damage inspection
Surviving Washed-away 0 2 1 km (c) Geospatial Information Authority of Japan (GSI) Figure 6. Example of mapping structural damage in Ishinomaki city by visual inspection of pre-and post-event aerial photos [23]. The aerial photo archives were provided by Geospatial Information Authority of Japan (GSI). With the ZENRIN building data (building footprint), the inspection was conducted to clarify if buildings were washed-away or surviving, by comparing pre-and post-tsunami aerial photos and looking at existence of each building's roof.
In addition, the use of Unmanned Aerial Vehicles (UAV) are becoming popular to obtain high-resolution images from the disaster areas. Despite its limited coverage compared to satellite images, it offers fast deployment and multiple angle footage providing valuable information for 3D mapping and higher accuracy on damage assessment for localized areas [73][74][75].

Using LiDAR data
Applications of Light Detection And Ranging (LiDAR) for damage detection are still few compared with other remote-sensing technologies. The main reason is the lack of LiDAR data before a disaster. Some attempts exist to monitor the tsunami affected areas focusing on accumulation of debris generated by the tsunami inundation by three dimensional mapping [42]. The authors proposed a fusion of feature extraction [40] with optical sensor data (aerial photos, satellite images) and volumetric estimation with LiDAR data. An object-based analysis of optical sensor data was used to classify/extract the debris area in the earth surface, while a digital surface model (DSM) of LiDAR data was used to estimate the volume of such debris to be used as the observation of debris removal efforts [76]. Figure 7 shows three-dimensional mapping of debris performed by integration of optical image analysis to extract tsunami debris area with the height information of objects on the ground obtained from LiDAR analysis.
A similar approach using LiDAR data can be used to estimate building damage, an example from earthquake-induced building damage can be found [77]. An example of building damage interpretation using LiDAR data is shown in Figure 8. Collapsed buildings were detected using a pair of digital surface models (DSMs), taken before and after the main shock by airborne LiDAR flights. The change in average elevation within a building footprint was found to be the most important factor.

Using Synthetic Aperture Radar (SAR) Data
The radar signal is primarily sensitive to surface structure. Figure 9 shows considerable patterns of land surfaces in tsunami affected area [46]. Change of signal properties in the radar data would be used to detect and evaluate the building damage. Man-made structures reveal high backscattering, i.e., high sigma naught values in the SAR data, because of double bounce, and completely washed-away or collapsed structures indicate lower backscattering. It is inferred that the change ratio in areas with high backscattering varies as washed-away or structurally destroyed. Several exceptions may be expected to characterize tsunami damage. For instance, the case shown in Figure 9B, high backscattering in slightly-damaged buildings with surrounding tsunami debris may result in false positive errors. The cases in Figure 9D,F may not reveal changes in areas with high backscattering, since the movement of buildings and accumulation of debris induce double-bounce scattering. These damage patterns may cause false negative errors. These considerable damage patterns are utilized to discuss the reasons for errors. A number of methods have been proposed for estimating building damage caused by tsunami disasters in recent decades [41,46,48,78]. Following the 2011 Tohoku tsunami disaster, approaches using high-resolution SAR data have been proposed to detect building damage at the building unit scale. Some new approaches were proposed to classify tsunami-induced building damage into multiple classes using pre-and post-event high-resolution radar (TerraSAR-X) data. To exploit the relationships of backscattering features in the tsunami affected areas, the methods for classifying tsunami-induced building damage into multiple classes were developed by analyzing the statistical relationship between the change ratios in areas with high backscattering and in areas with building damage. Figure 10 shows an example of building damage classification in the affected areas of the 2011 Tohoku tsunami [46]. The results provided an overall accuracy of 67.4%. It can be said that the method shows a good performance for classifying damaged buildings, but more discussion has to be done to consider if this performance is acceptable in the disaster response efforts.

From Change Detection to Machine Learning Algorithms
As described in the above sections, change detection is the basis for damage interpretation using remote sensing. Change detection determines the differences in the ground surface from images acquired at different time phases. Two kinds of change detection methods are generally applied [79]; (1) Post-classification comparison and (2) comparative analysis. In the first one, the result of classification of each image is then compared to identify changed and unchanged portions [24]. The latter, however, constructs a difference image from multiple time phases and then detects the change through analysis [80]. In both methods there is a "classification" task. High accuracy of classification is required to ensure the method gives suitable results. However, classifying objects in an image through visual inspection is impractical for rapid damage assessment due to the time consuming task. An alternative is to use real-time collaborative systems and produce classification by multiple experts at the same time. This is how some of the mapping products in the Copernicus Emergency Management System are built. This will help reduce production time, however it might incorporate bias in classifications due to the differences in expertise from different producers. Another option, to overcome previous limitations, is to automatize the whole classification process via machine learning algorithms.
Incorporated supplemental GIS information and human expert knowledge into digital image processing was performed in wetland classification [81]. They describe a two stage process: (1) training and (2) learning. The processes generate a "knowledge base" which is a set of rules that can be used by the expert to perform a final image classification.
In the next section, we will discuss how machine learning and deep learning algorithms are used to support the tsunami hazard and damage assessment.

Application of Machine Learning to Tsunami Damage Detection
This section summarizes the machine learning (ML) algorithms applied to identify the consequences of tsunamis attack to the coast. Here we report the type of remote sensing data used, describe the feature space, and a glimpse of the ML algorithms.

Consider a set of samples
computed from remote sensing, where x i ∈ R n is an n-dimensional vector, hereafter referred as feature vector. A machine learning algorithm refers to any procedure that assigns the samples a distinctive label/class. The reported methods are divided into two groups; unsupervised machine learning and supervised machine learning. In unsupervised learning, the algorithm learns only from the data. On the other hand, a supervised learning algorithm is calibrated from a subset of samples whose correct class is known beforehand.

Thresholding
Thresholding is a procedure used to segment an input image into two classes [12]. A framework is implemented to select automatically samples from which the threshold is defined [78]. Here, a one-dimensional feature vector is used. Each feature represents either a difference or a log-ratio image pixel computed from a pair of images acquired before and after a tsunami. The discriminant function can be expressed as follows; where T is the threshold value that needs to be calibrated. With this objective, the image is split into N subimages and sorted according to their standard deviation. It is expected to find large standard deviation in subimages that contains changed pixels. Therefore, a fraction of the subimages with the largest standard deviation are used to define T. By applying standard thresholding procedures [82,83], the method was successfully used to identify damages induced by the tsunami occurred on 26 December 2004 off the north of Sumatra Island, Indonesia. Two microwave intensity images obtained from the RADARSAT-1 satellite were used. The pre and post-event images were acquired respectively in April 1998 and January 2005. Both have the same acquisition parameters and were filtered to reduce the speckle noise. The seaside pixels were masked as a preprocessing step. Then, log-ratio between the images was computed. The image was divided into 272 subimages (509 × 512 pixels each), from which only 149 were used for further analysis. A total of six subimages were used to set the threshold T, where the Kittler and Illingworth algorithm, extended to a generalized Gaussian model [83] was used.

K-Means Clustering
The algorithm assigns a class k ∈ C to a sample x i by the following expression; where µ k denotes the average samples x i assigned to class k, and is tuned from an iterative procedure termed as maximization step. The number of classes is defined beforehand. In summary, given any starting value, the following two steps are repeated consecutively until reach convergence. First, a sample is assigned using Equation (2). Second, re-estimate the mean µ k as the average of samples assigned to k. Convergence is guaranteed with the maximization step algorithm.
There is an example that applied k-cluster to characterize the surface tsunami effects due to the 2011 Tohoku tsunami [53]. x i is composed of radiance pixels extracted from visible and near-infrared (VNIR) images from ASTER satellite. The resolution of the imagery was 15 m. Four components belonged to the bands of VNIR acquired on 24 February 2011; and the other four components to that acquired on 19 March 2011. A total of 30 different classes were set, from which 6 of them were identified as flooding debris, building damage, mud, stressed, and non stressed vegetation.

Expectation-Maximization (EM) Algorithm
While k-means algorithm updates µ k under a euclidean distance criterion, the expectationmaximization algorithm updates parameters of a distribution model. Here the distribution of X is assumed to represent a mixture of distributions.
where k is the number of classes defined beforehand. G i denotes a class and P(G i ) are the mixture proportions. The distributions are defined from an iterative optimization. For instance, assuming Gaussian distributions, the calibration of their parameters is as follows; where m i and S i denote the vector mean and variance. Further reading on EM algorithm can be found in [84,85]. The EM-algorithm is employed to evaluate the performance of features computed from fully polarimetric SAR images to identify damaged areas due to the 2011 Tohoku tsunami [86]. The images were acquired at 21 November 2010 and 8 April 2011, respectively. Three groups of polarimetric features were tested. The first group is composed of the difference of surface, double-bounce, and volume scattering computed from a decomposition method [87]. The second group of features is composed of entropy, anisotropy, and average scattering angle; all computed from an eigenvalue-eigenvector decomposition [88]. The third group includes the polarimetric orientation angle, the co-polarization coherence in the linear polarization basis and in the circular polarization basis. Positive change, negative change, and no change were set as the three classes.

Imagery, Hazard, and Fragility Function (IHF) Method
The authors proposed a modification of the standard logistic regression classifier within a framework that can be used automatically without the use of training data [49]. As the name suggests, the discriminant function has the form of a logistic function; where the vector w and the value ρ are usually calibrated from training data. A sample x i will be predicted as damaged if f (x i ) > 0.5; otherwise, it is predicted as non-damaged sample. It is pointed out that some large scale disasters, such as earthquakes and tsunamis, can be numerically modeled on near real-time [49]. Furthermore, in many cases there are available the so-called fragility functions. These are functions that provide relationship between the intensity of a hazard at an arbitrary location and the expected percentage of damage. Therefore, under such situations, the discriminant function can be calibrated from the following optimization problem; where d i denotes the intensity of the disaster at the location of sample x i . d i is estimated from a numerical model of the disaster, g(d i ) refers to an arbitrary fragility function. Figure 11 shows a scheme of the application of the IHF method to identify washed away buildings using TerraSAR-X images, estimation of the inundation depth map, and fragility functions as inputs. First, a pair of TerraSAR-X imagery, taken before (21 October 2010) and after (13 March 2011) the 2011 Tohoku tsunami, is used to computed two parameters for each building: an averaged difference and the correlation coefficient. These parameters were computed using the bounding box of the building footprint. Additionally, each sample building is associated with an estimation of the inundation depth (d i ) at its geolocation. With the inundation depth, the probability that a sample will be washed-away can be computed with a fragility function (g(d i )) [89]. Then, having the three source of information (i.e., remote sensing-based features, inundation depth, and a suitable fragility function), the IHF method can be applied. A total of M = 5222 samples, randomly selected but uniformly distributed with respect to their inundation depth, were used to solve Equation (6). The procedure were repeated four times. The resulting linear discriminant functions are shown in the right side of Figure 11, from which it is observed the four discriminant functions are almost the same. The complete dataset (31,262 samples) is also shown, but separated in different scatter plot according to the damage level surveyed by the Ministry of Land, Infrastructure, Transport, and Tourism (MLIT) [90]. It is clearly observed that most of the washed-away buildings are located to the left side of the discriminant function; while the buildings with different damage level are located on the other side. Figure 11. Scheme of the Imagery, Hazard, and Fragility Function (IHF) method applied to the 2011 Tohoku tsunami. TerraSAR-X data is used to construct the feature space, where each sample represents a building. Then each building is associated with a probability of of damage according to the inundation depth map and fragility functions. Finally, a discriminant function is calibrated from Equation (6). The example shows the resulting linear discriminant function that separates washed away buildings from the rest. The figure is modified from [49].

Support Vector Machine
Here, we introduce Support Vector Machine (SVM) as one of supervised learning models used for classification. The discriminant function is defined as follows; where Φ(x) is a function that maps the feature space to a space belonging to a kernel, Φ : R n → F. the approach lies on mapping the feature space to the space F and then define an hyperplane that separates the classes with a maximum margin. The vector w a vector in F that is perpendicular to the referred hyperplane. ρ ∈ R denotes an offset. The parameters w and ρ are computed from the following quadratic programming problem: The function Φ is usually chosen beforehand based on its associated kernel. Very often the Gaussian kernel is employed; Using SAR data was verified to identify the damaged buildings due to the 11 March 2011 Tohoku tsunami [91]. Three TerraSAR-X datasets, one was acquired before and two were acquired after the tsunami, and two scenes of ALOS PALSAR data, acquired before and after the tsunami, were used. The field survey performed by the Ministry of Land, Infrastructure, Transport, and Tourism (MLIT) [90] was used for training and evaluation. The amount and quality of the data provided by the MLIT made the 2011 tsunami a benchmark to test different algorithms. Two types of features were used. The first group is termed multi-temporal change features of a factor that consist of averaged difference, correlation coefficient [92]. These features were computed using images acquired before and after the tsunami event. The second group of features consist of statistical features, such as mode, mean, standard deviation, minimum, and maximum values. All features were computed at a building unit resolution. Different configurations were tested to evaluate the influence of the training samples, the acquisition date, the classification scheme, and the type of microwave sensor.
Not only two-level classifier; damaged and non-damaged, multi-level classifier was explored to improve the performance of the SVM as a multi-class classifier [48]. The study focused on the 2011 tsunami as well. Two microwave intensity images acquired on 20 October 2010 and 12 March 2011 by the TerraSAR-X data were used. The images were used to compute the averaged difference and the correlation coefficient using six different window sizes: 5 × 5, 9 × 9, 13 × 13, 17 × 17, 21 × 21, and 25 × 25. Thus, a total of 12 components constitutes each vector x i . As training data, the MLIT survey results were rearranged, under different configurations, in three groups labeled as washed away, moderate change, and slight change. Regarding the calibration of Equation (7), the Gaussian kernel were employed, Equation (10), and the parameters λ and γ were tuned using a 10-fold cross-validation.
A novel set of features was proposed to be used for damage detection [52]. These new features were computed from the gray level co-occurrence matrix (GLCM), one of many procedures available for texture analysis. The novelty in their study is that the GLCM was constructed in a three-dimensional domain, referred as 3DGLCM. That is, stacked multi-temporal imagery. Under this configuration, it was proved the 3DGLCM tend to be a near to a diagonal matrix in non-damaged areas; whereas the non-zero components are far from the diagonal in damaged areas (Figure 12). A SVM classifier was used to evaluate the performance of the new features to identify collapsed buildings due to the 2011 tsunami. The same datasets in [48] were used in this study as well. A total of eleven 3DGLCM-based features were used as components of x i . The regularization parameter, λ, and the γ of the Gaussian kernel were tuned using a 10-fold cross-validation. It was reported the results outperformed previous studies on the same events with the same imagery.

Decision Trees and Random Forest Classifiers
Decision tree, as the name implies, is a structure of tree-like model for decisions. It is a hierarchical model defined as a sequence of recursive splits. A node denotes a subset defined as follows; where the sets X j and X k are linked by a branch. Every node has associated a metric that measures its impurity. When the impurity metric of an arbitrary node is lower than certain threshold, all the samples that get that node are assigned to a class.Otherwise, a decision function, f j (x i ) > 0, is used to divide the samples into two sets. The advantage of decision tree is its interpretability. Further reading can be found in [93]. Additional information regarding on calibration of decision trees can be found in [94,95]. A parameter referred as C r was proposed to identify damaged buildings due to the 2011 Tohoku tsunami [46]. The microwave intensity pair of TerraSAR-X images were used in the referred study. The new parameter is related to the ratio of the area with high intensity at both pre-and post-event images and the area with high intensity at the pre-event image. Using the proposed parameter, a decision tree was calibrated to classify buildings as slightly damaged, collapsed, and washed away, the feature vector is composed of a single value, x i = C r . The nodes of the calibrated decision tree were defined as follows; The first node, X 1 , includes all the samples. Samples in the nodes X 2 , X 4 , and X 5 are respectively classified as washed away, slightly damaged, and collapsed; and the decision function of nodes X 1 and X 3 are f 1 (x i ) = x i − 0.67 and f 3 (x i ) = x i − 0.46, respectively.
As stated before, in spite of its simplicity, decision tree is popular because it provides a clear description of the model. However, more complex models can be built based on an ensemble of decision trees. The idea behind this method, termed as Random Forest [96], lies on several decision trees calibrated each with a random subset of the training data. Then, a combination of many predictions, each from a decision tree, improve the overall accuracy.
The evaluation was made in the use of a random forest [55] and two further modifications termed as rotation forest [97] and canonical rotation forest [98]. The classifiers were used to identify the building damage during the 2018 Sulawesi, Indonesia earthquake tsunami. A feature vector was composed of pixel values from several earth observation technologies, including microwave satellite images from ALOS-2/PALSAR-2 and Sentinel-1 and optical satellite images from PlanetScope and Sentinel-2. The damage map provided by the Copernicus Emergency Management Service were used as training data [99]. The proposed framework learned the damage patterns from a limited available human-interpreted building damage annotation and expands this information to map a larger affected area [55]. Table 3 summarized the studies reported in this chapter. A glimpse of the accuracy achieved in these referred studies is included. However, it worth mentioning some of the studies evaluated different scenarios an architectures and not all of them used the same metric to measure the resulted accuracy. Thus, the minimum and maximum accuracy is reported. Three different metrics are reported: the overall accuracy (OA), g-mean, and F1. The OA denotes the fraction of samples correctly classified. The g-mean = √ acc + × acc − is a combination of the accuracy of damaged (acc + ) and non-damaged (acc − ) samples. The F1 = 2/(U A −1 + PA −1 ) is a combination of the user accuracy (UA) and producer accuracy (PA). The UA is the percentage of the truth data correctly classified and the PA is the percentage of the predicted samples correctly classified. The F1 is usually computed for each class separately and the values reported in Table 3 is the averaged value.

Application of Deep Learning to Tsunami Damage Detection
Information technologies are rapidly increasing at an unprecedented rate with use of high-performance computing power [100] such that they are increasingly realistic and relevant to characterize the tsunami damage more efficiently. The foremost challenges and opportunities are how to leverage the leading information technologies to gain new insights from earth-observation data for grasping the tsunami damage more efficiently. We state that deep learning (DL) will play a key role in that effort. DL-driven breakthroughs have come initially in the field of computer visions with the purpose of automated image recognition [101], but scientists in disaster field have rapidly adopted and extended these techniques to detect the building damage caused by tsunami. The recent interest in DL among tsunami focused on automated detection of tsunami damage from the satellite images [50,54,102].
Here we introduce the concepts of deep convolutional neural networks (CNN) that consists of a series of layers to learn representations of data with multiple levels of abstraction [103,104], and its classical architecture and case studies of supervised deep learning algorithms for tsunami damage mapping.

Convolutional Neural Networks (CNN)
Convolutional Neural Network (CNN) is one of the most commonly applied deep learning algorithms, and it has been widely used in the disaster-related application [105][106][107][108][109]. The typical structure of CNN for tsunami damage detection from remote sensing image is shown in Figure 13. In general, CNN consists of convolutional layers, pooling layers, and fully connected layers. In the convolutional layers, a CNN utilizes different convolution operation achieved by various filters to convert an N-dimensional matrix of remote sensing image into yet another N-dimensional matrix of feature maps with the purpose of extracting the high-level features such as edges, from the input remote sensing image. Generally, a pooling layer follows a convolutional layer to play the role of reducing the dimensions of feature maps and network parameters. Convolutional layers and pooling layers are both translations invariant as they take the neighboring pixels into account while computation, max pooling, and average pooling are the most effective pooling strategies. Following the last pooling layer in the network as seen in Figure 13, there are several fully-connected layers [110] converting N-dimensional feature maps into a 1D feature vector that can be feed forward into certain number categories for image classification. The first step to train a CNN is to represent the input image with predefined parameters (weights and bias) in each layer, and then the prediction result is employed to calculate the loss cost with the ground truth labels, this procedure is called forward stage. The second step which is also called backward stage is to compute the gradients of each parameter with chain rules based on the loss cost, the results are prepared for the next forward computation. Finally, the robust CNN was trained after sufficient iterations of forward and backward stages.  The advantage of DL lies in that it enables to build deep architectures to learn more abstract information [103,104]. However, the large number of parameters introduced by deep computation may also lead to overfitting problem. This problem can be relieved by regularization techniques such as dropout [111]. During each training case, the dropout algorithm will randomly omit some of the feature detectors to prevent complex co-adaptations and enhance the generalization ability. Class imbalance problem often exists when CNN is applied to image recognition, data augmentation can be utilized to generate the additionally labeled dataset, the common data augmentation techniques are flipping, mirroring, and rotation [112].

Deep Learning Methods for Damage Detection-Case Studies from the 2011 Tohoku tsunami
The state of the art review indicates that deep learning based approaches for tsunami damage detection mainly focus on the 2011 Tohoku tsunami and earthquake as listed in Table 4. First is the case that CNN was used to detect whether the buildings inside the tsunami-affected area were washed-away or not from the pre-and post-tsunami aerial photos [54]. In this study, the building footprint information was used to select the target image patch area. The result achieved an overall accuracy of 94%∼96% and also discovered that the use of the pre-tsunami image does not benefit for the improvement of detection accuracy [54].
Second is using satellite data. As the aerial photos and the building footprint data are not always accessible, the authors proposed a rapid deep neural network based tsunami damage detection approach from only post-event SAR image [102]. As shown in Figure 14a, the built-up areas were extracted from the post-tsunami SAR data using Squeezenet, and then a wide residual network was applied to classify the built-up areas into washed-away, collapsed, and survived in patches. To build a CNN framework that enables to detect tsunami damage in an end-to-end manner, the authors introduced an operational tsunami damage mapping framework [102] based on the U-net Convolutional Network from high-resolution optical images as shown in Figure 14b. This work was built in Microsoft Cognitive Toolkit [113] framework that implemented in the Microsoft Azure platform.

Future Perspective of Deep Learning for Detecting Tsunami Damage
The vigorous development of ML and DL methods implies a great opportunity to implement a technical framework to systematically identify the impacts of natural disasters, whose frequency of occurrence has increased in the recent years. One such example is depicted in Figure 15. The 2011 tsunami event accumulated numerous tsunami damage and related remote sensing observation datasets unprecedentedly [23,46,114]. The developed damage recognition model expect to utilize for the future remote sensing image that covered the tsunami-affected areas for detecting the tsunami damage. It is, however, important to consider the acquisition conditions in order to transfer what was learned from the 2011 tsunami event. The incident angle of the satellite sensor, for example, has significant effect in the microwave image analysis. Another example is the effect of the time acquisition in the brightness of optical images. In order to avoid the effect of the acquisition conditions, most of the studies avoids the use of previous events and focus on the current event.
In this context, supervised ML and deep learning are very limited by the amount of training data and the capabilities to extract them in the aftermath of a large-scale disaster. Thus, studies on automatic collection of training samples is currently an important subject. Recent innovations in deep learning represented by transfer learning [115] have reduced dependence on massive disaster sample data. The synergy efforts of private sectors, e.g., the GeoAI package provided in the Microsoft Azure cloud platform that launched by Microsoft and ESRI [116], have increased capability of leveraging the deep learning and high-performance computation to achieve more efficient tsunami damage detection. There are additional issues regarding their potentials, prospects, and challenges that need to be pointed out. Data challenges come from both the damage characterization from the field survey and the remote sensing observation. The structure of affected buildings represent complex patterns, especially, the affected buildings are surrounded by debris and water [46], which increase the difficulty of damage characterization and detection. In addition, the more accurate damage detection requires observation from multi-sensors [55,117]. The resulting data are complex, with a multi-spatial resolution, spectral and temporal acquisition requiring innovative approaches. In addition, the label work of tsunami damage field survey is lagging behind remote sensing observation. Therefore labels are often highly subjective or biased toward the real damage status, limiting the effectiveness of algorithms that rely on training datasets. It is worth mentioning that, to the best of our knowledge, only the buildings with the highest damage level (i.e., collapsed and washed away buildings) have been identified with great accuracy (over about 80%). Further, previous experience in natural hazards indicates that the damage usually only account for a small portion of disaster-affected area [118,119], and therefore, the data challenges also come from the unbalance data collection. Nevertheless, it is worth mentioning that this era has provided us with a precious opportunity to explore the use of DL/ML to detect tsunami damage. Digital Globe initiated the open data program for disaster response which provides the near real-time high-resolution optical sensor observation for the 2018 Indonesia tsunami events. As a supplement, the Sentinel project from European space agency continuously conducting throughout the year in near real-time to provide the synthetic aperture radar observations for the 2018 Indonesia tsunami events.

Conclusions
We have an urgent need to obtain the best informative understanding of what extent the tsunami damage would occur. Tsunami damage estimation is both a data and technology-driven subject. Innovative discoveries increasingly will derive from the interpretation of large datasets, new developments in earth observation technologies, and procedures enabled by high-performance computing platforms. The amount of earth-observation datasets available to scientists has grown enormously, through high-resolution optical sensors and the new synthetic aperture radar sensing methods and its fusion.
Through the experiences of the past catastrophic tsunami disasters, tremendous progress was achieved to comprehend the impact of tsunami disaster. Advances made in the past events represent significant breakthrough in sophistication of mapping tsunami impacts. Many case studies of the 2011 Great East Japan (Tohoku) earthquake tsunami provided evidences regarding how earth observations, in combination with local, in-situ data and information sources, can support the decision-making process before, during, and after a tsunami disaster strikes. In particular, the use of SAR technology was mentioned in this paper where innovative damage detection methodologies were presented. We also provide evidences regarding how such remote sensing applications become better with the advances of machine learning and deep learning.
The Sendai Framework for Disaster Risk Reduction 2015-2030, that was adopted at the Third UN World Conference 2015, articulates the need to promote and enhance geospatial and space-based technologies, through international cooperation, including technology transfer, access to, sharing and use of data and information. There are still strong needs of remote sensing studies that includes establishing standardization of the methodologies, damage scale, and use of in-situ data. Sharing the outcomes of remote sensing research has to be made so that they can be easily utilized by responders and decision-makers. Discussions with users on acceptable accuracy of the results and its limitation are also the keys to achieve the goal. Therefore, verifying the results and methodologies is a critical step to achieve broader implementation of remote sensing technologies to tsunami disaster management.