Earthquake Aftermath from Very High-Resolution WorldView-2 Image and Semi-Automated Object-Based Image Analysis (Case Study: Kermanshah, Sarpol-e Zahab, Iran)

: This study aimed to classify an urban area and its surrounding objects after the destructive M7.3 Kermanshah earthquake (12 November 2017) in the west of Iran using very high-resolution (VHR) post-event WorldView-2 images and object-based image analysis (OBIA) methods. The spatial resolution of multispectral (MS) bands (~2 m) was ﬁrst improved using a pan-sharpening technique that provides a solution by fusing the information of the panchromatic (PAN) and MS bands to generate pan-sharpened images with a spatial resolution of about 50 cm. After applying a segmentation procedure, the classiﬁcation step was considered as the main process of extracting the aimed features. The aforementioned classiﬁcation method includes applying spectral and shape indices. Then, the classes were deﬁned as follows: type 1 (settlement area) was collapsed areas, non-collapsed areas, and camps; type 2 (vegetation area) was orchards, cultivated areas, and urban green spaces; and type 3 (miscellaneous area) was rocks, rivers, and bare lands. As OBIA results in the integration of the spatial characteristics of the image object, we also aimed to evaluate the efﬁciency of object-based features for damage assessment within the semi-automated approach. For this goal, image context assessment algorithms (e.g., textural parameters, shape, and compactness) together with spectral information (e.g., brightness and standard deviation) were applied within the integrated approach. The classiﬁcation results were satisfactory when compared with the reference map for collapsed buildings provided by UNITAR (the United Nations Institute for Training and Research). In addition, the number of temporary camps was counted after applying OBIA, indicating that 10,249 tents or temporary shelters were established for homeless people up to 17 November 2018. Based on the total damaged population, the essential resources such as emergency equipment, canned food and water bottles can be estimated. The research makes a signiﬁcant contribution to the development of remote sensing science by means of applying different object-based image-analyzing techniques and evaluating their efﬁciency within the semi-automated approach, which, accordingly, supports the efﬁcient application of these methods to other worldwide case studies.


Introduction
An earthquake is a sudden shaking of the ground, and it is known as a natural hazard that cannot be predicted. Therefore, the only means to reduce its consequences is to retrofit buildings in the pre-occurrence phase and quickly identify the damaged areas in the post-occurrence phase. Earthquake hazards yield unfavorable damage to buildings and other manmade structures, representing a serious threat to human life [1]. Rapid damage mapping in urban areas in the early hours after a hazardous event is an important task of the disaster response phase [2][3][4]. According to the collected statistical information, people being trapped in collapsed buildings poses a high potential of human loss after any strong earthquake [5,6]. Therefore, a quick evaluation of buildings in an urban area can be helpful to estimate earthquake aftermath in order to facilitate proper shelter for injured people and to increase the post-earthquake response of rescue teams [7,8].
The recent progress in remote sensing and Earth observation technologies provided a wide range of satellite images that can be employed for many applications, such as efficient natural hazard monitoring [9,10]. Facilitating the availability of a wide range of data requires a quick and cost-effective data driven image analysis approach, such as semi-automated object-based image analysis [11]. In recent years, with the development of remote sensing technology, Earth observation-based damage assessment has been widely investigated by many researchers. Remote sensing techniques are a cost-effective and prompt way to probe a specific area and utilize the obtained information for a quick disaster response [12,13].
Applying very high-resolution (VHR) multitemporal satellite imagery is one of the most common approaches to identifying and mapping destroyed buildings using remote sensing [14,15]. However, it is also understood that the VHR images are not available at all times due to the high cost of their production, especially in the pre-earthquake phase. Therefore, innovative and semi-automated methods, such as object-based image analysis (OBIA) and deep learning methods can be applied to extract and recognize damaged buildings from a single post-event VHR image in a quick and cost-effective manner. In the context of using a single image, commonly used spatial features in building damage identification include image texture [16][17][18][19] and post-earthquake buildings' morphological characteristics [20]. Due to the complexity and diversity of building damage caused by earthquakes [19,21], applying traditional and common methods in the field of satellite image processing cannot help crisis management to correctly identify different devastated areas in an urban environment.
In the present study, semi-automated OBIA was used to classify and distinguish destroyed and damaged buildings using satellite data processing. The OBIA approach represents a promising methodology, because it has a lot in common with human perception, starting with segmenting images into homogeneous regions that almost correspond to real-world objects [22]. Technically, the various characteristics in the calculated segments, such as shape, texture, layer-based values, and the context of the object, are considered as the main factors of the classification process in the OBIA technique [23]. One of the most significant characteristics of the OBIA approach is the possibility of detecting and classifying targets bigger than pixels as image objects, which allows for the integration of a variety of spatial and spectral features, such as textural parameters, shape, neighborhood, and relations for modelling tasks [24]. Additionally, OBIA offers the ability of using the intrinsic properties of objects and the use of contextual or spatial behavior through the neighborhood or topological relationships between objects [25].
OBIA is regarded as an effective and efficient technology for VHR image classification due to its clear and intuitive technical process [26]. The OBIA approach can be effectively paired with geographic information system (GIS) techniques, allowing for a more comprehensive mapping of land use classes for GIS studies [27,28]. Moreover, compared with the pixel-based classification techniques, with the OBIA approach, the segmented objects exhibit rich spectral and textural features and provide shape and contextual information, which can improve the classification performance for various types of objects [29,30].
The use of statistical indicators, parameters, and the characteristics of segments is one of the most common operations in improving the OBIA technique for the classification of high-resolution images. Parameter optimization has been a research subject for decades, with the most recent trend being the use of an automated, optimal parameter determination technique. However, defining appropriate segmentation parameters, even for a single image, is a significant challenge [31]. Furthermore, it should be noted that OBIA requires specific parameterization for different urban patterns to produce optimal segmentation products [32].
It has to be indicated that the range of satellite sensors and the volume of remote sensing data and products has increased significantly over the past decade. Intensive progress in Earth observation technologies yielded significant improvements in the spatial, spectral, and temporal resolutions of satellite images, which is increasingly taking remote sensing into the arena of big data technology. The diversity and number of application fields and the variety of methods and methodologies to process large amounts of data have steadily increased [33][34][35]. Thus, traditional image processing and classification methods faced incentive challenges. Based on the context of novel satellite images, it is well understood that the remote sensing community demands novel and efficient data driven approaches such as semi-automated approaches. In response to this demand, the OBIA with rule-based development capability leads to developing an efficient framework which can be applied for infinitive applications. Technically, rule-based classification is an exceptional object-identifying technique that presents data mining performance by dividing image contexts into intelligent image objects which can be accordingly used as the base of semi-automated classification [36,37]. Based on these statements, the objective of the present study is to explore the performance of OBIA in building damage assessment and temporary camp detection using post-earthquake WorldView-2 VHR satellite imagery. Our goal is to apply and identify the efficiency of OBIA algorithms for the development of a semi-automated framework that can be used by future researchers in other earthquake case studies. It is well understood that the proposed approach will support future research and studies for applying efficient methods for cost-effective data driven approaches for monitoring earthquake consequences. This study can be considered as state of the art and progressive research in the domain of remote sensing application for rapid estimation of essential resources for the affected people.

Study Area
Iran is widely acknowledged as a country where many natural disasters occur every year, posing a serious threat to people's lives. For example, on 26 December 2003, a destructive earthquake (M6.5) struck Bam city with 30,000 fatalities and 20,000 injuries [38]. On 11 August 2012, Ahar-Varzaghan twin earthquakes (M6.3 and M6.4) caused more than 300 deaths [39]. On 12 November 2017 (21:48 local time) in Sarpol-e Zahab, the deadliest earthquake (M7.3) of 2017 with more than 600 casualties, 7000 injured people and 15,000 homeless was recorded [40,41]. Given the reported earthquakes, the requirement to identify devastated areas for emergency assistance to the injured is one of the most critical tasks in cities that are prone to earthquakes. Based on seismic reports, the central government authorities were not able to estimate the extent of damage during the first 72 h of aftermath of the event, because of night-time earthquake, as well as the harsh topography of the affected area and the scattered settlement areas (e.g., towns and villages). These issues increased the number of casualties and imposed irreversible socioeconomic consequences [2].
As stated, Iran is a country that faces natural disasters, such as floods, landslides, and earthquakes, every year [42][43][44][45]. In the present study, we seek to identify the destroyed buildings in Sarpol-e Zahab located in the Kermanshah province, Iran. According to reliable data released from the Sarpol-e Zahab Municipality, the population of this city is currently 45,481 people, and the city covers an area of about 544 hectares [46]. The study area is located in the west of Iran in the Kermanshah province near the Iraqi border. Sarpol-e Zahab and the surrounding urban and rural areas were severely damaged by the M7.3 earthquake, which was the strongest seismic event since 1967 near the Iran-Iraq border ( Figure 1) [47][48][49]. From the geological point of view, the M7.3 Sarpol-e Zahab earthquake occurred in the northwest Zagros fold-thrust area ( Figure 2). The Zagros Mountains have been experiencing an uplift since the late Eocene because of the tectonic collision between the Arabian and Eurasian plates [50]. According to the USGS (United States Geological Survey) earthquake catalog, only a few earthquakes ranging from M4 to M5 have been recorded for a radius >200 km of the epicenter [51].  From the geological point of view, the M7.3 Sarpol-e Zahab earthquake occurred in the northwest Zagros fold-thrust area ( Figure 2). The Zagros Mountains have been experiencing an uplift since the late Eocene because of the tectonic collision between the Arabian and Eurasian plates [50]. According to the USGS (United States Geological Survey) earthquake catalog, only a few earthquakes ranging from M4 to M5 have been recorded for a radius >200 km of the epicenter [51]. From the geological point of view, the M7.3 Sarpol-e Zahab earthquake occurred in the northwest Zagros fold-thrust area ( Figure 2). The Zagros Mountains have been experiencing an uplift since the late Eocene because of the tectonic collision between the Arabian and Eurasian plates [50]. According to the USGS (United States Geological Survey) earthquake catalog, only a few earthquakes ranging from M4 to M5 have been recorded for a radius >200 km of the epicenter [51].

Dataset
The WorldView-2 is a commercial satellite that gathers information of 8 multispectral bands. The observation cycle and width of the footprint are 1.1 days and 16.4 km, respectively [52][53][54][55]. The spatial resolution of the acquired images is 2 m for multispectral (MS) bands: coastal blue (400-450 nm), blue (450-510 nm), green (510-580 nm), yellow (585-625 nm), red (630-690 nm), red edge (705-745 nm), NIR1 (770-895 nm), and NIR2 (860-1040 nm) [52]. The spatial resolution of the panchromatic (PAN) band (450-800 nm) of the sensor is about 0.5 m. According to Digital Globe [56], these 8 bands are uniquely designed to cover various demands and applications, such as damage monitoring, coastal line delineation, environmental purposes, and resource management [53,54]. In order to identify damaged buildings and the location of the temporary shelters and tents after the earthquake, we used a single four-band VHR image from the WorldView-2 satellite. Details of the obtained VHR image are given in Table 1. To validate the outcomes of VHR image classification, we also obtained a reference map provided by the United Nations. The reference map for Sarpol-e Zahab was generated in 2017 by the UNITAR (United Nations Institute for Training and Research) using various remote sensing and field measurements.

Methodology
The spectral information from WorldView-2 was used for different indices to create rulesets for the OBIA method. The general workflow of this study is shown in Figure 3. As shown in Table 1, the spectral bands of WorldView-2 images reveal that the spatial resolution is 2 m, but for the current case, the mentioned spatial resolution is not suitable to extract urban features. The pan-sharpening technique provides a solution to boost the spatial resolution of the MS band to 0.5 m by fusing the PAN and MS images to generate pan-sharpened images. In brief, the pan-sharpened images are promoted MS images with spatial resolution that is the same as that of the PAN image [57][58][59]. Therefore, as shown in the workflow (Figure 3), we fused the panchromatic band and spectral bands of WorldView-2 to increase the spatial resolution of the image because increasing the spatial resolution of the image is extremely convenient for detecting small objects in the image, such as temporary settlement tents and small buildings, as well as for identifying different patterns of destroyed buildings [59]. After preparing the image, object-based classification was performed. After achieving the classification results using object-based technique, to examine the performance of the applied technique we evaluated the results' accuracy using ground control points related to the study area received from the UN. The kappa coefficient was used to calculate the accuracy assessment. The OBIA oriented classification was performed in the eCognition software environment, and accuracy assessment was conducted using Microsoft Excel software.

Object-Based Image Analysis for Damage Mapping
A building's characteristics, such as the geometric features of buildings, the area, rectangular fit, and convexity, yield various unexpected complexity in collapsed building detection after earthquakes [60]. Where the urban morphology is combined with an adequate image resolution to allow for recognition, OBIA approaches have shown good performance in the extraction of objects (such as roofs and roads) at the settlement level [61,62]. To consider both context and pixel values in a classification process, OBIA tech-

Object-Based Image Analysis for Damage Mapping
A building's characteristics, such as the geometric features of buildings, the area, rectangular fit, and convexity, yield various unexpected complexity in collapsed building detection after earthquakes [60]. Where the urban morphology is combined with an adequate image resolution to allow for recognition, OBIA approaches have shown good performance in the extraction of objects (such as roofs and roads) at the settlement level [61,62]. To consider both context and pixel values in a classification process, OBIA techniques are a beneficial means, because they use various items, such as texture, form, compacity, and relationships of neighborhood pixels during the segmentation and feature extraction process [63][64][65][66][67][68]. Although the main goal of the study is to detect the damaged buildings and temporary camps, other urban element were classified to provide statistical information about the city. To begin classifying using the OBIA technique, image segmentation is the first requirement. In fact, image objects/segments are obtained during image segmentation to represent the boundary of the objects. Image segmentation enables users to access reliable results within the classifying process [69][70][71][72].
Regarding the variety of scales of urban features and objects, different scales of segmentation were applied to identify the features and objects. For instance, by selecting a low segmentation scale, tiny image objects can be created, which is a process that is called over-segmentation [73]. A low segmentation scale is implemented to identify tents and temporary residents as well as preserve the feature boundaries. Some examples of differences between urban objects in various classes are presented in Figure 4. niques are a beneficial means, because they use various items, such as texture, form, compacity, and relationships of neighborhood pixels during the segmentation and feature extraction process [63][64][65][66][67][68]. Although the main goal of the study is to detect the damaged buildings and temporary camps, other urban element were classified to provide statistical information about the city. To begin classifying using the OBIA technique, image segmentation is the first requirement. In fact, image objects/segments are obtained during image segmentation to represent the boundary of the objects. Image segmentation enables users to access reliable results within the classifying process [69][70][71][72].
Regarding the variety of scales of urban features and objects, different scales of segmentation were applied to identify the features and objects. For instance, by selecting a low segmentation scale, tiny image objects can be created, which is a process that is called over-segmentation [73]. A low segmentation scale is implemented to identify tents and temporary residents as well as preserve the feature boundaries. Some examples of differences between urban objects in various classes are presented in Figure 4. For this purpose, a segmentation scale of 25 was applied for the segmentation process. In order to obtain the optimal scale of segmentation, the cadaster map and field measurement for 120 building as sample were employed. For this goal, the segmentation was performed by several scales (10,15,25,30,35) and by comparing the area of obtained image objects of 120 sample buildings with image generated in each scale, we selected the 25 as optimal scale of segmentation. The segmented features in some parts of the image were illogical, which means that the features were not distinguished completely. To solve this problem, merging operations were used in the desired parts to obtain the correct border of the features. The scale levels for segmentation and merging were chosen regarding visual inspection and trial and error, as recommended by previous studies [74,75]. The numbers were validated by visual examinations to identify the shapes and patterns of the objects. In the present study, to implement the object-based technique, the following different rulesets were used: NDVI; mean and maximum of band red, green, blue, and NIR; the brightness index; standard deviation; and shape compactness. Determining the rules For this purpose, a segmentation scale of 25 was applied for the segmentation process. In order to obtain the optimal scale of segmentation, the cadaster map and field measurement for 120 building as sample were employed. For this goal, the segmentation was performed by several scales (10,15,25,30,35) and by comparing the area of obtained image objects of 120 sample buildings with image generated in each scale, we selected the 25 as optimal scale of segmentation. The segmented features in some parts of the image were illogical, which means that the features were not distinguished completely. To solve this problem, merging operations were used in the desired parts to obtain the correct border of the features. The scale levels for segmentation and merging were chosen regarding visual inspection and trial and error, as recommended by previous studies [74,75]. The numbers were validated by visual examinations to identify the shapes and patterns of the objects. In the present study, to implement the object-based technique, the following different rulesets were used: NDVI; mean and maximum of band red, green, blue, and NIR; the brightness index; standard deviation; and shape compactness. Determining the rules depends on human experience and reasoning to achieve a specific objective [74][75][76][77]. An explanation of each of the rulesets is given below.

Normalized Difference Vegetation Index
As the cultivated area, area of orchards, and urban green space were significantly different from each other, the NDVI was used to identify green lands in the study area. The use of this index after applying segmentation with a size of 250 provided satisfactory results in detecting vegetation in the study area. This index was implemented using the following equation [76]: where NIR represents the near-infrared band (which is band 4 in our case) and R equals the red band (which is band 3 in our case).

Mean and Maximum of bands
Due to the different spectral reflections of objects in different ranges of the electromagnetic spectrum, statistical indices such as the mean and maximum reflection of used bands can be applied to distinguish objects from each other. In the present study, using the maximum reflection in the visible bands (RGB), we were able to identify and extract buildings with bright or impenetrable roofs. The use of averages in the blue band also helped to identify bright objects in the used image.

Brightness index
The brightness index distinguishes and identifies the brightest and darkest parts of the image using the values reflected from it. Applying this index allowed us to identify the shadows of buildings and trees as the darkest part of the image and the tents as the brightest objects of the image. The calculation of this index is based on the following equation [77]: where W B K is brightness weight of the image k, which is between 0 and 1. K represents the number layers of the used image (4 in our case). W B is the sum of the brightness weights of all layers of the image k used to calculate W B = ∑ K k=1 W B K , and C k (v) represents the average intensity of the image layer k of the segment v.

Standard Deviation (StdDev)
This index indicates the measurement of standard deviation of the pixels that produce an object or a segment. The calculation of this index is based on the following equation [78]: where σk(v) is the calculated StdDev for object v in image k, Pv is a set of pixels created by object v, (x, y, x, t) is the coordinates of pixels of object v, and C k represents the calculated StdDev of a pixel in object v.

Shape compactness
This index describes the compactness ratio of objects. The compression of the image objects is obtained using Equation (4) [79] by dividing the area and perimeter of the object by the total number of pixels. In this criterion, the value range of the effects is between zero and infinity, which in a satisfactory situation is equal to 1.

Accuracy Assessment
When analyzing satellite images, it is crucial for the accuracy of any classification to be assessed [73]. Therefore, we measured the accuracy of our methodology concerning its suitability for the given application (identifying devastated buildings affected by the earthquake). In this study, we assessed the accuracy of the obtained map by evaluating the overall accuracy, user accuracy, producer accuracy, and kappa coefficient. The data used in the evaluation of accuracy indices include data extracted from the image used for processing, which has a spatial accuracy of 0.5 m, as well as data received from the UNITAR, which was collected after the earthquake ( Figure 5).

Accuracy Assessment
When analyzing satellite images, it is crucial for the accuracy of any classification to be assessed [73]. Therefore, we measured the accuracy of our methodology concerning its suitability for the given application (identifying devastated buildings affected by the earthquake). In this study, we assessed the accuracy of the obtained map by evaluating the overall accuracy, user accuracy, producer accuracy, and kappa coefficient. The data used in the evaluation of accuracy indices include data extracted from the image used for processing, which has a spatial accuracy of 0.5 m, as well as data received from the UNI-TAR, which was collected after the earthquake ( Figure 5). Overall and user and producer accuracies were obtained using the data gathered from the aforementioned image. Table 2 presents how the accuracy indices are calculated.

Description Equation Components
Overall Accuracy = 1 = Overall Accuracy Pi = number of correctly classified pixels N = total number of pixels compared  Overall and user and producer accuracies were obtained using the data gathered from the aforementioned image. Table 2 presents how the accuracy indices are calculated. Table 2. Accuracy assessment indices.

Description Equation Components
Overall Accuracy OA = 1 N ∑ P i OA = Overall Accuracy P i = number of correctly classified pixels N = total number of pixels compared

Results
The present investigation was conducted to identify buildings devastated by earthquakes and the temporary camps in Sarpol-e Zahab. For this purpose, a four-band single image from the WorldView-2 satellite was used, which was taken after the earthquake. The technique used to identify the urban objects was the OBIA method. In order to increase the spatial resolution of the used image, we pan-sharpened the image using QGIS software. Performing pan-sharpening had a huge impact on increasing the spatial accuracy to identify the urban objects rather than the simple image ( Figure 6).

Results
The present investigation was conducted to identify buildings devastated by earthquakes and the temporary camps in Sarpol-e Zahab. For this purpose, a four-band single image from the WorldView-2 satellite was used, which was taken after the earthquake. The technique used to identify the urban objects was the OBIA method. In order to increase the spatial resolution of the used image, we pan-sharpened the image using QGIS software. Performing pan-sharpening had a huge impact on increasing the spatial accuracy to identify the urban objects rather than the simple image ( Figure 6).

Object-Based Image Analysis
After the pre-processing step, to identify the devastated buildings and the camps, the OBIA method was utilized. Segmentation is a significant step in OBIA. It can be considered as a process of dividing an image into homogeneous and non-overlapping zones that are then identified as objects [69]; therefore, to segment the image used in the present study, the multiresolution method was used in which the small segments are merged. Due to the irregular pattern of the destroyed buildings, the segmentation process of these features requires a high precision. Therefore, in the obtained segmentation, all of the segments related to the destroyed buildings were merged with each other. To define the appropriate rulesets, five functions, namely, shape compactness, standard deviation (StdDev), brightness index, mean and maximum of bands, and normalized difference vegetation index (NDVI), were utilized in the present study. Technically, each of the used functions and rules demonstrated reliable performance in identifying certain features. For instance, the shape compactness is the most suitable ruleset for identifying irregularly and disordered patterns of the demolished buildings, and high accuracy has been observed when identifying vegetation with the NDVI.
The obtained results from the remote sensing and satellite image processing techniques have not demonstrated considerable accuracy. Utilizing object-based methods has somewhat solved the issues of uncertainty of these methods. In the present study, different indices were used to classify objects with the same statistical properties. The spectral

Object-Based Image Analysis
After the pre-processing step, to identify the devastated buildings and the camps, the OBIA method was utilized. Segmentation is a significant step in OBIA. It can be considered as a process of dividing an image into homogeneous and non-overlapping zones that are then identified as objects [69]; therefore, to segment the image used in the present study, the multiresolution method was used in which the small segments are merged. Due to the irregular pattern of the destroyed buildings, the segmentation process of these features requires a high precision. Therefore, in the obtained segmentation, all of the segments related to the destroyed buildings were merged with each other. To define the appropriate rulesets, five functions, namely, shape compactness, standard deviation (StdDev), brightness index, mean and maximum of bands, and normalized difference vegetation index (NDVI), were utilized in the present study. Technically, each of the used functions and rules demonstrated reliable performance in identifying certain features. For instance, the shape compactness is the most suitable ruleset for identifying irregularly and disordered patterns of the demolished buildings, and high accuracy has been observed when identifying vegetation with the NDVI.
The obtained results from the remote sensing and satellite image processing techniques have not demonstrated considerable accuracy. Utilizing object-based methods has somewhat solved the issues of uncertainty of these methods. In the present study, different indices were used to classify objects with the same statistical properties. The spectral reflection of various objects, especially in the urban environments, is different. That is, every single class has a different spectral signature to that of other classes; therefore, we analyzed all classes considering their statistical and spectral attributes. Figure 7 shows some of the computed results. It is well understood that object-based classification is based on objects generated by the segmentation. Moreover, after the trial-and-error period, corrections, such as the merging of some segments, are required. reflection of various objects, especially in the urban environments, is different. That is, every single class has a different spectral signature to that of other classes; therefore, we analyzed all classes considering their statistical and spectral attributes. Figure 7 shows some of the computed results. It is well understood that object-based classification is based on objects generated by the segmentation. Moreover, after the trial-and-error period, corrections, such as the merging of some segments, are required. The destroyed buildings in the image obtained from the WV-2 satellite do not have a proper and recognizable shape. The only elements that are effective in detecting them are the high density of the pixels and their heterogeneity. Accordingly, we used segmentation with a size of 40, and then the objects associated with the destroyed buildings were identified. After performing this step, these objects were merged in a class using the shape index. Other indices were not appropriate for the detection of the destroyed buildings. For instance, in Figure 7a, the brightness index was compared to the shape density index, The destroyed buildings in the image obtained from the WV-2 satellite do not have a proper and recognizable shape. The only elements that are effective in detecting them are the high density of the pixels and their heterogeneity. Accordingly, we used segmentation with a size of 40, and then the objects associated with the destroyed buildings were identified. After performing this step, these objects were merged in a class using the shape index. Other indices were not appropriate for the detection of the destroyed buildings. For instance, in Figure 7a, the brightness index was compared to the shape density index, and, eventually, the shape density index displayed better performance in identifying destroyed buildings. Furthermore, temporary camps after the earthquake are among the critical areas in which to provide aid to the injured people, and it is necessary to correctly identify these places. Since the tents allocated to temporary residence were in white, it was easy to identify them. Nevertheless, the brightness index cannot be used for this purpose, because this index distinguishes all high-brightness entities; therefore, the probability of uncertainty in detecting tents increases. For this purpose, a segment with a size of 20 and the standard deviation of the NIR band were used (Figure 7b). As the water bodies in the NIR band are dark, it is easy to identify the river. However, in urban environments, there is a problem with the complexity of objects, so class extraction is not easy. For this purpose, both the spectral (separation of the river from other long objects) and compression coefficient (separation of the river from other regular objects) thresholds were used (Figure 7c). The application of the normalized differential vegetation index (NDVI) is one of the most commonly used methods of extracting vegetation. In the present study, this operation was performed using red and near-infrared bands. Eventually, three types of vegetation were identified (Table 3 and Figure 7d). Moreover, the standard deviation of the red band was used to extract the residential areas. All performed statistical indices were based on trial and error in eCognition software (Figure 7e).
After implementing the defined rulesets, an identified class map was extracted with emphasis placed on identifying temporary camps and destroyed buildings as well as undestroyed buildings. In this study, nine classes were identified and mapped. The area of each land use was calculated in ha. Agricultural lands and urban buildings account for the largest area of obtained classes (Figure 8). The area of each obtained class was also calculated in the ArcMap software environment. The area of each class was calculated in ha. Agricultural lands and urban buildings account for the largest areas of obtained LULC (land use land cover) (Figure 9), while the area of destroyed buildings is equal to 54.75 ha. Since the infrastructure area of buildings in Sarpol-e Zahab city is on average equal to 90 square meters, according to this calculation, it can be estimated that about 6083 buildings were destroyed. The area of each obtained class was also calculated in the ArcMap software environment. The area of each class was calculated in ha. Agricultural lands and urban buildings account for the largest areas of obtained LULC (land use land cover) (Figure 9), while the area of destroyed buildings is equal to 54.75 ha. Since the infrastructure area of buildings in Sarpol-e Zahab city is on average equal to 90 square meters, according to this calculation, it can be estimated that about 6083 buildings were destroyed.

Accuracy Assessment
As previously stated, in the present study, four indices were used to evaluate the accuracy of the final results. Table 4 demonstrates the calculation error matrix for all classes. Generally, user's accuracy and producer's accuracy (among classes) were 92.92% and 92.94% for all LULC, respectively. Using the data obtained from the error matrix, the calculated overall accuracy for the acquired map was 94.26%, which is a reliable rate. Moreover, according to the purpose of the present study, the calculated producer accuracy for the class of destroyed buildings was 99.17%, and the user accuracy obtained for that was 95.33%, revealing a high rate of reliability (Table 4 and Figure 10). The kappa coefficient was determined, which is one of the most commonly used indices to compute the accuracy of satellite image classification results. In this regard, field data collected by the United Nations after the earthquake were used. The results showed that the obtained map presents a kappa coefficient of 94.05%.

Accuracy Assessment
As previously stated, in the present study, four indices were used to evaluate the accuracy of the final results. Table 4 demonstrates the calculation error matrix for all classes. Generally, user's accuracy and producer's accuracy (among classes) were 92.92% and 92.94% for all LULC, respectively. Using the data obtained from the error matrix, the calculated overall accuracy for the acquired map was 94.26%, which is a reliable rate. Moreover, according to the purpose of the present study, the calculated producer accuracy for the class of destroyed buildings was 99.17%, and the user accuracy obtained for that was 95.33%, revealing a high rate of reliability (Table 4 and Figure 10). The kappa coefficient was determined, which is one of the most commonly used indices to compute the accuracy of satellite image classification results. In this regard, field data collected by the United Nations after the earthquake were used. The results showed that the obtained map presents a kappa coefficient of 94.05%.

Human Settlement in Temporary Camps
One of the most important measures to reduce post-earthquake stress and concern is to provide temporary and safe housing and other essential demands for people whose houses have been destroyed. Thus, an object-based VHR image analysis will enable us to

Human Settlement in Temporary Camps
One of the most important measures to reduce post-earthquake stress and concern is to provide temporary and safe housing and other essential demands for people whose houses have been destroyed. Thus, an object-based VHR image analysis will enable us to estimate from "A" to "Z" for a proper disaster response. In the present study, one of the objectives was to identify the camps and tents provided in the study area. The results revealed that about 10,249 tents were established after the event. The estimated number of the tents is in a good agreement with the report of the Iranian Red Crescent Society for the city of Sarpol-e Zahab that stated that the number of distributed tents was about 10,000 right after the earthquake. The capacity of the tents distributed to provide shelter for the people of Sarpol-e Zahab was an average of four people per tent. According to the total population of Sarpol-e Zahab, which was equal to 45,481 people, about 40,996 people had temporarily settled and about 4485 people were homeless and required aid and tents. Figure 11 presents the size of each tent at the classified segmentation as well as a photo taken by the authors at the earthquake site. Furthermore, based on the empirical relationships gathered for previous Iranian earthquakes, essential demands of the damaged population, such as emergency toilets and baths, blankets, and canned food, are estimated to support more flexible and resilient disaster response according to Table 5.

Discussion
After an earthquake, the most crucial task is to quickly identify destroyed buildings so that immediate aid can be provided. The average revisit interval of WV-2 is 1.1 days,

Discussion
After an earthquake, the most crucial task is to quickly identify destroyed buildings so that immediate aid can be provided. The average revisit interval of WV-2 is 1.1 days, and it is capable of collecting up to 1 million km 2 Earth images daily, but the revisit interval of most optical satellite missions varies from days to weeks, which makes the prompt damage response difficult. High-resolution satellite data are usually commercial and limited to certain companies, which has become a problem in conducting a high-quality examination of the technique. The image used in the present study was a four-band image of the WV-2 satellite. To increase its spatial resolution, the panchromatic band of this satellite with a resolution of 0.5 m was applied, which supported us in applying a variety of object-based features and identifying properly distinguishable objects. Despite the enhancements, it is still difficult to identify some features. For instance, identifying destroyed buildings that do not have a regular shape or pattern, continuing the path of a river that is covered by a canopy of trees, and determining objects such as cars or small buildings that are covered by the shadow of towers and larger buildings remain challenges. Therefore, it must be noted that, in addition to the accuracy of the techniques and applied materials, user creativity is also required. Additionally, the segmentation process in the OBIA method of classification through the merging of adjacent pixels with similar characteristics and the formation of image objects minimized the pixel-based errors during classification [11]. Moreover, the lack of spectral bands in very high spatial quality images is a limitation of spectral indices. For instance, SWIR bands and some related indices might help in identifying complex features.
This study was a novel attempt to provide useful information from VHR optical images from "A" to "Z" for rescue teams using a cost-effective method. Here, we demonstrated the feasibility of VHR optical images not only for the detection of damages and changes but also for the provision of useful information regarding the number of affected people and their initial demands. This cost-effective, semi-automated method presented here will enable decision makers to predict consequences of an event and estimate the demands of the affected people. The detection of objects in the urban environment is a challenging problem in the field of optical remote sensing image processing, especially when it is required to distinguish the specific entities. Therefore, the use of an enhanced object-based technique facilitates the detection of the aftermath of urban elements, such as temporary tents and collapsed buildings. The presented semi-automated approach showed a satisfactory result for the detection of the urban classes, but not for the site specifications. Further analysis must be conducted to examine this method's transferability to other case studies. Indeed, transferability is an element of research that, in many cases, allows knowledge previously gained while solving a problem to be accordingly applied to a different case study with almost the same conditions, such as an earthquake in a different geographic location. In future work, a joint focus on object-based image analysis and deep learning algorithms might satisfy the transferability problem much more easily.

Conclusions
We aimed to identify urban objects using the OBIA method with emphasis placed on the extraction of earthquake-damaged buildings in the city of Sarpol-e Zahab, Iran. Because of inherent difficulties of the remote sensing techniques, we cannot classify the degree of damage as defined by EMS-98 (European Macroseismic Scale). Accordingly, we only detected and validated totally collapsed buildings. Disaster response using only optical images is not complete, even when VHR images are available. To decrease data latency, complementary datasets are required. For instance, synthetic aperture radar (SAR) imagery is independent of weather conditions in day or night. Another advantage of SAR data is the use of phase and intensity information of complex images in different polarizations, which is helpful for the detection of different mechanisms of urban and non-urban objects. Double-bounce information can quickly provide the location of builtup areas, while its decrease comparing the pre-event images can reveal the location of damaged buildings. As a complementary task, combining optical and SAR imagery is a crucial task for a thorough aftermath analysis. Although there are several datasets for both optical and SAR missions, in order to attain an effective disaster response, the priority should be given to those with shorter revisit intervals [81][82][83]. Accordingly, differential methods (with pre-events and post-events) may provide accurate results as described in previous studies [83,84]. Since the use of OBIA for an area struck with disaster in the night remains a question, the acquisition of VHR polarimetric SAR information in X, C, and L bands is recommended.
Due to the wide variety of methods and techniques, it is necessary to identify those that are efficient [9][10][11]. From the methodological perspective, results of this study showed that a rule-based OBIA can be considered as an efficient data driven approach in the domain of remote sensing. However, we also concluded that this approach with effective spectral and spatial algorithms as well as integration remote sensing and GIScience can be considered as an efficient modelling method. We expect that integration of OBIA with novel image classification methods such as deep learning will even increase the efficiency of OBIA and provide a robust image analysis methodology which will be considered in our future works. The main objective of this research was to use OBIA and identify efficient object-based methods. The results of this research suggest that integrated semi-automated spatial, geometrical, and spectral features can be applied to monitor building damages as the result of earthquakes. Based on the research outcome, we conclude that the current study can be considered a piece of progressive research in remote sensing sciences due to the proposal of an efficient approach. The results will support authorities and decision makers to gather the required information in the post-disaster step for crisis management and mitigation objectives.