Next Article in Journal
The Impact of Agricultural Labor Policies on Agricultural Enterprises: Evidence from Türkiye
Previous Article in Journal
Sustainability Assessment of Decentralized Hybrid Rainwater–Graywater Systems for Water Management in Arid and Semi-Arid Regions: A Systematic Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Automated Hurricane Damage Classification for Sustainable Disaster Recovery Using 3D LiDAR and Machine Learning: A Post-Hurricane Michael Case Study

1
Moss School of Construction, College of Engineering and Computing, Florida International University, Miami, FL 33199, USA
2
Department of Earth and Environment, Florida International University, Miami, FL 33199, USA
3
Department of Chemistry, Busitema University, Tororo P.O. Box 236, Uganda
*
Author to whom correspondence should be addressed.
Sustainability 2026, 18(1), 90; https://doi.org/10.3390/su18010090 (registering DOI)
Submission received: 12 November 2025 / Revised: 17 December 2025 / Accepted: 19 December 2025 / Published: 21 December 2025
(This article belongs to the Section Sustainable Urban and Rural Development)

Abstract

Accurate mapping of hurricane-induced damage is essential for guiding rapid disaster response and long-term recovery planning. This study evaluates the Three-Dimensional Multi-Attributes, Multiscale, Multi-Cloud (3DMASC) framework for semantic classification of pre- and post-hurricane Light Detection and Ranging (LiDAR) data, using Mexico Beach, Florida, as a case study following Hurricane Michael. The goal was to assess the framework’s ability to classify stable landscape features and detect damage-specific classes in a highly complex post-disaster environment. Bitemporal topo-bathymetric LiDAR datasets from 2017 (pre-event) and 2018 (post-event) were processed to extract more than 80 geometric, radiometric, and echo-based features at multiple spatial scales. A Random Forest classifier was trained on a 2.37 km2 pre-hurricane area (Zone A) and evaluated on an independent 0.95 km2 post-hurricane area (Zone B). Pre-hurricane classification achieved an overall accuracy of 0.9711, with stable classes such as ground, water, and buildings achieving precision and recall exceeding 0.95. Post-hurricane classification maintained similar accuracy; however, damage-related classes exhibited lower performance, with debris reaching an F1-score of 0.77, damaged buildings 0.58, and vehicles recording a recall of only 0.13. These results indicate that the workflow is effective for rapid mapping of persistent structures, with additional refinements needed for detailed damage classification. Misclassifications were concentrated along class boundaries and in structurally ambiguous areas, consistent with known LiDAR limitations in disaster contexts. These results demonstrate the robustness and spatial transferability of the 3DMASC–Random Forest approach for disaster mapping. Integrating multispectral data, improving small-object representation, and incorporating automated debris volume estimation could further enhance classification reliability, enabling faster, more informed post-disaster decision-making. By enabling rapid, accurate damage mapping, this approach supports sustainable disaster recovery, resource-efficient debris management, and resilience planning in hurricane-prone regions.

1. Introduction

Hurricanes are among the most destructive natural hazards globally, producing widespread structural damage and generating millions of cubic yards of debris [1,2,3]. Effective debris management is essential for minimizing risks to public safety, restoring infrastructure functionality, and accelerating recovery timelines [4]. Climate change is intensifying the frequency and severity of Atlantic hurricanes, placing greater demands on disaster response systems [5,6]. In 2024, the Atlantic hurricane season recorded 18 named storms, 11 hurricanes, and 5 major hurricanes [7]. A notable example is Hurricane Michael in 2018, which generated an estimated 37.4 million cubic yards of debris and triggered over $1 billion in Federal Emergency Management Agency (FEMA) Public Assistance grants, mirroring the impact patterns of past high-intensity hurricanes such as Katrina [8]. As such events become more frequent and severe, the development of predictive, automated debris detection systems is critical to improving resource allocation, accelerating clearance operations, and reducing logistical bottlenecks in post-disaster environments [9,10].
The evolution of debris estimation methods has progressed from labor-intensive manual surveys, often inconsistent due to subjectivity and access constraints, to aerial photography and satellite imagery, which expanded coverage and reduced human error [9]. The integration of GIS-based tools improved spatial analysis, and advances in LiDAR, UAV photogrammetry, and terrestrial laser scanning enabled high-resolution, rapid data collection for post-hurricane assessments [11,12]. More recently, multispectral and hyperspectral imaging have enhanced material differentiation, further refining debris classification [13,14]. FEMA has also advanced its debris assessment capabilities through models such as Hazus-MH, which uses storm history, terrain data, and structural vulnerability assessments to estimate debris volumes [15,16]. While effective at a regional scale, these models struggle to capture fine-scale structural complexity and the rapid environmental changes that occur immediately after a hurricane [17].
Artificial intelligence, particularly machine learning and deep learning, offers transformative potential for debris detection by combining automation with high classification accuracy [18]. Convolutional neural networks (CNNs) and other deep learning methods applied to imagery and 3D point clouds have demonstrated superior recall and processing efficiency compared to manual inspection or traditional feature-based models [19,20]. Integrating AI with LiDAR point clouds has improved debris and partially collapsed structure detection, increasing recall by 20–30% over conventional methods [21,22]. Despite these advances, a clear research gap remains because few operational workflows systematically combine multiscale point cloud feature extraction with supervised learning to produce high-fidelity, transferable debris classification across spatially distinct disaster zones.
Compared to conventional LiDAR classification workflows and deep learning approaches, the Three-Dimensional Multi-Attributes, Multiscale, Multi-Cloud (3DMASC) framework offers several advantages. First, it computes over 80 multiscale geometric, radiometric, and echo-based features, enabling robust characterization of complex disaster environments [23]. Second, its multi-cloud capability supports direct comparison of co-registered datasets from different sensors, preserving spectral and sampling characteristics [23]. Third, the method is interpretable, with feature importance rankings and SHapley Additive exPlanations (SHAP) analysis, both of which are often lacking in black-box machine learning and deep learning models [23,24]. These attributes make 3DMASC particularly suitable for disaster response scenarios where transparency, scalability, and adaptability to varying data sources are critical.
This study addresses this gap by developing and evaluating a machine learning-based workflow for hurricane debris classification using the 3DMASC framework in CloudCompare. The approach leverages a Random Forest classifier trained on bi-temporal LiDAR data to automate semantic labeling of both stable and damage-specific features, aiming to provide scalable and precise support for post-hurricane response operations. The primary hypothesis is that multiscale feature extraction combined with supervised learning can classify hurricane debris and structural damage from LiDAR point clouds with accuracy comparable to that achieved for persistent features. The guiding research question is whether a 3DMASC–Random Forest framework applied to bi-temporal LiDAR data can deliver accurate and transferable debris classification suitable for operational disaster response.
This work demonstrates the potential to significantly shorten disaster assessment timelines from weeks to hours, providing rapid situational awareness to prioritize high-impact clearance zones and support resource allocation. While the framework is highly effective for mapping stable features, additional refinements are needed for detailed damage classification before full operational deployment.

2. Methodology

2.1. The 3DMASC Framework

The 3DMASC [23] framework was used for semantic classification of 3D point cloud data. Implemented as an open-source plugin in CloudCompare, 3DMASC calculates over 80 handcrafted features, including geometric descriptors such as curvature and roughness, radiometric statistics such as intensity mean and standard deviation, echo-based metrics, and dimensionality measures from principal component analysis. These features are computed at multiple spherical neighborhood scales, allowing the model to capture both fine and broad structural patterns. The “Multi-Cloud” capability enables direct comparison between co-registered point clouds from different sensors, preserving their spectral and sampling characteristics while supporting ratio and difference computations. Classification is performed with a supervised Random Forest model trained on a labeled subset of points, typically with no more than 2000 samples per class. Output includes per-point semantic labels, probability scores, and feature importance rankings, with interpretability supported through SHAP analysis.

2.2. Study Area and Data Sources

The study focuses on Mexico Beach (Figure 1), a small coastal city in Bay County, Florida, USA (centered approximately at 29.94° N, 85.41° W), located along the Gulf of Mexico. The area lies on a low-lying coastal plain with elevations generally below 10 m above sea level. Sandy soils, sparse coastal vegetation, and a mix of residential, commercial, and recreational infrastructure characterize it. Its shoreline consists of sandy beaches backed by dune systems, interspersed with small inlets and marshes. The region is highly exposed to tropical cyclone impacts due to its location within the Gulf’s hurricane corridor.
On 10 October 2018, Hurricane Michael made landfall near Mexico Beach as a Category 5 storm with sustained winds of approximately 260 km/h and storm surge heights exceeding 4 m in some areas. The event resulted in catastrophic structural damage, widespread vegetation loss, and the deposition of substantial debris across the urban and natural landscape. Post-event surveys estimated more than 37 million cubic yards of hurricane-generated debris within the affected counties, with Mexico Beach among the most severely impacted locations.
The classification experiment utilized two bi-temporal topo-bathymetric LiDAR datasets to characterize pre- and post-hurricane conditions. The pre-hurricane dataset, “2017 Northwest Florida Water Management District (NWFWD) LiDAR: Lower Choctawhatchee,” was acquired by the NWFMD between 9 April and 17 May 2017. The post-hurricane dataset, “2018 U.S. Army Corps of Engineers (USACE) FEMA Post-Michael Topobathy LiDAR: Florida Panhandle,” was collected by the USACE between 24 October and 4 November 2018. Both datasets were obtained from the National Oceanic and Atmospheric Administration [7] and delivered in the Log ASCII Standard [19] format, containing XYZ coordinates, intensity values, and return numbers, but no Red, Green, or Blue (RGB) attributes.
To support visual interpretation and aid manual labeling, high-resolution aerial imagery was integrated into the workflow. Post-event imagery was sourced from the “2018 NOAA National Geodetic Survey (NGS) Disaster Supplemental Survey (DSS) Natural Color 8-Bit” dataset, while pre-event reference imagery was obtained from the “2017 NOAA NGS Oblique Natural Color 8-Bit” dataset. These datasets provided visual confirmation of structural conditions and debris presence where LiDAR features alone were insufficient.

2.3. Experimental Design and Data Preprocessing

Two non-overlapping areas were defined for training and testing. Zone A (training) covered approximately 2.37 km2, bounded by coordinates Xmin = −9,509,953.48, Ymin = 3,496,203.85 and Xmax = −9,508,209.76, Ymax = 3,497,560.60. Zone B (testing) covered approximately 0.95 km2, bounded by Xmin = −9,506,296.45, Ymin = 3,493,905.81 and Xmax = −9,505,274.10, Ymax = 3,494,851.72. Both zones shared similar environmental characteristics and urban structures, enabling the classifier to generalize effectively. The larger training zone was chosen to ensure exposure to diverse hurricane damage patterns and background features.
All point clouds were transformed into a standard projected coordinate system. Statistical outlier removal was applied to eliminate noise, and ground points were separated from off-ground points using a progressive morphological filter to separate ground from non-ground LiDAR returns, following the algorithm implementation described by [25]. The pre- and post-hurricane datasets were co-registered using the iterative closest point ICP) alignment tool in CloudCompare to ensure spatial consistency between epochs. ICP is a widely used method for aligning 3D point clouds by minimizing the distance between corresponding points [26].

2.4. Semantic Labeling

The training zone was subdivided into 100 uniform grids, each approximately 154 m × 154 m (23,700 m2) in size, to facilitate systematic sampling for semantic labeling. These grids were used solely to organize the selection of training data and to ensure spatial diversity, not to represent LiDAR acquisition resolution. The LiDAR point density remained consistent with the original airborne survey specifications. Within each grid, semantic labeling was performed by integrating LiDAR-derived features with the auxiliary imagery. Ten semantic classes were defined based on relevance to hurricane damage detection and their distinguishable geometry in 3D point clouds: Ground, Road, Vegetation, Building, Ocean/Water, Vehicle, Fence, Power Line, Damaged Building, and Debris. Descriptions of these classes are provided in Table 1. A maximum of 2000 points per class was used to maintain balanced training samples across categories with vastly different abundances. This threshold was selected after preliminary tests showed that increasing the sample count beyond 2000 per class did not improve classification accuracy but significantly increased training time. No explicit oversampling, cost-sensitive weighting, or synthetic data augmentation was applied; instead, we relied on the per-class cap to partially mitigate extreme class imbalance while preserving a simple, reproducible workflow. This threshold aligns with prior LiDAR classification studies [27,28]. It mitigates the extreme imbalance between stable classes (millions of points) and rare damage classes, preventing the Random Forest classifier from being dominated by overrepresented categories.

2.5. Feature Extraction

Feature computation in 3DMASC followed a standardized naming convention: [FeatureType]_[Scale]_[Statistic], where “Scale” refers to the spherical neighborhood diameter. The features selected for each class are shown in Table 2 and were chosen based on a combination of prior literature [29,30] and empirical evaluation using feature importance scores from preliminary Random Forest runs. Features with consistently high importance and low redundancy were retained for each class. This format allowed precise control over feature selection and supported multiscale analysis. A complete description of all 3DMASC features used in this study is provided in Supplementary Table S1.

2.6. Model Training, Classification, and Performance Evaluation

The Random Forest classifier was trained on the labeled dataset from Zone A (Figure 2). Cross-validation in Zone A was conducted using a 5-fold scheme, with the manually labeled grid cells partitioned into spatially disjoint folds to avoid spatial autocorrelation between training and test samples. In each fold, 80% of the points were used to train the Random Forest model, and 20% were used for validation. The final trained model was then applied to the independent test zone (Zone B) for evaluation. Classification performance was assessed using a confusion matrix, overall accuracy, and per-class precision, recall, and F1-scores.

2.7. Assumptions

The pre-hurricane airborne LiDAR was acquired in 2017, and the post-hurricane collection was conducted several weeks after Hurricane Michael in 2018. To quantify and address this temporal mismatch, we restricted the training labels to features and structures known to be stable across acquisitions (e.g., ground, water, buildings, power lines). We inspected all training tiles using high-resolution imagery to ensure consistency. Damage-related labels were manually verified against post-event National Agriculture Imagery Program (NAIP) and NOAA imagery to minimize mislabeling. Although temporal mismatch introduces unavoidable uncertainty in the exact geometry of vegetation and small objects, our approach mitigates its influence on classification outcomes.

3. Results and Discussion

3.1. Pre-Hurricane Classification Performance

The pre-hurricane classification focused exclusively on stable, persistent land cover and infrastructure categories, as damage-related classes were absent before Hurricane Michael (Figure 3). Using bi-temporal LiDAR data processed through the 3DMASC framework, the Random Forest classifier achieved an overall classification accuracy of 0.9711 when applied to approximately 20 million points in the training zone. These points correspond to the complete LiDAR coverage of Zone A after preprocessing and co-registration. The distribution follows the original acquisition pattern of the airborne LiDAR survey, which is generally uniform along flight lines. No random sampling was applied; instead, semantic labeling was performed on systematically defined grids as described in Section 2.4. These points were distributed across six semantic classes: ground, road, vegetation, water, power lines, and other permanent structures. The spatial configuration of the pre-hurricane sampling is shown in Figure 3.
To illustrate the classification output in the pre-hurricane setting, Figure 4 presents a side-by-side view of Zone A, showing the original pre-hurricane image alongside the corresponding segmented classification produced by the 3DMASC–Random Forest model. The reference image (Figure 4a) provides the unclassified visual context of the study area. In contrast, the segmented map (Figure 4b) depicts the predicted semantic classes, color-coded according to the class IDs defined in Table 1. This visual comparison shows the model’s ability to delineate stable features such as ground, water, and power lines, while also revealing localized areas of confusion between similar classes, particularly along boundaries between ground, roads, and vegetation.
The corresponding confusion matrix (Table 3) demonstrates consistently high performance for most stable classes. Ground (Class 2) achieved a precision of 0.998 and an F1-score of 0.86, although recall (0.75) was reduced due to confusion with roads and vegetation. Such misclassification is a well-documented challenge in LiDAR-based urban mapping, where similar elevation profiles and reflectance values between ground and adjacent surfaces can blur semantic boundaries. Previous work by [24] has shown that while multiscale feature extraction can mitigate these effects, it cannot eliminate class overlap in heterogeneous urban environments.
Roads (Class 3) displayed substantial precision (0.943) but extremely low recall (0.093), indicating that many true road segments were misclassified as ground (Table 3). This underdetection likely stems from surface blending caused by sediment deposition, pavement degradation, or occlusion by overhanging vegetation. A 2025 study by [31] found similar issues in satellite-based road damage detection, where debris and vegetation significantly reduced model performance.
Vegetation (Class 4) was underrepresented in the training dataset, resulting in poor precision and a near-zero F1-score. Although recall reached 0.67, vegetation was frequently misclassified as ground, particularly in sparsely vegetated or flat areas. This misclassification is consistent with findings by [32], who showed that low-stature vegetation and grasslands often return LiDAR signals similar to bare earth, especially in semiarid regions or sloped terrain. The absence of complementary RGB or multispectral data further magnifies this effect, making it challenging to distinguish vegetation from surrounding ground surfaces.
Water (Class 6) achieved near-perfect classification performance (F1-score: 0.99), aided by its flat geometry and consistently low-reflectance intensity values. These attributes are well-documented in LiDAR-based coastal and river mapping. Ref. [33] demonstrated that water surfaces exhibit uniform radiometric and geometric profiles, making them highly distinguishable in airborne laser bathymetry.
Power lines (Class 11) recorded high precision (0.99) and recall (0.93), yielding an F1-score of 0.96 (Table 3). Nonetheless, the confusion matrix revealed that some power line points were misclassified as ground (Class 2). This error is likely to occur where cables sag close to the ground or are partially obscured by nearby structures or vegetation, reducing their geometric distinctiveness. Ref. [34] reported similar challenges in airborne LiDAR filtering, noting that thin, elevated features such as power lines are difficult to detect in cluttered environments, especially when surrounded by vegetation or complex urban surfaces.
As expected, no predictions were made for damage-related classes, Damaged Buildings (Class 17) and Debris (Class 18), in the pre-hurricane dataset. All values for these categories were recorded as non-applicable (NaN), confirming that the classifier did not erroneously assign damage labels to non-impacted areas.

3.2. Post-Hurricane Classification Performance

Following Hurricane Michael, the classification task was expanded to include damage-specific and transient classes in addition to the stable features assessed in the pre-hurricane analysis. The post-hurricane sampling design is shown in Figure 5, with the trained Random Forest classifier applied to the independent test zone (Zone B). The corresponding post-disaster confusion matrix (Table 4) shows an overall classification accuracy of 0.97, indicating the robustness of the 3DMASC framework in segmenting a diverse set of semantic classes in a highly complex, post-disaster environment.
Stable classes such as ground, roads, vegetation, buildings, and water retained consistently high performance, with precision and recall scores exceeding 0.95 in all cases. Ocean/water (Class 6) achieved near-perfect classification (F1-score: 0.99), reflecting its distinctive planar geometry and low-reflectance intensity profile. The work by [33] on topo-bathymetric LiDAR confirms that water surfaces are highly separable due to their flat geometry and low radiometric reflectance, especially in shallow coastal and riverine environments. These properties hold across both pre- and post-event datasets, making water classification highly reliable.
In contrast, performance declined for damage-related or small, transient classes. Vehicles (Class 7) recorded the lowest recall (0.13), with many instances misclassified as ground or buildings. This limitation is standard in LiDAR-based urban mapping, particularly when small objects are underrepresented in the training dataset or lack distinctive geometric features. Ref. [35] reported similar challenges in detecting compact urban targets, noting that vehicles often suffer from low recall in object detection models due to occlusion and poor point density in dense urban scenes.
Damaged buildings (Class 17) achieved an F1-score of 0.58, with frequent confusion between debris and intact buildings. This disparity reflects the inherent structural ambiguity of post-disaster environments: partially collapsed roofs, fragmented walls, and debris piles often produce irregular or incomplete LiDAR returns that resemble other disturbance classes. Point cloud sparsity over rubble, combined with the absence of RGB attributes, further limits the classifier’s ability to capture fine-scale damage signatures. Class imbalance also contributes to lower performance, as damaged buildings were substantially underrepresented in the training data compared with intact structures. For instance, ref. [36] found that accurately separating damaged buildings from debris requires high-resolution 3D segmentation and contextual modeling, particularly when structural boundaries are irregular or obscured.
Debris (Class 18) achieved a moderate F1-score of 0.77, with most misclassifications occurring into ground and damaged building classes. This pattern reflects the inherent challenge of identifying scattered material fragments, which often lack consistent shape, structure, or elevation. Ref. [37] observed that terrain-like LiDAR returns from irregular surfaces can mimic bare earth, especially in the absence of complementary RGB or multispectral datasets.
Across these damage-related and small-object classes, two factors appear to drive the reduced performance. First, although we limited the maximum number of training samples to 2000 points per class, vehicles and partially collapsed structures occupy very small spatial extents and are represented by comparatively few independent object instances. Consequently, many local geometric configurations present in the validation area were not observed during training, a limitation consistent with prior findings on class imbalance in LiDAR-based classification workflows [5,6]. Second, the 3DMASC features associated with debris, damaged buildings, and surrounding ground often overlap because point densities over rubble are low, returns are noisy and irregular, and color information is unavailable. Similar challenges have been reported in disaster mapping studies where LiDAR-only datasets lack RGB or multispectral attributes, reducing feature separability between visually similar classes [14,37]. This combination of limited small-object samples and weak feature separation explains much of the confusion observed between vehicles, debris, and damaged buildings in the post-hurricane classification.
To evaluate the classifier’s ability to generalize beyond the training extent, we applied the model trained on Zone A to the independent post-hurricane test zone (Zone B). This validation area contains a representative mix of intact infrastructure and hurricane-induced damage, making it an effective benchmark for assessing performance on both persistent and transient classes. The classification output (Figure 6) shows that stable features, such as roads, buildings, and water bodies, were segmented with consistently high fidelity, indicating strong transferability of the model to new spatial contexts.

3.3. Environmental Complexities and Scene-Level Ambiguities

Several classification errors in the post-hurricane dataset can be traced to scene-level complexities typical of disaster-affected environments. One of the most persistent issues was confusion between roads and ground surfaces, particularly in areas affected by mudflows, sediment deposition, and debris accumulation [38]. These conditions altered both the geometric and radiometric properties of road surfaces, making them nearly indistinguishable from adjacent terrain. Similar blending effects have been documented in post-flood LiDAR analyses, where sediment-laden surfaces reduce the separability of engineered features from surrounding ground.
Another source of misclassification arose from inland water pooling on damaged land surfaces, which was sometimes labeled as ocean or other permanent water bodies. This occurred because temporary and permanent water bodies share similar flatness and low reflectance values, particularly in low-lying coastal zones where storm surge and rainfall-induced flooding persisted [39]. Such ambiguities are well documented in hydrologic remote sensing, where ephemeral water can mimic the spectral and geometric signatures of stable aquatic features.
A further challenge involved vegetated rubble, zones where fallen trees and partially collapsed structures formed hybrid configurations that confused the classifier. In these cases, buildings retained partial vertical form but were entangled with vegetation, leading to frequent mislabeling between building, vegetation, and debris classes [40]. Similar effects have been observed in landslide and wind damage mapping, where structural collapse and biomass accumulation produce mixed-class artifacts.
Temporal misalignment between LiDAR and supporting aerial imagery also introduced inconsistencies [41]. For example, structures labeled as damaged in the LiDAR dataset may have been cleared or altered by the time imagery was acquired, leading to semantic lag in the training labels. This issue is particularly problematic in rapidly evolving disaster zones and has been noted in studies comparing LiDAR- and imagery-based change detection systems.
Furthermore, mixed-class borders, such as transition zones between debris piles, damaged buildings, and intact structures, produced classification artifacts [42]. These edge effects are challenging to model, especially when transitional geometries are underrepresented in training data. Similar boundary-related errors have been reported in hybrid ensemble models for landslide susceptibility mapping, where overlapping features degrade classification accuracy. A visual example of these spatial ambiguities is provided in Figure 6, which shows the classifier’s performance in the post-hurricane validation area.
In addition, the current workflow operates purely at the point level, without explicit contextual descriptors that summarize neighborhood structure or object-level connectivity, and without explicit pre–post change metrics (e.g., elevation differences between epochs) as features. These omissions likely contribute to the difficulty of distinguishing small or partially collapsed objects whose local geometry is similar to the surrounding background, a limitation noted in previous work on multi-temporal LiDAR change detection and contextual modeling for structural damage assessment [35,36,42].
Although the framework achieved high accuracy for stable classes, the performance for damage-specific categories, debris (F1 = 0.77), and damaged buildings (F1 = 0.58), is below the threshold typically required for operational deployment in emergency management. These results indicate that the current approach is best suited for rapid preliminary assessments and situational awareness rather than detailed damage quantification. Further development, including integration of RGB/multispectral data, improved representation of small objects, and class balancing, is necessary before the method can be considered operationally ready.

3.4. Data Requirements for Applying the Workflow

Applying the 3DMASC–Random Forest framework requires bi-temporal airborne LiDAR datasets with sufficient point density to support multiscale geometric feature extraction, particularly for small or thin objects such as vehicles, fences, and power lines. Accurate co-registration between pre- and post-event point clouds is essential, as misalignment can distort curvature-, roughness-, and PCA-based descriptors. The datasets should include geometric and radiometric attributes such as XYZ coordinates, intensity, and return numbers. At the same time, RGB or multispectral imagery is not mandatory; it significantly improves class separability and supports accurate manual labeling. Consistent radiometric calibration and a uniform projected coordinate system are also crucial for maintaining feature stability across flight lines. The workflow performs best when applied to high-quality airborne LiDAR with minimal occlusion, and users working with lower-density data or variable acquisition geometries should expect reduced class separability and higher uncertainty. Finally, adequate computational resources are required to process large point clouds and efficiently extract more than 80 multiscale features.

3.5. Limitations

This study involved several data-related and methodological constraints that shape the performance and generalizability of the proposed workflow. A central limitation is the absence of RGB or multispectral information in the airborne LiDAR datasets. Without color-based cues, the classifier relied entirely on geometric and radiometric attributes, which are less discriminative in cluttered or structurally complex environments. This constraint contributed directly to several of the misclassification patterns observed after Hurricane Michael. Debris, damaged buildings, and bare ground often shared similar geometric signatures, especially in areas with sparse or irregular point sampling, leading to confusion between debris and ground, or between damaged and intact structures. The very low recall for vehicles similarly reflects the difficulty of distinguishing small objects whose shapes and intensities closely resemble those of surrounding surfaces when spectral information is unavailable.
Point cloud sparsity further limited the detectability of narrow or small features. Vehicles, fences, power lines, and other thin structures require dense point coverage for accurate representation, and the airborne LiDAR acquisitions did not consistently capture this detail. Class imbalance also affected model generalization were stable features dominated the landscape, whereas vehicles and partially collapsed structures were underrepresented both spatially and in the training dataset. Although capping the number of training points per class reduced the most extreme imbalance, it did not compensate for the low number of distinct objects available for training. More advanced balancing approaches, such as oversampling, cost-sensitive learning, or synthetic augmentation, were not applied here but may improve recognition of minority classes.
Several sources of uncertainty stem from the data alignment and labeling process. Temporal mismatches between LiDAR and auxiliary imagery meant that cleanup activities, temporary repairs, and natural processes such as erosion or vegetation die-off could alter the landscape between acquisitions, reducing correspondence between labels and actual post-event conditions. Transitional zones, such as areas where debris overlapped with damaged buildings or where vegetation mixed with structural collapse, were inherently ambiguous and underrepresented in the training data, leading to edge-related misclassifications. Manual labeling also introduced subjectivity, particularly in partially collapsed structures or vegetated rubble, where boundaries are difficult to delineate consistently.
Methodological constraints further shaped model performance. The workflow relies on pointwise 3DMASC descriptors and does not incorporate contextual, object-level, or explicit multi-temporal change features. This simplifies implementation but limits the ability to distinguish small or partially collapsed objects whose local geometry resembles nearby ground or intact structures. The Random Forest classifier, while robust and interpretable, lacks spatial reasoning and may be outperformed by graph-based methods or deep learning models in scenes with complex structural relationships. Additionally, the 3DMASC feature set was optimized for airborne topo-bathymetric LiDAR; its performance across other LiDAR modalities, such as UAV LiDAR, terrestrial laser scanning, or mobile mapping systems, remains untested and may require re-engineering of features.
Operational considerations present additional challenges. Extracting over 80 multiscale features is computationally intensive, and real-time or near-real-time deployment may require reduced feature sets, GPU acceleration, or model compression. The workflow also does not explicitly quantify classification uncertainty. Although probability scores were generated, they were not mapped or analyzed to identify regions of low confidence, which is critical for operational disaster response, where uncertainty-aware decision support is needed.
Finally, the workflow’s generalizability is constrained by the study area’s geographic specificity. Both training and testing were conducted in Mexico Beach, Florida, a homogeneous coastal environment with consistent terrain and building typologies. Performance in regions with different architectural patterns, vegetation structures, or topographic complexity remains untested and will require further validation, retraining, or adaptation before operational deployment in other settings.

3.6. Future Work

Future research should expand the evaluation of this framework across a broader range of disaster scenarios, geographic settings, and LiDAR acquisition conditions to assess its scalability and robustness. Testing the model on both high- and low-density point clouds, as well as datasets acquired under variable flight parameters or weather conditions, would better reflect real-world emergency deployments and reveal how reliably the workflow performs under operational constraints.
Enhancing the semantic outputs with quantitative damage metrics represents another key direction. Extending the classification framework to estimate debris volume would enable direct quantification of hurricane-generated waste, thereby linking semantic mapping to practical decision-making tasks such as debris removal planning and resource allocation.
Several methodological improvements could further strengthen discrimination of damage-related classes. Integrating multi-source data, including RGB or multispectral imagery, UAV photogrammetry, or thermal measurements, would introduce spectral cues essential for separating visually distinct but geometrically similar materials, such as debris, bare ground, and partially damaged structures. Incorporating contextual and multi-temporal change descriptors, such as object-level metrics, neighborhood connectivity, or explicit height differences between co-registered pre- and post-event point clouds, may also improve the detection of complex or mixed damage patterns.
Model development presents additional opportunities. While this study used a Random Forest classifier, more advanced ensemble methods and deep learning architectures tailored for 3D small-object detection (for example, point-based or voxel-based neural networks) may improve the recognition of vehicles and partially collapsed structures. Evaluating such models under the same bi-temporal LiDAR conditions would clarify their potential advantages and trade-offs.
Addressing class imbalance remains a significant challenge. Future work will investigate oversampling of minority classes, cost-sensitive learning, and synthetic data augmentation to improve recall for underrepresented categories such as vehicles, vegetation, and partially collapsed buildings. These techniques may reduce bias toward majority classes without requiring additional data acquisitions.
Finally, future research should examine strategies for increasing reliability across diverse operational environments. This includes mitigating the effects of temporal scene evolution, co-registration inaccuracies, and variations in LiDAR sensor characteristics. Approaches such as domain adaptation, uncertainty-aware prediction outputs, and sensor-specific feature calibration may help ensure consistent performance as the workflow is applied to heterogeneous landscapes and acquisition conditions.

4. Conclusions

This study demonstrated the applicability of the 3DMASC framework, coupled with a Random Forest classifier, for automated semantic classification of pre- and post-hurricane LiDAR data in a coastal disaster context. Using Mexico Beach, Florida, as a case study for Hurricane Michael, the method achieved high classification accuracy in both stable and disaster-affected environments. Pre-hurricane classification reached an overall accuracy of 0.9711, effectively mapping stable classes such as ground, water, and buildings, while also delivering high precision and recall for power lines despite minor misclassifications. Post-hurricane classification maintained a comparable accuracy of 97%, with stable features consistently segmented at high fidelity.
While damage-specific classes, such as debris and damaged buildings, were identified with moderate success, classification performance in these categories was challenged by scene-level ambiguities, structural overlap, and limited training samples. Misclassifications were most frequent in transitional zones and for small, underrepresented features such as vehicles, reflecting known limitations of LiDAR-only datasets without RGB or multispectral attributes.
Despite these challenges, the framework proved robust, interpretable, and transferable between spatially distinct zones, highlighting its potential for rapid situational awareness and preliminary damage assessments in coastal hurricane-affected regions where high-quality LiDAR data is available. However, this workflow was validated only for Mexico Beach, Florida, and its applicability to other regions remains untested. Broader generalization to diverse landscapes, particularly those lacking high-quality LiDAR or labeled training data, will require further validation.
Future improvements should focus on integrating multi-source datasets (RGB or multispectral imagery, UAV photogrammetry, thermal data), incorporating contextual and multi-temporal features, addressing class imbalance, and developing automated tools for debris volume estimation. These enhancements are essential for improving separation between visually similar classes, detecting small objects, and supporting operational-grade disaster mapping and long-term resilience analysis.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/su18010090/s1, Table S1. Detailed definitions of 3DMASC features used in Table 2.

Author Contributions

Writing—Original Draft: J.K.N.; Writing—Review and Editing: J.K.N., I.O. and L.L.; Visualization: J.K.N. and I.O.; Validation: J.K.N. and L.L.; Software: J.K.N. and L.L.; Project Administration: L.L.; Conceptualization: J.K.N. and L.L.; Methodology: J.K.N. and L.L.; Investigation: J.K.N.; Formal Analysis: J.K.N., I.O. and L.L.; Data Curation: J.K.N. and L.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors on request.

Acknowledgments

The authors are grateful to Florida International University’s Moss School of Construction Management and the Department of Earth and Environment, where the study was conducted.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Kyne, D. Cleanup after Katrina: An analysis on policy, process, priorities, problems, and politics. Disaster Prev. Manag. 2010, 19, 314–332. [Google Scholar] [CrossRef]
  2. Denhart, H. Deconstructing disaster: Economic and environmental impacts of deconstruction in post-Katrina New Orleans. Resour. Conserv. Recycl. 2010, 54, 194–204. [Google Scholar] [CrossRef]
  3. Atchue, J., III; Morse, A. Planning and implementing effective debris management is a crucial part of ensuring resilience. Int. J. Waste Manag. Technol. 2023, 6, 1–9. [Google Scholar]
  4. Babbitt, C.W. The role of clean technology research in disaster debris management. Clean Technol. Environ. Policy 2019, 21, 923–924. [Google Scholar] [CrossRef]
  5. Holland, G.; Bruyère, C.L. Recent intense hurricane response to global climate change. Clim. Dyn. 2014, 42, 617–627. [Google Scholar] [CrossRef]
  6. Marsooli, R.; Lin, N.; Emanuel, K.; Feng, K. Climate change exacerbates hurricane flood hazards along US Atlantic and Gulf Coasts in spatially varying patterns. Nat. Commun. 2019, 10, 3785. [Google Scholar] [CrossRef]
  7. NOAA. 2024 Atlantic Hurricane Season Summary. National Oceanic and Atmospheric Administration. Available online: https://www.nhc.noaa.gov/data/tcr/index.php?season=2024 (accessed on 12 August 2025).
  8. Roueche, D.B.; Nakayama, J.O.; Kijewski-Correa, T.; Prevatt, D.O. A unified multievent windstorm performance testbed for single-family residential buildings. Nat. Hazards Rev. 2024, 25, 04023060. [Google Scholar] [CrossRef]
  9. Cheng, C.-S.; Behzadan, A.; Noshadravan, A. A post-hurricane building debris estimation workflow enabled by uncertainty-aware AI and crowdsourcing. Int. J. Disaster Risk Reduct. 2024, 112, 104785. [Google Scholar] [CrossRef]
  10. Jiang, Y.; Yuan, Y. Emergency Logistics in a Large-Scale Disaster Context: Achievements and Challenges. Int. J. Environ. Res. Public Health 2019, 16, 779. [Google Scholar] [CrossRef]
  11. Kovanič, Ľ.; Topitzer, B.; Peťovský, P.; Blišťan, P.; Gergeľová, M.B.; Blišťanová, M. Review of Photogrammetric and Lidar Applications of UAV. Appl. Sci. 2023, 13, 6732. [Google Scholar] [CrossRef]
  12. Rawat, A.; Witt, E.; Roumyeh, M.; Lill, I. Advanced digital technologies in the post-disaster reconstruction process—A review leveraging small language models. Buildings 2024, 14, 3367. [Google Scholar] [CrossRef]
  13. Karaer, A.; Ulak, M.B.; Abichou, T.; Arghandeh, R.; Ozguven, E.E. Post-hurricane vegetative debris assessment using spectral indices derived from satellite imagery. Transp. Res. Rec. 2021, 2675, 504–523. [Google Scholar] [CrossRef]
  14. Jiang, S.; Friedland, C.J. Automatic urban debris zone extraction from post-hurricane very high-resolution satellite and aerial imagery. Geomat. Nat. Hazards Risk 2016, 7, 933–952. [Google Scholar] [CrossRef]
  15. Schneider, P.J.; Schauer, B.A. HAZUS—Its development and its future. Nat. Hazards Rev. 2006, 7, 40–44. [Google Scholar] [CrossRef]
  16. Haque, C.E.; Mahmud, K.H.; Walker, D. Understanding flood risk and vulnerability of a place: Estimating prospective loss and damage using the HAZUS model. Geographies 2022, 2, 453–475. [Google Scholar] [CrossRef]
  17. Tate, E.; Muñoz, C.; Suchan, J. Uncertainty and Sensitivity Analysis of the HAZUS-MH Flood Model. Nat. Hazards Rev. 2015, 16, 04014030. [Google Scholar] [CrossRef]
  18. Oyege, I.; Sibitenda, H.; Bhaskar, M.S.B. Deep learning applications for real-time and early detection of fall armyworm, African armyworm, and maize stem borer. Mach. Learn. Appl. 2024, 18, 100596. [Google Scholar] [CrossRef]
  19. Fountzilas, E.; Pearce, T.; Baysal, M.A.; Chakraborty, A.; Tsimberidou, A.M. Convergence of evolving artificial intelligence and machine learning techniques in precision oncology. npj Digit. Med. 2025, 8, 75. [Google Scholar] [CrossRef]
  20. Kulicki, M.; Cabo, C.; Trzciński, T.; Będkowski, J.; Stereńczak, K. Artificial Intelligence and Terrestrial Point Clouds for Forest Monitoring. Curr. For. Rep. 2024, 11, 5. [Google Scholar] [CrossRef]
  21. Zhang, Y.; Nie, G.; Wang, D. Structural health monitoring based on three-dimensional point cloud technology: A systematic review. Results Eng. 2025, 27, 106552. [Google Scholar] [CrossRef]
  22. Mohammadi, M.E.; Watson, D.P.; Wood, R.L. Deep Learning-Based Damage Detection from Aerial SfM Point Clouds. Drones 2019, 3, 68. [Google Scholar] [CrossRef]
  23. Letard, M.; Lague, D.; Le Guennec, A.; Lefèvre, S.; Feldmann, B.; Leroy, P.; Girardeau-Montaut, D.; Corpetti, T. 3DMASC: Accessible, explainable 3D point clouds classification. Application to Bi-spectral Topo-bathymetric lidar data. ISPRS J. Photogramm. Remote Sens. 2024, 207, 175–197. [Google Scholar] [CrossRef]
  24. Belgiu, M.; Drăguţ, L. Random forest in remote sensing: A review of applications and future directions. ISPRS J. Photogramm. Remote Sens. 2016, 114, 24–31. [Google Scholar] [CrossRef]
  25. Zhang, K.; Chen, S.-C.; Whitman, D.; Shyu, M.-L.; Yan, J.; Zhang, C. A progressive morphological filter for removing nonground measurements from airborne LIDAR data. IEEE Trans. Geosci. Remote Sens. 2003, 41, 872–882. [Google Scholar] [CrossRef]
  26. Besl, P.J.; McKay, N.D. Method for Registration of 3-D Shapes. In Proceedings of the Sensor Fusion IV: Control Paradigms and Data Structures, Boston, MA, USA, 14–15 November 1991; pp. 586–606. [Google Scholar]
  27. Ramezan, C.A.; Warner, T.A.; Maxwell, A.E.; Price, B.S. Effects of Training Set Size on Supervised Machine-Learning Land-Cover Classification of Large-Area High-Resolution Remotely Sensed Data. Remote Sens. 2021, 13, 368. [Google Scholar] [CrossRef]
  28. Majgaonkar, O.; Panchal, K.; Laefer, D.; Stanley, M.; Zaki, Y. Assessing LiDAR training data quantities for classification models. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2021, 46, 101–106. [Google Scholar] [CrossRef]
  29. Kuprowski, M.; Drozda, P. Feature selection for airbone LiDAR point cloud classification. Remote Sens. 2023, 15, 561. [Google Scholar] [CrossRef]
  30. Chehata, N.; Guo, L.; Mallet, C. Airborne lidar feature selection for urban classification using random forests. In Proceedings of the Laserscanning, Paris, France, 1–2 September 2009. [Google Scholar]
  31. Cha, J.; Lee, S.; Kim, H.-K. Deep Learning-Based Detection and Assessment of Road Damage Caused by Disaster with Satellite Imagery. Appl. Sci. 2025, 15, 7669. [Google Scholar] [CrossRef]
  32. Glenna, N.; Mitchella, J.; Spaetea, L.; Sankeya, T.; Shresthaa, R.; Hardegreeb, S. Modeling semiarid low-height vegetation with airborne LiDAR. Int. J. Remote Sens. 2011, 30, 23. [Google Scholar]
  33. Asgharian Pournodrati, L.; Baba, M.M.; Gangelhoff, J.; Sörgel, U. Multi-scale Transformer-based classification of bathymetric LiDAR data in shallow water environments. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2025, 48, 1–6. [Google Scholar] [CrossRef]
  34. Sithole, G.; Vosselman, G. Filtering of airborne laser scanner data based on segmented point clouds. In Proceedings of the ISPRS Workshop Laser Scanning 2005, Enschede, The Netherlands, 12–14 September 2005; pp. 66–71. [Google Scholar]
  35. Li, T.; Xiong, X.; Zhang, Y.; Fan, X.; Zhang, Y.; Huang, H.; Hu, D.; He, M.; Liu, Z. RE-YOLOv5: Enhancing Occluded Road Object Detection via Visual Receptive Field Improvements. Sensors 2025, 25, 2518. [Google Scholar] [CrossRef]
  36. Zahs, V.; Anders, K.; Kohns, J.; Stark, A.; Höfle, B. Classification of structural building damage grades from multi-temporal photogrammetric point clouds using a machine learning model trained on virtual laser scanning data. Int. J. Appl. Earth Obs. Geoinf. 2023, 122, 103406. [Google Scholar] [CrossRef]
  37. Puetz, A.; Olsen, R.; Helt, M. Terrain Categorization Using LIDAR and Multispectral Data; SPIE: Bellingham, WA, USA, 2007; Volume 6565. [Google Scholar]
  38. Lovekin, J.R.; Crandall, A.; Zhou, W.; Perman, E.A.; Knies, D. LiDAR-Based Delineation and Classification of Alluvial and High-Angle Fans for Regional Post-Wildfire Geohazard Assessment in Colorado, USA. GeoHazards 2025, 6, 45. [Google Scholar] [CrossRef]
  39. Mullen, C.; Penny, G.; Müller, M.F. A simple cloud-filling approach for remote sensing water cover assessments. Hydrol. Earth Syst. Sci. 2021, 25, 2373–2386. [Google Scholar] [CrossRef]
  40. Liu, H.-D.; Fan, L.; Luo, J.; Chang, Z. Synergistic optimization of mechanical strength and vegetative growth in ecological slope restoration materials through mix ratio design. Sci. Rep. 2025, 15, 20374. [Google Scholar] [CrossRef]
  41. Yadav, R.; Nascetti, A.; Ban, Y. Building Change Detection using Multi-Temporal Airborne LiDAR Data. arXiv 2022, arXiv:2204.12535. [Google Scholar] [CrossRef]
  42. van Delden, A.; Scholtus, S.; Burger, J. Quantifying the Effect of Classification Errors on the Accuracy of Mixed-Source Statistics; Statistics Netherlands: Hague, The Netherlands, 2015. [Google Scholar]
Figure 1. A map showing the geographic location of Mexico Beach, Florida, USA, and the delineated training and testing zones.
Figure 1. A map showing the geographic location of Mexico Beach, Florida, USA, and the delineated training and testing zones.
Sustainability 18 00090 g001
Figure 2. Workflow of the training sequence for hurricane damage classification.
Figure 2. Workflow of the training sequence for hurricane damage classification.
Sustainability 18 00090 g002
Figure 3. Spatial layout of the pre-hurricane sampling area in Mexico Beach, Florida, showing the systematic grid framework (100 rectangles, each ~154 m × 154 m) used for semantic labeling during training. These grids do not represent LiDAR acquisition resolution but were introduced to organize sampling and ensure spatial diversity. For geographic coordinates and scale, refer to Figure 1, which provides the full map of the study site with delineated training and testing zones.
Figure 3. Spatial layout of the pre-hurricane sampling area in Mexico Beach, Florida, showing the systematic grid framework (100 rectangles, each ~154 m × 154 m) used for semantic labeling during training. These grids do not represent LiDAR acquisition resolution but were introduced to organize sampling and ensure spatial diversity. For geographic coordinates and scale, refer to Figure 1, which provides the full map of the study site with delineated training and testing zones.
Sustainability 18 00090 g003
Figure 4. Pre-hurricane reference and segmentation for Zone A: (a) pre-hurricane raster image of the study area; (b) corresponding 3DMASC–Random Forest segmentation for the same extent, colored by semantic classes.
Figure 4. Pre-hurricane reference and segmentation for Zone A: (a) pre-hurricane raster image of the study area; (b) corresponding 3DMASC–Random Forest segmentation for the same extent, colored by semantic classes.
Sustainability 18 00090 g004
Figure 5. Spatial layout of the post-hurricane sampling area in Mexico Beach, Florida, showing the delineated training grids used for damage and debris classification following Hurricane Michael. For geographic coordinates and scale, refer to Figure 1, which provides the full map of the study site with delineated training and testing zones.
Figure 5. Spatial layout of the post-hurricane sampling area in Mexico Beach, Florida, showing the delineated training grids used for damage and debris classification following Hurricane Michael. For geographic coordinates and scale, refer to Figure 1, which provides the full map of the study site with delineated training and testing zones.
Sustainability 18 00090 g005
Figure 6. Classification results for the post-hurricane validation area (Zone B), showing the spatial distribution of predicted semantic classes.
Figure 6. Classification results for the post-hurricane validation area (Zone B), showing the spatial distribution of predicted semantic classes.
Sustainability 18 00090 g006
Table 1. Semantic classes used for hurricane damage classification based on their geometric and structural characteristics in 3D point cloud data.
Table 1. Semantic classes used for hurricane damage classification based on their geometric and structural characteristics in 3D point cloud data.
Class IDClass NameDescription
2GroundFlat, stable terrain surfaces with uniform elevation and reflectance
3RoadConstructed transport surfaces with smooth, linear geometry
4VegetationUpright, irregularly shaped natural features with complex 3D structure
5BuildingMan-made elevated structures with planar rooftops and defined edges
6Ocean/WaterHorizontally uniform, low-reflectance water surfaces
7Vehicles/CarsSmall, box-shaped mobile units with reflective surfaces
8Fence/BoundaryLinear, narrow, low-elevation structural dividers
11Power LineElevated, thin, tensioned line features suspended above ground
17Damaged BuildingStructures exhibiting irregular roof geometry, reduced vertical continuity, and fragmented elevation profiles compared to intact buildings.
18DebrisClusters of small, irregular points with high surface roughness and low vertical coherence, typically scattered across ground surfaces.
Table 2. Features selected for each semantic class during 3DMASC-based classification.
Table 2. Features selected for each semantic class during 3DMASC-based classification.
LabelFeatures Used
GroundZ_SC2_STD, Z_SC4_STD, PCA1_SC2, PCA2_SC2, ROUGH_SC2, INT_SC2_MEAN
RoadZ_SC2_STD, PCA1_SC2, ROUGH_SC2, CURV_SC2, INT_SC0
VegetationZ_SC2_STD, PCA3_SC2, ROUGH_SC4, CURV_SC4, INT_SC0
BuildingZ_SC2_STD, Z_SC4_STD, PCA1_SC4, ROUGH_SC2, CURV_SC2, INT_SC4_MEAN
Ocean/WaterZ_SC2_STD, PCA1_SC2, ROUGH_SC2, INT_SC0
Vehicles/CarsZ_SC2_STD, PCA1_SC2, PCA3_SC2, ROUGH_SC2, INT_SC2_MEAN
Fence/BoundaryPCA1_SC2, Z_SC2_STD, CURV_SC2, INT_SC0
Power LinePCA1_SC2, PCA2_SC2, PCA3_SC2, Z_SC2_MEAN, CURV_SC2, INT_SC0
Damaged BuildingZ_SC2_STD, Z_SC4_STD, ROUGH_SC4, CURV_SC4, PCA2_SC4, INT_SC2_MEAN
DebrisZ_SC2_STD, PCA3_SC2, ROUGH_SC2, ROUGH_SC4, CURV_SC4, INT_SC2_STD
Note: Abbreviations used in this table are explained in Supplementary Table S1.
Table 3. Confusion matrix for the pre-hurricane classification results, showing model performance across six stable semantic classes in the training zone with an overall accuracy of 0.9711.
Table 3. Confusion matrix for the pre-hurricane classification results, showing model performance across six stable semantic classes in the training zone with an overall accuracy of 0.9711.
Predicted
Real 234611PrecisionRecallF1-Score
210,190,06433,3853,270,552449500.99800.75480.8595
316,858553,06966000.94310.09280.1690
400264000.0000810.66670.000162
6343800536,54200.99170.99340.9925
117400034960.990.930.96
Table 4. Confusion matrix for the post-hurricane classification results, illustrating model performance across both stable and damage-related semantic classes in the independent test zone with an overall accuracy of 0.97.
Table 4. Confusion matrix for the post-hurricane classification results, illustrating model performance across both stable and damage-related semantic classes in the independent test zone with an overall accuracy of 0.97.
Predicted
2345678111718PrecisionRecallF1-Score
Real2227,475183318228310201113750.970.990.98
35394966000200000.960.90.93
4584027,2554900030000.960.960.96
5238052247,8330032104530.980.970.98
67200064040000010.990.99
776029401700000.890.130.23
86101187001900000.950.930.96
1154000000290101760.990.930.96
17318023800007666430.870.430.58
18509809104001396813,4140.840.720.77
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ndolo, J.K.; Oyege, I.; Lagos, L. Automated Hurricane Damage Classification for Sustainable Disaster Recovery Using 3D LiDAR and Machine Learning: A Post-Hurricane Michael Case Study. Sustainability 2026, 18, 90. https://doi.org/10.3390/su18010090

AMA Style

Ndolo JK, Oyege I, Lagos L. Automated Hurricane Damage Classification for Sustainable Disaster Recovery Using 3D LiDAR and Machine Learning: A Post-Hurricane Michael Case Study. Sustainability. 2026; 18(1):90. https://doi.org/10.3390/su18010090

Chicago/Turabian Style

Ndolo, Jackson Kisingu, Ivan Oyege, and Leonel Lagos. 2026. "Automated Hurricane Damage Classification for Sustainable Disaster Recovery Using 3D LiDAR and Machine Learning: A Post-Hurricane Michael Case Study" Sustainability 18, no. 1: 90. https://doi.org/10.3390/su18010090

APA Style

Ndolo, J. K., Oyege, I., & Lagos, L. (2026). Automated Hurricane Damage Classification for Sustainable Disaster Recovery Using 3D LiDAR and Machine Learning: A Post-Hurricane Michael Case Study. Sustainability, 18(1), 90. https://doi.org/10.3390/su18010090

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop