Next Article in Journal
Design and Implementation of a Multi-Band Active Radar Calibrator for SAR
Next Article in Special Issue
Investigating Fold-River Interactions for Major Rivers Using a Scheme of Remotely Sensed Characteristics of River and Fold Geomorphology
Previous Article in Journal
Ensemble-Based Cascaded Constrained Energy Minimization for Hyperspectral Target Detection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Augmented Reality Mapping of Rock Mass Discontinuities and Rockfall Susceptibility Based on Unmanned Aerial Vehicle Photogrammetry

1
State Key Laboratory of Hydraulic Engineering Simulation and Safety, Tianjin University, Tianjin 300350, China
2
Yalong River Hydropower Dev Co Ltd., Chengdu 610051, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2019, 11(11), 1311; https://doi.org/10.3390/rs11111311
Submission received: 11 April 2019 / Revised: 22 May 2019 / Accepted: 23 May 2019 / Published: 1 June 2019

Abstract

:
In rockfall hazard management, the investigation and detection of potential rockfall source areas on rock cliffs by remote-sensing-based susceptibility analysis are of primary importance. However, when the rockfall analysis results are used as feedback to the fieldwork, the irregular slope surface morphology makes it difficult to objectively locate the risk zones of hazard maps on the real slopes, and the problem of straightforward on-site visualization of rockfall susceptibility remains a research gap. This paper presents some of the pioneering studies on the augmented reality (AR) mapping of geospatial information from cyberspace within 2D screens to the physical world for on-site visualization, which directly recognizes the rock mass and superimposes corresponding rock discontinuities and rockfall susceptibility onto the real slopes. A novel method of edge-based tracking of the rock mass target for mobile AR is proposed, where the model edges extracted from unmanned aerial vehicle (UAV) structure-from-motion (SfM) 3D reconstructions are aligned with the corresponding actual rock mass to estimate the camera pose accurately. Specifically, the visually prominent edges of dominant structural planes were first explored and discovered to be a robust visual feature of rock mass for AR tracking. The novel approaches of visual-geometric synthetic image (VGSI) and prominent structural plane (Pro-SP) were developed to extract structural planes with identified prominent edges as 3D template models which could provide a pose estimation reference. An experiment verified that the proposed Pro-SP template model could effectively improve the edge tracking performance and quality, and this approach was relatively robust to the changes of sunlight conditions. A case study was carried out on a typical roadcut cliff in the Mentougou District of Beijing, China. The results validate the scalability of the proposed mobile AR strategy, which is applicable and suitable for cliff-scale fieldwork. The results also demonstrate the feasibility, efficiency, and significance of the geoinformation AR mapping methodology for on-site zoning and locating of potential rockfalls, and providing relevant guidance for subsequent detailed site investigation.

Graphical Abstract

1. Introduction

Slope movement is one of the most common geological and geomorphic natural hazards in China [1]. Rockfall is an essential type of slope movement phenomenon that frequently occurs on steep rock cliffs and represents a relevant hazard in mountainous areas, significantly influencing the safety of residents and infrastructure. Defining, zoning, and locating the most probable future rockfall source areas are of primary importance for rock slope hazard investigation, assessment and management. According to the spatial scale of geomorphological observations and investigations, rockfall analysis mainly covers three different scales: the regional scale (areas of square kilometers), block specific scale (areas of square meters), and cliff scale (hundreds to thousands of square meters) [2], and of which this paper focused on the cliff scale rockfall. In practice, the field investigations of cliffs are usually carried out in stages. The primary survey stage should start with rockfall source detection to define the points of initiation for future rockfalls [2]. Compared with traditional approaches that might consider all steep slopes as equivalent source areas; the problem with on-site detecting, zoning, and locating of the most likely rockfall source areas and unstable rock masses on the slope badly needs to be addressed [3]. On-site rockfall source detection can ensure a more effective and reliable hazard assessment, guide further detailed fieldwork and subsequent slope treatment, and help establish a smart rock slope environment. Therefore, this research problem is particularly worthy of academic attention.
Rockfall mechanisms at the cliff scale are strongly linked to the slope morphology and rock discontinuities [2], which are primarily the joints involved in our study areas. The use of remote sensing techniques over recent decades has dramatically increased the 3D data acquisition quality of both the slope and discontinuity geometry for rockfall analysis [4]. Based on the remote sensed high-resolution 3D point clouds and mesh models, more reliable landslide mappings, either in 2D or 3D, in scalar or vectors forms, can be generated [5]. With the goal of on-site locating potential rockfalls, the analysis results need to be further visualized in the fieldwork. The strategies for data visualization could be divided into two categories: displaying the data within the 2D screen of laptops and desktops, and the recently developed augmented reality (AR) visualization, which can put the 2D images and 3D models into the real world using the camera of mobile devices; including smartphones and tablets. However, due to the innate irregularity of rocky slope surface morphology, it is a nonintuitive and puzzling task to establish the corresponding correlation between the real slopes and the hazard maps displayed as virtual content within a 2D screen, as schematically illustrated in Figure 1 where the virtual information is 3D scalar point clouds that could convey the information of join sets and rockfall susceptibility with the colors of the points. Different from the conventional 2D visualization provided with the 2D screens, the AR methods can present a 3D holographic visualization, which can effectively feedback rockfall analysis results to the fieldwork and avoid subjective cognitive bias. This paper presents some pioneering research in introducing mobile augmented reality technology to the realm of rockfall hazard management and developed a methodology to visualize the slope information beyond the screen to the real slope environment and directly superimpose slope information onto the real slopes, as shown in Figure 1.
In order to understand which areas of a steep cliff are more likely to be affected by a future rockfall, the spatial distribution of rockfall probability is mainly quantified by the index ”susceptibility”, as previously reported [2,3]. Detailed 3D outcrop models are the fundamental data basis that significantly controls the resolution, accuracy, and reliability of the susceptibility analysis. Far beyond the conventional survey methods of fracture windows sketching and the assumption that simplifies the whole slope with a uniform dip and dip direction, close-range remote sensing technologies, including Light Detection and Ranging (LiDAR) and photogrammetry [5,6,7], are revolutionary in acquiring 3D virtual outcrops of rock slopes with high resolution and high accuracy for susceptibility analysis. Specifically, for steep slopes at the cliff scale, the overhanging areas are at high risk to rockfalls but are poorly resolved with the nadir imaging view of airborne remote sensing techniques, where the Terrestrial Laser Scanning (TLS), ground-based photogrammetry, and the recently developed unmanned aerial vehicle (UAV) photogrammetry are especially applicable [2,7].
In the last two decades, LiDAR has been widely applied to the generation of digital elevation models (DEMs) and rock discontinuity characterization for slope stability analysis [8,9]. The authors in [5] developed an algorithm for automated geometric analysis and classification of landslides based on LiDAR-derived DEMs. The exact parameters of outcropped structural planes, including positions, orientations, persistence, and spacing can be measured from laser-scanned 3D point clouds or 3D meshes of the cliff. Reference [10] researched the method for rock surface extraction from 3D rock-mass point clouds based on major orientation estimation. The research in [11] developed a semi-automatic method for rock mass joint recognition from 3D point clouds while the study in [12] further explored the automatic extraction of discontinuity orientation. Reference [13] performed geomechanical analysis on high-resolution 3D point clouds collected by TLS for slope stability condition assessment.
In recent years, more and more attention has been given to UAV photogrammetry, especially the SfM-MVS [14], which can also generate high-resolution 3D point clouds and 3D meshes of the cliff, with the advantage of low cost, flexible oblique view sensing and photo-realistic vision information when compared with LiDAR [15]. The authors in [16] first introduced SfM photogrammetry to rock outcrop modeling and presented an initial appraisal of the SfM-derived DEM compared with laser scanned models, which revealed the revolutionary potential of the SfM technique for geoscience applications. The study by [17] demonstrated that the photogrammetry method using low-cost image sensors and aerial platforms could create reliable and accurate rock surface models and [18] overviewed the UAV SfM workflows and applications in geomorphometry reconstruction. Reference [19] introduced the SfM generated models into 3D rockfall simulations. [20] applied UAV SfM photogrammetry to recognize the predominant discontinuities of rock slopes and quantified the block size for rockfall simulation. During rock discontinuity characterization for rockfall analysis, discontinuity orientation is the most critical input parameter for quantifying the main joint sets that serve to destabilize rock masses [2]. In fact, in case of discontinuity orientation measuring, since TLS and ground-based photogrammetry technologies suffer from the effects of occlusion when dealing with high slopes, UAV SfM virtual outcrops can yield better along-strike predictions, which is attributed to a broader coverage, the avoiding of hidden zones by UAV and the reliable reconstruction of a higher number of structural planes [21].
As overviewed by [22], modern UAV SfM technology can provide point clouds and textured models for structural geology with multidimensional information (x, y, z, point orientation, color, texture). However, to the best of our knowledge, current studies have mainly focused on the geometry information of the UAV SfM model, and have not made the most of the advantage in the captured photo-realistic model materials and textures of rock slopes. Our research explored the computer vision information derived from photo-realistic virtual outcrops, which show another dimension of the UAV SfM potentiality.
To detect potential rockfall source areas, current research has seen a surge of interest in combining kinematic analysis with 3D point clouds or 3D meshes from remote sensing to generate susceptibility maps according to the interaction between the discontinuities and the local cliff orientation. The authors in [3] studied the impacts of fracturing patterns on the rockfall susceptibility and erosion rate of slopes based on Airborne Laser Scanning (ALS), while [2] researched the cliff scale failure mechanisms of steep and overhanging rock slopes and developed some indexes to assess rockfall susceptibility using kinematic analysis based on TLS models. Intensive studies have focused on the calculation step of susceptibility maps; however, the realization of the straightforward on-site visualization of hazard maps to feedback the risk zones to fieldwork still remains a research gap. To address this problem, we first introduce advanced mobile augmented reality technology, and then used its advantages to directly superimpose the rockfall susceptibility maps onto corresponding real rock slopes.
Mobile augmented reality has only been applied to the geological and environmental context in the last few years. The authors in [23] developed a new AR software for fieldwork, such as the interpretation of folded structures in a mobile augmented reality environment. Reference [24] developed a novel AR system for abandoned mine hazard site investigation by visualizing mine drifts, boreholes and underground water in a mobile augmented reality environment. These applications are generally aimed at larger scenarios that allow for a higher degree of AR overlapping discrepancies as compared with the task of superimposing information onto fractured rock mass in this research, which has not been addressed in previous studies.
The concept of augmented reality is established based on the seamless alignment between virtual objects and the real world in an accurate and visually acceptable way, which belong to the scope of a motion tracking issue. The solution to this underlying problem is referred to as “camera pose estimation” with a simple definition: “given a set of correspondences between 3D features and their projections in the images plane, pose estimation consists in computing the position and orientation of the camera” [25].
In mobile augmented reality realizations, there exist various strategies for camera pose estimation [25] including sensor-based [23,24], marker-based [26], SLAM-based [27], and model-based methods [28,29], from which the most suitable approach should be selected for the scenario of rockfall AR mapping. The sensor-based method utilizes GPS, IMU, depth cameras, and other sensors embedded in mobile devices, and this type of approach cannot meet the accuracy demands for aligning 3D models with the intricate fractured rock mass surface. The marker-based method using fiducial markers as locating references [26] and requires a heavy workload of marker arrangement, which is unpractical in fieldwork scenarios. Simultaneous localization and mapping (SLAM) including visual-inertial odometry (VIO) simultaneously locates the camera and maps the environment in a random local coordinate system [27], which causes difficulties in transforming the camera location to another desired coordinate system, so this type of method alone cannot meet the requirements of rockfall AR mapping, but is suitable when combined with other approaches to extend and enhance the tracking performance. While the model-based method utilizes a predefined 3D template model of the real-world object to directly recognize and track the object itself, which allows more accurate alignment when compared with the sensor-based method [29], avoids fiducial markers correlated with the marker-based method and simplifies the coordinate system conversion problem when compared with the SLAM-based method [25]. Consequently, the model-based strategy is the most suitable for the AR mapping of rockfall susceptibility maps, which was adopted in this study.
The matching and alignment between predefined 3D models and real-world objects based on computer vision features is a critical problem to be addressed in model-based AR, and different types of features such as key point features, edge features, and color features, etc., need to be chosen in consideration of different types of target objects, which may be static rigid or articulated, textured or untextured, with or without sharp edges [30], etc. Concerning the specific target of the fractured rock mass, the visual characteristics can be preliminarily described as follows: first, the fractured rock mass is a rigid object and is generally static during the inspection process; second, it usually has many intricately and heterogeneously distributed sharp edges at the intersections of structural planes with no semantic straight lines and no outer contours; and third, it can present various textures and colors, and may sometimes be texture-less, depending on the type of rock mass material. Therefore, the key point features and color features can sometimes be unreliable for this type of tracking target, while the sharp edges of structural planes can be a superb candidate feature for model-based AR tracking.
However, current edge-based camera pose estimation studies have been mainly aimed at tracking the manufactured artifacts for Industry 4.0, which can be distinguished by regular humanmade geometries with sharp edges according to precise 3D CAD models. The research by [31] and [32] studied the binocular approach and the contour-color feature fusion approach for tracking texture-less benchmark toy objects, respectively. The authors in [33] researched the line-based AR initialization method for aircraft assembly and [34] studied fast pose estimation for texture-less industrial products based on B-Rep models. Reference [35] proposed the average shading gradients method for the AR registration of untextured geometries including statues and buildings, while [36] researched model-constrained SLAM localization methodology to track both polyhedral and curved objects including toys, furniture, vehicles, and machines. Unfortunately, to the best of our knowledge, in current studies of edge-based tracking for augmented reality, the rock mass has seldom been considered as the tracking target.
In the problem of the edge-based tracking of a rock mass target, there are mainly two aspects of research questions: where to track and how to get the 3D model for matching (detailed in Section 2.3). Regarding the first aspect, compared with the industrial applications aiming at well trackable geometries, the irregularly fractured morphology can appear heterogeneous in the spatial distribution of edge features, where the scales of extending, winding, and crisscrossing of the edges detected in camera images can vary tremendously in different parts of the rock mass [37,38]. Consequently, the slope region with the most favorable visual features for edge alignment needs to be delineated first. Regarding the second aspect, a priori 3D model corresponding to the rock mass to be used as a matching template does not exist. Furthermore, even the remote sensed and 3D reconstructed rock surface models would not directly work as a useful 3D template model, on account of the naturally formed irregular geometries of the rock mass [39]. Therefore, the fundamental problem to be solved is how to obtain an applicable 3D template model of the rock mass. In this paper, we first studied the model-based tracking of the fractured rock mass by focusing on the edge features formed by the structural plane intersections, and developed a tracking region selection and 3D template modeling solution which extracts the visually prominent structural planes from UAV SfM models by the proposed method VGSI and Pro-SP.
This paper is organized as follows. Section 1 overviews the previous studies on the rockfall analysis based on remote sensing and different augmented reality strategies sorted by the camera pose estimation technologies. Subsequently, the framework of the proposed methodology is elucidated in Section 2, and Section 2.1 to Section 2.6 explain the six parts of this framework, respectively. Section 2.1 presents the UAV-SfM remote sensing and modeling of a slope at risk to rockfall in the case study location; Section 2.2 clarifies the rock discontinuity characterization process; Section 2.3 elaborates the realization strategy of the proposed markerless mobile AR inspection methodology; Section 2.4 articulates the proposed rock mass 3D template modeling method, which lays the foundation for recognizing and tracking rock mass in model-based AR; Section 2.5 illustrates the kinematic schemes for stability analysis using the 3D point clouds; and Section 2.6 explains the rock slope information AR superimposing and AR mapping. Then, Section 3 presents the results, including the validation of the edge tracking performance using the proposed 3D templated modeling method, the realization of susceptibility AR mapping and on-site rockfall zoning and locating in augmented reality. Then, the results, potential, and limitations of the proposed methodology are discussed in Section 4. Finally, Section 5 concludes and presents future research prospects.

2. Materials and Methods

The research framework of the proposed methodology for rock slope information AR mapping and on-site rockfall zoning is shown in Figure 2. Two aspects of the problem need to be solved: how to generate high-resolution rock fall hazard maps and how to superimpose the hazard maps onto real rock slopes. This paper establishes a novel workflow that solves both problems by using the UAV photogrammetric remote sensing data.
The framework consists of six parts: (1) the photogrammetric remote sensing (Section 2.1); (2) the rock discontinuity characterization (Section 2.2); (3) the markerless mobile AR (Section 2.3); (4) the vision-feature-based 3D template modeling (Section 2.4); (5) the rockfall stability analysis (Section 2.5) and (6) the rock slope information AR superimposing and mapping (Section 2.6).
First, in the photogrammetric remote sensing section, based on multiple UAV imagery of the rock slope, the SfM-MVS technique reconstructs the 3D virtual slope with high resolution and accuracy in the form of both 3D point clouds and photo-realistic 3D mesh models. Compared with other remote sensing approaches, the photogrammetric method can acquire high-quality information on the specific dimension of computer vision such as photo-realistic model textures, which is indispensable for the following vision-based augmented reality, thus making photogrammetry a requisite choice for remote sensing in the proposed methodology. Second, in the rock discontinuity characterization section, joint sets and structural planes are detected from the 3D point clouds. The yielded structural plane model lays the foundation for analyzing both the visual features and rockfall susceptibilities of the slope rock mass. Regarding the first data flow direction, the vision-feature-based 3D template modeling section, the Canny edges detected in the images rendered from the photo-realistic 3D mesh were aligned with the projected structural plane models to distinguish visually prominent structural planes and establish a 3D template model for AR tracking. In the other direction, in the rockfall stability analysis section, the kinematic analysis was conducted using the joint set parameters to generate rockfall susceptibility maps. Furthermore, in the markerless mobile AR section, following the model-based camera pose estimation paradigm, the model features extracted from the 3D template model and the image features detected from the mobile device camera video streaming were matched to perform the model-based tracking. Moreover, the VIO technique based on the camera and IMU of the mobile devices was used to extend the AR tracking when the target object was out of view. Last, but not least, the rock slope information, including rock mass discontinuities and rockfall susceptibility maps were superimposed and AR mapped on real rock slopes. After integrating the methodology and techniques above-mentioned, the potential rockfall sources could finally be zoned and located in the augmented reality environment.

2.1. UAV SfM Modeling

The Structure-from-Motion Multi-View-Stereo (SfM-MVS) technique can flexibly produce 3D models with photo-realistic materials from a highly redundant bundle adjustment based on matching features in multiple overlapping photographs. Combining SfM photogrammetry with UAV oblique photos can significantly promote the routine use of 3D modeling in slope observation practice. Figure 3 shows the multi-rotor drone DJI Mavic Pro used in this research. The specifications of the built-in camera of DJI Mavic Pro are as follows: image sensor, Sony EXMOR 1/2.3”; sensor type, CMOS; effective pixels, 12.35Mpix; resolution, 4000 * 3000 pixels; pixel size, 1.55 μm. Figure 4 shows the location of the case study, which is a typical roadcut cliff at risk to rockfall, located in the Mentougou District of Beijing, China. UAV imagery is also presented in Figure 4. The average Ground Sampling Distance (GSD) of the UAV images is around 4mm. The date of the UAV photo shoot was Beijing time, January 5, 2019. The SfM-MVS modeling was performed using open source softwares, including VisualSFM [40,41], PMVS [42,43], and MeshLab [44,45,46,47]. The modeling output included high-resolution 3D point clouds and 3D meshes with photo-realistic materials of this slope.

2.2. Rock Discontinuity Characterization

It is interesting that the two different tasks of rockfall analysis and rock mass target AR tracking intersect at the focus of the three-dimensional morphology of exposed slope rock mass, i.e., the outcropping structural planes that geometrically affect the slope stability are just the flat surface planes dominating the visual edge features of slopes [20,48].
The recognition and characterization of rock mass discontinuity from 3D point clouds were carried out using the opensource software DSE and the approach proposed in [11], of which the procedure consists of three steps:
(i) Local curvature calculation
For each 3D point of the point cloud, the K-nearest neighbors (knn) were identified, the coplanarity condition of these k + 1 points were checked, and the normal vector was calculated with a plane adjustment. The normal vectors of the 3D points could both quantify the orientations of the joint sets and slope surface for the later rockfall kinematic analysis in Section 2.5.
(ii) Statistical analysis of the planes
Taking advantage of the stereographic projection of the plane poles, the density of the poles was calculated using kernel density estimation (KDE); the principal orientations of different discontinuity sets were identified through a semi-automatic selection of the local density peaks; point clouds were segmented into different discontinuity sets. The stereographic projection of the studied slope is plotted in Figure 5, where eight joint sets are clustered, and the segmented point clouds of these joint sets with different colors are shown in Figure 6. Joint set 1 was considered the main slope direction and five principal joint sets with relatively large percentages of point number in each joint set to the total point number, magnified in Figure 6, were considered in the rockfall analysis in Section 2.5.
(iii) Cluster analysis
The points belonging to a common discontinuity set were grouped according to planar clusters by employing the “Density-based Scan Algorithm with Noise”(DBSCAN) before these point clusters were fitted to flat surfaces representing the individual structural planes of the discontinuity set. Some of the dominant structural planes were then filtered based on the structural plane areas for 3D template modeling, as demonstrated in Section 2.4.

2.3. Edge-Based AR Tracking of Rock Mass Target

As overviewed in Section 1, there are various camera pose estimation approaches, among which the model-based tracking strategy focusing on edge features from a monocular vision sequence of mobile devices is relatively the most suitable choice in the rock slope context. As illustrated in the camera tracking process part of Figure 7, the principle behind the edge-based camera tracking process is to define a distance between the edges detected in the 2D image sequence captured by the mobile camera and the projected 3D edges underlying the corresponding 3D model and minimize the norm of the reprojection error [49].
Two phases of edge alignment are involved: tracking initialization and tracking maintenance. First, the tracking initialization phase requires an initial model pose to be predefined, according to which the 3D model is projected into the 2D image, then the projected 3D edges L ( q ) need to be aligned to the actual objects to generate an estimation of the initial camera pose as the starting point of the tracking. Subsequently, in the tracking maintenance phase, the edges in the camera images are detected and sampled into point x i , and a search is performed along the image edge sampled point x i normal to the projected edges L ( q ) to find strong gradients in the next frame. A nonlinear iterative optimization approach can be used to estimate the camera pose, which minimizes the errors between the selected points and the projected edges [30], that is:
q ^ = arg min q i d ( L ( q ) , x i )
q = ( t w c , θ u ) T , q s e ( 3 )
where t w c is the translation vector between the world frame and camera frame, θ and u are the angle and the axis of the rotation matrix between the world frame and camera frame, d ( L ( q ) , x i ) is the squared distance between the point x i and the projection of the 3D edges L ( q ) for the model pose q .
Obviously, according to the above-mentioned mathematical mechanism, the quality of the edge-based pose estimation is highly dependent on two factors: the quality of the detected image edges, which need to be clearly visible and detectable in the camera image sequences, and the quality of the projected model edges, which need to be prominent enough to depict the major edge features of the 3D target object.
As reviewed in Section 1, in current edge-based camera pose estimation studies, the tracking target is mainly the manufactured artifacts, such as toys, industrial machines, domestic appliance and vehicles, etc. including the standard mechanical parts in Figure 7. These types of tracking targets generally have regular geometric shapes with simple sharp edges [51] and are often designed and manufactured by building a corresponding 3D CAD model with an accurate depiction of the geometries [50]. Therefore, the two requirements of the 2D image edge quality and the 3D model edge quality are naturally met for these targets.
Unfortunately, we have not found previous research on edge-based pose estimation aimed at the specific tracking target of rock mass. When it comes to this specific challenging tracking target, two critical aspects of research questions came to light, as clarified in the tracking target objects part of Figure 7:
Research question one: Where do we track?
Unlike most industrial objects, which often have a relatively small size and clear outer contours; cliffs with large spatial scales mainly present sharp inner edges at the intersection lines of rock structural planes. However, irregularly fractured morphology can appear heterogeneous in the spatial distribution of edge features, where the scales of extending, winding and crisscrossing of the edges detected in camera images can vary tremendously in different parts of the rock mass. Consequently, the slope region with the most desirable visual features for edge alignment needs to be delineated first.
Research question two: How do we obtain a 3D model for matching?
As there are no previously provided corresponding 3D models (such as CAD models) available to serve as the 3D template model—and even if 3D reconstructions of rock mass are accessible through remote sensing, simply adopting all parts of the 3D rock mass models within the camera’s field of view can involve relatively small and cluttered model edges derived from small-scale fractures, which need to be avoided in the edge tracking paradigm. Consequently, appropriate 3D template models that can selectively represent the prominent sharp edges of rock mass need to be extracted from the remotely sensed 3D models.
Considering the research questions in Figure 7, two hypotheses regarding AR tracking for the fieldwork scenario were assumed: (a) only some local parts of the rock slope with meter scales would be selected for AR initialization, and (b) the predefined camera pose of the mobile device for AR initialization was approximately horizontal and perpendicular to the main slope surface. From practical experience, the first hypothesis is beneficial in a large-scale outdoor AR context, where regardless of how large the cliff is, the AR initialization is confined to the same spatial scale as a human observer. The second hypothesis is reasonable and in harmony with the actual fieldwork scenarios of rock slope investigation. Accordingly, the region of the rock mass target for model-based tracking is limited to a rectangular part of the rock slope referred to as the Tracking Initialization Rectangle, as indicated and magnified in Figure 8 with the yellow dashed rectangle.
To address research question one in Figure 7, we proposed a novel visual-geometric synthetic image (referred to as VGSI) method to search for a suitable initial camera pose facing the local region of rock mass within the Tracking Initialization Rectangle in Figure 8 with relatively preferred edge features for AR tracking. To address research question two in Figure 7, we proposed a novel prominent structural plane (referred to as Pro-SP) method for the 3D template modeling of a rock mass target in Figure 8. These two methods are elaborated in Section 2.4.
After settling the model-based pose estimation question, the overall procedure of the mobile augmented reality in the slope fieldwork context is demonstrated in Figure 9. First, we faced the mobile device to the local region of the rock slope selected by the VGSI approach and adjusted the camera to the predefined initialization pose to recognize the rock mass target and trigger the tracking; then we superimposed the AR content onto the rock slope, and look around the whole cliff augmented with overlapping virtual information such as the rock discontinuity point clouds in Figure 9. When the rock mass tracking region was within the camera’s field of view, the proposed 3D template model Pro-SP maintained the tracking of the real rock mass objects, assisted by the freely available AR engine Vuforia [52] and game engine Unity3D [53] for camera pose computation and screen rendering. When the target rock mass was out of sight, the extendible AR tracking was maintained based on SLAM, until the target came back into view, which could ensure a scalable AR experience. The extendible AR could be quickly realized using the VIO approach of the free AR SDK (ARKit [54] of IOS and ARCore [55] of Android) embedded in the popular mobile operating systems today.

2.4. VGSI and Pro-SP Method for Tracking Region Selecting and 3D Template Modeling

In the paradigm of edge tracking, the complete alignment between one hundred percent of the detected image edges and the projected model edges is not a requisite for launching an AR session, while the visually prominent edges of the real target objects and the corresponding 3D models play decisive roles in the feasibility, accuracy, and robustness of tracking performance. Therefore, the detection, recognition, and modeling of the prominent edges is the key to those two research questions described in the Introduction Section and Section 2.3. This section elaborates the proposed novel approaches of the visual-geometric synthetic image (VGSI) for selecting the rock mass tracking region involving more edges that are visually prominent and the prominent structural plane (Pro-SP) for the 3D template modeling of fractured rock mass containing prominent edges, as shown in Figure 10.
Benefiting from the powerful SfM-MVS modeling technology, the realistic visual appearance of real slope surfaces and the high-resolution geometry of the outcropping structural planes can both be reconstructed in a virtual environment, where the precise spatial overlapping of the photo-realistic 3D mesh model and the discontinuity point clouds of the same slope is achievable, and the pose of a virtual camera can be parametrically controlled. The SfM models provide the precondition for the coupled analysis of the visual-geometric information in the synthetic images captured by the virtual camera. However, previous studies that have applied UAV SfM to rock slope modeling and stability analysis have seldom mined the vision information from the photo-realistic model materials mapped onto the mesh geometry, partly because the idea of combining vision-based augmented reality with rock slope remote sensing has seldom been reported and studied.
In this paper, among the rock discontinuities, the term dominant structural plane refers to a flat and complete surface of which the area is large enough relative to the Tracking Initialization Rectangle [2,20]. Since the point cloud of the rock outcrop was subsampled with the same minimal space between points, based on the point clouds of individual structural planes segmented and clustered, the number of points in each structural plane could represent the relative size of it. The point cloud of the regional rock outcrop within the Tracking Initialization Rectangle was grouped into six joint sets in Section 2.3 as shown in Figure 8 using the approaches in Section 2.2. Then, dominant structural planes were filtered from all detected structural planes based on the threshold of the minimal number of points, as revealed in the bottom part of Figure 8.
In the proposed VGSI method, from the same position and orientation of a virtual camera, the Canny edge [48,56,57,58] image detected from rendering the photo-realistic mesh and the perspective projection image of dominant structural planes extracted from the point clouds were overlaid and aligned, resulting in a synthetic image as shown in Figure 10. Using advanced 3D rendering software, the rendered image in the top left corner of Figure 10 could achieve a fantastic extent of fidelity when compared with the on-site taken photo in the top right corner of Figure 10, which can meet the requirements of the following computer vision tasks.
According to the two factors of detected image edges and projected model edges that influence the edge tracking quality as explained in Section 2.3, three criteria of recognizing visually prominent edges in a visual-geometric synthetic image are defined as follows:
Criterion 1: Prominent edges should be correlated to dominant structural planes.
The cluttered edges derived from both the structural planes that are relatively too small to the tracking region and the rough slope surface should be excluded. Furthermore, the daylight shadowing edges should also be excluded from the synthetic images. Dominant structural planes can provide useful geometric guidance for recognizing which edges are the traces of structural plane intersections in the Canny edge image. Since the dominant structural planes were 3D expressed by the point clouds filtered with the number of points larger than the threshold, the recognized edges naturally correspond to a relatively large trace length to the tracking region, which makes them visually prominent. In our case study, the threshold for the minimal number of points was semi-automatically adjusted to 1500, according to the relative size of the filtered dominant structural planes to the Tracking Initialization Rectangle.
Criterion 2: Prominent edges should be robustly detectable in the camera image sequences.
The overlapped edges of the dominant structural planes that have large intersection angles with the imaging projection plane, the intermittent edges at the margin of the dominant structural planes that are not sharp enough to be detected as continuous curves, and the cluttered edges derived from the inner fissures, waviness or textures within dominant structural planes should be excluded from the synthetic images since they are not robustly detectable in the camera image sequences.
Criterion 3: Prominent edges should be distinctly extractable from the 3D template model.
The extraction of prominent edges including contour edges and inner edges from 3D template models is a simple geometric process. The contour ones are the edges of 3D meshes, which are effortless to extract wholly, and the inner ones are the edges of mesh faces in the 3D meshes traced along the high curvatures, which can be challenging to wholly extract when the corresponding edges of real rock mass are not sharp enough or when the 3D mesh is oversmoothed.
Then, the terminology prominent structural plane is defined in this paper as the dominant structural plane with edges that are recognized as visually prominent edges in the visual-geometric synthetic images.
In the Tracking Initialization Rectangle magnified in Figure 8, the relatively most tracking-suitable rock mass was searched using the visual-geometric synthetic images captured by the virtual camera. The orientation of the virtual camera was fixed, and the location of the virtual camera could only be moved along a horizontal line segment path. When the observer is faced at the main slope surface, the midpoint of the line segment was denoted as “0.0 m”, at which position the purple pyramid representing the camera sight is magnified in Figure 8; the right hand of the midpoint was denoted with positive distances and the left side with negative distances. In this way, a series of synthetic images were generated. Figure 11 displays some of the synthetic images at different positions on the camera path to reveal the difference in visual-geometric characters of the fractured rock mass. According to Criterion 1 and Criterion 2, to recognize visually prominent edges, Figure 11a shows a superior candidate camera position where more than 10 prominent structural planes were identified, which were mainly distributed in the middle and bottom part of this synthetic image. The top left corner of Figure 11a reveals some cluttered edges of small structural planes, and the top right corner of Figure 11a indicates some overlapped edges of structural planes almost perpendicular to the screen. In contrast, Figure 11b,c show inferior candidate camera positions with only a few structural planes that are not ideal for edge tracking. Figure 11b is almost covered with the cluttered edges of a rough slope surface isolating several dominant structural planes, and conversely, Figure 11c is almost covered by one single dominant structural plane of which the edges are not sharp, but the inner fissures and waviness generate much clutter. Finally, after a comparison of all the candidate positions in this case study, we selected the 0.0 m position on the camera path as the initialization camera pose and the corresponding part of the rock mass in the synthetic image of 0.0 m as the edge-based tracking region.
Subsequently, the proposed Pro-SP method focused on the local part of the rock mass within the intersection volume between the slope surface and the camera sight pyramid, i.e., the rock mass within the purple pyramid in Figure 8, which corresponded to the selected VGSI Figure 11a in this case study. The Pro-SP strategy extracts the prominent structural planes as the 3D template model including the following steps: the joint sets’ extraction from slope point clouds, individual structural plane clustering and dominant structural plane filtering from the joint sets’ point clouds, visual-geometric image synthesis and prominent edge recognition, and corresponding prominent structural plane extraction from the 3D mesh. The Pro-SP template model in the case study is presented in the bottom right corner of Figure 10, where 10 prominent structural planes belonging to joint set 2 at the center of view in Figure 11a were extracted

2.5. Kinematic Analysis of Rockfall Susceptibility

When calculating a rockfall susceptibility map at the cliff scale, the slope surface topography and the characterized fracture patterns measured from the 3D point cloud data can be utilized to search the geometries that serve to destabilize rock mass. The analysis process of rockfall susceptibility can be divided into two steps: the kinematic test and susceptibility index calculation.
In the kinematic test stage, the normal vector to every point of the rock slope point cloud is first calculated concerning its neighboring points. Then, the dip direction and dip of every point is obtained, and the overhanging areas of the cliff are identified, as shown in Figure 12. As a rule of thumb, the volumes located above the overhanging regions are less stable given the lack of support at the base. Accordingly, it is indispensable to take the overhanging areas into account in the field investigation and susceptibility analysis, where the utilization of the UAV SfM or ground-based remote sensing approaches is mandatory to overcome the occlusion problems of airborne acquisitions, and the kinematic failure mechanisms of overhanging conditions need to be separately considered. Subsequently, a well-known kinematic test can be conducted by a comparison between the orientations of the main joint sets and the local cliff areas to find out where planar, wedge and toppling failure is geometrically possible.
Planar failure:
For normal slopes (dz > 0), the point P where a planar failure mode is possible must satisfy geometric conditions for values of the dip direction ( α i ) and dip ( β i ) of joint sets i:
α i t < α p < α i + t
and
β i < β p
where α p is the dip direction of a given point P, β p is the slope angle of the topography at the given point P, and t is the tolerance of the dip direction. The tolerance t is set to 20° in our case study.
For overhanging slopes (i.e., non-normal slopes, dz < 0), all the possible slope steepness is risky to a planar rockslide, so only one condition is demanded [2]:
( 180 + α i ) t < α p < ( 180 + α i ) + t
Wedge failure:
A wedge failure is generated by two joints, respectively labeled i and j. For normal slopes (dz > 0), the point P where a wedge failure mode is possible must also satisfy geometric conditions for values of the trend ( α i j ) and plunge ( β i j ) of the wedge axis:
α i j t < α p < α i j + t
and
β i j < β p
For overhanging slopes (dz < 0) [2]:
( 180 + α i j ) t < α p < ( 180 + α i j ) + t
Toppling detachment:
For normal slopes (dz > 0):
( α i + 180 ) t < α p < ( α i + 180 ) + t
and
90 β i < β p
For overhanging slopes (dz < 0) [2]:
α i t < α p < α i + t
At this point, for each rock compartment of the slope at the cliff scale, we know whether a particular failure mechanism attributed to a given joint set (or a couple of joint sets, for wedge failures) is geometrically possible, but the differences in the quantitative values of failure probability further need to be revealed. Then, the rockfall susceptibility is calculated in these rockfall prone areas to define the slope areas with the orientations most favorable to rockfall. The planar, wedge, and toppling rockfall susceptibility indexes proposed in [2] were adopted in this study, where the susceptibility equations are defined based on the interaction between the occurrences of discontinuities and the occurrences of local cliff, which respectively treat individual conditions for normal and overhanging slopes and consider the parameters including the average spacing and average persistence of each joint set, the mean incidence angle between the joint set (or wedge axis) and the slope surface, the mean slope angle of each joint set (or wedge axis), and the angle between the two joint sets when dealing with the wedge failure. Finally, each point in the three-point cloud sub-groups in Figure 13 was colored according to the calculated rockfall susceptibility index value yielding three susceptibility maps, which were further AR visualized on the corresponding real slopes in Section 3.2.

2.6. Rock Slope Information AR Superimposing

Based on the methods above, the process of mobile AR superimposing of rock slope information can be explained as follows. First, video streams of a rock slope was captured using the built-in camera of a mobile device. Then, the relative six-degree-of-freedom poses of the camera to the rock slope was calculated in real-time based on the approaches presented in Section 2.3 and Section 2.4. Next, the real rock slope, the corresponding virtual 3D point clouds captured by UAV SfM, and the camera were put in the same local coordinate system, where the real rock slope and the virtual 3D point clouds could naturally overlap with each other, and the camera could capture and render this overlapping. Finally, the 3D point clouds, that conveyed the information of join sets obtained in Section 2.2 and rockfall susceptibility obtained in Section 2.5, were superimposed onto the rock slope in these video streams displayed on the screen of the mobile device. The superimposed rockfall susceptibility maps of planar failure, wedge failure, and toppling detachment using mobile augmented reality based on the proposed rock mass 3D template modeling and edge tracking method are shown in Section 3.2.

3. Results

3.1. Pro-SP Template Model Validation

The result of the on-site experiment is presented in this subsection to verify the edge tracking performance promotion of the proposed Pro-SP template modeling method and the influence of illuminance on the tracking performance. A photo of the equipment for on-site testing is shown in Figure 14. A USB camera Logitech C525 was used to capture video streams of the rock mass target and transmit them to the computer. The pose of the USB camera was fixed with a tripod during the experiment. The intensity of illuminance at the position of the USB camera was measured by a TES-1332A illuminance meter.
Then, a comparative experiment of three different template modeling approaches was conducted. Now that the 3D mesh model of this rock mass has already been obtained from the UAV SfM, it can be directly used as a 3D template, as illustrated in Figure 15a, just like the case of CAD models for cars or machines. This straightforward template modeling strategy is referred to as “ALL” in the following sections. Local parts of the 3D mesh can be used as the 3D template model, which is the central part of the rock mass in this case as illustrated in Figure 15b, and this strategy is referred to as “CENTER” in the following sections. The proposed template model is referred to as “Pro-SP’ and illustrated in Figure 15c.
The effectiveness and validity of the proposed Pro-SP method were quantitively verified, where edge tracking of a real rock slope was performed using the USB camera, respectively based on the 3D template model of ALL, CENTER and Pro-SP in Figure 15. During the generation of the projected model edges, two parameters were controlled including the Laplace Threshold, which indicates the minimum depth jump in space to be interpreted as an edge, and the Normal Threshold, which confines the curvature between adjacent surfaces. Figure 15 manifests the edges extracted from these three template models with the Laplace Threshold equal to 12mm and the Normal Threshold equal to 1. It was noticeable that the projected edges of the template model ALL included unnecessary contours and many fragmented line segments, and most of the desired prominent edges were not entirely extracted and projected from both the template models of ALL and CENTER. Adjusting the parameters of the Laplace Threshold and Normal Threshold had some influence on the extracted edges, but could not fix these problems.
To evaluate the matching performance, a valid indicator of the Inlier Ratio was measured [59]. The inlier ratio depicts the ratio of the matched parts of edges among all the projected model edges, and a higher value of the inlier ratio suggests that the edge tracking is stricter and more robust. The software VisLab [60], distributed by the Visometry GmbH and the Fraunhofer IGD in Germany [61,62,63], was used to calculate in real-time the Inlier Ratios of edge tracking. In order to obtain comparable inlier ratio results of the ALL, CENTER, and Pro-SP tracking performance; the following parameters were controlled: these methods were used to track the same rock mass target; the initial poses of these three 3D template models were adjusted to the same location and orientation; the USB cameras were faced straight to the target rock mass, thus maintaining a fixed pose relative to the slope during all of the tracking test processes; and the intensity of illuminance was kept adequate and relative steady. In our case, the pose of the USB camera was adjusted and fixed with the tripod, and the tests were carried out during the noon when the sunlight did not change rapidly, and the illuminance captured by the USB camera kept larger than 5000 LUX. There were only three variables permitted to change in the experiment setting: the 3D template model, the Laplace Threshold, and the Normal Threshold.
The recorded data of inlier ratio in the tracking process capitalizing the ALL, CENTER, and Pro-SP is depicted in the violin plots in Figure 16, respectively. The inlier ratio distributions of Pro-SP dramatically exceeded that of ALL and CENTER, with a higher mean value and a smaller standard deviation. In different scenes, the minimum threshold of the inlier ratio for triggering the edge tracking needs to be assumed and adjusted. Generally, to establish a reliable AR tracking session, a good value to start is setting the minimum initial inlier ratio to 0.6 [60], which means at least sixty percent of the projected model edges are matched. The average inlier ratio of Pro-SP was around 0.65, and almost all the inlier ratio data stayed above 0.6, regardless of the threshold parameters. In contrast, the average inlier ratio of ALL was between 0.45 and 0.6 in most cases, and the mean value of the CENTER was around 0.6. Moreover, the values of the standard deviation of ALL and CENTER were higher than Pro-SP in most cases, which suggests more discrete and less robust tracking performance, and were close in the rest of the cases. Consequently, the high and stable inlier ratios of the Pro-SP template model demonstrated that it could yield strict edge alignment to provide robust tracking with high quality. In contrast, ALL and CENTER are faced with severe problems in launching the AR tracking. Given the inlier ratio threshold of 0.6 for tracking initialization in this case study, according to the distributions of the inlier ratios, the AR application adopting the template model ALL did not recognize the rock mass target in most cases, while, the AR application adopting the CENTER template model sometimes worked and sometimes did not, which was influenced by the parameters for projected model edge generation. In the working cases of the CENTER application, the users may have to take a long time in adjusting the mobile camera pose to align the projected model edges with the real objects for initial recognition, on account of the random variation of inlier ratios floating up or down around the threshold of 0.6, which can result in a terrible AR experience. Furthermore, if we reduced the initialization threshold to a small enough value and forced the tracking based on ALL and CENTER to start, the low level of inlier ratio could lead to more severe jitter problems during the tracking process and could bring about false positives that recognize and track the wrong targets. Above all, the results of this test proved the improvements of the proposed Pro-SP template modeling method for edge tracking performance and quality.
Meanwhile, the influence of the illumination intensity on the tracking performance based on Pro-SP was tested using the same equipment and settings mentioned above. Since field inspections of rockfall hazard generally require to be performed with good sunlight and weather conditions, the continuous tests were conducted under different illumination during the daytime between the sunrise and the sunset. The values of illuminance captured by the USB camera, which were measured by the illuminance meter at the position of the camera, are shown in the x-axis of the violin plots in Figure 17. On the day of the experiment, the sunrise was at 05:07, and the sunset was at 19:15. The maximal illuminance in the noon of that day was around 6650 LUX, the illuminance before and after the sun went down below the mountains in the afternoon were respectively 850 LUX and 790 LUX, and the illuminance near the sunset at 19:05 was 141 LUX. The data trend in Figure 17 was suggestive of a positive correlation between the intensity of illumination and the average inlier ratio. The same setting of the minimum initial inlier ratio with the experiments above, which is 0.6, was adopted. The average inlier ratio transcended 0.7 when the sunlight was sufficient with an illuminance of 6650 LUX and stayed above 6.1 under the poor lighting condition of the dusk with an illuminance of 141 LUX. The standard deviation remained basically the same under different illuminance. The data indicate that the proposed method for rock mass recognition and edge tracking can still maintain effectiveness under the challenging low intensity of illumination after sunrise and before sunset, which was turned out to work well in on-site tests, and the results reveal a relative robustness of the Pro-SP method against the changing of lighting conditions.

3.2. On-Site Rockfall Zoning and Locating in Augmented Reality

As emphasized by [2], the relatively high susceptibility of failure can help define the points of initiation for future rockfalls, so the proposed AR mapping method can remarkably enhance the power of susceptibility maps and extend its potential value to fieldwork guidance, where the critical rock mass of the most likely potential failure can be directly and intuitively located and zoned on-site, as demonstrated in Figure 18 and Figure 19, where an 11-inch iPad Pro was used as the mobile device of the mobile AR application. According to the AR mapping result in Figure 18, some local areas of this slope are inclined to planar and toppling failure. In Figure 19a, a toppling detachment source area was AR zoned at a large structural plane that could clearly be seen as overhanging areas from visual inspection, so the upper rock masses showed a significant probability of detachment and should be considered as the relatively hazardous portions of the cliff, which verifies the validity of proposed AR rockfall locating. In Figure 19b, a planar failure source area was AR zoned where the slope surface was intricately fractured and the critically dangerous rock mass was almost impossible to be directly distinguished by the naked eye, which indicates the necessity of the AR mapping of in-depth analyzed rockfall hazard maps onto real rock slopes beyond simple visual observation.

4. Discussion

The proposed methodology of augmented reality mapping based on rock mass tracking could establish a cyber-physical fusion of geoinformation in the specific scene of rock slope hazard management. On one side, based on the UAV-SfM modeling, the physical cliffs were photo-realistically reconstructed, from which extremely detailed geometric-visual information of rock mass could be explored in the 3D cyberspace just like the on-site survey. On the other side, massive amounts of cyber information, e.g., rockfall susceptibility maps studied in this paper and rockfall simulation results in future research [20], could be embedded into the physical cliffs and extracted based on the visual features of rock mass by taking advantage of the context-aware augmented reality. Some of the potential users of such a tool could be the staff involved in rock slope hazard management. For example, a geological investigator can accurately locate the critical rockfall source areas on the rock slopes based on the primary investigation during on-site inspection, and can better plan the further detailed fieldwork that is directly targeted on the most dangerous rock mass. A designer will be able to preview the simulated rockfall damages and compare different slope supporting designs in the real slope environment. In addition, the constructors guided with the augmented reality information of slope supporting construction will benefit from more efficient communication based on the construction information models.
There are many different strategies and paradigms of augmented reality, among which the most suitable approach for the realization of rockfall AR mapping was explored in this paper. First, considering the large-scale outdoor scene of slope fieldwork, the vision-based mobile AR technology was selected, which could perceive a broad enough range, flexibly adjust perspectives and ensure rendering effects in outdoors. Furthermore, to settle the critical problem in mobile AR of camera pose estimation in the rock slope context [64], the model-based markerless tracking method was adopted which directly recognized the real slope rock mass, and the Simultaneous Localization and Mapping (SLAM) was used for assisting extended tracking. The advantages of this method include relative high overlapping accuracy, compact interaction process, and good feasibility and scalability without the workload of fiducial marker arrangement. Finally, to establish the 3D template model of rock mass target for model-based tracking, based on the Unmanned Aerial Vehicle (UAV) oblique photogrammetry and Structure-from-Motion & Multi-View-Stereo (SfM-MVS) 3D reconstruction, this paper proposed the novel methods visual-geometric synthetic image (VGSI) and prominent structural plane (Pro-SP).
The comparison of the three template modeling strategies in Section 3.1 are further discussed as follows:
When dealing with the specific rock mass target, the template ALL often resulted in poor tracking quality, and sometimes even the initialization of tracking could not be launched. According to Criterion 1 of recognizing visually prominent edges, many bad edges that were not correlated to the dominant structural planes were also extracted from the template model ALL, and further projected and matched with the image edges, which decreased the tracking quality. Furthermore, the manually defined contour edges did not exist in the real slopes, as revealed in Figure 15a, which confused the edge matching.
The template CENTER could, to some extent, overcome these two drawbacks of template ALL. However, even so, according to the Criterion 3 of recognizing visually prominent edges, the preferred edges were contained as inner edges of the 3D template mesh, which could suffer from incomplete extraction depending on the curvature and smoothness of the mesh surfaces.
Conspicuously, the Pro-SP template model excluded the mesh faces and edges that were not correlated to dominant structural planes and conveyed preferred edges as mesh contours that could be distinctly extracted, which could meet Criterion 1 and 3. Moreover, the Pro-SP also considered Criterion 2 in the VGSI analysis and further excluded the dominant structural planes without prominent edges, such as the ones almost perpendicular to the screen in Figure 11a, and the ones rounded with blunt edges that were not sharp enough (Figure 11c). From the perspective of prominent edge criterion, the proposed Pro-SP overcame the shortcomings of the ALL and CENTER strategies.
Different from the indoor lighting in industrial AR applications, the lighting conditions of outdoor AR inspections of rock slopes is controlled by sunlight, which keeps changing during the daytime. In this scene, the edge alignment method can be a more suitable selection for AR tracking realization, because edge alignment abstracts an image as edges with stable distributions and it is naturally robust against the changes in the sunlight conditions, which is not necessarily true of approaches that consider the internal pixels [29,65,66]. The result of the validation test in Figure 17 further demonstrated that the AR tracking based on the proposed template modeling method could maintain a good performance during daytime, which could effectively guarantee its practicability in actual fieldwork.
In the workflow of the proposed AR inspection of rock slope, although the AR initialization process explained in Figure 9 had some limitations, it turned out to be feasible and effective in on-site realization practice. The initial edge tracking will not start anywhere on the rock outcrop. Instead, only the selected local region of the slope rock mass will be recognized from a predefined camera pose, so users need to appropriately adjust the pose of their mobile devices to align the edges and trigger the AR tracking and further inspection. As elaborated in Section 2.3 and Section 2.4, not every part of the slope rock outcrop is suitable for edge tracking initialization. First, recognizing the local rock mass at a very high position of the cliff or just the whole cliff is usually impractical, since the movement of the mobile devices in the hand of a human observer can only change the relative pose of the camera to the towering cliff target in a very limited range, and setting a suitable predefined initial pose for edge alignment becomes tricky. Second, although the surface of the slopes with rockfall risk generally outcrop fractured rock mass, which presents strong visual edge features at the intersections of structural planes, some other parts of the outcrop have an irregular morphology with weak visual edge features. Our AR initialization strategy simplified the first problem by introducing two hypotheses (explained in Section 2.3) to confine the slope rock mass targets for edge tracking to a local rectangular region, referred to as the Tracking Initialization Rectangle (Figure 8), which had a similar altitude and scale with the observers. In practice, the predefined camera pose can be easily reached with the edge aligned. Furthermore, to solve the second problem, we developed the VGSI and Pro-SP methods to select candidate tracking regions with strong edge features. In our case study, we adopted one of these candidate tracking regions, as shown in Figure 11a, which also corresponded to the slope rock mass within the purple pyramid camera sight magnified in Figure 8. One initial tracking region was enough for our case, while the adoption of multiple candidate tracking regions of rock outcrop within the Tracking Initialization Rectangle will undoubtedly increase the flexibility of AR inspection, which deserves further research.
In our case study, the latest AR inspections were performed more than four months after the initial UAV acquisition, and the tracking performance was still good. The proposed method appeared to be relatively robust to the time span between the SfM modeling and the model-based tracking, under the precondition that the slope deformation of rock mass selected as the tracking target was hardly apparent. However, in other cases where a considerable slope deformation or even rockfall occurred in the tracking region, the initial SfM-derived template model could lose the match with the actual rock mass edge features and may stop working. One interesting future research direction is to study the robustness of the edge tracking to slope deformation through long-term observations and on-site testing on rapidly deforming slopes. In practical applications, if an edge alignment stops working due to slope deformation or rockfalls, a simple solution is to carry out the SfM modeling of the local part of the rock mass within the tracking region again, align the regional model with the original slope model using 3D registration, and use it as an updated new template model. This solution could be acceptable if the template model updating frequency is not too high.

5. Conclusions

This paper presented some pioneering studies on augmented reality (AR) mapping of geospatial information beyond cyberspace in 2D screens to the physical world, which directly recognizes the slope rock mass and superimposes corresponding rock discontinuities and rockfall susceptibility onto the real slopes with mobile devices for the on-site locating of potentially unstable rock masses. A novel method of edge-based tracking of the rock mass target for mobile augmented reality was proposed. The 3D template model of the rock mass for edge tracking was extracted from the unmanned aerial vehicle (UAV) structure-from-motion (SfM) reconstructions, including 3D point clouds and photo-realistic 3D meshes. The edges extracted from the template model were aligned with the corresponding actual rock mass to estimate the camera pose and realize AR 3D registration. Specifically, the prominent edges of dominant structural planes were first explored and discovered to be a robust visual feature for edge tracking, and the criteria for recognizing visually prominent edges coupling the visual and geometrical dimensions of the photogrammetric 3D models were established. First, the joint sets and individual structural planes were segmented and clustered from the 3D point clouds. Subsequently, a novel method called the visual-geometric synthetic image (VGSI) was proposed to search for tracking-suitable rock mass on the virtual outcrop of the slope. The Canny edges in the images rendered from photo-realistically meshes were detected, and the dominant structural planes segmented from joint point clouds were projected and overlaid with the Canny edges in synthetic images to identify the prominent edges. Then, a new approach called the prominent structural plane (Pro-SP) was proposed for the 3D template modeling of fractured rock mass by extracting structural planes with prominent edges from the 3D meshes. Meanwhile, the kinematic analysis was conducted according to the interaction between the occurrences of main joint sets and the occurrences of local cliff to generate rockfall susceptibility maps. Finally, the joint sets and susceptibility maps were superimposed onto the real slopes based on edge tracking using the Pro-SP template model.
The novel contributions of this paper include the following five aspects:
  • The methodology for the cyber-physical association of geospatial information in the augmented reality environment of rock slopes was established by superimposing cyber geoinformation onto the physical rock slope;
  • A novel method of potential rockfall on-site zoning and locating with mobile augmented reality was first proposed to enhance rockfall hazard management;
  • To improve the scalability of AR tracking in large outdoor scenes of cliff-scale slopes, a new strategy was developed by combining a model-based approach for tracking initialization and maintaining with the simultaneous localization and mapping (SLAM) approach for tracking extending, where the model-based approach recognizes the local part of the slope using UAV-SfM-reconstructed 3D models and the SLAM approach extends the tracking when the local target slope is outside the camera view;
  • To realize robust model-based tracking, the visual edge features of the fractured rock mass target for edge-based tracking was researched for the first time, where the visually prominent edges at the intersections of dominant structural planes of rock mass were discovered to be valid and reliable features for edge alignment;
  • The novel methods of VGSI and Pro-SP were proposed to search for tracking-suitable rock mass on the photogrammetric 3D models and build 3D template models with compelling edge features to be matched with real rock mass. Some pioneering research on coupling the visual dimension and the geometrical dimension of the SfM-reconstructed 3D models of the fractured rock mass was performed. The VGSI method identifies the prominent edges among the visually detectable edges in the photo-realistically rendered slope images and the geometrically extractable edges in segmented rock joint point clouds. The Pro-SP method extracts the structural planes that best convey the prominent edges as 3D template models, which significantly improved the edge tracking quality.
In future research, the terrestrial laser scanning technology can be combined with UAV photogrammetry to improve the accuracy of 3D virtual outcrop geometry, which can further facilitate the robustness of model-based AR tracking and the resolution of potential rockfall zoning. The influence of the rock slope deformation on the edge tracking deserves subsequent research. In the concept of the cyber-physical information model, beyond the rock discontinuities and rockfall susceptibility maps, more information of rock slopes can also be superimposed and overlapped on real cliffs [67,68]. And the proposed rock mass AR tracking paradigm can be further connected with geographic information system (GIS) and construction information model (CIM) systems to extend its potential.

Author Contributions

Conceptualization, Y.Z. and D.Z.; Data curation, Y.Z.; Formal analysis, Y.Z., M.L. and T.G.; Investigation, Y.Z. and M.L.; Methodology, Y.Z.; Visualization, Y.Z. and T.G.; Writing—original draft, Y.Z.; Writing—review & editing, P.Y., G.Z. and T.G.

Funding

This research was funded by the National Key R&D Program of China, grant number 2018YFC0406704; the Yalong River Joint Funds of the National Natural Science Foundation of China, grant number U1765205; and the National Key R&D Program of China, grant number 2018YFC0407101.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Emmer, A. Geographies and scientometrics of research on natural hazards. Geosciences 2018, 8, 382. [Google Scholar] [CrossRef]
  2. Matasci, B.; Stock, G.M.; Jaboyedoff, M.; Carrea, D.; Collins, B.D.; Guérin, A.; Matasci, G.; Ravanel, L. Assessing rockfall susceptibility in steep and overhanging slopes using three-dimensional analysis of failure mechanisms. Landslides 2017, 15, 859–878. [Google Scholar] [CrossRef]
  3. Matasci, B.; Jaboyedoff, M.; Loye, A.; Pedrazzini, A.; Derron, M.H.; Pedrozzi, G. Impacts of fracturing patterns on the rockfall susceptibility and erosion rate of stratified limestone. Geomorphology 2015, 241, 83–97. [Google Scholar] [CrossRef]
  4. Zhao, C.; Lu, Z. Remote sensing of landslidesa review. Remote Sens. 2018, 10, 279. [Google Scholar] [CrossRef]
  5. Pirasteh, S.; Li, J. Developing an algorithm for automated geometric analysis and classification of landslides incorporating lidar-derived dem. Environ. Earth Sci. 2018, 77, 414. [Google Scholar] [CrossRef]
  6. Pirasteh, S.; Li, J. Landslides investigations from geoinformatics perspective: Quality, challenges, and recommendations. Geomat. Nat. Hazards Risk 2017, 8, 448–465. [Google Scholar] [CrossRef]
  7. Francioni, M.; Salvini, R.; Stead, D.; Coggan, J. Improvements in the integration of remote sensing and rock slope modelling. Nat. Hazards 2018, 90, 975–1004. [Google Scholar] [CrossRef]
  8. Jaboyedoff, M.; Oppikofer, T.; Abellan, A.; Derron, M.-H.; Loye, A.; Metzger, R.; Pedrazzini, A. Use of lidar in landslide investigations: A review. Nat. Hazards 2012, 61, 5–28. [Google Scholar] [CrossRef]
  9. Olsen, M.J.; Wartman, J.; McAlister, M.; Mahmoudabadi, H.; O’Banion, M.S.; Dunham, L.; Cunningham, K. To fill or not to fill: Sensitivity analysis of the influence of resolution and hole filling on point cloud surface modeling and individual rockfall event detection. Remote Sens. 2015, 7, 12103–12134. [Google Scholar] [CrossRef]
  10. Liu, L.; Xiao, J.; Wang, Y. Major orientation estimation-based rock surface extraction for 3d rock-mass point clouds. Remote Sens. 2019, 11, 635. [Google Scholar]
  11. Riquelme, A.J.; Abelian, A.; Tomas, R.; Jaboyedoff, M. A new approach for semi-automatic rock mass joints recognition from 3d point clouds. Comput. Geosci. 2014, 68, 38–52. [Google Scholar] [CrossRef]
  12. Chen, J.; Zhu, H.; Li, X. Automatic extraction of discontinuity orientation from rock mass surface 3d point cloud. Comput. Geosci. 2016, 95, 18–31. [Google Scholar] [CrossRef]
  13. Mazzanti, P.; Schiliro, L.; Martino, S.; Antonielli, B.; Brizi, E.; Brunetti, A.; Margottini, C.; Mugnozza, G.S. The contribution of terrestrial laser scanning to the analysis of cliff slope stability in sugano (central italy). Remote Sens. 2018, 10, 1475. [Google Scholar] [CrossRef]
  14. Sayab, M.; Aerden, D.; Paananen, M.; Saarela, P. Virtual structural analysis of jokisivu open pit using ’structure-from-motion’ unmanned aerial vehicles (uav) photogrammetry: Implications for structurally-controlled gold deposits in southwest finland. Remote Sens. 2018, 10, 1926. [Google Scholar] [CrossRef]
  15. Giordan, D.; Hayakawa, Y.; Nex, F.; Remondino, F.; Tarolli, P. Review article: The use of remotely piloted aircraft systems (rpass) for natural hazards monitoring and management. Nat. Hazards Earth Syst. Sci. 2018, 18, 1079–1096. [Google Scholar] [CrossRef]
  16. Westoby, M.J.; Brasington, J.; Glasser, N.F.; Hambrey, M.J.; Reynolds, J.M. ‘Structure-from-motion’ photogrammetry: A low-cost, effective tool for geoscience applications. Geomorphology 2012, 179, 300–314. [Google Scholar] [CrossRef]
  17. Thoeni, K.; Santise, M.; Guccione, D.E.; Fityus, S.; Roncella, R.; Giacomini, A. Use of low-cost terrestrial and aerial imaging sensors for geotechnical applications. Aust. Geomech. J. 2018, 53, 101–122. [Google Scholar]
  18. Eltner, A.; Kaiser, A.; Castillo, C.; Rock, G.; Neugirg, F.; Abellan, A. Image-based surface reconstruction in geomorphometry—Merits, limits and developments. Earth Surface Dyn. 2016, 4, 359–389. [Google Scholar] [CrossRef]
  19. Thoeni, K.; Giacomini, A.; Lambert, C.; Sloan, S.W.; Carter, J.P. A 3d discrete element modelling approach for rockfall analysis with drapery systems. Int. J. Rock Mech. Min. Sci. 2014, 68, 107–119. [Google Scholar] [CrossRef]
  20. Sarro, R.; Riquelme, A.; Carlos Garcia-Davalillo, J.; Maria Mateos, R.; Tomas, R.; Luis Pastor, J.; Cano, M.; Herrera, G. Rockfall simulation based on uav photogrammetry data obtained during an emergency declaration: Application at a cultural heritage site. Remote Sens. 2018, 10, 1923. [Google Scholar] [CrossRef]
  21. Cawood, A.J.; Bond, C.E.; Howell, J.A.; Butler, R.W.H.; Totake, Y. Lidar, uav or compass-clinometer? Accuracy, coverage and the effects on structural models. J. Struct. Geol. 2017, 98, 67–82. [Google Scholar] [CrossRef]
  22. Bemis, S.P.; Micklethwaite, S.; Turner, D.; James, M.R.; Akciz, S.; Thiele, S.T.; Bangash, H.A. Ground-based and uav-based photogrammetry: A multi-scale, high-resolution mapping tool for structural geology and paleoseismology. J. Struct. Geol. 2014, 69, 163–178. [Google Scholar] [CrossRef]
  23. Gazcon, N.F.; Trippel Nagel, J.M.; Bjerg, E.A.; Castro, S.M. Fieldwork in geosciences assisted by argeo: A mobile augmented reality system. Comput. Geosci. 2018, 121, 30–38. [Google Scholar] [CrossRef]
  24. Suh, J.; Lee, S.; Choi, Y. UmineAR: Mobile-tablet-based abandoned mine hazard site investigation support system using augmented reality. Minerals 2017, 7, 198. [Google Scholar] [CrossRef]
  25. Bae, H.; Golparvar-Fard, M.; White, J. Image-based localization and content authoring in structure-from-motion point cloud models for real-time field reporting applications. J. Comput. Civil Eng. 2015, 29. [Google Scholar] [CrossRef]
  26. Cejka, J.; Bruno, F.; Skarlatos, D.; Liarokapis, F. Detecting square markers in underwater environments. Remote Sens. 2019, 11, 459. [Google Scholar] [CrossRef]
  27. Fan, Y.; Feng, Z.; Mannan, A.; Khan, T.U.; Shen, C.; Saeed, S. Estimating tree position, diameter at breast height, and tree height in real-time using a mobile phone with rgb-d slam. Remote Sens. 2018, 10, 1485. [Google Scholar] [CrossRef]
  28. Shan, G.-l.; Ji, B.; Zhou, Y.-f. A review of 3d pose estimation from a monocular image sequence. In Proceedings of the 2009 2nd International Congress on Image and Signal Processing, Tianjin, China, 17–19 Ocotber 2009; Volume 1;–9, pp. 1946–1950. [Google Scholar]
  29. Fua, V.L.a.P. Monocular model-based 3d tracking of rigid objects: A survey. Found. Trends Comput. Graph. Vis. 2005, 1, 1–89. [Google Scholar]
  30. Marchand, E.; Uchiyama, H.; Spindler, F. Pose estimation for augmented reality: A hands-on survey. IEEE Trans. Vis. Comput. Graph. 2016, 22, 2633–2651. [Google Scholar] [CrossRef]
  31. Park, J.; Seo, B.-K.; Park, J.-I. Binocular mobile augmented reality based on stereo camera tracking. J. Real-Time Image Process. 2017, 13, 571–580. [Google Scholar] [CrossRef]
  32. Li, C.; Gao, X. Adaptive contour feature and color feature fusion for monocular textureless 3d object tracking. IEEE Access 2018, 6, 30473–30482. [Google Scholar] [CrossRef]
  33. Han, P.; Zhao, G. Line-based initialization method for mobile augmented reality in aircraft assembly. Vis. Comput. 2017, 33, 1185–1196. [Google Scholar] [CrossRef]
  34. Wang, J.; Yan, W. Fast pose estimation for texture-less objects based on b-rep model. Eurasip J. Image Video Process. 2018, 2018, 117. [Google Scholar] [CrossRef]
  35. Ploetz, T.; Roth, S. Automatic registration of images to untextured geometry using average shading gradients. Int. J. Comput. Vis. 2017, 125, 65–81. [Google Scholar] [CrossRef]
  36. Loesch, A.; Bourgeois, S.; Gay-Bellile, V.; Gomez, O.; Dhome, M. Localization of 3d objects using model-constrained slam. Mach. Vis. Appl. 2018, 29, 1041–1068. [Google Scholar] [CrossRef]
  37. Wang, Y.; Wang, J.; Chen, X.; Chu, T.; Liu, M.; Yang, T. Feature surface extraction and reconstruction from industrial components using multistep segmentation and optimization. Remote Sens. 2018, 10, 1073. [Google Scholar] [CrossRef]
  38. Ni, H.; Lin, X.; Ning, X.; Zhang, J. Edge detection and feature line tracing in 3d-point clouds by analyzing geometric properties of neighborhoods. Remote Sens. 2016, 8, 710. [Google Scholar] [CrossRef]
  39. Malihi, S.; Zoej, M.J.V.; Hahn, M. Large-scale accurate reconstruction of buildings employing point clouds generated from uav imagery. Remote Sens. 2018, 10, 1148. [Google Scholar] [CrossRef]
  40. Wu, C. Towards linear-time incremental structure from motion. In Proceedings of the 2013 International Conference on 3d Vision (3dv 2013), Seattle, WA, USA, 29 June–1 July 2013; pp. 127–134. [Google Scholar]
  41. Wu, C.; Agarwal, S.; Curless, B.; Seitz, S.M. Multicore bundle adjustment. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Colorado Springs, CO, USA, 20–25 June 2011. [Google Scholar]
  42. Furukawa, Y.; Ponce, J. Accurate, dense, and robust multiview stereopsis. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 32, 1362–1376. [Google Scholar] [CrossRef]
  43. Furukawa, Y.; Curless, B.; Seitz, S.M.; Szeliski, R. Towards internet-scale multi-view stereo. In Proceedings of the 23rd IEEE Conference on Computer Vision and Pattern Recognition (CVPR), San Francisco, CA, USA, 13–18 June 2010; pp. 1434–1441. [Google Scholar]
  44. Meshlab. Available online: http://www.meshlab.net/ (accessed on 5 March 2019).
  45. Cignoni, P.; Corsini, M.; Dellepiane, M.; Ganovelli, F.; Ranzuglia, G. Meshlab: An open-source mesh processing tool. In Proceedings of the Sixth Eurographics Italian Chapter Conference, Salerno, Italy, 2–4 July 2008; pp. 129–136. [Google Scholar]
  46. Kazhdan, M.; Hoppe, H. Screened poisson surface reconstruction. ACM Trans. Graph. 2013, 32, 29. [Google Scholar] [CrossRef]
  47. Corsini, M.; Cignoni, P.; Scopigno, R. Efficient and flexible sampling with blue noise properties of triangular meshes. IEEE Trans. Vis. Comput. Graph. 2012, 18, 914–924. [Google Scholar] [CrossRef] [PubMed]
  48. Vasuki, Y.; Holden, E.-J.; Kovesi, P.; Micklethwaite, S. Semi-automatic mapping of geological structures using uav-based photogrammetric data: An image analysis approach. Comput. Geosci. 2014, 69, 22–32. [Google Scholar] [CrossRef]
  49. Morikubo, Y.; Hashimoto, N. Edge-based object tracking for dynamic projection mapping. In Proceedings of the International Workshop on Advanced Image Technology (IWAIT), Chiang Mai, Thailand, 7–9 January 2018. [Google Scholar]
  50. Han, P.; Zhao, G. Cad-based 3d objects recognition in monocular images for mobile augmented reality. Comput. Graph.-UK 2015, 50, 36–46. [Google Scholar] [CrossRef]
  51. Lu, T.; Ming, D.; Lin, X.; Hong, Z.; Bai, X.; Fang, J. Detecting building edges from high spatial resolution remote sensing imagery using richer convolution features network. Remote Sens. 2018, 10, 1496. [Google Scholar] [CrossRef]
  52. Vuforia.Developer Portal. Available online: https://developer.vuforia.com/ (accessed on 5 March 2019).
  53. Unity. Available online: https://unity3d.com/unity/ (accessed on 5 March 2019).
  54. Arkit. Available online: https://developer.apple.com/arkit/ (accessed on 5 March 2019).
  55. Arcore. Available online: https://developers.google.com/ar/ (accessed on 5 March 2019).
  56. Chen, R.; Li, X.; Li, J. Object-based features for house detection from rgb high-resolution images. Remote Sens. 2018, 10, 451. [Google Scholar] [CrossRef]
  57. Canny, J. A computational approach to edge detection. IEEE Trans. Pattern Anal. Mach. Intell. 1986, 8, 679–698. [Google Scholar] [CrossRef] [PubMed]
  58. Shi, X.; Jiang, J. Automatic registration method for optical remote sensing images with large background variations using line segments. Remote Sens. 2016, 8, 426. [Google Scholar] [CrossRef]
  59. Wang, H.-R.; Lei, J.; Li, A.; Wu, Y.-H. A geometry-based point cloud reduction method for mobile augmented reality system. J. Comput. Sci. Technol. 2018, 33, 1164–1177. [Google Scholar] [CrossRef]
  60. Vislab. Available online: https://visionlib.com/products/vislab/ (accessed on 5 March 2019).
  61. Wuest, H.; Engekle, T.; Wientapper, F.; Schmitt, F.; Keil, J. From CAD to 3D Tracking - Enhancing & Scaling Model-based Tracking for Industrial Appliances. In Proceedings of the 2016 IEEE International Symposium on Mixed and Augmented Reality (ISMAR-Adjunct), Merida, Yucatan, Mexico, 19–23 September 2016; pp. 346–347. [Google Scholar]
  62. Seo, B.-K.; Wuest, H. A Direct Method for Robust Model-Based 3D Object Tracking from a Monocular RGB Image. In Computer Vision—ECCV 2016 Workshops; Lecture Notes in Computer Science; Springer: Cham, Switzerland, 2016; Volume 9915, pp. 551–562. [Google Scholar] [CrossRef]
  63. Seo, B.-K.; Wuest, H. Robust 3D Object Tracking Using an Elaborate Motion Model. In Proceedings of the 2016 IEEE International Symposium on Mixed and Augmented Reality (ISMAR-Adjunct), Merida, Yucatan, Mexico, 19–23 September 2016; pp. 70–71. [Google Scholar] [CrossRef]
  64. Cai, G.; Su, S.; Leng, C.; Wu, Y.; Lu, F. A robust transform estimator based on residual analysis and its application on uav aerial images. Remote Sens. 2018, 10, 291. [Google Scholar] [CrossRef]
  65. Qiu, K.; Liu, T.; Shen, S. Model-based global localization for aerial robots using edge alignment. IEEE Robot. Autom. Lett. 2017, 2, 1256–1263. [Google Scholar] [CrossRef]
  66. Ling, Y.; Kuse, M.; Shen, S. Edge alignment-based visual-inertial fusion for tracking of aggressive motions. Auton. Robots 2018, 42, 513–528. [Google Scholar] [CrossRef]
  67. Dai, K.; Li, Z.; Tomas, R.; Liu, G.; Yu, B.; Wang, X.; Cheng, H.; Chen, J.; Stockamp, J. Monitoring activity at the daguangbao mega-landslide (china) using sentinel-1 tops time series interferometry. Remote Sens. Environ. 2016, 186, 501–513. [Google Scholar] [CrossRef]
  68. Zhang, Y.; Zhong, D.; Wu, B.; Guan, T.; Yue, P.; Wu, H. 3D parametric modeling of complex geological structures for geotechnical engineering of dam foundation based on t-splines. Comput. Aided Civil Infrastruct. Eng. 2018, 33, 545–570. [Google Scholar] [CrossRef]
Figure 1. The difficulty in building intuitive cognition of correlation between real rock slopes and corresponding slope analysis results can be solved by the proposed augmented reality mapping.
Figure 1. The difficulty in building intuitive cognition of correlation between real rock slopes and corresponding slope analysis results can be solved by the proposed augmented reality mapping.
Remotesensing 11 01311 g001
Figure 2. The research framework. The photogrammetry remote sensing part is presented in Section 2.1; the rock discontinuity characterization part is clarified in Section 2.2; the markerless mobile augmented reality (AR) part is elaborated in Section 2.3; the vision-feature-based 3D template modeling part is articulated in Section 2.4, the rockfall stability analysis part is illustrated in Section 2.5; and the rock slope information AR superimposing and AR mapping part is explained in Section 2.6.
Figure 2. The research framework. The photogrammetry remote sensing part is presented in Section 2.1; the rock discontinuity characterization part is clarified in Section 2.2; the markerless mobile augmented reality (AR) part is elaborated in Section 2.3; the vision-feature-based 3D template modeling part is articulated in Section 2.4, the rockfall stability analysis part is illustrated in Section 2.5; and the rock slope information AR superimposing and AR mapping part is explained in Section 2.6.
Remotesensing 11 01311 g002aRemotesensing 11 01311 g002b
Figure 3. The unmanned aerial vehicle (UAV) device (DJI Mavic Pro) used in this study for photogrammetric remote sensing.
Figure 3. The unmanned aerial vehicle (UAV) device (DJI Mavic Pro) used in this study for photogrammetric remote sensing.
Remotesensing 11 01311 g003
Figure 4. The case study place and the SfM-MVS modeling process. The studied cliff is 19 meters high.
Figure 4. The case study place and the SfM-MVS modeling process. The studied cliff is 19 meters high.
Remotesensing 11 01311 g004
Figure 5. (a) The stereographic projection of the rock discontinuity orientation derived from the virtual outcrop point cloud generated from drone images, and (b) A 3D visualization of these pole density isolines. J1 to J8 represent the eight clustered joint sets. The scale bar illustrates the values of pole density corresponding to different colors.
Figure 5. (a) The stereographic projection of the rock discontinuity orientation derived from the virtual outcrop point cloud generated from drone images, and (b) A 3D visualization of these pole density isolines. J1 to J8 represent the eight clustered joint sets. The scale bar illustrates the values of pole density corresponding to different colors.
Remotesensing 11 01311 g005
Figure 6. The segmented point clouds of joint sets. J1 to J8 represent the eight clustered joint sets corresponding to different colors. Only five joint sets J2, J3, J4, J5, and J7 were considered in further rock fall susceptibility analysis according to the percentage of point number in each joint set to the total point number.
Figure 6. The segmented point clouds of joint sets. J1 to J8 represent the eight clustered joint sets corresponding to different colors. Only five joint sets J2, J3, J4, J5, and J7 were considered in further rock fall susceptibility analysis according to the percentage of point number in each joint set to the total point number.
Remotesensing 11 01311 g006
Figure 7. The edge-based camera tracking mechanism and process presented in [30,50], the differences in edge features between the artifact tracking target and the rock mass tracking target, and the proposed methods to overcome these gaps.
Figure 7. The edge-based camera tracking mechanism and process presented in [30,50], the differences in edge features between the artifact tracking target and the rock mass tracking target, and the proposed methods to overcome these gaps.
Remotesensing 11 01311 g007
Figure 8. The rock mass tracking region Tracking Initialization Rectangle, the virtual camera sight pyramids at different camera positions and the dominant structural planes in the tracking region. J1 to J6 represent the six clustered joint sets corresponding to different colors. The shown cliff is 19 meters high.
Figure 8. The rock mass tracking region Tracking Initialization Rectangle, the virtual camera sight pyramids at different camera positions and the dominant structural planes in the tracking region. J1 to J6 represent the six clustered joint sets corresponding to different colors. The shown cliff is 19 meters high.
Remotesensing 11 01311 g008
Figure 9. The AR realization process combining model-based tracking and visual-inertial odometry (VIO).
Figure 9. The AR realization process combining model-based tracking and visual-inertial odometry (VIO).
Remotesensing 11 01311 g009
Figure 10. The workflow of the proposed visual-geometric synthetic image (VGSI) and prominent structural plane (Pro-SP) methods.
Figure 10. The workflow of the proposed visual-geometric synthetic image (VGSI) and prominent structural plane (Pro-SP) methods.
Remotesensing 11 01311 g010
Figure 11. (a), (b) and (c) illustrate the synthesis process of the images captured by three different virtual cameras at different camera positions defined in Figure 8, respectively. In each subfigure, the canny edge image detected from a rendered image of photo-realistic mesh models is shown in the top left corner, the image of the point clouds of dominant structural planes is shown in the top right corner, and the image synthesizing the canny edges and the point clouds is shown in the bottom. J1 to J6 represent the six clustered joint sets corresponding to different colors. The numbers labeled in these images indicate the recognized prominent structural planes (Pro-SP) in the visual-geometric synthetic images (VGSI), and the same number in different images indicate the same structural plane.
Figure 11. (a), (b) and (c) illustrate the synthesis process of the images captured by three different virtual cameras at different camera positions defined in Figure 8, respectively. In each subfigure, the canny edge image detected from a rendered image of photo-realistic mesh models is shown in the top left corner, the image of the point clouds of dominant structural planes is shown in the top right corner, and the image synthesizing the canny edges and the point clouds is shown in the bottom. J1 to J6 represent the six clustered joint sets corresponding to different colors. The numbers labeled in these images indicate the recognized prominent structural planes (Pro-SP) in the visual-geometric synthetic images (VGSI), and the same number in different images indicate the same structural plane.
Remotesensing 11 01311 g011aRemotesensing 11 01311 g011b
Figure 12. The normal parts of the cliff are shown with green point clouds on the 3D mesh of the rock slope in (a), and the overhanging parts of the cliff are shown with red point clouds on the 3D mesh of the rock slope in (b).
Figure 12. The normal parts of the cliff are shown with green point clouds on the 3D mesh of the rock slope in (a), and the overhanging parts of the cliff are shown with red point clouds on the 3D mesh of the rock slope in (b).
Remotesensing 11 01311 g012
Figure 13. The point cloud sub-groups as 3D maps of rockfall susceptibility to (a) planar failure, (b) wedge failure, and (c) toppling detachment, respectively. The color of every point illustrates the value of rockfall susceptibility, and the location of each point indicates the portions of the cliff that are geometrically possible for rockfalls.
Figure 13. The point cloud sub-groups as 3D maps of rockfall susceptibility to (a) planar failure, (b) wedge failure, and (c) toppling detachment, respectively. The color of every point illustrates the value of rockfall susceptibility, and the location of each point indicates the portions of the cliff that are geometrically possible for rockfalls.
Remotesensing 11 01311 g013
Figure 14. The equipment for on-site testing of edge tracking performance.
Figure 14. The equipment for on-site testing of edge tracking performance.
Remotesensing 11 01311 g014
Figure 15. Three different 3D template models of the target rock mass in Figure 11a. (a) is a possible 3D template model corresponding to the rock mass in all the view field, referred to as “ALL”. (b) is another a possible 3D template model corresponding to the rock mass in the center of the view field, referred to as “CENTER”. (c) is the proposed 3D template model: prominent structural planes, referred to as “Pro-SP”. The first line of figures illustrates the 3D meshes of these template models in the same pose, and the second line illustrates the edges extracted from these meshes. The third line shows the images captured by the USB camera in Figure 14, where the template model meshes are aligned with the real rock mass through adjusting the pose of the USB camera, and the fourth line shows the corresponding aligned edges. The Laplace Threshold was set to 12mm and the Normal Threshold was set to 1.
Figure 15. Three different 3D template models of the target rock mass in Figure 11a. (a) is a possible 3D template model corresponding to the rock mass in all the view field, referred to as “ALL”. (b) is another a possible 3D template model corresponding to the rock mass in the center of the view field, referred to as “CENTER”. (c) is the proposed 3D template model: prominent structural planes, referred to as “Pro-SP”. The first line of figures illustrates the 3D meshes of these template models in the same pose, and the second line illustrates the edges extracted from these meshes. The third line shows the images captured by the USB camera in Figure 14, where the template model meshes are aligned with the real rock mass through adjusting the pose of the USB camera, and the fourth line shows the corresponding aligned edges. The Laplace Threshold was set to 12mm and the Normal Threshold was set to 1.
Remotesensing 11 01311 g015
Figure 16. Violin plots of inlier ratio data of edge tracking based on three different template models with different values of Laplace Threshold and Normal Threshold.
Figure 16. Violin plots of inlier ratio data of edge tracking based on three different template models with different values of Laplace Threshold and Normal Threshold.
Remotesensing 11 01311 g016
Figure 17. Violin plots of inlier ratio data of edge tracking using Pro-SP under different intensities of illumination.
Figure 17. Violin plots of inlier ratio data of edge tracking using Pro-SP under different intensities of illumination.
Remotesensing 11 01311 g017
Figure 18. The screen shots of an iPad showing the AR mapping of rockfall susceptibility to the (a) planar failure, (b) wedge failure and (c) toppling detachment.
Figure 18. The screen shots of an iPad showing the AR mapping of rockfall susceptibility to the (a) planar failure, (b) wedge failure and (c) toppling detachment.
Remotesensing 11 01311 g018
Figure 19. On-site locating and zoning of potential rockfalls through mobile augmented reality using an iPad, including (a) a toppling detachment source area and (b) a planner failure source area.
Figure 19. On-site locating and zoning of potential rockfalls through mobile augmented reality using an iPad, including (a) a toppling detachment source area and (b) a planner failure source area.
Remotesensing 11 01311 g019

Share and Cite

MDPI and ACS Style

Zhang, Y.; Yue, P.; Zhang, G.; Guan, T.; Lv, M.; Zhong, D. Augmented Reality Mapping of Rock Mass Discontinuities and Rockfall Susceptibility Based on Unmanned Aerial Vehicle Photogrammetry. Remote Sens. 2019, 11, 1311. https://doi.org/10.3390/rs11111311

AMA Style

Zhang Y, Yue P, Zhang G, Guan T, Lv M, Zhong D. Augmented Reality Mapping of Rock Mass Discontinuities and Rockfall Susceptibility Based on Unmanned Aerial Vehicle Photogrammetry. Remote Sensing. 2019; 11(11):1311. https://doi.org/10.3390/rs11111311

Chicago/Turabian Style

Zhang, Yichi, Pan Yue, Guike Zhang, Tao Guan, Mingming Lv, and Denghua Zhong. 2019. "Augmented Reality Mapping of Rock Mass Discontinuities and Rockfall Susceptibility Based on Unmanned Aerial Vehicle Photogrammetry" Remote Sensing 11, no. 11: 1311. https://doi.org/10.3390/rs11111311

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop