Next Article in Journal
Effect of Muscle-Specific Fatigue on the Risk of Anterior Cruciate Ligament Injury in Females
Previous Article in Journal
RF Exposure Assessment for Various Poses of Patient Assistant in Open MRI Environment
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Research on Visual Positioning of a Roadheader and Construction of an Environment Map

1
School of Mechanical Electronic & Information Engineering, China University of Mining and Technology-Beijing, Beijing 100083, China
2
Fujian Provincial Key Laboratory of Information Processing and Intelligent Control (Minjiang University), Fuzhou 350121, China
3
School of Mechanics and Civil Engineering, China University of Mining and Technology-Beijing, Beijing 100083, China
4
China Energy Information Technology Co., Ltd., Beijing 100011, China
5
Intelligent Mine (Coal Industry) Engineering Research Center, Beijing 100011, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2021, 11(11), 4968; https://doi.org/10.3390/app11114968
Submission received: 29 April 2021 / Revised: 17 May 2021 / Accepted: 21 May 2021 / Published: 28 May 2021
(This article belongs to the Section Robotics and Automation)

Abstract

:
The autonomous positioning of tunneling equipment is the key to intellectualization and robotization of a tunneling face. In this paper, a method based on simultaneous localization and mapping (SLAM) to estimate the body pose of a roadheader and build a navigation map of a roadway is presented. In terms of pose estimation, an RGB-D camera is used to collect images, and a pose calculation model of a roadheader is established based on random sample consensus (RANSAC) and iterative closest point (ICP); constructing a pose graph optimization model with closed-loop constraints. An iterative equation based on Levenberg–Marquadt is derived, which can achieve the optimal estimation of the body pose. In terms of mapping, LiDAR is used to experimentally construct the grid map based on open-source algorithms, such as Gmapping, Cartographer, Karto, and Hector. A point cloud map, octree map, and compound map are experimentally constructed based on the open-source library RTAB-MAP. By setting parameters, such as the expansion radius of an obstacle and the updating frequency of the map, a cost map for the navigation of a roadheader is established. Combined with algorithms, such as Dijskra and timed-elastic-band, simulation experiments show that the combination of octree map and cost map can support global path planning and local obstacle avoidance.

1. Introduction

With the rapid development of the industrial Internet, big data, and artificial intelligence, the intellectualization and robotization of equipment have become an inevitable trend in industrial development. Based on literature data from the Web of Science database, a cluster analysis of hotspots in intelligent coal mine research, from 2010 to 2020, obtained using VOSviewer software, is shown in Figure 1. It reveals that the in-depth integration of intelligent technology and the development of the coal industry is inevitable. For China’s coal industry, the degree of automation and intelligence of fully mechanized mining faces has been significantly improved over years of innovative development, and has been applied practically in many mining areas [1,2]. However, the intelligent development of fully mechanized excavation faces is still in a bottleneck period due to the poor working environments and complex excavation processes; the imbalance between mining and tunneling seriously restricts the safety and efficiency of coal mine production. Therefore, it is urgent to carry out research on the intelligence of tunneling equipment and systems for a fully mechanized excavation face [3,4].
As the main equipment for coal roadway driving, autonomous positioning of roadheaders is a key link to realize intelligent tunneling, and many scholars have done a great deal of original work in this area. An indoor global position system (IGPS)-based pose measurement method for roadheaders was proposed by Tao et al. [5]. The position coordinates of a receiver can be obtained from a single transmitter station, which realized the autonomous measurement of the body pose in a multi-point time-sharing manner. A remote-controlled system for roadheaders was developed by Roman et al. [6]. The system had built-in active lighting infrared sensors, which could assist in the positioning of a roadheader without visual control by workers. Du et al. [7] studied the pose measurement method of roadheaders based on machine vision. A laser target was fixed on a roadheader to receive the laser beam. After the imaging features were collected using a monocular camera, the roadheader pose can accurately be solved by solving the pose model and performing error compensation. Mao et al. [8] explored a multi-source fusion measurement method based on ultrasound, inertial navigation, a laser, and geomagnetic information. The method could overcome the shortcoming of excessive accumulated errors in a single inertial navigation measurement and effectively monitor the roadheader pose. Fu et al. [9,10] developed a set of roadheader pose measurement systems based on an ultra-wideband. A time-of-arrival model was built after the distance between the positioning node of a roadheader and ultra-wideband (UWB) positioning base station was measured. After the positioning algorithm was solved, the roadheader’s pose information could be obtained. Yang et al. [11] presented a vision measurement method for the pose of roadheader’s cutting head, based on infrared LEDs. A 16-point infrared target was fixed on the roadheader, and then the target image was collected and processed using a monocular camera; the pose of the cutting head could then be estimated. Simulation and experimental results showed that the accuracy of pose measurements could meet the requirements of underground roadway construction. Yan et al. [12] investigated a roadheader pose measurement method based on multi-source information, such as a gyroscope, a total station, and an inertial measurement system, and built a positioning model. Experimental results showed that this method could meet precision requirements. Xue et al. [13] put forward a positioning method for roadheaders based on binocular vision technology. By collecting information with a binocular vision sensor, the coordinate model between a roadheader and the roadway space was established, and then the pose was calculated. Simulations and experiments showed that the accuracy of the calculated positioning parameters was good and the established positioning model was reasonable. Yang et al. [14] proposed a vision measurement method for the pose of a roadheader body based on laser beams. The laser beams were used for the image feature, and a measurement system of roadheader body pose, based on monocular vision, was established. Experimental results showed that this method could effectively solve body pose information of roadheaders. A pose measurement method for a roadheader, combining total station and strapdown inertial navigation, was proposed by Zhang et al. [15]. Total station and inertial navigation technology were used to measure the pose of roadheaders, respectively, and the pose was determined using fusion data, which reduced the influence of dust in the mine.
The above research covers the main schemes of present pose measurement of roadheaders. Total station has high installation flexibility, but its light path is easily blocked, and its real-time performance is poor and cannot adapt to the dynamic working conditions of a roadheader. An inertial sensor is generally used as an auxiliary measurement method because of its large cumulative error. UWB technology has a strong real-time performance, but its signal is greatly affected by the complex electromagnetic environment of coal mines. IGPS technology has a large measurement range, but the cost of installation and construction is high. Visual measurements based on target imaging are susceptible to the influence of light beam scattering, and the installation of a light source emitter is limited by complicated roadway structures. The composition of the combined measurement system is complex, and the multi-source data are difficult to synchronize accurately in time and space; thus, the data reliability is not high. State perception, trajectory planning, and motion control are the three keys to realize intelligent and robotized roadheaders. As a core technology of autonomous positioning and navigation for mobile devices, simultaneous localization and mapping (SLAM) is widely used in mobile robot navigation, unmanned driving, and in other fields [16,17,18]. For instance, the open source library, Real-Time Appearance-Based Mapping (RTAB-MAP), provides a positioning and mapping solution that is independent of time and scale, which can be applied to various setups [19]. According to different sensor types, SLAM technology is mainly divided into laser-based and vision-based [20]. The point cloud obtained by LiDAR is sparse, a created map lacks semantic information, and the linearization error is large. In contrast, vision sensors can obtain far more status information than LiDARs. Furthermore, the RGB-D (red, green, blue, and depth) camera in a vision sensor is based on active ranging technology, and it has the advantages of not being affected by ambient lighting changes and textures compared with monocular and binocular cameras.
In summary, SLAM technology is a better solution to realize autonomous positioning and navigation of roadheaders and roadway mapping in a GPS-free roadway environment. At present, there are no reports on the application of visual SLAM technology to coal roadway roadheaders, and research on this technology is still lacking. Therefore, the positioning requirements of roadheaders and the objective conditions, such as tunnel environment and sensor attributes, are comprehensively considered, a method of body pose estimation for roadheaders, based on visual SLAM, is proposed, and multiple schemes of environmental map construction based on laser and visual SLAM are tested. These can provide information for pose correction, navigation of roadheaders, and three-dimensional reconstruction of roadways.

2. Positioning Foundation of the Roadheader and Modeling

2.1. Definition of Coordinate System and Pose

The positioning problem of a roadheader is transformed into a state estimation problem in this paper; that is, to solve the optimal pose. To accurately describe the pose of a roadheader in three-dimensional (3D) space, a series of coordinate systems need to be established, an earth coordinate system OeXeYeZe, roadway coordinate system OtXtYtZt, world coordinate system OwXwYwZw, and camera coordinate system OcXcYcZc. The Oe system is fixed on the earth, which is the basis of the design of the roadway; the Ot system is the absolute reference of the location in the roadway, and its position is determined by the structure of the roadway; the Ow system is the reference frame for the pose description of an airborne camera when it is moving, and the Oc and Ow systems are related to rotation and translation.
As the input device of visual SLAM, an RGB-D camera can be installed in the sensing system of a roadheader through intrinsic safety technology. Considering that the camera is connected to the roadheader body, the pose of the roadheader’s body is expressed by the pose of the RGB-D camera relative to the Ow system. If the absolute pose of the roadheader body in the Ot system or the Oe system is required, it can be obtained using Euclidean coordinate transformation. In addition, the pose of the cutting head of a cantilever roadheader can be solved according to the Denavit–Hartenberg model and forward kinematics in robotics [21]. In sum, the pose acquisition of a roadheader body in the Ow system is the basis for realizing other functions.
At the initial moment, it is stipulated that the Oc and Ow systems are coincident. The Xw axis of the Ow system is located in the horizontal plane and is perpendicular to the roadway midline, pointing to the right of the heading direction. The Yw axis is located in the horizontal plane and is parallel to the roadway midline, pointing to the heading direction; the Zw axis is perpendicular to the horizontal plane and points to the roadway roof. After the roadheader moves, the pose of the Oc system changes relative to the Ow system, and the pose of the roadheader body (camera) is described by axis deflection angles α, β, and γ, and axis displacements tx, ty, and tz, as shown in Figure 2. α is defined as the included angle between the projection of the axis in the tunneling direction on the OwXwYw plane and the Yw axis, namely the yaw angle (Figure 2a). β is defined as the included angle between the projection of the axis in the tunneling direction on the OwYwZw plane and the Yw axis, namely the pitch angle (Figure 2b). γ is defined as the included angle between the projection of the transverse axis of the roadheader on the OwXwZw plane and the Xw axis, namely the roll angle (Figure 2c); tx, ty, and tz are, respectively, the displacement of the roadheader body in the direction of Xw, Yw, and Zw axes (Figure 2d–f).
According to rigid body kinematics, the motion of a roadheader can be decomposed into rotation and translation. The rotation of the body (camera) in the Ow system is described by rotation matrix R; Rcw means the rotation transformation from the Ow system to the Oc system. Specifically, point Pw(xw, yw, zw) in the Ow system is rotated and transformed to point Pc(xc, yc, zc) in the Oc system, which can be described as: (1) rotating at angle α along the Zw axis, as shown in Equation (1); (2) rotating at angle β along the Xw axis, as shown in Equation (2); and (3) rotating at angle γ along the Yw axis, as shown in Equation (3). R has a definite transformation relationship with the α, β, and γ, as shown in Equation (4) [22].
[ x c y c z c ] = [ cos α sin α 0 sin α cos α 0 0 0 1 ] [ x w y w z w ]
[ x c y c z c ] = [ 1 0 0 0 cos β sin β 0 sin β cos β ] [ x w y w z w ]
[ x c y c z c ] = [ cos γ 0 sin γ 0 1 0 sin γ 0 cos γ ] [ x w y w z w ]
R cw = Rot ( Z w , α ) Rot ( X w , β ) Rot ( Y w , γ ) = [ cos α cos γ sin α sin β sin γ sin α cos β cos α sin γ + sin α sin β cos γ sin α cos γ + cos α sin β sin γ cos α cos β sin α sin γ cos α sin β cos γ cos β sin γ sin β cos β cos γ ]
The translation of the body (camera) is described by 3D vector t, and tcw represents the translation transformation from the Ow system to the Oc system, as shown in Equation (5). After homogeneous transformation, Rcw and tcw can be unified into transformation matrix Tcw, as shown in Equation (6) [22]. In the absence of ambiguity, Rcw, tcw, and Tcw below are referred to as R, t, and T without subscripts.
t c w = [ t x t y t z ] T
T c w = [ R c w t c w 0 T 1 ]

2.2. SLAM Mathematical Model of Roadheader

In the moving process of the roadheader, the airborne camera observes the environmental signposts and estimates its own motion. As shown in Figure 3, the SLAM process of a roadheader equipped with a camera can be visualized, and the pose state and signposts are represented by nodes, and processes of camera movement and observation are represented by edges. Further, at time i, let the pose variable of the airborne camera be xi, the position variable of waypoint is yj, the motion input is ui, and the observation data of airborne camera are zi,j; then, the mathematical model of visual SLAM for the roadheader is shown as Equation (7).
{ x i = f p ( x i 1 , u i , w i ) i ( 1 , , m ) z i , j = g o ( y j , x i , v i , j ) j ( 1 , , n )
where fp and go are the process function and observation function, respectively; wi and vi,j are the motion noise and observation noise at time i, respectively.

3. Pose Estimation and Optimization of the Roadheader

3.1. Pose Estimation Scheme

The Pose estimation scheme of the roadheader (camera) is shown in Figure 4. Specifically, (1) the images of the roadway environment are collected by the airborne RGB-D camera and depth information is obtained; (2) feature extraction and matching are performed on the color images, and 3D matching pairs are formed after depth mapping; (3) combined with the random sample consensus (RANSAC) strategy and the iterative closest point (ICP) strategy, the pose of the roadheader is estimated; (4) the pose graph, based on keyframe and closed-loop constraints, is constructed, and the global pose optimization is carried out to obtain the optimal pose estimation of the roadheader.

3.2. Image Feature Extraction and Matching

Images collected by the RGB-D camera need feature calculations and data associations, and commonly used methods include the direct method, optical flow method, and feature point method [23]. Both the direct method and the optical flow method rely too heavily on the assumption of a constant gray level, so it is difficult for them to be directly used in a roadway environment; the feature point method, however, is less affected by motion and illumination and is relatively stable. In coal mine roadways, the forming quality of a section is affected by the real-time and accurate measurement of the roadheader pose. Compared with other features, ORB (oriented fast and rotated brief) features perform better in terms of real-time performance and accuracy, and ORB features are not sensitive to rotation and translation [24,25]. Consequently, ORB features are used for extraction and matching in this paper.
To verify the effect of ORB feature extraction and matching, two images from the public dataset, TUM-RGBD, were selected for the experiment. The experimental results are shown in Figure 5. Figure 5a shows the effect obtained after using the Brute-Force matcher and Figure 5b shows the effect after matching filtering, and the filtering basis is that the Hamming distance is less than twice the minimum distance.

3.3. Pose Solution

Based on image feature extraction and matching, an initial motion transform with a high confidence is iteratively solved using the RANSAC method, which aims to provide reliable initial values for pose estimation based on ICP. The process is as follows: (1) According to the matched 3D point set, the initial model is generated by randomly sampling some point pairs; (2) the model is evaluated according to the given threshold by calculating the point-pair error; and (3) through calculating the number of interior point pairs and the average value of the error, the model is updated and iterated based on the given threshold. After several iterations, a relatively optimal motion transformation can be obtained [26,27].
After depth mapping and RANSAC iteration, a set of 3D point sets with exact matching is obtained, i.e., P = {p1, p2, …, pn}, Q = {q1, q2, …, qn}, the body pose transformation between the i-th pair of 3D matching points in the point sets can be written as:
p i = R q i + t
According to the SLAM mathematical model of a roadheader (Equation (7)), there are noises in the motion equation and observation equation of the airborne camera; thus, the state estimation is uncertain. The inter-frame error function, e(3D-3D)i, is positioned as [28]:
e ( 3 D - 3 D ) i = p i ( R q i + t )
For n pairs of 3D matching points in the point set, the least squares expression of the objective function, F(R, t), can be expressed as [29,30]:
F ( R , t ) = arg min R , t 1 2 i = 1 n p i ( R q i + t ) 2
To solve Equation (10), let the de-centroid coordinates of point sets P and Q be p i and q i , respectively, that is:
{ p i = p i 1 n i = 1 n p i q i = q i 1 n i = 1 n q i
Combining Equations (10) and (11), the de-centroid form can be obtained as:
F ( R , t ) = arg min R , t 1 2 i = 1 n ( p i R q i 2 + 1 n i = 1 n p i ( R 1 n i = 1 n q i + t ) 2 )
Equation (12) can be decomposed into two functions, f1(R) and f2(t), namely:
f 1 ( R ) = arg min R 1 2 i = 1 n p i R q i 2
f 2 ( t ) = arg min t 1 2 i = 1 n 1 n i = 1 n p i R 1 n i = 1 n q i t 2
Further, Equation (13) can be expanded and simplified as:
g 1 ( R ) = arg min R [ t r ( R i = 1 n q i ( p i ) T ) ]
where, tr() is the trace operator of matrix.
To understand the R contained in Equation (15), a method based on singular value decomposition is adopted, that is, let [31,32]:
i = 1 n q i ( p i ) T = U D V T
where D, U, and V are all diagonal matrices, and matrix D is composed of singular values.
According to the principle of Cauchy–Schwarz’s inequality, when the objective function g1(R) in Equation (15) takes the minimum value and DUVT is full rank, R can be obtained, the expression is:
R = U V T
After R is solved, let f2(t) in Equation (14) be zero, and t can be solved, i.e.:
t = 1 n i = 1 n p i R 1 n i = 1 n q i
So far, through the construction of the least squares problem of pose solution and singular value decomposition, the pose parameters, R and t, of the roadheader body can be solved.

3.4. Pose Optimization

Pose optimization must be carried out because of the accumulated errors in pose solutions. Waypoints are regarded as constraints of pose optimization rather than as optimization variables; the optimization model of a body pose graph is established by selecting the image key frame. Additionally, closed-loop constraints are added to carry out global pose optimization, so as to ensure the effectiveness and real-time performance of pose estimation, and then the deviation correction is completed in time.

3.4.1. Pose Graph Construction and Closed-Loop Detection

The selection principle of key frames is based on the evaluation of the distance between frames. After R and t are given different weights, the motion transformation index representing the distance between frames can be calculated. If the index is less than a given threshold, the current frame is determined to be a key frame.
After selecting the key frame, the pose graph of the motion transformation constraint can be constructed. To further modify the computed pose, closed-loop detection is needed and closed-loop constraints are added to the pose graph. The key of closed-loop detection lies in the judgment of the similarity between two frames of images, i.e., if the similarity between the current frame and the historical key frame exceeds a certain value of similarity between the current frame and the previous-moment key frame, and the consistency of space and time is satisfied, a closed loop can be determined.
Based on the SLAM motion model of the roadheader, shown in Figure 3, the pose graph optimization model, which eliminates the waypoint variables and adds closed-loop constraints, is shown in Figure 6. The figure shows two processes of motion transformation and closed-loop detection, Ti(i−1) is the motion transformation constraint between pose nodes xi−1 and xi, and T(i+1)n is the closed-loop constraint between pose nodes xi+1 and xn.

3.4.2. Pose Graph Optimization

According to the pose graph optimization model shown in Figure 6, let the inter-frame transformation matrix obtained in the pose solution stage or closed-loop detection be Tij, and the pose estimation error be eij; then, eij can be expressed as [33]:
e i j = ln ( T i j 1 T i 1 T j )
where the sign ∨ means to transform the matrix into a vector.
To optimize the global pose graph, the least squares expression of the objective function, G(T), based on the error term is constructed, that is [34]:
G ( T ) = arg min T i j e i j T Σ i j 1 e i j
where Σij is the covariance matrix of the pose measurement, which is used to represent the uncertainty of the measurement.
Equation (20) is nonlinear and is solved using an iterative method in this paper. Let ∆x be the iterative increment, x* be the initial value of the pose estimation, and then the Taylor expansion form of the error, eij, at x* can be expressed as:
e i j ( x + Δ x ) = e i j + J i j Δ x
where Jij is the Jacobian matrix of eij at x*.
Combined with Equations (20) and (21), Equation (22) can be obtained after approximate processing, that is:
{ H Δ x = b H = J i j T Σ i j 1 J i j b = i j e i j T Σ i j 1 J i j
By introducing Lagrange multiplier ρ on the basis of Equation (22), the increment equation based on the Levenberg–Marquadt method is shown in Equation (23):
( H + ρ Ι ) Δ x = b
where I is the identity matrix.
After multiple iterations of Equation (23), when the termination condition of the iteration is satisfied, the optimized pose estimation can be obtained. Subsequently, a closed-loop control and deviation correction can be carried out for the traveling of the roadheader according to the optimized body pose, and trajectory planning of the cutting arm can be performed according to robotics to ensure the forming quality of the roadway section.

4. Map Construction and Test

4.1. Overview of Mapping Scheme

Although the acquired pose can provide a priori information for the moving of the roadheader, its navigation also depends on the environment map. Based on SLAM technology, a roadway map is established, which can help the roadheader identify obstacles and judge roadway edge shapes, and provide support for remote man–machine interactions, roadheader travel prediction, section forming control, and so on.
Common mapping schemes are shown in Table 1, including laser-based and vision-based schemes, and the real triangle symbol indicates that the algorithm supports the relevant interface. It can be seen from Table 1 that the RTAB-MAP algorithm can provide a variety of sensor input interfaces, and can output a dense 3D point cloud map, which is currently an excellent application-based mapping solution.
Comprehensively considering the underground environment of a coal mine as well as the existing conditions of a laboratory, a long corridor is adopted to simulate an underground roadway, and some mapping algorithms, listed in Table 1, are tested. The long corridor used for test, as shown in Figure 7, is as close as possible to the dimly lit, narrow and closed roadway environment.

4.2. Construction of Grid Map and Test

4.2.1. Principle of Laser Mapping

To construct an occupancy grid map based on laser SLAM, the key is to use the prior information to calculate the posterior probability of the map, and then identify the grid state, i.e., unknown, idle, or occupied [41,42]. The grid state schemas formed after two-dimensional (2D) LiDAR scanning are shown in Figure 8; OgXgYg and ObXbYb are the grid coordinate system and body coordinate system, respectively. The white area in the grid is idle, the black area in the grid is occupied, and the gray area in the grid is unknown, which is the target area for further exploration using SLAM technology.
The measured data after the laser beam is emitted by the 2D LiDAR is denoted as zL, and is as shown in Equation (24). According to probability theory, if s = 1 that means that the grid is occupied, and if s = 0 that means that the grid is idle; then, the updated rule of state S for any grid is shown in Equation (25).
z L = [ d θ ] T
where d represents the distance of the measuring point, and θ represents the included angle between the laser beam and the body orientation.
S i = S i 1 + log p ( z | s = 1 ) p ( z | s = 0 )
where p() is the probability calculation expression.
The map constructed using laser SLAM can provide a priori information for navigation and obstacle avoidance of the roadheader, but it is not safe to move in the occupancy grid map using only Boolean values [42]. To realize global navigation of the roadheader and to avoid dynamic obstacles in real time, an obstacle map layer and expansion layer should be added, based on the static map layer constructed using SLAM; that is, a cost map for navigation should be constructed. The value principle of a cost map is shown in Figure 9, where each grid state is represented by a gray value from 0 to 255. The dashed line frame in the figure represents the contours of the roadheader, and the grid value of 254 indicates that the obstacles coincide with the center of the body, which will inevitably lead to a collision. The single dotted circle represents the inner tangent circle of the body contour line, the grid value in the inner tangent circle is 253, and collision is also inevitable in this area. The double-dot dotted circle represents the outer tangential circle of the body contour line, the grid value range in the outer tangential circle is 128–252, and collisions may occur in this area; the three-point scribbled circle represents the buffer area, and the grid value range is 1–128, which should be avoided. The grid value in the three-point double-dashed circle is 0, and the area is allowed free passage. The gray block with a grid value of 255 represents an unknown area, which needs to be updated according to the sensor information [43].

4.2.2. Mapping Test Based on Laser SLAM

Laser SLAM is divided into filtering-based and optimization-based, the Gmapping algorithm [35] in Table 1 belongs to the former, and the algorithms, Cartographer [36], Karto [37], and Hector [38], belong to the latter. In this paper, the four abovementioned laser SLAM algorithms are used to carry out mapping tests of the simulated roadway along the long corridor shown in Figure 7.
The test results of the four algorithms are shown in Figure 10. Figure 10a shows a map of simulated roadway, built based on the Gmapping algorithm, and the Gmapping algorithm relies on the Rao–Blackwellized particle filter, which takes laser data and odometer data as inputs; the map constructed based on the Cartographer algorithm is shown in Figure 10b. The Cartographer algorithm consists of two phases of local and global SLAM, with the first phase establishing and maintaining grid submaps, and the second phase eliminating cumulative errors and carrying out global optimization. Figure 10c shows a map built based on the Karto algorithm, which is based on the graph optimization theory, and uses the graph mean to represent the map, and the front-end and back-end are carried out by a single thread. The map constructed based on the Hector algorithm is shown in Figure 10d; the Hector algorithm only creates a grid map by matching laser data without the need for odometer data.
The Gmapping algorithm based on the filtering framework has a lower computational complexity, but it relies too heavily on odometer data. Although the Hector algorithm is not limited to the odometer, it requires a high laser frequency. In addition, the Cartographer algorithm is relatively resource intensive. Consequently, based on the 2D occupy grid map constructed using the Karto algorithm, the cost map for the navigation of the roadheader is constructed by introducing parameters, such as the expansion radius of obstacles and the updating frequency of map. The simulation test results are shown in Figure 11, the red arrow represents the body pose.

4.3. Construction of 3D Map and Test

4.3.1. Principle of Vision Mapping

According to Table 1, the map constructed using the ORB-SLAM1/2 algorithm is a sparse point cloud map, which cannot be used for navigation and obstacle avoidance of the roadheader. In addition, the algorithm can easily miss a target when the airborne camera rotates and it is sensitive to noise. Considering the complex environment of underground roadways and the requirements of tunneling work, the performance of running and being in real-time, the accuracy of algorithm should be given more priority. The RTAB-MAP framework provides a map construction scheme that is independent of scale and time. As shown in Figure 12 [19], its internal memory management is implemented through three memory nodes, i.e., WM (working memory), LTM (long-term memory), and STM (short-term memory). The weight of the registration point is updated in STM, and the closed-loop detection of the registration point is completed in WM. When the number of registration points exceeds a certain threshold, the registration point with the longest storage time is placed in WM from STM. To ensure real-time Bayesian estimation of the number of registration points, two functional nodes, namely fetch and transfer, are defined between WM and LTM during closed-loop detection in WM. When the closed loop is detected in the WM, the adjacent location points in the closed loop will be retrieved from the LTM and placed in the WM. When the processing time for closed-loop detection exceeds a certain threshold, the anchor point with the lowest weight is transferred to the LTM [19]. Based on the above memory management mechanism, the RTAB-MAP framework avoids the problem of having map data that are too large, which can meet the engineering needs of long-term, real-time, and online mapping in a large-scale environment.
The RTAB-MAP algorithm can build a dense point cloud map. Considering the disadvantage of a large amount of data in a dense point cloud, a portable stored octree map can be formed after octree compression [44,45], which can be used for stereoscopic navigation of the roadheader. The data structure of the octree is shown in Figure 13; each parent node contains eight child nodes, among which the leaf layer node corresponds to the map with the highest resolution. In addition, each node points to a cube voxel, and the probability of the n-th node corresponding to the voxel being occupied is calculated according to Bayesian theory, as shown in Equation (26) [45].
p ( n | z 1 : t c ) = [ 1 + 1 p ( n | z t c ) p ( n | z 1 : t c ) 1 p ( n | z 1 : t - 1 c ) p ( n | z 1 : t - 1 c ) 1 p ( n ) p ( n ) ] 1
where zc represents the observation data of the airborne camera at the corresponding subscript time.

4.3.2. Mapping Test Based on Vision SLAM

Based on RTAB-MAP framework, the image data collected by RGB-D camera are input, the dense 3D point cloud map is constructed by mapping the test of the simulated roadway (shown in Figure 14), and the octree map is constructed via octree compression and is shown in Figure 15.
Octree map is convenient for navigation of the roadheader because of its portable characteristics. Based on the octree map shown in Figure 15, the multi-layer cost map for navigation is constructed by introducing parameters, such as the expansion radius of obstacles and the frequency of map updates, as shown in Figure 16. The red arrow in the figure represents the body pose, the darkly colored expansion area around the body is a local cost map, which is used for local path planning; the rest of the area is a global cost map, which is used for global path planning. In this test, the Dijskra algorithm is used for global path planning, and the red lines in the figure are the trajectory of the global path planning. The TEB (timed-elastic-band) algorithm is used for local path planning, and the blue line is the trajectory of local path planning.
As can be seen from Figure 16, when an obstacle is temporarily added in the navigation test, the dynamic obstacle expansion region can be marked by updating the cost map, and the obstacle avoidance based on local path planning is successfully realized.

4.3.3. Mapping Test Based on Fusion of Laser and Vision

Although the mapping accuracy of laser SLAM is high and the computational complexity is low, the map constructed is a sparse point cloud, lacks semantic information, and has large linearization errors. In contrast, the environmental information obtained by the camera is much more than that obtained by LiDAR, and the semantic information is rich. Moreover, the RGB-D camera has the advantage of not being affected by the change in ambient light and texture, but its mapping accuracy is affected by the cumulative error. Based on the above considerations, the two mapping schemes are complementary. Specifically, in the dynamic environment of a coal mine roadway with strong interference, the mapping scheme based on the laser and vision fusion has a stronger robustness.
The RTAB-MAP framework provides an interface between the camera and LiDAR [19], as shown in Figure 17. First, the laser and visual data are entered, then the multi-source information is initialized and aligned, and finally a map of the environment is output.
The mapping result of the simulated roadway, based on the fusion of laser and visual data, is shown in Figure 18. The map is in a composite form, and includes a dense 3D point cloud and a 2D occupied grid. Compared with Figure 14, it can be seen that the SLAM mapping scheme that integrates laser and vision has more advantages in detail processing of the point cloud, and the map is more refined. Additionally, the point cloud can be used for 3D reconstruction of the environment after gridding and texture mapping in the later stage, and the 2D occupancy grid map can be used for roadheader navigation after being converted into a cost map.

5. Conclusions

In this paper, according to the development requirements of intelligent and robotization of fully mechanized excavation faces, the method of estimating roadheader poses and constructing navigation maps by using vision and laser SLAM technology is proposed. The main conclusions are as follows:
(1)
The positioning method based on visual SLAM technology is proposed, the environment data are collected by an airborne RGB-D camera, and the RANSAC+ICP model is constructed to solve and realize the autonomous measurement of pose by establishing and solving the optimization model of pose graph with closed-loop constraints. The global consistent pose of the roadheader can be obtained, which can effectively solve the difficult positioning and orientation problem of roadheaders in a restricted space.
(2)
A series of open-source algorithms, such as Gmapping, Cartographer, Karto, Hector, and RTAB-MAP, are applied to construct simulated roadway maps, including 2D occupied grid, 3D point cloud, and octree, based on a static map created by SLAM. The cost map for roadheader navigation is further constructed by setting parameters, such as expansion radius of obstacles and the updating frequency of map. Furthermore, obstacle avoidance tests based on the Dijskra+TEB algorithm show that a combination of octree map and cost map can support global path planning and local dynamic obstacle avoidance.
(3)
Based on the open-source library RTAB-MAP framework, by fusing laser and vision data, a composite map composed of a dense 3D point cloud and a 2D occupancy grid is constructed, which could, not only be used for plane navigation of a roadheader, but also provide support for 3D reconstruction of an environment.

Author Contributions

Conceptualization, G.Z. and W.Z.; methodology, W.Z.; software, G.Z. and W.Z.; validation, W.Z., G.Z., and Z.Y.; formal analysis, G.Z., Z.Y., and T.P.; investigation, W.Z.; data curation, W.Z. and R.C.; writing—original draft preparation, W.Z.; writing—review and editing, W.Z. and G.Z.; visualization, W.Z. and R.C.; supervision, G.Z., Z.Y., and T.P.; funding acquisition, G.Z. and T.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Fundamental Research Funds for the Central Universities (Grant No. 2021YJSJD10); the National Key Research and Development Program of China (Grant No. 2017YFC0804307); Fujian Provincial Key Laboratory of Information Processing and Intelligent Control (Minjiang University) (Grant No. MJUKF-IPIC201905).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wang, G.; Xu, Y.; Ren, H. Intelligent and ecological coal mining as well as clean utilization technology in China: Review and prospects. Int. J. Min. Sci. Technol. 2019, 29, 161–169. [Google Scholar] [CrossRef]
  2. Wang, J.; Huang, Z. The Recent Technological Development of Intelligent Mining in China. Engineering 2017, 3, 439–444. [Google Scholar] [CrossRef]
  3. Tian, J.; Wang, S.; Wu, M. Kinematic models and simulations for trajectory planning in the cutting of Spatially-Arbitrary crosssections by a robotic roadheader. Tunn. Undergr. Space Technol. 2018, 78, 115–123. [Google Scholar] [CrossRef]
  4. Deshmukh, S.; Raina, A.K.; Murthy, V.M.S.R.; Trivedi, R.; Vajre, R. Roadheader—A comprehensive review. Tunn. Undergr. Space Technol. 2020, 95, 103148. [Google Scholar] [CrossRef]
  5. Tao, Y.F.; Zong, K.; Zhang, M.J.; Jia, W.H.; Fu, S.C.; Yang, J.J.; Wu, M. Aposition and orientationmeasurement method of single-station, multipoint and time-sharing forroadheaderbody based on iGPS. J. China Coal Soc. 2015, 40, 2611–2616. [Google Scholar]
  6. Roman, M.; Josif, B.; Nikolay, D. Development of Position System of a Roadheader on a Base of Active IR-sensor. In Proceedings of the 25th Daaam International Symposium on Intelligent Manufacturing and Automation, Oldest Technical Univ, Vienna, Australia, 26–29 November 2015; pp. 617–621. [Google Scholar]
  7. Du, Y.; Tong, M.; Liu, T.; Dong, H. Visual measurement system for roadheaders pose detection in mines. Opt. Eng. 2016, 55, 104107. [Google Scholar] [CrossRef]
  8. Mao, Q.H.; Zhang, X.H.; Ma, H.W.; Xue, X.S. Study on spatial position and posture monitoring system of boom-type roadheader based on multi sensor information. Coal Sci. Technol. 2018, 46, 41–47. [Google Scholar]
  9. Fu, S.-C.; Li, Y.; Zhang, M.; Zong, K.; Cheng, L.; Wu, M. Ultra-wideband pose detection system for boom-type roadheader based on Caffery transform and Taylor series expansion. Meas. Sci. Technol. 2017, 29, 015101. [Google Scholar] [CrossRef] [Green Version]
  10. Fu, S.; Li, Y.; Zong, K.; Liu, C.; Liu, D.; Wu, M. Ultra-wideband pose detection method based on TDOA positioning model for boom-type roadheader. AEU Int. J. Electron. Commun. 2019, 99, 70–80. [Google Scholar] [CrossRef]
  11. Yang, W.; Zhang, X.; Ma, H.; Zhang, G.-M. Infrared LEDs-Based Pose Estimation with Underground Camera Model for Boom-Type Roadheader in Coal Mining. IEEE Access 2019, 7, 33698–33712. [Google Scholar] [CrossRef]
  12. Yan, C.; Zhao, W.; Lu, X. A Multi-Sensor Based Roadheader Positioning Model and Arbitrary Tunnel Cross Section Automatic Cutting. Sensors 2019, 19, 4955. [Google Scholar] [CrossRef] [Green Version]
  13. Xue, X.S.; Zhang, X.H.; Mao, Q.H.; Zheng, J.K.; Wang, M. Localization and orientation method of roadheader robot based on binocular vision. J. Xi’an Univ. Sci. Technol. 2020, 40, 781–789. [Google Scholar]
  14. Yang, W.; Zhang, X.; Ma, H.; Zhang, G. Laser beams-based localization methods for Boom-type roadheader using underground camera non-uniform blur model. IEEE Access 2020, 8, 1. [Google Scholar] [CrossRef]
  15. Zhang, X.H.; Liu, B.X.; Zhang, C.; Yang, W.J.; Zhao, J.X. Roadheader positioning method combining total station and strapdown inertial navigation system. Ind. Mine Autom. 2020, 46, 1–7. [Google Scholar]
  16. Debeunne, C.; Vivet, D. A review of Visual-LiDAR fusion based simultaneous localization and mapping. Sensors 2020, 20, 20. [Google Scholar] [CrossRef] [Green Version]
  17. Zaffar, M.; Ehsan, S.; Stolkin, R.; Maier, K.M. Sensors, SLAM and Long-term Autonomy: A Review. In Proceedings of the 2018 NASA/ESA Conference on Adaptive Hardware and Systems (AHS), Edinburgh, UK, 6–9 August 2018; Institute of Electrical and Electronics Engineers (IEEE): New York, NY, USA, 2018; pp. 285–290. [Google Scholar]
  18. Cadena, C.; Carlone, L.; Carrillo, H.; Latif, Y.; Scaramuzza, D.; Neira, J.; Reid, I.D.; Leonard, J.J. Past, Present, and Future of Simultaneous Localization and Mapping: Toward the Robust-Perception Age. IEEE Trans. Robot. 2016, 32, 1309–1332. [Google Scholar] [CrossRef] [Green Version]
  19. Labbé, M.; Michaud, F. RTAB-Map as an open-source lidar and visual simultaneous localization and mapping library for large-scale and long-term online operation. J. Field Robot. 2019, 36, 416–446. [Google Scholar] [CrossRef]
  20. Saeedi, S.; Trentini, M.; Seto, M.; Li, H. Multiple-Robot Simultaneous Localization and Mapping: A Review. J. Field Robot. 2016, 33, 3–46. [Google Scholar] [CrossRef]
  21. Liu, Z.X.; Wang, S.A.; Xie, M.; Chen, H.Y.; Miao, C.; Xie, C.X. Characteristics and modeling of roadway surface topography under influences of cutting head geometry. China Mech. Eng. 2020, 31, 2583–2591. [Google Scholar]
  22. Xing, H.M.; Shi, L.W.; Tang, K.; Guo, S.X.; Hou, X.H.; Liu, Y.; Liu, H.K.; Hu, Y. Robust RGB-D camera and IMU fusion-based cooperative and relative close-range localization for multiple turtle-inspired amphibious spherical robots. J. Bionic Eng. 2019, 16, 442–454. [Google Scholar] [CrossRef]
  23. Sualeh, M.; Kim, G.-W. Simultaneous Localization and Mapping in the Epoch of Semantics: A Survey. Int. J. Control. Autom. Syst. 2019, 17, 729–742. [Google Scholar] [CrossRef]
  24. Endres, F.; Hess, J.; Sturm, J.; Cremers, D.; Burgard, W. 3-D Mapping With an RGB-D Camera. IEEE Trans. Robot. 2014, 30, 177–187. [Google Scholar] [CrossRef]
  25. Sharif, H.; Hoelzel, M. A comparison of prefilters in ORB-based object detection. Pattern Recognit. Lett. 2017, 93, 154–161. [Google Scholar] [CrossRef]
  26. Senthooran, I.; Murshed, M.; Barca, J.C.; Kamruzzaman, J.; Chung, H. An efficient RANSAC hypothesis evaluation using sufficient statistics for RGB-D pose estimation. Auton. Robot. 2018, 43, 1257–1270. [Google Scholar] [CrossRef] [Green Version]
  27. Lu, X.; Wang, H.; Tang, S.; Huang, H.; Li, C. DM-SLAM: Monocular SLAM in Dynamic Environments. Appl. Sci. 2020, 10, 4252. [Google Scholar] [CrossRef]
  28. Marchel, Ł.; Specht, C.; Specht, M. Testing the Accuracy of the Modified ICP Algorithm with Multimodal Weighting Factors. Energies 2020, 13, 5939. [Google Scholar] [CrossRef]
  29. Lee, Y.-J.; Song, J.-B.; Choi, J.-H. Performance Improvement of Iterative Closest Point-Based Outdoor SLAM by Rotation Invariant Descriptors of Salient Regions. J. Intell. Robot. Syst. 2013, 71, 349–360. [Google Scholar] [CrossRef]
  30. Wang, J.; Zhao, M.; Chen, W. MIM_SLAM: A Multi-Level ICP Matching Method for Mobile Robot in Large-Scale and Sparse Scenes. Appl. Sci. 2018, 8, 2432. [Google Scholar] [CrossRef] [Green Version]
  31. Leal, N.; Zurek, E.; Leal, E. Non-Local SVD Denoising of MRI Based on Sparse Representations. Sensors 2020, 20, 1536. [Google Scholar] [CrossRef] [Green Version]
  32. Ying, S.; Peng, J.; Du, S.; Qiao, H. A Scale Stretch Method Based on ICP for 3D Data Registration. IEEE Trans. Autom. Sci. Eng. 2009, 6, 559–565. [Google Scholar] [CrossRef]
  33. Feng, Y.Y.; Wang, Q.; Yang, G.C. Incremental 3-D pose graph optimization for SLAM algorithm without marginalization. Int. J. Adv. Robot. Syst. 2020, 17, 14. [Google Scholar]
  34. Jackson, J.; Brink, K.; Forsgren, B.; Wheeler, D.; McLain, T. Direct Relative Edge Optimization, A Robust Alternative for Pose Graph Optimization. IEEE Robot. Autom. Lett. 2019, 4, 1932–1939. [Google Scholar] [CrossRef]
  35. Grisetti, G.; Stachniss, C.; Burgard, W. Improved Techniques for Grid Mapping with Rao-Blackwellized Particle Filters. IEEE Trans. Robot. 2007, 23, 34–46. [Google Scholar] [CrossRef] [Green Version]
  36. Hess, W.; Kohler, D.; Rapp, H.; Andor, D. Real-time loop closure in 2D LIDAR SLAM. In Proceedings of the 2016 IEEE International Conference on Robotics and Automation (ICRA), Stockholm, Sweden, 16–21 May 2016; pp. 1271–1278. [Google Scholar]
  37. Vincent, R.; Limketkai, B.; Eriksen, M. Comparison of indoor robot localization techniques in the absence of GPS. In Detection and Sensing of Mines, Explosive Objects, and Obscured Targets XV; International Society for Optics and Photonics: Bellingham, WA, USA, 2010; p. 7664. [Google Scholar]
  38. Kohlbrecher, S.; Von Stryk, O.; Meyer, J.; Klingauf, U. A flexible and scalable SLAM system with full 3D motion estimation. In Proceedings of the 2011 IEEE International Symposium on Safety, Security, and Rescue, Kyoto, Japan, 1–5 November 2011; pp. 155–160. [Google Scholar]
  39. Mur-Artal, R.; Montiel, J.M.M.; Tardos, J.D. ORB-SLAM: A Versatile and Accurate Monocular SLAM System. IEEE Trans. Robot. 2015, 31, 1147–1163. [Google Scholar] [CrossRef] [Green Version]
  40. Mur-Artal, R.; Tardos, J.D. ORB-SLAM2: An Open-Source SLAM System for Monocular, Stereo, and RGB-D Cameras. IEEE Trans. Robot. 2017, 33, 1255–1262. [Google Scholar] [CrossRef] [Green Version]
  41. Qi, X.; Wang, W.; Yuan, M.; Wang, Y.; Li, M.; Xue, L.; Sun, Y. Building semantic grid maps for domestic robot navigation. Int. J. Adv. Robot. Syst. 2020, 17, 12. [Google Scholar] [CrossRef] [Green Version]
  42. O’Meadhra, C.; Tabib, W.; Michael, N. Variable Resolution Occupancy Mapping Using Gaussian Mixture Models. IEEE Robot. Autom. Lett. 2018, 4, 2015–2022. [Google Scholar] [CrossRef]
  43. Sun, J.; Chen, Z.H.; Wang, P.; Zhang, Q.B.; Bao, P. Multi-region coverage method based on cost map and minimal tree for mobile robot. Robot 2015, 37, 435–442. [Google Scholar]
  44. Schön, B.; Mosa, A.S.M.; Laefer, D.F.; Bertolotto, M. Octree-based indexing for 3D pointclouds within an Oracle Spatial DBMS. Comput. Geosci. 2013, 51, 430–438. [Google Scholar] [CrossRef] [Green Version]
  45. Jessup, J.; Givigi, S.N.; Beaulieu, A. Robust and Efficient Multirobot 3-D Mapping Merging with Octree-Based Occupancy Grids. IEEE Syst. J. 2015, 11, 1723–1732. [Google Scholar] [CrossRef]
Figure 1. Research hotspots of intelligent coal mines from 2010 to 2020.
Figure 1. Research hotspots of intelligent coal mines from 2010 to 2020.
Applsci 11 04968 g001
Figure 2. Pose parameters of roadheader body (camera).
Figure 2. Pose parameters of roadheader body (camera).
Applsci 11 04968 g002
Figure 3. SLAM motion model of the roadheader.
Figure 3. SLAM motion model of the roadheader.
Applsci 11 04968 g003
Figure 4. Vision-based pose estimation scheme of roadheader.
Figure 4. Vision-based pose estimation scheme of roadheader.
Applsci 11 04968 g004
Figure 5. Effect of ORB feature extraction and matching.
Figure 5. Effect of ORB feature extraction and matching.
Applsci 11 04968 g005
Figure 6. Optimized model of the pose graph.
Figure 6. Optimized model of the pose graph.
Applsci 11 04968 g006
Figure 7. Long corridor used for test.
Figure 7. Long corridor used for test.
Applsci 11 04968 g007
Figure 8. Construction principle of occupied grid map.
Figure 8. Construction principle of occupied grid map.
Applsci 11 04968 g008
Figure 9. Value principle of cost map.
Figure 9. Value principle of cost map.
Applsci 11 04968 g009
Figure 10. Grid map of simulated roadway based on laser SLAM.
Figure 10. Grid map of simulated roadway based on laser SLAM.
Applsci 11 04968 g010
Figure 11. Cost map of simulated roadway based on the Karto algorithm.
Figure 11. Cost map of simulated roadway based on the Karto algorithm.
Applsci 11 04968 g011
Figure 12. Memory management mechanism of RTAB-MAP.
Figure 12. Memory management mechanism of RTAB-MAP.
Applsci 11 04968 g012
Figure 13. Principle of octree compression.
Figure 13. Principle of octree compression.
Applsci 11 04968 g013
Figure 14. Dense 3D point cloud map of simulated roadway.
Figure 14. Dense 3D point cloud map of simulated roadway.
Applsci 11 04968 g014
Figure 15. Octree map of simulated roadway.
Figure 15. Octree map of simulated roadway.
Applsci 11 04968 g015
Figure 16. Navigation and obstacle avoidance based on the octree map and cost map.
Figure 16. Navigation and obstacle avoidance based on the octree map and cost map.
Applsci 11 04968 g016
Figure 17. The RTAB-MAP mapping process based on vision and laser.
Figure 17. The RTAB-MAP mapping process based on vision and laser.
Applsci 11 04968 g017
Figure 18. Composite map based on visual and laser data.
Figure 18. Composite map based on visual and laser data.
Applsci 11 04968 g018
Table 1. Common mapping schemes of laser/visual SLAM [19].
Table 1. Common mapping schemes of laser/visual SLAM [19].
Mapping AlgorithmInputOutput
CameraLaser RadarOdometerGrid MapFeature/Point Cloud Map
MonocularBinocularRGB-D2D3D 2D3DSparseDense
Gmapping [35]
Cartographe [36]
Karto [37]
Hector [38]
ORB-SLAM1 [39]
ORB-SLAM2 [40]
RTAB-MAP [19]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhang, W.; Zhai, G.; Yue, Z.; Pan, T.; Cheng, R. Research on Visual Positioning of a Roadheader and Construction of an Environment Map. Appl. Sci. 2021, 11, 4968. https://doi.org/10.3390/app11114968

AMA Style

Zhang W, Zhai G, Yue Z, Pan T, Cheng R. Research on Visual Positioning of a Roadheader and Construction of an Environment Map. Applied Sciences. 2021; 11(11):4968. https://doi.org/10.3390/app11114968

Chicago/Turabian Style

Zhang, Wentao, Guodong Zhai, Zhongwen Yue, Tao Pan, and Ran Cheng. 2021. "Research on Visual Positioning of a Roadheader and Construction of an Environment Map" Applied Sciences 11, no. 11: 4968. https://doi.org/10.3390/app11114968

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop