Next Article in Journal
Vertebral Center Points Locating and Cobb Angle Measurement Based on Deep Learning
Next Article in Special Issue
A Spatial Location Representation Method Incorporating Boundary Information
Previous Article in Journal
Fractional Sliding Mode Harmonic Control of an Active Power Filter
Previous Article in Special Issue
Real-Time Stereo Visual Odometry Based on an Improved KLT Method
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Point–Line-Aware Heterogeneous Graph Attention Network for Visual SLAM System

1
Beijing Key Laboratory of Petroleum Data Mining, Beijing 102249, China
2
Department of Computer Science and Technology, China University of Petroleum, Beijing 102249, China
3
Pipeline Technology and Safety Research Center, China University of Petroleum, Beijing 102249, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(6), 3816; https://doi.org/10.3390/app13063816
Submission received: 8 February 2023 / Revised: 13 March 2023 / Accepted: 15 March 2023 / Published: 16 March 2023
(This article belongs to the Special Issue Trajectory Analysis, Positioning and Control of Mobile Robots)

Abstract

:
Simultaneous localization and mapping (SLAM), as an important research topic in robotics, is useful but challenging to estimate robot pose and reconstruct a 3-D map of the surrounding environment. Despite recent success of several deep neural networks for visual SLAM, those methods cannot achieve robust results in complex industrial scenarios for constructing accurate and real-time maps due to the weak texture and complex geometric structure. This paper presents a novel and efficient visual SLAM system based on point–line-aware heterogeneous graph attention network, which combines points and line segments to solve the problem of the insufficient number of reliable features in traditional approaches. Firstly, a simultaneous feature extraction network is constructed based on the geometric relationships between points and points and points and lines. To further improve the efficiency and accuracy of the geometric association features of key regions, we design the point–line-aware attention module to guide the network to pay attention to the trivial features of both points and lines in images. Moreover, the network model is optimized by a transfer-aware knowledge distillation strategy to further improve the system’s real-time performance. Secondly, to improve the accuracy of the point–line matching, we design a point–line heterogeneous graph attention network, which combines an edge aggregation graph attention module and a cross-heterogeneous graph iteration module to conduct learning on the intragraph and intergraph. Finally, the point–line matching process is transformed into an optimal transport problem, and a near-iterative method based on a greedy strategy is presented to solve the optimization problem. The experiments on the KITTI dataset and a self-made dataset demonstrate the better effectiveness, accuracy, and adaptability of our method than those of the state of the art in visual SLAM.

1. Introduction

Simultaneous localization and map construction technology, as the key to autonomous movement of robots, is widely used in unmanned driving, virtual reality, mobile robots, and other fields [1,2,3,4,5]. Compared with laser SLAM, vision-based SLAM has a low power consumption, low cost, miniaturization, and other advantages, and its theoretical and application value is very prominent [6]. Visual SLAM constructs a map of the surrounding environment by obtaining the plane image information of the real world through a camera. The pose state of the camera is inferred by the extracted feature information or pixel grayscale. Visual SLAM methods can be classified into four categories: feature-point-based methods [6,7,8], feature-line-based methods [9,10,11,12], feature-plane-based methods [12,13], and the combination of the above methods [14,15,16,17,18,19,20], according to the kind of features used to estimate the trajectory. The existing empirical methods cannot deal effectively with complex industrial scenarios due to occlusion, illumination, and deformation issues. In recent years, research on visual SLAM based on deep learning has attracted widespread attention. Serra et al. [21] used deep convolution to extract a point description with the L2 norm as a similarity measure to enhance the robust matching between key points and their local image features. Since a single feature descriptor cannot generate an effective key-point detection, Shend et al. [22] proposed an end-to-end trainable matching network, RF-Net, based on a receptive field to achieve a more efficient key-point detection. In order to solve the problem of the low localization accuracy caused by the lack of key-point shape perception in the joint learning of feature detectors and descriptors, Luo et al. [23] applied a deformable convolutional network with a dense spatial transformation to enhance the dynamic receptive field and improve the ability to express local shapes. Sarlin et al. [24] proposed a graph neural network, SuperGlue, based on attention aggregation, which used the optimal transmission model for matching optimization and realized the pose estimation in both indoor and outdoor environments. Combining a convolutional network with a recurrent network, Tang et al. [25,26] proposed a geometric correspondence network (GCN), which used an end-to-end learning method to detect key points and generate descriptors for improving the accuracy of the pose estimation. Aimed at the problem of decreased localization accuracy caused by the partial occlusion of line segments, Pautrat et al. [27] used a self-supervised network for line detection and for extracting line-segment descriptors, which improved the robustness of line-segment matching. The above-mentioned feature detection algorithms based on deep learning fused multilevel features, which did not deeply explore the association and constraint relationship between point and line features.
In industrial production scenarios, there are many complex background objects such as various buildings, pipelines, production equipment, and safety signs that lack corner points or contain repeated textures. The point features in the image are not specific enough, which makes them unable to provide an accurate position estimation. In addition, the mismatch of line features greatly increases the time complexity of the computation. In general, although existing methods have achieved certain results in feature detection and matching tasks, due to the uneven light, single texture, and complex scene structure in industrial scenarios, the pose estimation is easily degraded. How to efficiently fuse point and line information to build a more stable visual SLAM system is still a difficult problem that needs further research. Our main contributions can be summarized as follows:
  • To solve the problem of weak point–line extraction ability in complex scenes, a point–line synchronous geometric feature extraction network, PL-Net, is proposed. We use an optimized residual block-feature pyramid network (ORB-FPN) to extract the feature map of the input image. In the point extraction branch, based on the point-aware module, the multiscale context is aggregated to obtain features with rich receptive fields. Moreover, the edge information is used for the line extraction branch to improve the accuracy of the line-segment detection. In order to make the network lightweight, a transfer-aware knowledge distillation method is proposed to compress the model for generating the point–line feature in the extraction task.
  • Targeting a high accuracy and efficiency, a heterogeneous attention graph neural network (HAGNN) is presented, which uses an edge-aggregated graph attention network (EAGAT) to iterate the vertices of the heterogeneous graph constructed from points and lines. To enhance the performance of the point–line matching, a cross-heterogeneous graph interaction (CHGI) is used for harmonizing heterogeneous information between graphs.
  • By transforming the point–line matching process into an optimal transport problem, a greedy inexact proximal point method for optimal transport, GIPOT, is proposed, which calculates the optimal feature assignment matrix to find the global optimal solution for the point–line matching problem.

2. SLAM System Framework

The framework of the SLAM system proposed in this paper is shown in Figure 1. Firstly, a new image is input into the PL-Net network to detect key points and line segments, and the corresponding descriptors are obtained through the point–line-aware attention module to enhance the feature expressiveness for both points and lines in images. Then, the point and line features of the two images are transferred to the point–line heterogeneous graphs, which are constructed by using the point and line features as the vertices and connecting a vertex to its neighbors within a fixed radius. Secondly, the attention network HAGNN obtains the enhanced features and inputs them into the GIPOT to generate point–line matching results and calculate the pose of the current frame. Finally, by reprojecting the features in the local map to the current frame, the projection error is calculated for the backend processing of SLAM to complete the map.

3. Methodology

3.1. Point–Line Feature Extraction Network

The point–line feature extraction network PL-Net is shown in Figure 2. Firstly, the ORB-FPN module was used to extract the features of each layer for the image, and the PG module completed the key-point extraction. Then, the center point map and displacement map were generated through the branch of the line-segment perception module. Finally, the point–line descriptor was generated through convolution and upsampling operations.

3.1.1. ORB-FPN Module

As shown in Figure 2, an optimized residual block (ORB) is designed based on the Nesterov acceleration gradient (NAG) algorithm to enhance the expressive ability of target features [26]. Then, we have:
y k + 1 = x k + β x k x k 1
x k + 1 = y k + 1 α f y k + 1
where x k and y k + 1 denote the output and input of the first layer of the network, and α and β represent the learning rate and momentum parameters, respectively. f y k + 1 is the gradient of the objective function f at y k + 1 , and f is a smooth function satisfying the Lipschitz property. When the momentum parameter β = 0 , the NAG algorithm is equivalent to the standard gradient descent algorithm. When β > 0 , it optimizes the combination of α and β to accelerate convergence.
In the neural network propagation process, the transmission of the signal from the first layer to the last layer is expressed as:
L i + 1 = σ U i L i
where L i + 1 is the features of the i + 1 th layer in the network, and σ represents the activation function. Suppose U is a symmetrical positive definite matrix; let V = U and μ = V L ; then, for the nonlinear activation function σ μ , there is a function g μ , when g μ = σ μ . We have:
i g V j T μ = U σ U T L = U σ U L
The objective function f μ is defined as:
f μ = μ 2 2 i g V j T μ
where V i is the ith column of V. Then,
f μ i = β i V σ V μ i
Equation (2) can be expressed by:
μ i + 1 = μ i + β μ i μ i 1 α 1 + β f μ i β f μ i 1
Recovering L by L = V 1 μ leads to:
L i + 1 = 1 + β 1 α α β L i + β 1 α L i 1 + α 1 + β σ U L i
where σ U L i is the ith layer feed-forward network, and the ORB module structure is shown in Figure 2.
In order to aggregate the FPN multiscale feature information, a dual attention module (DAM) was designed to perform the feature aggregation. As shown in Figure 3, firstly, in order to obtain the position and channel information of the feature, the shallow feature map x R W × H × D 1 was passed through a global pooling operation and compression to generate the position vector x p R W × H × 1 and the channel vector x c R 1 × 1 × D 1 , respectively. Then, the position vector x p computed the weight of each position with a sigmoid activation function and multiplied it with the feature map x p to generate the spatial position feature map F p R W × H × D 1 , which was defined as:
F p = σ ( x ) x
where x is the shallow feature map, and σ is the sigmoid activation function. Similarly, the convolution of the ReLU activation function, defined as f ( x ) = m a x ( 0 , x ) , and the sigmoid activation function was performed on the channel vector x c . The weight of each channel was calculated and multiplied by the feature map x to generate the feature map F c R W × H × D 1 , which was defined as follows:
F c = σ ( δ ( W 2 ( W 1 ( g p ( x ) ) ) ) x
where δ is the ReLU activation function, and W 1 and W 2 are convolution operations with sizes 1 × 1 × D 1 / 16 and 1 × 1 × D 1 , respectively. Finally, the final output feature map F was obtained by fusing the feature maps F p , F c , and x . Then,
F = [ ( x F p F c ) , x ]
where ⊕ represents the addition of the corresponding elements of two matrices.
The input image was passed through the ORB-FPN module, which denoted the output of the backbone as { C 2 , C 3 , C 4 , C 5 } with strides of { 4 , 8 , 16 , 32 } . { F 2 , F 3 , F 4 } were obtained after a 1 × 1 convolution with the same 128-dimensional channel features. Finally, the ORB module was added to enhance the acceptance domain of the output feature by using the backbone network on C 5 to separate the important context information. After interpolation and maximum pooling of the extracted context features and the generated three feature maps, an elementwise summation was performed to obtain features F.

3.1.2. Key-point Detection Module

As shown in Figure 2, the key-point detection module consisted of three point perception modules and two 3 × 3 convolutions with a stride length of 1. A batch normalization layer and ReLU layers were added between each convolutional layer. The output vector was processed through a sigmoid activation function, so that the pixel values of the saliency map were between 0 and 1. Then, through the key-point perception module and convolution processing, the convolution operation was used to discriminate whether the 8 × 8 area prediction contained key points. Finally, the key points were detected by using the nonmaximum suppression (NMS) method in the key-point generation module (PG).
The point-aware module was used to capture the relationship between key points. As shown in Figure 4, the key-point extraction branch embedded a context enhancement module, which improved the feature expression ability. The output feature y p R W × H × D was obtained by fusing the convolutional features of different scales, which took y p W × H × D as input. The above process was defined as:
y p = W 1 [ x p , B N ( W 1 x p ) , B N ( W 2 x p ) , B N ( W 3 W 1 x p ) ]
where x P is the input feature, W 1 , W 2 , and W 3 are convolution operations with sizes of 1 × 1 , 2 × 2 , and 3 × 3 , respectively. B N is a normalization, and [ ] is a splicing operation.

3.1.3. Line-Segment Detection Module

The line-segment detection module extracted the features of the plane image through the ORB-FPN module and then input it to the line-segment extraction module to generate a midpoint with two symmetrical endpoints as the line-segment detection result. The extraction of the line segment’s center point [28] used the classification model to judge whether the pixel was the center point of the line segment. Since the shape of the line segment was narrow and long, a large receptive field was required to classify the center of the line segment. Therefore, a hybrid convolution module was introduced by stacking three convolutional layers, a 3 × 3 deformable convolutional layer and two 3 × 3 dilated convolutional layers. The receptive field of the network was increased while reducing the parameters of the network. Then, three line-segment perception modules were used to enhance the feature representation ability. Finally, a deconvolution layer was used to restore the size of the output map to 512 × 512, which represented the center point of the line segment on the output feature map. The displacement regression task of the line-segment extraction branch was designed to predict the angle and length of the end point relative to the midpoint. It was composed of a 3 × 3 deformable convolution layer and two 3 × 3 convolutional layers with a stride of 1. The relevant displacement was indexed by the position through the output map. Finally, the CAL [29] was used for the line-segment generation, and the two endpoints of the line segment were defined as follows:
( x l s , y l s ) = ( x l c , y l c ) + α 2 ( cos θ , sin θ )
( x l e , y l e ) = ( x l c , y l c ) α 2 ( cos θ , sin θ )
where ( x l c , y l c ) is the coordinates of the root node, α is the length of the line segment, and θ is the rotation angle.
In the line-segment detection, a line-aware module was introduced to effectively extract line shape features. As shown in Figure 5, the line-segment-aware module adopted the improved self-attention mechanism, which took the feature x L R W × H × D as input and fused the self-attention features to generate the final output feature y L R W × H × D ; the above process was defined as follows:
y L = x L W 1 ( α ( W q x L × W k x L ) × W v x L )
where x L is the input feature, W 1 , W q , W k , and W v are the learned weight matrices, which were implemented as 1 × 1 convolutions, and ⊕ means that the corresponding elements of the two matrices are added.

3.1.4. Parallel Attention Module

To efficiently merge the two branches of point descriptors and line descriptors, we designed a parallel attention module (PAB) based on self-attention and channel attention. As shown in Figure 6, the output features of the key-point description branch contained line-edge information with a strong correlation. In order to improve the accuracy of the line descriptor, a lightweight attention mechanism was used to assign more weights for the features of useful regions. The output feature map of the point description branch was expressed as X E R C × H × W . A one-dimensional convolution of X E was performed to obtain the spatial attention map A E R C × H × W . The edge feature map X E S R C × H × W was calculated as: X E S = a ( X E A E ) + X E , where a is a learnable parameter that was initialized to 0. The CAEU module was designed to calculate a channel attention map, which recalibrated the weight of the channel and obtained the fused feature map X F S R C × H × W from X F S = X E S δ ( C o n v 1 × 1 ( C o n v 1 × 1 ( G A P ( X E S ) ) ) . Finally, the final output X F S C R 2 C × H × W of PAB was obtained by concatenating X T S and X F S together.

3.1.5. Network Output Distillation

In order to reduce the increasing computation cost caused by the introduction of the attention mechanism and the point–line perception module, we further compressed the point–line detection model PL-Net. A transfer-aware method is presented to transfer the information from the teacher model to the student network. The knowledge distillation strategy (KD) [30,31,32] was used to fine-tune the accuracy of the recovery model. As shown in Figure 2, the multiples tasks were combined with the teacher network and the student network to guide the training of the student feature extraction. In the training process, the adaptive weighted multitask distillation was realized, and X t e a , Y t e a , and Z t e a represented the key point of the teacher model, the center point of the line segment, and the output of the line-segment regression feature layer, respectively. The student model corresponded to the feature layers X s t u , Y s t u , and Z s t u . The mean squared error (MSE) function was applied to the multitask distillation. The loss function of the training distillation was:
L p S = X t e a X s t u 2 L r o o t S = Y t e a Y s t u 2 L d i s S = Y t e a Y s t u 2
where L p S , L r o o t S , and L d i s S denote the key point, the line segment’s center point, and the distillation loss of the line-segment regression task, respectively. Then, the weighted distillation loss was defined as:
L M S E s = l ω l L l s
where L M S E s is the multitask distillation loss, and ω l represents the weight value of the verification loss.

3.1.6. Interlayer Knowledge Distillation

Different from the network output distillation, the student network was further enhanced by performing an interlayer knowledge distillation between the output of the teacher model and the student model. The aware modules of each student layer were associated with the relevant target-layer-aware modules for knowledge transfer. The layer’s knowledge distillation loss was defined as:
L F M D = ( s l , t l ) C D i s t ( T r a n s t ( F t l t ) , T r a n s s ( F s l s ) )
Then, the overall loss was obtained as:
L total = L M S E s + β L F M D
where T r a n s ( · ) means to convert the feature map of the perception module into a specific manual representation through the attention map. C is the perception module, and F t l t and F s l s are the feature layers of the lth layer of the student model and the teacher model, respectively. The distance function D i s t ( · ) was used for computing the distillation loss of the feature maps, and β was a hyperparameter.

3.2. Heterogeneous Graph Attention Network

As shown in Figure 7, consider two images I and I , and the number of two feature sets m and n belonging to them, respectively. Let d R D be the feature descriptor, where D is the dimension of the descriptor. We utilized an attention graph neural network to integrate the contextual cues and enhance its feature expression ability.
Position encoders were used for the two input features, and the key points and lines positions were embedded into a high-dimensional vector by adding position encoding to F ^ 1 and F ^ 2 ; thus, we had:
f i = d i + MLP ( P i )
MLP ( P i ) = W a σ ( W b P i )
where P i is the position of the feature, d i is the descriptor information, and σ ( · ) is the ReLU activation function.

3.2.1. Edge-Clustering Graph Attention Module

We propose an edge-aggregated graph attention network (EAGAT) based on GAT [33], which uses edge information for feature enhancement during the aggregation process. In order to make full use of the information of edge features, these different types of links used different attention mechanisms. For features of the same nature (points and points, lines and lines), the self-attention mechanism was used for the aggregation. For features of different nature (points and lines), the cascade method was used for the aggregation. Let the feature of the vertex v i in the graph be f i , defined as:
f i = σ ( j N s s o f t m a x ( W a f i W β f j d k W γ a j i ) W ε f j + j N d s o f t m a x ( [ W a f i | | W β f j | | W γ a j i ] ) W ε ) | | f i
where W a , W β , W γ , and W ε R f η × f η represent the weight parameters, N s is the feature set of the same nature, N d represents the feature set of different properties, and σ is the ReLU activation function.

3.2.2. Cross-Heterogeneous Graph Iteration Module

Due to the affinity learning problem of message passing between graphs in graph matching, a point–line heterogeneous graph message-passing method is proposed to enhance the node features through an interactive correlation. The edge features and node features were aggregated in two ways. For nodes of the same nature (points and points, lines and lines), we used linear attention for the aggregation, and for nodes of different properties (points and lines), we used the aggregation method of the self-attention mechanism. Then, f s i R f η is expressed as:
f s i = LN { j N d i ( s o f t m a x [ ( W v f j ) ( W k f j ) T ] + W q f s i ) } + LN { j N s i ( W l ( ( W q f s i ) ( W k f j ) T + W k f j ) + f s i }
where v l / r R f η is the feature node of the two graphs, W ( · ) is the weight parameter, and LN represents a layer normalization (LN).

3.3. Greedy near Iterative Matching Module

The output H i of the last layer in the graph neural network is the feature of graph I . H j is the feature of graph I , and the point–line distance matrix G R + N 1 × N 2 can be expressed as:
G = f a f f ( H i , H j ) , i v 1 , j v 2
where f a f f is the weighted bilinear function, defined as:
f a f f ( H i , H j ) = exp ( H i T K H j T τ )
where the feature is an n-dimensional vector, namely i v 1 , j v 2 and H i T , H j T R n × n , K R n × n is the weight matrix of the affinity function, and τ is the regularization parameter. In the matching process, due to the inconsistency of point and line types, a direct fusion may cause a mismatch of point and line types. For this reason, we regarded the unit block diagonal matrix as the initial coupling matrix Γ ( 1 ) so that the relationship between point features and line features in the iterative process minimized the matching cost. Then, we had:
Γ ( 1 ) 11 p 0 0 11 L
where P is the number of points after completion, and L is the number of lines after completion. Then, the point–line discrete distribution Sinkhorn distance [34,35] was defined as:
W ( u , v ) = min Γ ( u , v ) C , Γ + λ h ( Γ )
where u and v are probability vectors, W ( u , v ) is the distance between u and v, the matrix C = c i j R + n × n is the cost matrix, and c i j is the distance between u i and v j . The regularization term h ( Γ ) = i , j Γ i j ln Γ i j . The proximal point iteration method [35] was used to solve Equation (25). According to the proximal point iteration method, it is defined as the Bregman divergence:
D h x , y = i = 1 n x i ln x i y i i = 1 n x i i = 1 n y i
After introducing the near-end point iteration, Equation (27) can be rewritten as:
Γ ( t + 1 ) = argmin Γ ( u , v ) C , Γ + β t D h ( Γ , Γ ( t ) )
Substituting the Bregman divergence Equation (28) into Equation (29), it becomes
Γ ( t + 1 ) = argmin Γ ( u , v ) C , Γ + β t h ( Γ )
where C = C β t ln Γ ( t ) ; we used the greedy strategy to update the best row or column and defined the distance matrix
ρ ( x , y ) = y x + log x y
According to Equations (29) and (30), the affinity matrix was updated to find the best matching relationship, and the specific algorithm flow is shown in Algorithm  1.
Algorithm 1: GIPOT( μ , v, G).
Applsci 13 03816 i001

3.4. Loss Function

In order to realize the matching of points and lines, the point–line extraction loss, descriptor extraction loss, and point–line matching loss were used as the loss functions during model training.

3.4.1. Point–Line Extraction Loss

In the training stage of the point and line extraction branch, the output included the root node’s confidence map, key-point map, and displacement map. The losses of these three tasks were combined into Equation (32), defined as follows:
L P L E = L r o o t + L p + L d i s
The ground truth of the root-point confidence map was constructed by marking the root-point positions on a zero map. A weighted binary cross-entropy loss L r o o t was used to supervise this task, which was defined as:
L r o o t = i R ˜ i log R i + ( 1 R ˜ i ) log ( 1 R i )
where R ˜ i and R i are the prediction and true label of the root node of the line segment, respectively. The true value of the key point was marked with the ORB feature. L p ( X , Y ) was defined as:
L p ( X , Y ) = 1 H c w c l p T i j , T i j ˜
l p T i j , T i j ˜ = log ( exp ( T k i j ) 1 64 exp ( T k i j ) )
The displacement part of the line segment relative to the root node was used to locate the length and angle of the line segment, which used the L1 loss and L1 smoothing loss, respectively defined as:
L d i s = i = 1 m θ i θ ^ i + 0.5 ( ρ i ρ ^ i ) 2 i f ρ i ρ ^ i < 1 θ i θ ^ i + ρ i ρ ^ i 0.5 o t h e r w i s e
where θ i and ρ i represent the actual line segment’s length and angle, and θ ^ i and ρ ^ i are the predicted line segment’s length and angle, respectively.

3.4.2. Point–Line Descriptor Loss

Denote the original image as I, and apply the homography transformation for I to form a new image I . Since the homography transformation is known, the corresponding relationship between key points and line segments on I and I can be obtained. Therefore, the loss function can be defined as:
L d ( θ , d a θ , d + θ , d θ ) = [ m + d a θ d + θ 2 min d θ d θ d a θ d θ 2 ] +
where the parameter m was set to 0.5, d a θ is the descriptor on I of the anchor point, d + θ is the matching descriptor on I of the positive sample, and d θ is the set of nonmatching descriptors on I of the negative sample.

3.4.3. Matching Loss

For a matching network using the L2 loss, the loss function can be expressed as:
L m = 1 M g t ( i , j ) M g t 1 σ 2 ( i ) M ( i , j ) M g t ( i , j ) 2
where σ 2 ( i ) is the confidence variance of feature i. M ( i , j ) is the matching probability of feature i and feature j, and M g t is the real-valued matrix obtained by the homography transformation.

3.4.4. Normalization

The total loss function was the sum of the above loss functions:
L s u m = λ 1 L P L E + λ 2 L d + λ 3 L m
where λ 1 , λ 2 and λ 3 represent the coefficients of each loss function, respectively. λ 1 , 2 , 3 = { 0.25 , 0.25 , 0.5 } .

4. Experiments and Evaluation

4.1. Model Training Details

A wide range of experiments were performed on different datasets to demonstrate the efficacy of our method. Our approach was evaluated with several evaluation criteria by comparing to the typical SLAM methods on the KITTI dataset [36]. We used the training set of Wireframe [37] with the ground truth to train our models and the other compared methods. The training process used data enhancement techniques such as random Gaussian noise, motion blur, and brightness level changes to improve the network’s robustness ability for changes in lighting and viewing angles. For end-to-end training of the point–line matching network, our network was implemented in Pytorch [38] using the Adam [39] optimizer to train the network with an initial learning rate of 1 × 10 5 and a decay of the learning rate by 20 at each epoch. We trained our model on a GeForce GTX2080Ti GPU.

4.2. KITTI Dataset Evaluation

We tested the proposed algorithm on the KITTI dataset [36]. The quantitative evaluations for the different SLAM systems were the absolute trajectory error (ATE) [40] and the relative pose error (RPE) [41] based on translations and rotations. Table 1 shows the performance of this system was better than ORB-SLAM2, especially in sequences with strong lighting, motion blur, and low texture areas, such as 06 and 09. It can be seen that the multifeature fusion not only improved the accuracy of the algorithm but also avoided the degradation problem that may occur in the pose solution algorithm when using a single feature.
Figure 8 shows the comparison results of ORB-SLAM2 and the method in this paper on KITTI’s partial sequences. It can be seen that the algorithm of this paper was equivalent to ORB-SLAM2 as a whole. However, in sequences such as 00, 02, and 09, it showed that our method had the best results. On the other hand, the ORB-SLAM2 lost track easily and did not have a whole trajectory.
Figure 9 depicts the variation curve of the pose estimation error with the number of training iterations. When we used more than 80 iterations, the rotation and translation errors were relatively small while the performance improved, which showed that the algorithm in this paper had a good convergence.
Figure 10 visualizes the statistical error property between our system and ORB-SLAM2 on sequence 09 of KITTI. It shows that our model achieved obviously a better performance than ORB-SLAM2 in terms of the root-mean-square error (controlled within 2 m), extreme value error, and the sum of the squared errors. Figure 10b shows the columnar statistical comparison of postural errors; our model was better than ORB-SLAM2 in the APE distribution range of the algorithm.

4.3. Real Data Evaluation

As shown in Figure 11, the feasibility of the algorithm was verified on a physical and virtual oil and gas station, where the virtual simulation platform with a quadruped robot was built with Unity3D.
To show the effects of point and line features on the SLAM system, we intercepted two frames of images for extracting the features by PL-Net and matching by HAGNN. As shown in Figure 12, it can be seen that the combination of point and line could make the SLAM algorithm obtain richer and more diverse feature information.
The virtual simulation platform used a quadruped robot to inspect the oil field equipment. As shown in Figure 13, this trajectory was compared with the ground truth by using the evaluation package to obtain the RPE and APE. It can be seen that the area with a larger error was basically distributed at the corner of the trajectory. The RMSE was 17.5 m. Overall, the trajectory of our method was consistent with the ground truth with a high accuracy.

4.4. GIPOT Experiment

In order to illustrate the convergence of GIPOT with different β , the Wasserstein distance of two one-dimensional Gaussian distributions was measured as an evaluation index. As shown in Figure 14a, the blue equation was 0.5 N ( 70 , 8 ) + 0.5 ( 35 , 10 ) , the red equation was 0.4 N ( 80 , 9 ) + 0.6 N ( 40 , 10 ) , where N ( μ , σ 2 ) is the probability density function of the one-dimensional Gaussian distribution, μ and σ 2 are the mean and variance, respectively. Figure 14b shows the convergence of GIPOT under different conditions. Compared with the Sinkhorn method, the convergence of the GIPOT iteration was quicker when β was large. GIPOT could converge to the exact Wasserstein distance with a complexity comparable to that of Sinkhorn.

4.5. Ablation Study

We used two datasets to demonstrate our proposed knowledge distillation. The result of the teacher network and the student network are shown in Figure 15. Compared with the true value, both models could identify key points and line segments with high precision. Although there were some small missing line segments and connection errors in the results of the student network, the expression of the line-segment structure in the environment was basically accurate. The quantitative comparison is shown in Table 2. Although the performance of the student network was slightly lower than that of the teacher network, the operation speed was increased by 73%.
To verify the role of the multifeature fusion in the SLAM system, the root-mean-square error (RMSE) of the ATE index under different feature combinations was calculated. The experimental results are given in Table 3, where the point–line feature combination method used in this paper significantly improved the accuracy of the pose estimation. To evaluate our design decisions, we evaluated four different variants with results. This ablation study, presented in Table 4, showed that all HAGNN blocks were useful and brought substantial performance gains. “No EAGAT” replaced all EAGAT layers with CHGI layers, and the matching accuracy decreased by 9.7%. “No CHGI” replaced all CHGI layers with EAGAT layer, and the precision of the resulting matching decreased by 22.6%, “No HAGNN” replaced the graph neural network with a single linear projection, and the precision of the resulting matching decreased by 26.1%.
The proposed HAGNN was compared with two feature matching methods: the nearest neighbor (NN) method and SuperGlue. As show in Table 5, it can be seen clearly that HAGNN had a significantly higher pose estimation accuracy than all competitors, which showed a higher feature expression ability.

5. Conclusions

In this paper, we proposed a point–line-aware heterogeneous graph attention network for a visual SLAM system. Combining the point- and line-aware attention modules based on an attention-driven mechanism, the geometric association features of key regions was further extracted, and the model was simplified by a transfer-aware knowledge distillation strategy. By improving the accuracy of image point–line matching, a point–line heterogeneous graph attention network was proposed, which realized the feature aggregation by conducting learning on the intragraph and intergraph. Based on the optimal transport theory, we proposed a greedy inexact proximal point method that could effectively solve the point–line matching problem. Experiments on a public dataset and a self-made dataset showed qualitatively and quantitatively that our model had stronger robustness and a better generalization ability. One limitation of our feature matching was that it was not easy to estimate the pose error due to the interference of dynamic objects. Thus, in a future study, we will introduce the cross-frame semantic information in the network for dynamic environments.

Author Contributions

Conceptualization, Y.L. and H.S.; methodology, Y.L.; software, H.S.; validation, S.D.; formal analysis, H.S.; investigation, Y.L.; resources, H.S.; data curation, Y.L.; writing—original draft preparation, H.S.; writing—review and editing, Y.L.; visualization, H.S.; supervision, Y.L.; project administration, Y.L.; funding acquisition, Y.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by NSFC 61972353, NSF IIS-1816511, OAC-1910469 and Strategic Cooperation Technology Proiects of CNPC and CUPB: ZLZX2020-05.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Yousif, K.; Bab-Hadiashar, A.; Hoseinnezhad, R. An overview to visual odometry and visual SLAM: Applications to mobile robotics. Intell. Ind. Syst. 2015, 1, 289–311. [Google Scholar] [CrossRef]
  2. Demim, F.; Nemra, A.; Boucheloukh, A.; Louadj, K.; Hamerlain, M.; Bazoula, A. Robust SVSF-SLAM algorithm for unmanned vehicle in dynamic environment. In Proceedings of the 2018 International Conference on Signal, Image, Vision and Their Applications (SIVA), Guelma, Algeria, 26–27 November 2018; pp. 1–5. [Google Scholar]
  3. Demim, F.; Nemra, A.; Boucheloukh, A.; Kobzili, E.; Hamerlain, M.; Bazoula, A. SLAM based on adaptive SVSF for cooperative unmanned vehicles in dynamic environment. IFAC-PapersOnLine 2019, 52, 73–80. [Google Scholar] [CrossRef]
  4. Kuo, X.Y.; Liu, C.; Lin, K.C.; Lee, C.Y. Dynamic attention-based visual odometry. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 14–19 June 2020; pp. 36–37. [Google Scholar]
  5. Teixeira, B.; Silva, H.; Matos, A.; Silva, E. Deep learning for underwater visual odometry estimation. IEEE Access 2020, 8, 44687–44701. [Google Scholar] [CrossRef]
  6. Lee, T.; Kim, C.-H.; Cho, D.-i.D. A monocular vision sensor-based efficient SLAM method for indoor service robots. IEEE Trans. Ind. Electron. 2018, 66, 318–328. [Google Scholar] [CrossRef]
  7. Zhang, L.; Koch, R. An efficient and robust line segment matching approach based on LBD descriptor and pairwise geometric consistency. J. Vis. Commun. Image Represent. 2013, 24, 794–805. [Google Scholar] [CrossRef]
  8. Cho, H.; Kim, E.K.; Kim, S. Indoor SLAM application using geometric and ICP matching methods based on line features. Robot. Auton. Syst. 2018, 100, 206–224. [Google Scholar] [CrossRef]
  9. Han, R.; Li, W.F. Line-feature-based SLAM algorithm. Acta Autom. Sin. 2006, 32, 43–46. [Google Scholar]
  10. Zhou, D.; Dai, Y.; Li, H. Ground-plane-based absolute scale estimation for monocular visual odometry. IEEE Trans. Intell. Transp. Syst. 2019, 21, 791–802. [Google Scholar] [CrossRef] [Green Version]
  11. Yang, S.; Scherer, S. Monocular object and plane slam in structured environments. IEEE Robot. Autom. Lett. 2019, 4, 3145–3152. [Google Scholar] [CrossRef] [Green Version]
  12. Cho, H.; Yeon, S.; Choi, H.; Doh, N. Detection and compensation of degeneracy cases for imu-kinect integrated continuous slam with plane features. Sensors 2018, 18, 935. [Google Scholar] [CrossRef] [Green Version]
  13. Gomez-Ojeda, R.; Moreno, F.A.; Zuniga-Noël, D.; Scaramuzza, D.; Gonzalez-Jimenez, J. PL-SLAM: A stereo SLAM system through the combination of points and line segments. IEEE Trans. Robot. 2019, 35, 734–746. [Google Scholar] [CrossRef] [Green Version]
  14. Zhang, X.; Wang, W.; Qi, X.; Liao, Z.; Wei, R. Point-plane slam using supposed planes for indoor environments. Sensors 2019, 19, 3795. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Grant, W.S.; Voorhies, R.C.; Itti, L. Efficient Velodyne SLAM with point and plane features. Auton. Robot. 2019, 43, 1207–1224. [Google Scholar] [CrossRef]
  16. Sun, Q.; Yuan, J.; Zhang, X.; Duan, F. Plane-Edge-SLAM: Seamless fusion of planes and edges for SLAM in indoor environments. IEEE Trans. Autom. Sci. Eng. 2020, 18, 2061–2075. [Google Scholar] [CrossRef]
  17. Chen, X.; Cai, Y.; Tang, Y. A visual SLAM algorithm based on line point invariants. Robot 2020, 42, 485–493. [Google Scholar]
  18. Li, H.; Hu, Z.; Chen, X. PLP-SLAM: A visual SLAM method based on point-line-plane feature fusion. Jiqiren 2017, 39, 214–220. [Google Scholar]
  19. Simo-Serra, E.; Trulls, E.; Ferraz, L.; Kokkinos, I.; Fua, P.; Moreno-Noguer, F. Discriminative learning of deep convolutional feature point descriptors. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 118–126. [Google Scholar]
  20. Fu, K.; Liu, Z.; Wu, X.; Sun, C.; Chen, W. An Effective End-to-End Image Matching Network with Attentional Graph Neural Networks. In Proceedings of the 2022 IEEE 17th Conference on Industrial Electronics and Applications (ICIEA), Chengdu, China, 15–18 April 2022; pp. 1628–1633. [Google Scholar]
  21. Zhang, J.; Marszałek, M.; Lazebnik, S.; Schmid, C. Local features and kernels for classification of texture and object categories: A comprehensive study. Int. J. Comput. Vis. 2007, 73, 213–238. [Google Scholar] [CrossRef] [Green Version]
  22. Sarlin, P.E.; DeTone, D.; Malisiewicz, T.; Rabinovich, A. Superglue: Learning feature matching with graph neural networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 4938–4947. [Google Scholar]
  23. Tang, J.; Ericson, L.; Folkesson, J.; Jensfelt, P. GCNv2: Efficient correspondence prediction for real-time SLAM. IEEE Robot. Autom. Lett. 2019, 4, 3505–3512. [Google Scholar] [CrossRef] [Green Version]
  24. Tang, J.; Folkesson, J.; Jensfelt, P. Geometric correspondence network for camera motion estimation. IEEE Robot. Autom. Lett. 2018, 3, 1010–1017. [Google Scholar] [CrossRef]
  25. Pautrat, R.; Lin, J.T.; Larsson, V.; Oswald, M.R.; Pollefeys, M. SOLD2: Self-supervised occlusion-aware line description and detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 11368–11378. [Google Scholar]
  26. Li, H.; Yang, Y.; Chen, D.; Lin, Z. Optimization algorithm inspired deep neural network structure design. In Proceedings of the Asian Conference on Machine Learning, Beijing, China, 14–16 November 2018; pp. 614–629. [Google Scholar]
  27. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30, 1–11. [Google Scholar]
  28. Duan, K.; Bai, S.; Xie, L.; Qi, H.; Huang, Q.; Tian, Q. Centernet: Keypoint triplets for object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 6569–6578. [Google Scholar]
  29. Zhang, H.; Luo, Y.; Qin, F.; He, Y.; Liu, X. ELSD: Efficient Line Segment Detector and Descriptor. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 2969–2978. [Google Scholar]
  30. Hinton, G.; Vinyals, O.; Dean, J. Distilling the knowledge in a neural network. arXiv 2015, arXiv:1503.02531. [Google Scholar]
  31. Zhang, K.; Zhang, C.; Li, S.; Zeng, D.; Ge, S. Student network learning via evolutionary knowledge distillation. IEEE Trans. Circuits Syst. Video Technol. 2021, 32, 2251–2263. [Google Scholar] [CrossRef]
  32. Valverde, F.R.; Hurtado, J.V.; Valada, A. There is more than meets the eye: Self-supervised multi-object detection and tracking with sound by distilling multimodal knowledge. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 19–25 June 2021; pp. 11612–11621. [Google Scholar]
  33. Veličković, P.; Cucurull, G.; Casanova, A.; Romero, A.; Lio, P.; Bengio, Y. Graph attention networks. arXiv 2017, arXiv:1710.10903. [Google Scholar]
  34. Cuturi, M. Sinkhorn distances: Lightspeed computation of optimal transport. Adv. Neural Inf. Process. Syst. 2013, 26, 1–9. [Google Scholar]
  35. Xie, Y.; Wang, X.; Wang, R.; Zha, H. A fast proximal point method for computing exact wasserstein distance. In Proceedings of the Uncertainty in Artificial Intelligence, Online, 3–6 August 2020; pp. 433–453. [Google Scholar]
  36. Geiger, A.; Lenz, P.; Stiller, C.; Urtasun, R. Vision meets robotics: The kitti dataset. Int. J. Robot. Res. 2013, 32, 1231–1237. [Google Scholar] [CrossRef] [Green Version]
  37. Huang, K.; Wang, Y.; Zhou, Z.; Ding, T.; Gao, S.; Ma, Y. Learning to parse wireframes in images of man-made environments. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 626–635. [Google Scholar]
  38. Chen, B.B.; Gao, Y.; Guo, Y.B.; Liu, Y.; Zhao, H.H.; Liao, H.J.; Wang, L.; Xiang, T.; Li, W.; Xie, Z. Automatic differentiation for second renormalization of tensor networks. Phys. Rev. B 2020, 101, 220409. [Google Scholar] [CrossRef]
  39. Kingma, D.P.; Ba, J.A.; Adam, J. A method for stochastic optimization. arXiv 2020, arXiv:1412.6980. [Google Scholar]
  40. Schubert, D.; Goll, T.; Demmel, N.; Usenko, V.; Stückler, J.; Cremers, D. The TUM VI benchmark for evaluating visual-inertial odometry. In Proceedings of the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 1–5 October 2018; pp. 1680–1687. [Google Scholar]
  41. Nistér, D.; Stewénius, H. Scalable Recognition with a Vocabulary Tree. In Proceedings of the Computer Vision and Pattern Recognition, New York, NY, USA, 17–22 June 2006. [Google Scholar]
Figure 1. System overview. The system has four components: 1. The point–line feature extraction network (PL-Net) extracts key points, line segment and their descriptors (Section 3.1). 2. A heterogeneous graph attention network (HAGNN) is added with the positional encoding, which has N EAGAT and GBGI layers (Section 3.2). 3. A greedy near iterative matching (GIPOT) module is used to match the transformed features, which computes the affinity matrix M ˜ a f f and the assignment matrix. (Section 3.3). 4. The backend optimization includes local mapping, loop detection, and mapping.
Figure 1. System overview. The system has four components: 1. The point–line feature extraction network (PL-Net) extracts key points, line segment and their descriptors (Section 3.1). 2. A heterogeneous graph attention network (HAGNN) is added with the positional encoding, which has N EAGAT and GBGI layers (Section 3.2). 3. A greedy near iterative matching (GIPOT) module is used to match the transformed features, which computes the affinity matrix M ˜ a f f and the assignment matrix. (Section 3.3). 4. The backend optimization includes local mapping, loop detection, and mapping.
Applsci 13 03816 g001
Figure 2. The framework of our proposed network PL-Net. We input two images of size 512 × 512 and then obtain key points, line segment, and their descriptors through the ORB-FPN, extraction module, and descriptor module. The gray box in the upper-right corner shows the structure of the ORB-FPN. The upper gray branch is responsible for extracting points and lines. The lower red branch is used to extract the point–line descriptor.
Figure 2. The framework of our proposed network PL-Net. We input two images of size 512 × 512 and then obtain key points, line segment, and their descriptors through the ORB-FPN, extraction module, and descriptor module. The gray box in the upper-right corner shows the structure of the ORB-FPN. The upper gray branch is responsible for extracting points and lines. The lower red branch is used to extract the point–line descriptor.
Applsci 13 03816 g002
Figure 3. The dual attention module. The attention mechanism is adopted to adaptively aggregate different features, where the weights are normalized with the softmax function.
Figure 3. The dual attention module. The attention mechanism is adopted to adaptively aggregate different features, where the weights are normalized with the softmax function.
Applsci 13 03816 g003
Figure 4. The point-aware module. BN means batch normalization.
Figure 4. The point-aware module. BN means batch normalization.
Applsci 13 03816 g004
Figure 5. The line-aware module.
Figure 5. The line-aware module.
Applsci 13 03816 g005
Figure 6. The parallel attention block (PAB). The PAB is designed to transfer important information to the line branch output ( X F S C ). The pink box in the lower-left corner shows the structure of the CAEU.
Figure 6. The parallel attention block (PAB). The PAB is designed to transfer important information to the line branch output ( X F S C ). The pink box in the lower-left corner shows the structure of the CAEU.
Applsci 13 03816 g006
Figure 7. The overall structure of HAGNN. The first stage embeds the key points and line positions I and I into a high-dimensional vector, which generates G x and G y . Graph architecture learning is used to construct the graph. The two graphs can generate discriminative features through EAGAT and CBGI. The second stage computes the affinity matrix and the assignment matrix between two sets H i and H j and uses the assignment matrix to find matches and filter nonmatches.
Figure 7. The overall structure of HAGNN. The first stage embeds the key points and line positions I and I into a high-dimensional vector, which generates G x and G y . Graph architecture learning is used to construct the graph. The two graphs can generate discriminative features through EAGAT and CBGI. The second stage computes the affinity matrix and the assignment matrix between two sets H i and H j and uses the assignment matrix to find matches and filter nonmatches.
Applsci 13 03816 g007
Figure 8. Comparison of trajectories estimated by our SLAM method, ORB-SLAM2, and the ground truth on the KITTI dataset.
Figure 8. Comparison of trajectories estimated by our SLAM method, ORB-SLAM2, and the ground truth on the KITTI dataset.
Applsci 13 03816 g008
Figure 9. The variation curve of the pose estimation error with the number of training iterations on sequence 09 of KITTI. Each color represents a different number of sampling points.
Figure 9. The variation curve of the pose estimation error with the number of training iterations on sequence 09 of KITTI. Each color represents a different number of sampling points.
Applsci 13 03816 g009
Figure 10. Comparison of the statistical error property between our system and ORB-SLAM2 on sequence 09 of KITTI. (a) Quantitative index chart of ATE. (b) Histogram of ATE.
Figure 10. Comparison of the statistical error property between our system and ORB-SLAM2 on sequence 09 of KITTI. (a) Quantitative index chart of ATE. (b) Histogram of ATE.
Applsci 13 03816 g010
Figure 11. The experimental robot platform. (a) The physical experimental robot platform with robot hardware including a camera, depth camera, and IMU. An additional GPS/RTK was used for the ground-truth estimation. (b) The virtual simulation platform.
Figure 11. The experimental robot platform. (a) The physical experimental robot platform with robot hardware including a camera, depth camera, and IMU. An additional GPS/RTK was used for the ground-truth estimation. (b) The virtual simulation platform.
Applsci 13 03816 g011
Figure 12. The effect of point–line feature tracking. (a) The line-segment extraction results. (b) The point and line-segment extraction results. (c) The point and line matching results. We visualized the matching results with RGB color.
Figure 12. The effect of point–line feature tracking. (a) The line-segment extraction results. (b) The point and line-segment extraction results. (c) The point and line matching results. We visualized the matching results with RGB color.
Applsci 13 03816 g012
Figure 13. Simulation platform experiment. (a) The simulation platform. (b) The error mapped onto trajectory. (c) RMSE of ATE in meters after translation and scale alignment.
Figure 13. Simulation platform experiment. (a) The simulation platform. (b) The error mapped onto trajectory. (c) RMSE of ATE in meters after translation and scale alignment.
Applsci 13 03816 g013
Figure 14. The difference graph of the Wasserstein distance. (a) GIPOT under different conditions and convergence trajectory graph. We also plotted the ones for the Sinkhorn method for comparison. (b) The average time of GIPOT and Sinkhorn iterations under different conditions.
Figure 14. The difference graph of the Wasserstein distance. (a) GIPOT under different conditions and convergence trajectory graph. We also plotted the ones for the Sinkhorn method for comparison. (b) The average time of GIPOT and Sinkhorn iterations under different conditions.
Applsci 13 03816 g014
Figure 15. Qualitative evaluation of PL-Net point–line detection knowledge distillation method on the Wireframe dataset and the YorkUrban dataset.
Figure 15. Qualitative evaluation of PL-Net point–line detection knowledge distillation method on the Wireframe dataset and the YorkUrban dataset.
Applsci 13 03816 g015
Table 1. Comparison between ATE and RPE on different SLAM algorithms.
Table 1. Comparison between ATE and RPE on different SLAM algorithms.
SeqORB-SLAM2Our
ATE (m) RPE trans (%) RPE rot (deg/m)ATE (m) RPE trans (%) RPE rot (deg/m)
001.26652.50.3631.2332.90.122
014.2963.40.4202.6164.80.044
0212.7904.30.10712.7213.60.077
030.4030.80.0720.3852.00.055
040.4662.20.0550.1922.10.040
050.3482.30.1440.4021.70.056
061.1843.90.0890.5721.80.042
070.4391.30.0760.4361.60.046
083.12212.10.0762.8743.90.054
093.31915.00.1041.5372.20.054
100.9272.60.0900.9892.10.060
Table 2. Quantitative evaluation of PL-Net point-line detection knowledge distillation method on Wireframe dataset and YorkUrban dataset.
Table 2. Quantitative evaluation of PL-Net point-line detection knowledge distillation method on Wireframe dataset and YorkUrban dataset.
MethodWireframe DatasetYorkUrban DatasetFPS
F H sAP LAP F H sAP LAP
Student77.558.959.864.625.932.012.5
Teacher80.657.661.367.227.634.37.2
Table 3. Results of ablation experiment in term of the RMSE of ATE (Unit: m).
Table 3. Results of ablation experiment in term of the RMSE of ATE (Unit: m).
SeqP-SLAML-SLAMPL-SLAMORB-SLAM2LSD-SLAMPTAM
001.2036.2331.2331.2665.3472.842
013.93412.3672.6164.2963.358
027.68912.72112.79013.742
030.3935.4570.3850.4037.4312.302
040.34713.8240.1920.4662.773
050.8630.4020.3481.2930.456
060.8840.5721.1841.024
070.2550.4360.4390.423
083.1222.8743.1223.358
092.6254.7831.5373.31911.3952.048
100.4475.8240.9890.9272.8410.768
Table 4. Ablation of HAGNN.
Table 4. Ablation of HAGNN.
KnownUnknown
Match PrecisionMatching ScoreMatch PrecisionMatching Score
No EAGAT79.629.555.315.6
No CHGI66.725.348.218.5
No HAGNN63.219.451.210.3
Full89.334.278.323.8
Table 5. Experimental results of the pose estimation. Matching PL-Net features with HAGNN resulted in a significantly higher pose accuracy (AUC), precision (P), and matching score (MS) than with handcrafted or other learned methods.
Table 5. Experimental results of the pose estimation. Matching PL-Net features with HAGNN resulted in a significantly higher pose accuracy (AUC), precision (P), and matching score (MS) than with handcrafted or other learned methods.
FeatureMatcherPose Estimation AUCPMS
@ 5 @ 10 @ 20
SIFTNN7.8910.2235.3043.41.7
SIFTSuperGlue23.6836.4449.4474.17.2
SuperPointNN9.8018.9930.8822.54.9
SuperPointSuperGlue34.1844.3264.1684.911.1
LSD + LBDNN5.437.8328.5432.51.3
SOLD 2 NN18.3413.2223.5163.66.2
SuperPoint + SOLD 2 Ours35.8644.7364.4385.312.3
OursOurs36.6744.2664.7386.612.7
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lian, Y.; Sun, H.; Dong, S. Point–Line-Aware Heterogeneous Graph Attention Network for Visual SLAM System. Appl. Sci. 2023, 13, 3816. https://doi.org/10.3390/app13063816

AMA Style

Lian Y, Sun H, Dong S. Point–Line-Aware Heterogeneous Graph Attention Network for Visual SLAM System. Applied Sciences. 2023; 13(6):3816. https://doi.org/10.3390/app13063816

Chicago/Turabian Style

Lian, Yuanfeng, Hao Sun, and Shaohua Dong. 2023. "Point–Line-Aware Heterogeneous Graph Attention Network for Visual SLAM System" Applied Sciences 13, no. 6: 3816. https://doi.org/10.3390/app13063816

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop