Next Article in Journal
Machine Learning Supporting Virtual Reality and Brain–Computer Interface to Assess Work–Life Balance Conditions for Employees
Previous Article in Journal
Enhanced Readout Reliability in Phase Change Memory with a Dual-Sensing-Margin Offset-Compensated Sense Amplifier
Previous Article in Special Issue
ORPP—An Ontology for Skill-Based Robotic Process Planning in Agile Manufacturing
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A GGCM-E Based Semantic Filter and Its Application in VSLAM Systems

School of Information and Communication Engineering, Hainan University, Haikou 570228, China
*
Author to whom correspondence should be addressed.
Electronics 2024, 13(22), 4487; https://doi.org/10.3390/electronics13224487
Submission received: 9 September 2024 / Revised: 3 November 2024 / Accepted: 11 November 2024 / Published: 15 November 2024
(This article belongs to the Special Issue Application of Artificial Intelligence in Robotics)

Abstract

:
Image matching-based visual simultaneous localization and mapping (vSLAM) extracts low-level pixel features to reconstruct camera trajectories and maps through the epipolar geometry method. However, it fails to achieve correct trajectories and mapping when there are low-quality feature correspondences in several challenging environments. Although the RANSAC-based framework can enable better results, it is computationally inefficient and unstable in the presence of a large number of outliers. A Faster R-CNN learning-based semantic filter is proposed to explore the semantic information of inliers to remove low-quality correspondences, helping vSLAM localize accurately in our previous work. However, the semantic filter learning method generalizes low precision for low-level and dense texture-rich scenes, leading the semantic filter-based vSLAM to be unstable and have poor geometry estimation. In this paper, a GGCM-E-based semantic filter using YOLOv8 is proposed to address these problems. Firstly, the semantic patches of images are collected from the KITTI dataset, the TUM dataset provided by the Technical University of Munich, and real outdoor scenes. Secondly, the semantic patches are classified by our proposed GGCM-E descriptors to obtain the YOLOv8 neural network training dataset. Finally, several semantic filters for filtering low-level and dense texture-rich scenes are generated and combined into the ORB-SLAM3 system. Extensive experiments show that the semantic filter can detect and classify semantic levels of different scenes effectively, filtering low-level semantic scenes to improve the quality of correspondences, thus achieving accurate and robust trajectory reconstruction and mapping. For the challenging autonomous driving benchmark and real environments, the vSLAM system with respect to the GGCM-E-based semantic filter demonstrates its superiority regarding reducing the 3D position error, such that the absolute trajectory error is reduced by up to approximately 17.44%, showing its promise and good generalization.

1. Introduction

Visual simultaneous localization and mapping (vSLAM) is becoming a widely increasing research community, partly because of its affordable applications in autonomous navigation in complex and real environments [1] and partly because of leveraging vision for the semantic perception that results in computational affordability of mapping [2]. Traditional vSLAM methods refer to epipolar geometry for calculating camera motion, reconstructing 3D scenes, and then resolving camera/robot trajectories through huge filtering or optimization [3]. vSLAM shows promising performance in ideal scenarios while presenting many challenges for these conditions: (1) VSLAM algorithms proposed on the assumption of static scene usually reconstruct incorrect camera trajectories and 3D map points in dynamic targets presented scenarios [4]. (2) Traditional vSLAM algorithms reconstruct trajectories using pixels or low-level feature correspondences from keyframes, suffering from the accumulation of errors caused via a trajectory drift, especially due to uncertainty scale in monocular vSLAM [5]. (3) VSLAM algorithms relying on pixel features are not sufficiently robust in complex scenes such as brightness changes, object occlusion, low texture, and repeated texture in real environments, thus resulting in frame loss, wrong localization, and poor trajectory reconstruction. To overcome such challenges, Random Sampling Consistency (RANSAC)-based outlier detection has been developed in the past, which treats dynamic objects as outliers and removes them from the trajectory reconstructing and mapping. However, the typically epipolar constraints method generates feature correspondences using statistical or other matching criteria in feature space–image space, which is infeasible since identifying outliers in high-level semantic image space is difficult [6].
Compared to traditional image processing techniques and deep learning networks, ref. [7] reveals semantic patterns associated with different scenes to realize scene perception of complex environments. We can conclude that referring to semantic information can enhance the resilience of vSLAM in complex environments since semantic feature can be more stable than pixel features. In our previous study, a semantic filter based on Fast R-CNN learning was proposed to explore the semantic information of outliers [8]. However, the semantic information was defined by the proportion of inliers and outliers, ignoring the texture-rich environments, requiring further detailed semantic analysis. Experimental results show that removing regions containing complex textures (e.g., dense foliage) can effectively reduce the feature matching error and thus the VSLAM localization error to some extent. However, the Fast-RCNN filter is not effective in this case, and the prediction accuracy only approximates to 60%. Consequently, semantic levels of complex scenes for high-quality camera trajectories and map reconstructions in vSLAM systems are urgently required.
This paper, focusing on the texture-rich semantic scenes, presents a further detailed analysis of semantic levels, proposing a GGCM-E (Gray-Level Gradient Co-occurrence Matrix–Energy)-based semantic filter. With respect to YOLOv8 networks, the GGCM-E-based semantic filter can effectively improve the accuracy of camera trajectories and maps in vSLAM systems. The semantic filter at the vSLAM front end can effectively identify the semantic regions in the frames and classify them in terms of texture complexity by means of the GGCM-E descriptor, after which the semantic regions can be effectively targeted to filter out the low-level and dense texture-rich texture scenes to reduce low-quality feature correspondences, thus resulting in accurate camera motion and map points in many challenging scenes, such as dense texture-rich outdoor and indoor environments. To be precise, we propose the following contributions in this paper.
  • A novel method to analyze the semantic level using GGCM-E features is proposed. This method calculates GGCM-E features of semantic patches according to visual complexity of texture images and then classifies the semantic patch in terms of its semantic level. These classified semantic patches are feasible for the semantic description of dense texture-rich scenes, helping generate semantic filter for low-level semantic scene filtering.
  • A semantic filter trained using YOLOv8 networks is applied to the ORB-SLAM3 system, assisting ORB-SLAM3 to achieve accurate and robust trajectory reconstruction and mapping. Through filtering out low-level semantic scenes, the semantic filter effectively addresses outlier issues in the semantic image space, thereby improving the localization accuracy of the ORB-SLAM3 system.
Figure 1 shows the overview of the proposed semantic filter-based ORB-SLAM3 framework. The green part is the proposed semantic filter module, which pre-processes the frames by identifying semantic patches, extracting the GGCM-E descriptor, classifying the semantic level, and filtering out low-level semantic patches to produce high-quality frames for the visual odometry. The rest of this paper is organized as follows: Section 2 briefly summarizes the related works. The proposed semantic filtering is described in detail in Section 3. Section 4 presents the experimental results and analysis. Conclusions are drawn in Section 5.

2. Related Works

Traditional image matching-based vSLAM relies on feature descriptors to obtain matches and then estimates the fundamental matrix using epipolar geometry, reconstructing camera trajectories and maps in numerous environments. Epipolar geometry estimation can become computationally inefficient and inaccurate when there are low-quality correspondences. Although the RANSAC method shows superiority in eliminating low-quality correspondences [9], due to the computational complexity and noise-sensitivity, it can fail in many situations, such as dynamic model mismatch and texture-rich scenes outlier interference, and cumulative drift error. Therefore, improving the correspondences qualities to eliminate outliers is required in vSLAM systems, promoting the development of many semantic vSLAM methods. According to the various methodologies for implementing deep neural networks in vSLAM, these methods can be classified into the following two major categories.

2.1. Improved vSLAM Using Deep Neural Network-Based Specific Modules

Quach et al. propose SupSLAM using a new feature called SuperPoint to replace the ORB descriptor, releasing a pre-trained deep neural network to obtain accurate feature correspondences with a high computational efficiency [10]. Xiao Zhang et al. proposed the SL-SLAM method throughout the SLAM system with Lightglue, a SOTA (state-of-the-art) matching method that can generate local feature matches associated with a deep neural network [11]. The deep neural network-based improved vSLAM exhibits enhanced resilience in complex scenes, such as sparse textures and fluctuating light conditions. However, it is worth noting that a significant investment of learning time is required for the end-to-end approach, such as entailing camera motion estimation through a deep neural network to map from scenes to camera trajectories. The end-to-end improved vSLAM can be subdivided into supervised and unsupervised approaches [12,13,14], where the supervised approach relies on camera motion or ground-truth depth maps, and the unsupervised approach refers to the mutual constraints between camera motion and depth maps [15,16,17]. Note that depth maps are important in end-to-end work due to their benefits in network training and accurate camera trajectory estimation. However, this approach necessitates huge training data with considerable computational resources and storage space, rendering a challenge to real-time requirements of the tracking module. Consequently, it is necessary to perform research on hybrid approaches associated with deep learning-based vSLAM to enhance the functionality of vSLAM modules.

2.2. Improved vSLAM Using Deep Neural Networks + vSLAM Modules

In real environments, especially outdoors, dynamic targets are inevitably encountered, resulting in localization errors in vSLAM that are caused by low-quality feature correspondences. Zhong et al. proposed a Detect-SLAM method composed of single-shot multibox detector (SSD) network-based dynamic target detection [18], eliminating dynamic feature correspondences through a feature matcher and key-point-based probabilistic propagation of movement. However, due to failure to consider the state of the dynamic category target, whether it is in motion or at rest, this method also deals with static objects, i.e., parked cars, and excludes feature correspondences in highly dynamic scenes, thus reducing huge high-quality feature correspondences to result in accuracy degradation and a trajectory shift. Qiuyu Zang et al. employed a two-step process to detect potentially dynamic objects using YOLOv5 networks and exclude dynamic scenes through the Lucas–Kanade optical flow and RANSAC algorithms [19]. Berta proposed a DynaSLAM consisting of Mask-RCNN-based semantic segmentation for ORB-SLAM2 to filter out dynamic feature correspondences [20,21]. Currently, the DIG-SLAM applies YOLOv7-based instance segmentation to describe targets in terms of their movement consistency [22], estimating camera motion by the left static point and line features. These semantic segmentation and instance segmentation-based vSLAMs require significant computation and time–cost frame processing, making it difficult to realize real-time vSLAM.
Consequently, many point–line–plane features associated with approaches to save computation have been developed [23]. However, these approaches behave poorly in many challenging environments with respect to low-level feature correspondences, such as in low-texture and dense texture-rich scenes. Thanks to the deep neural network, acquiring semantic information can help vSLAM to perceive high-level scene context and yield feature correspondences dependence to enhance the resilience of vSLAM. Many semantic approaches regard semantic information of specific environmental landmarks [24], enabling accurate target measurements, resulting in precious localization. TextSLAM is proposed to cope with target textual objects, e.g., door numbers and billboards [25]. By extracting relevant semantic features, vSLAM localizes accurately and robustly on a map even under challenging scenes, i.e., image blurring, different viewpoints, and illumination changes. Qian developed the SmSLAM + LCD (Semantic SLAM + Loop Closure Detection) approach by integrating high-level 3D semantic information with low-level feature correspondences into a unified semantic vSLAM framework [26], enabling accurate closed-loop detection and effective drift reduction. In indoor environments , ref. [27] employed a semantic association method using parking lines to construct maps for self-driving vehicles and indoor parking. Kang et al. proposed a distinct semantic labeling to compare the semantic categories of map point–reprojection pairs between images by applying normalized information distance in information theory to improve the camera trajectory estimation in vSLAM [28]. Taking these into consideration, vSLAM associated with semantic information can improve its resilience to complex environments, realizing high-accuracy localization and reducing cumulative error.
According to human visual attention mechanisms, the semantic context of feature correspondences in the image space strongly influences the image matching-based vSLAM [29]. Not all feature correspondences have the same semantic labels in a vSLAM system. We have conducted extensive experiments to conclude that semantic labels, with respect to the outlier ratio, can reveal the semantic level of feature correspondences. Filtering out low-level semantic scenes can improve the quality of feature correspondences, thus helping vSLAM perform better in dynamic and static scenarios [8,30]. However, the outlier ratio-based semantic filter without texture information of semantic scenes has low accuracy, thus requiring further semantic analysis when dealing with complex environments. Considering the latest target detection algorithm-YOLOv8 networks, a novel semantic filter approach based on the GGCM-E feature is proposed, effectively improving semantic filter accuracy and the performance of vSLAM systems. Experiments demonstrate that, for dense texture-rich scenes, the semantic-based vSLAM approach can result high-quality feature correspondences, achieving accurate and robust trajectory reconstruction and mapping.

3. Methodology

3.1. Semantic Patches of Inliers

The goal of this paper is to filter out low-level semantic scenes and assist accurate vSLAM. In deep learning-based methods, data describing target objects can be pre-collected, labeled, and used to train a network. The model can predict objects with similar features in scenes. As illustrated in Figure 2, we initially focus on collecting semantic patches of inliers using challenging KITTI sequences, and we capture real environments [31]. Firstly, ORB brute force matching is performed on key frames calculated from these sequences, resulting high-ratio potential mismatches. Referring to the RANSAC algorithm, low-quality correspondences are filtered out using epipolar geometry. Then, semantic patches of n × n size centered on the inlier are extracted. The prior study indicates that the semantic filter with n = 33 shows the best performance in the vSLAM system. To extend the generalization on complex scenes, a YOLOv8-based network is pre-trained using semantic patches with the size of n = 33, since the YOLOv8 network provides high accuracy, low computation, and high flexibility in target detection. Finally, these semantic patches are classified into different classes according to GGCM-E features.

3.2. Semantic Patch Labeling Based on GGCM-E Features

In our previous study, semantic patches were labeled according to outlier ratios, assigning the centered inlier a semantic level, which is its quality with respect to feature correspondences. However, the outlier ratio-based semantic patches labeling cannot perceive high-level semantic information, such that the relevant semantic filter has low accuracy in segmenting within complex environments. Consequently, we conducted further research to analyze the semantic information in semantic patches. Our previous research demonstrates that the trajectory error in semantic filter-based vSLAM can be significantly reduced when filtering out dense texture-rich scenes, such as scenarios with trees. The reason we envisage is that in the region of dense, complex textures, the number of ORB feature points extracted is very high. However, because of the complex similarity of the texture, this will lead to a mismatch between the ORB feature points of consecutive frames, and this kind of mismatch cannot be entirely eliminated by just relying on the RANSAC algorithm. Then, it generates a localization error. In order to investigate how the feature correspondences in dense texture-rich scenes affect the accuracy in vSLAM systems, we propose a GGCM-E feature descriptor to classify semantic patches, which can be implemented to classify the texture complexity level of semantic image patches and then detect the regions of dense and complex textures, enabling us to predict low-level semantic scenes associated with deep learning networks.
The Gray-level Gradient Co-occurrence Matrix (GGCM) is a statistical method that focuses on the semantic information within images in terms of the grayscale and gradient. As we know, grayscale images can describe the semantic contexts for different visual tasks, and gradient images proved to be more discriminant and reliable when it comes to texture-less targets in complex scenes. Consequently, the GGCM feature can reveal the semantic information with respect to pixel levels, providing a comprehensive prediction in dense texture-rich scenes.
Firstly, semantic patches of inliers are divided into l equal-sized S- r e g i o n s , s ( i , j ) , to process GGCM feature calculations. Then, the number of pixel pairs with coordinates ( x , y ) = ( F ( i , j ) , G ( i , j ) ) is accumulated to obtain h ( x , y ) for the covariance matrix H. Note that the notion h ( x , y ) is defined as the entire pixel-pair number in terms of the x- v a l u e pixel in the normalized grayscale image F ( i , j ) (Equation (1)) and the y- v a l u e pixel in the normalized gradient image G ( i , j ) (Equation (2)). Finally, the normalization of the GGCM is conducted to obtain h ( x , y ) . Equation (1) defines the normalized grayscale image of the S- r e g i o n s :
F ( i , j ) = f ( i , j ) · ( L g y 1 ) f max + 1
where f ( i , j ) denotes the gray value of pixel ( i , j ) in the original image, F ( i , j ) denotes the gray value of pixel ( i , j ) in the normalized grayscale image, f m a x is the maximum gray value in the original image, and L g y is the maximum gray value after normalization, with L g y being set to l.
After that, the normalized gradient image is calculated as in Equation (2):
G ( i , j ) = g ( i , j ) · ( L g t 1 ) g max + 1
where g ( i , j ) is the gradient value of pixel ( i , j ) in the original image, G ( i , j ) is the gradient value of pixel ( i , j ) in the normalized grayscale image, g m a x is the maximum gradient value in the original image, and L g t is the maximum gradient value after normalization, with L g t being set to l. Note that, g ( i , j ) is derived from the square sum of the Sobel’s operator in the horizontal and vertical directions, respectively, according to Equation (3):
g ( i , j ) = g x 2 + g y 2
Here,
g x = f ( i + 1 , j ) f ( i , j )
g y = f ( i , j + 1 ) f ( i , j )
Count the pixel pairs with respect to x = F ( i , j ) and y = G ( i , j ) in the normalized grayscale image F ( i , j ) and the normalized gradient image G ( i , j ) .
x = F ( i , j ) , x 1 , L g y
y = G ( i , j ) , y 1 , L g t
Until now, the normalized GGCM h (x, y) can be derived, as follows:
h ( x , y ) = h ( x , y ) x = 0 L g y y = 0 L g t h ( x , y )
E = x y h ( x , y ) 2
According to the normalized GGCM, we define the energy of the statistical property as Equation (9):
As demonstrated in Figure 3, when the energy is less than a threshold T h , the S- r e g i o n can be characterized by complex textures, such that a significant change within the gradient value and the gray value occurs, presenting the relevant GGCM-E feature with 1- v a l u e result r. On the contrary, if the energy is larger than or equal to the preset threshold, the S- r e g i o n contains many backgrounds, indicating that the texture is sparse and that non-existed GGCM-E features change with respect to the gradient value and the gray value, thus presenting a 0- v a l u e result r. We illustrated the T h parameter in Figure 3. We set this value to 0.25. The reason for this is that when the proportion of flat and texture-free area reaches 50% of the S- r e g i o n , the probability that its corresponding certain value in the gray-level gradient co-occurrence matrix is normalized is 0.5. Moreover, when calculating the energy characteristics of the whole matrix, the sum-of-squares operation will be performed, and the energy value corresponding to this region will be 0.25. So, when we calculate the energy value of the gray-level gradient covariance matrix to be less than 0.25, the flat untextured region is less than half of the area of this S- r e g i o n . Alternatively, it means that the area of the textured region is more than half of the area of the S- r e g i o n , and we default to the S- r e g i o n with a complex texture, GGCM -E feature number plus 1. Then, by calculating all S- r e g i o n s , the sum of the number of GGCM-E features is derived, which is used as the basis for classifying the image complexity of different image semantic patches. When the sum of the number of GGCM-E features of a certain semantic patch is l, this indicates that all the S- r e g i o n s exhibit GGCM-E characteristics, i.e., each S- r e g i o n exhibits complex textures. Thus, the semantic patch is the one that has the most complex texture semantic patch, which is suitable to be divided into a individual class to study. After the S r e g i o n s transformation and the energy calculation, GGCM-E features are obtained in terms of the r result. The GGCM-E feature calculation is summarized in Algorithm 1. Ultimately, these semantic patches will be classified in accordance with the number of GGCM-E features, which serves to ascertain the level of texture complexity.
To the end, a sequence of semantic patches labeled according to the GGCM-E features is obtained as the training data before they are input into the YOLOv8 network.
Algorithm 1 Compute GGCM_E feature values.
Input: Semantic patches of Inliers
Output: GGCM_E feature values
 1:
L g y l ,   L g t l ,   G G C M _ E 0
 2:
for S in p a t c h  do
 3:
      Spatch [ L : R , U : D ] // Extract S from the patch.
 4:
       ( g x , g y ) Sobel gradients of S
 5:
       g ( i , j ) g x 2 + g y 2 // Compute Sobel gradient magnitude.
 6:
       f _ m a x max ( S )
 7:
       F ( i , j ) f ( i , j ) / f _ m a x × ( L g y 1 ) + 1 // Normalize the gray level matrix.
 8:
       g _ m a x max ( g ( i , j ) )
 9:
       G ( i , j ) g ( i , j ) / g _ m a x × ( L g t 1 ) + 1 // Normalize the gradient matrix.
10:
      for pixel ( i , j ) in S do
11:
             F x y round ( F ( i , j ) )
12:
             G x y round ( G ( i , j ) )
13:
             h ( x , y )     h ( F x y , G x y ) + 1 // Fill the gray-level gradient co-occurrence matrix.
14:
      end for
15:
       h ( x , y )     h ( x , y ) / h ( x , y )
16:
       E n e r g y     ( h ( x , y ) 2 ) // Compute energy of the normalized h ( x , y ) .
17:
      if  E n e r g y < T h  then
18:
             G G C M _ E G G C M _ E + 1 // Compute GGCM_E feature value.
19:
      end if
20:
end for

4. Analysis of Experiments and Results

4.1. Semantic Filter Model Training

After calculating the GGCM-E features of all the semantic patches, we classified them and utilized the YOLOv8 neural network for training. The initial approach was to categorize all semantic patches into one class. However, the experimental results indicated that this approach was not optimal. A superior classification method is devised based on the formulation of Equation (11). The formula’s interpretation is that the label with the greatest number of GGCM-E features is extracted as a standalone class. In contrast, the remaining classes are classified as a group comprising every c feature. The final total number of classes is, therefore, represented by the value of C. The value of l is set to 9, 16, and 25, respectively, and the model is trained according to one class within our own designed classes, as referenced in Equation (11), on this dataset in a separate training phase. The performance of the obtained model is demonstrated in Table 1.
c = l 1
C = l 1 l 1 + 1
Table 1 demonstrates that the optimal results are achieved when l is set to 16, T h is set to 0.25, and the semantic patches are classified into six classes. The experimental results serve to validate the scientific validity of the GGCM-E classification methodology. Semantic patches are obtained from three sources: the open source KITTI dataset, the TUM dataset provided by the Technical University of Munich, and our collection of TUM images of real scenes. The KITTI dataset comprises authentic images of urban, rural, and highway scenes, including intricate and dense texture scenes, such as leaves and houses, which are replete with scene images. It also furnishes the truth values for the subsequent comparison of the filter on ORB-SLAM3. The TUM dataset, provided by the Technical University of Munich, and our own TUM dataset, collected in the real world, contain a variety of complex and dense texture scenes, including wooden boards, tiles, messy desktops, and so on. Using datasets with different scenes is beneficial for improving the filter’s generalization ability. The total dataset contains approximately 11,500 image frames and 230,000 labeled file representations of semantic image patches. We divide the labeled semantic patches into six classes with training-to-validation ratios of 8:2, such as semantic patches with label1 to label3 are one-class filters, semantic patches with label4 to label6 are two-class filters, semantic patches with label7 to label9 are three-class filters, semantic patches with label10 to label12 are four-class filters, semantic patches with label13 to label15 are five-class filters, and semantic patches with label16 are six-class filters. As mentioned above, due to its highest accuracy, a YOLOv8 network-based semantic filter is then used to train for various challenging environments. We use the mechanism of transfer learning to train the network. For the backbone network, we adopt the pre-trained parameters, where the initial learning rate (lr0) is set to 0.01, the batch size is set to 8, and the pre-trained weights are set to YOLOv8l.pt. We train the network with 600 epochs in total to evaluate the loss convergence. We implement the training on a GPU server with four V100-PCIE-16GB and 320GB RAM in the programming environment Linux 20.04.6.
Table 2 summarizes the semantic filter accuracy using YOLOv8 networks and GGCM-E feature-based semantic patches’ information. Notably, the most common labels for the semantic filter are 13, 14, and 15 (five-class filter), respectively, such that the five-class semantic patch ratio is 36.9%. From the calculation in Algorithm 1, the 13–15 labeled semantic patches contain complex texture-rich scenes. We then compare the precision (P), recall (R), and mean average precision (MAP@0.5) on the six classes of semantic patches with respect to the relative semantic filter module. As presented in the results, the average precision, recall, and MAP@0.5 of the six classes based the semantic filter are 77.6%, 76.4%, and 82.4%, respectively.
We can conclude from the semantic filtering on the challenging KITTI dataset and real outdoor dataset that, as shown in Figure 4 and Figure 5, the GGCM-E-based semantic filter performs well with moderate detection confidence. Notably, the two-class, three-class, and four-class semantic patches predominantly depict scenes containing discernible edges and more regular textures, whereas five-class and six-class semantic patches are primarily blurred or dense texture-rich scenes, providing interesting results with regard to the research on the five-class- and six-class-based semantic filter.

4.2. Experiments on Semantic Filter-Based ORB-SLAM3

To verify the performance of the semantic filter-based vSLAM, we implement the proposed method on top of ORB-SLAM3 using monocular input data. KITTI07 and KITTI05 sequences are selected as the dataset for evaluation in the target domain because they contain dense texture-rich scenes. As discovered by Raul Muir Altar and Juan D. Tardos, accurate trajectory estimation can be obtained in terms of the average trajectory by running the ORB-SLAM3 system many times [21]. We implement the semantic filter-based ORB-SLAM3 11 times to evaluate the trajectory estimation. The comparison of the absolute attitude error (ATE) between ORB-SLAM3 and the proposed semantic filter-based ORB-SLAM3 is summarized in Table 3 and Table 4.
As presented in Table 3 and Table 4, after filtering out the five-class and the six-class semantic scenes, the ATEs are effectively reduced. The six-class semantic filter-based vSLAM achieves the top performance, such that the root mean square error (RMSE) is reduced by about 17.4%, and the maximum and minimum localization errors are the lowest in the experiments with the KITTI07 sequence. The RMSE of the experiment with the KITTI05 sequence is diminished by approximately 13.7%, and the minimum localization error is the lowest. This indicates that vSLAM is more accurate when the low-level semantic scenes in patches labeled 13, 14, 15, and 16 are filtered out. As described previously, these labeled semantic patches contain blurry and dense texture-rich scenes, resulting in trajectory errors. We hence conclude that filtering out these semantic patches leads to the least loss in localization precision and trajectory estimation. Consequently, the six-class semantic patches are low-level semantic scenes, requiring further research on their performance on vSLAM systems. Accordingly, the six-class area was selected as the filtering region. Moreover, further comprehensive experiments were conducted in diverse scenarios through GGCM-E+ORB-SLAM3. As demonstrated in Figure 6, the trajectory provided by GGCM-E semantic filter-based vSLAM effectively completes loop closure with highly accurate ground-truth disparities.

4.3. Extensive Experiments on the Six-Class Semantic Filter-Based ORB-SLAM3

To further illustrate the performance obtained with the six-class semantic filter, we evaluate the six-class semantic filter-based ORB-SLAM3 on other KITTI sequences (sequences 00 to 09). We compare the estimated camera trajectories to their ground-truth trajectories while playing the sequences at the frame rate at which they are recorded.
In Figure 7, we compare the mapping results provided by the semantic filter-based ORB-SLAM3 with respect to the ground truth. As the figures show, we successfully detect most loop closures, i.e., sequences 00 and 09, and we have the most similar trajectory compared to the ground truth. Note that several accumulated trajectory drifts yield at sequence 00 because of the large camera rotation in the monocular vSLAM. Nevertheless, due to the large-scale loop closure that occurred in sequence 00, this drift was corrected after the loop was detected and closed. A qualitative comparison of the absolute position error (APE) is shown in Figure 8, from which the accuracy of our semantic filter can be qualitatively judged. In accordance with the results shown in the figures, our approach is capable of filtering out low-level semantic scenes, performing best on most of the sequences. For some parts of sequence 05, our approach yields several bad positions at 200 s and 250 s. This occurs due to the rapid increase in low-level semantic scenes generated by image rotation, e.g., the 2000th and 2500th frames (FPS = 10) are recorded using large rotations. Note that when the camera rotation is large in the semantic filter-based vSLAM, the high-level semantic decreases, leading to more high-quality feature correspondences being filtered out by the semantic filter, thus resulting in high position errors.
Table 5 presents the comparison of RMSE metrics of the ATEs of various vSLAM methods. We compare the results obtained by ORB-SLAM3, GGCM-E+ORB-SLAM3, and our previously proposed Inliers-Outliers+ORB-SLAM3 semantic filter, where the semantic filter is generated using a Faster R-CNN learning-based method [8]. It can be concluded from these results that the semantic filter based on GGCM-E features is more effective than the inlier–outlier semantic filter. After filtering out low-level and dense texture-rich semantic scenes, the quality of feature correspondences was effectively improved, thus decreasing the RMSE of camera trajectories in vSLAM systems, such that the largest reduction was by about 17.44%.

4.4. Experiments on Dense Texture-Rich Sequences

To further explore the semantic filter-based vSLAM on dense texture-rich environments, we take seven sequences (named DTR sequences) consisting of large dense texture-rich scenes from the TUM dataset for evaluating our method. As illustrated in Figure 9, for the fre1_plant sequence, a flower pot near the wooden floor is recorded with a considerable duration to generate large dense texture-rich scenes. The fre1_room sequence contains many dense and complex scenes, such as the curtains, and wooden panels. The fre2_360_hemisphere sequence is recorded with the indoor environment, where the wall tiles, ceiling, and other surfaces exhibit dense texture-rich patterns. The same scene is recorded as the fre2_large_with_loop sequence, obtaining dense texture-rich scene-based loop closure for the camera motion trajectory. As shown by the experimental results summarized in Table 6, the semantic filter-based vSLAM works well on full sequences, effectively decreasing the RMSE with respect to the absolute trajectory error, such that the top reduction on the fre2_360_hemisphere sequence is approximately 11.47%. In addition, we compare our method with PLP-SLAM because of its resilience to repetitive and dense texture environments [32], which is achieved by utilizing point and line features for tracking. We can conclude from the relevant results that our method performs best on most of the dense texture-rich scenes, yielding high-accuracy position estimation. Despite the fact that our method estimates the position well with vSLAM, its ability on scenes consisting of large rotation, i.e., a fre1_plant sequence, is slightly lower than that of PLP-SLAM.
A comparison of the camera trajectories in the DTR sequence between ORB-SLAM3 and our method is demonstrated in Figure 10, where our proposed method successfully reconstructed the landmarks around the camera path with a lower trajectory error. The above experiments demonstrate that our proposed semantic filter is stable for filtering out low-level semantic scenes, especially for dense texture-rich scenes, thus enabling vSLAM systems to operate well in various challenging environments.

4.5. Experiments on Other Semantic Filter-Based vSLAM Systems

To evaluate the generalization of the proposed semantic filter, we implement it in other vSLAM systems and conduct several experiments on KITTI sequences. Considering the case of dynamic target filtering, DynaSLAM is selected as the baseline algorithm for a comparison with [20]. The results presented in Figure 11 show that the GGCM-E semantic filter-based DynaSLAM works well concerning camera trajectory estimation, especially around the straight camera path, improving the camera trajectory with the help of a semantic filter, indicating the superiority of the semantic filter on low-level semantic scene filtering, even after dynamic target filtering.
In Table 7, the GGCM-E filter was integrated with Structure-SLAM [33], LDSO [34], and DynaSLAM, and a comparative assessment was conducted on KIITTI sequences. The experimental outcomes indicated that the filter exhibited enhanced performance for Structure-SLAM and DynaSLAM, and the RMSE was diminished across all sequences, which substantiates the filter’s role in feature-based vSLAM. The impact of generality on the GGCM-E+LDSO on 05 and 06 sequences is contradictory. The RMSE increased somewhat, and the error increased compared to LDSO. We hypothesize that this may be due to the fact that LDSO is a direct-method SLAM and due to the localization accuracy being less correlated with the local complex texture region.
In Figure 12, the present study compares the error details of Structure-SLAM, LDSO, and DynaSLAM with the applied GGCM-E semantic filter in accordance with the KITTI07 experiment, respectively. The results demonstrate that the majority of SLAMs with the GGCM-E semantic filter have a notable reduction in the high error portion. The minimum errors obtained are effectively reduced, although the maximum errors are higher with GGCM-E+LDSO. It is notable that all SLAMs with GGCM-E semantic filters exhibit a reduction in APEs compared to those without such filters. The above experiments demonstrate that the proposed semantic filter is capable of filtering out low-level dense texture-rich scenes, being suitable for various vSLAM systems in challenging complex scenes. Consequently, filtering out low-level semantic information in dense texture-rich scenes using a GGCM-E-based semantic filter will generate high-quality feature correspondence, thus helping vSLAM achieve accurate and robust trajectory reconstruction and mapping.

5. Conclusions

In this paper, we propose a GGCM-E-based semantic filter with YOLOv8 networks. The proposed semantic filter can automatically filter out low-level semantic scenes. According to our previous research, different semantic patches centered on the inliers of various frames are labeled in terms of GGCM-E features, generating six classes of semantic filters. High-quality correspondences are obtained by applying these semantic filters to vSLAM systems, improving image matching-based camera trajectory estimation and mapping. The localization accuracy of the proposed semantic filter-based pipeline within different vSLAM frameworks is also evaluated. Extensive experiments on the KITTI dataset and the TUM dataset are performed. The results demonstrate that the position error can be significantly reduced when the semantic filter is used in feature-based vSLAM systems. For dynamic outdoor and indoor environments, the respective ATE can be reduced by up to approximately 17.44% and 11.47%, respectively. A comparison of the ATE results with respect to the semantic filter-based DynaSLAM, Structure-SLAM, and ORB-SLAM3 shows the promising generalization capabilities of the proposed method with regard to helping various vSLAM systems localize accurately and robustly on a map. In the future, we will include machine learning methods that can access semantic level-based information in different semantic scenes.

Author Contributions

Conceptualization, Y.L. and C.S.; methodology, Y.L.; validation, Y.L.; formal analysis, Y.L. and J.W.; resources, Y.L. and J.W.; data curation, Y.L.; writing—original draft preparation, Y.L.; writing—review and editing, C.S.; supervision, C.S.; project administration, C.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Hainan Provincial Natural Science Foundation of China (No. 624RC480) and the Scientific Research Foundation for Hainan University (No. KYQD(ZR)-21013).

Data Availability Statement

Data are contained within this article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Haomin, L.; Guofeng, Z.; Hujun, B. A survey of monocular simultaneous localization and mapping. J. Comput.-Aided Des. Comput. Graph. 2016, 28, 855–868. [Google Scholar]
  2. Engelhard, N.; Endres, F.; Hess, J.; Sturm, J.; Burgard, W. Real-time 3D visual SLAM with a hand-held RGB-D camera. In Proceedings of the RGB-D Workshop on 3D Perception in Robotics at the European Robotics Forum, Vasteras, Sweden, 6–8 April 2011; Volume 180, pp. 1–15. [Google Scholar]
  3. Strasdat, H.; Montiel, J.M.; Davison, A.J. Visual SLAM: Why filter? Image Vis. Comput. 2012, 30, 65–77. [Google Scholar] [CrossRef]
  4. Davison, A.J.; Reid, I.D.; Molton, N.D.; Stasse, O. MonoSLAM: Real-time single camera SLAM. IEEE Trans. Pattern Anal. Mach. Intell. 2007, 29, 1052–1067. [Google Scholar] [CrossRef] [PubMed]
  5. Mur-Artal, R.; Montiel, J.M.M.; Tardos, J.D. ORB-SLAM: A versatile and accurate monocular SLAM system. IEEE Trans. Robot. 2015, 31, 1147–1163. [Google Scholar] [CrossRef]
  6. Toft, C.; Stenborg, E.; Hammarstrand, L.; Brynte, L.; Pollefeys, M.; Sattler, T.; Kahl, F. Semantic match consistency for long-term visual localization. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 383–399. [Google Scholar]
  7. Wang, X.; Shrivastava, A.; Gupta, A. A-fast-rcnn: Hard positive generation via adversary for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2606–2615. [Google Scholar]
  8. Shao, C.; Zhang, L.; Pan, W. Faster R-CNN learning-based semantic filter for geometry estimation and its application in vSLAM systems. IEEE Trans. Intell. Transp. Syst. 2021, 23, 5257–5266. [Google Scholar] [CrossRef]
  9. Luong, Q.T.; Faugeras, O.D. The fundamental matrix: Theory, algorithms, and stability analysis. Int. J. Comput. Vis. 1996, 17, 43–75. [Google Scholar] [CrossRef]
  10. Quach, C.H.; Phung, M.D.; Le, H.V.; Perry, S. SupSLAM: A robust visual inertial SLAM system using SuperPoint for unmanned aerial vehicles. In Proceedings of the 2021 8th NAFOSTED Conference on Information and Computer Science (NICS), Hanoi, Vietnam, 21–22 December 2021; IEEE: New York, NY, USA, 2021; pp. 507–512. [Google Scholar]
  11. Xiao, Z.; Li, S. SL-SLAM: A robust visual-inertial SLAM based deep feature extraction and matching. arXiv 2024, arXiv:2405.03413. [Google Scholar]
  12. Wang, S.; Clark, R.; Wen, H.; Trigoni, N. End-to-end, sequence-to-sequence probabilistic visual odometry through deep neural networks. Int. J. Robot. Res. 2018, 37, 513–542. [Google Scholar] [CrossRef]
  13. Zhou, H.; Ummenhofer, B.; Brox, T. Deeptam: Deep tracking and mapping. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 822–838. [Google Scholar]
  14. Li, Y.; Wang, G.; Ji, X.; Xiang, Y.; Fox, D. Deepim: Deep iterative matching for 6d pose estimation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 683–698. [Google Scholar]
  15. Li, R.; Wang, S.; Long, Z.; Gu, D. Undeepvo: Monocular visual odometry through unsupervised deep learning. In Proceedings of the 2018 IEEE International Conference on Robotics and Automation (ICRA), Brisbane, Australia, 21–25 May 2018; IEEE: New York, NY, USA, 2018; pp. 7286–7291. [Google Scholar]
  16. Yin, Z.; Shi, J. Geonet: Unsupervised learning of dense depth, optical flow and camera pose. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 1983–1992. [Google Scholar]
  17. Ranjan, A.; Jampani, V.; Balles, L.; Kim, K.; Sun, D.; Wulff, J.; Black, M.J. Competitive collaboration: Joint unsupervised learning of depth, camera motion, optical flow and motion segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 12240–12249. [Google Scholar]
  18. Zhong, F.; Wang, S.; Zhang, Z.; Wang, Y. Detect-SLAM: Making object detection and SLAM mutually beneficial. In Proceedings of the 2018 IEEE Winter Conference on Applications of Computer Vision (WACV), Lake Tahoe, NV, USA, 12–15 March 2018; IEEE: New York, NY, USA, 2018; pp. 1001–1010. [Google Scholar]
  19. Zang, Q.; Zhang, K.; Wang, L.; Wu, L. An adaptive ORB-SLAM3 system for outdoor dynamic environments. Sensors 2023, 23, 1359. [Google Scholar] [CrossRef] [PubMed]
  20. Bescos, B.; Fácil, J.M.; Civera, J.; Neira, J. DynaSLAM: Tracking, mapping, and inpainting in dynamic scenes. IEEE Robot. Autom. Lett. 2018, 3, 4076–4083. [Google Scholar] [CrossRef]
  21. Mur-Artal, R.; Tardós, J.D. Orb-slam2: An open-source slam system for monocular, stereo, and rgb-d cameras. IEEE Trans. Robot. 2017, 33, 1255–1262. [Google Scholar] [CrossRef]
  22. Liang, R.; Yuan, J.; Kuang, B.; Liu, Q.; Guo, Z. DIG-SLAM: An accurate RGB-D SLAM based on instance segmentation and geometric clustering for dynamic indoor scenes. Meas. Sci. Technol. 2023, 35, 015401. [Google Scholar] [CrossRef]
  23. Tourani, A.; Bavle, H.; Sanchez-Lopez, J.L.; Voos, H. Visual slam: What are the current trends and what to expect? Sensors 2022, 22, 9297. [Google Scholar] [CrossRef] [PubMed]
  24. Chen, W.; Shang, G.; Ji, A.; Zhou, C.; Wang, X.; Xu, C.; Li, Z.; Hu, K. An overview on visual slam: From tradition to semantic. Remote Sens. 2022, 14, 3010. [Google Scholar] [CrossRef]
  25. Li, B.; Zou, D.; Huang, Y.; Niu, X.; Pei, L.; Yu, W. TextSLAM: Visual SLAM with Semantic Planar Text Features. IEEE Trans. Patt. Anal. Mach. Intell. 2023, 46, 593–610. [Google Scholar] [CrossRef] [PubMed]
  26. Qian, Z.; Fu, J.; Xiao, J. Towards accurate loop closure detection in semantic SLAM with 3D semantic covisibility graphs. IEEE Robot. Autom. Lett. 2022, 7, 2455–2462. [Google Scholar] [CrossRef]
  27. Qin, T.; Chen, T.; Chen, Y.; Su, Q. Avp-slam: Semantic visual mapping and localization for autonomous vehicles in the parking lot. In Proceedings of the 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Las Vegas, NV, USA, 25–29 October 2020; IEEE: New York, NY, USA, 2020; pp. 5939–5945. [Google Scholar]
  28. Kang, J.; Nam, C. A Measure of Semantic Class Difference of Point Reprojection Pairs in Camera Pose Estimation. IEEE Trans. Ind. Inform. 2023, 20, 201–212. [Google Scholar] [CrossRef]
  29. Kobyshev, N.; Riemenschneider, H.; Van Gool, L. Matching features correctly through semantic understanding. In Proceedings of the 2014 2nd International Conference on 3D Vision, Tokyo, Japan, 8–11 December 2014; IEEE: New York, NY, USA, 2014; Volume 1, pp. 472–479. [Google Scholar]
  30. Shao, C.; Zhang, C.; Fang, Z.; Yang, G. A deep learning-based semantic filter for RANSAC-based fundamental matrix calculation and the ORB-SLAM system. IEEE Access 2019, 8, 3212–3223. [Google Scholar] [CrossRef]
  31. Geiger, A.; Lenz, P.; Urtasun, R. Are we ready for autonomous driving? the kitti vision benchmark suite. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; IEEE: New York, NY, USA, 2012; pp. 3354–3361. [Google Scholar]
  32. Shu, F.; Wang, J.; Pagani, A.; Stricker, D. Structure plp-slam: Efficient sparse mapping and localization using point, line and plane for monocular, rgb-d and stereo cameras. In Proceedings of the 2023 IEEE International Conference on Robotics and Automation (ICRA), London, UK, 29 May–2 June 2023; IEEE: New York, NY, USA, 2023; pp. 2105–2112. [Google Scholar]
  33. Li, Y.; Brasch, N.; Wang, Y.; Navab, N.; Tombari, F. Structure-slam: Low-drift monocular slam in indoor environments. IEEE Robot. Autom. Lett. 2020, 5, 6583–6590. [Google Scholar] [CrossRef]
  34. Gao, X.; Wang, R.; Demmel, N.; Cremers, D. LDSO: Direct sparse odometry with loop closure. In Proceedings of the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 1–5 October 2018; IEEE: New York, NY, USA, 2018; pp. 2198–2204. [Google Scholar]
Figure 1. ORB-SLAM3 framework with the proposed semantic filter module.
Figure 1. ORB-SLAM3 framework with the proposed semantic filter module.
Electronics 13 04487 g001
Figure 2. Framework of the proposed semantic filter approach.
Figure 2. Framework of the proposed semantic filter approach.
Electronics 13 04487 g002
Figure 3. Computation of GGCM-E features.
Figure 3. Computation of GGCM-E features.
Electronics 13 04487 g003
Figure 4. Semantic filtering on the KITTI frame.
Figure 4. Semantic filtering on the KITTI frame.
Electronics 13 04487 g004
Figure 5. Semantic filtering on our captured outdoor frame.
Figure 5. Semantic filtering on our captured outdoor frame.
Electronics 13 04487 g005
Figure 6. The trajectory of KITTI07 with respect to the ground truth using GGCM-E semantic filter.
Figure 6. The trajectory of KITTI07 with respect to the ground truth using GGCM-E semantic filter.
Electronics 13 04487 g006
Figure 7. Comparison of trajectories between the proposed method and ground truth in the KITTI dataset.
Figure 7. Comparison of trajectories between the proposed method and ground truth in the KITTI dataset.
Electronics 13 04487 g007
Figure 8. Comparison on APEs with respect to ground truth of the ORB-SLAM3 and the semantic filter.
Figure 8. Comparison on APEs with respect to ground truth of the ORB-SLAM3 and the semantic filter.
Electronics 13 04487 g008
Figure 9. Dense texture-rich sequences in TUM dataset (DTR sequences).
Figure 9. Dense texture-rich sequences in TUM dataset (DTR sequences).
Electronics 13 04487 g009
Figure 10. Comparison of camera trajectories in DTR sequences.
Figure 10. Comparison of camera trajectories in DTR sequences.
Electronics 13 04487 g010
Figure 11. Comparison of the trajectory with respect to the ground truth of DynaSLAM and GGCM-E+DynaSLAM on KITTI00 sequences.
Figure 11. Comparison of the trajectory with respect to the ground truth of DynaSLAM and GGCM-E+DynaSLAM on KITTI00 sequences.
Electronics 13 04487 g011
Figure 12. Comparison of the APEs of semantic filter-based Structure-SLAM, LDSO and DynaSLAM on KITTI07 sequences.
Figure 12. Comparison of the APEs of semantic filter-based Structure-SLAM, LDSO and DynaSLAM on KITTI07 sequences.
Electronics 13 04487 g012
Table 1. Comparison of YOLOV8 semantic filters based on different parameters l and different classification methods.
Table 1. Comparison of YOLOV8 semantic filters based on different parameters l and different classification methods.
lClassesThPRMAP@0.5
990.2551.6%37.8%38.2%
50.2563.4%72.0%70.2%
16160.2557.5%45.9%49.1%
60.2072.8%69.0%75.6%
60.2577.6%76.4%82.4%
60.3073.9%71.3%78.2%
25250.2553.7%46.9%40.1%
70.2575.5%72.9%78.3%
Table 2. YOLOv8-based detection results between different semantic filters.
Table 2. YOLOv8-based detection results between different semantic filters.
GGCM-E Feature (Label)RatioClassPRMAP@0.5
1.2.31.4%167.7%83.6%82.0%
4.5.68.9%283.0%82.2%88.7%
7.8.97.3%379.1%78.1%85.5%
10.11.1216.4%479.8%76.7%82.4%
13.14.1536.9%578.2%69.7%78.3%
1628.8%677.4%67.9%77.6%
Table 3. Comparison of ATEs of KITTI07 for ORB-SLAM3 and our proposed method.
Table 3. Comparison of ATEs of KITTI07 for ORB-SLAM3 and our proposed method.
KITTI07RMSE/mStd/mMax/mMin/m
ORB-SLAM32.581.115.420.46
One-class filter + ORB-SLAM32.591.196.130.64
Two-class filter + ORB-SLAM33.011.224.970.17
Three-class filter + ORB-SLAM32.891.344.880.69
Four-class filter + ORB-SLAM32.751.035.520.82
Five-class filter + ORB-SLAM32.250.824.300.15
Six-class filter + ORB-SLAM32.121.053.650.08
All-class filter + ORB-SLAM33.040.946.230.32
Table 4. Comparison of ATEs of KITTI05 for ORB-SLAM3 and our proposed method.
Table 4. Comparison of ATEs of KITTI05 for ORB-SLAM3 and our proposed method.
KITTI05RMSE/mStd/mMax/mMin/m
ORB-SLAM37.242.8314.520.42
One-class filter + ORB-SLAM37.292.8215.191.30
Two-class filter + ORB-SLAM37.993.2015.960.45
Three-class filter + ORB-SLAM37.612.8414.380.50
Four-class filter + ORB-SLAM37.142.4813.180.75
Five-class filter + ORB-SLAM36.392.5913.860.41
Six-class filter + ORB-SLAM36.252.7914.230.15
All-class filter + ORB-SLAM36.662.8915.470.27
Table 5. Comparison of the RMSE with respect to the ATEs of different vSLAM methods—KITTI sequences (m).
Table 5. Comparison of the RMSE with respect to the ATEs of different vSLAM methods—KITTI sequences (m).
 00020506070809
ORB-SLAM38.2827.617.2415.332.5858.357.82
Inliers-Outliers+ORB-SLAM37.8626.236.8816.092.3257.557.58
GGCM-E+ORB-SLAM38.1525.136.2513.672.1357.127.79
Improvement *1.57%8.98%13.7%4.6%17.44%2.11%0.38%
* Improvement in the ATEs of “GGCM-E+ORB-SLAM3” with the original ORB-SLAM3.
Table 6. Comparison of the RMSE with respect to the ATEs of different vSLAM methods—dense texture-rich sequences in TUM dataset (cm).
Table 6. Comparison of the RMSE with respect to the ATEs of different vSLAM methods—dense texture-rich sequences in TUM dataset (cm).
 ORB-SLAM3PLP-SLAMGGCM-E+ORB-SLAM3Improvement *
fre1_desk2.182.061.9311.46%
fre1_plant1.631.551.601.84%
fre1_room8.037.687.269.58%
fre2_desk0.860.900.851.16%
fre2_360_hemisphere16.5615.3414.6611.47%
fre2_large_with_loop19.9719.3918.606.86%
fre3_long_office _household0.931.010.894.30%
* Improvement in the ATEs of “GGCM-E+ORB-SLAM3” with the original ORB-SLAM3.
Table 7. Comparison of the RMSE with respect to the ATEs of different vSLAM methods—KITTI sequences (m).
Table 7. Comparison of the RMSE with respect to the ATEs of different vSLAM methods—KITTI sequences (m).
 00020506070809
Structure-SLAM6.6325.6913.6223.88.2.99109.3813.40
GGCM-E+Structure-SLAM6.3624.7511.8420.572.75103.9112.06
LDSO9.5634.785.5013.672.64134.4821.15
GGCM-E+LDSO9.3532.625.7313.842.59131.4420.09
DynaSLAM9.6429.757.0415.792.8654.056.98
GGCM-E+DynaSLAM8.5025.916.6614.482.4652.836.62
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, Y.; Shao, C.; Wang, J. A GGCM-E Based Semantic Filter and Its Application in VSLAM Systems. Electronics 2024, 13, 4487. https://doi.org/10.3390/electronics13224487

AMA Style

Li Y, Shao C, Wang J. A GGCM-E Based Semantic Filter and Its Application in VSLAM Systems. Electronics. 2024; 13(22):4487. https://doi.org/10.3390/electronics13224487

Chicago/Turabian Style

Li, Yuanjie, Chunyan Shao, and Jiaming Wang. 2024. "A GGCM-E Based Semantic Filter and Its Application in VSLAM Systems" Electronics 13, no. 22: 4487. https://doi.org/10.3390/electronics13224487

APA Style

Li, Y., Shao, C., & Wang, J. (2024). A GGCM-E Based Semantic Filter and Its Application in VSLAM Systems. Electronics, 13(22), 4487. https://doi.org/10.3390/electronics13224487

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop