Next Article in Journal
Unleashing the Potential of Residual and Dual-Stream Transformers for the Remote Sensing Image Analysis
Previous Article in Journal
A Multitask Network for the Diagnosis of Autoimmune Gastritis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Beyond Handcrafted Features: A Deep Learning Framework for Optical Flow and SLAM

by
Kamran Kazi
1,2,
Arbab Nighat Kalhoro
1,2,
Farida Memon
2,
Azam Rafique Memon
2,3,* and
Muddesar Iqbal
3,*
1
Institute of Information and Communication Technologies, Mehran University of Engineering and Technology, Jamshoro 76062, Pakistan
2
Department of Electronic Engineering, Mehran University of Engineering and Technology, Jamshoro 76062, Pakistan
3
Renewable Energy Laboratory, College of Engineering, Prince Sultan University, Riyadh 11586, Saudi Arabia
*
Authors to whom correspondence should be addressed.
J. Imaging 2025, 11(5), 155; https://doi.org/10.3390/jimaging11050155
Submission received: 9 April 2025 / Revised: 7 May 2025 / Accepted: 12 May 2025 / Published: 15 May 2025
(This article belongs to the Section Visualization and Computer Graphics)

Abstract

:
This paper presents a novel approach for visual Simultaneous Localization and Mapping (SLAM) using Convolution Neural Networks (CNNs) for robust map creation. Traditional SLAM methods rely on handcrafted features, which are susceptible to viewpoint changes, occlusions, and illumination variations. This work proposes a method that leverages the power of CNNs by extracting features from an intermediate layer of a pre-trained model for optical flow estimation. We conduct an extensive search for optimal features by analyzing the offset error across thousands of combinations of layers and filters within the CNN. This analysis reveals a specific layer and filter combination that exhibits minimal offset error while still accounting for viewpoint changes, occlusions, and illumination variations. These features, learned by the CNN, are demonstrably robust to environmental challenges that often hinder traditional handcrafted features in SLAM tasks. The proposed method is evaluated on six publicly available datasets that are widely used for bench-marking map estimation and accuracy. Our method consistently achieved the lowest offset error compared to traditional handcrafted feature-based approaches on all six datasets. This demonstrates the effectiveness of CNN-derived features for building accurate and robust maps in diverse environments.

1. Introduction

Simultaneous Localization and Mapping (SLAM) is a critical task in robotics [1,2] and computer vision, enabling autonomous systems to navigate unknown environments by simultaneously estimating the robot’s trajectory and building a map [3]. Traditionally, SLAM relies heavily on handcrafted features—such as Scale-Invariant Feature Transform (SIFT), Speeded Up Robust Features (SURF), and Oriented FAST and Rotated BRIEF (ORB)—to extract visual cues from the environment. While these methods have been effective, they exhibit significant limitations under challenging conditions, such as changes in viewpoint, occlusions, and varying illumination conditions.
To overcome these limitations, deep learning techniques, particularly Convolutional Neural Networks (CNNs), offer a promising alternative by learning robust feature representations directly from data. In this paper, we propose a novel deep learning-based approach to SLAM that harnesses the power of CNNs to improve the robustness and accuracy of map estimation. Specifically, we extract features from intermediate layers of three widely used CNN architectures—ConvNeXtXLarge, EfficientNetV2L, and NASNetLarge—and evaluate their effectiveness in reducing odometry error compared to traditional methods.
Our approach involves estimating the odometry error [4] across various layers and filters within these CNNs to identify the feature combination that minimizes the error. We then compare the results with odometry errors calculated using ORB features, which are commonly used in ORB-SLAM, one of the most well-known and widely used SLAM algorithms. The filter combination with the lowest odometry error is subsequently applied to estimate the map of the environment, following the key steps outlined in the ORB-SLAM framework.
The other contributions of the proposed methods include the following:
  • This research uses CNN features to minimize reprojection error in estimating trajectory maps.
  • This gives hope of obtaining minimal reprojection errors without even optimizing the map; hence, the needfor loop closure detection and map optimization can be eliminated in the future.
  • This paper explores the ConvNeXtXLarge, EfficientNetV2L, and NASNetLarge models, performing a detailed layer-wise and filter-wise analysis to identify features that yield minimal odometry error in SLAM tasks.
  • The CNN-based features show improved robustness to common environmental challenges in SLAM, such as viewpoint changes, occlusions, and illumination variations, which often degrade the performance of traditional handcrafted features.
  • The use of CNN features in place of handcrafted features results in a significant improvement by reducing odometry error, which suggests that CNN features can be used to estimate efficient trajectory maps without the need for map optimization in the future.
The experimental results conducted on six publicly available benchmark datasets demonstrate that our deep learning-based feature extraction method significantly reduces odometry error compared to traditional handcrafted features, particularly ORB. This reduction in error not only improves the accuracy of map estimation [5] but also enhances the system’s robustness in the face of environmental challenges such as occlusions, dynamic objects, and changing lighting conditions. Our proposed method represents a step forward in the integration of deep learning [6] with SLAM, offering a more reliable solution for autonomous navigation in diverse real-world environments.
The remainder of this paper is organized as follows: In Section 2, existing work related to the proposed topic is discussed. Section 3 elaborates on the proposed method. It details the analysis of various layers and their underlying filters and methods, information which is used to find the odometry error. Subsequently, we present a comparative analysis with existing SLAM models to highlight the reduction in odometry error achieved by our method in Section 4. Furthermore, we evaluate the proposed approach on six publicly available datasets, encompassing diverse environments and challenges. The extensive evaluation demonstrates the robustness and effectiveness of the proposed approach in real-world scenarios. Finally, we conclude by summarizing the key findings and discussing potential avenues for future research in Section 5.

2. Related Work

Simultaneous Localization and Mapping (SLAM) is a crucial task in robotics, enabling robots to build a map [1,7,8] of their environment while simultaneously estimating their location within that map. Traditionally, SLAM has relied on handcrafted features like SIFT or ORB for feature extraction. However, these features have limitations in handling viewpoint changes, illumination variations, and occlusions, which can significantly impact the accuracy and robustness of the generated map.
Recent research has explored the potential of leveraging Convolution Neural Networks (CNNs) for feature extraction in SLAM. CNNs, trained on massive image datasets, learn high-level features that are often more invariant to these challenges compared to handcrafted features. This literature review examines existing research on trajectory mapping with a focus on approaches that utilize features extracted from intermediate layers of CNNs.
Several studies have demonstrated the potential benefits of CNN features for SLAM. The authors of [9] introduced DXSLAM, which utilizes a combination of local and global CNN features to improve map consistency. Similarly, the authors of [10] presented VSO, which integrates semantic constraints obtained from CNNs for enhanced pose and map optimization. These works highlight the ability of CNN features to capture richer and more robust information compared to traditional handcrafted methods.
Lin et al. [11] developed an approach based on object-level semantic mapping and topology-aware features. Their system uses semantic object detection [12,13] and topological relationships to improve loop closure detection [7,14,15]. The topological features improve robustness by abstracting the environment at a higher level than pixel-based or handcrafted feature methods, making it less sensitive to local environmental changes. Chen et al. [16] proposed SuMa++, a SLAM system that incorporates LiDAR-based semantic features for more accurate map building and localization. They use deep learning-based semantic segmentation to extract meaningful features from the environment, which improves the performance of traditional LiDAR-based SLAM. Sualeh and Kim [17] provided a comprehensive survey of semantic SLAM approaches. Their review covers a range of feature extraction methods, from handcrafted features like ORB and SIFT to more recent deep learning-based features. They highlight how the shift towards semantic features has improved loop closure detection and place recognition, especially in dynamic and cluttered environments. Mahmoud and Atia [18] proposed a method that combines semantic segmentation with layout estimation to enhance visual SLAM. Their use of CNN-based semantic features allows the system to better understand the scene, particularly in environments with occlusions or varying lighting conditions. This improves the performance of the visual SLAM system, particularly in challenging indoor environments.
Despite the promising results, challenges remain in utilizing CNN features for SLAM. One key challenge is the computational cost associated with extracting features from CNNs. Compared to handcrafted methods, this process can be computationally expensive, potentially impacting real-time performance, especially on resource-constrained platforms [19]. Additionally, the “black-box” nature of CNNs presents another challenge. Understanding how specific intermediate layers contribute to the robustness of feature extraction remains an open question. Furthermore, selecting layers that provide robust features that are reliable when facing environmental changes such as different viewpoints, different times of the day, and different lighting and weather conditions is a real challenge.
This research addresses these limitations by proposing a novel approach for trajectory mapping. We utilize features extracted from an intermediate layer of a pre-trained CNN for optical flow estimation. Our approach involves an extensive search to identify the optimal layer and filter combination that minimizes offset error while maintaining robustness to viewpoint changes, occlusions, and illumination variations. This strategy aims to achieve accurate and robust map creation in a single session, paving the way for improved SLAM performance in diverse and challenging environments.
Many researchers [20] have created their algorithms based on ORB SLAM [21,22] approaches. The authors of [20] discussed an approach to reducing map points by discarding redundant map points to optimize SLAM. They modified the existing ORB SLAM 2 method and optimized the method for resource-limited hardware platforms.
Various authors [23,24,25] have used landmarks instead of features. According [23], the landmark-based approaches are better for dynamic environments. The authors of [25] made use of detected objects for camera pose recovery in SLAM, with augmented reality being identified as an important application. Their main contributions included the improvement of the re-localization ability of the SLAM system using object landmarks at a higher level. Their system is also able to track and reconstruct objects in 3D by making use of ellipsoids. Experimentation was performed for objects as well as point-based strategies.
A visual SLAM algorithm is presented in [26] that utilizes multi-map fusion to enable fast relocation or local map construction in a scenario where current position and tracking are lost. Fusion of all previous maps is performed during loop back. Moreover, for location recognition, a modified version of the bag of visual words method is used to ensure loop errors are avoided as much as possible, followed by robust map fusion. They did not use bag-of-words-based methods, which are not efficient in complex environments. These cannot match scenes from different viewpoints and in different illumination conditions.
The authors of [27] proposed a robust long-term robotic mapping system using ground segmentation, outlier-robust registration, hierarchical multi-session SLAM, and instance-aware static map building to handle dynamic environments.

3. Methodology

The major task of the proposed research in this paper is to find features that are robust, consistent, and reliable in all or in most of natural conditions. However, considering all natural conditions is nearly ideal and not possible in realistic scenario. Almost all other researchers target only a single environmental issue, keeping other natural conditions constant, which makes their algorithms suitable for specific natural conditions, leading to failures in their methods when environmental conditions change.
The proposed method in this paper handles multiple natural conditions like illumination changes, appearance changes, viewpoint changes, and the presence of dynamic objects. The proposed method has an estimated offset error from thousands of filters from various layers of a CNN network and incorporates a layer that can extract features that are much less disturbed by environmental changes and appearance changes. This layer is then used to create a map of the environment.

3.1. Selection of Pre-Trained CNN Models for Feature Extraction

Our proposed approach utilizes features extracted from an intermediate layer of a pre-trained Convolution Neural Network (CNN) for optical flow estimation. This is in contrast to traditional SLAM methods that rely on handcrafted features like SIFT or ORB. While handcrafted features have been widely used, they exhibit limitations in handling variations in the environment. For instance, viewpoint changes can significantly alter handcrafted features, leading to inaccurate map creation [28]. Illumination variations and occlusions further hinder these methods [29].
In contrast, CNNs, trained on massive image datasets, learn high-level features that are demonstrably more robust to these challenges [9]. These learned features encode richer information about the scene, leading to improved map consistency [9] and more accurate pose and map optimization [10]. By leveraging CNN features, we aim to achieve robust and accurate trajectory mapping even in environments with viewpoint changes, illumination variations, and occlusions.
To exploit the potential of CNN features, we have to use the features extracted from the CNN to find the optical flow and create a map of the environment. Now, the real challenge is which filter from which layer is most useful and can be reliable in case of scene appearance changes or in case of continuously changing natural conditions. So, to find the answer to this question, we extracted features from multiple filters (i.e., kernels) from various layers of three popular CNN models that are pre-trained on the Imagenet dataset and that can be customized using keras/tensorflow. Out of 38 available CNN models, we have extracted features from three models with unique architectures that have superior Top-1 accuracy. The models that we have picked and their parameters are given in Table 1. The filters that were not used include layers such as the Max Pooling layer, Normalization Layer, Flatten Layer, and other non-convolution layers.

3.2. Analysing Filters and Estimating Odometry Error

The convolution layers of each model are analyzed and features from each filter of each convolution layer are extracted to find the odometry error, and the filter with the minimum odometry error is selected for creating the estimated map of the environment. Figure 1 shows the input image in (a), whereas (b), (c), and (d) represent images taken from intermediate layers, resized to the same dimensions as the input to visualise the impact through different layers.
A total of 313,746 filters were analyzed from the three chosen models and compared with the most popular ORB features that are used by many SLAM algorithms for trajectory estimation.
E t = | | T c T p | |
E r = c o s 1 ( t r a c ( R c · R p ) 1 2 )
where T c and T p represent the current translation vector and previous translation vector, respectively. Similarly, R c and R p represent the current and previous rotation matrices. Symbol | | represents the Euclidean distance between two vector points and t r a c represents the sum of diagonals of matrices.
We estimated the odometry errors from 313,746 filters of the three models; some of the filters and their errors are mentioned in Table 2.
The most suitable filter was filter #10 from the convnext_xlarge_stemlayer of the ConvNeXtXLarge model.

3.3. Datasets Used for Testing the Proposed Method

The proposed method used six publicly available datasets that are mostly used for validating trajectory mappings of environments obtained using the structure of motion and by estimating optical flow. These datasets include the following:
  • Kitti Sequence 00 [30].
  • Kitti Sequence 02 [30].
  • Kitti Sequence 05 [30].
  • Kitti Sequence 06 [30].
  • Kitti Sequence 08 [30].
  • Kitti Sequence 09 [30].
A glimpse of these datasets is shown in Figure 2 and the details are given below:

Kitti Sequence Datasets

The Kitti Sequence Datasets [30] contain 22 stereo sequences, out of which 11 sequences have ground truth available. The sequences were taken in the afternoon time and had various lighting variations and shade levels. Some of the sequences have occlusions, slight viewpoint changes, illumination changes, and the presence of road traffic. The Kitti sequences 00, 02, 05, 06, 08, and 09 are used to validate the proposed approach in this paper, as these sequences have loops, which help in optimizing the maps. More details on the datasets are given in Table 3.

3.4. Estimating Transformation Matrix from CNN Features

To create the environment map, it is essential to find the optical flow. The optical flow can determine how a scene is moving. The transformation matrix determines the scene movement. Assume I t is the most recent image frame at current time t and I t 1 is the previous image frame acquired at previous time t 1 . The features acquired from convnext_xlarge_stem layer # 2 and filter # 10 as F I t from image frame I t are compared with features acquired from the same layer and filter F I t 1 from previous image frame I t 1 .

3.5. Plotting the Map of the Trajectory

In visual SLAM using CNNs, trajectory mapping is essential for understanding the path estimated through frames captured by the camera, facilitating accurate map creation. This section outlines the process of generating local trajectory maps from sequences of frames, leveraging CNN-derived features to enhance robustness against viewpoint changes, occlusions, and illumination variations. Each sequence contributes to a cohesive trajectory map, refined through feature extraction from intermediate layers of a pre-trained CNN model.
To compute the full trajectory, it is necessary to determine the relative transformation T k between consecutive image frames I t and I t 1 , where t is any time > 0 . These relative transformations are concatenated to reconstruct the complete path incrementally, pose by pose, as the camera moves through the environment. At regular intervals (every m poses), the trajectory is refined to provide a more accurate estimate of the local path. This refinement is achieved by minimizing the offset error using CNN-derived features, enhancing stability in challenging environments. In this approach, the extracted CNN features serve as the basis for robust trajectory mapping, enabling a more consistent and optimized map across a range of viewpoints and lighting conditions.
To plot the trajectory map, optical flow between consecutive frames is calculated using CNN-extracted features. When the initial pose P o s i n i t is unknown, robust CNN-derived features from frames F I t and F I t 1 are employed to estimate both the Homography (Equation (3)) H I t I t 1 and the Fundamental matrix (Equation (4)) F I t I t 1 . This process establishes the initial pose, enabling trajectory mapping that remains accurate and consistent as the mapping data accumulate. Through the CNN’s learned features, this approach provides a solid foundation for further enhancements in map optimization and adaptability to dynamic environments, setting a new standard for accuracy and robustness in SLAM applications.
I t = H I t I t 1 I t 1
I t 1 T F I t I t 1 I t 1 = 0
To find out which matrix to use for initial pose estimation [31], we need to find score S M , as also carried out by [21].
S M = ( ρ M ( d I t I t 1 2 ( I I t i , I I t 1 i , M ) ) + ρ M ( d I t 1 I t 2 ( I I t i , I I t 1 i , M ) )
where M is either H for homography or F for fundamental matrix. d I t I t 1 2 and d I t 1 I t 2 are the symmetric transfer errors [21,32] from one frame to another frame.
ρ M ( d 2 ) = Γ d 2 if d 2 < T M 0 if d 2 T M
where T M is the outlier rejection threshold based on the χ 2 test. Γ is equal to T H , so that both models score equally for the same d in their inlier region, again to make the process homogeneous [21].
To find which model to use, we use the following [21]:
R H = S H S H + S F
When R H > 0.40 , the homography matrix is considered to take the lead, which adequately captures the planar and low-parallax cases. Otherwise, the fundamental matrix takes the lead.

3.5.1. Estimating Transformation Matrix from Homography

In the case of the homography matrix, eight motion hypotheses using the method of Faugeras et al. [33] are discussed. Their method proposes various tests to find a valid solution. To find the valid solution, we need to find out whether the feature points found in the image frames at time t and t 1 are in front of the camera or if some points go behind the camera in a recent frame at t.
Z = ( R 1 r F g I t 1 × R 3 r ) · T ( R 1 r F g I t 1 × R 3 r ) · F g I t 1
P 1 3 d = ( F g I t × Z , F g I t 1 × Z , Z )
P 2 3 d = ( R T · P 1 3 d R T · T )
The n r above any matrix refers to the n-th row of that matrix. P 1 3 d and P 2 3 d represent 3d points; if any of these is negative, the points are not in front of the camera in both frames. R and T represent the rotation and translation matrix estimated by decomposing the homography matrix.
The initial value is always the current value R c = R and T c = T .
T c = T c + s c a l e R f · T if T z > T x & T z > T y T c = T c o t h e r w i s e
R c = R · R c if T z > T x & T z > T y R c = R c o t h e r w i s e
T c x , T c y , and T c z represent the x, y, and z positions in the estimated map, respectively. For a 2D map, the z-axis can be ignored.

3.5.2. Estimating Transformation Matrix from Fundamental Matrix

When the intrinsic parameters of the camera in both image frames at time t and t 1 are known, the essential matrix E can be determined from the fundamental matrix F using
E = K T F K
where K is the camera calibration matrix with the following intrinsic camera parameters:
K = f x 0 c x 0 f y c y 0 0 0
where f x and f y are the focal length of the camera and c x and c y are the center points of the camera.
The essential matrix E is then decomposed to obtain singular values w, left singular values u, and right singular values v t .

3.6. Generating the Trajectory Map

To create the estimated map, based on Equation (7), either the fundamental matrix F or homography matrix H will be used. In the case of the homography matrix, the transformation matrices R and T can be obtained from Equation (12). But in the case where the fundamental matrix is chosen by Equation (7), the essential matrix is obtained and then the essential matrix is decomposed and values w, u, and v t are obtained. These values will help in finding the transformation matrices R and T using following equations:
Using the Singular Value Decomposition (SVD) method, the essential matrix E is decomposed into singular values and rotation and translation vectors are obtained.
The rotation matrix can be calculated as R = u × w × v t and the translation matrix as T = u . To calculate the updated translation and rotation on upcoming frames at t + 1 , Equations (11) and (12) can be used, respectively.

3.7. When Initial Pose Is Already Detected

When the initial pose is already detected, we only need to find the essential matrix from features of the current image frame I t and features of the previous image frame I t 1 using f I t E f I t 1 = 0 . The rotation and translation matrices are then determined using the method already discussed in Section 3.6.

4. Experimental Results

We now determine the accuracy of the estimated map generated using features obtained from convnext_xlarge_stem layer # 2 and filter # 10. The maps are estimated on six datasets and the estimated trajectories are shown in Figure 3. The proposed method is also compared with two other methods, which also allow for estimating the offset odometry error. The results are listed in Table 4. The estimated trajectory maps of all datasets using Deep-VO [34] are given in Figure 4 and estimated maps using TartanVO [35] are given in Figure 5. The proposed system was tested on an i7-core laptop with 16 GB RAM and it can process from 7 to 13 frames per second.

Offset Error Estimation on Datasets

To estimate the offset error between the ground truth and the estimated map using proposed method, the following equation is used:
E x = | | P x g P x e | |
E y = | | P y g P y e | |
The offset error in X a x i s is E x and the offset error in y a x i s is E y , which is the sum of all individual frames. Meanwhile, P x g and P y g are the ground truth positions of the current frame in the x a x i s and y a x i s , respectively, and P x e and P y e are the estimated positions of the current frame in the x a x i s and y a x i s , respectively.
a v e r a g e o f f s e t e r r o r = E x + E y N f
where N f is the total number of frames.

5. Conclusions

The proposed research introduces a pioneering approach to feature selection in SLAM by conducting an in-depth analysis of layers and filters across three state-of-the-art CNN architectures. This systematic exploration identifies features that exhibit unprecedented robustness to environmental variations—a capability that has not been fully realized in prior SLAM methodologies. Unlike existing approaches, our method achieves significantly lower offset error without requiring map optimization, demonstrating a fundamental shift in how SLAM systems can achieve higher accuracy and reliability. By addressing a long-standing challenge in the field, this work sets a new benchmark for integrating deep learning into SLAM, paving the way for more resilient and precise trajectory mapping in complex environments.

Author Contributions

Conceptualization, K.K. and A.N.K.; methodology, K.K. and A.R.M.; data collection, K.K.; validation, K.K., F.M., A.R.M., and M.I.; formal analysis, K.K.; writing—original draft preparation, K.K.; writing—review and editing, A.N.K., F.M., and A.R.M.; visualization, A.R.M. and M.I.; supervision, A.N.K., F.M., A.R.M., and M.I. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by research grants [PID-000085_01_03: Unmanned Ground Vehicles for Disaster Management and Recovery] from the PSDSARC at Prince Sultan University, Saudi Arabia.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data used in this paper are already publicly available.

Acknowledgments

The authors would like thank the Institute of Information & Communication Technology, Mehran University of Engineering & Technology, Jamshoro, Sindh, Pakistan, as well as acknowledge the support of Prince Sultan University, who paid the Article Processing Charges (APCs) for this publication.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Athira, K.A.; Divya Udayan, J.U.S. A Systematic Literature Review on Multi-Robot Task Allocation. ACM Comput. Surv. 2024, 57, 1–28. [Google Scholar] [CrossRef]
  2. Haider, Z.; Sardar, M.Z.; Azar, A.T.; Ahmed, S.; Kamal, N.A. Exploring reinforcement learning techniques in the realm of mobile robotics. Int. J. Autom. Control 2024, 18, 655–697. [Google Scholar] [CrossRef]
  3. Azar, A.T.; Sardar, M.Z.; Ahmed, S.; Hassanien, A.E.; Kamal, N.A. Autonomous Robot Navigation and Exploration Using Deep Reinforcement Learning with Gazebo and ROS. Lect. Notes Data Eng. Commun. Technol. 2023, 2023, 287–299. [Google Scholar] [CrossRef]
  4. Liu, J.; Gao, W.; Xie, C.; Hu, Z. Implementation and observability analysis of visual-inertial-wheel odometry with robust initialization and online extrinsic calibration. Robot. Auton. Syst. 2024, 176, 104686. [Google Scholar] [CrossRef]
  5. Junaedy, A.; Masuta, H.; Sawai, K.; Motoyoshi, T.; Takagi, N. Real-Time 3D Map Building in a Mobile Robot System with Low-Bandwidth Communication. Robotics 2023, 12, 157. [Google Scholar] [CrossRef]
  6. Petrakis, G.; Partsinevelos, P. Keypoint Detection and Description Through Deep Learning in Unstructured Environments. Robotics 2023, 12, 137. [Google Scholar] [CrossRef]
  7. Memon, A.R.; Iqbal, M.; Almakhles, D. DisView: A Semantic Visual IoT Mixed Data Feature Extractor for Enhanced Loop Closure Detection for UGVs During Rescue Operations. IEEE Internet Things J. 2024, 11, 36214–36224. [Google Scholar] [CrossRef]
  8. Iqbal, M.; Memon, A.R.; Almakhles, D.J. Accelerating Resource-Constrained Swarm Robotics with Cone-Based Loop Closure and 6G Communication. IEEE Trans. Intell. Transp. Syst. 2025; early access. [Google Scholar] [CrossRef]
  9. Li, D.; Shi, X.; Long, Q.; Liu, S.; Yang, W.; Wang, F.; Wei, Q.; Qiao, F. DXSLAM: A Robust and Efficient Visual SLAM System with Deep Features. In Proceedings of the 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Las Vegas, NV, USA, 24 October–24 January 2021; pp. 4958–4965. [Google Scholar]
  10. Lianos, K.N.; Schönberger, J.L.; Pollefeys, M.; Sattler, T. VSO: Visual Semantic Odometry; Springer: Berlin/Heidelberg, Germany, 2018; pp. 246–263. [Google Scholar]
  11. Lin, S.; Wang, J.; Xu, M.; Zhao, H.; Chen, Z. Topology Aware Object-Level Semantic Mapping Towards More Robust Loop Closure. IEEE Robot. Autom. Lett. 2021, 6, 7041–7048. [Google Scholar] [CrossRef]
  12. Waheed, S.R.; Suaib, N.M.; Rahim, M.S.M.; Khan, A.R.; Bahaj, S.A.; Saba, T. Synergistic Integration of Transfer Learning and Deep Learning for Enhanced Object Detection in Digital Images. IEEE Access 2024, 12, 13525–13536. [Google Scholar] [CrossRef]
  13. Tahir, N.U.A.; Zhang, Z.; Asim, M.; Chen, J.; Elaffendi, M. Object Detection in Autonomous Vehicles under Adverse Weather: A Review of Traditional and Deep Learning Approaches. Algorithms 2024, 17, 103. [Google Scholar] [CrossRef]
  14. Memon, A.R.; Wang, H.; Hussain, A. Loop closure detection using supervised and unsupervised deep neural networks for monocular SLAM systems. Robot. Auton. Syst. 2020, 126, 103470. [Google Scholar] [CrossRef]
  15. Ul Islam, Q.; Khozaei, F.; Salah Al Barhoumi, E.M.; Baig, I.; Ignatyev, D. Advancing Autonomous SLAM Systems: Integrating YOLO Object Detection and Enhanced Loop Closure Techniques for Robust Environment Mapping. Robot. Auton. Syst. 2024, 185, 104871. [Google Scholar] [CrossRef]
  16. Chen, X.; Milioto, A.; Palazzolo, E.; Giguere, P.; Behley, J.; Stachniss, C. SuMa++: Efficient LiDAR-based Semantic SLAM. In Proceedings of the 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Macau, China, 3–8 November 2019; pp. 4530–4537. [Google Scholar] [CrossRef]
  17. Sualeh, M.; Kim, G.W. Simultaneous Localization and Mapping in the Epoch of Semantics: A Survey. Int. J. Control. Autom. Syst. 2019, 17, 729–742. [Google Scholar] [CrossRef]
  18. Mahmoud, A.; Atia, M. Improved Visual SLAM Using Semantic Segmentation and Layout Estimation. Robotics 2022, 11, 91. [Google Scholar] [CrossRef]
  19. Giubilato, R.; Vayugundla, M.; Schuster, M.J.; Sturzl, W.; Wedler, A.; Triebel, R.; Debei, S. Relocalization with Submaps: Multi-Session Mapping for Planetary Rovers Equipped with Stereo Cameras. IEEE Robot. Autom. Lett. 2020, 5, 580–587. [Google Scholar] [CrossRef]
  20. Müller, C.J. Map Point Selection for Hardware Constrained Visual Simultaneous Localisation and Mapping; Technical Report; Stellenbosch University: Stellenbosch, South Africa, 2024. [Google Scholar]
  21. Mur-Artal, R.; Montiel, J.M.M.; Tardos, J.D. ORB-SLAM: A Versatile and Accurate Monocular SLAM System. IEEE Trans. Robot. 2015, 31, 1147–1163. [Google Scholar] [CrossRef]
  22. Mur-Artal, R.; Tardos, J.D. ORB-SLAM2: An Open-Source SLAM System for Monocular, Stereo, and RGB-D Cameras. IEEE Trans. Robot. 2017, 33, 1255–1262. [Google Scholar] [CrossRef]
  23. Memon, A.R.; Liu, Z.; Wang, H. Viewpoint-Invariant Loop Closure Detection Using Step-Wise Learning with Controlling Embeddings of Landmarks. IEEE Trans. Intell. Transp. Syst. 2022, 23, 20148–20159. [Google Scholar] [CrossRef]
  24. Mahattansin, N.; Sukvichai, K.; Bunnun, P.; Isshiki, T. Improving Relocalization in Visual SLAM by using Object Detection. In Proceedings of the 19th International Conference on Electrical Engineering/Electronics, Computer, Telecommunications and Information Technology, ECTI-CON 2022, Prachuap Khiri Khan, Thailand, 24–27 May 2022; pp. 17–20. [Google Scholar] [CrossRef]
  25. Zins, M.; Simon, G.; Berger, M.O. OA-SLAM: Leveraging Objects for Camera Relocalization in Visual SLAM. In Proceedings of the 2022 IEEE International Symposium on Mixed and Augmented Reality (ISMAR), Singapore, 17–21 October 2022; pp. 720–728. [Google Scholar]
  26. Ming, D.; Wu, X. Research on Monocular Vision SLAM Algorithm for Multi-map Fusion and Loop Detection. In Proceedings of the 2022 6th International Conference on Automation, Control and Robots (ICACR), Shanghai, China, 23–25 September 2022. [Google Scholar]
  27. Lim, H. Outlier-robust long-term robotic mapping leveraging ground segmentation. arXiv 2024, arXiv:2405.11176. [Google Scholar]
  28. Milford, M.J.; Wyeth, G.F. SeqSLAM: Visual route-based navigation for sunny summer days and stormy winter nights. In Proceedings of the IEEE International Conference on Robotics and Automation, Saint Paul, MN, USA, 14–18 May 2012; pp. 1643–1649. [Google Scholar]
  29. Xu, Z.; Rong, Z.; Wu, Y. A survey: Which features are required for dynamic visual simultaneous localization and mapping? Vis. Comput. Ind. Biomed. Art 2021, 4, 20. [Google Scholar] [CrossRef]
  30. Geiger, A.; Lenz, P.; Urtasun, R. Are we ready for autonomous driving? The KITTI vision benchmark suite. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 3354–3361. [Google Scholar]
  31. Yu, Z.; Jiang, X.; Liu, Y. Pose estimation of an aerial construction robot based on motion and dynamic constraints. Robot. Auton. Syst. 2024, 172, 104591. [Google Scholar] [CrossRef]
  32. Hartley, R.; Zisserman, A. Multiple View Geometry in Computer Vision, 2nd ed.; Cambridge University Press: Cambridge, UK, 2004. [Google Scholar]
  33. Faugeras, O.; Lustman, F. Motion and Structure From Motion in a Piecewise Planar Environment. Int. J. Pattern Recognit. Artif. Intell. 1988, 2, 485–508. [Google Scholar] [CrossRef]
  34. Wang, S.; Clark, R.; Wen, H.; Trigoni, N. Deepvo: Towards end-to-end visual odometry with deep recurrent convolutional neural networks. In Proceedings of the 2017 IEEE International Conference on Robotics and Automation (ICRA), Singapore, 29 May–3 June 2017; pp. 2043–2050. [Google Scholar]
  35. Wang, W.; Zhu, D.; Wang, X.; Hu, Y.; Qiu, Y.; Wang, C.; Hu, Y.; Kapoor, A.; Scherer, S. TartanAir: A Dataset to Push the Limits of Visual SLAM. In Proceedings of the 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Las Vegas, NV, USA, 24 October–24 January 2021. [Google Scholar]
Figure 1. Features obtained from intermediate layer of ConvNeXtXLarge dataset. (a) Input image from Kitti Sequence 00; (b) image taken from Layer 0, Filter 0 of ConvNeXtXLarge model; (c) image taken from Layer 1, Filter 0 of ConvNeXtXLarge model; (d) image taken from Layer 3, Filter 0 of ConvNeXtXLarge model.
Figure 1. Features obtained from intermediate layer of ConvNeXtXLarge dataset. (a) Input image from Kitti Sequence 00; (b) image taken from Layer 0, Filter 0 of ConvNeXtXLarge model; (c) image taken from Layer 1, Filter 0 of ConvNeXtXLarge model; (d) image taken from Layer 3, Filter 0 of ConvNeXtXLarge model.
Jimaging 11 00155 g001aJimaging 11 00155 g001b
Figure 2. Glimpse of datasets used to validate trajectory mapping. (a) Kitti Sequence 00; (b) Kitti Sequence 02; (c) Kitti Sequence 05; (d) Kitti Sequence 06; (e) Kitti Sequence 08; (f) Kitti Sequence 09.
Figure 2. Glimpse of datasets used to validate trajectory mapping. (a) Kitti Sequence 00; (b) Kitti Sequence 02; (c) Kitti Sequence 05; (d) Kitti Sequence 06; (e) Kitti Sequence 08; (f) Kitti Sequence 09.
Jimaging 11 00155 g002aJimaging 11 00155 g002b
Figure 3. Estimated trajectory maps of Kitti Sequence Datasets without any optimizations. (a) Kitti Sequence 00; (b) Kitti Sequence 02; (c) Kitti Sequence 05; (d) Kitti Sequence 06; (e) Kitti Sequence 08; (f) Kitti Sequence 09.
Figure 3. Estimated trajectory maps of Kitti Sequence Datasets without any optimizations. (a) Kitti Sequence 00; (b) Kitti Sequence 02; (c) Kitti Sequence 05; (d) Kitti Sequence 06; (e) Kitti Sequence 08; (f) Kitti Sequence 09.
Jimaging 11 00155 g003
Figure 4. Estimated trajectory maps of Kitti Sequence Datasets using DeepVO [34]. (a) Kitti Sequence 00; (b) Kitti Sequence 02; (c) Kitti Sequence 05; (d) Kitti Sequence 06; (e) Kitti Sequence 08; (f) Kitti Sequence 09.
Figure 4. Estimated trajectory maps of Kitti Sequence Datasets using DeepVO [34]. (a) Kitti Sequence 00; (b) Kitti Sequence 02; (c) Kitti Sequence 05; (d) Kitti Sequence 06; (e) Kitti Sequence 08; (f) Kitti Sequence 09.
Jimaging 11 00155 g004
Figure 5. Estimated trajectory maps of Kitti Sequence Datasets using TartanVO [35]. (a) Kitti Sequence 00; (b) Kitti Sequence 02; (c) Kitti Sequence 05; (d) Kitti Sequence 06; (e) Kitti Sequence 08; (f) Kitti Sequence 09.
Figure 5. Estimated trajectory maps of Kitti Sequence Datasets using TartanVO [35]. (a) Kitti Sequence 00; (b) Kitti Sequence 02; (c) Kitti Sequence 05; (d) Kitti Sequence 06; (e) Kitti Sequence 08; (f) Kitti Sequence 09.
Jimaging 11 00155 g005
Table 1. Models used to find filter with environment-invariant features.
Table 1. Models used to find filter with environment-invariant features.
ModelAccuracyParametersLayersFiltersUsedNot Used
Top-1Top-5
ConvNeXtXLarge86.7%-350.1 Million297332,80636,096296,710
EfficientNetV2L85.7%97.5%119.0 Million10311,253,254261,216992,038
NASNetLarge82.5%96.0%88.9 Million1041502,40716,434485,973
Table 2. Some of the layer and filter combinations and their corresponding estimated translation and rotation errors.
Table 2. Some of the layer and filter combinations and their corresponding estimated translation and rotation errors.
Average Offset Error
S. #Layer Name–Layer #–Filter #Translation (m)Rotation (deg)
ConvNeXtXLarge Dataset
1Input Layer–00–0024.54127.21
2Input Layer–00–0132.22125.14
3Input Layer–00–0215.01108.37
4convnext_xlarge_stem–2–0Pose not detectedPose not detected
5convnext_xlarge_stem–2–1442.27126.01
6convnext_xlarge_stem–2–2Pose not detectedPose not detected
7convnext_xlarge_stem–2–3Pose not detectedPose not detected
8convnext_xlarge_stem–2–431.6097.35
9convnext_xlarge_stem–2–551.54130.41
10convnext_xlarge_stem–2–63,580,159.43104.76
11convnext_xlarge_stem–2–7Pose not detectedPose not detected
12convnext_xlarge_stem–2 – 82809.92127.59
13convnext_xlarge_stem–2–9Pose not detectedPose not detected
14convnext_xlarge_stem–2–107.2878.02
15convnext_xlarge_stem–2–11545,608.38104.77
16convnext_xlarge_stem–2–12Pose not detectedPose not detected
17convnext_xlarge_stem–2–13660,566.05105.23
18convnext_xlarge_stem–2–14Pose not detectedPose not detected
19convnext_xlarge_stem–2–15955,025.13104.47
20convnext_xlarge_stem–2–164,441,462.76104.42
21convnext_xlarge_stem–2–17Pose not detectedPose not detected
22convnext_xlarge_stem–2–18270,065,492.593.92
23convnext_xlarge_stem–2–19305,075.56106.48
24convnext_xlarge_stem–2–20Pose not detectedPose not detected
EfficientNetV2L Model
1Input Layer–00–0010.38124.41
2Input Layer–00–0113.50113.19
3Input Layer–00–0221.73147.26
4stem_conv–02–00293,395.01105.52
5stem_conv–02–01Pose not detectedPose not detected
6stem_conv–02–02Pose not detectedPose not detected
NASNetLarge Model
1input_1 Layer–00–004.85105.52
2input_1 Layer–00–0117.3696.62
3input_1 Layer–00–022,454,372.65113.61
4stem_bn1– 02–001,149,842.15110.19
5stem_bn1–02–018.5896.36
6stem_bn1–02–029.25105.60
The layer and filter marked in bold shows minimum translation and rotation error.
Table 3. Details of datasets used for creating map of environment.
Table 3. Details of datasets used for creating map of environment.
DatasetNo. of ImagesImage SizeGround Truth Available?Contains Loops?Widely Used?
Kitti Sequence 0045411241 × 376YesYesYes
Kitti Sequence 0246611241 × 376YesYesYes
Kitti Sequence 0527611226 × 370YesYesYes
Kitti Sequence 0611011226 × 370YesYesYes
Kitti Sequence 0840711226 × 370YesYesYes
Kitti Sequence 0915911226 × 370YesYesYes
Table 4. Estimated offset error on datasets.
Table 4. Estimated offset error on datasets.
DatasetNumber of FramesAverage Offset Error
Deep-VO [34]TartanVO [35]Ours
Kitti Sequence 00454113,227.92124,361.71476.40
Kitti Sequence 02466113,005.68317,924.601451.84
Kitti Sequence 0527617250.48133,977.30286.65
Kitti Sequence 0611017774.2345,526.16310.44
Kitti Sequence 08407123,942.91203,107.88374.23
Kitti Sequence 0915914379.55110,614.15536.42
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kazi, K.; Kalhoro, A.N.; Memon, F.; Memon, A.R.; Iqbal, M. Beyond Handcrafted Features: A Deep Learning Framework for Optical Flow and SLAM. J. Imaging 2025, 11, 155. https://doi.org/10.3390/jimaging11050155

AMA Style

Kazi K, Kalhoro AN, Memon F, Memon AR, Iqbal M. Beyond Handcrafted Features: A Deep Learning Framework for Optical Flow and SLAM. Journal of Imaging. 2025; 11(5):155. https://doi.org/10.3390/jimaging11050155

Chicago/Turabian Style

Kazi, Kamran, Arbab Nighat Kalhoro, Farida Memon, Azam Rafique Memon, and Muddesar Iqbal. 2025. "Beyond Handcrafted Features: A Deep Learning Framework for Optical Flow and SLAM" Journal of Imaging 11, no. 5: 155. https://doi.org/10.3390/jimaging11050155

APA Style

Kazi, K., Kalhoro, A. N., Memon, F., Memon, A. R., & Iqbal, M. (2025). Beyond Handcrafted Features: A Deep Learning Framework for Optical Flow and SLAM. Journal of Imaging, 11(5), 155. https://doi.org/10.3390/jimaging11050155

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop