Abstract
Navigation systems are developing rapidly; nevertheless, tasks are becoming more complex, significantly increasing the number of challenges for robotic systems. Navigation can be separated into global and local navigation. While global navigation works according to predefined data about the environment, local navigation uses sensory data to dynamically react and adjust the trajectory. Tasks are becoming more complex with the addition of dynamic obstacles, multiple robots, or, in some cases, inspection of places that are not physically reachable by humans. Cognitive tasks require not only detecting an object but also evaluating it without direct recognition. For this purpose, sensor fusion methods are employed. However, sensors of different physical nature sometimes cannot directly extract required information. As a result, AI methods are becoming increasingly popular for evaluating acquired information and for controlling and generating robot trajectories. In this work, a review of sensors for mobile robot localization is presented by comparing them and listing advantages and disadvantages of their combinations. Also, integration with path-planning methods is looked into. Moreover, sensor fusion methods are analyzed and evaluated. Furthermore, a concept for channel robot navigation, designed based on the research literature, is presented. Lastly, discussion and conclusions are drawn.
1. Introduction
In the rapid development of the automation and robotics world, making of autonomous vehicles and mobile robots is a big step toward operational efficiency, safety, and autonomy. At the heart of this technological revolution is the intricate domain of sensor fusion, a paradigm that merges data from different sensors to make cohesive and accurate perceptions of operational environments. This paper goes into the realm of sensor-fusion-based navigation systems for autonomous robots, spotlighting diverse methodologies that underpin their functionality and emerging trends that shape their evolution.
Navigational autonomy in robots is paramount for their effective deployment across a spectrum of applications, from industrial automation to exploration in inaccessible terrains. Traditional navigation methodologies, while foundational, often grapple with complexities and dynamic changes intrinsic to real-world environments. Bridging this gap, advanced navigation systems harness the synergy of global and local navigation methods [1]. Global navigation operates on the premise of pre-acquired environmental knowledge, facilitating formulation and adherence to predetermined paths. In contrast, local navigation equips mobile robots with agility to dynamically refine their paths in real time, utilizing an arsenal of external sensors—ranging from infrared and ultrasonic sensors to LASER, LIDAR, and cameras [2]. This sensorial diversity, when orchestrated by sophisticated software algorithms, enables autonomous correction of robot orientation and trajectory, ensuring navigational resilience against unforeseen obstacles and alterations in environment [3].
The dichotomy of global and local navigation methods embodies methodological diversity in robotic navigation, allowing robots to chart optimal paths and fulfil their designated tasks within varied environmental contexts. Nevertheless, reliance on prior environmental knowledge or capability for real-time path adjustment underscores limitations of classic navigation approaches [4,5]. These systems often operate within a deterministic framework, wherein navigation paths are predetermined, or a non-deterministic framework that allows for probabilistic path planning based on sensor input and environmental interaction [6].
A non-deterministic framework becomes very relevant in applications that require navigation in hazardous and physically difficult to reach places for humans—for example, inspection of narrow underground channels. That kind of working environment lacks global reference points that could be used for a deterministic framework. Furthermore, there is the probability of encountering unexpected obstacles. For these reasons, integration of sensors for robot localization is a must.
Amidst these methodologies, optical data-based localization emerges as a critical area of focus, leveraging visual information to enhance a robot’s environmental awareness and decision-making capability. However, reliance on optical data introduces unique challenges for navigation, including the need for sophisticated object recognition algorithms and the ability to define navigational paths without explicit recognition cues [7,8].
As we delve deeper into the big picture of research in sensor-fusion-based navigation, this paper aims to elucidate myriad localization methods that empower mobile robots to traverse and interact with their surroundings effectively. By analyzing limitations of classic localization approaches and addressing challenges posed by optical data reliance, we seek to highlight the transformative potential of sensor fusion in crafting more adaptable, reliable, and sophisticated autonomous navigation solutions primarily focused on local path planning.
In anthropocentric terms, localization methods can be classified into vision-based and non-vision-based approaches, which makes the distinction easier to grasp. Vision-based methods rely on imaging cameras to capture visual information, similar to human sight, which is then analyzed in various ways to understand and navigate the environment. Non-vision-based methods, in contrast, use sensors like LIDAR, radar, ultrasonic etc., which perceive the environment through means that are alien to human senses, such as detecting distances through sound waves or localizing oneself through RFID tags.
The manuscript is organized to provide a comprehensive review of sensor-fusion-based navigation systems. Section 2: A literature search method details systematic processes, databases, and inclusion criteria used to gather relevant studies. Section 3: Navigation methods review global and local approaches, discussing their principles, strengths, and limitations. Section 4: Analysis of non-vision-based localization systems highlights technologies like ultrasonic, infrared, LiDAR, and radar sensors, while Section 5: Analysis of vision-based localization systems examines both standalone and hybrid configurations, focusing on integration and challenges. Section 6: Essential sensor fusion systems classify fusion architectures into cooperative, complementary, and competitive approaches, exploring key methodologies. Section 7: A solution for channel robot navigation presents exemplary cost efficient sensor fusion based local navigation system intended for mobile robots functioning in channels that cannot be physically reached by a human, combining RGB cameras, laser pointers, and pseudo-LiDAR. Finally, Section 8: Discussion and conclusions summarize key findings, emerging trends, and future directions in sensor fusion for robotics.
2. Literature Search Method
The literature search method was based on the systemic process presented in article [9], which focuses on preferred reporting items of systematic reviews and meta-analyses (PRISMA) statement. Four main databases were utilized, including MDPI, IEEE Xplore, Google Scholar, and Science direct. Other specific databases were also used if there was no other way to access a required paper. The main criteria focusing on autonomous robot navigation topic were formed for the inclusion in this survey, such as:
- Focused on sensor application
- Focused on path planning
- Focused on mapping techniques
- Focused on sensor fusion method adaptions
- Focused on machine learning adaptions
Additional criteria for narrowing the main topic:
- Articles that are older than 5 years were excluded with some exceptions if specific points needed more investigation.
- Articles that do not focus on mobile robot navigation were excluded except if specific technology being investigated needed more input.
- Articles focusing on railways and sea navigation were not taken into consideration with the exception of several articles presenting air navigation systems.
The main keywords that were selected for research on sensor fusion and autonomous mobile robot included in this manuscript were: “Sensor fusion”, “YOLO”, “Mobile robot”, “Kalman filter”, “Sensors for navigation”, “Path planning methods”, “LiDAR and camera fusion”, and “ML based sensor fusion”. A simplified workflow of the concluded survey for this manuscript is shown in Figure 1.
Figure 1.
A systematic literature review workflow.
4. Analysis of Non-Vision-Based Localization Systems
Non-vision-based localization technologies play a crucial role in the field of robotics, especially in environments where visual data may be unreliable or unavailable. These technologies encompass a variety of methods and sensors designed to enhance a robot’s ability to localize and navigate itself within its environment, leveraging alternative sensory data to achieve precise and reliable navigation. The same is true of the common non-vision-based technologies, which are shown on the left side of Figure 2.
One significant branch of non-vision-based localization focuses on target localization. This involves determining the position of specific targets within an environment, utilizing technologies such as Ultra-Wideband (UWB), Bluetooth Low Energy (BLE), and Radio-Frequency Identification (RFID). UWB technology, known for its high accuracy and reliability, is widely used in indoor positioning systems due to its ability to provide precise location information even in complex environments [21]. BLE, on the other hand, is commonly employed for proximity detection and location tracking, benefiting from its low power consumption and widespread use in consumer electronics [22]. RFID systems offer another layer of versatility, allowing for the identification and tracking of objects through electromagnetic fields [23]. These technologies collectively enhance the ability of robots to locate and interact with various targets, crucial for applications such as inventory management and asset tracking.
Robot localization, another critical aspect of non-vision-based localization, involves methods that enable robots to determine their own position within an environment. Infrared (IR) sensors are versatile tools used in both target and robot localization, providing reliable distance measurements and object detection capabilities [24]. Tactile sensors, which detect physical contact with objects, are particularly useful in cluttered environments where precise positioning is essential [25]. Ultrasonic sensors, employing sound waves to measure distances, are effective for obstacle detection and navigation in various conditions, including occluded vision due to fog or smoke or underwater environments [26]. Lidar (Light Detection and Ranging) systems stand out due to their ability to create high-resolution maps of the environment using laser pulses, offering unparalleled accuracy and detail [27,28]. Radar systems, which use radio waves, provide robust performance in diverse environmental conditions, making them indispensable for applications requiring reliable distance, angle, and velocity measurements [29]. To unravel and compare non-vision sensors for robot localization, methods proposed in the literature were analyzed and presented in Table 2.
Table 2.
Non-vision-based robot localization technology review.
From Table 2, we can see a variety of solutions to effectively achieve local navigation by incorporating proximity and contact sensors of different physical nature to detect obstacles. Due to field view limitations, it is noticeable that ultrasonic and IR distance sensors are usually used in combinations to compensate for those disadvantages. LiDAR and radar sensors have higher accuracy and field of view but require more efficient mapping techniques to increase performance. Further comparison of analyzed sensors is shown in Table 3.
Table 3.
Non-vision robot localization technology comparison.
As shown in Table 3, tactile sensors computationally lack proximity evaluation capabilities but are very computationally efficient. They are a great addition not only for obstacle detection purposes but also for collaborative function with human operators. Also, it is worth mentioning that in recent studies, tactile sensors vary in complexity and can even become a system of several sensors to measure contact and deformation phenomena. For example, in article [46], an optical tactile sensing system is presented, which can measure force distribution for arial mobile robot purposes.
The integration of these non-vision-based navigation technologies into robotic systems addresses several challenges associated with visual data reliance. For instance, varying lighting conditions and the need for sophisticated object recognition algorithms can complicate vision-based navigation. Non-vision-based systems, leveraging a combination of sensory inputs such as IR, tactile, ultrasonic, lidar, and radar, can navigate and localize effectively without the constraints of visual data. This adaptability is particularly advantageous in environments like warehouses, underwater explorations, and subterranean locales such as mines or tunnels where visual cues are limited or non-existent.
5. Analysis of Vision-Based Localization Systems
5.1. Standalone Vision Navigation Systems
Vision capability is an essential feature for mobile robot navigation systems. Many cameras were proven to work in this scenario with corresponding advantages and disadvantages, some of which are shown one the right side of Figure 2.
Camera devices can be separated into single and 3D cameras. Single camera can take 2D images. Most commonly used single cameras frequently used in robotic systems are RGB cameras based on CCD or CMOS sensors, which represent each taken pixel in an extensive spectrum of colors extracted from red, green, and blue color space [47]. They are highly applied for navigation. In article [48], an RGB camera is used to detect road lines according to the color so vehicles could follow the path in combination with other sensors. Other notable examples of single cameras are NIR cameras, which are less sensitive to visible light, meaning images are not corrupted by reflections [49]. Also, a fisheye camera is a powerful omni-directional perception sensor. It is used in navigation systems because of its wide field of view. In article [50], a fisheye camera is used to take images from 180 angle using the ASIFT algorithm to extract features of obstacles. Another edition for visions devices is the polarized camera, which had polarization systems able to extract orientation of the light oscillations reflected from perceived surfaces [51]. It is very convenient for detecting objects in crowded environments by filtering unwanted reflections and glare and enhancing the image contrast.
Depth measurement capability allows not only color recognition but also evaluation of object 3D geometry. One of the most frequently used cameras for this purpose is the RGB-D camera, which emits a predefined pattern of infrared light rays and the depth of each pixel is calculated by the reflection of rays [52]. Similarly, time of flight IR cameras work by illuminating present objects with modulated light and observing reflections, allowing the robot to perceive depth [53], although color cannot be perceived with this camera. Another increasing in popularity is the event-based camera, frequently employing DVS sensors, which capture pixel intensity changes, and robust compared to other cameras [54]. These cameras can also calculate depth by event capture, although it is computationally demanding and methods for efficiency are needed.
All mentioned cameras have corresponding advantages and disadvantages. To evaluate their properties and functionality, some of the researched methods for mobile robots and other types of navigation that integrate cameras in their systems will be analyzed. The researched methods are presented in Table 4.
Table 4.
Vision based robot localization technology review.
From Table 4, we can see a wide application of cameras for navigation purposes. Several techniques to effectively use vision devices for recognition were mentioned. One of the most popular techniques improving rapidly is you only look once (YOLO) and its advanced versions, which can work with high accuracy and speed in real time. It converts a target detection problem into a regression problem, dividing images in grids and making predictions for each grid cell separately [61]. YOLO incorporates convolutional neural network (CNN) principles to train and predict image data [62]. A typical YOLO network architecture is shown in Figure 3.
Figure 3.
Typical YOLO network architecture consisting of 24 convolutional layers [63].
The first 24 convolutional layers extract features from the image, and the two fully connected layers predict the output bounding boxes and class probabilities directly from image pixels. Models from YOLO-v1 to the newly developed YOLO-v9 improved significantly. Going from YOLO-v1 to YOLO-v8 increased processing speed from 45 to 280 FPS and increased detection accuracy of 53.9% [64]. As stated in article [65] newly developed YOLO-v9 further increases detection accuracy by reducing information loss, which is encountered in sequential feature extraction process by utilization of programmable gradient information.
To select the most effective visions system for a specific project, it is important to know not only image recognition methods but also properties of devices. From the research papers, the properties of the most used cameras were summed up for comparison in Table 5.
Table 5.
Vision-based robot localization technology comparison.
From Table 5, it can be seen that CCD and CMOS cameras are the most cost efficient and have established methods for efficient object recognition tasks. They lack depth capability compared to other cameras in the table. Nevertheless, if it is convenient for a project because of the advantages mentioned, it is possible to measure depth with these cameras to a certain accuracy. For example, in a previously mentioned article [57], the triangulation principle was used to detect changes in lase pointer projection to estimate distance. Also, using similar triangulation principles, two cameras positioned at slightly different positions can measure depth by matching taken images [78]. By measuring the required time for reflected light to go from the source and come back, the concept of ToF sensors is designed. ToF cameras working in an infrared range are very convenient for accurate depth estimation. On the other hand, RGB-D can not only estimate depth based on similar principles but also detect a wide range of colors, but it is moderately more expensive than previous cameras and requires smarter algorithms for more efficient matching of color and depth. For example, in article [79], adaptive color-depth matching is proposed using a transformer-based framework to enhance computational performance. Lastly, event-based cameras enhance capabilities of object detection even more with high dynamic range and depth measuring capabilities. Nevertheless, these cameras are more expensive and challenge current AI-based methods for more effective performance.
5.2. Hybrid Visions Localization Systems
As previously explained, depth and field of view estimation with cameras is limited and, in some cases, expensive. Moreover, certain surfaces introduce challenges for detection. For this reason, in robotic navigation systems, cameras are commonly integrated in combination with other distance measurement sensors to enhance overall perception of working environments. Some of the common fusion combinations are shown in Figure 2.
As navigation environments are becoming more complex with dynamic obstacles and crowded spaces, infrastructures having more than one sensor became the staple of localization, combining sensors that can detect different physical phenomena. To obtain a better understanding of the advantages and disadvantages of hybrid systems, proposed methods in the literature were analyzed. Some of the methods are shown in Table 6.
Table 6.
Hybrid robot localization technology review.
Going through analyzed approaches of hybrid sensor methods in Table 6, it is clear that richer data can be acquired from working environments. Combining distance and visual sensors enables significantly more accurate object detection, which is achieved by mapping accurate distance data with visual data. On the other hand, all presented methods deal with high computational resources. To increase performance of mapping sensor data, several methods were established in time. One of the most widely used is simultaneous localization and mapping (SLAM), which utilizes data from the camera, distance, and other sensors and concurrently estimates sensor poses to generate a comprehensive 3D representation of the surrounding environments [86]. LiDAR and visual SLAM are well-known techniques, but the need to fuse different sensors established new algorithms. For example, in article [87], LiDAR inertial camera SLAM is proposed, enabling accurate tracking and photorealistic map reconstruction using 3D Gaussian splatting.
The core of hybrid sensors systems are fusion methods including Kalman filters, particle filters, and AI methods, which drastically affect the performance of the system. These methods will be introduced further in the next chapter. It is also important to choose the right devices for the project according to sensor properties, which affect the overall performance of the system. A comparison between different hybrid sensors combinations is presented in Table 7.
Table 7.
Hybrid robot localization technology comparison.
From Table 7, it can be seen that depth capabilities of RGB, RGB-D, and DVS cameras are enhancing significantly in fusion with distance sensors. To achieve the highest accuracy, DVS and Lidar solutions show a lot of promise, because DVS cameras also have low sensitivity to disturbances. If cost-efficient solutions with range capabilities are needed, then combining ultrasonic or radar sensors with cameras is a way to go. Combination of tactile sensors with cameras might not provide range but can be used for force-sensitive applications to detect and inspect object geometry and even material properties.
6. Essential Sensor Fusion Systems
Sensor fusion is an essential part of navigation because standalone systems based on one or two sensors cannot cope with increasing complexity of working environments and required tasks. As mentioned before, the addition of cooperating or competitive sensors allows an increase in the overall properties of the system including field of view and accuracy, taking into account different physical phenomena to generate better understanding about working environments and internal processes. To maximize the performance of sensor fusion, it is important to choose appropriate architecture depending on required tasks and chosen sensors. For better understanding, sensor fusion is classified by several factors in the literature. One of the main factors that regularly appears in the literature [100,101] defines how early sensor data are interconnected during data processing steps. It can be interpreted as abstraction level. Sensor fusion level according to abstraction can be classified as:
- Low-level—indicates that raw sensor data are directly sent to fusion module. This way no data are lost because of noise introduced by postprocessing, meaning some relevant data would not be overlooked. For example, in article [102], LiDAR 3D point cloud points are augmented by semantically strong image features significantly increasing the number of detected 3D bounding boxes. Nevertheless, high computational resources are required to compute raw data. Also, fusion modules are less adaptive because adding new sensors requires adjustments to the new sensor format.
- Medium-level (Feature)—involves extracting some key features from raw sensors. Due to this, bandwidth is reduced before carrying data fusion and similar efficiency of extracting relevant data is achieved. Also, this structure is more adaptive and adjustable. This is a very commonly used fusion method, then optimization is important. For instance, in article [103], encoder, color image, and depth image are first pre-processed before fusion. Unnecessary noise is removed from images to filter only required regions, and encoder provides orientation, ultimately creating a system capable of object recognition and robot localization.
- High-level—according to this structure, each sensor is postprocessed and carries out its task independently, and then high-level fusion of detected objects or trajectories by each sensor is performed. This type of fusion has high modularity and simplicity. On the other hand, key sensor data at lower levels are lost.
Another way to classify sensor fusion architectures is by relationship among the sources as listed in article [104], separating into three groups:
- Complementary—sensor information does not directly depend on one another but then combined can provide a more complete picture of observed phenomena.
- Competitive (redundant)—same or similar information is received from sensors to reduce uncertainties and errors, which could appear if using sensors separately.
- Cooperative—involves combined information extraction that cannot be acquired using one sensor. Involves active sensor collaboration exchanging insight and or intermediate data and increasing accuracy and reliability of overall fusion system.
To design proposed architectures and realize sensor fusion, specific methods and algorithms are required including Kalman, particle filters, novel neural network approaches, etc. To obtain a better understanding of sensor fusion architectures and methods used for mobile robot navigation and classification tasks, proposed solutions in literature are analyzed and presented in in Table 8, Table 9 and Table 10 below.
Table 8.
Low-level cooperative sensor fusion methods for mobile robot navigation.
Table 9.
Mid-level cooperative sensor fusion methods for mobile robot navigation.
Table 10.
High-level complementary sensor fusion methods for mobile robot navigation.
Low-level fusion architecture is useful for systems that require maximizing acquired data from the sensors with no loss for higher accuracy. Systems presented in the Table are designed for obstacle and human detection tracking tasks. These tasks must be performed with upmost accuracy to ensure safety in for all elements in the working environment. To integrate low-level fusion architecture in modern systems, which require real-time capabilities and communication between various software and hardware elements, optimization is necessary to reduce computational load.
High-level sensor fusion requires significant computing resources, and often these facilities are located remotely and connected via a fast network; therefore, known realized cases are less numerous except the previous ones.
Comparing analyzed sensor fusion approaches, it can be seen that for mobile robot navigation systems, which mainly focus on robot and target localization, cooperative mid-level sensor fusion architectures are dominant. Navigation requires not only accuracy but also efficiency to perform localization tasks faster. Due to this, mid-level sensor fusion architectures are convenient. Nevertheless, the system has to evaluate more phenomena which are not directly dependent on one another, and complementary fusion becomes handy. This is especially common in vehicle-to-everything communication. For example, in article [119], high-level fusion structure is presented where LiDAR and Radar is tasked with distance measurement and obstacle detection, and the camera complements the system by classification of objects. There are also plenty of modular-type sensor fusion architectures in autonomous robotic systems. For example, in article [120], a human detection system is designed with complementary sensor fusion. There, LiDAR is used to detect the lower part of a human and camera for the upper part of pose recognition.
To realize the designed structure of sensor fusion, the next step is to choose appropriate methods and algorithms to interconnect sensor data for correct estimation of system state. Going through the analyzed approaches in Table 9, several methods can be distinguished, which will be presented below.
6.1. Sensor Fusion Using Kalman Filter
Kalman filter is a common method for sensor fusion because it can estimate parameters of a constantly changing system in real time, minimizing error covariance [121], although standard Kalman filter is not suitable for non-linear systems. Nowadays, several advanced Kalman filter methods are used for robotic systems, which were briefly mentioned before. For example, extended Kalman filter (EKF) is commonly used for non-linear systems [122]. First it constructs linear estimation, but then it is subsequentially updated. It is especially useful for merging sensor data with varying measurement models like GPS, IMU, and vision systems. Nevertheless, subsequential update of linear estimation requires calculating partial derivatives in each step, significantly increasing computational load. Unscanned Kalman filter (UKF) was created to work around the shortcomings of EKF, which can be applied for non-linear systems without direct laterization using sigma point approach for calculation mean and covariance. This method is very useful for accelerometer, GNSS and rotation sensordata fusion as presented in article [123].
Going further, cubature Kalman filter (CKF) was built upon its predecessors, which can deal with non-linear data with accuracy and reliability by performing high-dimensional state estimation. Nevertheless, it showed to suffer from error accumulation in long-term operations. In article [124], utilization of trifocal tensor geometry (TTG) for the CKF algorithm was suggested to increase filter estimation accuracy for long-term visual inertial odometry application.
Another recent filter showing great results for tracking large-scale moving objects is probabilistic Kalman filter (PKF). It simplifies conventional state variables, thus reducing computational load and making non-uniform modelling more effective. For example, in article [125], PKF-based non-uniform formulation is proposed for tackling escape problems in multi-object tracking and introducing a first fully GPU-based tracker paradigm. Non-uniform motion is modelled as uniform motion by transforming a time variable into a related displacement variable allowing to integrate deacceleration strategy into a control input model.
6.2. Sensor Fusion Using Particle Filter
It is another class of estimation algorithms that involves a probabilistic approach to estimate the state of the system. Particle filter (PF) stands out because of its ability to deal with non-linear system models and non-Gaussian noise. They also show great potential for localization and object detection tasks. For example, in article [126], PF is used for two ultrasonic sensors and radar fusion for a system that is able to navigate in unknown environments with static and dynamic obstacles. However, basic PF approaches are not suitable for real-time applications especially if the required number particles for accurate estimation is very high [127]. In article [128], an enhanced particle filter-weighted differential evolution (EPF-WDE) scheme is proposed, which is used to manage a non-linear and multidimensional system involving a variety of smartphone sensors with notable gains in accuracy and convergence.
6.3. Deep Learning for Sensor Fusion
Navigation systems are becoming increasingly complex with a large number of sensors with different physical nature. This amounts to large amount of imperfect raw data. Multi-modal sensor fusion architecture is essential in these cases, and deep learning (DP) techniques are emerging to tackle these tasks. DP is very effective because of non-linear mapping capabilities. Furthermore, DP models have deep layers that can generate high-dimensional representations, which are more comprehensive compared to previously mentioned methods. Also, it is very flexible and can be applied to a variety of applications [129]. In article [130], an adaptive-network-based fuzzy interface system (ANFIS) is proposed for LiDAR and inertial navigation system GNSS/INS fusion to localize indoor mobile robot. It incorporated human-like decision making with neural networks, which enables learning from data and improving performance, and it resulted in a lower standard deviation error compared to more classical EKF method. Another deep learning-based high-level fusion method for LiDAR and camera data is presented in article [131]. The author proposes high-order Attention Mechanism Fusion Networks (HAMFNs) for multi-scale learning and image expression analysis. It is capable of more accurate perception of surrounding the objects’ state, which is essential in autonomous driving.
8. Discussion and Conclusions
Going through the analyzed literature concerning mobile robot navigation, it is clear that hybrid sensor localization systems will be applied even more in the future. The combination of vision and distance sensors enhances the ability of object detection with accurate distance, color, and dynamic behavior estimation. Sensor hardware is also improving, in some cases creating modules combining several functions—for example, modules incorporating infrared, RGB, and ToF functionality. Furthermore, dynamic vision sensors (DVS) are rapidly improving with significant advantages over standard cameras with low latency, high dynamic range, and ability to estimate depth. Polarization filters also proved advantageous for vision technology by enhancing vision contrast and allowing detection of more difficult to perceive objects. Tactile sensor technology improves independently from optical navigation technologies. The soft structure of tactile sensors is able to inspect contact with obstacles with high accuracy, allowing navigation in very narrow spaces or even evaluating terrain properties and slippage.
Nevertheless, hardware technology advancement is relatively stable in comparison to software development, which will ring the main advantage in the future to enhanced performance of multi sensor systems. AI methods are proving to be effective in all stages of multi-sensor systems starting from post-processing individual sensor data to fusing and mapping the overall picture of the environment. For vision technology, new versions of YOLOv8 and YOLOv9 object detection systems are being built upon further for distinguishing small details in the image from large datasets. Deep learning architectures are progressively improving. CNN networks are commonly used for LiDAR and camera mapping. Nevertheless, transformer networks are being researched, which can increase performance of classification and mapping tasks especially when working with large datasets.
Advancement AI techniques for multi-sensor data processing, mapping, and path planning also allow use of cost-efficient sensors by enhancing software performance. The concept of a cost-efficient multi-sensor autonomous channel robot was presented incorporating a laser-RGB camera scanner, pseudo-LiDAR, and inertial sensors odometry. Future work will focus on incorporating deep learning methods for data fusion and path planning according to the research conducted in this survey.
Author Contributions
Conceptualization, V.U. and V.B.; methodology, A.D.; software, M.N.; validation, M.N., A.D. and V.U.; formal analysis, V.U.; investigation, V.U. and V.B., resources, V.B.; writing—original draft preparation, V.U.; writing—review and editing, V.B. and A.D.; visualization, V.U.; supervision, V.B.; project administration, A.D.; funding acquisition, V.B. All authors have read and agreed to the published version of the manuscript.
Funding
This research received no external funding.
Data Availability Statement
No new data were created or analyzed in this study. Data sharing is not applicable to this article.
Conflicts of Interest
The authors declare no conflicts of interest.
References
- Sekaran, J.F.; Sugumari, T. A Review of Perception-Based Navigation System for Autonomous Mobile Robots. Recent Pat. Eng. 2022, 17, e290922209298. [Google Scholar] [CrossRef]
- Chen, J.; Wang, H.; Yang, S. Tightly Coupled LiDAR-Inertial Odometry and Mapping for Underground Environments. Sensors 2023, 23, 6834. [Google Scholar] [CrossRef]
- Tatsch, C.; Bredu, J.A.; Covell, D.; Tulu, I.B.; Gu, Y. Rhino: An Autonomous Robot for Mapping Underground Mine Environments. In Proceedings of the 2023 IEEE/ASME International Conference on Advanced Intelligent Mechatronics (AIM), Seattle, WA, USA, 28–30 June 2023; pp. 1166–1173. [Google Scholar] [CrossRef]
- Yang, L.; Li, P.; Qian, S.; Quan, H.; Miao, J.; Liu, M.; Hu, Y.; Memetimin, E. Path Planning Technique for Mobile Robots: A Review. Machines 2023, 11, 980. [Google Scholar] [CrossRef]
- Liu, Y.; Zhang, W.; Li, F.; Zuo, Z.; Huang, Q. Real-Time Lidar Odometry and Mapping with Loop Closure. Sensors 2022, 22, 4373. [Google Scholar] [CrossRef] [PubMed]
- Qin, H.; Shao, S.; Wang, T.; Yu, X.; Jiang, Y.; Cao, Z. Review of Autonomous Path Planning Algorithms for Mobile Robots. Drones 2023, 7, 211. [Google Scholar] [CrossRef]
- Jwo, D.J.; Biswal, A.; Mir, I.A. Artificial Neural Networks for Navigation Systems: A Review of Recent Research. Appl. Sci. 2023, 13, 4475. [Google Scholar] [CrossRef]
- Liu, C.; Lekkala, K.; Itti, L. World Model Based Sim2Real Transfer for Visual Navigation. arXiv 2023, arXiv:2310.18847. [Google Scholar]
- Almeida, J.; Rufino, J.; Alam, M.; Ferreira, J. A Survey on Fault Tolerance Techniques for Wireless Vehicular Networks. Electronics 2019, 8, 1358. [Google Scholar] [CrossRef]
- Alshammrei, S.; Boubaker, S.; Kolsi, L. Improved Dijkstra Algorithm for Mobile Robot Path Planning and Obstacle Avoidance. Comput. Mater. Contin. 2022, 72, 5939–5954. [Google Scholar] [CrossRef]
- Li, J.; Qin, H.; Wang, J.; Li, J. OpenStreetMap-Based Autonomous Navigation for the Four Wheel-Legged Robot Via 3D-Lidar and CCD Camera. IEEE Trans. Ind. Electron. 2022, 69, 2708–2717. [Google Scholar] [CrossRef]
- Martins, O.O.; Adekunle, A.A.; Olaniyan, O.M.; Bolaji, B.O. An Improved Multi-Objective a-Star Algorithm for Path Planning in a Large Workspace: Design, Implementation, and Evaluation. Sci. Afr. 2022, 15, e01068. [Google Scholar] [CrossRef]
- Wang, H.; Qi, X.; Lou, S.; Jing, J.; He, H.; Liu, W. An Efficient and Robust Improved A* Algorithm for Path Planning. Symmetry 2021, 13, 2213. [Google Scholar] [CrossRef]
- Abdulsaheb, J.A.; Kadhim, D.J. Classical and Heuristic Approaches for Mobile Robot Path Planning: A Survey. Robotics 2023, 12, 93. [Google Scholar] [CrossRef]
- Wang, H.; Fu, Z.; Zhou, J.; Fu, M.; Ruan, L. Cooperative Collision Avoidance for Unmanned Surface Vehicles Based on Improved Genetic Algorithm. Ocean Eng. 2021, 222, 108612. [Google Scholar] [CrossRef]
- Guo, N.; Li, C.; Wang, D.; Song, Y.; Liu, G.; Gao, T. Local Path Planning of Mobile Robot Based on Long Short-Term Memory Neural Network. Autom. Control Comput. Sci. 2021, 55, 53–65. [Google Scholar] [CrossRef]
- Zohaib, M.; Pasha, S.M.; Javaid, N.; Iqbal, J. IBA: Intelligent Bug Algorithm—A Novel Strategy to Navigate Mobile Robots Autonomously. Commun. Comput. Inf. Sci. 2013, 414, 291–299. [Google Scholar] [CrossRef]
- van Breda, R.; Smit, W.J. Applicability of Vector Field Histogram Star (Vfh*) on Multicopters. In Proceedings of the International Micro Air Vehicle Competition and Conference 2016, Beijing, China, 17–21 October 2016; pp. 62–69. [Google Scholar]
- Kobayashi, M.; Motoi, N. Local Path Planning: Dynamic Window Approach with Virtual Manipulators Considering Dynamic Obstacles. IEEE Access 2022, 10, 17018–17029. [Google Scholar] [CrossRef]
- Mishra, D.K.; Thomas, A.; Kuruvilla, J.; Kalyanasundaram, P.; Prasad, K.R.; Haldorai, A. Design of Mobile Robot Navigation Controller Using Neuro-Fuzzy Logic System. Comput. Electr. Eng. 2022, 101, 108044. [Google Scholar] [CrossRef]
- Durodié, Y.; Decoster, T.; Van Herbruggen, B.; Vanhie-Van Gerwen, J.; De Poorter, E.; Munteanu, A.; Vanderborght, B. A UWB-Ego-Motion Particle Filter for Indoor Pose Estimation of a Ground Robot Using a Moving Horizon Hypothesis. Sensors 2024, 24, 2164. [Google Scholar] [CrossRef]
- Wu, H.; Liu, H.; Roddelkopf, T.; Thurow, K. BLE Beacon-Based Floor Detection for Mobile Robots in a Multi-Floor Automation Laboratory. Transp. Saf. Environ. 2023, 6, tdad024. [Google Scholar] [CrossRef]
- Tripicchio, P.; D’Avella, S.; Unetti, M.; Motroni, A.; Nepa, P. A UHF Passive RFID Tag Position Estimation Approach Exploiting Mobile Robots: Phase-Only 3D Multilateration Particle Filters With No Unwrapping. IEEE Access 2024, 12, 58778–58788. [Google Scholar] [CrossRef]
- Özcan, M.; Aliew, F.; Görgün, H. Accurate and Precise Distance Estimation for Noisy IR Sensor Readings Contaminated by Outliers. Measurement 2020, 156, 107633. [Google Scholar] [CrossRef]
- Hu, H.; Zhang, C.; Pan, C.; Dai, H.; Sun, H.; Pan, Y.; Lai, X.; Lyu, C.; Tang, D.; Fu, J.; et al. Wireless Flexible Magnetic Tactile Sensor with Super-Resolution in Large-Areas. ACS Nano 2022, 16, 19271–19280. [Google Scholar] [CrossRef] [PubMed]
- Khaleel, H.Z.; Oleiwi, B.K. Ultrasonic Sensor Decision-Making Algorithm for Mobile Robot Motion in Maze Environment. Bull. Electr. Eng. Inform. 2024, 13, 109–116. [Google Scholar] [CrossRef]
- De Heuvel, J.; Zeng, X.; Shi, W.; Sethuraman, T.; Bennewitz, M. Spatiotemporal Attention Enhances Lidar-Based Robot Navigation in Dynamic Environments. IEEE Robot. Autom. Lett. 2024, 9, 4202–4209. [Google Scholar] [CrossRef]
- Cañadas-Aránega, F.; Blanco-Claraco, J.L.; Moreno, J.C.; Rodriguez-Diaz, F. Multimodal Mobile Robotic Dataset for a Typical Mediterranean Greenhouse: The GREENBOT Dataset. Sensors 2024, 24, 1874. [Google Scholar] [CrossRef] [PubMed]
- Brescia, W.; Gomes, P.; Toni, L.; Mascolo, S.; De Cicco, L. MilliNoise: A Millimeter-Wave Radar Sparse Point Cloud Dataset in Indoor Scenarios. In Proceedings of the MMSys ‘24: Proceedings of the 15th ACM Multimedia Systems Conference, Bari, Italy, 15–18 April 2024; pp. 422–428. [Google Scholar] [CrossRef]
- Ou, X.; You, Z.; He, X. Local Path Planner for Mobile Robot Considering Future Positions of Obstacles. Processes 2024, 12, 984. [Google Scholar] [CrossRef]
- Wang, C.; Zang, X.; Song, C.; Liu, Z.; Zhao, J.; Ang, M.H. Virtual Tactile POMDP-Based Path Planning for Object Localization and Grasping. Meas. J. Int. Meas. Confed. 2024, 230, 114480. [Google Scholar] [CrossRef]
- Armleder, S.; Dean-Leon, E.; Bergner, F.; Guadarrama Olvera, J.R.; Cheng, G. Tactile-Based Negotiation of Unknown Objects during Navigation in Unstructured Environments with Movable Obstacles. Adv. Intell. Syst. 2024, 6, 21. [Google Scholar] [CrossRef]
- Al-Mallah, M.; Ali, M.; Al-Khawaldeh, M. Obstacles Avoidance for Mobile Robot Using Type-2 Fuzzy Logic Controller. Robotics 2022, 11, 130. [Google Scholar] [CrossRef]
- Wondosen, A.; Shiferaw, D. Fuzzy Logic Controller Design for Mobile Robot Outdoor Navigation. arXiv 2017, arXiv:2401.01756. [Google Scholar]
- Sabra, M.; Tayeh, N. Maze Solver Robot. 2024. Available online: https://hdl.handle.net/20.500.11888/18671 (accessed on 15 December 2024).
- Kim, K.; Kim, J.; Jiang, X.; Kim, T. Static Force Measurement Using Piezoelectric Sensors. J. Sens. 2021, 2021, 6664200. [Google Scholar] [CrossRef]
- Kong, Y.; Cheng, G.; Zhang, M.; Zhao, Y.; Meng, W.; Tian, X.; Sun, B.; Yang, F.; Wei, D. Highly Efficient Recognition of Similar Objects Based on Ionic Robotic Tactile Sensors. Sci. Bull. 2024, 69, 2089–2098. [Google Scholar] [CrossRef]
- Zhang, S.; Yang, Y.; Sun, F.; Bao, L.; Shan, J.; Gao, Y.; Fang, B. A Compact Visuo-Tactile Robotic Skin for Micron-Level Tactile Perception. IEEE Sens. J. 2024, 24, 15273–15282. [Google Scholar] [CrossRef]
- Verellen, T.; Kerstens, R.; Steckel, J. High-Resolution Ultrasound Sensing for Robotics Using Dense Microphone Arrays. IEEE Access 2020, 8, 190083–190093. [Google Scholar] [CrossRef]
- Okuda, K.; Miyake, M.; Takai, H.; Tachibana, K. Obstacle Arrangement Detection Using Multichannel Ultrasonic Sonar for Indoor Mobile Robots. Artif. Life Robot. 2010, 15, 229–233. [Google Scholar] [CrossRef]
- Nair, S.; Joladarashi, S.; Ganesh, N. Evaluation of Ultrasonic Sensor in Robot Mapping. In Proceedings of the 2019 3rd International Conference on Trends in Electronics and Informatics (ICOEI), Tirunelveli, India, 23–25 April 2019. [Google Scholar]
- Li, Q.; Zhu, H. Performance Evaluation of 2D LiDAR SLAM Algorithms in Simulated Orchard Environments. Comput. Electron. Agric. 2024, 221, 108994. [Google Scholar] [CrossRef]
- Belkin, I.; Abramenko, A.; Yudin, D. Real-Time Lidar-Based Localization of Mobile Ground Robot. Procedia Comput. Sci. 2021, 186, 440–448. [Google Scholar] [CrossRef]
- Wang, H.; Yin, Y.; Jing, Q. Comparative Analysis of 3D LiDAR Scan-Matching Methods for State Estimation of Autonomous Surface Vessel. J. Mar. Sci. Eng. 2023, 11, 840. [Google Scholar] [CrossRef]
- Adams, M.; Jose, E.; Vo, B.-N. Robotic Navigation and Mapping with Radar; Artech: Morristown, NJ, USA, 2012; ISBN 9781608074839. [Google Scholar]
- Aucone, E.; Sferrazza, C.; Gregor, M.; D’Andrea, R.; Mintchev, S. Optical Tactile Sensing for Aerial Multi-Contact Interaction: Design, Integration, and Evaluation. IEEE Trans. Robot. 2024, 41, 364–377. [Google Scholar] [CrossRef]
- Omar, E.Z.; Al-Tahhan, F.E. A Novel Hybrid Model Based on Integrating RGB and YCrCb Color Spaces for Demodulating the Phase Map of Fibres Using a Color Phase-Shifting Profilometry Technique. Optik 2024, 306, 171792. [Google Scholar] [CrossRef]
- Maitlo, N.; Noonari, N.; Arshid, K.; Ahmed, N.; Duraisamy, S. AINS: Affordable Indoor Navigation Solution via Line Color Identification Using Mono-Camera for Autonomous Vehicles. In Proceedings of the IEEE 9th International Conference for Convergence in Technology (I2CT), Pune, India, 5–7 April 2024. [Google Scholar]
- Lan, H.; Zhang, E.; Jung, C. Face Reflection Removal Network Using Multispectral Fusion of RGB and NIR Images. IEEE Open J. Signal Process. 2024, 5, 383–392. [Google Scholar] [CrossRef]
- Zhang, Y.; Ma, Y.; Wu, Y.; Liu, L. Achieving Widely Distributed Feature Matches Using Flattened-Affine-SIFT Algorithm for Fisheye Images. Opt. Express 2024, 32, 7969. [Google Scholar] [CrossRef]
- Li, Y.; Moreau, J.; Ibanez-guzman, J. Emergent Visual Sensors for Autonomous Vehicles. IEEE Trans. Intell. Transp. Syst. 2023, 24, 4716–4737. [Google Scholar] [CrossRef]
- Tychola, K.A.; Tsimperidis, I.; Papakostas, G.A. On 3D Reconstruction Using RGB-D Cameras. Digital 2022, 2, 401–421. [Google Scholar] [CrossRef]
- Varghese, G.; Reddy, T.G.C.; Menon, A.K.; Paul, A.; Kochuvila, S.; Varma Divya, R.; Bhat, R.; Kumar, N. Multi-Robot System for Mapping and Localization. In Proceedings of the 2023 8th International Conference on Robotics and Automation Engineering (ICRAE), Singapore, 17–19 November 2023; pp. 79–84. [Google Scholar] [CrossRef]
- Ghosh, D.K.; Jung, Y.J. Two-Stage Cross-Fusion Network for Stereo Event-Based Depth Estimation. Expert Syst. Appl. 2024, 241, 122743. [Google Scholar] [CrossRef]
- Kim, T.; Lim, S.; Shin, G.; Sim, G.; Yun, D. An Open-Source Low-Cost Mobile Robot System with an RGB-D Camera and Efficient Real-Time Navigation Algorithm. IEEE Access 2022, 10, 127871–127881. [Google Scholar] [CrossRef]
- Canovas, B.; Nègre, A.; Rombaut, M. Onboard Dynamic RGB-D Simultaneous Localization and Mapping for Mobile Robot Navigation. ETRI J. 2021, 43, 617–629. [Google Scholar] [CrossRef]
- Abukhalil, T.; Alksasbeh, M.; Alqaralleh, B.; Abukaraki, A. Robot Navigation System Using Laser and a Monocular Camera. J. Theor. Appl. Inf. Technol. 2020, 98, 714–724. [Google Scholar]
- Tsujimura, T.; Minato, Y.; Izumi, K. Shape Recognition of Laser Beam Trace for Human-Robot Interface. Pattern Recognit. Lett. 2013, 34, 1928–1935. [Google Scholar] [CrossRef]
- Romero-Godoy, D.; Sánchez-Rodríguez, D.; Alonso-González, I.; Delgado-Rajó, F. A Low Cost Collision Avoidance System Based on a ToF Camera for SLAM Approaches. Rev. Tecnol. Marcha 2022, 35, 137–144. [Google Scholar] [CrossRef]
- Iaboni, C.; Lobo, D.; Choi, J.W.; Abichandani, P. Event-Based Motion Capture System for Online Multi-Quadrotor Localization and Tracking. Sensors 2022, 22, 3240. [Google Scholar] [CrossRef]
- Jiang, P.; Ergu, D.; Liu, F.; Cai, Y.; Ma, B. A Review of Yolo Algorithm Developments. Procedia Comput. Sci. 2021, 199, 1066–1073. [Google Scholar] [CrossRef]
- Zhang, J.; Zhang, Y.; Liu, J.; Lan, Y.; Zhang, T. Human Figure Detection in Han Portrait Stone Images via Enhanced YOLO-V5. Herit. Sci. 2024, 12, 119. [Google Scholar] [CrossRef]
- Plastiras, G.; Kyrkou, C.; Theocharides, T. Efficient Convnet-Based Object Detection for Unmanned Aerial Vehicles by Selective Tile Processing. In Proceedings of the ICDSC ‘18: Proceedings of the 12th International Conference on Distributed Smart Cameras, Eindhoven, The Netherlands, 3–4 September 2018. [Google Scholar] [CrossRef]
- Hussain, M. YOLOv1 to v8: Unveiling Each Variant-A Comprehensive Review of YOLO. IEEE Access 2024, 12, 42816–42833. [Google Scholar] [CrossRef]
- Verma, T.; Singh, J.; Bhartari, Y.; Jarwal, R.; Singh, S.; Singh, S. SOAR: Advancements in Small Body Object Detection for Aerial Imagery Using State Space Models and Programmable Gradients. arXiv 2024, arXiv:2405.01699. [Google Scholar]
- Minz, P.S.; Saini, C.S. RGB Camera-Based Image Technique for Color Measurement of Flavored Milk. Meas. Food 2021, 4, 100012. [Google Scholar] [CrossRef]
- Sohl, M.A.; Mahmood, S.A. Low-Cost UAV in Photogrammetric Engineering and Remote Sensing: Georeferencing, DEM Accuracy, and Geospatial Analysis. J. Geovisualization Spat. Anal. 2024, 8, 14. [Google Scholar] [CrossRef]
- Haruta, M.; Kikkawa, J.; Kimoto, K.; Kurata, H. Comparison of Detection Limits of Direct-Counting CMOS and CCD Cameras in EELS Experiments. Ultramicroscopy 2022, 240, 113577. [Google Scholar] [CrossRef] [PubMed]
- Ünal, Z.; Kızıldeniz, T.; Özden, M.; Aktaş, H.; Karagöz, Ö. Detection of Bruises on Red Apples Using Deep Learning Models. Sci. Hortic. 2024, 329, 113021. [Google Scholar] [CrossRef]
- Furmonas, J.; Liobe, J.; Barzdenas, V. Analytical Review of Event-Based Camera Depth Estimation Methods and Systems. Sensors 2022, 22, 1201. [Google Scholar] [CrossRef]
- Hidalgo-Carrio, J.; Gehrig, D.; Scaramuzza, D. Learning Monocular Dense Depth from Events. In Proceedings of the 2020 International Conference on 3D Vision (3DV), Fukuoka, Japan, 25–28 November 2020; pp. 534–542. [Google Scholar] [CrossRef]
- Fan, L.; Li, Y.; Jiang, C.; Wu, Y. Unsupervised Depth Completion and Denoising for RGB-D Sensors. In Proceedings of the 2022 International Conference on Robotics and Automation (ICRA), Philadelphia, PA, USA, 23–27 May 2022; pp. 8734–8740. [Google Scholar] [CrossRef]
- Miranda, J.C.; Arnó, J.; Gené-Mola, J.; Lordan, J.; Asín, L.; Gregorio, E. Assessing Automatic Data Processing Algorithms for RGB-D Cameras to Predict Fruit Size and Weight in Apples. Comput. Electron. Agric. 2023, 214, 108302. [Google Scholar] [CrossRef]
- Guo, S.; Yoon, S.C.; Li, L.; Wang, W.; Zhuang, H.; Wei, C.; Liu, Y.; Li, Y. Recognition and Positioning of Fresh Tea Buds Using YOLOv4-Lighted + ICBAM Model and RGB-D Sensing. Agriculture 2023, 13, 518. [Google Scholar] [CrossRef]
- Osvaldová, K.; Gajdošech, L.; Kocur, V.; Madaras, M. Enhancement of 3D Camera Synthetic Training Data with Noise Models. arXiv 2024, arXiv:2402.16514. [Google Scholar]
- Hou, C.; Qiao, T.; Dong, H.; Wu, H. Coal Flow Volume Detection Method for Conveyor Belt Based on TOF Vision. Meas. J. Int. Meas. Confed. 2024, 229, 114468. [Google Scholar] [CrossRef]
- Horaud, R.; Hansard, M.; Evangelidis, G.; Ménier, C. An Overview of Depth Cameras and Range Scanners Based on Time-of-Flight Technologies. Mach. Vis. Appl. 2016, 27, 1005–1020. [Google Scholar] [CrossRef]
- Condotta, I.C.F.S.; Brown-Brandl, T.M.; Pitla, S.K.; Stinn, J.P.; Silva-Miranda, K.O. Evaluation of Low-Cost Depth Cameras for Agricultural Applications. Biol. Syst. Eng. 2020, 173, 105394. [Google Scholar] [CrossRef]
- Zhu, X.-F.; Xu, T.; Wu, X.-J. Adaptive Colour-Depth Aware Attention for RGB-D Object Tracking. IEEE Signal Process. Lett. 2024; early access. [Google Scholar] [CrossRef]
- Mac, T.T.; Lin, C.Y.; Huan, N.G.; Nhat, L.D.; Hoang, P.C.; Hai, H.H. Hybrid Slam-Based Exploration of a Mobile Robot for 3d Scenario Reconstruction and Autonomous Navigation. Acta Polytech. Hung. 2021, 18, 197–212. [Google Scholar] [CrossRef]
- Gomez-Rosal, D.A.; Bergau, M.; Fischer, G.K.J.; Wachaja, A.; Grater, J.; Odenweller, M.; Piechottka, U.; Hoeflinger, F.; Gosala, N.; Wetzel, N.; et al. A Smart Robotic System for Industrial Plant Supervision. In Proceedings of the 2023 IEEE SENSORS, Vienna, Austria, 29 October–1 November 2023; pp. 1–13. [Google Scholar] [CrossRef]
- Huang, X.; Dong, X.; Ma, J.; Liu, K.; Ahmed, S.; Lin, J.; Qiu, B. The Improved A* Obstacle Avoidance Algorithm for the Plant Protection UAV with Millimeter Wave Radar and Monocular Camera Data Fusion. Remote Sens. 2021, 13, 3364. [Google Scholar] [CrossRef]
- Chaki, N.; Devarakonda, N.; Cortesi, A.; Seetha, H. Proceedings of International Conference on Computational Intelligence and Data Engineering: ICCIDE 2021 (Lecture Notes on Data Engineering and Communications Technologies); Springer: Berlin/Heidelberg, Germany, 2021; ISBN 9789811671814. [Google Scholar]
- Saucedo, M.A.V.; Patel, A.; Sawlekar, R.; Saradagi, A.; Kanellakis, C.; Agha-Mohammadi, A.A.; Nikolakopoulos, G. Event Camera and LiDAR Based Human Tracking for Adverse Lighting Conditions in Subterranean Environments. IFAC-PapersOnLine 2023, 56, 9257–9262. [Google Scholar] [CrossRef]
- Le, N.M.D.; Nguyen, N.H.; Nguyen, D.A.; Ngo, T.D.; Ho, V.A. ViART: Vision-Based Soft Tactile Sensing for Autonomous Robotic Vehicles. IEEE/ASME Trans. Mechatron. 2023, 29, 1420–1430. [Google Scholar] [CrossRef]
- Cai, Y.; Ou, Y.; Qin, T. Improving SLAM Techniques with Integrated Multi-Sensor Fusion for 3D Reconstruction. Sensors 2024, 24, 2033. [Google Scholar] [CrossRef] [PubMed]
- Lang, X.; Li, L.; Zhang, H.; Xiong, F.; Xu, M.; Liu, Y.; Zuo, X.; Lv, J. Gaussian-LIC: Photo-Realistic LiDAR-Inertial-Camera SLAM with 3D Gaussian Splatting. arXiv 2024, arXiv:2404.06926. [Google Scholar]
- Bhattacharjee, T.; Shenoi, A.A.; Park, D.; Rehg, J.M.; Kemp, C.C. Combining Tactile Sensing and Vision for Rapid Haptic Mapping. In Proceedings of the 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Hamburg, Germany, 28 September–2 October 2015; pp. 1200–1207. [Google Scholar] [CrossRef]
- Álvarez, D.; Roa, M.A.; Moreno, L. Visual and Tactile Fusion for Estimating the Pose of a Grasped Object. Adv. Intell. Syst. Comput. 2020, 1093 AISC, 184–198. [Google Scholar] [CrossRef]
- Naga, P.S.B.; Hari, P.J.; Sinduja, R.; Prathap, S.; Ganesan, M. Realization of SLAM and Object Detection Using Ultrasonic Sensor and RGB-HD Camera. In Proceedings of the 2022 International Conference on Wireless Communications Signal Processing and Networking (WiSPNET), Chennai, India, 24–26 March 2022; pp. 167–171. [Google Scholar]
- Chen, X.; Wang, S.; Zhang, B.; Luo, L. Multi-Feature Fusion Tree Trunk Detection and Orchard Mobile Robot Localization Using Camera/Ultrasonic Sensors. Comput. Electron. Agric. 2018, 147, 91–108. [Google Scholar] [CrossRef]
- Lin, Z.; Gao, Z.; Chen, B.M.; Chen, J.; Li, C. Accurate LiDAR-Camera Fused Odometry and RGB-Colored Mapping. IEEE Robot. Autom. Lett. 2024, 9, 2495–2502. [Google Scholar] [CrossRef]
- You, H.; Xu, F.; Ye, Y.; Xia, P.; Du, J. Adaptive LiDAR Scanning Based on RGB Information. Autom. Constr. 2024, 160, 105337. [Google Scholar] [CrossRef]
- Jing, J. Simulation Analysis of Fire-Fighting Path Planning Based On SLAM. Highlights Sci. Eng. Technol. 2024, 85, 434–442. [Google Scholar] [CrossRef]
- Tan, C.J.; Ogawa, S.; Hayashi, T.; Janthori, T.; Tominaga, A.; Hayashi, E. 3D Semantic Mapping Based on RGB-D Camera and LiDAR Sensor in Beach Environment. In Proceedings of the 2024 1st International Conference on Robotics, Engineering, Science, and Technology (RESTCON), Pattaya, Thailand, 16–18 February 2024; pp. 21–26. [Google Scholar]
- Qiao, G.; Ning, N.; Zuo, Y.; Zhou, P.; Sun, M.; Hu, S.; Yu, Q.; Liu, Y. Spatio-Temporal Fusion Spiking Neural Network for Frame-Based and Event-Based Camera Sensor Fusion. IEEE Trans. Emerg. Top. Comput. Intell. 2024, 8, 2446–2456. [Google Scholar] [CrossRef]
- Zuo, Y.F.; Xu, W.; Wang, X.; Wang, Y.; Kneip, L. Cross-Modal Semidense 6-DOF Tracking of an Event Camera in Challenging Conditions. IEEE Trans. Robot. 2024, 40, 1600–1616. [Google Scholar] [CrossRef]
- Yadav, R.; Vierling, A.; Berns, K. Radar+RGB Attentive Fusion for Robust Object Detection in Autonomous Vehicles. arXiv 2020, arXiv:2008.13642. [Google Scholar]
- Yao, S.; Guan, R.; Huang, X.; Li, Z.; Sha, X.; Yue, Y.; Lim, E.G.; Seo, H.; Man, K.L.; Zhu, X.; et al. Radar-Camera Fusion for Object Detection and Semantic Segmentation in Autonomous Driving: A Comprehensive Review. IEEE Trans. Intell. Veh. 2024, 9, 2094–2128. [Google Scholar] [CrossRef]
- Aeberhard, M.; Kaempchen, N. High-Level Sensor Data Fusion Architecture for Vehicle Surround Environment Perception. 2015. Available online: https://www.researchgate.net/publication/267725657 (accessed on 10 December 2024).
- Thakur, A.; Mishra, S.K. An In-Depth Evaluation of Deep Learning-Enabled Adaptive Approaches for Detecting Obstacles Using Sensor-Fused Data in Autonomous Vehicles. Eng. Appl. Artif. Intell. 2024, 133, 108550. [Google Scholar] [CrossRef]
- Rövid, A.; Remeli, V.; Szalay, Z. Raw Fusion of Camera and Sparse LiDAR for Detecting Distant Objects Fusion von Kameradaten Und Spärlichem LiDAR-Rohsignal Zur Erkennung Entfernter Objekte. At-Automatisierungstechnik 2020, 68, 337–346. [Google Scholar] [CrossRef]
- Li, F.; Li, W.; Chen, W.; Xu, W.; Huang, L.; Li, D.; Cai, S.; Yang, M.; Xiong, X.; Liu, Y. A Mobile Robot Visual SLAM System with Enhanced Semantics Segmentation. IEEE Access 2020, 8, 25442–25458. [Google Scholar] [CrossRef]
- Hassani, S.; Dackermann, U.; Mousavi, M.; Li, J. A Systematic Review of Data Fusion Techniques for Optimized Structural Health Monitoring. Inf. Fusion 2024, 103, 102136. [Google Scholar] [CrossRef]
- Thakur, A.; Pachamuthu, R. LiDAR and Camera Raw Data Sensor Fusion in Real-Time for Obstacle Detection. In Proceedings of the 2023 IEEE Sensors Applications Symposium (SAS), Ottawa, ON, Canada, 18–20 July 2023. [Google Scholar]
- Risti, D.; Gao, G.; Leu, A. Low-Level Sensor Fusion-Based Human Tracking. Ristić-Durrant 2016, 15, 17–32. [Google Scholar]
- Puriyanto, R.D.; Mustofa, A.K.; Dahlan, U.A.; Author, C. Design and Implementation of Fuzzy Logic for Obstacle Avoidance in Differential Drive Mobile Robot. J. Robot. Control. JRC 2024, 5, 132–141. [Google Scholar] [CrossRef]
- Kim, T.; Member, S.; Kang, G.; Member, S. Development of an Indoor Delivery Mobile Robot for a Multi-Floor Environment. IEEE Access 2024, 12, 45202–45215. [Google Scholar] [CrossRef]
- Azhar, G.A.; Kusuma, A.C.; Izza, S. Differential Drive Mobile Robot Motion Accuracy Improvement with Odometry-Compass Sensor Fusion Implementation. ELKHA 2023, 15, 24–31. [Google Scholar]
- Li, J.; Liu, Y.; Wang, S.; Wang, L.; Sun, Y.; Li, X. Visual Perception System Design for Rock Breaking Robot Based on Multi-Sensor Fusion. Multimed. Tools Appl. 2024, 83, 24795–24814. [Google Scholar] [CrossRef]
- Park, G. Optimal Vehicle Position Estimation Using Adaptive Unscented Kalman Filter Based on Sensor Fusion. Mechatronics 2024, 99, 103144. [Google Scholar] [CrossRef]
- Jiang, P.; Hu, C.; Wang, T.; Lv, K.; Guo, T.; Jiang, J.; Hu, W. Research on a Visual/Ultra-Wideband Tightly Coupled Fusion Localization Algorithm. Sensors 2024, 24, 1710. [Google Scholar] [CrossRef]
- Hu, K.; Chen, Z.; Kang, H.; Tang, Y. 3D Vision Technologies for a Self-Developed Structural External Crack Damage Recognition Robot. Autom. Constr. 2024, 159, 105262. [Google Scholar] [CrossRef]
- Sarmento, J.; Neves dos Santos, F.; Silva Aguiar, A.; Filipe, V.; Valente, A. Fusion of Time-of-Flight Based Sensors with Monocular Cameras for a Robotic Person Follower. J. Intell. Robot. Syst. Theory Appl. 2024, 110, 30. [Google Scholar] [CrossRef]
- Zheng, X.; Ji, S.; Pan, Y.; Zhang, K.; Wu, C. NeurlT: Pushing the Limit of Neural Inertial Tracking for Indoor Robotic IoT. arXiv 2024, arXiv:2404.08939. [Google Scholar]
- Li, C.; Chen, K.; Li, H.; Luo, H. Engineering Applications of Artificial Intelligence Multisensor Data Fusion Approach for Sediment Assessment of Sewers in Operation. Eng. Appl. Artif. Intell. 2024, 132, 107965. [Google Scholar] [CrossRef]
- Gao, H.; Li, X.; Song, X. A Fusion Strategy for Vehicle Positioning at Intersections Utilizing UWB and Onboard Sensors. Sensors 2024, 24, 476. [Google Scholar] [CrossRef] [PubMed]
- Ming, Z.; Berrio, J.S.; Shan, M.; Worrall, S. OccFusion: A Straightforward and Effective Multi-Sensor Fusion Framework for 3D Occupancy Prediction. arXiv 2024, arXiv:2403.01644v3. [Google Scholar]
- Kocic, J.; Jovicic, N.; Drndarevic, V. Sensors and Sensor Fusion in Autonomous Vehicles. In Proceedings of the 2018 26th Telecommunications Forum (TELFOR), Belgrade, Serbia, 20–21 November 2018. [Google Scholar] [CrossRef]
- Luo, R.C.; Chang, N.W.; Lin, S.C.; Wu, S.C. Human Tracking and Following Using Sensor Fusion Approach for Mobile Assistive Companion Robot. In Proceedings of the 2009 35th Annual Conference of IEEE Industrial Electronics, Porto, Portugal, 3–5 November 2009; pp. 2235–2240. [Google Scholar] [CrossRef]
- Khodarahmi, M.; Maihami, V. A Review on Kalman Filter Models. Arch. Comput. Methods Eng. 2023, 30, 727–747. [Google Scholar] [CrossRef]
- Arpitha Shankar, S.I.; Shivakumar, M. Sensor Fusion Based Multiple Robot Navigation in an Indoor Environment. Int. J. Interact. Des. Manuf. 2024, 18, 4841–4852. [Google Scholar] [CrossRef]
- Yara, R.; Konstantinos, T.; Roland, H.; John, C.; Eleni, C.; Markus, R. Unscented Kalman Filter–Based Fusion of GNSS, Accelerometer, and Rotation Sensors for Motion Tracking. J. Struct. Eng. 2024, 150, 5024002. [Google Scholar] [CrossRef]
- Nguyen, T.; Mann, G.K.I.; Vardy, A.; Gosine, R.G. CKF-Based Visual Inertial Odometry for Long-Term Trajectory Operations. J. Robot. 2020, 2020, 7362952. [Google Scholar] [CrossRef]
- Liu, C.; Li, H.; Wang, Z. FastTrack: A Highly Efficient and Generic GPU-Based Multi-Object Tracking Method with Parallel Kalman Filter. Int. J. Comput. Vis. 2023, 132, 1463–1483. [Google Scholar] [CrossRef]
- Damsgaard, B.; Gaasdal, S.S.; Bonnerup, S. Multi-Sensor Fusion with Radar and Ultrasound for Obstacle Avoidance on Capra Hircus 1.0. In Proceedings of the 11th Student Symposium on Mechanical and Manufacturing Engineering, Parit Raja, Malaysia, 25–26 August 2021; pp. 1–8. [Google Scholar]
- Cai, Y.; Qin, T.; Ou, Y.; Wei, R. Intelligent Systems in Motion: A Comprehensive Review on Multi-Sensor Fusion and Information Processing From Sensing to Navigation in Path Planning. Int. J. Semant. Web Inf. Syst. 2023, 19, 1–35. [Google Scholar] [CrossRef]
- Jamil, H.; Jian, Y. An Evolutionary Enhanced Particle Filter-Based Fusion Localization Scheme for Fast Tracking of Smartphone Users in Tall Complex Buildings for Hazardous Situations. IEEE Sens. J. 2024, 24, 6799–6812. [Google Scholar] [CrossRef]
- Tang, Q.; Liang, J.; Zhu, F. A Comparative Review on Multi-Modal Sensors Fusion Based on Deep Learning. Signal Process. 2023, 213, 109165. [Google Scholar] [CrossRef]
- Thepsit, T.; Konghuayrob, P.; Saenthon, A.; Yanyong, S. Localization for Outdoor Mobile Robot Using LiDAR and RTK-GNSS/INS. Sensors Mater. 2024, 36, 1405–1418. [Google Scholar] [CrossRef]
- Jiang, H.; Lu, Y.; Zhang, D.; Shi, Y.; Wang, J. Deep Learning-Based Fusion Networks with High-Order Attention Mechanism for 3D Object Detection in Autonomous Driving Scenarios. Appl. Soft Comput. 2024, 152, 111253. [Google Scholar] [CrossRef]
- Fu, G.; Menciassi, A.; Dario, P. Development of a Low-Cost Active 3D Triangulation Laser Scanner for Indoor Navigation of Miniature Mobile Robots. Rob. Auton. Syst. 2012, 60, 1317–1326. [Google Scholar] [CrossRef]
- KlanÄnik, S.; BaliÄ, J.; PlaninšiÄ, P. Obstacle Detection with Active Laser Triangulation. Adv. Prod. Eng. Manag. 2007, 2, 79–90. [Google Scholar]
- França, J.G.D.M.; Gazziro, M.A.; Ide, A.N.; Saito, J.H. A 3D Scanning System Based on Laser Triangulation and Variable Field of View. In Proceedings of the IEEE International Conference on Image Processing 2005, Genova, Italy, 14 September 2005; Volume 1, pp. 425–428. [Google Scholar] [CrossRef]
- Fu, G.; Corradi, P.; Menciassi, A.; Dario, P. An Integrated Triangulation Laser Scanner for Obstacle Detection of Miniature Mobile Robots in Indoor Environment. IEEE/ASME Trans. Mechatron. 2011, 16, 778–783. [Google Scholar] [CrossRef]
- Schlarp, J.; Csencsics, E.; Schitter, G. Design and Evaluation of an Integrated Scanning Laser Triangulation Sensor. Mechatronics 2020, 72, 102453. [Google Scholar] [CrossRef]
- Ding, D.; Ding, W.; Huang, R.; Fu, Y.; Xu, F. Research Progress of Laser Triangulation On-Machine Measurement Technology for Complex Surface: A Review. Meas. J. Int. Meas. Confed. 2023, 216, 113001. [Google Scholar] [CrossRef]
- So, E.W.Y.; Munaro, M.; Michieletto, S.; Antonello, M.; Menegatti, E. Real-Time 3D Model Reconstruction with a Dual-Laser Triangulation System for Assembly Line Completeness Inspection. Adv. Intell. Syst. Comput. 2013, 194 AISC, 707–716. [Google Scholar] [CrossRef]
- Dong, H.; Weng, C.Y.; Guo, C.; Yu, H.; Chen, I.M. Real-Time Avoidance Strategy of Dynamic Obstacles via Half Model-Free Detection and Tracking with 2D Lidar for Mobile Robots. IEEE/ASME Trans. Mechatron. 2021, 26, 2215–2225. [Google Scholar] [CrossRef]
- Gul, F.; Rahiman, W.; Nazli Alhady, S.S. A Comprehensive Study for Robot Navigation Techniques. Cogent Eng. 2019, 6, 1632046. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).