A Deep Learning-Driven Semantic Mapping Strategy for Robotic Inspection of Desalination Facilities
Abstract
1. Introduction
- Investigate and analyze the limitations of existing semantic navigation frameworks.
- Design and construct a specialized multimodal dataset representing real desalination plant conditions.
- Develop an intelligent semantic mapping system that fuses LiDAR-based geometry, RGB-D vision, and odometry data to automatically detect, label, and localize critical elements within desalination facilities.
- Implement and validate the proposed system within the ROS environment, demonstrating its ability to perform semantic mapping in support of industrial monitoring.
2. Related Work
2.1. LiDAR-Based Systems
2.2. Vision-Based Systems
2.3. Hybrid Systems
3. Semantic Map Production System
3.1. Operational Stages of the Developed System
3.2. Sensor Fusion-Based System for Semantic Mapping
3.3. Desalination Plant Dataset
3.4. Object Detection Subsystem
| Algorithm 1: Semantic map production |
| 01: Define w, h as the 2D navigation area with width = w and height = h 02: Define mr as the mobile robot in the desalination plant environment 03: Define mr(x, y) as the current 2D location of mr obtained from wheel odometry 04: Define mr_maxX as the farthest visited point along the x-axis 05: Define mr_maxY as the farthest visited point along the y-axis 06: Define yoloD as the custom-trained object detection model (desalination dataset) 07: Define depth_to_object(k) as the depth distance (in meters) to detected object k 08: Define navigation_fun() as the navigation function in the area of interest 09: Define semantic_table as a data structure storing objects and their 2D coordinates 10: while (h< mr_maxX < w): 11: while (obs_dist > 100): //ensure safe distance from obstacles 12: if object_detected(yoloD, depth_to_object(k)) == True: 13: semantic_table.add(detected_object, mr(x, y)) 14: Else navigation_fun() 15: end while 16: end while |
4. Experimental Results
4.1. Development Environment
4.2. Results
- Model classification efficiency: This refers to the evaluation of object classification models based on key performance metrics, including precision, recall, accuracy, and mean average precision (mAP). Precision measures how many times the model correctly detected objects. Recall refers to how many of the actual objects present in the images were correctly detected by the model. The classification accuracy refers to how often the classification model assigns the correct class label to the object it detects. Finally, the Mean Average Precision (mAP) is used to measure the performance of the implemented object detection models.
- Recognized objects ratio (): This refers to the ratio of the overall number of objects that has been correctly classified in the desalination plant environment with comparison to the total existing objects in the same area, as presented below:where j is the index number of a certain class, y is the total number of detected objects in the kth class, m is the total number of existing objects in the kth class, and n is the total number of objects in the simulated environment.
- Localization Error (LE) of detected objects: This estimates the average localization error between the estimated 2D location of an object and the real 2D position of that object, using the Euclidian distance formula.
- Map construction accuracy: This refers to the accuracy of the semantic map generated by integrating LiDAR data with object detection outputs. Hence, by combining geometric information from LiDAR with identified objects and their locations, a comprehensive semantic map is produced that precisely represents both the environment’s structure and its semantic elements.
5. Discussion
6. Conclusions & Future Work
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Liang, C.J.; Cheng, M.H. Trends in robotics research in occupational safety and health: A scientometric analysis and review. Int. J. Environ. Res. Public Health 2023, 20, 5904. [Google Scholar] [CrossRef] [PubMed]
- Song, W.K.; Song, W.J.; Kim, Y.; Kim, J. Usability test of KNRC self-feeding robot. In Proceedings of the 2013 IEEE 13th International Conference on Rehabilitation Robotics (ICORR), Seattle, WA, USA, 24–26 June 2013; pp. 1–5. [Google Scholar]
- Bemelmans, R.; Gelderblom, G.J.; Jonker, P.; de Witte, L. Socially assistive robots in elderly care: A systematic review into effects and electiveness. J. Am. Med. Dir. Assoc. 2012, 13, 112–114. [Google Scholar] [CrossRef] [PubMed]
- Kanda, T.; Shiomi, M.; Miyashita, Z.; Ishiguro, H.; Hagita, N. An active guide robot in a shopping mall. In Proceedings of the 4th ACM/IEEE International Conference on Human-Robot Interaction, HRI’09, La Jolla, CA, USA, 11–13 March 2009. [Google Scholar]
- Salman, R.; Willms, I. A mobile security robot equipped with UWB-radar for super-resolution indoor positioning and localisation applications. In Proceedings of the 2012 International Conference on Indoor Positioning and Indoor Navigation, IPIN 2012—Conference Proceedings, Sydney, Australia, 13–15 November 2012. [Google Scholar]
- Lowry, S.; Sünderhauf, N.; Newman, P.; Leonard, J.J.; Cox, D.; Corke, P.; Milford, M.J. Visual place recognition: A survey. IEEE Trans. Robot. 2015, 32, 1–19. [Google Scholar] [CrossRef]
- Alterovitz, R.; Koenig, S.; Likhachev, M. Robot planning in the real world: Research challenges and opportunities. AI Mag. 2016, 37, 76–84. [Google Scholar] [CrossRef]
- Radhakrishnan, S.; Gueaieb, W. A state-of-the-art review on topology and differential geometry-based robotic path planning—Part I: Planning under static constraints. Int. J. Intell. Robot. Appl. 2024, 8, 435–454. [Google Scholar] [CrossRef]
- Pendleton, S.D.; Andersen, H.; Du, X.; Shen, X.; Meghjani, M.; Eng, Y.H.; Rus, D.; Ang, M.H. Perception, planning, control, and coordination for autonomous vehicles. Machines 2017, 5, 6. [Google Scholar] [CrossRef]
- Urrea, C.; Kern, J. Recent Advances and Challenges in Industrial Robotics: A Systematic Review of Technological Trends and Emerging Applications. Processes 2025, 13, 832. [Google Scholar] [CrossRef]
- Konieczna-Fuławka, M.; Koval, A.; Nikolakopoulos, G.; Fumagalli, M.; Santas Moreu, L.; Vigara-Puche, V.; Müller, J.; Prenner, M. Autonomous mobile inspection robots in deep underground mining—The current state of the art and future perspectives. Sensors 2025, 25, 3598. [Google Scholar] [CrossRef]
- Bretschneider, L.; Bollmann, S.; Houssin-Agbomson, D.; Shaw, J.; Howes, N.; Nguyen, L.; Robinson, R.; Helmore, J.; Lichtenstern, M.; Nwaboh, J.; et al. Concepts for drone based pipeline leak detection. Front. Robot. AI 2024, 11, 1426206. [Google Scholar] [CrossRef]
- Long, L.; Cao, S.; Rao, F.; Fu, Q.; Leng, Y.; Xu, Q.; Zhang, X. Research and experiment of a novel leak detection robot. Ind. Robot. Int. J. Robot. Res. Appl. 2025, 52, 914–922. [Google Scholar] [CrossRef]
- Alqobali, R.; Alshmrani, M.; Alnasser, R.; Rashidi, A.; Alhmiedat, T.; Alia, O.M.D. A survey on robot semantic navigation systems for indoor environments. Appl. Sci. 2023, 14, 89. [Google Scholar] [CrossRef]
- Hopkinson, C.; Chasmer, L.; Gynan, C.; Mahoney, C.; Sitar, M. Multisensor and multispectral lidar characterization and classification of a forest environment. Can. J. Remote Sens. 2016, 42, 501–520. [Google Scholar] [CrossRef]
- McDaniel, M.W.; Nishihata, T.; Brooks, C.A.; Iagnemma, K. Ground plane identification using LIDAR in forested environments. In Proceedings of the 2010 IEEE International Conference on Robotics and Automation, Anchorage, AK, USA, 3–7 May 2010; pp. 3831–3836. [Google Scholar]
- Sun, N.; Yang, E.; Corney, J.; Chen, Y. Semantic path planning for indoor navigation and household tasks. In Towards Autonomous Robotic Systems: 20th Annual Conference, TAROS 2019, London, UK, 3–5 July 2019, Proceedings, Part II 20; Springer International Publishing: Berlin/Heidelberg, Germany, 2019; pp. 191–201. [Google Scholar]
- Álvarez-Aparicio, C.; Guerrero-Higueras, A.M.; Rodríguez-Lera, F.J.; Ginés Clavero, J.; Martín Rico, F.; Matellán, V. People detection and tracking using LIDAR sensors. Robotics 2019, 8, 75. [Google Scholar] [CrossRef]
- Dewan, A.; Oliveira, G.L.; Burgard, W. Deep semantic classification for 3d lidar data. In Proceedings of the 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Vancouver, BC, Canada, 24–28 September 2017; pp. 3544–3549. [Google Scholar]
- Alenzi, Z.; Alenzi, E.; Alqasir, M.; Alruwaili, M.; Alhmiedat, T.; Alia, O.M.D. A Semantic Classification Approach for Indoor Robot Navigation. Electronics 2022, 11, 2063. [Google Scholar] [CrossRef]
- Joo, S.H.; Manzoor, S.; Rocha, Y.G.; Bae, S.H.; Lee, K.H.; Kuc, T.Y.; Kim, M. Autonomous navigation framework for intelligent robots based on a semantic environment modeling. Appl. Sci. 2020, 10, 3219. [Google Scholar] [CrossRef]
- Zhang, J.; Wang, W.; Qi, X.; Liao, Z. Social and robust navigation for indoor robots based on object semantic grid and topological map. Appl. Sci. 2020, 10, 8991. [Google Scholar] [CrossRef]
- Banerjee, S.; Bhowmick, B.; Roychoudhury, R.D. Object goal navigation based on semantics and rgb ego view. arXiv 2022, arXiv:2210.11543. [Google Scholar] [CrossRef]
- Qi, X.; Wang, W.; Yuan, M.; Wang, Y.; Li, M.; Xue, L.; Sun, Y. Building semantic grid maps for domestic robot navigation. Int. J. Adv. Robot. Syst. 2020, 17, 1729881419900066. [Google Scholar] [CrossRef]
- Wei, S.; Wang, S.; Li, H.; Liu, G.; Yang, T.; Liu, C. A Semantic Information-Based Optimized vSLAM in Indoor Dynamic Environments. Appl. Sci. 2023, 13, 8790. [Google Scholar] [CrossRef]
- Liang, J.; Song, W.; Shen, L.; Zhang, Y. Indoor semantic map building for robot navigation. In Proceedings of the 2019 IEEE 8th Joint International Information Technology and Artificial Intelligence Conference (ITAIC), Chongqing, China, 24–26 May 2019; pp. 794–798. [Google Scholar]
- Crespo, J.; Barber, R.; Mozos, O.M. Relational model for robotic semantic navigation in indoor environments. J. Intell. Robot. Syst. 2017, 86, 617–639. [Google Scholar] [CrossRef]
- Zhao, X.; Zuo, T.; Hu, X. OFM-SLAM: A visual semantic SLAM for dynamic indoor environments. Math. Probl. Eng. 2021, 2021, 5538840. [Google Scholar] [CrossRef]
- Kowalewski, S.; Maurin, A.L.; Andersen, J.C. Semantic mapping and object detection for indoor mobile robots. IOP Conf. Ser. Mater. Sci. Eng. 2019, 517, 012012. [Google Scholar] [CrossRef]
- Kim, W.; Seok, J. Indoor semantic segmentation for robot navigating on mobile. In Proceedings of the 2018 Tenth International Conference on Ubiquitous and Future Networks (ICUFN), Prague, Czech Republic, 3–6 July 2018; pp. 22–25. [Google Scholar]
- Dang, T.V.; Bui, N.T. Multi-scale fully convolutional network-based semantic segmentation for mobile robot navigation. Electronics 2023, 12, 533. [Google Scholar] [CrossRef]
- Deng, W.; Huang, K.; Chen, X.; Zhou, Z.; Shi, C.; Guo, R.; Zhang, H. Semantic RGB-D SLAM for rescue robot navigation. IEEE Access 2020, 8, 221320–221329. [Google Scholar] [CrossRef]
- Wu, Y.; Zhang, Y.; Zhu, D.; Deng, Z.; Sun, W.; Chen, X.; Zhang, J. An object slam framework for association, mapping, and high-level tasks. IEEE Trans. Robot. 2023, 39, 2912–2932. [Google Scholar] [CrossRef]
- Rollo, F.; Raiola, G.; Zunino, A.; Tsagarakis, N.; Ajoudani, A. Artifacts mapping: Multi-modal semantic mapping for object detection and 3d localization. In Proceedings of the 2023 European Conference on Mobile Robots (ECMR), Coimbra, Portugal, 4–7 September 2023; pp. 1–8. [Google Scholar]
- Zhao, Y.; Xiao, P.; Wang, J.; Guo, R. Semi-Autonomous Navigation Based on Local Semantic Map for Mobile Robot. J. Shanghai Jiaotong Univ. Sci. 2025, 30, 27–33. [Google Scholar] [CrossRef]
- Maggio, D.; Chang, Y.; Hughes, N.; Trang, M.; Griffith, D.; Dougherty, C.; Cristofalo, E.; Schmid, L.; Carlone, L. Clio: Real-time task-driven open-set 3d scene graphs. IEEE Robot. Autom. Lett. 2024, 9, 8921–8928. [Google Scholar] [CrossRef]
- Aitken, J.M.; Evans, M.H.; Worley, R.; Edwards, S.; Zhang, R.; Dodd, T.; Mihaylova, L.; Anderson, S.R. Simultaneous localization and mapping for inspection robots in water and sewer pipe networks: A review. IEEE Access 2021, 9, 140173–140198. [Google Scholar] [CrossRef]
- Alotaibi, A.; Alatawi, H.; Binnouh, A.; Duwayriat, L.; Alhmiedat, T.; Alia, O.M.D. Deep Learning-Based Vision Systems for Robot Semantic Navigation: An Experimental Study. Technologies 2024, 12, 157. [Google Scholar] [CrossRef]
- Alqobali, R.; Alnasser, R.; Rashidi, A.; Alshmrani, M.; Alhmiedat, T. A Real-Time Semantic Map Production System for Indoor Robot Navigation. Sensors 2024, 24, 6691. [Google Scholar] [CrossRef]
- Alhmiedat, T.; Alia, O.M.D. Utilizing a deep neural network for robot semantic classification in indoor environments. Sci. Rep. 2025, 15, 21937. [Google Scholar] [CrossRef]


















| Parameter | Value |
|---|---|
| Total number of images | 765 |
| Image resolution | 640 × 640 |
| Number of classes | 3 |
| Overall dataset size | 29 MBytes |
| Image format | JPG |
| Class | Count | Percentage |
|---|---|---|
| Vertical pipe | 4777 | 50.77% |
| Horizontal pipe | 1067 | 11.34% |
| Pump | 3565 | 37.88% |
| Total | 9409 | 100.0% |
| Subset | Percentage | Count |
|---|---|---|
| Training subset | 80% | 612 |
| Testing subset | 10% | 76 |
| Validation subset | 10% | 77 |
| Subset | Percentage |
|---|---|
| Epochs | 100 |
| Batch size | 16 |
| Optimizer | SGD |
| Learning rate | 0.01 |
| Weight decay | 0.0005 |
| Subset | Vertical | Horizontal | Pump |
|---|---|---|---|
| Training Instances | 3873 | 815 | 2854 |
| Testing Instances | 445 | 131 | 371 |
| Validation Instances | 459 | 121 | 340 |
| Object | Simulated Quantity | Real Quantity |
|---|---|---|
| Vertical pipe | 27 | 6 |
| Horizontal pipe | 11 | 8 |
| Person | 3 | 2 |
| Table | 3 | - |
| Cone | 12 | - |
| Parameter | Specification |
|---|---|
| RGB-D camera | Oad-D-Pro (Luxonis, Denver, CO, USA) |
| RGB-D camera resolution | 1280 × 800 |
| LiDAR unit | RPLiDAR A1 (Shanghai, China) |
| LiDAR range | 8 m |
| Frame rate (FPS) | 2 |
| Inertial resolution | 150 CPR |
| Robot speed | 0.3 m/s |
| Raspberry Pi model | Raspberry Pi 4 4GB |
| LiDAR sample rate | 8000 |
| Category | Main Packages/Libraries | Function |
|---|---|---|
| Mapping & navigation | gmapping, cartographer, move_base, amcl | SLAM, localization & path planning |
| Sensor interfaces | rplidar_ros, realsense2_camera, robot_localization | LiDAR, RGB-D, & odometry integration |
| Vision & perception | cv_bridge, pcl_ros, ultralytics, image_transport | Object detection and depth processing |
| Data fusion | tf, tf2, message_filters | Coordinate transformation & synchronization |
| Visualization | rviz, rqt, plotjuggler | Visualization and debugging tools |
| Class | Precision | Recall | Accuracy | mAP@0.50 |
|---|---|---|---|---|
| Horizontal pipe | 71.70% | 76.40% | 70.53% | 63.50% |
| Vertical pipe | 57.80% | 53.40% | 52.97% | 47.70% |
| Pump | 36.50% | 75.00% | 57.37% | 60.00% |
| Average | 55.33% | 68.28% | 60.29% | 57.06% |
| Class | Precision | Recall | Accuracy | mAP@0.50 |
|---|---|---|---|---|
| Horizontal pipe | 90.00% | 91.10% | 91.00% | 88.00% |
| Vertical pipe | 80.00% | 83.00% | 81.00% | 90.00% |
| Pump | 84.00% | 88.50% | 86.00% | 87.00% |
| Average | 84.60% | 87.53% | 85.00% | 88.33% |
| Object Class | Exist | Detected | % of Detected |
|---|---|---|---|
| Horizontal pipe | 13 | 12 | 92.30% |
| Vertical pipe | 28 | 25 | 89.28% |
| Person | 4 | 4 | 100.0% |
| Table | 3 | 2 | 66.66% |
| Cone | 14 | 12 | 85.71% |
| Object Class | Total Detected | Average LE for Each Class (m) | SD (m) |
|---|---|---|---|
| Horizontal-pipe | 12 | 1.81 | 0.25 |
| Vertical-pipe | 25 | 1.17 | 0.22 |
| Person | 4 | 1.21 | 0.15 |
| Table | 2 | 1.42 | 0.13 |
| Cone | 12 | 1.96 | 0.19 |
| Average | 88.70% | 1.51 | 0.18 |
| Object Class | Exist | Detected | % of Detected |
|---|---|---|---|
| Horizontal pipe | 8 | 7 | 87.50% |
| Vertical pipe | 6 | 4 | 66.66% |
| Person | 2 | 2 | 100.0% |
| Object Class | Total Detected | Average LE for Each Class (m) | SD (m) |
|---|---|---|---|
| Horizontal-pipe | 7 | 1.95 | 0.36 |
| Vertical-pipe | 4 | 2.89 | 0.24 |
| Person | 2 | 1.39 | 0.06 |
| Average | 81.25% | 2.07 | 0.22 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Alotaibi, A.; Alrashidi, R.; Alatawi, H.; Duwayriat, L.; Binnouh, A.; Alhmiedat, T.; Al-Qerem, A. A Deep Learning-Driven Semantic Mapping Strategy for Robotic Inspection of Desalination Facilities. Machines 2025, 13, 1129. https://doi.org/10.3390/machines13121129
Alotaibi A, Alrashidi R, Alatawi H, Duwayriat L, Binnouh A, Alhmiedat T, Al-Qerem A. A Deep Learning-Driven Semantic Mapping Strategy for Robotic Inspection of Desalination Facilities. Machines. 2025; 13(12):1129. https://doi.org/10.3390/machines13121129
Chicago/Turabian StyleAlotaibi, Albandari, Reem Alrashidi, Hanan Alatawi, Lamaa Duwayriat, Aseel Binnouh, Tareq Alhmiedat, and Ahmad Al-Qerem. 2025. "A Deep Learning-Driven Semantic Mapping Strategy for Robotic Inspection of Desalination Facilities" Machines 13, no. 12: 1129. https://doi.org/10.3390/machines13121129
APA StyleAlotaibi, A., Alrashidi, R., Alatawi, H., Duwayriat, L., Binnouh, A., Alhmiedat, T., & Al-Qerem, A. (2025). A Deep Learning-Driven Semantic Mapping Strategy for Robotic Inspection of Desalination Facilities. Machines, 13(12), 1129. https://doi.org/10.3390/machines13121129

