Next Article in Journal
GAPO: A Graph Attention-Based Reinforcement Learning Algorithm for Congestion-Aware Task Offloading in Multi-Hop Vehicular Edge Computing
Previous Article in Journal
Analytical Investigation of Primary Waveform Distortion Effect on Magnetic Flux Density in the Magnetic Core of Inductive Current Transformer and Its Transformation Accuracy
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Smart Safety Helmets with Integrated Vision Systems for Industrial Infrastructure Inspection: A Comprehensive Review of VSLAM-Enabled Technologies

by
Emmanuel A. Merchán-Cruz
1,*,
Samuel Moveh
1,
Oleksandr Pasha
2,
Reinis Tocelovskis
2,
Alexander Grakovski
1,
Alexander Krainyukov
1,
Nikita Ostrovenecs
1,
Ivans Gercevs
1 and
Vladimirs Petrovs
1
1
Engineering Faculty, Transport and Telecommunication Institute, Lauvas Iela 2, LV-1019 Riga, Latvia
2
3D Engineering SIA, Mežu Iela 41-33, LV-3405 Liepaja, Latvia
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(15), 4834; https://doi.org/10.3390/s25154834
Submission received: 2 July 2025 / Revised: 1 August 2025 / Accepted: 4 August 2025 / Published: 6 August 2025

Abstract

Smart safety helmets equipped with vision systems are emerging as powerful tools for industrial infrastructure inspection. This paper presents a comprehensive state-of-the-art review of such VSLAM-enabled (Visual Simultaneous Localization and Mapping) helmets. We surveyed the evolution from basic helmet cameras to intelligent, sensor-fused inspection platforms, highlighting how modern helmets leverage real-time visual SLAM algorithms to map environments and assist inspectors. A systematic literature search was conducted targeting high-impact journals, patents, and industry reports. We classify helmet-integrated camera systems into monocular, stereo, and omnidirectional types and compare their capabilities for infrastructure inspection. We examine core VSLAM algorithms (feature-based, direct, hybrid, and deep-learning-enhanced) and discuss their adaptation to wearable platforms. Multi-sensor fusion approaches integrating inertial, LiDAR, and GNSS data are reviewed, along with edge/cloud processing architectures enabling real-time performance. This paper compiles numerous industrial use cases, from bridges and tunnels to plants and power facilities, demonstrating significant improvements in inspection efficiency, data quality, and worker safety. Key challenges are analyzed, including technical hurdles (battery life, processing limits, and harsh environments), human factors (ergonomics, training, and cognitive load), and regulatory issues (safety certification and data privacy). We also identify emerging trends, such as semantic SLAM, AI-driven defect recognition, hardware miniaturization, and collaborative multi-helmet systems. This review finds that VSLAM-equipped smart helmets offer a transformative approach to infrastructure inspection, enabling real-time mapping, augmented awareness, and safer workflows. We conclude by highlighting current research gaps, notably in standardizing systems and integrating with asset management, and provide recommendations for industry adoption and future research directions.

1. Introduction

Aging infrastructure and complex industrial assets require regular inspections to ensure safety and reliability [1]. Traditional inspection methods are often labor-intensive, time-consuming, and potentially hazardous to personnel (e.g., climbing structures or entering confined spaces) [2,3]. There is a growing need for more efficient and safer inspection techniques. In recent years, wearable technology, particularly smart safety helmets with integrated sensors, has garnered attention as a means to augment inspectors’ capabilities and reduce risks [4,5,6]. These helmets serve as advanced personal protective equipment (PPE) that not only shield the worker but also embed cameras and other sensors to monitor and analyze the environment in real time. Industrial environments such as construction sites, oil and gas plants, and power generation facilities are high-risk settings where situational awareness is critical. Every year, thousands of workplace injuries and fatalities are recorded in such sectors, often during inspection and maintenance activities [7,8]. By providing hands-free visual monitoring and on-demand information, smart helmets aim to enhance safety and reduce human error [9].
Concurrently, infrastructure owners face increasing pressure to maintain aging bridges, tunnels, and factories with limited resources. Manual inspection techniques (visual checks, pen-and-paper note-taking, etc.) can be inefficient and may not detect subtle issues until they worsen [10]. Digital inspection technologies promise better coverage and data fidelity. Drones and robots have been used for remote inspection of hard-to-reach areas, but there remain scenarios where a human inspector’s presence and judgment are indispensable [11,12]; for example, examining complex machinery internals or performing detailed nondestructive tests onsite. In these cases, a smart helmet with an integrated vision system can empower the inspector by streaming a video to remote experts, overlaying augmented reality (AR) guidance, and simultaneously building a visual map of the inspected asset. Such capabilities improve inspection accuracy while minimizing exposure to hazards [13,14].
A particularly enabling technology for these advanced helmets is visual SLAM (Simultaneous Localization and Mapping). VSLAM algorithms allow a moving camera to map an environment and track its own position within that map [15,16,17]. Originally developed in robotics, VSLAM enables a helmet-mounted camera to function as the “eyes” of the inspector, continuously capturing the surroundings and creating a 3D model of structural elements (beams, pipes, walls, etc.) as the inspector moves [18,19]. This map can serve multiple purposes, guiding the inspector (by providing navigation cues in complex facilities or GPS-denied areas like mines), identifying structural defects via 3D point cloud analysis, and registering inspection data (photos, thermal images, measurements) to specific locations on the asset. The importance of VSLAM is underscored by its ability to work where GNSS fails in indoor or cluttered environments, offering a self-contained positioning method based on vision [20,21].
As cameras are lightweight, low-cost, and informative, they are increasingly favored over heavy, expensive LiDAR sensors for many applications [22]. Visual methods can thus dramatically augment infrastructure inspection capabilities, provided that robust algorithms and hardware are in place.

1.1. Research Objectives

This review aims to address the following research questions about VSLAM-enabled smart safety helmets for industrial infrastructure inspection.
Primary objective: to systematically analyze and evaluate smart safety helmets equipped with visual SLAM technologies, examining their technological capabilities, deployment effectiveness across industrial inspection scenarios, and implementation challenges to guide future research and industrial adoption.
Specific research questions:
RQ1 (Technology Characterization): how can current helmet-mounted vision and mapping technologies be examined and categorized, especially VSLAM algorithms and camera system configurations?
RQ2 (Performance Assessment): how well do these systems perform across different industrial inspection scenarios, and what quantifiable benefits have been demonstrated in real-world deployments?
RQ3 (Implementation Analysis): what are the key technical, human factor, and regulatory challenges that affect widespread adoption, along with identified best practices for successful deployment?
RQ4 (Future Directions): what are the current research gaps, emerging technological trends, and priority areas for future research development?
Review scope: This analysis focuses on peer-reviewed literature, patents, and industry reports spanning the last decade (2015–2025), particularly emphasizing the recent period when advances in wearable computing and visual SLAM have significantly accelerated. This review specifically targets helmet-integrated vision systems designed for industrial inspection applications, excluding purely consumer or recreational uses.
Target audience: this review is intended for researchers in computer vision and robotics, industrial engineers involved in infrastructure inspection, technology developers working on wearable systems, and decision makers evaluating smart helmet adoption for industrial applications.
The research questions are systematically addressed through the following paper structure: RQ1 (Technology Characterization) is examined in Section 3 (Camera System Typologies) and Section 4 (VSLAM Technologies and Implementations), where we analyze and categorize camera configurations, VSLAM algorithms, and sensor fusion approaches. RQ2 (Performance Assessment) is addressed in Section 5 (Industrial Applications and Use Cases), which evaluates system effectiveness across different inspection domains and presents quantified benefits from real-world deployments. RQ3 (Implementation Analysis) is covered in Section 6 (Challenges and Limitations), examining technical, human factor, and regulatory barriers, while Section 7 (Collaborative and Multi-agent Systems) explores advanced deployment approaches. Finally, RQ4 (Future Directions) is addressed in Section 8 (Conclusion and Recommendations), which synthesizes current research gaps and provides forward-looking guidance for technology development and industrial adoption.

1.2. Historical Evolution of Smart Helmet Technology

The concept of integrating cameras and sensors into helmets has evolved considerably over the past few decades. Early instances in the late 20th century were rudimentary, often driven by research in wearable computing and augmented reality (AR). In 1997, Steve Mann’s seminal work on wearable computing described a “wearable tetherless computer with video capability” as a first step toward personal imaging [23,24,25]. These initial prototypes were cumbersome, involving helmets or headgear connected to backpack PCs and basic cameras, yet they demonstrated the potential of hands-free recording and overlaying information onto a user’s view. Around the same time, the military and law enforcement experimented with helmet-mounted cameras for training and field documentation. However, these systems were largely for offline recording; the idea of real-time processing was limited by the computing hardware of the era.
The early 2000s saw significant developments. Notably, a 2004 U.S. patent by Hartwell and Brug described an “Integrated Smart Helmet” with built-in electronics for safety and convenience features [26]. This helmet included a GPS for location, sensors for environment interaction, a small display, and communications capability, essentially an early blueprint of a multi-sensor smart helmet [27]. Another landmark was the OmniVision helmet camera [28], which patented a completely integrated helmet camera unit with a single-chip CMOS sensor, wireless video transmission, and remote control features. By minimizing size and weight, this invention made it feasible to embed a camera directly in a helmet without encumbering the wearer, reflecting a shift toward truly wearable designs. Such innovations were primarily applied to sports (e.g., helmet cams for skydiving or motocross) and firefighting or policing (for incident recording). They laid the groundwork for treating the helmet as a sensor hub rather than just protection. In the late 2000s and early 2010s, parallel advances in computer vision and SLAM began to influence wearable systems.
On one hand, the computer vision community achieved real-time monocular SLAM on PCs [29]; for example, MonoSLAM was one of the first real-time camera SLAM systems, and PTAM (Parallel Tracking and Mapping) demonstrated robust AR tracking on a handheld device. On the other hand [30], wearable tech advanced with projects like Microsoft’s Wearable AR prototypes and early industrial smart glasses (e.g., Google Glass released in 2013, albeit more of a consumer device). The convergence was evident in research prototypes by 2011. The authors of [31] discussed the optimal placement of wearable vision sensors, including head-mounted stereo rigs. By 2012, a team led by Gutierrez-Gomez, Puig, and Guerrero [32] achieved full 3D visual odometry using an omnidirectional camera on a helmet. Their system recovered metric scale and produced 3D trajectories from a single wide-angle camera, showing that even without dedicated depth sensors, a helmet camera could perform SLAM in principle. Around the same time, the term “smart helmet” started appearing in the literature for systems that improve situational awareness. For instance, a 2013 study by [33] introduced a smart helmet for motorcyclists that could detect accidents and enforce helmet use (though not vision-based, it reflects growing interest in helmet-mounted intelligence).
The mid-2010s saw an inflection point with the emergence of robust VSLAM algorithms and the rise in industrial AR solutions. ORB-SLAM (2015) provided a versatile, high-accuracy monocular SLAM solution with loop closure [34], which was soon extended to stereo and RGB-D versions. Around this time, commercial enterprises brought smart helmets to the market: DAQRI introduced its smart helmet in 2014, integrating an Intel processor, multiple cameras (including a depth camera), and a transparent AR visor [35]. The device was specifically aimed at industry, enabling workers to see overlaid building plans and sensor data in real time, effectively allowing them to “see through” structures with thermal vision and X-ray-like visualization of hidden infrastructure. This marked a shift from using helmet cameras purely for recording or remote viewing to using them for real-time augmented analysis.
Academic research on visual–inertial SLAM and sensor fusion was rapidly maturing, making its way into wearable formats. By the late 2010s, algorithms like VINS-Mono could tightly integrate camera and IMU data for reliable tracking in real time, and ORB-SLAM3 demonstrated accurate visual–inertial SLAM with monocular or stereo cameras [36]. These advances directly feed into helmet systems; a helmet can now leverage an IMU (often already present in headsets for orientation) alongside the camera to achieve far more robust localization, even under fast head movements or low-texture scenes. It showed ORB-SLAM3’s visual–inertial mode to be 5–10× more accurate than pure vision in challenging scenarios. We also see multi-camera helmets being proposed to capture full 360° views. A notable patent in 2019 by Fiala described a helmet with a multi-camera array covering the entire spherical view, combined with fiducial marker tracking for AR applications [37]. By the early 2020s, the concept of a “smart helmet” had expanded [38]: it now encapsulates devices with multi-modal sensors (visual, thermal, gas, etc.), on-board processing (sometimes cloud-connected), and advanced software (VSLAM, object recognition, and communication tools).

2. Methodology

2.1. Search Strategy and Paper Selection Criteria

To capture the relevant literature for this review, we employed a multi-pronged search strategy as seen in [39,40]. We focused on two major scientific databases (Scopus and IEEE Xplore), queried using keywords such as
I.
“smart helmet” OR “wearable AR helmet”.
II.
“helmet camera SLAM” OR “visual SLAM inspection”.
III.
“wearable vision system” OR” industrial augmented reality”.
And combinations thereof. Given the interdisciplinary nature of the topic, our search spanned domains of robotics, computer vision, construction engineering, and occupational safety. We also searched patent databases (e.g., Google Patents) for patents with terms like “smart helmet”, “helmet camera system”, and “helmet AR”, which yielded insight into commercially driven innovations.
To ensure high relevance and quality, we followed the PRISMA guidelines to document and structure our paper selection. Figure 1 presents the flow diagram, which visually summarizes the identification, screening, eligibility, and inclusion stages of our literature review covering both white and gray literature sources. Our initial search (spanning publications from 2015 to 2025) yielded over 392 papers, and after removing duplicates, we were left with 338.
During the screening process, a number of reports were excluded for various reasons. For scientific articles and conference papers, the following exclusion criteria were applied:
-
Not helmet-mounted or wearable: excluded if the focus was not on helmet-mounted or wearable systems.
-
No VSLAM or mapping capability: excluded if the work did not address visual SLAM or relevant mapping functionalities central to this review.
-
Non-industrial or irrelevant application domain: excluded if focusing solely on non-industrial domains (e.g., gaming and recreational VR/AR) unless they introduced transferable technologies.
-
Insufficient technical or methodological depth: excluded if lacking empirical results, technical reproducibility, or in-depth methodological description.
Gray literature sources were screened using the following additional criteria:
-
Promotional or marketing focus: excluded if the material was primarily promotional, lacked objectivity, or was produced for marketing purposes without substantive technical content.
-
Lack of empirical or practical evidence: excluded if the source did not provide real-world deployment evidence, technical results, or data relevant to inspection/maintenance use cases.
-
Redundant with peer-reviewed publication: excluded if peer-reviewed information was available.
We included a limited number of industry reports and product whitepapers to illustrate real-world adoption, such as Real Wear devices, or use cases such as RINA’s marine inspections using the Kiber Helmet [41,42].
After filtering, about 100 sources were identified as highly relevant. Each was reviewed in detail, and from these, we extracted technical details, experimental results, and key insights.
The sections that follow will show a brief discussion of the classification framework and performance evaluation metrics for VSLAM-enabled helmet systems.

2.2. Classification Framework for Technologies

Table 1 shows the organization of the diverse literature classification framework. It shows the adopted classification of the three high-level facets, which are camera system topology, VSLAM and sensor fusion techniques, and finally application and development, while Table 2 shows the taxonomy framework for the three high-level facets with examples.
Evaluating the performance of VSLAM-enabled helmet systems requires considering both technical SLAM metrics and application-specific metrics, as can be seen in Table 3.

3. Camera System Typologies

Helmet-mounted vision systems can employ different camera configurations, each with its own technical implications for SLAM and inspection capabilities. We identify three main typologies: monocular, stereo, and omnidirectional (360°) camera systems. This section reviews each type, discussing their specifications, implementations in research and industry, strengths and limitations for infrastructure inspection, and how they perform within VSLAM frameworks.

3.1. Monocular Camera Systems

A monocular system uses a single camera (or single camera module) on the helmet. This is the simplest hardware configuration, often leveraging a forward-facing RGB camera mounted near the visor or on the side of the helmet. Technically, monocular SLAM is well-studied; it uses a single video stream to both track motion and infer environment structure up to scale [71]. Many early smart helmet prototypes started with monocular cameras due to their low cost, small size, and ease of integration [72,73]. For instance, the Omni Vision patent (2004) integrated a single CMOS camera in a helmet primarily to capture and transmit video [28]. Modern industrial examples include certain AR headsets attached to helmets (like early versions of HoloLens or RealWear devices) [74], which have a single main camera for video streaming and documentation. Table 4 gives a summary of Monocular Camera Typologies with details.

3.2. Stereo Vision Systems

Stereo camera systems consist of two cameras (typically a matched pair) mounted on the helmet with a fixed baseline between them, enabling binocular vision and direct depth perception through triangulation [76]. Many modern AR headsets and some smart helmets use stereo vision as it provides the system with a sense of 3D structure akin to human eyes. For example, the DAQRI smart helmet was reported to include “four HD cameras for 360° awareness” and a “3D depth sensor”, effectively using multiple stereo pairs to cover all directions, with Intel RealSense technology providing active stereo depth [77]. Another example is Microsoft’s HoloLens (integrated into some industrial hardhats), which employs stereo gray cameras for environment tracking along with a depth sensor [78]. In research, stereo helmets have been built by mounting calibrated camera pairs on the front or sides of a helmet to achieve real-time depth mapping of the environment [79]. A detailed summary of the stereo vision systems′ topologies can be seen in Table 5 below, highlighting depth perception and accuracy, notable implementations, strengths, limitations, and VSLAM characteristics of the stereo vision system

3.3. Omnidirectional (360°) Camera Systems

Omnidirectional vision systems provide a full 360° view of the environment around the helmet, either in the horizontal plane or full spherical coverage (360° horizontal and 180° vertical). This can be achieved by multiple cameras arranged to cover all directions (multi-camera array) or by using specialized optics like fisheye lenses or catadioptric mirrors to capture a wide field in one image [88]. On a helmet, achieving true spherical vision often means mounting several cameras (e.g., front, back, sides, and maybe top) and stitching their outputs into a panoramic view. Some research prototypes have used catadioptric cameras (a camera pointed at a curved mirror) mounted atop a helmet to obtain omnidirectional images [89]. The impetus for omnidirectional systems is to ensure that no feature or object around the user goes unobserved. This is very useful for SLAM (minimizing the loss of tracking) and for situational awareness (the user or remote observers can see everything around, not just where the helmet is facing) [90]. The full summary, explaining the details of typical configurations, image stitching and processing, complete environment capture benefits, notable VSLAM implementations, challenges specific to 360° systems, VSLAM performance, and the inspection and field use benefits of omnidirectional (360°) camera typologies can be seen in Table 6.

3.4. Comparative Analysis

Each camera typology offers distinct advantages and faces unique limitations. Here we compare them across several dimensions relevant to industrial deployment
Table 7 summarizes these comparisons. Generally, it can be observed that monocular helmets win on simplicity, cost, and ease, but require more algorithmic aid (like inertial fusion) and careful movement to obtain complete data. Stereo helmets hit a middle ground, giving depth and better accuracy upfront, which improves reliability and measurement utility with moderate added complexity. Omnidirectional helmets provide maximum data and robustness, at the cost of complexity in hardware and software; they shine in specialized scenarios requiring that level of detail. It is worth noting that some hybrid approaches exist too: for instance, a helmet might have a front stereo pair and an additional rear monocular camera, giving depth to what is ahead and at least visual coverage behind. Such hybrid configurations can be seen as a compromise to achieve near-360 coverage without full panorama stitching.
Figure 2 presents a comparative estimation of three common camera system configurations, monocular, stereo, and omnidirectional—frequently employed in helmet-mounted VSLAM systems for industrial inspection and mapping tasks. The radar chart in subfigure (a) synthesizes the performance across five critical criteria derived from the peer-reviewed literature: coverage and accuracy, processing need, ease of deployment, cost, and application scope. These criteria were selected to reflect both technical capabilities and real-world usability in constrained and dynamic industrial settings.
Omnidirectional systems demonstrate a superior performance in coverage and accuracy as well as application scope, owing to their ability to capture full 360° imagery and support rich environmental reconstruction. However, this advantage comes at a cost: their deployment on wearable platforms often requires multiple cameras and complex synchronization, which increases both computational load and system cost. Consequently, they rank lowest in terms of ease of deployment and affordability.
At the opposite end of the spectrum, monocular systems exhibit excellent ease of deployment and cost-efficiency, making them attractive for lightweight, mobile scenarios. Their lower scores in coverage and accuracy stem from a limited field of view and the need for continuous motion to construct complete environmental maps. Stereo configurations occupy an intermediate position, offering a compromise between spatial accuracy and deployment feasibility. Their dual-camera setup supports effective depth estimation while avoiding the bulk and complexity of omnidirectional rigs.
The scenario-based mapping shown in subfigure (b) illustrates how these trade-offs translate into real-world use. In confined environments such as tunnels or pipelines, monocular systems are preferred due to their compact form and minimal processing requirements. Bridge inspection tasks, which demand accurate 3D measurement and structural assessment, benefit from stereo vision’s depth-sensing capabilities. For large-scale facility mapping, particularly for digital twin applications, omnidirectional systems offer comprehensive environmental coverage and contextual awareness, despite higher deployment complexity.
Together, these visualizations underscore that no single camera configuration universally outperforms the others. Instead, optimal system selection hinges on the operational context, balancing practical constraints with performance goals such as localization precision, data fidelity, and system robustness. As the next sections will demonstrate, advancements in VSLAM algorithm design and sensor fusion strategies are increasingly mitigating the limitations of simpler camera setups, enabling a robust performance even in challenging inspection scenarios.

4. VSLAM Technologies and Implementations

To convert raw video data from helmet-mounted cameras into actionable information, such as maps, accurate localization, and environmental insights, advanced algorithms are required. Visual Simultaneous Localization and Mapping (VSLAM) lies at the heart of smart helmet systems, enabling real-time trajectory estimation and environmental mapping. This section provides an overview of major VSLAM algorithm categories, a comparative analysis of their strengths and weaknesses, and reviews implementation strategies specifically adapted to wearable inspection contexts.

4.1. Core VSLAM Algorithm Categories

The fundamental approaches used in wearable VSLAM are summarized in Table 8, highlighting the characteristics, example algorithms, benefits, and key considerations.
These algorithmic approaches continue to evolve as wearable platforms and sensor suites become more sophisticated. Feature-based and direct methods remain the mainstays for many commercial and research systems, but hybrid and deep-learning-augmented methods are gaining traction as hardware capabilities advance.

4.2. Algorithm Impact and Trends (2015–2025)

Building upon the systematic literature review described earlier in Section 2.1, data from the final corpus was analyzed to determine how the impact of each algorithm category evolved over the past decade. The findings, shown in Table 9 and visualized in Figure 2, represent normalized annual impact scores (0–10) based on frequency of adoption, citation trends, and benchmark performances (KITTI, TUM RGB-D, and EuRoC MAV).
Figure 3 shows a plot of the past decade on how the landscape of VSLAM (Visual Simultaneous Localization and Mapping) algorithms for wearable systems has undergone a significant transformation, marked by evolving strengths across various algorithm categories. Initially, feature-based SLAM methods such as ORB-SLAM held dominant influence, particularly between 2015 and 2018. Their robustness, relatively low computational demands, and compatibility with mobile hardware made them the go-to choice for early smart helmets and AR devices. As wearable environments grew more dynamic and demanding, the limitations of feature-only methods, particularly under challenging lighting or texture conditions, led to a gradual decline in their impact. By 2025, while still used in lightweight systems, their influence will have tapered.
Direct methods like LSD-SLAM and DSO saw a moderate rise around 2017 due to their ability to operate in low-feature environments by leveraging image intensity gradients. However, their susceptibility to lighting variations, lack of robustness during fast motion, and computational intensity restricted their wider adoption in wearable platforms. As a result, their impact steadily declined in subsequent years. The integration of visual–inertial SLAM has proven to be one of the most impactful advancements, particularly from 2018 onward. Systems that combine camera data with IMU readings like VINS-Mono, OKVIS, and ORB-SLAM3 in visual–inertial mode offer significantly improved stability, scale observability, and drift reduction. In wearable applications, where rapid head movements and temporary vision occlusions are common, visual–inertial fusion has become essential. Since 2020, it has remained one of the most influential SLAM approaches in real-world helmet and AR system deployments. Parallel to sensor fusion advancements, deep-learning-augmented SLAM began gaining traction after 2018. By incorporating learned features, monocular depth prediction, and semantic segmentation, these approaches improved SLAM robustness in texture-poor or dynamic environments. Though initially hindered by computational constraints, the rise in mobile AI accelerators has enabled more on-device inference, allowing wearable systems to benefit from the adaptability and resilience of AI-enhanced pipelines. By 2025, deep-learning-based SLAM is projected to be the most impactful category, especially in complex inspection and industrial scenarios.
According to [104], neural representation approaches such as Neural Radiance Fields (NeRF) have recently emerged as powerful tools for generating dense, photorealistic 3D reconstructions. While not yet widely adopted for real-time SLAM on wearable devices, their role in post-processing and digital twin generation is rapidly expanding. These methods are especially promising for environments requiring high-fidelity documentation, such as tunnels, heritage sites, or industrial facilities. From 2022 to 2025, their impact has grown significantly and is expected to play a central role in future mapping pipelines. Therefore, it can be said that VSLAM in wearable systems has transitioned from classical, handcrafted methods to more integrated, sensor-rich, and learning-based paradigms. Visual–inertial SLAM remains a core technology for real-time tracking, while deep learning and neural scene representations are shaping the future of mapping and spatial understanding in wearable contexts.

4.3. Sensor Fusion Approaches for Enhanced Robustness

To overcome the limitations of purely visual approaches, wearable VSLAM often integrates complementary sensors to enhance robustness, accuracy, and reliability. Table 10 summarizes the sensor fusion strategies deployed in helmet systems, including their practical implications:
  • Visual–inertial fusion (Camera + IMU): Widely regarded as the gold standard, combining visual data with inertial measurements significantly reduces positional drift, enhances robustness to rapid head movements, and maintains tracking during brief visual occlusions. However, this fusion demands precise calibration and careful sensor synchronization, crucial for high-performance outcomes.
  • Visual–LiDAR fusion: LiDAR sensors provide an accurate geometric structure even under poor visibility, greatly enhancing spatial accuracy. While extremely beneficial for high-fidelity maps (digital twins), LiDAR’s cost, weight, and energy consumption make it practical mainly in specialized industrial helmets or backpack-mounted configurations.
  • Visual–GNSS fusion: integrating GNSS (GPS/RTK) positioning is beneficial for georeferencing outdoor inspections, limiting drift over extensive areas. However, GNSS accuracy degrades indoors or in covered structures, limiting widespread use without supplemental positioning methods.
  • Additional sensors (thermal, barometer, and magnetometer): Sensors like thermal imaging help inspections in low-visibility conditions (smoke and darkness), while barometers and magnetometers support vertical and directional orientation. Despite niche benefits, environmental interferences such as magnetic distortions or thermal-image texture limitations require careful consideration in practical deployment.

4.4. Processing Architectures and Computational Strategies

The computational demands of real-time VSLAM in wearable contexts necessitate the thoughtful selection of processing architectures. Table 11 reviews these architectures and their trade-offs:
  • Edge computing: On-device processing provides immediate, latency-free operation critical for safety-sensitive tasks. Examples like HoloLens use specialized hardware (CPUs, GPUs, ASICs) optimized for energy efficiency. However, battery life, heat management, and device bulk remain significant engineering challenges.
  • Cloud or remote processing: Offloading computation to cloud infrastructures enables the use of sophisticated algorithms (deep learning, photogrammetry). Systems such as the Kiber Helmet demonstrate practical use in remote-expert scenarios. However, cloud reliance introduces latency issues, data security concerns, and dependency on reliable network connectivity.
  • Hybrid architectures (edge–cloud synergy): Combining local edge processing with remote cloud computation strikes a balance, performing real-time tasks on-device and leveraging cloud capabilities for complex analysis or data-intensive tasks. This requires sophisticated partitioning strategies and robust synchronization mechanisms to handle real-world operational constraints effectively.
  • Hardware acceleration and optimization: GPUs, DSPs, and custom processors significantly accelerate real-time VSLAM tasks, enabling low latency and reduced power consumption. However, integrating specialized hardware involves complexity in software development, cost considerations, and challenges in device integration.
Recent advances in edge AI [105,106] are beginning to transform how wearable systems handle VSLAM. Instead of relying solely on traditional embedded processors, newer helmet platforms are now incorporating dedicated AI accelerators, such as NPUs (Neural Processing Units) or optimized deep learning inference engines, directly into the device. This allows deep neural networks used for tasks like feature extraction or object recognition to run locally, with minimal latency and reduced reliance on cloud resources. These edge AI chips are designed for low power consumption, making them ideal for wearable platforms that demand real-time processing without sacrificing battery life. Integrating edge AI into smart helmets opens the door to more intelligent, context-aware decision making in the field, even when internet access is limited or unreliable.

5. Industrial Implementations

Having reviewed the technological underpinnings, we now turn to real-world deployments of smart helmets with integrated vision systems. This section examines key infrastructure inspection workflows, bridges and structural assets, industrial plants, underground infrastructure (e.g., tunnels, mines, and sewers), and power generation facilities that benefit from smart helmet adoption.
Table 12 outlines the typical inspection methods for each domain, highlighting the enhancements enabled by smart helmets and their associated benefits.
To assess the transformative impact of these systems, we analyzed the academic literature and industry reports from the last decade (2015–2025), focusing on how smart helmets enhance three core capabilities:
  • Spatial awareness—helping workers understand their position in complex or GPS-denied environments.
  • Augmented visualization—overlaying relevant data in the user’s field of view.
  • Automation—providing guided procedures, live alerts, or enabling remote expert support.
Using a rating scale from 1 to 5, where 5 indicates high impact, we scored each domain based on the relevance and effectiveness of smart helmet features. The results are shown in Table 13 and visualized in Figure 3.
Figure 4 illustrates how smart helmet technologies are reshaping inspection practices. Spatial awareness shows the highest impact across all domains, particularly in underground infrastructure (scoring 5), where GPS-denied and hazardous conditions necessitate precise real-time localization. Augmented visualization is notably effective in industrial and energy environments (both scoring 4), where overlaid data enhances situational understanding. Automation sees its strongest effect in industrial settings (score of 4), reflecting the use of smart helmets for procedural guidance, sensor alerts, and remote support. While bridge inspections benefit most from spatial awareness, their automation scores remain lower due to the manual nature of structural assessments.
Following this impact assessment, Table 14 presents a series of illustrative industrial implementations demonstrating real-world implementations across various industries. These examples show measurable improvements in safety, inspection accuracy, and operational efficiency.

6. Challenges and Limitations

Despite the promise and demonstrated benefits of VSLAM-enabled smart helmets, there are several Challenges and Limitations hindering their widespread adoption and optimal performance. These span technical issues (hardware and algorithm constraints), human factors (usability and safety for the wearer), and broader regulatory or standardization concerns. Recognizing these challenges is important for guiding future improvements and managing expectations in deployment. Table 15 below highlights the key challenges faced by VSLAM. These challenges were divided into three main categories based on the findings made from the review.
One often overlooked challenge in deploying VSLAM-enabled smart helmets is the ethical dimension, especially around privacy and surveillance. These systems typically involve continuous video capture, environmental scanning, and sometimes biometric data collection (e.g., eye tracking or physiological monitoring). In industrial settings, this can raise concerns among workers about being monitored or recorded without consent and may also conflict with data protection regulations. As noted in [109,110], organizations must take steps to ensure transparent use policies, secure data handling, and, where possible, the anonymization or on-device processing of sensitive information to reduce privacy risks. Designing helmet systems with privacy by design, not as an afterthought, will be crucial for broader acceptance and ethical compliance.
Table 15 highlights a comprehensive set of challenges hindering the adoption of VSLAM-enabled smart helmets across three broad categories: technical, human, and regulatory. Technically, the most pressing issues include short battery life (2–3 h), insufficient processing power for real-time complex tasks like AI or 3D reconstruction, and environmental vulnerabilities such as dust, lighting, or vibration, which impair sensor accuracy. Additionally, the large volume of data generated poses storage, transmission, and privacy challenges, while SLAM algorithms struggle in degraded or feature-poor environments, often requiring fallback systems. Connectivity in remote industrial zones further complicates real-time operations dependent on cloud services. From a human perspective, helmet discomfort, cognitive overload from excessive AR content, difficult user interfaces in noisy environments, and steep training requirements can limit user acceptance. Cultural resistance due to fears of surveillance or job replacement adds another layer of complexity. Regulatory issues also pose significant barriers, including the high cost and time needed for safety certifications, legal concerns over continuous video/audio recording, and a lack of standardization for AR data formats. These multifaceted challenges highlight the need for holistic solutions that address not only the technology itself but also human factors and institutional frameworks to ensure successful deployment and integration in industrial settings.

7. Collaborative and Multi-Agent Systems

Figure 5 shows the network diagram for Collaborative and Multi-agent Systems in smart helmet technology. The network illustrates a dynamic ecosystem where multiple inspectors, drones, ground robots, and remote experts are interconnected through a shared cloud platform. In this system, each smart helmet collects spatial data and communicates with other agents in real time or asynchronously when connectivity permits. Drones and robots contribute by scanning hard-to-reach or hazardous areas, feeding visual and spatial information into the shared map. Remote experts can view these maps and provide live AR guidance, such as virtual annotations or 3D models that helmet users see in their display.
A promising area for future exploration is the development of multi-user or collaborative SLAM systems [46,111]. In large or hazardous environments, such as underground mines, power stations, or disaster zones, multiple workers may be navigating the same space with helmet systems. Enabling these devices to share maps, localize each other, or fuse mapping data in real time could dramatically improve efficiency and safety. For example, if one helmet detects a blocked passage or hazard, that information could be instantly shared with others nearby. While collaborative SLAM has seen early-stage research in robotics, applying it to wearable VSLAM in the field introduces new challenges around synchronization, network stability, and privacy, but also represents an exciting leap forward in operational coordination.
The centralized cloud platform acts as the integration hub, synchronizing maps, annotations, and decision-making tools across all users. The result is a highly coordinated and intelligent inspection workflow where data is pooled, expertise is shared instantly, and inspection efficiency is significantly increased. This visualization reinforces how collaboration transforms smart helmets from standalone tools into part of a scalable, connected system capable of real-time spatial intelligence and cooperative problem solving.

8. Conclusions and Recommendations

The integration of VSLAM-enabled technologies in smart safety helmets for industrial infrastructure inspection represents a promising and rapidly maturing field that addresses critical challenges in workplace safety, inspection efficiency, and data quality. The convergence of advanced computer vision algorithms, robust sensor fusion techniques, and wearable computing platforms has created unprecedented opportunities for enhancing human capabilities in hazardous and complex industrial environments. However, successful deployment requires careful consideration of technical limitations, human factors, and organizational readiness, factors that collectively determine whether these promising technologies can achieve widespread industrial adoption and deliver their anticipated benefits in practice.

8.1. Research Question Synthesis

This review systematically addressed four key research questions about VSLAM-enabled smart safety helmets for industrial infrastructure inspection:
RQ1 (Technology Characterization): how can current helmet-mounted vision and mapping technologies be examined and categorized?
Our analysis identified three primary Camera System Typologies—monocular, stereo, and omnidirectional, each with distinct trade-offs in cost, complexity, and capability (Section 3). We categorized VSLAM approaches into four main types: feature-based methods (dominant in early systems like ORB-SLAM), direct methods (better for low-texture environments), hybrid approaches (combining multiple techniques for enhanced robustness), and emerging deep-learning-enhanced methods (showing promise but limited current field deployment). Sensor fusion strategies were classified into visual–inertial (most mature and widely adopted), visual–LiDAR (high accuracy but costly and power-intensive), visual–GNSS (for outdoor georeferenced applications), and multi-modal approaches incorporating thermal and environmental sensors. The analysis revealed a clear technological evolution from simple recording devices to sophisticated sensor-fused platforms capable of real-time environmental understanding and spatial intelligence.
RQ2 (Performance Assessment): how well do these systems perform across different industrial inspection scenarios?
Our estimation across four key domains—bridge/structural inspection, industrial plants, underground infrastructure, and power generation facilities—highlights important benefits (Section 5). Underground infrastructure showed the highest impact for spatial awareness (5/5 rating), where GPS-denied environments benefit most from SLAM-based localization. Industrial plants and power facilities benefited most from augmented visualization and automation capabilities (4/5 ratings each), enabling real-time overlay of operational data and remote expert guidance. However, performance varies significantly with environmental conditions, with VSLAM systems struggling in low-texture, dynamic lighting, or feature-poor scenarios without appropriate sensor fusion support.
RQ3 (Implementation Analysis): what challenges affect widespread adoption, and what are the best practices?
Three major challenge categories were identified (Section 6): Technical challenges (battery life limitations of 2–3 h, processing power constraints, and environmental robustness requirements), human factors (ergonomic discomfort, cognitive load management, extensive training requirements, and cultural resistance to surveillance technology), and regulatory issues (safety certification complexity, data privacy compliance, and lack of industry standardization). Proposed best practices for successful deployment include involving end-users in UI/UX design processes, starting implementations in stable indoor environments, ensuring robust connectivity or comprehensive offline capability, providing extensive training programs tailored to diverse skill levels, and establishing clear data governance policies with transparent usage guidelines. Research demonstrates that involving end-users throughout the design and development phases creates more usable equipment and systems [112], with worker involvement in safety-related decision making playing an important role in reducing occupational injuries and achieving safety ownership [113]. The most critical success factor identified was addressing human acceptance through demonstrable value delivery and trust-building rather than focusing solely on technical capabilities.
RQ4 (Future Directions): what are the current research gaps and priority areas for development?
Key research gaps include the absence of standardized benchmarks for helmet-based VSLAM evaluation, limited long-term user acceptance studies in industrial settings, an insufficient analysis of unsuccessful deployments and failure modes, and a need for more robust SLAM algorithms under challenging environmental conditions. Priority development areas encompass energy-efficient edge AI processing architectures, improved sensor fusion techniques for harsh industrial environments, semantic SLAM capabilities for automated defect recognition, collaborative multi-helmet mapping systems, and seamless integration with digital twin workflows. The field is transitioning from proof-of-concept demonstrations toward mature industrial tools, requiring focus on system reliability, industry standardization, and total cost of ownership optimization rather than pure technical advancement. Techniques such as geometry-aware 3D learning for semantic segmentation and precise feature detection [114] or compressed sensing methods leveraging dual-frequency LiDAR [115] offer promising pathways to increase the accuracy and efficiency of spatial mapping in future helmet-mounted VSLAM systems.
This research reveals that while VSLAM-enabled smart helmets have the potential to transform industrial inspection, successful deployment requires holistic consideration of technical capabilities, human factors, and organizational readiness. The technology is quickly maturing beyond research prototypes to viable industrial tools; however, widespread adoption depends on addressing integration challenges and demonstrating clear return on investment in real-world operational environments.

8.2. Strategic Implementation Guidance and Recommendations

Addressing these challenges requires a comprehensive approach that balances immediate implementation needs with longer-term technological development. Based on our systematic analysis, we identified ten critical categories, summarized in Table 16, where targeted actions, in our opinion, can accelerate successful deployment and advance the field toward mature, reliable industrial applications.
The implementation strategy should recognize that organizations need immediate, actionable guidance for current technology adoption while researchers require clear priorities for advancing the field’s capabilities.

Author Contributions

Conceptualization, E.A.M.-C., S.M. and A.G.; methodology, E.A.M.-C., S.M. and A.K.; validation, O.P., R.T. and A.K.; formal analysis, E.A.M.-C., S.M., N.O. and I.G.; investigation, A.G., N.O., I.G. and O.P.; data curation, S.M. and N.O.; writing—original draft preparation, E.A.M.-C. and S.M.; writing—review and editing, E.A.M.-C., S.M., A.G., A.K., O.P. and R.T.; visualization, A.G. and A.K.; project administration, V.P.; funding acquisition, V.P. All authors have read and agreed to the published version of the manuscript.

Funding

This article was prepared within the framework of the project “Development of a smart helmet for real-time processing of digital twin data in the oil and gas industry, with a promising solution for use in the construction industry”, founded by the Centrālā Finanšu un Līgumu Aģentūra—CFLA—Latvia, project identification 2.2.1.3.i.0/1/24/A/CFLA/003.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Acknowledgments

GenAI tools, namely OpenAI ChatGPT and Grammarly v1.2, were used to improve smooth transitions among sections and ensure style consistency and grammatical correctness throughout the manuscript. All suggestions for edits provided were revised before integration to ensure accuracy and prevent bias. All intellectual content and data analysis were conducted solely by the authors.

Conflicts of Interest

Authors Oleksandr Pasha and Reinis Tocelovskis are active researchers employed by 3D Engineering SIA, providing industrial insights on the potential use cases of the technologies surveyed in this manuscript. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest. The funding body had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Torres-Barriuso, J.; Lasarte, N.; Piñero, I.; Roji, E.; Elguezabal, P. Digitalization of the Workflow for Drone-Assisted Inspection and Automated Assessment of Industrial Buildings for Effective Maintenance Management. Buildings 2025, 15, 242. [Google Scholar] [CrossRef]
  2. Cannella, F.; Di Francia, G. PV Large Utility Visual Inspection via Unmanned Vehicles. In Proceedings of the 2024 20th IEEE/ASME International Conference on Mechatronic and Embedded Systems and Applications (MESA), Genova, Italy, 2 September 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 1–8. [Google Scholar]
  3. Halder, S.; Afsari, K. Robots in Inspection and Monitoring of Buildings and Infrastructure: A Systematic Review. Appl. Sci. 2023, 13, 2304. [Google Scholar] [CrossRef]
  4. Lee, P.; Kim, H.; Zitouni, M.S.; Khandoker, A.; Jelinek, H.F.; Hadjileontiadis, L.; Lee, U.; Jeong, Y. Trends in Smart Helmets With Multimodal Sensing for Health and Safety: Scoping Review. JMIR Mhealth Uhealth 2022, 10, e40797. [Google Scholar] [CrossRef] [PubMed]
  5. Nnaji, C.; Awolusi, I.; Park, J.; Albert, A. Wearable Sensing Devices: Towards the Development of a Personalized System for Construction Safety and Health Risk Mitigation. Sensors 2021, 21, 682. [Google Scholar] [CrossRef]
  6. Rane, N.; Choudhary, S.; Rane, J. Leading-Edge Wearable Technologies in Enhancing Personalized Safety on Construction Sites: A Review. SSRN J. 2023, 4641480. [Google Scholar] [CrossRef]
  7. Babalola, A.; Manu, P.; Cheung, C.; Yunusa-Kaltungo, A.; Bartolo, P. Applications of Immersive Technologies for Occupational Safety and Health Training and Education: A Systematic Review. Saf. Sci. 2023, 166, 106214. [Google Scholar] [CrossRef]
  8. Obeidat, M.S.; Dweiri, H.Q.; Smadi, H.J. Unveiling Workplace Safety and Health Empowerment: Unraveling the Key Elements Influencing Occupational Injuries. J. Saf. Res. 2024, 91, 126–135. [Google Scholar] [CrossRef]
  9. Bogdan, S.; Alisher, U.; Togmanov, M. Development of System Using Computer Vision to Detect Personal Protective Equipment Violations. In Proceedings of the 2024 5th International Conference on Communications, Information, Electronic and Energy Systems (CIEES), Veliko Tarnovo, Bulgaria, 20 November 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 1–4. [Google Scholar]
  10. Thomas, L.; Lajitha, C.S.; Mathew, J.T.; Jithin, S.; Mohan, A.; Kumar, K.S. Design and Development of Industrial Vision Sensor (IVIS) for Next Generation Industrial Applications. In Proceedings of the 2022 IEEE 19th India Council International Conference (INDICON), Kochi, India, 24 November 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 1–6. [Google Scholar]
  11. Rizia, M.; Reyes-Munoz, J.A.; Ortega, A.G.; Choudhuri, A.; Flores-Abad, A. Autonomous Aerial Flight Path Inspection Using Advanced Manufacturing Techniques. Robotica 2022, 40, 2128–2151. [Google Scholar] [CrossRef]
  12. Sarieva, M.; Yao, L.; Sugawara, K.; Egami, T. Synchronous Position Control of Robotics System for Infrastructure Inspection Moving on Rope Tether. J. Rehabil. Med. 2019, 31, 317–328. [Google Scholar] [CrossRef]
  13. Wang, Z.; Li, M.; Da, H. Research on the Application of AR in the Safety Inspection of Opening and Closing Stations. In Proceedings of the 2024 7th International Conference on Computer Information Science and Application Technology (CISAT), Hangzhou, China, 12 July 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 1193–1197. [Google Scholar]
  14. Kong, J.; Luo, X.; Lv, Y.; Xiao, J.; Wang, Y. Research on Intelligent Inspection Technology Based on Augmented Reality. In Proceedings of the 2024 Boao New Power System International Forum—Power System and New Energy Technology Innovation Forum (NPSIF), Qionghai, China, 8 December 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 688–695. [Google Scholar]
  15. Kanwal, A.; Anjum, Z.; Muhammad, W. Visual Simultaneous Localization and Mapping (vSLAM) of Driverless Car in GPS-Denied Areas. Eng. Proc. 2021, 12, 49. [Google Scholar]
  16. Mishra, B.; Griffin, R.; Sevil, H.E. Modelling Software Architecture for Visual Simultaneous Localization and Mapping. Automation 2021, 2, 48–61. [Google Scholar] [CrossRef]
  17. Bala, J.A.; Adeshina, S.A.; Aibinu, A.M. Advances in Visual Simultaneous Localisation and Mapping Techniques for Autonomous Vehicles: A Review. Sensors 2022, 22, 8943. [Google Scholar] [CrossRef] [PubMed]
  18. Chen, Z. Research on Localization and Mapping Algorithm of Industrial Inspection Robot Based on ORB-SLAM2. In Proceedings of the 2023 International Conference on Electronics and Devices, Computational Science (ICEDCS), Marseille, France, 22 September 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 310–315. [Google Scholar]
  19. Zhang, X.; Zhang, Y.; Zhang, H.; Cao, C. Research on Visual SLAM Technology of Industrial Inspection Robot Based on RGB-D. In Proceedings of the 2023 3rd International Symposium on Computer Technology and Information Science (ISCTIS), Chengdu, China, 7 July 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 51–55. [Google Scholar]
  20. Shahraki, M.; Elamin, A.; El-Rabbany, A. Event-Based Visual Simultaneous Localization and Mapping (EVSLAM) Techniques: State of the Art and Future Directions. J. Sens. Actuator Netw. 2025, 14, 7. [Google Scholar] [CrossRef]
  21. Chen, W.; Shang, G.; Ji, A.; Zhou, C.; Wang, X.; Xu, C.; Li, Z.; Hu, K. An Overview on Visual SLAM: From Tradition to Semantic. Remote Sens. 2022, 14, 3010. [Google Scholar] [CrossRef]
  22. Umezu, N.; Koizumi, S.; Nakagawa, K.; Nishida, S. Potential of Low-Cost Light Detection and Ranging (LiDAR) Sensors: Case Studies for Enhancing Visitor Experience at a Science Museum. Electronics 2023, 12, 3351. [Google Scholar] [CrossRef]
  23. Martin, T.; Starner, T.; Siewiorek, D.; Kunze, K.; Laerhoven, K.V. 25 Years of ISWC: Time Flies When You’re Having Fun. IEEE Pervasive Comput. 2021, 20, 72–78. [Google Scholar] [CrossRef]
  24. Mann, S. Phenomenal Augmented Reality: Advancing Technology for the Future of Humanity. IEEE Consum. Electron. Mag. 2015, 4, 92–97. [Google Scholar] [CrossRef]
  25. Thomas, B.H. Have We Achieved the Ultimate Wearable Computer? In Proceedings of the 2012 16th International Symposium on Wearable Computers, Newcastle, UK, 18–22 June 2012; IEEE: Piscataway, NJ, USA, 2012; pp. 104–107. [Google Scholar]
  26. Hartwell, P.G.; Brug, J.A. Smart Helmet. U.S. Patent 6,798,392, 28 September 2004. [Google Scholar]
  27. García, A.F.; Biain, X.O.; Lingos, K.; Konstantoudakis, K.; Hernández, A.B.; Iragorri, I.A.; Zarpalas, D. Smart Helmet: Combining Sensors, AI, Augmented Reality, and Personal Protection to Enhance First Responders’ Situational Awareness. IT Prof. 2023, 25, 45–53. [Google Scholar] [CrossRef]
  28. Foster, R.R.; Gallinat, G.D.; Nedinsky, R.A. Completely Integrated Helmet Camera. U.S. Patent 6,819,354, 16 November 2004. [Google Scholar]
  29. Trujillo, J.-C.; Munguia, R.; Urzua, S.; Guerra, E.; Grau, A. Monocular Visual SLAM Based on a Cooperative UAV–Target System. Sensors 2020, 20, 3531. [Google Scholar] [CrossRef]
  30. Ometov, A.; Shubina, V.; Klus, L.; Skibińska, J.; Saafi, S.; Pascacio, P.; Flueratoru, L.; Gaibor, D.Q.; Chukhno, N.; Chukhno, O.; et al. A Survey on Wearable Technology: History, State-of-the-Art and Current Challenges. Comput. Netw. 2021, 193, 108074. [Google Scholar] [CrossRef]
  31. Mayol-Cuevas, W.W.; Tordoff, B.J.; Murray, D.W. On the Choice and Placement of Wearable Vision Sensors. IEEE Trans. Syst. Man Cybern.-Part A Syst. Hum. 2009, 39, 414–425. [Google Scholar] [CrossRef]
  32. Gutiérrez-Gómez, D.; Puig, L.; Guerrero, J.J. Full Scaled 3D Visual Odometry from a Single Wearable Omnidirectional Camera. In Proceedings of the 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems, Vilamoura-Algarve, Portugal, 7 October 2012; pp. 4276–4281. [Google Scholar]
  33. Mohd Rasli, M.K.A.; Madzhi, N.K.; Johari, J. Smart Helmet with Sensors for Accident Prevention. In Proceedings of the 2013 International Conference on Electrical, Electronics and System Engineering (ICEESE), Kuala Lumpur, Malaysia, 4 December 2013; pp. 21–26. [Google Scholar]
  34. Zhang, Y.; Wu, Y.; Tong, K.; Chen, H.; Yuan, Y. Review of Visual Simultaneous Localization and Mapping Based on Deep Learning. Remote Sens. 2023, 15, 2740. [Google Scholar] [CrossRef]
  35. Asgari, Z.; Rahimian, F.P. Advanced Virtual Reality Applications and Intelligent Agents for Construction Process Optimisation and Defect Prevention. Procedia Eng. 2017, 196, 1130–1137. [Google Scholar] [CrossRef]
  36. Campos, C.; Elvira, R.; Rodriguez, J.J.G.; Montiel, J.M.; Tardos, J.D. ORB-SLAM3: An Accurate Open-Source Library for Visual, Visual–Inertial, and Multimap SLAM. IEEE Trans. Robot. 2021, 37, 1874–1890. [Google Scholar] [CrossRef]
  37. Mark, F. Fiducial Marker Patterns, Their Automatic Detection In Images, And Applications Thereof. U.S. Patent 20200090338A1, 23 February 2021. [Google Scholar]
  38. Écorchard, G.; Košnar, K.; Přeučil, L. Wearable Camera-Based Human Absolute Localization in Large Warehouses. In Proceedings of the Twelfth International Conference on Machine Vision (ICMV 2019), Amsterdam, The Netherlands, 31 January 2020; Osten, W., Nikolaev, D.P., Eds.; SPIE: Cergy-Pontoise, France, 2020; p. 96. [Google Scholar]
  39. Ballor, J.P.; McClain, O.L.; Mellor, M.A.; Cattaneo, A.; Harden, T.A.; Shelton, P.; Martinez, E.; Narushof, B.; Moreu, F.; Mascareñas, D.D.L. Augmented Reality for Next Generation Infrastructure Inspections. In Model Validation and Uncertainty Quantification, Proceedings of the Conference Society for Experimental Mechanics Series, Florida, FL, USA, 12−15 March 2018; Springer International Publishing: Cham, Switzerland, 2019; pp. 185–192. ISBN 978-3-319-74792-7. [Google Scholar]
  40. Choi, Y.; Kim, Y. Applications of Smart Helmet in Applied Sciences: A Systematic Review. Appl. Sci. 2021, 11, 5039. [Google Scholar] [CrossRef]
  41. RINA Innovative Solutions for Remote Inspection: RINA and Kiber Team Up. Available online: https://www.rina.org/en/media/press/2019/11/12/vr-media (accessed on 1 August 2025).
  42. VRMedia. 5 Reasons Why You Should Choose Augmented Reality Solutions; VRMedia s.r.l.: Pisa, Italy, 2020. [Google Scholar]
  43. Xu, Z.; Lin, Z.; Liu, H.; Ren, Q.; Gao, A. Augmented Reality Enabled Telepresence Framework for Intuitive Robotic Teleoperation. In Proceedings of the 2023 IEEE International Conference on Robotics and Biomimetics (ROBIO), Koh Samui, Thailand, 4 December 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 1–6. [Google Scholar]
  44. Barz, M.; Daiber, F.; Bulling, A. Prediction of Gaze Estimation Error for Error-Aware Gaze-Based Interfaces. In Proceedings of the Proceedings of the Ninth Biennial ACM Symposium on Eye Tracking Research & Applications, Charleston, SC, USA, 14 March 2016; ACM: New York, NY, USA, 2016; pp. 275–278. [Google Scholar]
  45. Kumar, S.; Micheloni, C.; Piciarelli, C.; Foresti, G.L. Stereo Localization Based on Network’s Uncalibrated Camera Pairs. In Proceedings of the 2009 Sixth IEEE International Conference on Advanced Video and Signal Based Surveillance, Genova, Italy, 2–4 September 2009; IEEE: Piscataway, NJ, USA, 2009; pp. 502–507. [Google Scholar]
  46. Zhang, T.; Zhang, L.; Chen, Y.; Zhou, Y. CVIDS: A Collaborative Localization and Dense Mapping Framework for Multi-Agent Based Visual-Inertial SLAM. IEEE Trans. Image Process. 2022, 31, 6562–6576. [Google Scholar] [CrossRef]
  47. Mante, N.; Weiland, J.D. Visually Impaired Users Can Locate and Grasp Objects Under the Guidance of Computer Vision and Non-Visual Feedback. In Proceedings of the 2018 40th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Honolulu, HI, USA, 18–21 July 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 1–4. [Google Scholar]
  48. Elmadjian, C.; Shukla, P.; Tula, A.D.; Morimoto, C.H. 3D Gaze Estimation in the Scene Volume with a Head-Mounted Eye Tracker. In Proceedings of the Workshop on Communication by Gaze Interaction, Warsaw, Poland, 15 June 2018; ACM: New York, NY, USA; pp. 1–9. [Google Scholar]
  49. Hasegawa, R.; Sakaue, F.; Sato, J. 3D Human Body Reconstruction from Head-Mounted Omnidirectional Camera and Light Sources. In Proceedings of the 18th International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications, Nagoya, Japan, 1 January 2023; SCITEPRESS—Science and Technology Publications: Lisbon, Portugal, 2023; pp. 982–989. [Google Scholar]
  50. Yang, Y.; Yuan, S.; Yang, J.; Nguyen, T.H.; Cao, M.; Nguyen, T.-M.; Wang, H.; Xie, L. AV-FDTI: Audio-Visual Fusion for Drone Threat Identification. J. Autom. Intell. 2024, 3, 144–151. [Google Scholar] [CrossRef]
  51. Jiang, L.; Xiong, Y.; Wang, Q.; Chen, T.; Wu, W.; Zhou, Z. SADNet: Generating Immersive Virtual Reality Avatars by Real-time Monocular Pose Estimation. Comput. Animat. Amp, Virtual 2024, 35, e2233. [Google Scholar] [CrossRef]
  52. Carozza, L.; Bosche, F.; Abdel-Wahab, M. [Poster] Visual-Lnertial 6-DOF Localization for a Wearable Immersive VR/AR System. In Proceedings of the 2014 IEEE International Symposium on Mixed and Augmented Reality (ISMAR), Munich, Germany, 10–12 September 2014; IEEE: Piscataway, NJ, USA, 2014; pp. 257–258. [Google Scholar]
  53. Piao, J.-C.; Kim, S.-D. Adaptive Monocular Visual-Inertial SLAM for Real-Time Augmented Reality Applications in Mobile Devices. Sensors 2017, 17, 2567. [Google Scholar] [CrossRef]
  54. Cutolo, F.; Cattari, N.; Carbone, M.; D’Amato, R.; Ferrari, V. Device-Agnostic Augmented Reality Rendering Pipeline for AR in Medicine. In Proceedings of the 2021 IEEE International Symposium on Mixed and Augmented Reality Adjunct (ISMAR-Adjunct), Bari, Italy, 4–8 October 2021; pp. 340–345. [Google Scholar]
  55. Javed, Z.; Kim, G.-W. Publisher Correction to: PanoVILD: A Challenging Panoramic Vision, Inertial and LiDAR Dataset for Simultaneous Localization and Mapping. J. Supercomput. 2022, 78, 13864. [Google Scholar] [CrossRef]
  56. Lai, T. A Review on Visual-SLAM: Advancements from Geometric Modelling to Learning-Based Semantic Scene Understanding Using Multi-Modal Sensor Fusion. Sensors 2022, 22, 7265. [Google Scholar] [CrossRef]
  57. Berrio, J.S.; Shan, M.; Worrall, S.; Nebot, E. Camera-LIDAR Integration: Probabilistic Sensor Fusion for Semantic Mapping. IEEE Trans. Intell. Transp. Syst. 2022, 23, 7637–7652. [Google Scholar] [CrossRef]
  58. Huang, P.; Zeng, L.; Chen, X.; Luo, K.; Zhou, Z.; Yu, S. Edge Robotics: Edge-Computing-Accelerated Multirobot Simultaneous Localization and Mapping. IEEE Internet Things J. 2022, 9, 14087–14102. [Google Scholar] [CrossRef]
  59. Wu, X.; Li, Y.; Long, J.; Zhang, S.; Wan, S.; Mei, S. A Remote-Vision-Based Safety Helmet and Harness Monitoring System Based on Attribute Knowledge Modeling. Remote Sens. 2023, 15, 347. [Google Scholar] [CrossRef]
  60. Jiang, J.; Yang, Z.; Wu, C.; Guo, Y.; Yang, M.; Feng, W. A Compatible Detector Based on Improved YOLOv5 for Hydropower Device Detection in AR Inspection System. Expert Syst. Appl. 2023, 225, 120065. [Google Scholar] [CrossRef]
  61. Zang, Q.; Zhang, K.; Wang, L.; Wu, L. An Adaptive ORB-SLAM3 System for Outdoor Dynamic Environments. Sensors 2023, 23, 1359. [Google Scholar] [CrossRef] [PubMed]
  62. Sahili, A.R.; Hassan, S.; Sakhrieh, S.M.; Mounsef, J.; Maalouf, N.; Arain, B.; Taha, T. A Survey of Visual SLAM Methods. IEEE Access 2023, 11, 139643–139677. [Google Scholar] [CrossRef]
  63. Zhang, Q.; Li, C. Semantic SLAM for Mobile Robots in Dynamic Environments Based on Visual Camera Sensors. Meas. Sci. Technol. 2023, 34, 085202. [Google Scholar] [CrossRef]
  64. Zhao, Y.; Vela, P.A. Good Feature Matching: Toward Accurate, Robust VO/VSLAM With Low Latency. IEEE Trans. Robot. 2020, 36, 657–675. [Google Scholar] [CrossRef]
  65. Servières, M.; Renaudin, V.; Dupuis, A.; Antigny, N. Visual and Visual-Inertial SLAM: State of the Art, Classification, and Experimental Benchmarking. J. Sens. 2021, 2021, 2054828. [Google Scholar] [CrossRef]
  66. Zhao, Z.; Wu, C.; Kong, X.; Li, Q.; Guo, Z.; Lv, Z.; Du, X. Light-SLAM: A Robust Deep-Learning Visual SLAM System Based on LightGlue Under Challenging Lighting Conditions. IEEE Trans. Intell. Transp. Syst. 2025, 26, 9918–9931. [Google Scholar] [CrossRef]
  67. Beghdadi, A.; Mallem, M. A Comprehensive Overview of Dynamic Visual SLAM and Deep Learning: Concepts, Methods and Challenges. Mach. Vis. Appl. 2022, 33, 54. [Google Scholar] [CrossRef]
  68. Jacobs, J.V.; Hettinger, L.J.; Huang, Y.-H.; Jeffries, S.; Lesch, M.F.; Simmons, L.A.; Verma, S.K.; Willetts, J.L. Employee Acceptance of Wearable Technology in the Workplace. Appl. Ergon. 2019, 78, 148–156. [Google Scholar] [CrossRef]
  69. Hyzy, M.; Bond, R.; Mulvenna, M.; Bai, L.; Dix, A.; Leigh, S.; Hunt, S. System Usability Scale Benchmarking for Digital Health Apps: Meta-Analysis. JMIR mHealth uHealth 2022, 10, e37290. [Google Scholar] [CrossRef] [PubMed]
  70. Li, J.; Wu, W.; Yang, B.; Zou, X.; Yang, Y.; Zhao, X.; Dong, Z. WHU-Helmet: A Helmet-Based Multisensor SLAM Dataset for the Evaluation of Real-Time 3-D Mapping in Large-Scale GNSS-Denied Environments. IEEE Trans. Geosci. Remote Sens. 2023, 61, 1–16. [Google Scholar] [CrossRef]
  71. Herrera-Granda, E.P.; Torres-Cantero, J.C.; Peluffo-Ordoñez, D.H. Monocular Visual SLAM, Visual Odometry, and Structure from Motion Methods Applied to 3D Reconstruction: A Comprehensive Survey. Heliyon 2023, 10, e37356. [Google Scholar] [CrossRef]
  72. Pearkao, C.; Suthisopapan, P.; Jaitieng, A.; Homvisetvongsa, S.; Wicharit, L. Development of an AI—Integrated Smart Helmet for Motorcycle Accident Prevention: A Feasibility Study. J. Multidiscip. Healthc. 2025, 18, 957–968. [Google Scholar] [CrossRef]
  73. Campero-Jurado, I.; Márquez-Sánchez, S.; Quintanar-Gómez, J.; Rodríguez, S.; Corchado, J. Smart Helmet 5.0 for Industrial Internet of Things Using Artificial Intelligence. Sensors 2020, 20, 6241. [Google Scholar] [CrossRef]
  74. Kamali Mohammadzadeh, A.; Eghbalizarch, M.; Jahanmahin, R.; Masoud, S. Defining the Criteria for Selecting the Right Extended Reality Systems in Healthcare Using Fuzzy Analytic Network Process. Sensors 2025, 25, 3133. [Google Scholar] [CrossRef]
  75. Tan, C.S. A Smart Helmet Framework Based on Visual-Inertial SLAM and Multi-Sensor Fusion to Improve Situational Awareness and Reduce Hazards in Mountaineering. Int. J. Softw. Sci. Comput. Intell. 2023, 15, 1–19. [Google Scholar] [CrossRef]
  76. Zhu, Y.; Zhang, D.; Zhou, Y.; Jin, W.; Zhou, L.; Wu, G.; Li, Y. A Binocular Stereo-Imaging-Perception System with a Wide Field-of-View and Infrared- and Visible Light-Dual-Band Fusion. Sensors 2024, 24, 676. [Google Scholar] [CrossRef] [PubMed]
  77. Keselman, L.; Woodfill, J.I.; Grunnet-Jepsen, A.; Bhowmik, A. Intel(R) RealSense(TM) Stereoscopic Depth Cameras. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Honolulu, HI, USA, 21–26 July 2017; pp. 1267–1276. [Google Scholar]
  78. Hübner, P.; Clintworth, K.; Liu, Q.; Weinmann, M.; Wursthorn, S. Evaluation of HoloLens Tracking and Depth Sensing for Indoor Mapping Applications. Sensors 2020, 20, 1021. [Google Scholar] [CrossRef] [PubMed]
  79. Schischmanow, A.; Dahlke, D.; Baumbach, D.; Ernst, I.; Linkiewicz, M. Seamless Navigation, 3D Reconstruction, Thermographic and Semantic Mapping for Building Inspection. Sensors 2022, 22, 4745. [Google Scholar] [CrossRef] [PubMed]
  80. Zaremba, A.; Nitkiewicz, S. Distance Estimation with a Stereo Camera and Accuracy Determination. Appl. Sci. 2024, 14, 11444. [Google Scholar] [CrossRef]
  81. Biswanath, M.K.; Hoegner, L.; Stilla, U. Thermal Mapping from Point Clouds to 3D Building Model Facades. Remote Sens. 2023, 15, 4830. [Google Scholar] [CrossRef]
  82. Li, J.; Yang, B.; Chen, Y.; Wu, W.; Yang, Y.; Zhao, X.; Chen, R. Evaluation of a compact helmet-based laser scanning system for aboveground and underground 3D mapping. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2022, XLIII-B2-2022, 215–220. [Google Scholar] [CrossRef]
  83. Hausamann, P.; Sinnott, C.B.; Daumer, M.; MacNeilage, P.R. Evaluation of the Intel RealSense T265 for Tracking Natural Human Head Motion. Sci. Rep. 2021, 11, 12486. [Google Scholar] [CrossRef]
  84. Zhang, G.; Yang, S.; Hu, P.; Deng, H. Advances and Prospects of Vision-Based 3D Shape Measurement Methods. Machines 2022, 10, 124. [Google Scholar] [CrossRef]
  85. Muhovič, J.; Perš, J. Correcting Decalibration of Stereo Cameras in Self-Driving Vehicles. Sensors 2020, 20, 3241. [Google Scholar] [CrossRef]
  86. Di Salvo, E.; Bellucci, S.; Celidonio, V.; Rossini, I.; Colonnese, S.; Cattai, T. Visual Localization Domain for Accurate V-SLAM from Stereo Cameras. Sensors 2025, 25, 739. [Google Scholar] [CrossRef]
  87. Tani, S.; Ruscio, F.; Bresciani, M.; Costanzi, R. Comparison of Monocular and Stereo Vision Approaches for Structure Inspection Using Autonomous Underwater Vehicles. In Proceedings of the OCEANS 2023—Limerick, Limerick, Ireland, 5–8 June 2023; pp. 1–7. [Google Scholar]
  88. Marcato Junior, J.; Antunes De Moraes, M.V.; Garcia Tommaselli, A.M. Experimental Assessment of Techniques for Fisheye Camera Calibration. Bol. Ciênc. Geod. 2015, 21, 637–651. [Google Scholar] [CrossRef]
  89. Murillo, A.C.; Gutierrez-Gomez, D.; Rituerto, A.; Puig, L.; Guerrero, J.J. Wearable Omnidirectional Vision System for Personal Localization and Guidance. In Proceedings of the 2012 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops, Providence, RI, USA, 16–21 June 2012; IEEE: Piscataway, NJ, USA, 2012; pp. 8–14. [Google Scholar]
  90. Gamallo, C.; Mucientes, M.; Regueiro, C.V. Omnidirectional Visual SLAM under Severe Occlusions. Robot. Auton. Syst. 2015, 65, 76–87. [Google Scholar] [CrossRef]
  91. Lin, B.-H.; Shivanna, V.M.; Chen, J.-S.; Guo, J.-I. 360° Map Establishment and Real-Time Simultaneous Localization and Mapping Based on Equirectangular Projection for Autonomous Driving Vehicles. Sensors 2023, 23, 5560. [Google Scholar] [CrossRef]
  92. Houba, R.; Constein, H.P.; Arend, R.; Kampf, A.; Gilen, T.; Pfeil, J. Camera System for Capturing Images and Methods Thereof. U.S. Patent 10187555B2, 22 January 2019. [Google Scholar]
  93. Ardouin, J.; Lécuyer, A.; Marchal, M.; Riant, C.; Marchand, E. FlyVIZ: A Novel Display Device to Provide Humans with 360° Vision by Coupling Catadioptric Camera with Hmd. In Proceedings of the 18th ACM Symposium on Virtual Reality Software and Technology, Toronto, ON, Canada, 10 December 2012; ACM: New York, NY, USA, 2012; pp. 41–44. [Google Scholar]
  94. Campos, M.B.; Tommaselli, A.M.; Honkavaara, E.; Prol, F.D.; Kaartinen, H.; El Issaoui, A.; Hakala, T. A Backpack-Mounted Omnidirectional Camera with Off-the-Shelf Navigation Sensors for Mobile Terrestrial Mapping: Development and Forest Application. Sensors 2018, 18, 827. [Google Scholar] [CrossRef]
  95. Chen, H.; Hu, W.; Yang, K.; Bai, J.; Wang, K. Panoramic Annular SLAM with Loop Closure and Global Optimization. Appl. Opt. 2021, 60, 6264. [Google Scholar] [CrossRef] [PubMed]
  96. Bonetto, E.; Goldschmid, P.; Pabst, M.; Black, M.J.; Ahmad, A. iRotate: Active Visual SLAM for Omnidirectional Robots. Robot. Auton. Syst. 2022, 154, 104102. [Google Scholar] [CrossRef]
  97. Li, J.; Zhao, Y.; Ye, W.; Yu, K.; Ge, S. Attentive Deep Stitching and Quality Assessment for 360$^{\circ }$ Omnidirectional Images. IEEE J. Sel. Top. Signal Process. 2020, 14, 209–221. [Google Scholar] [CrossRef]
  98. Zhang, W.; Wang, S.; Dong, X.; Guo, R.; Haala, N. BAMF-SLAM: Bundle Adjusted Multi-Fisheye Visual-Inertial SLAM Using Recurrent Field Transforms. In Proceedings of the 2023 IEEE International Conference on Robotics and Automation (ICRA), London, UK, 29 June 2023; pp. 6232–6238. [Google Scholar]
  99. Maohai, L.; Han, W.; Lining, S.; Zesu, C. Robust Omnidirectional Mobile Robot Topological Navigation System Using Omnidirectional Vision. Eng. Appl. Artif. Intell. 2013, 26, 1942–1952. [Google Scholar] [CrossRef]
  100. Browne, M.P.; Larroque, S. Electronic See-through Head Mounted Display with Minimal Peripheral Obscuration. In Proceedings of the Optical Architectures for Displays and Sensing in Augmented, Virtual, and Mixed Reality (AR, VR, MR), San Francisco, CA, USA, 19 February 2020; Kress, B.C., Peroz, C., Eds.; SPIE: Cergy-Pontoise, France; p. 45. [Google Scholar]
  101. Hottner, L.; Bachlmair, E.; Zeppetzauer, M.; Wirth, C.; Ferscha, A. Design of a Smart Helmet. In Proceedings of the Seventh International Conference on the Internet of Things, Linz, Austria, 22 October 2017; ACM: New York, NY, USA, 2017; pp. 1–2. [Google Scholar]
  102. Akturk, S.; Valentine, J.; Ahmad, J.; Jagersand, M. Immersive Human-in-the-Loop Control: Real-Time 3D Surface Meshing and Physics Simulation. In Proceedings of the 2024 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Abu Dhabi, United Arab Emirates, 14 October 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 12176–12182. [Google Scholar]
  103. Shivaanivarsha, N.; Vijayendiran, A.G.; Sriram, A. LoRa WAN Based Smart Safety Helmet with Protection Mask for Miners. In Proceedings of the 2024 International Conference on Communication, Computing and Internet of Things (IC3IoT), Chennai, India, 17 April 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 1–6. [Google Scholar]
  104. Zhang, Y.; Li, J.; Zhao, X.; Liao, Y.; Dong, Z.; Yang, B. NeRF-Based Localization and Meshing with Wearable Laser Scanning System: A Case Study in Underground Environment. ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci. 2024, X-4–2024, 501–506. [Google Scholar] [CrossRef]
  105. Jin, X.; Li, L.; Dang, F.; Chen, X.; Liu, Y. A Survey on Edge Computing for Wearable Technology. Digit. Signal Process. 2022, 125, 103146. [Google Scholar] [CrossRef]
  106. Ramasubramanian, A.K.; Mathew, R.; Preet, I.; Papakostas, N. Review and Application of Edge AI Solutions for Mobile Collaborative Robotic Platforms. Procedia CIRP 2022, 107, 1083–1088. [Google Scholar] [CrossRef]
  107. Szajna, A.; Stryjski, R.; Woźniak, W.; Chamier-Gliszczyński, N.; Kostrzewski, M. Assessment of Augmented Reality in Manual Wiring Production Process with Use of Mobile AR Glasses. Sensors 2020, 20, 4755. [Google Scholar] [CrossRef] [PubMed]
  108. Jang, J.; Lee, K.-H.; Joo, S.; Kwon, O.; Yi, H.; Lee, D. Smart Helmet for Vital Sign-Based Heatstroke Detection Using Support Vector Machine. J. Sens. Sci. Technol. 2022, 31, 433–440. [Google Scholar] [CrossRef]
  109. Ferraris, D.; Fernandez-Gago, C.; Lopez, J. A Trust-by-Design Framework for the Internet of Things. In Proceedings of the 2018 9th IFIP International Conference on New Technologies, Mobility and Security (NTMS), Malaga, Spain, 26 February 2018; pp. 1–4. [Google Scholar]
  110. Merchán-Cruz, E.A.; Gabelaia, I.; Savrasovs, M.; Hansen, M.F.; Soe, S.; Rodriguez-Cañizo, R.G.; Aragón-Camarasa, G. Trust by Design: An Ethical Framework for Collaborative Intelligence Systems in Industry 5.0. Electronics 2025, 14, 1952. [Google Scholar] [CrossRef]
  111. Chen, W.; Wang, X.; Gao, S.; Shang, G.; Zhou, C.; Li, Z.; Xu, C.; Hu, K. Overview of Multi-Robot Collaborative SLAM from the Perspective of Data Fusion. Machines 2023, 11, 653. [Google Scholar] [CrossRef]
  112. Horberry, T.; Burgess-Limerick, R. Applying a Human-Centred Process to Re-Design Equipment and Work Environments. Safety 2015, 1, 7–15. [Google Scholar] [CrossRef]
  113. Ajmal, M.; Isha, A.S.; Nordin, S.M.; Al-Mekhlafi, A.-B.A. Safety-Management Practices and the Occurrence of Occupational Accidents: Assessing the Mediating Role of Safety Compliance. Sustainability 2022, 14, 4569. [Google Scholar] [CrossRef]
  114. Wang, H.; Zhang, G.; Cao, H.; Hu, K.; Wang, Q.; Deng, Y.; Gao, J.; Tang, Y. Geometry-Aware 3D Point Cloud Learning for Precise Cutting-Point Detection in Unstructured Field Environments. J. Field Robot. 2025, 42, 514–531. [Google Scholar] [CrossRef]
  115. Li, X.; Hu, Y.; Jie, Y.; Zhao, C.; Zhang, Z. Dual-Frequency LiDAR for Compressed Sensing 3D Imaging Based on All-Phase Fast Fourier Transform. J. Opt. Photonics Res. 2023, 1, 74–81. [Google Scholar] [CrossRef]
Figure 1. PRISMA flow diagram of the literature search and selection process.
Figure 1. PRISMA flow diagram of the literature search and selection process.
Sensors 25 04834 g001
Figure 2. (a) Key trade-offs among three camera systems: monocular, stereo, and omnidirectional, (b) commonly used in industrial inspection and mapping tasks.
Figure 2. (a) Key trade-offs among three camera systems: monocular, stereo, and omnidirectional, (b) commonly used in industrial inspection and mapping tasks.
Sensors 25 04834 g002
Figure 3. Impact strength (on a scale from 0 to 10) of various VSLAM algorithm categories in wearable systems over the last 10 years (2015–2025).
Figure 3. Impact strength (on a scale from 0 to 10) of various VSLAM algorithm categories in wearable systems over the last 10 years (2015–2025).
Sensors 25 04834 g003
Figure 4. The transformative impact of smart helmet systems across four key infrastructure inspection domains.
Figure 4. The transformative impact of smart helmet systems across four key infrastructure inspection domains.
Sensors 25 04834 g004
Figure 5. Network diagram for Collaborative and Multi-agent Systems.
Figure 5. Network diagram for Collaborative and Multi-agent Systems.
Sensors 25 04834 g005
Table 1. Classification framework with taxonomy of three high-level facets.
Table 1. Classification framework with taxonomy of three high-level facets.
FacetDescriptionRelevant Reference
Camera System TypologyCategorizes helmet-mounted vision systems into monocular, stereo, and omnidirectional types based on camera configuration. Influences VSLAM algorithm choice and depth perception.[43,44,45]
VSLAM and Sensor Fusion TechniquesClassifies SLAM algorithms and sensor fusion techniques (e.g., visual–inertial and LiDAR fusion). Covers processing architectures like edge computing and cloud offloading.[46,47,48,49]
Application and DeploymentCategorizes industrial applications: inspections, maintenance, AR guidance, and remote collaboration. Maps combinations of camera systems and SLAM to domains.[50,51,52]
Table 2. Taxonomy framework for subcategories and examples.
Table 2. Taxonomy framework for subcategories and examples.
FacetSubcategoryDescriptionExample from References
Camera
System
Typology
MonocularSingle forward-facing cameraAR headset using monocular camera [53]
StereoDual cameras for depth estimationStereo vision in medical AR applications [54]
Omnidirectional360° coverage with multiple or fisheye lensesPanoramic vision sensors for navigation [55]
VSLAM and
Sensor
Fusion
SLAM algorithmsFeature-based, direct, and deep learning SLAMLearning-based VSLAM [56]
Sensor fusionVisual–inertial, LiDAR, and GNSS/IMULiDAR + vision integration [57]
Processing architecturesOn-device vs. cloud offloadingEdge computing for real-time SLAM [58]
Application and
Deployment
Industrial sectorsConstruction, plants, and tunnelsConstruction monitoring with helmet-enabled vision [59]
Use casesAR guidance, fault detection, and
inspection
Detection and guidance in plant inspection [60]
Table 3. Performance evaluation metrics for VSLAM-enabled helmet systems.
Table 3. Performance evaluation metrics for VSLAM-enabled helmet systems.
Evaluation CategoryMetric TypeDescriptionExample/Typical ValuesReference
SLAM Accuracy and RobustnessATE/RPE/precision–recall/cloud errorATE (Absolute Trajectory Error), RPE (Relative Pose Error) measure trajectory drift; cloud-to-cloud and Hausdorff distance measure map fidelityATE < 0.1–0.3 m; cloud error few cm like 2–5 cm[36,61,62,63]
Real-Time
Performance
Frame rate, latency, and CPU/GPU usageEvaluates SLAM throughput and feasibility for mobile hardware; goal is real-time performance on embedded processors15–30 FPS; <100 ms latency[64,65]
Environmental RobustnessQualitative stress conditionsPerformance under motion blur, low light, smoke, vibration, etc.Performance maintained in <20 lux lighting, moderate blur[66,67]
Human FactorsUsability surveys, ergonomics, and cognitive loadSUS score, operator feedback, and fatigue levels after useSUS score > 70 (acceptable usability); low fatigue over 2 h[68,69]
Benchmark
Datasets
Dataset usageUse of EuRoC MAV, WHU Helmet, etc., to evaluate accuracy and robustnessThe WHU Helmet dataset yielded <5 cm error[70]
Table 4. Summary of Monocular Camera Typologies.
Table 4. Summary of Monocular Camera Typologies.
CategoryDetails
Technical Specs and
Capabilities
-
Resolution HD to 4K
-
Shutter Type: global shutter preferred to reduce motion blur
-
Field of View (FOV): wide (~120°) beneficial for SLAM; introduces distortion
-
Form factor compact/shallow cameras (e.g., smartphone-sized) can be embedded in helmet brims
-
Camera Types: range from webcams to high-end machine vision cameras
-
Depth Estimation: no direct depth; scale/depth inferred from motion
-
SLAM Algorithms: feature-based (e.g., ORB-SLAM) and direct methods (e.g., LSD-SLAM)
-
Aids for Scale: use of known markers or initial pacing distance
Implementations and Use Cases
-
Research:
  •  The 2012 wearable SLAM by Gutiérrez-Gómez et al. (omnidirectional monocular camera)
  •  Smart helmet (camera + barometer + IMU) [75] for mountaineering SLAM
-
Industry:
  •  Live video streaming helmets (e.g., Kiber Helmet)
  •  Hardware present for SLAM even if not actively used
Strengths
-
Cost and Weight: low, as only one camera and minimal hardware
-
Comfort: lightweight setup improves wearability
-
Calibration: simpler than multi-camera setups (intrinsics + possibly extrinsics)
-
Power Efficiency: lower power draw → longer battery life
-
Algorithm Maturity: ORB-SLAM and others are well-optimized for single-camera input
-
Inspection Usability: high-res imagery useful for visual inspection and AI-based analysis
-
Form Factor: easy to conceal in compact mounts (headlamp style) for unobtrusive appearance
Limitations
-
No Absolute Scale: without external reference, only relative scaling possible
-
Poor Low-Texture Performance: may struggle in blank or repetitive environments
-
Rotation Issues: pure head rotation provides no parallax (no depth)
-
Feature Dependence: needs texture/features; poor in low light or featureless spaces
-
SLAM Drift: can accumulate error over distance or in poor environments
-
No Depth Perception: cannot directly measure object sizes or distances
VSLAM Performance (Monocular)
-
Accuracy: typically 1–2% of distance travelled (e.g., 5 cm drift over 30 m)
-
SLAM Quality: effective loop closure and mapping in small/indoor environments
-
Environment Sensitivity: performance degrades in dark/reflective/featureless areas
-
Algorithms: ORB-SLAM, LSD-SLAM (direct method useful in texture-poor settings)
-
Commercial Examples: Google ARCore (initially monocular camera + IMU)
-
Scale Solutions: use of known markers, pacing distance, or inertial fusion for metric
Table 5. Summary of a stereo vision system typologies.
Table 5. Summary of a stereo vision system typologies.
CategoryDetails
Depth Perception and
Accuracy
-
Primary Advantage: direct metric depth via triangulation from left/right camera images [80]
-
Depth Estimation: small errors (e.g., ~1% error at 1–2 m range with 10 cm baseline)
-
Application: ideal for dimensioning tasks (clearances, misalignments, and defect sizing)
-
Three-Dimensional Measurement: enables point-to-point measurements (e.g., crack length in AR view)
Notable Implementations
-
Integrated stereo RGB and thermal cameras for 3D point clouds, georeferencing, and structural alignment [81]
-
Industrial: “Helmet-based laser scanning + stereo vision” in mines, mapping tunnel geometry in real time with stereo and IMU [82]
-
Research: researchers adapt existing stereo rigs for helmets to capture depth images and enable on-the-fly measurements [83]
StrengthsHigh precision in dimensional measurements (acts like a portable 3D scanner) [84]
-
Robustness: higher robustness and lower drift compared to monocular SLAM (due to depth info)
-
Textureless Areas: depth recovery via disparity even in low-texture regions
-
Three-Dimensional Object Detection: enables obstacle avoidance and tripping hazard warnings (based on 3D perception)
-
Defect Detection: improved crack differentiation (real crack vs shadow) and volume measurements
-
AR Integration: better alignment of virtual and real-world objects in AR applications (due to absolute scale)
Limitations
-
Complexity: requires precise calibration for intrinsic and extrinsic camera parameters [85]
-
Calibration Drift: can be affected by mechanical shocks or temperature changes
-
Processing Demand: stereo image processing requires higher computational resources, reducing frame rates or increasing power usage
-
Baseline Trade-Off: larger baseline yields better depth accuracy at longer ranges, but limits near-field precision and can be physically cumbersome on helmets
-
Textureless/Featureless Areas: both stereo cameras face challenges in low-light or featureless conditions (similar to monocular issues)
-
Bulk and Cost: additional hardware increases bulk and cost, with two cameras and potentially an IR projector
-
Power Consumption: more demanding on power, especially in real-time high-res stereo setups
VSLAM CharacteristicsPerformance: stereo SLAM systems (e.g., ORB-SLAM2/3) outperform monocular in terms of localization accuracy and map fidelity [86]
-
Trajectory Estimation: stereo systems can double the accuracy of monocular VSLAM [87]
-
Dense Point Clouds: stereo systems generate dense point clouds for real-time visualization of 3D structure (e.g., in building scans)
-
Fusion with IMU: combining stereo with visual–inertial SLAM leads to very stable tracking
-
Real-Time Mesh Building: stereo depth can be used to build meshes (e.g., KinectFusion with stereo depth as RGB-D source)
Table 6. Summary of omnidirectional (360°) camera typologies.
Table 6. Summary of omnidirectional (360°) camera typologies.
CategoryDetails
Typical Configurations
-
Two Fisheye Cameras: back-to-back for spherical view (similar to consumer 360° cameras) [91]
-
Three to Six Normal Cameras: overlapping FOV to achieve ~360° coverage
-
Patented Designs: the authors of [92] proposed outward-facing multi-cam helmet for spherical vision
-
Catadioptric mirror + camera for full-view display inside helmet [93]
-
Industrial Use: Insta360 or GoPro Max on hardhats for full-environment capture
Image Stitching and Processing
-
Challenges: real-time stitching is compute-heavy; misalignment leads to visual artifacts [94]
-
SLAM Integration: some SLAM systems process individual camera streams (multi-cam SLAM)
-
ORB-SLAM3: supports multi-camera input with fixed extrinsics
-
Fisheye SLAM: some convert to spherical models for SLAM using specialized projections
-
Calibration Complexity: requires intrinsics + extrinsics for all cameras or accurate mirror modeling
Complete Environment
Capture Benefits
-
Omnidirectional Advantage: captures entire surroundings, reducing SLAM failure on rotation [95]
-
Loop Closure: enhanced due to wide-angle/global visibility [96]
-
Persistent Features: features remain in view longer, minimizing tracking loss
-
Awareness: can alert user to behind-the-head events (e.g., steam leaks)
-
Remote Viewers: experts can freely navigate 360° feed without needing user to reposition [97]
Notable VSLAM
Implementations
-
Fisheye visual–inertial SLAM, improved robustness indoors [98]
-
A 360° mapping in GNSS-denied environments [70]
-
Topological SLAM using omnidirectional vision for robust relocalization [99]
Challenges Specific to 360° Systems
-
Data Volume: multiple streams = high bandwidth/storage use
-
Sync Issues: requires hardware synchronization for temporal consistency
-
Exposure Variance: different lighting conditions per camera complicate processing
-
Form Factor Risks: exposed cameras/mirrors prone to damage; needs rugged design
-
Bulk and Cost: more cameras = higher cost and larger form factor
-
Lighting Inconsistencies: can cause seam visibility in stitched images
-
Real-Time Constraints: may require resolution/frame rate trade-offs to stay in real time
VSLAM Performance
-
Robust Tracking: SLAM continues even if some cameras are occluded
-
Loop Closure: strong performance due to comprehensive feature capture
-
Accuracy: sub-centimeter reported in 20 m loops under structured conditions
-
Mapping Efficiency: full environment mapped in one pass; supports fast inspection
-
Thermal + RGB: experimental helmets using both modalities for rich defect detection (e.g., hotspots)
Inspection and Field Use
Benefits
Wide-Area Documentation: captures scenes behind or beside user
-
Efficient Surveying: one pass = full data capture of environment
-
Advanced Applications: enables panoramic thermal maps and geometric overlays
-
Human-Viewable Output: panoramas for reports (tunnel interiors and plant rooms)
-
Analytical Potential: object tagging, defect localization, and 3D overlays
Table 7. Comparative analysis of camera systems.
Table 7. Comparative analysis of camera systems.
CriteriaMonocularStereoOmnidirectional
Mapping Accuracy and Coverage
-
Directional view only; limited coverage.
-
User must move to different angles.
-
Accurate locally but prone to scale drift.
-
Adequate for systematic inspections.
-
High depth accuracy (~120–150°).
-
Limited coverage unless more pairs added.
-
Low trajectory error (<1%).
-
Full 360° coverage in one pass.
-
Best for wide/irregular spaces.
-
Excellent tracking and map completeness.
-
High depth accuracy.
Processing
Requirements
-
Lightest load; single feed.
-
Runs on smartphone-grade processors at ~30 FPS.
-
Double data and disparity computation.
-
In total, ~15–20 FPS on same hardware.
-
High load (e.g., 4× with 4 cams).
-
May need GPU (e.g., Jetson) or belt unit.
-
May record for post-processing.
Practical Deployment
-
Small, lightweight, easy to mount.
-
Safest in confined spaces.
-
Simple to protect.
-
Needs fixed baseline mount.
-
May protrude from helmet.
-
Moderate complexity.
-
Complex integration (multi-units or panoramic).
-
Risk of imbalance or safety concerns.
-
More failure points.
Cost–Benefit
-
Most affordable (~hundreds USD).
-
High return for simple tasks.
-
Ideal for budget scenarios.
-
Moderate cost (~2× monocular + sync).
-
Good balance of cost and depth.
-
Most expensive (multi-cams + software).
-
Best for high-stakes or one-pass mapping (e.g., digital twins).
Table 8. Summary of the core VSLAM algorithm categories in wearable systems.
Table 8. Summary of the core VSLAM algorithm categories in wearable systems.
CategoryDescriptionExample AlgorithmsStrengthsChallenges/Considerations for Wearables
Feature-Based
(Indirect)
Extracts and matches visual features like corners/edges across frames
-
ORB-SLAM, ORB-SLAM2/3
-
PTAM
-
Marker-based AR (AprilTags, QR codes)
-
Efficient on limited hardware—good in textured, structured environments
-
Mature graph optimization methods (e.g., bundle adjustment)
-
Marker-based alignment for stability
-
Struggles with motion blur, low-texture scenes
-
Performance tied to feature quality
-
Can require adaptation for lighting or scale issues [99,100,101]
Direct MethodsOperates on raw pixel intensities to estimate motion by minimizing photometric error
-
LSD-SLAM
-
DSO (Direct Sparse Odometry)
-
Better in low-texture areas
-
Can be more accurate with photometric consistency
-
Sensitive to lighting changes
-
Can be computationally intensive
-
Semi-dense/dense mapping needs tuning for mobile hardware
Hybrid
Approaches
Combines feature and direct methods or integrates inertial/other sensors
-
SVO (Semi-Direct VO)
-
ORB-SLAM3 with IMU
-
Learned depth with classical SLAM back-end
-
Combines strengths (e.g., robust tracking + fast pose estimation)
-
Visual–inertial systems improve robustness
-
Leverages structured scenes (lines and planes) in wearables [102]
-
Complexity in integration
-
Sensor synchronization
-
May still be compute-heavy for real-time use [82,103]
Deep-Learning-EnhancedApplies deep networks for features, depth, or end-to-end SLAM
-
SuperPoint, LF
-
Net (learned features)
-
PoseNet (pose regression)—semantic SLAM systems
-
More robust features under tough conditions
-
Potential for semantic understanding
-
Can improve relocation/loop closure
-
High compute demand
-
Real-time use on helmets is limited unless offloaded
-
Cloud dependency poses latency/reliability issues
Table 9. Data used for the visualization of impact strength of various VSLAM algorithm categories in wearable systems over the last 10 years (2015–2025).
Table 9. Data used for the visualization of impact strength of various VSLAM algorithm categories in wearable systems over the last 10 years (2015–2025).
YearFeature-Based SLAMDirect MethodsVisual–Inertial SLAMDL-Augmented SLAMNeural Representation SLAM
201585310
201696510
201797720
201897831
201996942
202085953
202184964
202274975
202363987
202452999
20254191010
Table 10. Summary of sensor fusion approaches in VSLAM for wearable (helmet) systems.
Table 10. Summary of sensor fusion approaches in VSLAM for wearable (helmet) systems.
Fusion ApproachFused SensorsKey Algorithms/
Systems
BenefitsChallenges/Considerations for Wearables
Visual–Inertial Odometry (VIO)Camera + IMU (accelerometer + gyroscope)
-
VINS-Mono
-
OKVIS
-
ORB-SLAM3 (VI mode)
-
High-frequency motion prediction
-
Smooth 6-DoF tracking with low drift
-
Scale recovery for monocular SLAM
-
Requires precise time and spatial calibration
-
Sensitive to IMU bias/drift without visual correction
-
Needs synchronization routines and noise modeling
Visual–LiDAR FusionCamera + LiDAR (2D or 3D)
-
WHU-helmet project
-
Accurate structure from LiDAR, rich texture from vision
-
Heavy and power-hungry sensors
Visual–GNSS FusionCamera + GPS/GNSS (RTK or standard)
-
GNSS-tagged SLAM maps
-
Outdoor inspection systems
-
Anchors SLAM maps globally
-
Limits drift over large areas
-
Useful for georeferencing and switching indoor/outdoor
-
GNSS unreliable indoors
-
Standard GPS has low accuracy (~3–5 m)
-
RTK requires base station and adds complexity
Other Sensors (Barometer, Magnetometer, etc.)Barometer, magnetometer, digital compass
-
Altitude tracking in SLAM
-
Compass-based orientation alignment
-
Mountaineering helmet research with barometer
-
Adds vertical awareness (elevation gain/loss)
-
Compass can aid map orientation (e.g., aligning to north)
-
Magnetic interference common in industrial settings
-
Barometer only gives relative change, not absolute height
-
Requires fusion logic and filtering
Thermal/RGB/Depth FusionMulti-modal cameras (RGB + thermal, RGB-D)
-
Inspection helmets (heat mapping, etc.)
-
Enables mapping in low-light/smoke conditions
-
Thermal data useful for inspection (e.g., heat leaks)
-
Thermal images lack visual texture
-
Not useful alone for SLAM
-
Needs co-registration with RGB for accuracy
Fusion LevelVaries by system
-
EKF for VIO (older)
-
Tightly coupled graph optimization (modern SLAM)
-
Loose coupling (parallel visual and LiDAR SLAM)
-
Tight fusion improves consistency
-
Flexible trade-off with computation
-
Tight fusion more complex and computationally intensive
-
Loose fusion simpler but less accurate
-
Real-time needs may favor lightweight or partial fusion
Table 11. Summary of processing architectures for smart helmet systems.
Table 11. Summary of processing architectures for smart helmet systems.
Architecture TypeKey FeaturesExamplesBenefitsChallenges/Considerations for Helmets
Edge
Computing (On-Device)
All processing is conducted on or very near the helmet (CPU, GPU, and SoC in helmet or wearable module)
-
Microsoft HoloLens (CPU, GPU, and HPU)
-
Jetson TX2/AGX Xavier-based helmets
-
Snapdragon AR devices
-
Low latency, real-time feedback
-
Works offline
-
Greater reliability for safety-critical use
-
Limited by battery, power, and heat— needs efficient, optimized algorithms—bulkier if high compute power is required
Cloud/Remote ProcessingProcessing is offloaded to cloud or remote servers. Helmet streams data (video and sensor) for processing
-
Kiber Helmet (telepresence)
-
Cloud-based photogrammetry
-
Remote AI anomaly detection
-
Virtually unlimited compute
-
Enables complex algorithms (e.g., 3D recon and deep learning)
-
Reduces helmet weight and heat
-
Depends on connectivity (WiFi/5G)
-
Latency may hurt real-time use
-
Data privacy/security concerns
-
Not suitable for all environments
Hardware
Acceleration
Use of GPUs, DSPs, NPUs, FPGAs, or ASICs to speed up SLAM, vision, and AI tasks
-
NVIDIA Jetson GPU
-
Snapdragon Hexagon DSP—HoloLens HPU (custom ASIC
-
Google Tango Motion Processor
-
Enables real-time performance
-
Lower power than general-purpose CPU
-
Specialized acceleration for SLAM or AI
-
Often proprietary/hard to program
-
Adds design complexity
-
Thermal and integration challenges
Real Time vs. Post-ProcessingTrade-off between immediate on-site feedback and later high-fidelity analysis
-
Edge SLAM + Cloud Map Refinement
-
Record-then-process pipelines
-
ROS-based prototyping (RTAB-Map)
-
Flexible workload distribution
-
Enables high-quality reports later
-
Balances power use and responsiveness
-
Limited real-time capability on low-end edge
-
Risk of missing real-time feedback in post-only systems
Edge–Cloud Synergy
(Hybrid)
Combines local and remote processing for balanced performance and responsiveness
-
Live tracking on edge, deep analysis in cloud—Remote expert guidance + local AR
-
Local anomaly detection + cloud verification
-
Efficient use of resources
-
Can reduce latency for key tasks
-
Adapts to available network
-
Still needs careful task partitioning
-
Risk of data sync issues
-
Complexity in system integration
Software FrameworksMiddleware/platforms for managing sensor data and processing pipelines
-
ROS (Robot Operating System)
-
Custom C++/CUDA pipelines
-
Modular and extensible
-
Easier prototyping
-
Community support (for ROS)
-
ROS not always optimized for embedded use
-
Custom code can be hard to maintain
-
Must be robust for field use
Power and Battery OptimizationStrategy to manage compute and power load to extend operational time
-
Power-saving modes (pause processing when still)
-
Edge vs. cloud trade-offs for power
-
Prolongs battery life—keeps helmet cooler
-
Supports longer inspections
-
Processing must scale with motion/context
-
Extra logic to manage compute schedule
Security and
Privacy
Ensures inspection data is handled securely
-
On-premise edge servers
-
Encrypted cloud channels
-
Local-only processing
-
Data protection in sensitive sites
-
Compliance with regulations
-
Prevents leakage of proprietary data
-
Limits use of public cloud
-
Adds cost for secure infrastructure
-
Needs user authentication, access control
Table 12. Infrastructure inspection workflows for several classes of infrastructure.
Table 12. Infrastructure inspection workflows for several classes of infrastructure.
Inspection TypeTraditional MethodsSmart Helmet EnhancementsBenefits
Bridge and Structural InspectionsManual visual inspection, rope/lift access, paper notes, handheld camerasSLAM-based geo-tagged mapping, hands-free video/data capture, AR overlays of crack maps/BIM, voice or gaze-based defect taggingIncreased safety (hands-free), accurate location-based defect logs, overlay of design data, reduced post-processing
Industrial Plant InspectionClipboard-based rounds, manual gauge checks, heavy reference manualsCamera-based OCR of gauges, AR display of machine data, voice-guided procedures, live remote expert supportFaster decision making, reduced cognitive load, real-time alerts, reduced need for expert travel
Underground Infrastructure (Tunnels, Mines, Sewers)Flashlights, GPS-denied mapping, environmental monitors, manual sketchingSLAM with LiDAR/360° cameras, AR navigation cues, real-time 3D thermal/visual mapping, environmental data overlaysImproved situational awareness, immediate model generation, navigation aid in zero visibility, hazard alerts in context
Power Generation and Energy FacilitiesPaper-based records, handheld inspections, physical marking, high-risk confined space entryLocation-aware data capture, AR component labeling, visual overlays of future installations, interior mapping for turbines/boilersSafer confined space entry, error prevention in maintenance, planning aid for upgrades, objective image logs
Table 13. Data used for visualizing the transformative impact of smart helmet systems across four key infrastructure inspection domains.
Table 13. Data used for visualizing the transformative impact of smart helmet systems across four key infrastructure inspection domains.
Transformation
Category
Bridge and StructuralIndustrial PlantUnderground InfrastructurePower Generation
Spatial Awareness4354
Augmented Visualization3434
Automation2433
Table 14. Illustrative industrial smart helmet/headset implementations.
Table 14. Illustrative industrial smart helmet/headset implementations.
Industrial
Implementation
Application
Domain
Helmet/Headset
Role
Reported BenefitsTechnologies Used
RINA Marine
Surveys (Italy) [41]
Marine vessel inspectionRemote expert support through live helmet video and AR markupReduction in inspection time, reduced travel cost, improved safety, and accuracyKiber 3 Helmet, live video streaming, AR annotations
Boeing Wiring Assembly (USA) [107]Aerospace manufacturing, harness assemblyAR-guided step-by-step wiring layout on work surfaceReduction in assembly time, near-zero error rateAR wearable headsets, Visual Task Guidance
WHU Underground Mine Mapping (China) [70,82]Mining/tunnel infrastructure mappingReal-time and post-processed 3D SLAM mapping of tunnelsSurvey-grade 3D point clouds, mapping performed in minutes vs. hours, centimeter-level accuracyHelmet with LiDAR, IMU, cameras, and SLAM algorithms
Worker Safety Monitoring (South Korea) [108]Construction site safety complianceReal-time detection of hazards, PPE compliance, and environmental monitoringNear elimination of heatstroke cases, reduction in unsafe behavior, and safety incident preventionAI-enabled smart helmets, environmental sensors, cameras
Table 15. Key challenges faced by VSLAM-enabled smart helmets in three main categories: technical challenges, human factors, and regulatory and standardization issues.
Table 15. Key challenges faced by VSLAM-enabled smart helmets in three main categories: technical challenges, human factors, and regulatory and standardization issues.
CategoryChallengeDescription
Technical
Challenges
Battery Life ConstraintsHigh-performance cameras, processors, and transmitters consume significant power, leading to limited battery life (2–3 h), which is insufficient for full work shifts. Solutions include using belt-mounted packs, but that adds a tether and may not be practical in remote settings.
Processing Power LimitationsWearable devices have limitations in processing real-time complex algorithms like 3D reconstruction or AI detection, causing trade-offs between functionality and performance (e.g., reduced mapping or slower responsiveness).
Environmental Factors Affecting PerformanceHarsh environments (lighting variability, dust, rain, vibration, and EMI) degrade sensor performance, leading to potential failures in SLAM. Additional ruggedization or fallback sensors (LiDAR and radar) are required but add cost and weight.
Data Management and StorageHigh data volumes (video, 3D scans, and HD images) challenge storage and bandwidth, especially in remote locations. Managing data lifecycle, archiving, and compliance adds complexity. Privacy concerns and encryption are needed.
Localization in Degraded SituationsSLAM can fail in low-feature or dynamic environments, causing tracking errors or map failures. Solutions include manual reinitialization or fallback systems, but they impact user experience.
Connectivity in Real-time SystemsSmart helmets dependent on cloud services face network challenges in industrial environments with limited connectivity, requiring solutions for offline autonomy and local networks.
Human FactorsErgonomics and User AcceptanceHelmets can be uncomfortable (heavy, warm, and restrictive), leading to resistance from users. Balance between comfort, safety standards, and system functionality is essential to encourage adoption.
Training RequirementsWorkers need extensive training to use the system effectively, which may be challenging for non-tech-savvy workers. Designing intuitive interfaces and minimizing learning curves is key.
Cognitive Load during OperationAR systems can overwhelm users with excessive information, causing distractions or fatigue. The interface must provide contextually relevant information without causing mental overload.
User Interface ConsiderationsHands-free control via voice or gestures can fail in noisy or restrictive environments. Clear feedback and simple, quick task execution are necessary to avoid user frustration and ensure adoption.
Cultural and Workflow ChangesResistance to technology due to concerns over autonomy or surveillance can hinder adoption. Change management, trust-building, and clear communication of the system’s role as an assistant rather than a replacement are critical.
Regulatory and Standardization IssuesSafety Certifications for Industrial EnvironmentsHelmets must meet strict safety standards (e.g., impact resistance and ATEX for explosive environments). Certification processes are expensive and time-consuming, which limits deployment in hazardous environments.
Data Privacy and Security ConcernsContinuous data collection (video/audio) raises privacy concerns. Companies must establish clear data usage policies and secure communication to avoid misuse and comply with privacy laws.
Integration with Existing Safety ProtocolsSmart helmets must integrate with traditional safety protocols (e.g., lockout–tagout) and may need regulatory changes for digital tools to be accepted as equivalent to physical safety measures.
Lack of Industry Standards for AR Data and InteroperabilityLack of standardization for AR inspection data formats leads to vendor lock-in and interoperability challenges. Industry guidelines are evolving but are not yet fully established.
Ethical and Legal Issues of RecordingLegal implications of continuous video/audio recording, especially without consent, require clear policies and user consent. Intellectual property protection and privacy concerns for workers must be addressed.
Workforce Acceptance and Labor RelationsResistance to helmet technology due to concerns over surveillance or loss of autonomy requires thoughtful communication and potential labor agreements to ensure trust and positive adoption.
Table 16. Recommendations and future research direction.
Table 16. Recommendations and future research direction.
CategoryRecommendationsFuture Research Directions
SLAM and PerceptionUse proven SLAM solutions (e.g., ORB-SLAM3) and begin with stable indoor deploymentsDevelop robust SLAM that adapts to challenging conditions using event cameras, radar, and sensor fusion
AI and Semantic AnalysisStart integrating pre-trained models for basic object detection and labelingBuild semantic SLAM systems that understand environments (pipes, defects, etc.) and support automated defect tagging
Human–Computer InteractionInvolve inspectors in UI/UX design and train them on AR usage effectivelyStudy AR content delivery to reduce distraction and cognitive load, personalize content via adaptive AI
Hardware and IntegrationSelect ergonomic helmet designs with reliable battery life; plan for regular calibrationResearch miniaturized components, edge AI chips, and seamless AR displays like micro-LED or visor projection
Data and Digital TwinsCentralize data storage and begin fusing helmet outputs with BIM/IoT dashboardsEnable dynamic digital twins updated in real time with wearable and IoT data
Collaboration ToolsDeploy remote expert systems (e.g., RealWear, Kiber) for guidance and training Research shared spatial maps, swarm SLAM, and AR/VR platforms for team coordination and multi-user annotation
Networking and InfrastructureEnsure secure, robust connectivity (WiFi or 5G) in inspection areasDevelop decentralized SLAM, compression methods for SLAM data, and robust map merging protocols
Standardization and PolicyDefine clear IT and privacy policies; maintain devices and data proactivelyCreate benchmarks and datasets for helmet-based inspections; collaborate on open standards and IP-aware design
Industry AdoptionLaunch pilot projects with clear metrics; document “quick win” successes to justify scalingStudy cost–benefit and impact of helmet deployment in large-scale industrial maintenance cycles
XR for TrainingUse AR helmets to assist live tasks; link training to operations when possibleExplore using helmet-collected 3D models for VR-based training scenarios and study training-transfer effectiveness
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Merchán-Cruz, E.A.; Moveh, S.; Pasha, O.; Tocelovskis, R.; Grakovski, A.; Krainyukov, A.; Ostrovenecs, N.; Gercevs, I.; Petrovs, V. Smart Safety Helmets with Integrated Vision Systems for Industrial Infrastructure Inspection: A Comprehensive Review of VSLAM-Enabled Technologies. Sensors 2025, 25, 4834. https://doi.org/10.3390/s25154834

AMA Style

Merchán-Cruz EA, Moveh S, Pasha O, Tocelovskis R, Grakovski A, Krainyukov A, Ostrovenecs N, Gercevs I, Petrovs V. Smart Safety Helmets with Integrated Vision Systems for Industrial Infrastructure Inspection: A Comprehensive Review of VSLAM-Enabled Technologies. Sensors. 2025; 25(15):4834. https://doi.org/10.3390/s25154834

Chicago/Turabian Style

Merchán-Cruz, Emmanuel A., Samuel Moveh, Oleksandr Pasha, Reinis Tocelovskis, Alexander Grakovski, Alexander Krainyukov, Nikita Ostrovenecs, Ivans Gercevs, and Vladimirs Petrovs. 2025. "Smart Safety Helmets with Integrated Vision Systems for Industrial Infrastructure Inspection: A Comprehensive Review of VSLAM-Enabled Technologies" Sensors 25, no. 15: 4834. https://doi.org/10.3390/s25154834

APA Style

Merchán-Cruz, E. A., Moveh, S., Pasha, O., Tocelovskis, R., Grakovski, A., Krainyukov, A., Ostrovenecs, N., Gercevs, I., & Petrovs, V. (2025). Smart Safety Helmets with Integrated Vision Systems for Industrial Infrastructure Inspection: A Comprehensive Review of VSLAM-Enabled Technologies. Sensors, 25(15), 4834. https://doi.org/10.3390/s25154834

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop