Next Article in Journal
Enhanced Infrared Sparse Pattern Extraction and Usage for Impact Evaluation of Basalt-Carbon Hybrid Composites by Pulsed Thermography
Next Article in Special Issue
Validation of Zulu Watch against Polysomnography and Actigraphy for On-Wrist Sleep-Wake Determination and Sleep-Depth Estimation
Previous Article in Journal
Influence of Strength Level on the Acute Post-Activation Performance Enhancement Following Flywheel and Free Weight Resistance Training
Previous Article in Special Issue
3D Printed Robot Hand Structure Using Four-Bar Linkage Mechanism for Prosthetic Application
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Advanced Assistive Maintenance Based on Augmented Reality and 5G Networking

1
Department of Information Engineering, University of Padova, 35131 Padua, Italy
2
Dipartimento di Elettronica, Informazione e Bioingegneria, Politecnico di Milano, 20133 Milan, Italy
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(24), 7157; https://doi.org/10.3390/s20247157
Submission received: 13 November 2020 / Revised: 5 December 2020 / Accepted: 9 December 2020 / Published: 14 December 2020

Abstract

:
Internet of Things (IoT) applications play a relevant role in today’s industry in sharing diagnostic data with off-site service teams, as well as in enabling reliable predictive maintenance systems. Several interventions scenarios, however, require the physical presence of a human operator: Augmented Reality (AR), together with a broad-band connection, represents a major opportunity to integrate diagnostic data with real-time in-situ acquisitions. Diagnostic information can be shared with remote specialists that are able to monitor and guide maintenance operations from a control room as if they were in place. Furthermore, integrating heterogeneous sensors with AR visualization displays could largely improve operators’ safety in complex and dangerous industrial plants. In this paper, we present a complete setup for a remote assistive maintenance intervention based on 5G networking and tested at a Vodafone Base Transceiver Station (BTS) within the Vodafone 5G Program. Technicians’ safety was improved by means of a lightweight AR Head-Mounted Display (HDM) equipped with a thermal camera and a depth sensor to foresee possible collisions with hot surfaces and dangerous objects, by leveraging the processing power of remote computing paired with the low latency of 5G connection. Field testing confirmed that the proposed approach can be a viable solution for egocentric environment understanding and enables an immersive integration of the obtained augmented data within the real scene.

1. Introduction

Due to the growing cost of specialized technicians, the increasing complexity of maintenance work and the rising criticality of long machine downtime, Augmented Reality (AR) is playing an ever-growing role in advanced and remotely assisted maintenance [1,2]. In many cases, modern industrial service applications require specialized interventions, and apprenticeship programs are expensive and time-consuming. New employees need on-site oversight from more experienced technicians, which implies increasing costs and complicating the labor model, compromising service efficiency and effectiveness. Traditional actions taken by different societies are the following:
  • dedicated systems for service technicians to communicate with one another and the back-office;
  • access to data dashboards and detailed reports;
  • employment of journeymen for on-boarding of new technicians.
Nowadays, however, these solutions alone are not enough: the whole system can be made more efficient by adopting newer models that enable the assistive guidance of experienced workers without being physically on-site, and deliver guidance to more apprentices at the same time. An effective back-office interaction with field employees is thus becoming a crucial point for an efficient maintenance policy. Moreover, thanks to emerging technologies such as AR helmets and 5G networking, traditional interactions by means of cellphones or work-specific messenger applications are overtaken by more immersive and interactive approaches.
In this paper, we present a complete advanced assistive maintenance framework (Figure 1), deployed and tested in a real-world scenario. The on-site operator is provided with an AR helmet equipped with a thermal camera and a depth sensor. All these devices are connected in a WLAN based on the IEEE 802.11ac protocol and send data to a customer-premises equipment (CPE). The CPE provides a 5G connection to a remote control room, which hosts a Multi-access Edge Computer (MEC) that enables performing complex processing tasks for the egocentric vision system in real-time. A further remote call with a fully shared augmented reality environment was also implemented to allow an advanced support by an expert from a control room. The MEC approach represents in fact an affordable compromise to bring applications from centralized data centers in the cloud to the network edge, closer to the consumers and the data acquired by multiple sensors, and is acknowledged as the main brick to accomplish 5G 3GPP specifications [3].
The rest of the paper is organized as follows. Section 2 overviews the field of AR for egocentric vision systems. Section 3 frames the proposed system within the 5G panorama. Section 4 and Section 5 describe the algorithms developed to process the acquired 3D scene and integrate the information from heterogeneous sensors (depth and thermal cameras), respectively. Section 6 illustrates the prototyping and testing phase, along with a discussion of the obtained results. Finally, Section 7 concludes the paper and draws the overall conclusions.

2. AR for Egocentric Vision

The first Head-Mounted Displays (HMDs) date back to the 1960s, when the head tracking capability was first introduced by Sutherland [4]; then their use grew continuously, first in the Virtual Reality (VR) world for networked applications such as DIVE [5] or MASSIVE [6] and then in the 1990s, for the networked AR systems such as Studierstube [7] or Shared Space [8] (further historical details can be found in [9]).
Nowadays, we dispose of a wide variety of VR and AR devices, from cheap, smartphone-based ones, such as the papercraft FOV2GO from the University of Southern California [10], to more advanced and professional devices [11] such as Microsoft HoloLens, Magic Leap One and Google Glass. In addition to assistive maintenance [1,2], AR has proven its effectiveness in a wide range of applications, such as training for high-precision tasks [12], indoor navigation in large buildings (such as hospitals, airports and train stations) [13], education [14], image-guided surgery [15] and telemedicine [16].
Within the different features of AR for assistive maintenance, telepresence, i.e., the possibility to interact with a remote specialist along with real and virtual objects, is among the most exciting and requested capabilities. Telepresence can be carried out in different forms, from the old and simplest voice call to the most advanced telexistence, where a full holographic representation of the remote assistant is able to interact in real-time with virtual objects in the real scene. Of course, the higher the interaction complexity, the larger the required amount of exchanged data. Furthermore, the transmission latency must be as low as possible to ensure an immersive and realistic interaction.
The reasons for an unnatural and, in some cases, disturbing fusion of real and virtual elements in HMDs are to be found in the following: delays between the helmet’s movement and the refresh of the virtual scene; oversimplified 3D mesh surfaces; wrong ray-traced illumination of virtual objects. Recent smartglasses, such as Microsoft HoloLens, cope well with these aspects thanks to the integration of the internal Inertial Measurement Unit (IMU) with the on-board processing of multiple cameras and depth-cameras in a Simultaneous Localization and Mapping (SLAM) approach [17]. The user’s head position is continuously tracked with high accuracy so that no delay or misplacement in the rendering of virtual objects is noticeable. However, an additional physical cause of disturbing effects is related to the imperfect eyes trick induced by images projected on the holographic lenses. The human visual system allows for the perception of a three-dimensional world from plain and ambiguous stereo images obtained from the eyes. There are different theories on how this is accomplished; one of the most credited is based on the Cue Theory [18], where 10 depth cues interact with lighting and scene structure in order to provide a complete three-dimensional understanding [19]:
  • binocular disparity;
  • binocular convergence;
  • accommodative focus;
  • atmospheric haze;
  • motion parallax;
  • linear perspective and foreshortening;
  • occlusion;
  • height in the visual field;
  • shading;
  • texture gradient.
AR allows to trick our visual system into merging virtual objects seamlessly with the real environment by acting on some of the aforementioned cues. Still, HMDs suffer from a series of drawbacks due to forcing the user to a fixed accommodative length of the crystalline lens and to the limited field of view, causing AR to feel unnatural. Furthermore, according to [20,21,22], the aforementioned problems cause an underestimation of the egocentric depth, at least in the user’s neighborhood, with a mean error ranging from 0.4 to 0.8 meters for distances between 3 and 7 m [20]. This last aspect could represent a serious hazard in industrial environments.
The proposed system deals with some of the most common purposes of AR assistive maintenance:
  • indoor navigation;
  • remote interaction with a remote specialist;
  • contact with hot surfaces avoidance;
  • obstacle avoidance.
For all of them, a wrong perception of the distances of virtual objects in the HMD could result in a faulty service and introduce further risks to the operator’s safety. According to [23,24], it is possible to limit the erroneous distances perception by applying a linear rescaling of depth maps acquired by the depth camera, just for the egocentric AR visualization system. Virtual objects placed in this stretched frame of reference are then perceived as properly overlapped to their real counterpart. The transformation links the new depth d to the original one d according to (1):
d = d + 0.1 0.9
The description of the indoor navigation task is kept out of this paper, since it is based on the approach provided in [25]: the 3D point clouds and surfaces acquired by HoloLens provide an accurate indoor localization, from which navigation is obtained by visualizing holographic AR signage on the HMD. No audio indications were considered, due to the possible occurrence of noisy environments in industrial plants.

3. 5G Networking

The fifth generation of mobile communication networks (5G) sets some stringent requirements related to ultra-low-latency, ultra-responsive and ultra-reliable services. The only way of accomplishing these requests is to distribute the computational power from the central cloud to a Multi-access Edge Computing (MEC). The MEC approach allows moving storage and relevant computation tasks closer to the access network, reducing the traffic volume and achieving real-time performance for AR applications, together with context-awareness and enhanced privacy protection [3,26].
AR and VR technologies are expected to be among the most relevant applications of the 5G/MEC paradigm, with a predicted market of $179 billion by 2021 [27]. On the other hand, this implies a large growth in network traffic. One of the most widely proposed countermeasures consists in local caching at the Edge [28], which comes particularly handy in an assistive maintenance scenario as it allows to keep frequently accessed documents, 3D models and holographic videos directly on the MEC, reducing the load in backhaul links and cloud servers.
Caching also reduces the end-to-end latency, allowing in our case to upload 3D mesh surfaces at an average speed of 156 MB/s, equivalent to approximately 1.5 M triangles/s. Assuming an average environment mapping composed of 20 K triangles (empirical measurements are reported in Table 1), an entire new map can be uploaded (e.g., while the operator enters a new room) in less than 20 ms. Entering different rooms, along with the associated holographic augmented elements, is perceived seamlessly by the user as the whole procedure is carried out within the HoloLens’ framerate. It is worth mentioning that the WiFi transmission between the 5G modem/router and HoloLens is performed with a 802.11ac connection in the 5 GHz band [29], whose bandwidth is large enough to not introduce any significant delay.
Once the HMD enters a new room, it starts exploring changes with respect to the previously stored map and periodically transmits the new 3D mesh back to the MEC, where the Unity development platform [30] runs the analyses described in Section 4. Together with HoloLens’ data, RGB and thermal images (Section 5) are transmitted to the MEC as well, but their incidence is negligible (around 6.5 MB/s) with respect to the overall data stream.

4. Scene Processing

The core feature of our assistive maintenance framework consists of a scene understanding software that: (i) makes use of depth sensors to acquire 3D data from the environment; (ii) processes such data on an edge computing platform to detect and localize potentially dangerous objects; (iii) highlights these objects in real-time to the operator by projecting 3D holograms onto the HMD.
The HMD adopted for deploying the system is Microsoft HoloLens. These AR glasses are equipped with a Time-of-Flight (ToF) sensor that constantly acquires depth information from its surroundings and progressively builds and updates a 3D representation of the environment, in the form of a triangular mesh. The on-board computational power is, however, not sufficient to perform real-time analysis on the recorded 3D data; therefore, the environmental mesh is transmitted over a 5G network to an MEC, where the whole processing task is carried out by a set of specifically designed algorithms (implemented in Unity). The output consists of a set of obstacle coordinates, which are sent back to the AR glasses and visualized on the integrated holographic display.
Providing a formal definition for obstacles is no trivial task. Generally speaking, we may associate the concept of an obstacle to an object impeding or limiting the execution of a given action. In the specific case of on-field assisted maintenance, we chose to narrow our focus to isolated objects on the ground (or hanging from the ceiling) that stand in the operator’s way and, if unnoticed, may hamper its movements or result in potential security threats. Given a 3D mesh of the environment, we call obstacle any part of the mesh within a predefined size range that results disconnected from the scene once removed the floor (or ceiling) plane. The following paragraphs provide a detailed description of the algorithms developed to preprocess the environmental mesh, extract the ground-plane and localize the isolated components.

4.1. Mesh Simplification

AR devices typically acquire significant amounts of 3D data, with environmental meshes easily counting tens (or even hundreds) of thousands of vertices. While the actual number of bits remains limited to a few megabytes and thus does not impair the required low latency, the computational complexity on the MEC side may increase rapidly and non-linearly with the number of 3D points. To prevent the formation of a bottleneck at the edge computing side, we added a preprocessing step that simplifies the received mesh before feeding it to the obstacle detection module.
Mesh simplification consists of a class of algorithms that transform a given mesh into an approximated version with fewer vertices and faces. A variety of approaches have been proposed by the scientific community, addressing different application-specific quality criteria [31]. Our implementation falls within the vertex clustering subclass [32], which is considered to be among the most efficient, once paying the price of removing small-shape details. While the spatial resolution of the original mesh is fundamental to the HMD for an accurate positioning of holograms in the real world, a rougher representation of the environment is more than enough for the task of detecting macroscopic obstacles.
In Unity notation, a triangular mesh is an object M = ( V , T ) , where V is an array of 3D vertices and T is a concatenation of triplets of vertex indices, denoting triangular faces. The proposed algorithm receives a triangular mesh, M, and outputs a simplified version of the same, herein referred to as M = ( V , T ) , with respect to a given quantization parameter, δ .
The simplification procedure starts with quantizing the three-dimensional space into a lattice of cubes with δ -sized edges. The algorithm then scans through the vertices in V, retaining only one point per cube and keeping track of the discarded ones. Finally, new triangles are built by looping over T and remapping the indices of the removed vertices into those of the retained ones. A pseudocode representation is provided in Algorithm 1. Note that this implementation produces redundant copies of some triangular faces, which are deliberately not removed from T , since they do not impair Unity’s performance and, therefore, would result in a waste of computational time.
The simplified mesh M is now passed to the ground-plane extraction step.
Algorithm 1 Mesh simplification algorithm. Requires: V, array of 3-D vertices; T, concatenation of triplets of vertex indices, denoting triangles; δ , spatial quantization step.
 1: procedure Mesh Simplify ( M = ( V , T ) , δ )
 2:    n 0 ▹ New number of vertices
 3:   for i in range of | V | do
 4:      x , y , z V [ i , : ] · δ 1 ▹ Quantized vertex coordinates
 5:     if no vertex quantized to ( x , y , z ) yet then
 6:       V [ n ] V [ n ]
 7:       p r e v i o u s [ x , y , z ] n
 8:       o l d 2 n e w [ i ] n
 9:       n n + 1
 10:    else
 11:      o l d 2 n e w [ i ] p r e v i o u s [ x , y , z ]
 12:     end if
 13:   end for
 14:   for j in range of | T | do▹ Make new triangles
 15:      T [ j ] o l d 2 n e w [ T [ j ] ]
 16:   end for
 17:   return M = ( V , T ) ▹ Simplified mesh
 18: end procedure

4.2. Ground-Plane Extraction

Estimating the coordinates of the ground-plane represents a critical step for obstacle detection, as it defines the elevation level from which to start detecting isolated mesh components. The algorithm for this task was designed around the specific features of the employed HMD.
HoloLens’ world coordinates are initialized to ( x , y , z ) = 0 at the point in space where its operating system is booted. Also, the embedded IMU allows the device to orientate its y-axis along the direction of the gravity vector. Therefore, the ground-plane is always perpendicular to the y-axis (under the assumption of a non-skewed floor) and fully described by the one-parameter equation y = y G .
Once the MEC receives the first (simplified) mesh from the AR device, y G is estimated in real-time by analyzing the histogram of vertices’ y-coordinates (Figure 2). Given that the depth sensor cannot detect any point below the ground-plane (except for overestimation errors), and with the floor being the largest flat surface in the bottom part of the scene, we expect to observe a considerably higher bin among those related to the lowest elevations.
The number of histogram bins is set to
k = y max y min δ ,
where y max and y min are the y-coordinates of the extreme vertices of the mesh along the vertical dimension.
Let m y be the count of the histogram bin associated to the elevation interval
y δ 2 , y + δ 2
and let
Y L = [ y min , y min + p ( y max y min ) ]
be a predefined range of elevations, from the lowest vertex y min up to a given percentage p of the full span. The ground-plane parameter y G is estimated as
y G = arg max y Y L m y
Once the elevation of the ground-plane is calculated, the algorithm removes any vertex v from the environmental mesh such that y v y G + δ . Every object lying on the floor, therefore, becomes an independent submesh.

4.3. Objects Detection

The problem of detecting isolated submeshes within the global mesh may be reinterpreted as a graph theory problem, leveraging the concept of connected components.
Let M = ( V , T ) be a mesh object as defined in Section 4.1. One can build an undirected graph G = ( V , E ) where every vertex in G corresponds to a vertex in M; an edge e E exists between two vertices if the same are connected by a triangle side in M. By doing this, isolated submeshes in M’s domain translate into connected components in G’s one.
Definition 1.
A connected component (or simply component) of an undirected graph is a subgraph in which any two vertices are connected to each other by a path of edges, and which is connected to no additional vertices in the supergraph.
Computing all the components of an undirected graph is achievable in linear time, O ( | V | + | E | ) , using either breadth-first or deep-first search [33]. While looping through all vertices in V, a new search is started whenever the loop encounters a vertex v that has not yet been included into a previously found component. The search will then find the entire component containing v before returning.
Our implementation of the algorithm (using deep-first search) is provided in Algorithm 2. It is straightforward to see how this approach returns in fact every connected component of the mesh-graph: along with actual objects and obstacles, walls themselves, for instance, constitute a component. We observed, however, that uninteresting outcomes can be easily filtered out by an adjustable threshold on components’ dimensions (see Section 6.2).
The set of identified obstacles is now sent back over the 5G link to the AR device, in the form of a list of 3D coordinates, allowing for such points to be highlighted to the operator by positioning holograms corresponding to their spatial locations.
Algorithm 2 Connected components retrieval algorithm for undirected graphs. Requires: G = ( V , E ) , undirected graph. Ensures: S, set of G’s connected components.
 1: procedure Components( G = ( V , E ) )
 2:    S
 3:   for v V do
 4:       if v not visited then
 5:         C ▹ New component
 6:        RecursDFS( v , C )
 7:       end if
 8:   end for
 9:   return S
 10: end procedure
 11: function RecursDFS( v , C )▹ Deep-First Search
 12:  mark v as visited
 13:   C C { v }
 14:  for x adjacent to v do
 15:   if x not visited then
 16:    RecursDFS( x , C )
 17:   end if
 18:  end for
 19: end function

5. Thermal Sensing Integration

The additional challenge of our framework consisted in the integration of an external thermal sensor cooperating with the scene understanding software in order to detect and signal the presence of harmful heat sources. To do so, we installed a commercial thermal camera on the helmet and calibrated it with the AR device’s embedded depth sensor.
We approached the problem of cross-calibrating the two-sensor system as a stereo calibration [34]. Since ToF sensors allow the acquisition of reflectivity information alongside with depth (Section 6) and the adopted thermal camera also embeds a standard RGB one, by treating reflectivity maps as gray-scale images, we can straightforwardly calibrate the system using a checkerboard pattern [35].
The calibration parameters we are interested in are as follows:
  • K D , K T , the intrinsic parameters of the two cameras;
  • R , t , the extrinsics, i.e., the rototranslation of one camera with respect to the other;
  • F , the fundamental matrix of the two-view system.
From these, we can write down the projective models for the two cameras as
x D = P D X = K D [ I | 0 ] X ,
x T = P T X = K T [ R | t ] X ,
where X = ( X , Y , Z , 1 ) is a real-world 3D point in homogeneous coordinates; x = ( x , y , 1 ) is the projected point on the image plane; P D , P T are the respective camera matrices. Note, that we are adopting the center of projection of the depth camera as the origin of our world coordinates.
The two-sensor system is now described by the epipolar geometry [36], which models the relations between the observed 3D points in the scene and their projections onto the two camera planes, namely π D and π T (Figure 3a). In particular, given a generic world-point X , the plane identified by X and the two centers of projection, C D , C T , is called the epipolar plane. The projections of X onto π D and π T , namely x D and x T , lie themselves on the epipolar plane.
These geometric relations are utilized in the following paragraph to describe the algorithm designed to calculate the 3D coordinates of a heat-point given its 2D projection on the thermal camera plane. We leave the criterion by which to select such heat-points out of the discussion, as it is usually application-dependent, and we focus instead on the procedure to be applied for a generic given pixel on the thermal image.

5.1. Heat-Point Triangulation

Let x ^ T be a point on π T . Our aim is to calculate the 3D point X ^ that gets projected into x ^ T . To reconstruct the 3D position of a point in π T , we first need to find its corresponding point x ^ D on π D . From epipolar geometry, we know that such a point lies on the epipolar line, i.e., the projection on π D of the optical ray through C T and X ^ . Given the fundamental matrix, F , obtained by calibration, the parametric representation of the epipolar line in homogeneous coordinates is
l D = F x ^ T
For each point x D = ( x D , y D , 1 ) on line l D , we can now invert the relation in (3),
r = K D 1 x D
obtaining the directing vector of the optical ray r = ( r X , r Y , r Z ) .
Since each point in π D is associated to a depth value d—equal to the distance of the 3D point from the center of projection—we are able to unequivocally reconstruct X ’s coordinates through trigonometric relations. From Figure 3b, one can derive the following identities:
tan θ = r Z r X 2 + r Y 2 , tan α = r Y r X
and hence the 3D coordinates of X are given by
X ˜ = d cos θ cos α d cos θ sin α d sin θ , X = X ˜ 1
where X ˜ denotes X in Cartesian coordinates.
Given the set of reconstructed X points, it is possible to reproject them on π T using (4), obtaining a set of x T points lying on the epipolar line l T . By taking account for reprojection errors, we can finally select X ^ among the reconstructed points as the one whose projection x T = P T X minimizes the distance from the original point x ^ T on the thermal camera plane,
X ^ = arg min X P T X x ^ T
Note that the total number of operations required by the algorithm is proportional to the number of points on the epipolar line, which is quadratically less complex than searching through the whole image. This feature helps speed up the triangulation of a 3D point dramatically, allowing a real-time response from the detection of a heat-point on the thermal plane to the reprojection and signaling of the same on the AR viewer.

6. Prototyping and Testing

Assessing the performance of the proposed assistive maintenance framework first required the realization of a prototype. The employed AR device (Microsoft HoloLens) was installed on the bottom end of a safety helmet (see Figure 4a) of the kind that is usually adopted for on-field maintenance operations. The helmet was also equipped with a FLIR AX8 thermal camera, calibrated along with HoloLens’ depth sensor as a two-view stereo system (see Section 5.1).
Microsoft HoloLens embeds a ToF depth sensor with two infrared illuminators, namely short-throw (up to ∼1 m) and long-throw (up to ∼3 m), operating on different fields of view and acquiring depth and reflectivity information simultaneously with a resolution of 450 × 448 pixels (see Figure 5 for sample frames).
FLIR AX8 thermal camera acquires 80 × 60 thermal images with a temperature range from −10 C to 150 C and a field of view (FOV) of 48 × 37 ; it acquires standard RGB images as well with the same FOV and a resolution of 640 × 480 pixels (see Figure 6 for sample frames).

6.1. Testing Environment

Obstacle detection tests were conducted with four human testers inside a Vodafone Base Transceiver Station (BST) to simulate real-case maintenance tasks.
Each tester was located inside the BTS wearing the prototype AR helmet. HoloLens was connected wirelessly to the CPE and two computers were connected to the same network: a terminal for the remote assistant in the control room; the MEC, hosting the developed algorithms for obstacle detection. Two applications were running on HoloLens: a video call app, allowing the communication between the on-field tester and the remote assistant; the obstacle visualization app, which receives 3D coordinates from the MEC and projects holograms onto the AR viewer in correspondence of the spatial locations of the detected objects. Both on-board and MEC applications were developed in Unity.
Thermal sensing integration tests were conducted during the Milano Digital Week 2019—held in March 2019, in Milan, Italy—on the occasion of the first presentation of the prototype. Tests were carried out by utilizing an incandescent lamp that was being detected by the thermal camera and whose coordinates were correctly estimated by the algorithm described in Section 5.1 and displayed by a 3D holographic marker on the AR viewer. The choice of which points to reproject from the thermal camera plane for 3D triangulation was done according to a temperature threshold of 60 C.

6.2. Obstacle Detection Results

For the purpose of implementing a reliable framework for assistive maintenance, we addressed a particular focus to the real-time response of the obstacle detection system. In order to evaluate it, we used as Key Performance Indicator (KPI) the mesh processing time in relation to the number of points and triangles.
We set the quantization parameter for mesh simplification (also used as the histogram bin size for ground-plane extraction) to δ = 10 cm (see Section 4.1). The percentage of elevations used to detect the histogram peak associated to the ground-plane was set to p = 10 % . Among the retrieved mesh-graph components, we considered obstacles to be those lying within 1 m from the floor-plane.
Table 1 reports some results on the measured computation times for the tasks of mesh simplification, ground-plane extraction and object detection, in relation to the number of vertices (before and after reduction), triangles and detected obstacles. Measurements were collected sequentially, as the mesh acquired by the AR device grows while the environment is being explored. From these results, it is evident how mesh simplification helps to stabilize the total computation time: while the number of vertices in the non-processed mesh grows from thousands to tens of thousands, the overall time required for processing the scene remains below 100 ms, on average. Additionally, the fraction of computation time needed for detecting the mesh-graph components is most of the time negligible. It may surprise to observe that the number of detected obstacles do not follow a strictly non-decreasing trend, but this is in fact perfectly in line with the way HoloLens creates and refines environmental meshes: since new acquisitions are constantly being incorporated into the current model, it happens quite often to see an object being first acquired as multiple disconnected components and then being completed within the next few frames. The bandwidth and the latency of the 5G network, together with their counterpart in the 802.11ac WLAN, did not introduce any delay both in the real-time bidirectional data streaming to and from the HoloLens. An exhaustive 5G transmission speed and latency tests were not performed since the whole use-case was tested in an experimental 5G setup with limited performances with respect to the final operative deployment; however, the scene processing together with the remote call interaction and the proper hologram placement was always seamless and no disruptive delays occurred.

6.3. Cross-Sensor Calibration

Cross-calibration of depth and thermal cameras was carried out with a standard 7 × 9 checkerboard pattern, 26.5 mm per square, moved at a distance of about 1 m from the prototype helmet. Pictures of the pattern were taken using the depth sensor’s reflectivity camera and the thermal sensor’s RGB camera. Calibration was performed in accordance with the pipeline established by [35]: (i) estimate the intrinsic and extrinsic parameters in closed-form solution; (ii) refine all parameters, including lens distortion, with a non-linear minimization. We used the implementation available in Matlab R2019a.
The procedure returned the following reprojection errors, in unit of pixels:
ϵ D = 0.1571 , ϵ T = 0.2542 ,
with an average error ϵ ¯ = 0.2056 . A visualization of the estimated extrinsic parameters is provided in Figure 7, where one can verify how the thermal camera (in red) correctly resembles its collocation on top of the helmet, above the AR viewer and the associated depth camera (in blue).
Given the obtained sub-pixel reprojection error, we can conclude that a standard stereo calibration procedure is suitable as well for cross-calibrating heterogeneous sensors, combing visible and reflectivity information.

6.4. Remote Assistance

In view of a fully-capable on-field adoption, the system was also equipped with an augmented reality remote assistance service that allows the on-field operator to interact with the remote expert. The two share a real-time audio and video streaming based on Skype for HoloLens® with virtual objects and 3D markers in overlay, as depicted in Figure 8.

7. Conclusions

Augmented reality and 5G networking are paving the way to a whole new set of possibilities in the progressive digitization of modern industries. Remote assistive maintenance based on egocentric environment-aware devices represents a viable solution to improve workers’ safety in dangerous locations and the communication efficiency between on-site operators and remote assistants.
In this paper, we presented a fully-developed assistive maintenance system based on a wearable AR device equipped with environment-understanding and thermal sensors. A set of algorithms was specifically designed to detect potentially dangerous objects and heat-sources, while the holographic viewer enables the visualization of such information directly in the operator’s field of view. Finally, deploying the analysis software on an edge computing platform connected to the device via a 5G link, allows for all the required processing to be performed in real-time.

Author Contributions

Conceptualization, M.M. and S.M.; Formal analysis, S.V. and M.M.; Funding acquisition, M.M. and S.T.; Investigation, S.V. and M.M.; Methodology, M.M. and S.M.; Resources, M.M.; Software, S.V.; Supervision, M.M.; Writing—original draft, S.V., M.M. and S.M.; Writing—review and editing, S.V., M.M., S.M. and S.T. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Acknowledgments

The 5G Programme is founded by the Italian Ministry of Economic Development (MISE). The activity was accomplished in collaboration with Vodafone Italia S.p.A.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Mosiello, G.; Kiselev, A.; Loutfi, A. Using Augmented Reality to Improve Usability of the User Interface for Driving a Telepresence Robot. Paladyn J. Behav. Robot. 2013, 4, 174–181. [Google Scholar] [CrossRef]
  2. Brizzi, F.; Peppoloni, L.; Graziano, A.; Di Stefano, E.; Avizzano, C.A.; Ruffaldi, E. Effects of Augmented Reality on the Performance of Teleoperated Industrial Assembly Tasks in a Robotic Embodiment. IEEE Trans. Hum. Mach. Syst. 2018, 48, 197–206. [Google Scholar] [CrossRef]
  3. Han, B.; Wong, S.; Mannweiler, C.; Crippa, M.R.; Schotten, H.D. Context-Awareness Enhances 5G Multi-Access Edge Computing Reliability. IEEE Access 2019, 7, 21290–21299. [Google Scholar] [CrossRef]
  4. Sutherland, I.E. A head-mounted three dimensional display. In Proceedings of the December 9–11, 1968, Fall Joint Computer Conference, Part I; Association for Computing Machinery: New York, NY, USA, 1968; pp. 757–764. [Google Scholar]
  5. Carlsson, C.; Hagsand, O. DIVE A multi-user virtual reality system. In Proceedings of the IEEE Virtual Reality Annual International Symposium, Seattle, WA, USA, 18–22 September 1993; pp. 394–400. [Google Scholar]
  6. Greenhalgh, C.; Benford, S. MASSIVE: A collaborative virtual environment for teleconferencing. ACM Trans. Comput. Hum. Interact. 1995, 2, 239–261. [Google Scholar] [CrossRef]
  7. Schmalstieg, D.; Fuhrmann, A.; Hesina, G.; Szalavári, Z.; Encarnação, L.M.; Gervautz, M.; Purgathofer, W. The Studierstube Augmented Reality Project. Presence Teleoperators Virtual Environ. 2002, 11, 33–54. [Google Scholar] [CrossRef]
  8. Billinghurst, M.; Poupyrev, I.; Kato, H.; May, R. Mixing realities in Shared Space: An augmented reality interface for collaborative computing. In Proceedings of the International Conference on Multimedia and Expo, New York, NY, USA, 30 July–2 August 2000; pp. 1641–1644. [Google Scholar] [CrossRef]
  9. Orlosky, J.; Kiyokawa, K.; Takemura, H. Virtual and Augmented Reality on the 5G Highway. J. Inf. Process. 2017, 25, 133–141. [Google Scholar] [CrossRef] [Green Version]
  10. Bolas, M.; Hoberman, P.; Phan, T.; Luckey, P.; Iliff, J.; Burba, N.; McDowall, I.; Krum, D.M. Open virtual reality. In Proceedings of the 2013 IEEE Virtual Reality (VR), Lake Buena Vista, FL, USA, 18–20 March 2013; pp. 183–184. [Google Scholar]
  11. Lavingia, K.; Tanwar, S. Augmented Reality and Industry 4.0. In A Roadmap to Industry 4.0: Smart Production, Sharp Business and Sustainable Development; Nayyar, A., Kumar, A., Eds.; Springer International Publishing: Cham, Switzerlands, 2020; pp. 143–155. [Google Scholar]
  12. Condino, S.; Fida, B.; Carbone, M.; Cercenelli, L.; Badiali, G.; Ferrari, V. Wearable Augmented Reality Platform for Aiding Complex 3D Trajectory Tracing. Sensors 2020, 20, 1612. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. Huang, B.C.; Hsu, J.; Chu, E.T.H.; Wu, H.M. ARBIN: Augmented Reality Based Indoor Navigation System. Sensors 2020, 20, 5890. [Google Scholar] [CrossRef] [PubMed]
  14. Mateu, J.; Lasala, M.J.; Alamán, X. Developing mixed reality educational applications: The virtual touch toolkit. Sensors 2015, 15, 21760–21784. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Chien, J.C.; Tsai, Y.R.; Wu, C.T.; Lee, J.D. HoloLens-Based AR System with a Robust Point Set Registration Algorithm. Sensors 2019, 19, 3555. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Wang, S.; Parsons, M.; Stone-McLean, J.; Rogers, P.; Boyd, S.; Hoover, K.; Meruvia-Pastor, O.; Gong, M.; Smith, A. Augmented reality as a telemedicine platform for remote procedural training. Sensors 2017, 17, 2294. [Google Scholar] [CrossRef]
  17. Newcombe, R.; Izadi, S.; Molyneaux, D.; Hilliges, O.; Kim, D.; Shotton, J.D.J.; Kohli, P.; Fitzgibbon, A.; Hodges, S.E.; Butler, D.A. Mobile Camera Localization Using Depth Maps. U.S. Patent No. 8,711,206, 29 April 2014. [Google Scholar]
  18. Howard, I.P. Seeing in Depth, online-ausg ed.; Oxford Scholarship Online; Oxford University Press: New York, NY, USA; Oxford, UK, 2008. [Google Scholar]
  19. Landy, M.S.; Maloney, L.T.; Johnston, E.B.; Young, M. Measurement and modeling of depth cue combination: In defense of weak fusion. Vis. Res. 1995, 35, 389–412. [Google Scholar] [CrossRef] [Green Version]
  20. Swan, J.E.; Jones, A.; Kolstad, E.; Livingston, M.A.; Smallman, H.S. Egocentric depth judgments in optical, see-through augmented reality. IEEE Trans. Vis. Comput. Graph. 2007, 13, 429–442. [Google Scholar] [CrossRef]
  21. Cutting, J.E. How the eye measures reality and virtual reality. Behav. Res. Methods Instruments Comput. 1997, 29, 27–36. [Google Scholar] [CrossRef] [Green Version]
  22. Kishishita, N.; Kiyokawa, K.; Orlosky, J.; Mashita, T.; Takemura, H.; Kruijff, E. Analysing the effects of a wide field of view augmented reality display on search performance in divided attention tasks. In Proceedings of the 2014 IEEE International Symposium on Mixed and Augmented Reality (ISMAR), Munich, Germany, 10–12 September 2014; pp. 177–186. [Google Scholar]
  23. Loomis, J.M.; da Silva, J.A.; Fujita, N.; Fukusima, S.S. Visual space perception and visually directed action. J. Exp. Psychol. Hum. Percept. Perform. 1992, 18, 906–921. [Google Scholar] [CrossRef] [PubMed]
  24. Swan, J.E.; Livingston, M.A.; Smallman, H.S.; Brown, D.; Baillot, Y.; Gabbard, J.L.; Hix, D. A Perceptual Matching Technique for Depth Judgments in Optical, See-Through Augmented Reality. In Proceedings of the IEEE Virtual Reality Conference (VR 2006), Alexandria, VA, USA, 25–29 March 2006; pp. 19–26. [Google Scholar]
  25. Rehman, U.; Cao, S. Augmented-Reality-Based Indoor Navigation: A Comparative Analysis of Handheld Devices Versus Google Glass. IEEE Trans. Hum. Mach. Syst. 2016, 47, 140–151. [Google Scholar] [CrossRef]
  26. Pham, Q.V.; Fang, F.; Ha, V.N.; Piran, M.J.; Le, M.; Le, L.B.; Hwang, W.J.; Ding, Z. A survey of multi-access edge computing in 5G and beyond: Fundamentals, technology integration, and state-of-the-art. IEEE Access 2020, 8, 116974–117017. [Google Scholar] [CrossRef]
  27. ABI Research; Qualcomm. Augmented and Virtual Reality: The First Wave of 5G Killer Apps. 2017. Available online: https://www.qualcomm.com/news/onq/2017/02/01/vr-and-ar-are-pushing-limits-connectivity-5g-our-rescue (accessed on 11 December 2020).
  28. Erol-Kantarci, M.; Sukhmani, S. Caching and Computing at the Edge for Mobile Augmented Reality and Virtual Reality (AR/VR) in 5G. In Ad Hoc Networks; Zhou, Y., Kunz, T., Eds.; Springer International Publishing: Cham, Switzerlands, 2018; pp. 169–177. [Google Scholar]
  29. IEEE Computer Society. IEEE Std 802.11ac-2013 (Amendment to IEEE Std 802.11-2012, as amended by IEEE Std 802.11ae-2012, IEEE Std 802.11aa-2012, and IEEE Std 802.11ad-2012); Institute of Electrical and Electronics Engineers: New York, NY, USA, 2013. [Google Scholar]
  30. Unity Technologies. Unity Real-Time Development Platform. Available online: https://unity.com (accessed on 11 December 2020).
  31. Cignoni, P.; Montani, C.; Scopigno, R. A comparison of mesh simplification algorithms. Comput. Graph. 1998, 22, 37–54. [Google Scholar] [CrossRef]
  32. Rossignac, J.; Borrel, P. Multi-resolution 3D approximations for rendering complex scenes. In Modeling in Computer Graphics; Springer: Berlin/Heidelberg, Germany, 1993; pp. 455–465. [Google Scholar]
  33. Hopcroft, J.; Tarjan, R. Algorithm 447: Efficient algorithms for graph manipulation. Commun. ACM 1973, 16, 372–378. [Google Scholar] [CrossRef]
  34. Klette, R. Cameras, Coordinates, and Calibration. In Concise Computer Vision; Springer: Berlin/Heidelberg, Germany, 2014; Chapter 6; pp. 215–240. [Google Scholar]
  35. Zhang, Z. A flexible new technique for camera calibration. IEEE Trans. Pattern Anal. Mach. Intell. 2000, 22, 1330–1334. [Google Scholar] [CrossRef] [Green Version]
  36. Hartley, R.; Zisserman, A. Epipolar Geometry and the Fundamental Matrix. In Multiple View Geometry in Computer Vision; Cambridge University Press: Cambridge, UK, 2003; Chapter 9; pp. 239–259. [Google Scholar]
Figure 1. The proposed Advanced Assistive Maintenance framework. An AR viewer, equipped with a depth sensor and a thermal camera, sends multi-sensor information over an IEEE 802.11ac WLAN to a customer-premises equipment (CPE). The CPE is then connected through a 5G link to a Multi-access Edge Computer (MEC) that processes the received data to detect heat-sources and obstacles. Coordinates of detected points are sent back to the helmet and visualized in real-time. The same connection is also used for an interactive video call.
Figure 1. The proposed Advanced Assistive Maintenance framework. An AR viewer, equipped with a depth sensor and a thermal camera, sends multi-sensor information over an IEEE 802.11ac WLAN to a customer-premises equipment (CPE). The CPE is then connected through a 5G link to a Multi-access Edge Computer (MEC) that processes the received data to detect heat-sources and obstacles. Coordinates of detected points are sent back to the helmet and visualized in real-time. The same connection is also used for an interactive video call.
Sensors 20 07157 g001
Figure 2. Histogram of vertex elevations in the acquired environmental mesh. The ground–plane is estimated as the middle–point of the most numerous bin (10 cm wide) among the first 10% of the mesh’s elevation range.
Figure 2. Histogram of vertex elevations in the acquired environmental mesh. The ground–plane is estimated as the middle–point of the most numerous bin (10 cm wide) among the first 10% of the mesh’s elevation range.
Sensors 20 07157 g002
Figure 3. Epipolar geometry relations between the calibrated depth and thermal cameras (a) and 3D coordinates reconstruction of a given point X with known depth value (distance from the center of projection of the depth camera, C D ) and optical ray r (b).
Figure 3. Epipolar geometry relations between the calibrated depth and thermal cameras (a) and 3D coordinates reconstruction of a given point X with known depth value (distance from the center of projection of the depth camera, C D ) and optical ray r (b).
Sensors 20 07157 g003
Figure 4. Prototype helmet equipped with AR smartglasses and thermal camera (a); thermal image combined with the warning hologram reprojected in the HMD display (b).
Figure 4. Prototype helmet equipped with AR smartglasses and thermal camera (a); thermal image combined with the warning hologram reprojected in the HMD display (b).
Sensors 20 07157 g004
Figure 5. Example frames from HoloLens’ ToF depth camera. The sensor acquires 450 × 448 depth and reflectivity maps, in short-throw (ST), i.e., up to ∼1 m, and long-throw (LT), i.e., up to ∼3 m.
Figure 5. Example frames from HoloLens’ ToF depth camera. The sensor acquires 450 × 448 depth and reflectivity maps, in short-throw (ST), i.e., up to ∼1 m, and long-throw (LT), i.e., up to ∼3 m.
Sensors 20 07157 g005
Figure 6. Example frames from the FLIR AX8 thermal camera. The sensor acquires 80 × 60 thermal maps and 640 × 480 RGB images.
Figure 6. Example frames from the FLIR AX8 thermal camera. The sensor acquires 80 × 60 thermal maps and 640 × 480 RGB images.
Sensors 20 07157 g006
Figure 7. Extrinsic parameters visualization for the cross–calibration of ToF–sensor (blue) and thermal camera (red) using a moving checkerboard pattern, combining ToF reflectivity with thermal RGB information.
Figure 7. Extrinsic parameters visualization for the cross–calibration of ToF–sensor (blue) and thermal camera (red) using a moving checkerboard pattern, combining ToF reflectivity with thermal RGB information.
Sensors 20 07157 g007
Figure 8. A snapshot of a remote call with real-time augmented reality interaction: (left) the on-field operator follows the instructions provided by the remote assistant; (right) the monitor of the remote assistant displays a replica of the AR scene as seen by the operator; the devices involved in the maintenance intervention are highlighted by superimposing and visualizing the respective holograms on the HMD.
Figure 8. A snapshot of a remote call with real-time augmented reality interaction: (left) the on-field operator follows the instructions provided by the remote assistant; (right) the monitor of the remote assistant displays a replica of the AR scene as seen by the operator; the devices involved in the maintenance intervention are highlighted by superimposing and visualizing the respective holograms on the HMD.
Sensors 20 07157 g008
Table 1. Measured computation times for scene-processing tasks, namely: mesh simplification, ground-plane extraction and objects detection (via connected components retrieval). The acquisition is sequential: the number of mesh vertices (first column) grows as the environment is being explored by the Head-Mounted Display (HDM). Thanks to mesh simplification, the overall computation time remains stable on average as the mesh size increases.
Table 1. Measured computation times for scene-processing tasks, namely: mesh simplification, ground-plane extraction and objects detection (via connected components retrieval). The acquisition is sequential: the number of mesh vertices (first column) grows as the environment is being explored by the Head-Mounted Display (HDM). Thanks to mesh simplification, the overall computation time remains stable on average as the mesh size increases.
Mesh ParametersComputation Time
No. VerticesNo. TrianglesNo. Reduced VerticesNo. Detected ObstaclesMesh Simplif. (ms)Ground-Plane (ms)Components (ms)Total (ms)
147524145181258639
485576561419752201082
792812,7642375104229<181
934415,1092914103332<175
10,01416,4443054143935<188
19,94933,8696164964611125
20,12520,125182863860<198
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Verde, S.; Marcon, M.; Milani, S.; Tubaro, S. Advanced Assistive Maintenance Based on Augmented Reality and 5G Networking. Sensors 2020, 20, 7157. https://doi.org/10.3390/s20247157

AMA Style

Verde S, Marcon M, Milani S, Tubaro S. Advanced Assistive Maintenance Based on Augmented Reality and 5G Networking. Sensors. 2020; 20(24):7157. https://doi.org/10.3390/s20247157

Chicago/Turabian Style

Verde, Sebastiano, Marco Marcon, Simone Milani, and Stefano Tubaro. 2020. "Advanced Assistive Maintenance Based on Augmented Reality and 5G Networking" Sensors 20, no. 24: 7157. https://doi.org/10.3390/s20247157

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop