Next Article in Journal
Heart Rate Variability in Psychology: A Review of HRV Indices and an Analysis Tutorial
Previous Article in Journal
Respiration Monitoring via Forcecardiography Sensors
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

First Step toward Gestural Recognition in Harsh Environments

1
Magic Lab, Department of Industrial Engineering and Management, Ben Gurion University of the Negev, P.O. Box 653, Beer-Sheva 8410501, Israel
2
Autonomous Systems Lab, University of California, Santa Cruz, CA 95064, USA
3
Department of Asian Studies, Faculty of Humanities, Haifa University, Haifa 3498838, Israel
*
Authors to whom correspondence should be addressed.
Sensors 2021, 21(12), 3997; https://doi.org/10.3390/s21123997
Submission received: 25 April 2021 / Revised: 31 May 2021 / Accepted: 1 June 2021 / Published: 9 June 2021
(This article belongs to the Special Issue Advanced Sensor Technology and Human-Computer Interaction)

Abstract

:
We are witnessing a rise in the use of ground and aerial robots in first response missions. These robots provide novel opportunities to support first responders and lower the risk to people’s lives. As these robots become increasingly autonomous, researchers are seeking ways to enable natural communication strategies between robots and first responders, such as using gestural interaction. First response work often takes place in harsh environments, which hold unique challenges for gesture sensing and recognition, including in low-visibility environments, making the gestural interaction non-trivial. As such, an adequate choice of sensors and algorithms needs to be made to support gestural recognition in harsh environments. In this work, we compare the performances of three common types of remote sensors, namely RGB, depth, and thermal cameras, using various algorithms, in simulated harsh environments. Our results show 90 to 96% recognition accuracy (respectively with or without smoke) with the use of protective equipment. This work provides future researchers with clear data points to support them in their choice of sensors and algorithms for gestural interaction with robots in harsh environments.

1. Introduction

Mobile ground and aerial robots (also known as drones) offer unique potential to support people in first response situations, such as in firefighting, search and rescue, police, and military situations. These robots are now sent to explore areas, perform tasks with or instead of humans, and thus reduce the risk to people’s lives. Recently, we have been seeing a technological shift, with robots presenting increasing levels of autonomy [1,2,3], meaning that they can now directly interact with first responders. Such technological advances have led human–computer and human–robot interaction (HCI/HRI) researchers to explore natural ways for first responders to communicate with these robots, with one of the most prevalent ways being gestural interaction (e.g., using finger, hand, or full-body gestures to command and instruct a robotic agent) [4,5,6]. However, the conditions experienced by first responders in their missions present non-trivial challenges for gestural interaction sensing and recognition, such as when working in low visibility (e.g., in smoke or darkness). These situations are referred to as “harsh environments”, which further require first responders to wear heavy protective equipment, presenting yet additional constraints for gesture sensing and recognition.
While there is a large body of prior work on gestural interaction, including sensing techniques and algorithms for recognition, the literature about gestural recognition in harsh environments is sparse. Research is needed to understand which sensors and algorithms are best suited to support gestural interaction in harsh environments. In this manuscript, we build upon the preliminary work conducted by De Cillis et al. [7], which investigated the use of a depth camera to sense three types of gestures that were then trained and tested with two people (one and one) in simulated harsh conditions. We take their approach further and compare a depth camera with two different types of visual sensors: RGB and thermal cameras. We also investigate a range of algorithms to enable gesture recognition with each of the three sensors, using user-defined gestures that were specifically chosen by operational firefighters for interacting with drones in firefighting missions. We trained our system with 13 people and tested it with two people in several conditions, including simulated harsh environments. This work provides a review of possible technological solutions that can be embedded in ground robots and drones for gesture recognition in harsh environments, including the advantages and drawbacks of the different sensors and algorithms.
In this manuscript, we first present a literature review and then describe a set of gestures adapted to human–drone interaction (HDI) in firefighting. We then present our implementation of a gesture recognition system, based on a range of algorithms, which are then tested with the three different types of sensors. We present the system’s evaluation in laboratory settings, simulating different conditions of harsh environments. Our results show major differences between the three types of sensors regarding the recognition accuracy for different simulated environments. For example, we found that while the overall recognition accuracy dropped under smokey conditions, the accuracy stayed high for RGB and thermal cameras, while the depth sensor failed to recognize gestures. Our results present opportunities for future research on remote sensing in harsh environments and inform future HCI and HRI researchers and practitioners on the human and technical aspects to consider when integrating gestural recognition systems in emergency responses. This work contributes to a larger understanding of gestural recognition in harsh environments, as well as a comparison between three common sensors and various algorithms that can be used for remote sensing in harsh environments.

2. Related Work

In the following section, we review the related prior work, first on ground and aerial robots, then on gestural recognition, and finally on remote sensors, all in harsh environments.

2.1. Ground and Aerial Robots in Harsh Environments

“Harsh environment” is a broad term that can refer to “any environment that is hazardous to agents (human or robot, etc.) within it. For example, they can be characterized by high levels of radiation, high explosive risk, extreme temperatures or pressures, and lack of oxygen” [8]. Harsh environments can also be defined as environments that are “unknown, unstructured, dynamic, cluttered, hazardous and/or limited in resources (such as the availability of communications, GPS and visibility)” [9]. These environments present unique challenges for integrating robots, such as open fire sources, dynamic obstacles, and faulty sensor readings [10]. In particular, harsh environments present many challenges for the remote sensing community, with issues such as low visibility due to smoke or darkness, which can disturb visual sensing, or noise, which can prevent audio sensing. Furthermore, such environments present additional challenges for controlling robots and drones—for example, when GPS signals might be unavailable—or even for a user to directly interact with the robotic agent while under pressure, with a high workload, and while potentially wearing heavy protective equipment.
The HCI and HRI communities have investigated and proposed technological solutions for drones and robots to support first responders in their missions, in harsh environments, such as in firefighting [11], search and rescue [12], police [13] work, and in the military [14]. In particular, in firefighting, drones and robots have been used for a variety of applications, such as searching for victims and checking the environment and surroundings in search and rescue missions [15], detecting and extinguishing fires [16,17], and also in mapping and detecting gas in both indoor and outdoor environments [18,19]. A recent report summarized the benefits of using drones for firefighting missions, including visual access to low-visibility areas, increased safety for firefighters during emergency missions, optimized data gathering processes, time-saving, and reduction of human error in inspections of a zone [20].
As autonomous drones and robots emerge [2,21], researchers are interested in the transition of these robots from a tool to a teammate [22,23]. This paradigm shift drives the need for collocated interaction strategies between robots and humans [24], with one of the best strategies for such collocated interaction being gestures [25,26,27]. In this work, we propose to explore remote sensing and algorithms for gestural interaction with robots in harsh environments, using the example of firefighting missions given the amount of prior supportive literature from the fields of HCI and HRI. The following subsection describes the current state-of-the-art for gestural interaction in harsh environments.

2.2. Gestural Recognition in Harsh Environments

Gestural interaction in harsh environments holds major constraints for sensing visual stimuli and making sense of the information. Indeed, low-visibility conditions (e.g., smoke, darkness) can disturb the sensing. In addition, humans interacting with the system may be wearing heavy protective equipment, so that finger, hand, and body tracking may not be achieved with regular image segmentation (e.g., skeleton extraction). The existing body of work on gestural recognition in harsh environments is sparse. After extensively surveying the literature for first responders, we only found one prior work investigating supporting firefighters with robots using gestures [7]. The work proposed using a depth camera to sense three types of gestures (simple, medium to challenging, and hard gestures) that were then tested in simulated harsh conditions in a lab (i.e., darkness and smoke, indoor and with uniform). Unfortunately, while the work contributed a framework for gesture recognition in HRI, the manuscript did not adequately report on the gestures that were implemented, their suitability for firefighting, the conditions of the evaluation, or the exact simulated environment. Therefore, further research is needed to investigate which sensors and algorithms are most appropriate in harsh conditions and to propose adapted settings for laboratory simulation of harsh environments.
To conduct this research, we further report on the large body of literature on gestural recognition and remote sensing, focusing on the types of gestures and the recognition algorithms. Gestural interaction can be based on various body parts, from full-body [28], to arm [29], hand, and foot gestures [30], and even finger movements [31]. Escalera et al. [32] created a taxonomy for the various components involved in conducting research in gesture recognition and proposed that two categories of gestures can be distinguished: 1. Static gestures that represent a still pose in a single image, and 2. Dynamic gestures are represented by a sequence of images. Gestural interaction distinguishes between gestures that give commands to an agent and gestures that are meant to initialize or terminate the interaction, as is the case in our work. For example, prior work used the familiar arm-waving gesture as a signal to attract a UAV’s attention [26,33]. In our work, we propose to use a set of gestures that were elicited following a user-centered design process with firefighters [23].
In terms of interaction, many prior works investigated gestural interaction systems with computers, robots, and drones [34,35,36,37]. The literature offers a variety of approaches for gestural recognition, which are described in a recent survey [38], where the authors categorized algorithms into three categories: hand-crafted motion features, deep learning-based methods, and depth information-based methods. It is, however, clear from their work that the different methods are not equivalent and hold advantages and drawbacks based on the types of gestures to recognize, situations, and chosen sensors. As firefighting scenes may have smoke and generally low visibility, they need special care with regard to the choice of sensor. In the following section, we discuss sensors for remote gesture sensing in harsh environments.

2.3. Remote Sensing in Harsh Environments

Different types of visual sensors exist, each relying on different technologies, working across a range of the light spectrum. As such, different sensor types will perform differently in harsh environments, some working in daylight, others in darkness, or even able to sense the environmental temperature. In their work, Berman and Stern [39] created a taxonomy for sensors for gesture recognition systems and showed that optical sensors are traditionally used for recognizing gestures. Three main types of optical sensors are used in the literature [32]: RGB [40], thermal [41], and depth [42] sensors. In harsh environments, only a few works have compared sensors, such as Starr and Lattimer [43], who evaluated seven different types of sensors in fire and smoke conditions—in terms of performance, but not for gesture sensing. They divided their experiment into two categories of harsh environments: 1. light smoke with high temperature, and 2. dense smoke with low temperature. In the former, RGB and thermal sensors worked but the depth camera failed, and in the latter, the RGB sensor failed below 1 m visibility and encountered attenuation past 8 m visibility; the depth sensor showed poor results even with a visibility over 8 m, while the thermal sensor was not affected by the conditions. These results suggest that a thermal sensor may be the best choice for a gestural interaction system in firefighting, where thick smoke and darkness can be expected, and that RGB and depth sensors will be more likely to fail when smoke is present.
In our work, we propose to compare RGB, depth, and thermal sensors that can later be embedded into a drone or robot, and to evaluate their performance for gesture recognition in simulated firefighting conditions. We expect that the heavy protective equipment and smoke conditions will decrease the recognition accuracy, compared to optimal conditions. Furthermore, we will investigate how the simulated harsh environment conditions will affect the various sensors and algorithms that we implement. The next section describes the chosen set of gestures.

3. Gestures for Gestural Communication in Harsh Environments

In this section, we describe our choice of gestures to be implemented and evaluated. We opted for gestures that were previously elicited in a user study with firefighters [23], which focused on gestures for collocated interaction with a drone. We first describe the tasks allocated to the drone, as identified by the firefighters, and their matching gestures to instruct the drone. The four main drone tasks and associated user-defined gestures, as described in [23], are presented below:
  • Mapping. The firefighters envisioned sending the drone to map a floor or even an entire building, finding out how many rooms are inside a building, what the structure of the building looks like, and creating a map of it.
  • Identifying Hazardous Materials. The firefighters proposed that the drone could identify the type of hazardous material (such as gas or chemicals) and the source and size of the leak, which is a scenario that they encounter in industrial buildings.
  • Detecting Fires. The firefighters proposed that the drone could identify fire spots. This is in line with current drone usage in the country.
  • Finding Survivors. The firefighters suggested sending the drone to search for trapped civilians or people otherwise in danger. They would like the drone to gather information on the number of people as well as their condition and location. They proposed that the drone could directly provide help to citizens who have fallen into holes or deep tunnels. The firefighters further mentioned a situation that they referred to as “fear for human life and aid to civilians”. This corresponds to situations where there is a concern for civilians, who are locked in their homes and are not responding to the door, who usually have not been seen by their relatives or neighbors, who called the firefighters to break in. Participants mentioned that a drone could fly inside an apartment through a window to help them to conclude on whether they need to break into the apartment or not.
The drone tasks are summarized in Table 1, and the elicited gestures chosen for each task (see Figure 1) include: Detecting Fire: lifting the left hand and rotating it clockwise with two fingers up; Identifying hazardous materials: raising two hands and crossing them in front of the face; Finding survivors: lifting the right hand and rotating it clockwise with two fingers up (same as Detecting Fire but with the right hand); Mapping: drawing a “frame” in front of the body using both hands symmetrically with two fingers pointing up. The firefighters chose gestures that are continuous and repeated, meaning that they will be performed until the desired command is detected (i.e., when they receive feedback that the gesture was properly understood).
Based on the physical load of the equipment that firefighters carry, one could expect the resulting gestures to be one-handed, small, and fast, saving firefighters from any additional physical and cognitive efforts to communicate with the drone. However, Alon et al. [23] revealed that, regardless of the constraints, the firefighters preferred large, noticeable, and continuous gestures, some of them two-handed. They further highlighted that these gestures are consistent with those of other first responders and military environments [44,45]. We are thus confident that these chosen gestures are representative of wider first response environments.

4. Gesture Recognition System Prototyping

We designed and implemented a gesture recognition system prototype for the user-defined gestures shown in Figure 1. In addition to the set of gestures, we trained the system to recognize if a person is present in front of the camera and whether they are walking or standing still. Our goal was to survey which technology is best suited for gesture recognition in harsh conditions. In this section, we first present the implemented sensors and algorithms that were used in our experiment. We then describe the data collection and training of the system, and finally, we describe the evaluation process.

4.1. Visual Sensors

In terms of hardware for image acquisition, we compared three types of sensors, namely RGB, depth, and thermal cameras, as these are the most widespread sensors in the gestural recognition literature (e.g., RGB [46], depth [42], and thermal [47]). We used an Intel Realsense D435 [48] camera for RGB (frame rate: 30 Hz, resolution: 1920 × 1080, field of view: 69° × 42° (H × V)) and depth (frame rate: 90 Hz, resolution: 1280 × 720, field of view: 87° × 58° (H × V)) images, and a Therm-App HZ [49] for the thermal image (frame rate: 25 Hz, resolution: 384 × 288, field of view: 55° × 41° (H × V)). Figure 2 shows resulting frames from the depth and thermal sensors.

4.2. Algorithms

Before a system can be implemented in a robot or a drone, it needs to be validated in static conditions [37,50]. We apply this strategy to our work and present the system in static conditions. We empirically determined that the longest duration for a gesture took an average of 3 s to perform (see Section 4.3), and as such, we sampled data at a rate of 20 frames every 3 s (i.e., 6.7 Hz). The prediction was based upon the changes within the series of frames extracted in the 3 s window. We then implemented three algorithms, which we further describe below. The first one is based on skeleton extraction and the second one on a histogram of oriented gradient (HOG). Both algorithms use support vector machine (SVM) for the classification as this supervised learning model is widely used in the gesture recognition literature, both with skeleton features [51] and with HOG [52], and as it is suitable for high-dimensional representation. For the SVM, we used the scikit-learn API in Python, with linear kernel and C (regularization parameter) equal to 10. The third algorithm consists of a bespoke algorithm, “Frame Vote”, that we developed, and which can use either HOG and SVM or convolutional neural networks (CNN). Our intention was to compare several approaches to the traditional classification method of HOG + SVM often used in UAVs [53].
  • Skeleton Extraction. Key points of the person are extracted using the Realsense SDK (Cubemos), which provides a vector of 18 (x,y) coordinates for each frame. Each gesture consists of 20 frames which form a vector of 720 values (20 frames × 18 coordinates × 2 (x,y) values). Skeleton extraction is a known approach in the literature for extracting features for body gesture recognition (e.g., [54]) and it can be implemented with various APIs, such as OpenPose [55]. Compared to other feature extraction techniques, such as using 3D CNNs, which are time-consuming and may be difficult to train, the features of the human skeleton are concise and are based on a pre-trained CNN [54].
  • Histogram of oriented gradient (HOG). Each frame is represented by a HOG [56], so that each gesture forms a vector of 20 concatenated HOG vectors. We used scikit-image for extracting the HOG features with the number of pixels per cell = 10 × 10 (size of a cell (in pixels)), number of orientation bins = 9, number of cells in each block = 2 × 2, and using the L2 block normalization method [57]. To improve the accuracy, we used background subtraction using motion analysis by applying BackgroundSubtractorMOG2 [58] (OpenCV) prior to the HOG extraction. The following parameters were used: history = 150 (number of last frames that affect the background model), varThreshold = 40 (variance threshold for the pixel-model match), and detectShadows = False.
  • Frame Vote. We designed this algorithm to analyze each frame individually, using a different paradigm compared to the previous two algorithms. For each frame, we run a machine learning algorithm (HOG + SVM or CNN) that classifies the gesture. After 20 frames, the predicted gestures are compared and the one with the most occurrences is selected (majority vote). The algorithm is illustrated in Figure 3. The advantage of this method is that it does not take into account the sequence of the gesture, and thus may hold better if the sensing continuity is not entirely reliable, as may be the case in harsh environments. This is also a good fit for the continuous and repeated gestures elicited in the previous study, as the order in which the gestures are performed has less importance.
The algorithm works as follows: a label (i.e., gesture) is predicted for each frame y i ^ . Denoting by n i the number of frames labeled y i ^ , we then assign the label y ^ = y ^ argmax i n i to the gesture. The HOG and SVM configurations are as described above. For CNN, we used the ResNet50V2 architecture from Keras API [59,60], as it provides a good trade-off between computation time and accuracy [45,61]. We loaded weights pre-trained on ImageNet [62]. The last layer of the model is with Softmax activation and 7 outputs (corresponding to 7 situations described in Section 4.3). The model is compiled using Keras with TensorFlow backend. We used categorical cross-entropy loss function and Stochastic Gradient Descent (SGD) optimizer with the default learning rate of 0.01. We split the overall amount of training data into 80% training and 20% validation, and trained the network with 10 epochs and with a batch size of 8. Training and validation loss decreased quickly and remained close to 0 after 10 epochs (respectively, 0.0028 and 0.0236) (see Figure 4). Training and validation accuracy reached 98.48% and 99.11%, respectively, after two epochs only, and after 10 epochs, training accuracy remained over 99.9%, while validation accuracy stayed stable at around 99.4% (see Figure 5).

4.3. Data Collection and Training

We collected data from 13 people (5 f, 8 m) from the research lab, who performed the four user-defined gestures. Each person was also asked to both walk and stand at various distances in front of the camera from a distance of 1.5 m to a distance of 5 m. Participants performed omrieach gestures in front of the sensors and the data collection was conducted across four different indoor locations. Lighting conditions for the four locations ranged between 100 and 700 lux. The distance from the participants to the background ranged between 0.2 and 4 m. The temperature for all locations was an ambient temperature of 25 °C to 28 °C.
We gathered and trained the system to recognize when there was no person in front of it, so that it could differentiate this situation from the others. We gathered data points as follows for each sensor: 13 people × 7 situations (4 user-defined gestures + walking + standing + no person) × 15 trials × 20 frames = 27,300 frames, which were then used to train the algorithms. As part of the data collection process, we measured the average completion time for the user-defined gestures across participants (for one repetition, in seconds): the longest duration is “Mapping” ( μ = 3.05 , S D = 0.65 ), followed by “Identifying Hazardous Materials” ( μ = 2.26 , S D = 0.7 ), “Finding Survivors” ( μ = 1.13 , S D = 0.27 ), and “Detecting Fire” ( μ = 1.04 , S D = 0.26 ).
Note that these data were collected under the baseline conditions (i.e., without firefighting equipment and without added smoke). Our idea was to train the system under regular conditions and then determine if, and to what extent, a system trained under regular conditions would perform over different types of harsh conditions.

4.4. Evaluation

The aim of the evaluation phase was to evaluate how the sensors and algorithms would perform under different harsh conditions in real-time gesture recognition. The system was evaluated with two people (2 m) from the research lab who did not participate in the training phase. Each of them performed 6 situations (4 user-defined gestures + walk + stand) in front of the sensors. Each participant performed a total of 50 repetitions of the 6 situations in a random order (the ratio of situations tested was balanced). We further evaluated the system when no person was standing in front of the visual sensor. The evaluation was conducted under the following three conditions: 1. with the person wearing firefighting protective equipment, 2. with the person wearing firefighting protective equipment and with smoke in the room (Figure 6b), and 3. without equipment or smoke (i.e., baseline). The smoke was cold, dense smoke generated by a dedicated smoke machine in an indoor environment (derived from [7]), and all tests were run in controlled laboratory settings. Figure 7 illustrates the evaluation setup. Each of the 3 laboratory conditions was evaluated with all 3 sensors and the following algorithms: RGB (skeleton extraction + SVM), Depth (“Frame Vote” (HOG + SVM), “Frame Vote” (CNN), HOG + SVM), Thermal (“Frame Vote” (HOG + SVM), “Frame Vote” (CNN), HOG + SVM, skeleton extraction + SVM), resulting in a total of 24 settings.
The evaluation was conducted in conditions similar to the training phase. Lighting conditions were 500 lux, the distance between the participants and the sensors was 1.5 to 5 m, the distance from participants to the background ranged between 0.2 and 2 m, the participants performed the gestures in front of the sensors, and the ambient temperature was between 25 °C and 28 °C.

4.5. Results and Interpretation

We calculated the recognition accuracy as the ratio between the number of gestures correctly recognized over the total number of trials. These results are reported in Table 2. The accuracy was the highest overall with the skeleton extraction algorithm. When tested in baseline conditions (no smoke or equipment), we achieved, on the best algorithm, high accuracy for each sensor: 98% for RGB, 92% for depth, and 94% for thermal. With firefighting equipment on, the recognition accuracy remained high for RGB and thermal, with 96%, and lowered for depth, with 84%. When smoke was added, the recognition reached up to 90% with RGB and 88% with thermal but the depth sensor did not work in this condition. Although RGB performed slightly better than thermal, we know from the literature that its performance will degrade in darkness or under thicker smoke. We find that these results are consistent with the literature about the effect of smoke on the three sensors, and that they hold in gesture recognition applications [43].
In terms of algorithms, while the HOG + SVM technique achieved better accuracy than Frame Vote when tested without equipment and smoke, Frame Vote achieved better accuracy than HOG + SVM in smoke. Regarding the use of protective gear, the depth sensor was the most affected, but the other sensors’ recognition was not significantly affected. The skeleton extraction + SVM algorithm appeared the most robust to protective gear, with only a 2% decrease for the RGB sensor, and with an increase of 2% in accuracy for the thermal sensor. In addition, we have shown that skeleton extraction of key points can be done with thermal images, although the API was built for RGB data.
Figure 8, Figure 9 and Figure 10 present confusion matrices corresponding to 3 evaluated conditions out of the 24. We present the following three settings: RGB sensor with skeleton extraction, without smoke, and without equipment (Figure 8); depth sensor with HOG + SVM, without smoke, and without equipment (as in Figure 9); and thermal sensor with skeleton extraction + SVM, without smoke, and without equipment (as in Figure 10). The confusion matrices are used to demonstrate the recognition accuracy of each gesture and to describe the type of error made by the system.
Figure 8 presents the confusion matrix from the RGB sensor with skeleton extraction, without smoke, and without equipment, which shows an overall measured accuracy of 98%. We found that 3 out of 7 situations are recognized perfectly: “No Person”, “Detecting Fire”, and “Finding Survivors”. “Stand” was accurately recognized 96% of the time and was confused with “Walk” 4% of the time and “Walk” was accurately recognized 94% of the time and was confused with “Stand” 6% of the time. “Mapping” was accurately recognized 99% of the time and was confused with either “Stand” or “Detecting Fire” (<1% each). Finally, “Identifying Hazardous Materials” was accurately recognized 99% of the time and was confused with “Finding Survivors” (<1%).
Figure 9 presents the confusion matrix from the depth sensor with HOG + SVM, without smoke, and without equipment, which shows an overall measured accuracy of 92%. We found that 2 out of 7 situations are recognized perfectly: “Walk” and “Detecting Fire”. “No Person” was accurately recognized 67% of the time and was confused with “Stand” 33% of the time, while “Stand” was accurately recognized 79% of the time and was confused with “Walk” and “Identifying Hazardous Materials” (approximately 11% each). The major difference in the accuracy of these 2 situations, between this setting and RGB with skeleton extraction + SVM, can be explained by the fact that we used background subtraction based on motion analysis (see Section 4.2). When the participant is standing still, it appears as if there is no person in the scene after applying background subtraction. As such, this configuration introduced a confusion between the 2 situations, which further reveals a clear disadvantage of applying this background subtraction to static gestures. “Finding Survivors” was accurately recognized 94% of the time and was confused with “Mapping” 6% of the time. “Mapping” was accurately recognized 94% of the time and was confused with “Stand” 6% of the time. Finally, “Identifying Hazardous Materials” was accurately recognized 88% of the time and was confused with “Stand” 12% of the time.
Figure 10 presents the confusion matrix of the thermal sensor with skeleton extraction + SVM, without smoke, and without equipment, which shows an overall measured accuracy of 94%. In this condition, 5 situations out of 7 are recognized perfectly: “No person”, “Stand”, “Detecting Fire”, “Finding Survivors”, and “Mapping”. “Walk” was accurately recognized 89% of the time and was confused with “Stand” 11% of the time. This is not surprising given that standing corresponds to a subset of frames of walking. “Identifying Hazardous Materials” was the least recognized situation, with 76% accuracy, and was mainly confused with “Mapping” (12%) but also with “Detecting Fire” and “Walk” (5.9% each). When inspecting the user-defined gestures, we found similarities between the “Identifying Hazardous Materials” and “Mapping” gestures that present overlapping frames (when both hands are either in front of the face or moving from the side of the body to the face), which may explain the confusion between the two gestures.
Comparing these 3 confusion matrices, which all represent one laboratory setting (without smoke and without equipment), enables us to better understand some of the advantages and drawbacks of the algorithms or to identify differences and similarities in the gestures. For example, our findings suggest that the “Mapping”, “Detecting Fire”, and “Finding Survivors” gestures were overall successfully recognized and differentiated from the other gestures, with 94% to 100% accuracy in the presented settings. However, the “Identifying Hazardous Materials” gesture was recognized, with lower accuracy (76% to 99% in the presented settings), which suggests that this gesture might not be appropriate in the first response context, which is not tolerant to mistakes. It is important to note that when these gestures were elicited with firefighters, no algorithmic perspective was considered. However, as first responders are trained personnel, they can be taught specific gestures, and the “Identifying Hazardous Materials” could be modified in an iterative process with the firefighters to reduce its overlap with other gestures. In addition, we see that skeleton extraction + SVM achieved higher accuracy for the gesture “Identifying Hazardous Materials” with the RGB sensor (99%) than with the thermal sensor (76%), which can be explained by the fact that the API was trained with RGB data only. These findings highlight the need to train the skeleton extraction approach with additional types of sensor data for a wider range of needs.

5. Discussion

In this work, we compared how different remote visual sensors and algorithms perform under different conditions related to harsh environments. In terms of sensors, our results showed that the RGB camera performs relatively well in a smoky environment, although prior work showed that it is expected to fail in dense smoke when visibility decreases below 1 m. Depth sensors, on the contrary, which are designed to work well in darkness, did not allow for gesture recognition in smokey conditions, even with light smoke. The thermal sensor, which also performs well in darkness, maintained high accuracy in our experiment, even under smoky conditions and with the use of firefighting equipment. Based on our findings, and combined with the fact that thermal sensors can effectively detect fires [63], we propose that thermal sensors would be the best compromise for gestural recognition in harsh conditions. Nonetheless, we envision that future robotic systems will be required to perform well in both regular and harsh conditions, and in both daylight and darkness. As such, we propose that having multiple sensors and using sensor fusion algorithms would be a safer solution. While this solution should be easily adaptable on ground robots, weight limitations on aerial robots will probably still constrain researchers to a limited amount of sensors and computing power, for which we suggest that thermal sensors are most suitable.
However, thermal sensors also present limitations, such as high costs (over RGB and depth sensors). They are also less prevalent than RGB and depth cameras in existing robotic systems. Furthermore, RGB cameras typically provide higher-resolution images, while depth cameras provide the additional advantage that they can measure the distance between the robot and its environment, so that they can offer multiple usages. We therefore suggest that the decision regarding which sensor to embed on a robotic system shall take a holistic perspective in terms of the actual system’s needs.
In terms of algorithms for gestural recognition in harsh environments, we found that the skeleton extraction + SVM approach provided good results (with up to 98% accuracy in baseline conditions and 90% accuracy in smoke and with equipment), even though the system was not trained in harsh environment settings. Since there are no publicly available relevant datasets with gestures adapted to the firefighting context, we propose to compare our results to prior work using similar settings. For example, when comparing the results from the RGB sensor with skeleton extraction + SVM, without equipment, and without smoke, to recent results from Liu and Szirányi [54] in similar settings with skeleton extraction (OpenPose) + DNN, we find that our results are on par with the state-of-the-art. Indeed, their model can recognize 10 body gestures at 99.8% accuracy using their test set, against 98% in our system. While they achieved 1.8% higher accuracy, only 2 of their 10 gestures were dynamic, the rest being static. This may then explain the small difference in accuracy; however, further research should specifically compare static vs. dynamic gestures.
One of the goals of our experiment was to verify the effect of the three harsh environmental conditions on the sensors and algorithms’ performances. We have shown that, although the system was trained in normal conditions (i.e., without smoke and equipment), the presence of protective equipment on the person did not affect the RGB and thermal sensors’ recognition rates, and that the skeleton extraction + SVM was the least affected method. Moreover, when smoke was added, all three sensors suffered from a decrease in accuracy, with skeleton extraction + SVM being, once again, the least affected method.

6. Limitations and Future Work

This work investigated the performance of three common sensors for gesture recognition in simulated harsh environments. While we found interesting differences between the sensors, future work should consider additional sensors operating at different wavelengths or using different technology, as recent work compared LIDAR perfomance in smoke to six other sensors [43]. Our evaluation considered different harsh environments simulated in an indoor controlled laboratory setup. In one setting, we generated cold and dense smoke with a dedicated machine. Future research could vary the density and temperature of the smoke (e.g., light smoke with a high temperature [43]) and consider smoke generated from various burning materials. Future research could further expand to other algorithms, such as by using RNN or LSTM for the classification of gestures, or by expanding the “Frame Vote” algorithm by introducing weights to the frames.
In our work, we chose to test the gesture recognition system using a fixed setup, before embedding the sensor and computation on a robot or a drone, as per prior work [37,50]. As such, future work will investigate additional constraints related to embedding the system and testing it in real-world conditions that will affect the recognition accuracy. These include environmental factors such as various types of lighting (indoors or outdoors), different orientations and angles between people and sensor, managing jitter, evaluating the effect of occlusion due to objects such as debris, and considering various heat sources in the scene, since the performance of thermal sensors will be affected by them (e.g., open fire) [64].
We further expect that when training the system with data collected in additional environmental settings, we will see a lesser decrease in the accuracy between the baseline (i.e., no equipment or smoke) and harsh conditions (e.g., equipment and smoke). Finally, future research should implement the provided methods and conduct empirical research with first responders in simulated or real harsh conditions.

7. Conclusions

This manuscript presented research into gestural recognition in harsh environments tailored for first responders. In particular, we investigated appropriate sensors and algorithms for the remote sensing of body gestures from ground or aerial robot. We designed and prototyped a gestural interaction system using three types of sensors, namely RGB, depth, and thermal, and a series of algorithms and feature extraction techniques. We trained the system (N = 13) and evaluated it (N = 2) in three simulated harsh environments. Our results highlight differences between sensors and algorithms that can inform future researchers and practitioners on the human and technical aspects to consider when integrating gestural recognition systems in harsh environments.

Author Contributions

Conceptualization, O.A., S.R., C.F. and J.R.C.; methodology, O.A. and J.R.C.; software, O.A.; validation, O.A. and J.R.C.; formal analysis, O.A.; investigation, O.A. and J.R.C.; resources, O.A. and J.R.C.; data curation, O.A.; writing—original draft preparation, O.A. and J.R.C.; writing—review and editing, O.A., S.R., C.F. and J.R.C.; visualization, O.A.; supervision, J.R.C.; project administration, O.A., S.R. and J.R.C.; All authors have read and agreed to the published version of the manuscript.

Funding

This research received partial external funding from Rafael Ltd.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Acknowledgments

We thank Woody Gover for his help in performing the lab experiment and Netanel Barkay and Lior Melamed for lending us the equipment needed for the experiment.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
HRIHuman–robot interaction
HCIHuman–computer interaction
HDIHuman–drone interaction
SVMSupport vector machine
HOGHistogram of oriented gradient
CNNConvolutional neural networks
RNNRecurrent neural network
LSTMLong short-term memory

References

  1. Basiratzadeh, S.; Mir, M.; Tahseen, M.; Shapsough, A.; Dorrikhteh, M.; Saeidi, A. Autonomous UAV for Civil Defense Applications. In Proceedings of the 3rd World Congress on Computer Applications and Information Systems, Kuala Lumpur, Malaysia, 6–8 September 2016. [Google Scholar]
  2. Apvrille, L.; Tanzi, T.; Dugelay, J.L. Autonomous drones for assisting rescue services within the context of natural disasters. In Proceedings of the 2014 XXXIth URSI General Assembly and Scientific Symposium (URSI GASS), Beijing, China, 16–23 August 2014; pp. 1–4. [Google Scholar]
  3. Croizé, P.; Archez, M.; Boisson, J.; Roger, T.; Monsegu, V. Autonomous measurement drone for remote dangerous source location mapping. Int. J. Environ. Sci. Dev. 2015, 6, 391. [Google Scholar] [CrossRef] [Green Version]
  4. Martinson, E.; Lawson, W.E.; Blisard, S.; Harrison, A.M.; Trafton, J.G. Fighting fires with human robot teams. In Proceedings of the IROS, Vilamoura, Portugal, 7–12 October 2012; pp. 2682–2683. [Google Scholar]
  5. Chavez, A.G.; Mueller, C.A.; Doernbach, T.; Chiarella, D.; Birk, A. Robust gesture-based communication for underwater human-robot interaction in the context of search and rescue diver missions. arXiv 2018, arXiv:1810.07122. [Google Scholar]
  6. Cacace, J.; Finzi, A.; Lippiello, V.; Furci, M.; Mimmo, N.; Marconi, L. A control architecture for multiple drones operated via multimodal interaction in search & rescue mission. In Proceedings of the 2016 IEEE International Symposium on Safety, Security, and Rescue Robotics (SSRR), Lausanne, Switzerland, 23–27 October 2016; pp. 233–239. [Google Scholar]
  7. De Cillis, F.; Oliva, G.; Pascucci, F.; Setola, R.; Tesei, M. On field gesture-based human-robot interface for emergency responders. In Proceedings of the 2013 IEEE International Symposium on Safety, Security, and Rescue Robotics (SSRR), Linkoping, Sweden, 21–26 October 2013; pp. 1–6. [Google Scholar]
  8. Wong, C.; Yang, E.; Yan, X.T.; Gu, D. Autonomous robots for harsh environments: A holistic overview of current solutions and ongoing challenges. Syst. Sci. Control. Eng. 2018, 6, 213–219. [Google Scholar] [CrossRef]
  9. Wong, C.; Yang, E.; Yan, X.T.; Gu, D. An overview of robotics and autonomous systems for harsh environments. In Proceedings of the 2017 23rd International Conference on Automation and Computing (ICAC), Huddersfield, UK, 7–8 September 2017; pp. 1–6. [Google Scholar]
  10. Bausys, R.; Cavallaro, F.; Semenas, R. Application of sustainability principles for harsh environment exploration by autonomous robot. Sustainability 2019, 11, 2518. [Google Scholar] [CrossRef] [Green Version]
  11. Roldán-Gómez, J.J.; González-Gironda, E.; Barrientos, A. A Survey on Robotic Technologies for Forest Firefighting: Applying Drone Swarms to Improve Firefighters’ Efficiency and Safety. Appl. Sci. 2021, 11, 363. [Google Scholar] [CrossRef]
  12. Ruangpayoongsak, N.; Roth, H.; Chudoba, J. Mobile robots for search and rescue. In Proceedings of the IEEE International Safety, Security and Rescue Rototics, Workshop, Kobe, Japan, 6–9 June 2005; pp. 212–217. [Google Scholar]
  13. Jones, H.L.; Rock, S.M.; Burns, D.; Morris, S. Autonomous robots in swat applications: Research, design, and operations challenges. In Proceedings of the AUVSI’02, Orlando, FL, USA, 9–11 July 2002. [Google Scholar]
  14. Patil, D.; Ansari, M.; Tendulkar, D.; Bhatlekar, R.; Pawar, V.N.; Aswale, S. A Survey On Autonomous Military Service Robot. In Proceedings of the 2020 International Conference on Emerging Trends in Information Technology and Engineering (ic-ETITE), Vellore, India, 24–25 February 2020; pp. 1–7. [Google Scholar]
  15. Matsuno, F.; Tadokoro, S. Rescue robots and systems in Japan. In Proceedings of the 2004 IEEE International Conference on Robotics and Biomimetics, Shenyang, China, 22–26 August 2004; pp. 12–20. [Google Scholar]
  16. Kinaneva, D.; Hristov, G.; Raychev, J.; Zahariev, P. Early forest fire detection using drones and artificial intelligence. In Proceedings of the 2019 42nd International Convention on Information and Communication Technology, Electronics and Microelectronics (MIPRO), Opatija, Croatia, 20–24 May 2019; pp. 1060–1065. [Google Scholar]
  17. Tan, C.F.; Liew, S.M.; Alkahari, M.R.; Ranjit, S.S.S.; Said, M.R.; Chen, W.; Sivakumar, D. Fire fighting mobile robot: State of the art and recent development. Aust. J. Basic Appl. Sci. 2013, 7, 220–230. [Google Scholar]
  18. Rossi, M.; Brunelli, D. Autonomous Gas Detection and Mapping With Unmanned Aerial Vehicles. IEEE Trans. Instrum. Meas. 2016, 65, 765–775. [Google Scholar] [CrossRef]
  19. Hrabia, C.E.; Hessler, A.; Xu, Y.; Brehmer, J.; Albayrak, S. Efffeu project: Efficient operation of unmanned aerial vehicles for industrial fire fighters. In Proceedings of the 4th ACM Workshop on Micro Aerial Vehicle Networks, Systems, and Applications, Munich, Germany, 10–15 June 2018; pp. 33–38. [Google Scholar]
  20. Radu, V.T.; Kristensen, A.S.; Mehmood, S. Use of Drones for Firefighting Operations. Master’s Thesis, University Aalborg, Aalborg, Denmark, 2019. [Google Scholar]
  21. Hassanein, A.; Elhawary, M.; Jaber, N.; El-Abd, M. An autonomous firefighting robot. In Proceedings of the 2015 International Conference on Advanced Robotics (ICAR), Istanbul, Turkey, 27–31 July 2015; pp. 530–535. [Google Scholar]
  22. Fraune, M.R.; Khalaf, A.S.; Zemedie, M.; Pianpak, P.; NaminiMianji, Z.; Alharthi, S.A.; Dolgov, I.; Hamilton, B.; Tran, S.; Toups, Z. Developing Future Wearable Interfaces for Human-Drone Teams through a Virtual Drone Search Game. Int. J. Hum. Comput. Stud. 2021, 147, 102573. [Google Scholar] [CrossRef]
  23. Alon, O.; Fyodorov, C.; Rabinovich, S.; Cauchard, J.R. Drones in Firefighting: A User-Centered Design Perspective. In Proceedings of the Accepted for publication at MobileHCI’21: ACM International Conference on Mobile Human-Computer Interaction, Toulouse, France, 27–30 September 2021; pp. 1–17. [Google Scholar]
  24. Cauchard, J.R.; Khamis, M.; Garcia, J.; Kljun, M.; Brock, A.M. Toward a roadmap for human-drone interaction. Interactions 2021, 28, 76–81. [Google Scholar] [CrossRef]
  25. Waldherr, S.; Romero, R.; Thrun, S. A Gesture Based Interface for Human-Robot Interaction. Auton. Robot. 2000, 9, 151–173. [Google Scholar] [CrossRef]
  26. Cauchard, J.R.; Zhai, K.Y.; Landay, J.A. Drone & me: An exploration into natural human-drone interaction. In Proceedings of the 2015 ACM International Joint Conference on Pervasive and Ubiquitous Computing, Osaka, Japan, 7–11 September 2015; pp. 361–365. [Google Scholar]
  27. Jane, L.E.; Ilene, L.E.; Landay, J.A.; Cauchard, J.R. Drone & Wo: Cultural Influences on Human-Drone Interaction Techniques. In Proceedings of the 2017 CHI Conference on Human Factors in Computing Systems—CHI’17, Denver, CO, USA, 6–11 May 2017; Association for Computing Machinery: New York, NY, USA, 2017; pp. 6794–6799. [Google Scholar] [CrossRef] [Green Version]
  28. Cooney, M.D.; Becker-Asano, C.; Kanda, T.; Alissandrakis, A.; Ishiguro, H. Full-body gesture recognition using inertial sensors for playful interaction with small humanoid robot. In Proceedings of the 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems, Taipei, Taiwan, 18–22 October 2010; pp. 2276–2282. [Google Scholar] [CrossRef]
  29. Sigalas, M.; Baltzakis, H.; Trahanias, P. Gesture recognition based on arm tracking for human-robot interaction. In Proceedings of the 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems, Taipei, Taiwan, 18–22 October 2010; pp. 5424–5429. [Google Scholar]
  30. Lv, Z.; Halawani, A.; Feng, S.; Li, H.; Réhman, S.U. Multimodal hand and foot gesture interaction for handheld devices. ACM Trans. Multimed. Comput. Commun. Appl. (TOMM) 2014, 11, 10. [Google Scholar] [CrossRef]
  31. Wen, H.; Ramos Rojas, J.; Dey, A.K. Serendipity: Finger gesture recognition using an off-the-shelf smartwatch. In Proceedings of the 2016 CHI Conference on Human Factors in Computing Systems, San Jose, CA, USA, 7–12 May 2016; pp. 3847–3851. [Google Scholar]
  32. Escalera, S.; Athitsos, V.; Guyon, I. Challenges in multi-modal gesture recognition. In Gesture Recognition; Springer: Berlin/Heidelberg, Germany, 2017; pp. 1–60. [Google Scholar]
  33. Monajjemi, M.; Bruce, J.; Sadat, S.A.; Wawerla, J.; Vaughan, R. UAV, do you see me? Establishing mutual attention between an uninstrumented human and an outdoor UAV in flight. In Proceedings of the 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Hamburg, Germany, 28 September–2 October 2015; pp. 3614–3620. [Google Scholar]
  34. Wu, M.; Balakrishnan, R. Multi-finger and whole hand gestural interaction techniques for multi-user tabletop displays. In Proceedings of the 16th Annual ACM Symposium on User Interface Software and Technology, Vancouver, BC, Canada, 2–5 November 2003; pp. 193–202. [Google Scholar]
  35. Broccia, G.; Livesu, M.; Scateni, R. Gestural Interaction for Robot Motion Control. In Proceedings of the Eurographics Italian Chapter Conference, Llandudno, UK, 11–15 April 2011; pp. 61–66. [Google Scholar]
  36. Ng, W.S.; Sharlin, E. Collocated interaction with flying robots. In Proceedings of the 2011 Ro-Man, Atlanta, GA, USA, 31 July–3 August 2011; pp. 143–149. [Google Scholar]
  37. Cauchard, J.R.; Tamkin, A.; Wang, C.Y.; Vink, L.; Park, M.; Fang, T.; Landay, J.A. Drone.Io: A Gestural and Visual Interface for Human-Drone Interaction. In Proceedings of the 2019 14th ACM/IEEE International Conference on Human-Robot Interaction (HRI)—HRI’19, Daegu, Korea, 11–14 March 2019; pp. 153–162. [Google Scholar]
  38. Rodríguez-Moreno, I.; Martínez-Otzeta, J.M.; Sierra, B.; Rodriguez, I.; Jauregi, E. Video activity recognition: State-of-the-art. Sensors 2019, 19, 3160. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  39. Berman, S.; Stern, H. Sensors for gesture recognition systems. IEEE Trans. Syst. Man, Cybern. Part C (Appl. Rev.) 2011, 42, 277–290. [Google Scholar] [CrossRef]
  40. Freeman, W.T.; Roth, M. Orientation histograms for hand gesture recognition. In Proceedings of the International Workshop on Automatic Face and Gesture Recognition, Zurich, Switzerland, 26–28 June 1995; Volume 12, pp. 296–301. [Google Scholar]
  41. Vandersteegen, M.; Reusen, W.; Van Beeck, K.; Goedemé, T. Low-Latency Hand Gesture Recognition With a Low-Resolution Thermal Imager. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 14–19 June 2020; pp. 98–99. [Google Scholar]
  42. Ren, Z.; Meng, J.; Yuan, J. Depth camera based hand gesture recognition and its applications in human-computer-interaction. In Proceedings of the 2011 8th International Conference on Information, Communications & Signal Processing, Singapore, 13–16 December 2011; pp. 1–5. [Google Scholar]
  43. Starr, J.W.; Lattimer, B. Evaluation of navigation sensors in fire smoke environments. Fire Technol. 2014, 50, 1459–1481. [Google Scholar] [CrossRef]
  44. Department of the Army. Training Circular 3-21.60 (FM 21-60) Visual Signals; Technical Report; CreateSpace Publishing: Scotts Valley, CA, USA, 2017; pp. 1–92. [Google Scholar]
  45. Kawatsu, C.; Koss, F.; Gillies, A.; Zhao, A.; Crossman, J.; Purman, B.; Stone, D.; Dahn, D. Gesture recognition for robotic control using deep learning. In Proceedings of the NDIA Ground Vehicle Systems Engineering and Technology Symposium, Novi, MI, USA, 8–10 August 2017. [Google Scholar]
  46. Chen, Q.; Georganas, N.D.; Petriu, E.M. Real-time vision-based hand gesture recognition using haar-like features. In Proceedings of the 2007 IEEE Instrumentation & Measurement Technology Conference IMTC, Warsaw, Poland, 1–3 May 2007; pp. 1–6. [Google Scholar]
  47. Appenrodt, J.; Al-Hamadi, A.; Michaelis, B. Data gathering for gesture recognition systems based on single color-, stereo color-and thermal cameras. Int. J. Signal Process. Image Process. Pattern Recognit. 2010, 3, 37–50. [Google Scholar]
  48. Intel. Realsense Depth Camera D435. Available online: https://www.intelrealsense.com/depth-camera-d435/ (accessed on 7 June 2021).
  49. Opgal. Therm-App Hz. Available online: https://therm-app.com/ (accessed on 7 June 2021).
  50. Brock, A.M.; Chatain, J.; Park, M.; Fang, T.; Hachet, M.; Landay, J.A.; Cauchard, J.R. FlyMap: Interacting with Maps Projected from a Drone. In Proceedings of the 7th ACM International Symposium on Pervasive Displays—PerDis’18, Munich, Germany, 6–8 June 2018; Association for Computing Machinery: New York, NY, USA, 2018. [Google Scholar] [CrossRef] [Green Version]
  51. De Smedt, Q.; Wannous, H.; Vandeborre, J.P. Heterogeneous hand gesture recognition using 3D dynamic skeletal data. Comput. Vis. Image Underst. 2019, 181, 60–72. [Google Scholar] [CrossRef] [Green Version]
  52. Žemgulys, J.; Raudonis, V.; Maskeliūnas, R.; Damaševičius, R. Recognition of basketball referee signals from videos using Histogram of Oriented Gradients (HOG) and Support Vector Machine (SVM). Procedia Comput. Sci. 2018, 130, 953–960. [Google Scholar] [CrossRef]
  53. De Oliveira, D.C.; Wehrmeister, M.A. Using deep learning and low-cost RGB and thermal cameras to detect pedestrians in aerial images captured by multirotor UAV. Sensors 2018, 18, 2244. [Google Scholar] [CrossRef] [Green Version]
  54. Liu, C.; Szirányi, T. Real-Time Human Detection and Gesture Recognition for On-Board UAV Rescue. Sensors 2021, 21, 2180. [Google Scholar] [CrossRef] [PubMed]
  55. Cao, Z.; Hidalgo, G.; Simon, T.; Wei, S.E.; Sheikh, Y. OpenPose: Realtime multi-person 2D pose estimation using Part Affinity Fields. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 43, 172–186. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  56. Dalal, N.; Triggs, B. Histograms of oriented gradients for human detection. In Proceedings of the 2005 IEEE computer society conference on computer vision and pattern recognition (CVPR’05), San Diego, CA, USA, 20–25 June 2005; Volume 1, pp. 886–893. [Google Scholar]
  57. Dadi, H.S.; Pillutla, G.M. Improved face recognition rate using HOG features and SVM classifier. IOSR J. Electron. Commun. Eng. 2016, 11, 34–44. [Google Scholar] [CrossRef]
  58. OpenCV. BackgroundSubtractorMOG2. Available online: https://docs.opencv.org/master/d7/d7b/classcv_1_1BackgroundSubtractorMOG2.html (accessed on 7 June 2021).
  59. Keras API Reference. Resnet. Available online: https://keras.io/api/applications/resnet/ (accessed on 7 June 2021).
  60. He, K.; Zhang, X.; Ren, S.; Sun, J. Identity mappings in deep residual networks. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 8–16 October 2016; pp. 630–645. [Google Scholar]
  61. Bianco, S.; Cadene, R.; Celona, L.; Napoletano, P. Benchmark Analysis of Representative Deep Neural Network Architectures. IEEE Access 2018, 6, 64270–64277. [Google Scholar] [CrossRef]
  62. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 25, 1097–1105. [Google Scholar] [CrossRef]
  63. Hendel, I.G.; Ross, G.M. Efficacy of Remote Sensing in Early Forest Fire Detection: A Thermal Sensor Comparison. Can. J. Remote Sens. 2020, 46, 414–428. [Google Scholar] [CrossRef]
  64. Brunner, C.; Peynot, T.; Vidal-Calleja, T.; Underwood, J. Selective combination of visual and thermal imaging for resilient localization in adverse conditions: Day and night, smoke and fire. J. Field Robot. 2013, 30, 641–666. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Set of four user-defined gestures used to support the following tasks: (a) detection of fire spots (their number and location), (b) identification of gas/chemical leak source and size, (c) search for trapped citizens or people in danger and collection of information on their number and location, (d) mapping of the floor or building. Image courtesy of [23].
Figure 1. Set of four user-defined gestures used to support the following tasks: (a) detection of fire spots (their number and location), (b) identification of gas/chemical leak source and size, (c) search for trapped citizens or people in danger and collection of information on their number and location, (d) mapping of the floor or building. Image courtesy of [23].
Sensors 21 03997 g001
Figure 2. Example of images acquired in the evaluation phase. (a) Depth image of a person wearing a normal outfit. (b) Thermal image of a person wearing a firefighter’s protective equipment.
Figure 2. Example of images acquired in the evaluation phase. (a) Depth image of a person wearing a normal outfit. (b) Thermal image of a person wearing a firefighter’s protective equipment.
Sensors 21 03997 g002
Figure 3. Description of the bespoke “Frame Vote” algorithm. Each frame is classified as one of the possible gestures. The final classification is made by comparing which gesture was most often classified within 20 frames. In this example, two out of four frames have been classified as “Mapping”, so the algorithm chooses that gesture over the other two.
Figure 3. Description of the bespoke “Frame Vote” algorithm. Each frame is classified as one of the possible gestures. The final classification is made by comparing which gesture was most often classified within 20 frames. In this example, two out of four frames have been classified as “Mapping”, so the algorithm chooses that gesture over the other two.
Sensors 21 03997 g003
Figure 4. Training (blue) and validation (orange) loss through epochs for the CNN.
Figure 4. Training (blue) and validation (orange) loss through epochs for the CNN.
Sensors 21 03997 g004
Figure 5. Training (blue) and validation (green) accuracy through epochs for the CNN.
Figure 5. Training (blue) and validation (green) accuracy through epochs for the CNN.
Sensors 21 03997 g005
Figure 6. (a) Firefighting equipment including helmet, jacket, trousers, and gloves and (b) experiment conducted to evaluate the system’s accuracy with full equipment and smoke.
Figure 6. (a) Firefighting equipment including helmet, jacket, trousers, and gloves and (b) experiment conducted to evaluate the system’s accuracy with full equipment and smoke.
Sensors 21 03997 g006
Figure 7. Examples of the setup used in the evaluation of our system. The person is wearing protective firefighting equipment and smoke is starting to be generated. (a) The Realsense camera feed is shown on the laptop. (b) The raw thermal sensor feed is shown on the left side of the laptop’s screen and, on the right side of the laptop’s screen, the feed is shown after background subtraction.
Figure 7. Examples of the setup used in the evaluation of our system. The person is wearing protective firefighting equipment and smoke is starting to be generated. (a) The Realsense camera feed is shown on the laptop. (b) The raw thermal sensor feed is shown on the left side of the laptop’s screen and, on the right side of the laptop’s screen, the feed is shown after background subtraction.
Sensors 21 03997 g007
Figure 8. Confusion matrix for the RGB sensor with skeleton extraction + SVM, without smoke, and without equipment.
Figure 8. Confusion matrix for the RGB sensor with skeleton extraction + SVM, without smoke, and without equipment.
Sensors 21 03997 g008
Figure 9. Confusion matrix for the depth sensor with HOG + SVM, without smoke, and without equipment.
Figure 9. Confusion matrix for the depth sensor with HOG + SVM, without smoke, and without equipment.
Sensors 21 03997 g009
Figure 10. Confusion matrix for the thermal sensor with skeleton extraction + SVM, without smoke, and without equipment.
Figure 10. Confusion matrix for the thermal sensor with skeleton extraction + SVM, without smoke, and without equipment.
Sensors 21 03997 g010
Table 1. Drone tasks used as referents for the elicitation study. The corresponding user-defined gestures are presented in Figure 1.
Table 1. Drone tasks used as referents for the elicitation study. The corresponding user-defined gestures are presented in Figure 1.
TaskDescriptionChosen Gesture
Detecting FireDetect number and location of fire spots(a)
Identifying Hazardous MaterialsIdentify source and size of gas or chemical leak(b)
Finding SurvivorsSearch for trapped citizens or people in danger and collect information on their number and location(c)
MappingCreate a map of the floor or building based on exploration(d)
Table 2. Recognition accuracy results for each sensor and algorithm evaluated (see Section 4.4).
Table 2. Recognition accuracy results for each sensor and algorithm evaluated (see Section 4.4).
Recognition Accuracy
SensorAlgorithmwithout Smoke
without Equipment
without Smoke
with Equipment
with Smoke
with Equipment
Depth“Frame Vote” (HOG + SVM)78%70%N/A
“Frame Vote” (CNN)81%72%N/A
HOG + SVM92%84%N/A
RGBskeleton extraction + SVM98%96%90%
Thermal“Frame Vote” (HOG + SVM)90%86%70%
“Frame Vote” (CNN)90%84%71%
HOG + SVM86%86%56%
skeleton extraction + SVM94%96%88%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Alon, O.; Rabinovich, S.; Fyodorov, C.; Cauchard, J.R. First Step toward Gestural Recognition in Harsh Environments. Sensors 2021, 21, 3997. https://doi.org/10.3390/s21123997

AMA Style

Alon O, Rabinovich S, Fyodorov C, Cauchard JR. First Step toward Gestural Recognition in Harsh Environments. Sensors. 2021; 21(12):3997. https://doi.org/10.3390/s21123997

Chicago/Turabian Style

Alon, Omri, Sharon Rabinovich, Chana Fyodorov, and Jessica R. Cauchard. 2021. "First Step toward Gestural Recognition in Harsh Environments" Sensors 21, no. 12: 3997. https://doi.org/10.3390/s21123997

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop