Next Article in Journal
Study of Metal–Semiconductor–Metal CH3NH3PbBr3 Perovskite Photodetectors Prepared by Inverse Temperature Crystallization Method
Next Article in Special Issue
Assessment of the Potential of Wrist-Worn Wearable Sensors for Driver Drowsiness Detection
Previous Article in Journal
Influence of Electrostatic Forces on the Vibrational Characteristics of Resonators for Coriolis Vibratory Gyroscopes
Previous Article in Special Issue
Adaptive Binarization of QR Code Images for Fast Automatic Sorting in Warehouse Systems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Towards Mixed-Initiative Human–Robot Interaction: Assessment of Discriminative Physiological and Behavioral Features for Performance Prediction

by
Caroline P. C. Chanel
*,
Raphaëlle N. Roy
,
Frédéric Dehais
and
Nicolas Drougard
ISAE-SUPAERO, Université de Toulouse, 31400 Toulouse, France
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(1), 296; https://doi.org/10.3390/s20010296
Submission received: 2 December 2019 / Revised: 31 December 2019 / Accepted: 2 January 2020 / Published: 5 January 2020
(This article belongs to the Special Issue Human-Machine Interaction and Sensors)

Abstract

:
The design of human–robot interactions is a key challenge to optimize operational performance. A promising approach is to consider mixed-initiative interactions in which the tasks and authority of each human and artificial agents are dynamically defined according to their current abilities. An important issue for the implementation of mixed-initiative systems is to monitor human performance to dynamically drive task allocation between human and artificial agents (i.e., robots). We, therefore, designed an experimental scenario involving missions whereby participants had to cooperate with a robot to fight fires while facing hazards. Two levels of robot automation (manual vs. autonomous) were randomly manipulated to assess their impact on the participants’ performance across missions. Cardiac activity, eye-tracking, and participants’ actions on the user interface were collected. The participants performed differently to an extent that we could identify high and low score mission groups that also exhibited different behavioral, cardiac and ocular patterns. More specifically, our findings indicated that the higher level of automation could be beneficial to low-scoring participants but detrimental to high-scoring ones, and vice versa. In addition, inter-subject single-trial classification results showed that the studied behavioral and physiological features were relevant to predict mission performance. The highest average balanced accuracy (74%) was reached using the features extracted from all input devices. These results suggest that an adaptive HRI driving system, that would aim at maximizing performance, would be capable of analyzing such physiological and behavior markers online to further change the level of automation when it is relevant for the mission purpose.

1. Introduction

In the near future, human and artificial agents will be deployed to handle several missions in different contexts working as teammates. Some examples of contexts for which mixed human–machine systems would be relevant are search and rescue missions [1]; dirty, dull, and dangerous (3D) environments [2]; Mars rovers [3]; and multi-UAVs operation [4,5].
On the one hand, robots can operate in hostile environments, can quickly compute actions, and reduce mission costs while maximizing production. On the other hand, humans can handle unexpected events, they have intuition, are creative and capable of making complex (e.g., ethical or moral) decisions. However, the human operator may experience high workload [6] and cognitive fatigue [7,8] that in turn can impair their attentional [9] and executive abilities [10]. In this cybernetic system context, several challenges have arisen such as the implementation of human–robots monitoring and supervision systems to optimize their interaction and adapt the role of the human operator [11] and the role of the artificial agents [4,12,13].
A promising avenue to deal with this issue is to consider Mixed-Initiative for Human–Robots Interaction (MI-HRI). This type of interaction has been defined by [14] as “A collaboration strategy for human–robot teams where humans and robots opportunistically seize (relinquish) initiative from (to) each other as a mission is being executed, […] the initiative is mixed only when each member is authorized to intervene and seize control of it”. More specifically, this approach aims at identifying which human or artificial agent is the most capable of achieving a given task considering the evolution of the mission (e.g., failure of the robot, cognitive fatigue or psychological stress). Technically, this presupposes to implement monitoring techniques to evaluate the states of the different artificial and human agents and to design a decision system to dynamically adapt their respective role and authority [12,15]. Whereas the state assessment of deterministic agents (i.e., robots) is relatively simple, the evaluation of the human mental states remains challenging. This latter point is critical as poor diagnostics can lead to inadequate adaptations of the overall system [16]. Indeed, triggering higher level of automation, as well as not triggering it when necessary, can dramatically degrade performance [17,18]. Thus, one can argue that it could be beneficial to identify relevant markers of human performance as long mixed-initiative systems are concerned.
To our knowledge, the literature is still scarce concerning the evaluation of human performance, in particular, for Mixed-Initiative Human–Robot Interaction (MI-HRI) systems [12,19,20]. Interestingly, some authors [11,21,22] demonstrated that behavioral metrics could be defined to assess human operators’ workload and performance when interacting with a varying number of robots. In the same vein, ref. [23] proposed a methodology based on time series clustering to identify operators’ profile during their training process by exploiting several performance metrics. Thereby, the use of eye-tracking data has also been shown to be a suitable technique to estimate the human operator’s state [24] and to dynamically reallocate tasks between the human and unmanned aerial vehicles [4]. Physiological features were explored in [25] to recognize affective states from human operators interacting with a robot, and in [26] to directly adapt the interface being used during the interaction. A relevant approach is also presented in [19], whereby physiological data collected from a galvanic skin device were used to estimate the user engagement during human–robot interaction.
The use of physiological features to predict the operator’s performance during MI-HRI appears to be a relevant approach. The work in [27] proposed a rich review on several behavioral and physiological markers well-proofread in different scenarios, being general, and promising for online computation, thus, particularly useful for MI-HRI. For instance, features derived from eye-tracking (ET) devices, such as the number and duration of fixations, blinks [28], and cardiac features collected with an electrocardiogram (ECG), such as Heart Rate (HR) and Heart-Rate Variability (HRV), have been shown to be useful for workload assessment in several experimental contexts [27,29] and/or cognitive engagement estimation [30], while being quite promising for online computing [31]. Both ECG and ET devices are noninvasive devices, and easy to use in real life settings, which make them highly relevant for human–machine teaming applications. Particularly, as ET provides the eye gaze position, it is highly relevant to measure the human operator eye gaze activity in the context of supervisory control and multi-tasking management [32].
To progress towards systems that would dynamically drive MI-HRI to enhance systems’ performance, a first step is to characterize reliable behavioral and psycho-physiological features that could account for human performance. For this purpose, one can exploit a real world like human–robot mission, in which the human interacts with an autonomous robot. We thus built an experimental protocol using the robotic mission of the Firefighter Robot Game (http://robot-isae.isae.fr) developed by [33] and exhaustively described in [34].
This experimental set-up involved a user interface (see Figure 1) that allowed the participants to cooperate with a robot to fight fires. The robot could provide two levels of automation. In the high level of automation (i.e., autonomous robot mode) the robot navigates and extinguishes fires autonomously and the volunteers have to supervise the robot’s status and manage the water tank supply level. In the low level of automation (i.e., manual control), the participants have to perform the task manually such as handling the robot, extinguishing fires and managing the water tank supply level. The level of automation was randomly changed during the mission every ten seconds so as to assess its impact on the overall performance of the human–robot system (i.e., the number of extinguished fires). Different behavioral and psycho-physiological measures were collected during the experiment. First, participants’ actions (keystrokes and mouse clicks) on the user interface were monitored. An eye tracker (ET) was also used to measure the number of fixations on the main areas of interest (AOIs), and an electrocardiogram allowed to derive the Heart Rate (HR) and Heart-Rate Variability (HRV).
Thus, the objective of this study was to relate these different behavioral and psycho-physiological metrics to the mission performance. We used a two-step approach to demonstrate that participants’ behavioral, ECG and ET features would allow to discriminate between high and low mission-performer given the level of automation of the robot (manual control versus autonomous mode). First, we conducted statistical analysis for each of the dependent variables at the group level. Second, we implemented inter-subject single-trial classification algorithms that combine these different features to predict mission performance.

2. Materials and Methods

2.1. Participants

A total of 18 participants (5 females; mean age: 27.88 ; sd.: 4.76 ), all students from the ISAE-SUPAERO engineering school, were recruited by means of electronic mail announcements and oral advertising. All participants had normal (or corrected) vision abilities, and subscribed to social security and civil responsibility insurances. They all gave their informed consent. The project was validated by the local ethical committee of the University of Toulouse (CERNI number 2018-070).

2.2. Firefighter Robot Mission Task

This human–robot mission immerses the human operator in a scenario where s/he plays the role of a remote fireman who has to cooperate with a robot that is evolving in a virtual environment composed with trees that can randomly self-ignite. The goal of the mission is to fight as many fires as possible in a limited amount of time. Through the graphical user interface (GUI) shown in Figure 1, the human operator gets the video streaming from the robot’s camera (top right) as well as the position of the robot in the map (bottom center) and s/he monitors the robot’s internal states, namely, the temperature, the automation level, the battery charge level, and the amount of water in the on-board tank (bottom right). The human operator also has to manage an external water tank (bottom left) during all the mission using the buttons on the left-side of the interface: a tap, which can move horizontally by controlling a wheel (top buttons), fills the external tank when it is in the middle (which is an unstable equilibrium). To actually fill this external tank, the button bellow (black tap) turns on the tap for few seconds. Moreover the participants have to fix water leaks that may randomly occur on this external water tank by using the last button (black wrench).
The amount of water (i.e., the water in the robot on-board tank) that can be expulsed from the robot to put out fires is not unlimited: the robot has to be in the vicinity of a water filling station represented by a blue square on the ground to refill its on-board tank. The water of the filling station comes from the external water tank that must therefore not be empty. Despite the difficulty in filling this external tank and the occurrence of leaks, the human operator has to ensure that enough water will be available to the robot. In addition, the battery charge level of the robot decreases over time. It can be recharged at a charging station represented by a red square in the environment. If the battery is empty and the robot is not on the red square, the mission fails and ends prematurely. The temperature of the robot increases when it is too close to the flames and, it can also end the mission if too high.
In autonomous operation mode, the robot is able to navigate autonomously, and is programmed to extinguish the nearest fire, or to refill water or to recharge battery when necessary. This mode relates to a high automation condition, where the human operator has to supervise the robot’s parameters, and to manage the filling and repairing of the external tank. The robot’s manual control relates to a low automation level condition, because participants have to perform additional tasks, such as manually controlling the robot, managing its resources, and extinguishing fires. In this low automation level, the robot needs to be controlled by the keyboard arrows (navigation), and the participant has to press the keyboard space bar to shoot water. Note that, for the aim of this study, the robot’s automation level is changed randomly every ten seconds during the mission, following a uniform probability distribution over automation levels. In other words, every 10 s there was 50 % of chance that automation level changes or remains the same. As we study the impact of the robot automation level, a random change of the automation level is the best choice to avoid to create a bias related to a mode change policy; 10-s time windows were chosen as data collected during pilot experiments, as well as previous study [29], demonstrated to compute HR and HRV metrics in a robust manner for online purpose.
Each mission lasted at most 10 min. This design ensures a compromise between (i) the number of 10-s time windows for further statistical analysis and classification purpose. 60 × 10-s time windows per mission were expected, resulting on around 30 expected episodes of 10 s for each automation level condition. However, in our robotic scenario some missions could prematurely abort (robot battery fully discharged for instance), thus the average number of 10-s condition windows would be a little lower than 30; (ii) to induce time pressure for engaging participants (a countdown timer was displayed, see Figure 1 top-left) while avoiding fatigue issues given that the participants had to perform several missions. The complete experimental protocol is described in the following section.

2.3. Experimental Protocol

Participants were first welcomed in the experimental facilities at ISAE-SUPAERO. After reading and giving their signed consent, a training mission was proposed during which no data were recorded. The next step was to install the ECG device and to proceed with the calibration of the eye-tracking system. Once data acquisition was carefully checked, the participants had to fill the Karolinska Sleepiness Scale (KSS) survey [35] (see the paragraph “Subjective Feedback” Section 2.5.1). Finally, the participants had to perform four missions, each of them following the protocol below.
  • Rest period: the participants focused on a black cross in the center of the screen during 1 min. This allows to obtain a cardiac baseline.
  • Mission: the mission is launched and its maximum duration is 10 min.
  • Questionnaire: when the mission is finished, the participants are asked to fill the NASA-TLX survey [36] to get their subjective feedback concerning the effort made to carry out the mission (see the paragraph “Subjective Feedback” Section 2.5.1).
  • Break: a 2 min break is proposed to the participants.
After the end of the four missions, the participants are asked to fill again the KSS survey. The experiment was finalized by checking data records and removing the ECG equipment from the participant. The experimental protocol is summarized in Figure 2. As the participants have to perform four missions, the experiment lasted about 1h30 including training, device set up, surveys, and scenarios.

2.4. Data Collection

All robotic and mission state variables, as well as human operators’ actions on the ground station (keyboard inputs and mouse clicks), eye-tracking, and ECG data were recorded using Lab Streaming Layer (LSL) (https://github.com/sccn/labstreaminglayer). LSL is a middleware developed by the Swartz Center for Computational Neuroscience (UCSD), for unified collection of experimental measurements. This framework allows time-synchronization and online access to data while avoiding networking issues. LSL also provides a software, called LabRecorder, that allows to select and record all data of interest.
The robot position, battery level, amount of water available on board, temperature and automation level as well as the amount of water in the external tank, the state of the forest, elapsed time, leak position and participant actions on the interface were streamed via LSL each second [34]. With regard to the ECG equipment, data stream software was developed by Mega Electronics for LSL (software available at https://github.com/bwrc/faros-streamer). Faros ECG raw data (500 Hz) and Inter-Beat Interval (IBI)—or RR peaks intervals—were recorded. A Python script was developed based on the SMI SDK and LSL to produce streams containing the SMI measured gaze and fixation events.

2.5. Data Analysis

2.5.1. Data Processing

Performance Scores

The global mission performance was assessed using the total number of extinguished fires (see Section 2.2). Note that the human operator must properly manage the filling of the external tank, monitor the robot’s parameters (water reserve, battery, temperature, and automation level) and ensure manual control when necessary in order to reach a high mission score. In summary, the global mission score is intrinsically dependent on all of these tasks.
However, to evaluate the impact of the robot’s automation level, we also computed the total number of extinguished fires effectively done by the robot in autonomous mode and the total number of extinguished fires effectively performed by the human operator in manual control. In other words, for a given mission and a given automation level, the total number of fires is computed as the sum of fires extinguished during all 10-s time windows.

Subjective Feedback

The subjective feedback is composed by the scores collected with the NASA Task Load Index (or in short NASA-TLX) survey [36] and the Karolinska sleepiness scale (or KSS) questionnaire [35]. The NASA-TLX is a survey meant to assess the subjective task workload. It computes an overall workload score based on a weighted average of ratings of six dimensions: mental demand, physical demand, temporal demand, performance, effort, and frustration. The html version (available at https://www.keithv.com/software/nasatlx/), which automatically computes the score, was used in our experiment. Note that a high NASA-TLX score is associated with a high subjective workload. The KSS is a survey that is generally used to measure the subjective sleepiness (i.e., fatigue). The 9-graded KSS was used in our experiment. A high KSS score can be associated with a significant drowsiness.

Behavioral Data

The behavioral data concerns the human operator’s actions (keystrokes and mouse clicks) on the interface recorded every second. Only keystrokes and mouse clicks related to the different sub-tasks were considered (e.g., typing error was filtered). Thus, to assess on which tasks the participant’s actions were focused, we were interested in the average number of keystrokes/clicks performed and related with navigation and external tank sub-tasks given the robot’s level of automation, mission, and participant.
Additional metrics could also be computed based on the behavioral data. As stated before, the global mission score depends on a correct water level management of the external tank because the robot needs to recharge in water from it to extinguish fires. Therefore, given the fact that this sub-task needs to be performed by the human operator during the whole mission, we are also interested in the average variation of the water level of the external tank given the robot’s automation level, mission and participant. Since the time period between two automation level changes is a multiple of 10 s, the water level variations v R of the external tank are defined on successive (nonoverlapping) 10 s windows as the difference between the level at the end of the window l e R and the level at the beginning l b R : v = l e l b .

Eye-Tracking

Gaze movements and fixations events were recorded using the SMI Red 250 Hz ET system. We set a temporal threshold to 80 ms to identify instream fixation events. Five areas of interests (AOI, see Figure 3a) were defined on a reference image: AOI 1 refers to the current mission score and remaining time information; AOI 2 refers to the external tank task; AOI 3 represents the robot video stream; AOI 4 is the map of the zone; and, the AOI 5 summarizes the robot’s state variables related to the battery, the temperature, the amount of water, and the robot’s automation level. The Eye-Tracking device automatically detects fixations, i.e., when the gaze fixes more than 80 ms on a given screen point, and returns gaze location for each of them. Let us denote by f n the timestamp of the nth fixation and by p o s n = ( x n , y n ) R 2 its position on the screen. Given the time window w = [ w a , w b ] where ( w a , w b ) R 2 with w a < w b , we can define the set of considered indices W = { n N | f n w } N .
Now, given one of the screen zones (AOIs) shown in Figure 3a, here denoted by A O I i , the total number of fixations in A O I i is computed as follows,
F I X i ( W ) = n W 𝟙 { p o s n A O I i } ,
where i { 1 , , 5 } and 𝟙 A = 1 if A is true and 0 otherwise. This total number of fixations in each AOI F I X i is computed in this study on the 10-s time windows given the robot’s automation level, mission, and participant.

ECG Data

An ECG was used to collect participants’ cardiac activity at a sampling rate of 500Hz with the Faros Emotion 360 system. This ECG device filters the data and automatically extracts and streams the inter-beat intervals (IBI) or R-R intervals. Heart Rate (HR) was computed in beats per minute (bpm) using the serie of R-R intervals [37,38]. Given that I B I n = R n R n 1 , with R n as the timestamp of the nth R peak (i.e., highest positive peak), and given time window w = [ w a , w b ] with ( w a , w b ) R 2 with w a < w b , we can define the set of considered indices W = { n N | R n 1 w a n d R n w } N .
Thus, the mean HR can then be computed as follows,
H R ( W ) = 60 · # W n W I B I n ,
where, # W states for the cardinal of the set W .
The Heart Rate Variability (HRV) can also be computed in the time domain as the variability of the Inter-Beat Intervals (IBI):
H R V ( W ) = 1 # W 1 n W I B I n 60 H R ( W ) 2 .
Note that data collected during the rest periods (one minute before each mission, see Section 2.3) is used as a baseline for normalizing HR and HRV temporal metrics (computed with the Equations (2) and (3)) for a given mission and participant. Thus, for a given mission and participant, the set of indices associated with the rest sessions are denoted by W r e s t , and the normalized HR and HRV are computed as follows,
H R n o r m ( W ) = H R ( W ) H R ( W r e s t ) ,
H R V n o r m ( W ) = H R V ( W ) H R V ( W r e s t ) .
Thanks to this normalization, the ECG-based metrics of each mission performed by the 18 participants become inter-participant comparable physiological data. Thus, based on the Equations (4) and (5), we computed (i) the normalized HR ( H R n o r m ) and the normalized HRV ( H R V n o r m ) for the overall mission (10-min time window) per participant and mission; and (ii) the average of normalized HRs and HRVs computed on 10-s time windows, given the robot’s automation level, the mission, and the volunteer.

2.5.2. Statistical Analysis

Given the noncompliance of the normality assumption for all the computed metrics, all statistical analyses were ran using nonparametric tests from R packages v3.4 (available at https://cran.r-project.org/).

Mission Performance Analysis

A Spearman rank correlation was computed between the NASA-TLX and the mission scores (i.e., total number of extinguished fires). A Spearman rank correlation matrix was also computed between the physiological markers (i.e., ECG-based metrics), and the total mission score and the NASA-TLX score.

Behavioral and Physiological Markers’ Analysis Depending on the Robot Automation Level and the Performance Group

The main objective of this study is to assess whether the participants’ physiological and behavioral markers are impacted by the automation level in a different way if they perform the mission well or poorly. It was possible to characterize, using the median score of the 72 missions (4 missions × 18 participants), two well-balanced groups of mission performance: high score and low score groups. Such an analysis is based on the actual participant’s performance for a given mission, rather than average individual’s performance over all missions.
Thus, a Wilcoxon test was applied to compare the difference between the markers from the two groups (i.e., high and low score groups). However, when considering performance group and automation level, a pairwise Wilcoxon test for multiple comparisons with p-value adjustement was performed.
Moreover, nonparametric tests for multivariate data and many factor levels [39] were applied to behavioral (i.e., keystrokes/clicks) and eye-tracking (i.e., number of fixations on AOIs) markers. The factor levels concern the interaction between robot’s automation level and performance group. This test was chosen because the average number of actions on the interface (keystrokes/clicks) related to each sub-task, as well as, the average number of fixations in different AOIs, are dependent variables. Note that a pairwise Wilcoxon test for multiple comparisons with p-value was applied between factors levels as a post hoc test. These post-hoc tests were applied only to the dependent variables for which the Krushkal–Wallis test presented significant results.

2.5.3. Classification

Finally, two-dimensional classifiers were built to evaluate the performance predictive power of the behavioral and physiological features. More precisely, inter-subject single-trial classification models were studied exploiting the dataset composed by the ECG, the ET, and the keystrokes/clicks features, computed for 10-s time windows of a given automation level (see the paragraphs “Behavioral Data”, “Eye-tracking” and “ECG data” respectively in Section 2.5.1). Each sample was labeled by the related mission performance (high or low score). The total dataset is composed by 3591 samples, with 1459 samples for the low-scoring group and 2132 for the high-scoring one. Note that this unbalanced dataset can be explained by the fact that some low-scoring missions ended prematurely. In addition, several combinations of features were also studied in order to evaluate their contribution on the performance prediction with respect to the input device. The different combinations evaluated consisted in arrangements of ECG (i.e., HR and HRV), ET (i.e., number of fixation on each AOI), keystrokes/clicks (i.e., the number of keyboard inputs and clicks related the external tank and robot navigation) input features, all conditioned by the automation level of the robot. Thus, the automation level of the robot constitutes an additional common input feature for all tests. For instance, when ECG features are leaved out, classifications models were trained with only ET, keystrokes/clicks and the automation level of the robot as input features.
Well-known classifiers, such as, k-Nearest Neighbors (kNN), Linear and Quadratic Discriminant Analyses (LDA, QDA), Support Vector Machine (SVM), Gaussian Process (GP), Decision Trees (DT), Random Forest (RF), Neural Network (NN), AdaBoost (ADA), and Naive Bayes (NB), from the scikit-learn library version 0.20.3 were used. A grid search algorithm was applied for hyper-parameters optimization using 80% of data samples for all concerned algorithms given the balanced accuracy score. Then, each classifier was 5-fold cross-validated (CV = 5 ) twenty times ( 20 × 5 -fold); based on these runs, the average balanced accuracy and the 95% confidence interval were computed.

3. Results

3.1. Subjective Feedback

The Spearman’s rank correlation disclosed a significant negative correlation between the NASA-TLX and the global mission score ( ρ = 0.42 , N = 72 , p < 0.001 ). Figure 4 plots the final mission score versus the NASA-TLX score for each participant and mission.
A paired Wilcoxon test revealed no statistical difference between the KSS score between the beginning and the end of the experiment.

3.2. Behavioral Markers

3.2.1. Operator’s Contribution to Mission Score Given the Performance Group

Note that the robot is able to shoot water alone when in autonomous mode. Thus the actual operator’s score contribution relates to the number of fires extinguished when the robot is manually controlled. Figure 5a shows the scores reached considering all missions given the robot automation level. A Wilcoxon test showed that manual control scores are significantly higher then scores in autonomous mode ( p < 0.05 ) regardless of the performance group. However, a deeper analysis performed to evaluate the total number of extinguished fires depending on the robot automation level and performance group brought different results. A pairwise Wilcoxon test for multiple comparisons with p-adjustment showed that when the mission is part of the high score mission group, the participants have extinguished more fires in manual control than the robot in autonomous mode ( p < 0.001 ). As expected, during these high score missions, more fires were extinguished than during the low score missions, either in manual control ( p < 0.001 ) or in autonomous mode ( p < 0.001 ). Interestingly, for the low score group, the robot extinguished more fires in autonomous mode ( p = 0.05 ) than when the human operator was manually controlling the robot (see Figure 5b).

3.2.2. Operator’s Actions on the Interface Given the Robot Automation Level and the Performance Group

External Tank Water Level Variation With Respect to the Performance Group

A pairwise Wilcoxon test for multiple comparisons with p-adjustment showed that the high-scoring group reached a higher average variation on the external tank water level than the low-scoring group in autonomous mode ( p < 0.001 ). As expected, both the low score group ( p < 0.001 ) and high score group ( p < 0.001 ) presented a positive variation on tank level during the autonomous mode compared to manual control.

Keystrokes/Clicks Given the Robot Automation Level and Performance Group

A nonparametric test for multivariate data [39] was performed for the averaged number of keystrokes/clicks related to the external tank and navigation task. The four factor levels considered are the combinations of the two performance groups and the two levels of automation. A significant effect of this interaction was observed. The McKeon approximation for the Lawley Hotelling test gives a statistics of F ( 6.000 , 183.5663 ) = 105.336 , with p < 0.001 . These results are illustrated by Figure 6.
Post-hoc tests showed that
  • when the robot is in autonomous mode, participants use more actions (keystrokes/clicks) related to the external tank during high score missions than during low score missions ( p = 0.01 );
  • when the robot is in manual control, participants use more the navigation-related keystrokes during high-score missions than during low score missions ( p < 0.001 ); and
  • in autonomous mode participants use more the navigation-related keystrokes during low score missions than high score missions ( p < 0.01 ). Note that it represents behavioral errors since the robot does not need to be driven in autonomous mode (high automation level).

3.3. Eye-Tracking Markers

The same nonparametric test for multivariate data was performed for the number of fixations on the different AOIs. A significant effect of this interaction was observed (McKeon approximation for the Lawley Hotelling test: F ( 15 , 251.7024 ) = 34.694 , p < 0.001 ). These results are illustrated by Figure 7. Post hoc tests applied to relative effects between participants’ group and the robot automation level showed that
  • participants have more fixated AOI 1 (chronometer and score) during high score missions in both modes (with p < 0.001 for autonomous, and p < 0.001 for manual) than during low score missions;
  • participants have more fixated AOI 3 (robot video stream) during high score missions in both modes (autonomous p = 0.01 , manual p < 0.001 ) than during low score missions;
  • participants have more fixated AOI 4 (robot status) during high score missions in both modes ( p < 0.001 for both) than during low score missions.
  • participants have more fixated AOI 2 in autonomous mode ( p < 0.05 ) during high score missions than during low score missions; and
  • participants have more fixated AOI 2 in manual control during low score missions ( p < 0.001 ) than during high score missions. Note that it possibly demonstrates that these participants kept paying attention to the external tank sub-task instead of taking the robot’s manual control.

3.4. ECG Markers

3.4.1. ECG Markers, Mission Performance and NASA-TLX Score

A Spearman’s rank correlation matrix with adjusted p-value was computed between the table of normalized HR and HRV markers ( H R n o r m and H R V n o r m , cf. Equations (4) and (5)), and the table of mission and NASA-TLX scores. Note that the time window used here is the total duration of the mission (the associated set of R peak indices is denoted by W ).
Results showed that H R V n o r m ( W ) is correlated with NASA-TLX scores ( ρ = 0.42 , p = 0.001 ). However, only trends were observed between H R V n o r m ( W ) and mission score ( ρ = 0.24 , p = 0.07 ), and between H R n o r m ( W ) and mission score ( ρ = 0.27 , p = 0.06 ).

3.4.2. ECG Markers’ Analysis with Respect to Performance Group

A Wilcoxon test showed that participants have a significantly lower H R V n o r m ( W ) during high score mission than during low score missions ( W = 474 , p < 0.05 ). Also, H R n o r m ( W ) of the participants during high score missions is significantly higher than H R n o r m ( W ) during low score missions ( W = 883 , p < 0.01 ). These markers are illustrated in Figure 8 in function of the mission score.

3.4.3. ECG Markers’ Analysis Depending on the Performance Group and the Robot Automation Level

A nonparametric test for multivariate data was performed on averaged ECG features ( H R n o r m and H R V n o r m ) computed on 10-s time windows given the robot’s automation level and performance group. A significant effect was found (McKeon approximation for the Lawley Hotelling: F ( 6 , 183.5 ) = 2.323 , p < 0.05 ). Post hoc tests show a trend for H R V n o r m ( p = 0.09 for both automation levels) and H R n o r m ( p = 0.07 for both automation levels): the high score group presents a lower H R V n o r m and a higher H R n o r m compared to the low score group regardless of the robot’s automation level (cf. Section 3.1; Figure 9).

3.5. Classification

The average balanced accuracy and the 95% confidence interval results are illustrated in Figure 10 and summarized in Table 1. Interestingly, the best results found for all classifiers were obtained when all features—ECG, ET and keystrokes/clicks all conditioned on the automation level—were used as input features. The balanced accuracy results ranged from ( 0.74 ± 0.03 ) to ( 0.68 ± 0.04 ). When keystrokes/clicks features were removed, the results ranged from ( 0.72 ± 0.03 ) to ( 0.66 ± 0.03 ). When ET features were removed we obtained balanced accuracy results between ( 0.68 ± 0.02 ) and ( 0.61 ± 0.03 ). Balanced accuracy results ranged between ( 0.68 ± 0.03 ) and ( 0.62 ± 0.03 ) when ECG features were removed. Finally, when each ECG, ET or keystrokes/clicks features conditioned on the automation level of the robot were separately considered, the results were much less attractive.
The results indicated that the different classifiers (LDA, SVM, etc.) disclosed similar performance accuracy for a given set of input features (please report to mean and confidence interval in Table 1 and Figure 10). The highest average balanced accuracy was reached when all features were used as input features in this inter-subject single-trial approach, reflecting feature importance. These results demonstrated that all features were relevant to be used together for performance prediction. If we were to use a particular classifier, we would prefer the classifiers that have the best prediction results, i.e., here Support Vector Machine (SVM) and Gaussian Processes (GP). If it was necessary to train new classifiers several times, it might be better to use GP which has a lower computation time while a good prediction accuracy.

4. Discussion

The main objective of this study was to identify behavioral and psycho-physiological markers underpinning human’s performance that could be further used to drive the level of automation of robots in a mixed-initiative human–robot interaction (MI-HRI) paradigm. To assess discriminative physiological and behavioral features potentially relevant for MI-HRI, we designed an experimental scenario whereby the volunteers had to cooperate with a robot to fight forest fires. The level of automation of the robot (manual control versus autonomous mode) was randomly changed so as to measure its impact on the global performance of the human–robot system.
As previous works demonstrated that even experts may exhibit different abilities to perform similar tasks [40], we observed a variability among mission scores (as illustred in Figure 4). This observation is the cause of our choice in defining two mission score groups, separated by the median score (see Figure 8) that ensures two balanced groups. In our experiment, the high score group fought more fires as a consequence of a better navigation with the robot and better ability to manage the water tank in a given mission. Consistently, these two groups characterized by their mission score also differed at the psycho-physiological level. In the high score group, participants exhibited higher HR and lower HRV than in low score ones. Such an increase of HR and decrease of HRV are known to reflect a higher influence of the orthosympathetic nervous system on the autonomous nervous system, thus suggesting an increased catabolic activity to support the mobilization of cerebral resources to face the situation [41]. Conversely, poor behavioral performance associated with higher HRV may reveal task disengagement, and consequently an inability to face multitasking demands [42].
The distinction between high and low score missions allowed to conduct further analyses to assess the impact of the level of automation on the respective participants’ performance. This analysis disclosed that in the low score mission group, the autonomous mode of the robot improved the performance of the participants (i.e., more extinguished fires) whereas it had the opposite effect on the participants during the high score missions (see results in Section 3.2.1 and Figure 5b). Indeed, in the latter group, the participants fought more fires when manually controlling the robots. Moreover, this experiment revealed that the interaction with the robot could lead to conflict with automation. For instance, in low score missions, the participants experienced poor situation awareness of the current level of automation of the robot (see Section 3.3 concerning ET results), a phenomenon known as “mode error” [43] that led them to exhibit typical perseverative behaviors [10]. Indeed, during these low score missions, the participants erroneously attempted to manually control the robot in autonomous mode as revealed by their navigation related keystrokes. Conversely they also failed to take over when necessary as they hit less the navigation related keys when the robot was in manual control (low level of automation) than the participants during missions of the high score group. A reasonable explanation is that they were particularly focused on managing the external tank as revealed by a higher fixation rate on the external tank related AOI. Additionally, they glanced less at the robot’s parameters related AOI, therefore increasing the risk of ignoring important events (e.g., critical temperature or battery level) leading to abort the mission (cf. Section 3.3).
The statistical analysis allowed to assess promising physiological and behavioral features to predict performance depending on the robot automation level. Based on these results, ECG, ET and behavioral features were then used to train several classification algorithms in an inter-subject single-trial approach. Classification results showed that the highest average balanced accuracy was reached when all features were used together (see Figure 10) independently of the used classification algorithm. These results demonstrated the predictive power of these features. Recent works in robot teleoperation contexts have studied the use of ECG and/or ET features [24,44] to monitor human operator mental workload level or to detect degraded mental states. Thereby, and differently from them, our aim in this study was to highlight promising performance features that would help the HRI driving system to enhance human–robot interaction. In this sense, note that such a performance prediction would then be exploited by an intelligent HRI driving system to chose (or not) to change the automation level of the robot given the expected long-term performance.

4.1. Conclusion and Limitations

This paper is a first step towards MI-HRI driving systems. One first limitation is that we only manipulated two levels of automation. Such an approach already represented a challenging problem and proposed interesting results concerning useful discriminative features to predict mission performance. To go further, an interesting point would be to consider more levels of automation and their impact on the physiological and on the behavioral performance features. Sliding autonomy [45] literature gives examples where human operators could deliberatly choose the control mode [46], could be solicited to identify targets [4,5], or indicate to the system their intention of taking manual control [34,47]. Interestingly, the authors of [48] studied the impact on the human operator performance considering three levels of automation in a multi-robot scenario. They showed that a moderate automation level improved participants’ performance, and that higher levels of automation induced task disengagement. However, they also suggested that, due to individual differences that were quantified with an eye-tracking device, the appropriate automation level should depend on the underlying characteristics of the operator. Our results also go in this direction since they show that the automation level should be chosen in order to favor the upcoming mission performance given the human operator’s features pattern for instance.
This study has also other limitations related to the complexity of the tasks and the number of measures that were collected and analyzed (e.g., eye tracking, physiological, and subjective measures). It thus limits our interpretation and the ability to clearly isolate the relationship between the degree of autonomy of the robot and all the considered measures that account for our participants physiological and behavioral performance. This is an important issue as long as the sample size of volunteers was rather small ( n = 18) and homogeneous (i.e., same level of education and young people) that were not totally representative of the population of operators interacting with robots. Moreover, it is also recommended that gender be better balanced in our sample of participants. We should therefore conduct experiments with a larger sample of participants with different levels of studies and with different ages especially as long as several variables (e.g., age, gender, and body mass index) may affect physiological and behavioral response.
Additionally, in MI-HRI systems, which is the context of this work, initiative is mixed only when any agent (human or robot) is able to seize the initiative from the other [14,47]. It means that, in the MI context, allowing the operator to choose the control mode should only be possible if the interaction driving system considers it beneficial for the mission. For instance, consider an event that can only be handled by the human operator, and assume that the operator did not notice it. The interaction driving system would switch the robot control mode to autonomous mode, while informing the human operator that a significant event has occurred, even if the human operator intends to perform manual control. In another context, the operator’s intention to take control of the robot can be used by the interaction driving system to pass control to the operator to optimize the operation.

4.2. Practical Implications

This study demonstrates that an adaptive HRI driving system, aiming to maximize human–robot performance has to be carefully implemented. Indeed, this experiment highlights that the “human is still vital” [49], and is a key agent that needs to be taken into account when designing systems with operators in the control loop. For instance, a mixed-initiative interaction should switch the robot to manual control with ongoing high-scoring operators-like profiles to maintain their engagement and skills. Conversely, it should be preferred to trigger an autonomous mode with ongoing low-scoring operators-like profiles to prevent them from overloading. However, a decision system that drives human–robot interaction may not simply consist in regulating the level of automation. It is also critical to maintain the operators’ situation awareness when such mode transitions occur so that they are not only detected but also understood [43]; otherwise, conflict between the robot and the human could arise, therefore compromising mission success [49]. Several solutions have been put forward such as the use of cognitive countermeasures [50] and enhanced human–robot dialog [51] to improve automation transparency [52].

4.3. Future Research

As shown in Figure 7, Figure 8 and Figure 9, a high variability of ET and ECG features values is observed. This modulation is possibly related to the different events or different contexts during the mission. For instance, when leaks appear on the filling tank level it is normal to expect a high number of fixations on the related AOI. Thus, future research should consider the issue of how to aggregate the relevant features found with the mission context variables and automation level to better predict the mission performance. For instance it could be interesting to see how AOIs fixation pattern and HR/HRV features can relate with the external tank level, the number of leaks, the number of fires, the robot’s parameters and automation level. In particular, clustering algorithms could help identifying possible groups of features-context variables that would affect performance following the literature on cognitive states estimation. However, such a study would require a huge amount of data because of the need of sufficient representativeness (curse of dimensionality) appearing when many variables are considered. As in [34], a crowdsourcing platform can be exploited for this purpose.
Beyond that, and based on the achieved results of this paper, the next step in this research is to implement online methods to continuously infer the human operators’ performance like-profile (low versus high performer) and to launch adaptation based on the classification results obtained possibly used to feed a sequential decision-making process. More precisely, we believe that such an online performance profile prediction could be used to feed a stochastic sequential decision-making model dedicated to trigger adaptation [4,12].
Note that the conventional use of classifiers gives, in general, a one-shot label, performing an online tagging of the current input features. The certainty of this output label can be approximated by the accuracy of the classifier. For instance, different labels could be given as responses to similar features due to an inaccurate classification. It is known that classification errors can possibly bring the system to launch inappropriate actions. A promising solution to avoid launching inappropriate actions (e.g., actions based only on short term reasoning, called greedy actions) is to exploit the Partially Observable Markov Decision Process (POMDP) framework [4,12] to drive the human–robot interactions because: (i) it can integrate the uncertainties related to system states transitions (e.g., human actions) and related to observation functions (e.g., output labels from classifiers that are not always accurate); (ii) it can maintain a belief state distribution over possible human states (or operators-like profile); this belief state is updated each time an action is performed and an observation received using the Bayes rule; and (iii) its solution is an optimal policy—a function mapping reachable belief states to actions—that aims to maximize the expected long-term gains (i.e., entire system’s performance). In this sense, the belief state update process should estimate the system state (human, robot and mission states) during the interaction and choose to adapt the automation level of the robot with respect to the expected upcoming mission performance.

Author Contributions

Study conception and design: N.D., C.P.C.C., and R.N.R. Data acquisition: C.P.C.C. and N.D. Data Analysis: C.P.C.C. and N.D. Data Interpretation and Writing: C.P.C.C., N.D., R.N.R., and F.D. All authors have read and agreed to the published version of the manuscript.

Funding

This work has been financially supported by the Dassault Aviation Chair (https://www.isae-supaero.fr/fr/isae-supaero/mecenat-relations-avec-la-fondation-isae-supaero/chaire-dassault-aviation-casac) (CASAC 2016–2021).

Acknowledgments

Many thanks are given to Corentin Chauffaut for the constructive discussions about this work.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Finzi, A.; Orlandini, A. A Mixed-Initiative Approach to Human-Robot Interaction in Rescue Scenarios; American Association for Artificial Intelligence: Menlo Park, CA, USA, 2005. [Google Scholar]
  2. Nagatani, K.; Kiribayashi, S.; Okada, Y.; Otake, K.; Yoshida, K.; Tadokoro, S.; Nishimura, T.; Yoshida, T.; Koyanagi, E.; Fukushima, M.; et al. Emergency response to the nuclear accident at the Fukushima Daiichi Nuclear Power Plants using mobile rescue robots. J. Field Robot. 2013, 30, 44–63. [Google Scholar] [CrossRef]
  3. Bresina, J.L.; Morris, P.H. Mixed-initiative planning in space mission operations. AI Mag. 2007, 28, 75. [Google Scholar]
  4. Gateau, T.; Chanel, C.P.C.; Le, M.H.; Dehais, F. Considering human’s non-deterministic behavior and his availability state when designing a collaborative human-robots system. In Proceedings of the 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Daejeon, Korea, 9–14 October 2016; pp. 4391–4397. [Google Scholar]
  5. Bevacqua, G.; Cacace, J.; Finzi, A.; Lippiello, V. Mixed-Initiative Planning and Execution for Multiple Drones in Search and Rescue Missions. In Proceedings of the Twenty-Fifth International Conference in Automated Planning and Scheduling (ICAPS), Jerusalem, Israel, 7–11 June 2015; pp. 315–323. [Google Scholar]
  6. Dehais, F.; Duprès, A.; Blum, S.; Drougard, N.; Scannella, S.; Roy, R.N.; Lotte, F. Monitoring Pilot’s Mental Workload Using ERPs and Spectral Power with a Six-Dry-Electrode EEG System in Real Flight Conditions. Sensors 2019, 19, 1324. [Google Scholar] [CrossRef] [Green Version]
  7. Dehais, F.; Dupres, A.; Di Flumeri, G.; Verdiere, K.; Borghini, G.; Babiloni, F.; Roy, R. Monitoring Pilot’s Cognitive Fatigue with Engagement Features in Simulated and Actual Flight Conditions Using an Hybrid fNIRS-EEG Passive BCI. In Proceedings of the 2018 IEEE International Conference on Systems, Man, and Cybernetics (SMC), Miyazaki, Japan, 7–10 October 2018; pp. 544–549. [Google Scholar]
  8. Roy, R.N.; Bonnet, S.; Charbonnier, S.; Campagne, A. Mental fatigue and working memory load estimation: interaction and implications for EEG-based passive BCI. In Proceedings of the 2013 35th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Osaka, Japan, 3–7 July 2013; pp. 6607–6610. [Google Scholar]
  9. Dehais, F.; Rida, I.; Roy, R.N.; Iversen, J.; Mullen, T.; Callan, D. A pBCI to Predict Attentional Error Before it Happens in Real Flight Conditions. In Proceedings of the 2019 IEEE International Conference on Systems, Man and Cybernetics (SMC), Bari, Italy, 6–9 October 2019; pp. 4155–4160. [Google Scholar]
  10. Dehais, F.; Hodgetts, H.M.; Causse, M.; Behrend, J.; Durantin, G.; Tremblay, S. Momentary lapse of control: A cognitive continuum approach to understanding and mitigating perseveration in human error. Neurosci. Biobehav. Rev. 2019, 100, 252–262. [Google Scholar] [CrossRef] [Green Version]
  11. Crandall, J.W.; Goodrich, M.A.; Nielsen, C.W.; Olsen, D.R., Jr. Validating human–robot interaction schemes in multitasking environments. IEEE Trans. Syst. Man Cybern. Part A Syst. Hum. 2005, 35, 438–449. [Google Scholar] [CrossRef]
  12. de Souza, P.E.U.; Chanel, C.P.C.; Dehais, F. MOMDP-based target search mission taking into account the human operator’s cognitive state. In Proceedings of the IEEE 27th International Conference on Tools with Artificial Intelligence (ICTAI), Mare, Italy, 9–11 November 2015; pp. 729–736. [Google Scholar]
  13. Nikolaidis, S.; Ramakrishnan, R.; Gu, K.; Shah, J. Efficient Model Learning from Joint-Action Demonstrations for Human-Robot Collaborative Tasks. In Proceedings of the ACM/IEEE International Conference on Human-Robot Interaction, Portland, OR, USA, 2–5 March 2015; ACM: New York, NY, USA, 2015; pp. 189–196. [Google Scholar] [CrossRef] [Green Version]
  14. Jiang, S.; Arkin, R.C. Mixed-Initiative Human-Robot Interaction: Definition, Taxonomy, and Survey. In Proceedings of the 2015 IEEE International Conference on Systems, Man, and Cybernetics (SMC), Hong Kong, China, 9–12 October 2015; pp. 954–961. [Google Scholar]
  15. de Souza, P.E.U. Towards Mixed-Initiative Human-Robot Interaction: A Cooperative Human-Drone Team Framework. Ph.D. Thesis, Université de Toulouse, Délivré par l’Institut Supérieur de l’Aéronautique et de l’Espace (ISAE-SUPAERO), Toulouse, France, 2017. [Google Scholar]
  16. Vanderhaegen, F. Dissonance engineering: A new challenge to analyse risky knowledge when using a system. Int. J. Comput. Commun. Control 2014, 9, 776–785. [Google Scholar] [CrossRef]
  17. Scerbo, M.W. Adaptive Automation; Oxford University Press: Oxford, UK, 2008; pp. 239–252. [Google Scholar]
  18. de Visser, E.; Parasuraman, R. Adaptive aiding of human-robot teaming: Effects of imperfect automation on performance, trust, and workload. J. Cogn. Eng. Decis. Mak. 2011, 5, 209–231. [Google Scholar] [CrossRef]
  19. Mower, E.; Feil-Seifer, D.J.; Mataric, M.J.; Narayanan, S. Investigating implicit cues for user state estimation in human-robot interaction using physiological measurements. In Proceedings of the RO-MAN 2007—The 16th IEEE International Symposium on Robot and Human Interactive Communication, Jeju Island, Korea, 26–29 August 2007; pp. 1125–1130. [Google Scholar]
  20. Gateau, T.; Ayaz, H.; Dehais, F. In silico versus over the clouds: On-the-fly mental state estimation of aircraft pilots, using a functional near infrared spectroscopy based passive-BCI. Front. Hum. Neurosci. 2018, 12, 187. [Google Scholar] [CrossRef] [Green Version]
  21. Adams, J.A. Multiple robot/single human interaction: Effects on perceived workload. Behav. Inf. Technol. 2009, 28, 183–198. [Google Scholar] [CrossRef]
  22. Harriott, C.E.; Zhang, T.; Adams, J.A. Evaluating the applicability of current models of workload to peer-based human-robot teams. In Proceedings of the 2011 6th ACM/IEEE International Conference on Human-Robot Interaction (HRI), Ulm, Germany, 15–16 September 2011. [Google Scholar]
  23. Rodríguez-Fernández, V.; Menéndez, H.D.; Camacho, D. Analysing temporal performance profiles of UAV operators using time series clustering. Expert Syst. Appl. 2017, 70, 103–118. [Google Scholar] [CrossRef] [Green Version]
  24. Regis, N.; Dehais, F.; Rachelson, E.; Thooris, C.; Pizziol, S.; Mickaël Causse, C.T. Formal Detection of Attentional Tunneling in Human Operator-Automation Interactions. IEEE Trans. Hum. Mach. Syst. 2014, 44, 326–336. [Google Scholar]
  25. Rani, P.; Liu, C.; Sarkar, N.; Vanman, E. An empirical study of machine learning techniques for affect recognition in human–robot interaction. Pattern Anal. Appl. 2006, 9, 58–69. [Google Scholar] [CrossRef]
  26. Singh, G.; Bermùdez i Badia, S.; Ventura, R.; Silva, J.L. Physiologically Attentive User Interface for Robot Teleoperation—Real Time Emotional State Estimation and Interface Modification Using Physiology, Facial Expressions and Eye Movements. In Proceedings of the 11th International Joint Conference on Biomedical Engineering Systems and Technologies (BIOSTEC), Funchal, Madeira, Portugal, 19–21 January 2018. [Google Scholar]
  27. Heard, J.; Harriott, C.E.; Adams, J.A. A Survey of Workload Assessment Algorithms. IEEE Trans. Hum. Mach. Syst. 2018. [Google Scholar] [CrossRef]
  28. Ahlstrom, U.; Friedman-Berg, F.J. Using eye movement activity as a correlate of cognitive workload. Int. J. Ind. Ergon. 2006, 36, 623–636. [Google Scholar] [CrossRef]
  29. Chanel, C.P.C.; Wilson, M.D.; Scannella, S. Online ECG-based Features for Cognitive Load Assessment. In Proceedings of the IEEE International Conference on Systems, Man and Cybernetics, Bari, Italy, 6–9 October 2019. [Google Scholar]
  30. Fallahi, M.; Motamedzade, M.; Heidarimoghadam, R.; Soltanian, A.R.; Miyake, S. Effects of mental workload on physiological and subjective responses during traffic density monitoring: A field study. Appl. Ergon. 2016, 52, 95–103. [Google Scholar] [CrossRef]
  31. Tjolleng, A.; Jung, K.; Hong, W.; Lee, W.; Lee, B.; You, H.; Son, J.; Park, S. Classification of a Driver’s cognitive workload levels using artificial neural network on ECG signals. Appl. Ergon. 2017, 59, 326–332. [Google Scholar] [CrossRef]
  32. Valéry, B.; Matton, N.; Scannella, S.; Dehais, F. Global difficulty modulates the prioritization strategy in multitasking situations. Appl. Ergon. 2019, 80, 1–8. [Google Scholar] [CrossRef] [Green Version]
  33. Drougard, N.; Chanel, C.P.C.; Roy, R.N.; Dehais, F. Mixed-initiative mission planning considering human operator state estimation based on physiological sensors. In Proceedings of the IROS-2017 Workshop on Human-Robot Interaction in Collaborative Manufacturing Environments (HRI-CME), Vancouver, BC, Canada, 24 September 2017. [Google Scholar]
  34. Charles, J.A.; Chanel, C.P.C.; Chauffaut, C.; Chauvin, P.; Drougard, N. Human-Agent Interaction Model Learning based on Crowdsourcing. In Proceedings of the 6th International Conference on Human-Agent Interaction, Southampton, UK, 15–18 December 2018; pp. 20–28. [Google Scholar]
  35. Åkerstedt, T.; Gillberg, M. Subjective and objective sleepiness in the active individual. Int. J. Neurosci. 1990, 52, 29–37. [Google Scholar] [CrossRef]
  36. Hart, S.G.; Staveland, L.E. Development of NASA-TLX (Task Load Index): Results of empirical and theoretical research. In Advances in Psychology; Elsevier: Amsterdam, The Netherlands, 1988; Volume 52, pp. 139–183. [Google Scholar]
  37. Malik, M.; Bigger, J.T.; Camm, A.J.; Kleiger, R.E.; Malliani, A.; Moss, A.J.; Schwartz, P.J. Heart rate variability: Standards of measurement, physiological interpretation, and clinical use. Eur. Heart J. 1996, 17, 354–381. [Google Scholar] [CrossRef] [Green Version]
  38. Sörnmo, L.; Laguna, P. Bioelectrical Signal Processing in Cardiac and Neurological Applications; Elsevier Academic Press: Amsterdam, The Netherlands, 2005. [Google Scholar]
  39. Harrar, S.W.; Bathke, A.C. Nonparametric methods for unbalanced multivariate data and many factor levels. J. Multivar. Anal. 2008, 99, 1635–1664. [Google Scholar] [CrossRef] [Green Version]
  40. Wickens, C.D.; Dehais, F. Expertise in Aviation; Oxford University Press: Oxford, UK, 2019. [Google Scholar]
  41. Ko, S.M.; Ji, Y.G. How we can measure the non-driving-task engagement in automated driving: comparing flow experience and workload. Appl. Ergon. 2018, 67, 237–245. [Google Scholar] [CrossRef] [PubMed]
  42. Durantin, G.; Gagnon, J.F.; Tremblay, S.; Dehais, F. Using near infrared spectroscopy and heart rate variability to detect mental overload. Behav. Brain Res. 2014, 259, 16–23. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  43. Sarter, N.B.; Woods, D.D. How in the world did we ever get into that mode? Mode error and awareness in supervisory control. Hum. Factors 1995, 37, 5–19. [Google Scholar] [CrossRef]
  44. Landi, C.T.; Villani, V.; Ferraguti, F.; Sabattini, L.; Secchi, C.; Fantuzzi, C. Relieving operators’ workload: Towards affective robotics in industrial scenarios. Mechatronics 2018, 54, 144–154. [Google Scholar] [CrossRef]
  45. Huang, H.M.; Messina, E.; Albus, J. Toward a Generic Model for Autonomy Levels for Unmanned Systems (ALFUS); Technical Report; National Inst of Standards and Technology: Gaithersburg, MD, USA, 2003.
  46. Tang, F.; Ito, E. Human-assisted navigation through sliding autonomy. In Proceedings of the 2017 2nd International Conference on Robotics and Automation Engineering (ICRAE), Shanghai, China, 29–31 December 2017; pp. 26–30. [Google Scholar]
  47. Tang, F.; Mohammed, M.; Longazo, J. Experiments of human-robot teaming under sliding autonomy. In Proceedings of the 2016 IEEE International Conference on Advanced Intelligent Mechatronics (AIM), Banff, AB, Canada, 12–15 July 2016; pp. 113–118. [Google Scholar]
  48. Wright, J.; Chen, J.; Barnes, M. Human-automation interaction for multiple robot control: The effect of varying automation assistance and individual differences on operator performance. Ergonomics 2018, 61, 1033–1045. [Google Scholar] [CrossRef]
  49. Parasuraman, R.; Wickens, C.D. Humans: Still Vital After All These Years of Automation. Hum. Factors 2008, 50, 511–520. [Google Scholar] [CrossRef]
  50. Dehais, F.; Causse, M.; Vachon, F.; Tremblay, S. Cognitive conflict in human–automation interactions: A psychophysiological study. Appl. Ergon. 2012, 43, 588–595. [Google Scholar] [CrossRef] [Green Version]
  51. Dehais, F.; Peysakhovich, V.; Scannella, S.; Fongue, J.; Gateau, T. Automation surprise in aviation: Real-time solutions. In Proceedings of the 33rd Annual ACM Conference on Human Factors in Computing Systems, Seoul, Korea, 18–23 April 2015; pp. 2525–2534. [Google Scholar]
  52. Lyons, J.B. Being transparent about transparency: A model for human-robot interaction. In Proceedings of the 2013 AAAI Spring Symposium Series, Palo Alto, CA, USA, 25–27 March 2013. [Google Scholar]
Figure 1. Screenshot of the graphical user interface (GUI) during a mission. Top left: remaining time and score. Top right: robot camera feedback. Bottom left: water stock management task. Bottom right: robot status.
Figure 1. Screenshot of the graphical user interface (GUI) during a mission. Top left: remaining time and score. Top right: robot camera feedback. Bottom left: water stock management task. Bottom right: robot status.
Sensors 20 00296 g001
Figure 2. Experimental protocol time line.
Figure 2. Experimental protocol time line.
Sensors 20 00296 g002
Figure 3. (a) Screenshot of the graphical user interface showing the areas of interest (AOIs) defined for the study; (b) ECG-data acquisition during the experiment.
Figure 3. (a) Screenshot of the graphical user interface showing the areas of interest (AOIs) defined for the study; (b) ECG-data acquisition during the experiment.
Sensors 20 00296 g003
Figure 4. Final mission score versus NASA-TLX score. Each plotted point relates to one of the 72 missions.
Figure 4. Final mission score versus NASA-TLX score. Each plotted point relates to one of the 72 missions.
Sensors 20 00296 g004
Figure 5. Score analysis given the automation level of the robot and the performance group.
Figure 5. Score analysis given the automation level of the robot and the performance group.
Sensors 20 00296 g005
Figure 6. Average number of keystrokes/clicks per sub-task (filling the external tank and driving the robot) during autonomous mode and manual control given performance groups. These graphical representations reappear in the next figures. They show a large number of quantiles, which makes it easier to visualize nonparametric distributions.
Figure 6. Average number of keystrokes/clicks per sub-task (filling the external tank and driving the robot) during autonomous mode and manual control given performance groups. These graphical representations reappear in the next figures. They show a large number of quantiles, which makes it easier to visualize nonparametric distributions.
Sensors 20 00296 g006
Figure 7. Average number of fixations on AOIs during autonomous mode and manual control, given the performance group. AOI 1: remaining time; AOI 2: external tank; AOI 3: robot video stream; AOI 4: map; and AOI 5: robot’s parameters.
Figure 7. Average number of fixations on AOIs during autonomous mode and manual control, given the performance group. AOI 1: remaining time; AOI 2: external tank; AOI 3: robot video stream; AOI 4: map; and AOI 5: robot’s parameters.
Sensors 20 00296 g007
Figure 8. Normalized HR and HRV markers by performance group plotted in function of the final mission score. Each plotted point relates to one of the 72 missions.
Figure 8. Normalized HR and HRV markers by performance group plotted in function of the final mission score. Each plotted point relates to one of the 72 missions.
Sensors 20 00296 g008
Figure 9. ECG features during autonomous mode and manual control given participants’ performance group.
Figure 9. ECG features during autonomous mode and manual control given participants’ performance group.
Sensors 20 00296 g009
Figure 10. Average balanced accuracy for 20×5-fold cross-validations and the 95% confidence interval. Input features were conditioned on the automation level of the robot, which constitutes an additional common input feature for all tests. kNN: k-Nearest Neighbors; LDA: Linear Discriminant Analyses; QDA: Quadratic Discriminant Analyses; SVM: Support Vector Machine; GP: Gaussian Process; DT: Decision Trees; RF: Random Forest; NN: Neural Network; ADA: AdaBoost; and NB: Naive Bayes.
Figure 10. Average balanced accuracy for 20×5-fold cross-validations and the 95% confidence interval. Input features were conditioned on the automation level of the robot, which constitutes an additional common input feature for all tests. kNN: k-Nearest Neighbors; LDA: Linear Discriminant Analyses; QDA: Quadratic Discriminant Analyses; SVM: Support Vector Machine; GP: Gaussian Process; DT: Decision Trees; RF: Random Forest; NN: Neural Network; ADA: AdaBoost; and NB: Naive Bayes.
Sensors 20 00296 g010
Table 1. Average balanced accuracy results for 20×5-fold cross-validations and the 95 % confidence interval of two-dimensional inter-subject single-trial classifications including different combinations of input features.
Table 1. Average balanced accuracy results for 20×5-fold cross-validations and the 95 % confidence interval of two-dimensional inter-subject single-trial classifications including different combinations of input features.
Different Combinations of Input Features
ClassifiersECG, ET and Keys/ClicksECG and ETECG and Keys/ClicksET and Keys/ClicksETECGKeys/Clicks
kNN 72 % ± 3 % 70 % ± 4 % 67 % ± 3 % 66 % ± 4 % 63 % ± 3 % 63 % ± 3 % 59 % ± 3 %
LDA 70 % ± 3 % 66 % ± 3 % 61 % ± 3 % 65 % ± 3 % 62 % ± 3 % 54 % ± 3 % 59 % ± 3 %
QDA 72 % ± 3 % 69 % ± 3 % 66 % ± 3 % 67 % ± 3 % 63 % ± 3 % 63 % ± 3 % 60 % ± 3 %
SVM 74 % ± 3 % 71 % ± 3 % 67 % ± 3 % 67 % ± 3 % 63 % ± 3 % 64 % ± 3 % 59 % ± 3 %
GP 74 % ± 3 % 72 % ± 3 % 68 % ± 2 % 68 % ± 3 % 65 % ± 3 % 64 % ± 3 % 60 % ± 2 %
DT 68 % ± 4 % 68 % ± 4 % 65 % ± 4 % 63 % ± 4 % 63 % ± 4 % 63 % ± 4 % 60 % ± 2 %
RF 72 % ± 3 % 70 % ± 3 % 65 % ± 4 % 65 % ± 3 % 62 % ± 3 % 64 % ± 3 % 60 % ± 3 %
NN 73 % ± 3 % 72 % ± 3 % 68 % ± 3 % 68 % ± 4 % 64 % ± 3 % 64 % ± 3 % 60 % ± 3 %
ADA 73 % ± 3 % 70 % ± 3 % 67 % ± 3 % 65 % ± 2 % 63 % ± 3 % 63 % ± 3 % 59 % ± 3 %
NB 69 % ± 3 % 68 % ± 3 % 65 % ± 3 % 62 % ± 3 % 60 % ± 4 % 62 % ± 4 % 50 % ± 2 %
ECG features: HR and HRV for 10-s time windows. ET features: total number of fixations in AOI1, in AOI2, in AOI3, in AOI4, and in AOI5 for 10-s time windows. Keys/clicks features: total number of keystrokes/clicks related to external tank task and to the robot navigation task for 10-s time window. Recalling all input features were conditioned on the automation level of the robot, which constitutes an additionnal common input feature for all tests. kNN: k-Nearest Neighbors; LDA: Linear Discriminant Analyses; QDA: Quadratic Discriminant Analyses; SVM: Support Vector Machine; GP: Gaussian Process; DT: Decision Trees; RF: Random Forest; NN:Neural Network; ADA: AdaBoost; and, NB: Gaussian Naive Bayes.

Share and Cite

MDPI and ACS Style

Chanel, C.P.C.; Roy, R.N.; Dehais, F.; Drougard, N. Towards Mixed-Initiative Human–Robot Interaction: Assessment of Discriminative Physiological and Behavioral Features for Performance Prediction. Sensors 2020, 20, 296. https://doi.org/10.3390/s20010296

AMA Style

Chanel CPC, Roy RN, Dehais F, Drougard N. Towards Mixed-Initiative Human–Robot Interaction: Assessment of Discriminative Physiological and Behavioral Features for Performance Prediction. Sensors. 2020; 20(1):296. https://doi.org/10.3390/s20010296

Chicago/Turabian Style

Chanel, Caroline P. C., Raphaëlle N. Roy, Frédéric Dehais, and Nicolas Drougard. 2020. "Towards Mixed-Initiative Human–Robot Interaction: Assessment of Discriminative Physiological and Behavioral Features for Performance Prediction" Sensors 20, no. 1: 296. https://doi.org/10.3390/s20010296

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop