Next Article in Journal
Improved Adaptive Kalman-Median Filter for Line-Scan X-ray Transmission Image
Next Article in Special Issue
Attention Detection by Heartbeat and Respiratory Features from Radio-Frequency Sensor
Previous Article in Journal
PD Flexible Built-In High-Sensitivity Elliptical Monopole Antenna Sensor
Previous Article in Special Issue
Measuring Heart Rate Variability Using Facial Video
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Automatic System for Continuous Pain Intensity Monitoring Based on Analyzing Data from Uni-, Bi-, and Multi-Modality

Department of Neuro-Information Technology, Institute for Information Technology and Communications, Otto-von-Guericke University Magdeburg, 39106 Magdeburg, Germany
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(13), 4992; https://doi.org/10.3390/s22134992
Submission received: 6 June 2022 / Revised: 26 June 2022 / Accepted: 30 June 2022 / Published: 1 July 2022
(This article belongs to the Special Issue Monitoring Technologies in Healthcare Applications)

Abstract

:
Pain is a reliable indicator of health issues; it affects patients’ quality of life when not well managed. The current methods in the clinical application undergo biases and errors; moreover, such methods do not facilitate continuous pain monitoring. For this purpose, the recent methodologies in automatic pain assessment were introduced, which demonstrated the possibility for objectively and robustly measuring and monitoring pain when using behavioral cues and physiological signals. This paper focuses on introducing a reliable automatic system for continuous monitoring of pain intensity by analyzing behavioral cues, such as facial expressions and audio, and physiological signals, such as electrocardiogram (ECG), electromyogram (EMG), and electrodermal activity (EDA) from the X-ITE Pain Dataset. Several experiments were conducted with 11 datasets regarding classification and regression; these datasets were obtained from the database to reduce the impact of the imbalanced database problem. With each single modality (Uni-modality) experiment, we used a Random Forest [RF] baseline method, a Long Short-Term Memory (LSTM) method, and a LSTM using a sample weighting method (called LSTM-SW). Further, LSTM and LSTM-SW were used with fused modalities (two modalities = Bi-modality and all modalities = Multi-modality) experiments. Sample weighting was used to downweight misclassified samples during training to improve the performance. The experiments’ results confirmed that regression is better than classification with imbalanced datasets, EDA is the best single modality, and fused modalities improved the performance significantly over the single modality in 10 out of 11 datasets.

1. Introduction

Pain is an indicator of the human body’s health condition; it is a warning mechanism that gives a trusted message to indicate people to pay serious medical attention and respond quickly. Recent studies have shown that behavioral cues and physiological signals are very informative for pain detection [1,2,3]. Therefore, several automated methods have been proposed using these cues and signals to overcome the limitations of current clinical applications for pain recognition, such as their unreliability and non-continuous monitoring for patients, especially those who cannot self-report their pain (infants, intensive care patients, or people suffering from dementia). This paper focuses on introducing an automated system that can be used to complement the current clinical methods for better pain management by analyzing behavioral and physiological data from five modalities of the X-ITE Pain Dataset. This multimodal pain database provides additional valuable information for more advanced discriminating pain or pain intensities versus no pain because it includes different qualities and durations of pain [4].
Visual cues (facial expressions), vocalization cues (verbally and non-verbally), and physiological signals (bio-signals) express pain physically; these cues and signals play an important role in assessing pain in individuals [3]. Facial expressions are one of the more important cues in predicting pain based on analyzing human behaviors during pain [5,6]; across different nationalities, genders, cultures, ages, and genders, pain is expressed similarly. Further, some individuals express pain non-verbally by moaning, crying, groaning, and sighing (vocalization cues), and some express pain verbally by mentioning pain or using offensive words, and some express their pain [7,8]. Additionally, the physiological signals are informative indicators of pain, they cause alterations in tissues and organs (e.g., skin, heart, and muscles’ electrical properties). The most widely used signals are electrocardiogram (ECG), facial electromyography (EMG), and electrodermal activity (EDA) [9,10].
For objective pain intensity assessment, an increasing number of studies [10,11,12,13,14] has investigated behavioral cues and physiological signals with machine learning models. The findings demonstrated that the EDA modality mostly outperformed other physiological single modalities in terms of accurate pain assessment; EMG was the second best modality. Further, the temporal integration of extracted features from each modality improved the performance of pain assessment [14,15,16]. The temporal integration was represented for each modality by a time series statistics descriptor that was calculated from several statistical measures with their first and second derivatives per time series. The purpose of this paper is to confirm the above works’ findings, utilize temporal integration, and combine the best two modalities (EMG and EDA) and all modalities to improve the pain intensity monitoring performance; various studies [10,13,14,17,18] motivated us to use the combined modalities. The combination of both behavioral and physiological pain cues/signals is potentially good for developing objective pain assessment regarding the Odhner et al. [19] study. We also compared the results of the single modalities versus fused modalities (two and all modalities) to introduce the best automated system for objective continuous monitoring of pain intensity, which could be highly beneficial for reliable and economic pain intensity assessment.
The current work is organized as follows. Section 1.1 describes the contribution of this work. Section 2 provides an overview of pain recognition methods based on behavioral and physiological responses and describes their relevance to this paper. Section 3 shows a description of the materials and methods for automatic monitoring of continuous pain intensity using fused modalities: The X-ITE Pain Database preprocessing is shown in Section 3.1, automatic monitoring methods for continuous pain intensity regarding classification and regression are explained in Section 3.2, and the experimental setup is explained in Section 3.3. Section 4 presents a comparison between models’ results when using single and fused modalities regarding classification and regression, followed by a discussion in Section 5 and a conclusion in Section 6.

1.1. Contributions

Werner et al. [14] and Walter et al. [16] reported the results of using phasic (short) and tonic (long) stimulation samples from facial expressions, audio, and psychological data of 7 s, which were cut out from the continuous recording of the main stimulation phase in X-ITE Pain Database. This paper goes beyond their studies, as it reports the first continuous pain monitoring results based on analyzing data from the best two fused modalities (EMG and EDA) and all fused modalities (facial expressions, audio, ECG, EMG, EDA) of the same database. In line with this paper, the facial expressions modality results were reported in [20,21] and both the facial expressions and EDA modalities’ results were reported in [22]. The findings from this paper outperformed those from our recent paper [20,21,22]. We used the continuous recording of each modality of most of the data collection experiments, which is about 1 h and a half per subject. Further, we compared classification and regression results when using RF, LSTM, and LSTM-SW. We did not apply RF with fused modalities experiments because it was not needed due to its lower performance compared to LSTM and LSTM-SW during best single modality (EDA) experiments. We preferred to continue experiments with promising methods. Additionally, we compared models’ performances achieved using EDA (best single modality) with fused modalities (the best two modalities (EMG and EDA) and all modalities) to obtain the best method of recognizing pain intensity for the two pain stimulus types (phasic and tonic) in three pain intensities (low, moderate, and severe) and two qualities (heat and electrical stimuli). The findings of this work are the baseline results for future research related to real-time pain intensity monitoring systems using the EDA and fused modalities ((EMG and EDA) and all modalities) in the X-ITE Pain Database.

2. Related Work

Many studies have focused on various possible objective indicators for pain. The behavioral and physiological pain indicators are commonly used for pain assessment. Facial expressions and vocalizations (from audio) are considered behavioral pain indicators; heart rate variability, muscle activity, and electrodermal activity are considered physiological pain indicators. Recent studies reported that the fusion of these indicators increased the accuracy of distinguishing no pain versus pain [14,16,19]. Ekman and Friesen [23] decomposed facial expressions into individual facial Action Units (AUs) with the Facial Action Coding System (FACS). Examples of facial changes associated with pain are brow lower, cheek raise, lids tight, nose wrinkle, nasolabial deepen, upper lip raise, lip corner pull, lip stretch, lips apart, jaw drop, lids droop, eyes closed, and blink [1,23,24,25,26]. Vocalizations include verbal, non-verbal, and breathing behaviors. They define during pain experienced as the utterance of sounds, noises, and words using the vocal apparatus. Verbal vocalizations include protests and complaints by mentioning pain or using offensive words [27]; non-verbal vocalizations include moaning, crying, groaning, and gasps [7]; the changes in breathing patterns such as sighing have also been considered as vocalizations in regard to Waters et al. [28].
Clinical studies [29,30,31,32,33] have provided strong empirical evidence for the correlation between individual physiological signals and pain. The pain process starts from the sensory receptors (also called nociceptors) by noxious thermal, chemical, or mechanical stimuli, which can be activated to the body from an external or internal source. The information regarding detecting harmful stimuli and converting these into electrical signals is transduced via nociceptors and transmitted through the spinal cord to the brain. Then, specific parts of the brain are responsible for responding to pain signals, which are the prosencephalon, mesencephalon, and cortex [34,35,36,37]. ECG signals captures the changes in the electrical activity of the heartbeats (low-frequency/high-frequency ratio) and the heart rate interval. The Heart Rate Variability (HRV) is calculated on ECG data; the changes of HRV in the low-frequency power increase during painful stimulation [38,39]. EMG signals measures the changes in electrical properties of the muscle; EMG activity is often measured at the zygomaticus (mouth corner raiser), trapezius (back of the neck), and corrugator superscillii (brow lowerer) muscles [40]. EDA signals record the changes in the electrical activity of the skin when using two electrodes connected to the index and ring fingers. In response to a pain stimulus, EDA is a good measure because of intense body activity after experiencing pain; when a painful stimulus is applied, the sympathetic nervous system (SNS) activates the finger’s sweat glands to produce more sweat, and this, in turn, increases skin conductance [41,42,43].
Many promising methods were introduced for automatically recognizing pain using behavioral and physiological pain indicators because they carry valuable information about different pain levels in patients. The automatic pain recognition consists of three main steps: (1) face detection and registration/vocalizations preprocessing/physiological preprocessing, (2) feature extraction, and (3) pain recognition. Several automatic methods using facial expressions were introduced for recognizing frame-level and sequence-level pain intensity. Frame-level methods ignore temporal information and are thus limited in describing relevant dynamic information that is beneficial for pain intensity recognition. Further, occlusion, such as self-occlusion, oxygen mask, and pacifier, is another limitation of using such methods. Thus, many recent works focus on sequence-level pain recognition because it is more effective in describing such information [44,45,46]. According to the ability of RF [47] for pain detection using facial expressions, audio, or physiological indicators [13,14,20,21,22,45,48,49], we introduced in this work RF as a baseline method for single modalities experiments and compared its performance to the proposed deep learning methods (LSTM and LSTM-SW) that analyze data encoding temporal information.
Most methods of pain recognition used a single modality [44,50] used video, [51,52] used audio signal, and [10,12,33] used physiological signals. The recent methods used multiple modalities [3,4,14,16,53,54,55] that can improve the performance and flexibility of pain recognition. Thiam et al. [54] proposed multi-modal methods to develop pain intensity classification. They introduced a supervised deep learning method and a self-supervised method for recognizing pain intensity based on physiological signals. The self-supervised method automatically generated physiological data and simultaneously performed a fine-tuning of the deep learning model, which had been previously trained on a significantly smaller amount of data. Thus, they were able to significantly improve the data efficiency. Hinduja et al. [55] introduced a multi-modal method for pain recognition by fusing facial expressions and physiological signals (heart rate, respiration, blood pressure, and electrodermal activity) from the BP4D+ database. The fusion improved accuracy when evaluation included all subjects or same gender compared to using only one modality (facial expressions or physiological signals). Pouromran et al. [10] focused on pain intensity estimation using the BioVid dataset; they built different machine learning models for continuous pain estimation: Linear Regression, Support Vector Regression (SVR), Neural Networks, k-nearest neighbors, Random Forest, and XGBoost. They used the extracted features from a single modality and the combination features from multiple modalities. The EDA single model outperformed multi-modal results for pain intensity recognition, and SVR gave the best predictive performance across different sensors.
The current automated methods for pain assessment have not yielded estimation accuracies acceptable in clinical settings due to several reasons: (1) Half of the available pain databases for research purposes contain response data from a single modality (e.g., facial expressions). (2) The multimodal pain databases have a significant impact on the performance of automatic pain assessment systems; to the common belief that the quality and duration of pain are need to provide additional valuable information for more advanced discriminating pain or pain intensities versus no pain, the X-ITE Pain Database has been made to complement existing databases and the analysis of pain regarding quality and length. (3) There are some studies to assess pain reported promising results of using more than three modalities; they show that to obtain a reliable pain assessment system, it is recommended to analyze pain and detecting valid pain patterns from multiple modalities, including both behavioral cues and physiological signals. (4) Few of the current pain assessment methods focus on continuous monitoring, which is more necessary for pain assessment for prompt pain detection and immediate intervention. Therefore, this paper addresses those limitations and proposes an automatic and multimodal system for continuous monitoring of pain intensity using five sensor modalities (facial expressions, audio, ECG, EMG, and EDA) from the X-ITE Pain Database. A combination of behavioral cues and physiological signals was used with appropriate machine learning methods. LSTM [56] and LSTM using sample weighting (LSTM-SW) were utilized with a single or multiple fused modalities (facial expression, audio, ECG, EMG, EDA, (EMG and EDA), or all modalities) to predict continuous phasic or tonic pain intensity versus no pain. LSTM was proposed to learn long-term dependency among longer time periods by storing information from previous periods. The sample weighting method [48] was applied to reduce the weight of misclassified samples by duplicating some training samples with more facial responses if their classification scores were above 0.3 to improve the pain intensity recognition performance.

3. Materials and Methods

Figure 1 shows the methodology of the proposed automatic system for continuous monitoring of pain intensity. We pre-processed the input data from the five modalities (facial expressions, audio, ECG, EMG, and EDA) to extract useful features. Then, we calculated temporal integration features from time series data coming from those five modalities. After that, we prepared the experimental data by further processing the obtained data from the temporal integration process; such processing was suggested to overcome imbalanced database and outlier problems; for more details, see Section 3.1. Regarding classification and regression, three types of experiments were introduced to analyze the experimental data for continuous monitoring of pain intensity: (1) Uni-modality (data from single modalities) experiments, (2) Bi-modality (fusing data from two modalities) experiments, and (3) Multi-modality (fusing data from five modalities) experiments. We applied RF as baseline methods (Random Forest classifier (RFc) and Random Forest regression (RFr)) with single modalities experiments, Long Short-Term Memory (LSTM) and LSTM using the sample weighting method (LSTM-SW) with single and fused modalities experiments; see Section 3.2. For more details about the conducted experiments, see Section 3.3.

3.1. X-ITE Pain Database Pre-Processing

In this work, a multimodal Experimentally Induced Thermal and Electrical (X-ITE) Pain Database [4] includes data that were recorded when healthy participants (subjects) were exposed to two different pain qualities (heat and electric) in three intensities (low, medium, and high) and two different pain stimuli durations (phasic (5 s = short) and tonic (1 min = long)). The subjects were aged between 18 and 50 years (at least 90% Caucasian participants = 67 men and 67 women). The heat pain stimulus was stimulated at participants’ forearm using a thermal stimulator (Medoc PATHWAY Model ATS, Medoc—Advanced Medical Systems, Ramat Yishay, Israel). The electrical pain stimulus was stimulated at participants when electrodes were attached to participants’ index and middle fingers using an electrical stimulator (Digitimer DS7A, Digitimer Ltd., Letchworth Garden City, UK). A person calibration was applied prior to the main stimulation phase experience, which means the participant self-reported the pain experienced during several stimuli using the numeric rating scale. Three intensities of both pain stimuli (heat and electricity) of each duration pain (phasic or tonic) were selected individually based on participants’ personal pain sensitivity (tolerances). For each phasic stimulus, the three pain intensities (times two pain modalities) were repeated 30 times for 5 s duration, applied in randomized order with pauses of 8–12 s. The 1-min tonic stimuli were applied once per intensity, followed by a pause of 5 min. The EDA includes both background tonic (Skin Conductance Level: SCL = 1-min pain stimuli) and rapid phasic components (Skin Conductance Responses: SCRs = 5-min pain stimuli). There were three phases of how tonic heat and electrical pain intensity stimuli were applied: the two lower intensities were applied randomly during the phasic stimulus period, and the highest intensity was applied at the end of the experiment. The entire experiment (preparation and actual experiment) took about 3 h per participant; approximately 1 and a half hours was the duration of the actual experiment for each participant, which was used in this work. For more details see Gruss et al. [4]. The five modalities were analyzed to objectively monitor the phasic and tonic pain intensity during the application of the thermal and electrical pain stimuli and no pain.
In line with [14,16,20,21,22,48], the subset of the same 127 human healthy participants (subjects) was selected, including samples only, for which data were available from five sensor modalities (frontal facial RGB video, audio, ECG, EMG, and EDA). The data from the five modalities from X-ITE Pain Database were pre-processed by (1) processing the frontal facial RGB video using OpenFace [57] for detecting the face from each frame for each participant (subject) and for extracting facial features (FF) and head pose, (2) processing the audio signal using openSMILE [58], (3) applying the QRS-detection algorithm by Hamilton et al. [59] with the ECG signals, (4) processing the three EMG channels with a zero-phase third-order Butterworth band-pass filter, and (5) keeping EDA without filtering. The FF that were used included 3 head poses (Yaw, Pitch, and Roll), AU1 (binary occurrence output), and 17 AU intensity outputs of OpenFace, which are AU1 (Inner Brow Raiser), AU2 (Outer Brow Raiser), AU4 (Brow Lowerer), AU5 (Upper Lid Raiser), AU6 (Cheek Raiser), AU7 (Lid Tightener), AU9 (Nose Wrinkler), AU10 (Upper Lip Raiser), AU12 (Lip Corner Puller), AU14 (Dimpler), AU15 (Lip Corner Depressor), AU17 (Chin Raiser), AU20 (Lip stretcher), AU23 (Lip Tightener), AU25 (Lips Part), AU26 (Jaw Drop), and AU45 (Blink). The low-level descriptor (LLD) was extracted from audio modality, comprising 4 energy features (sum of auditory spectrum (loudness), sum of RASTA-filtered auditory spectrum, root-mean-square (RMS) energy, and zero-crossing rate), 6 voicing features (F0 (Subharmonic Summation (SHS)\and Viterbi smoothing), probability of voicing, logarithmic harmonics-to-noise ratio (HNR), Jitter (local and delivered duty paid (DDP)), Shimmer (local)), and 14 spectral features (Mel Frequency Cepstral Coefficients (MFCCs)). With processed ECG, R-to-R intervals were used to determine heart rate. Afterward, we interpolated the heart rate signal linearly to match the sampling of the EMG and EDA (1000 Hz). The 21-dimensional facial expression/24-dimensional audio/1-dimensional ECG/3-dimensional EMG/1-dimensional EDA time series were selected at a frame rate of 25 frames per second (fps).
The temporal integration for each modality was represented by a time series statistics descriptor [44,60] to describe the changes of features from each modality, which are called Facial Activity Descriptor (FAD), Audio Descriptor (Audio-D), ECG Descriptor (ECG-D), EMG Descriptor (EMG-D), and EDA Descriptor (EDA-D). Each second in each descriptor was summarized by four statistics of the time series itself and its first and second derivative, including minimum, maximum, mean, and standard deviation, yielding a 12 × 21-dimensional, 12 × 24-dimensional, 12 × 1-dimensional, 12 × 3-dimensional, and 12 × 1-dimensional descriptor for facial, audio, ECG, EMG, and EDA features, respectively. A person-specific standardization of the features [44] was applied with all descriptors in order to focus on the within-subject response variation rather than the differences between subjects. For each subject, the mean and standard deviation were calculated, then each feature value was subtracted by the mean and divided by the standard deviation that belonged to the same subject. RF, LSTM, and LSTM-SW were used with all descriptors. The descriptor for each modality we use was at the same frame rate (1/25 s = 25 fps) in line with [20,21,22]. A sliding window strategy was used to obtain input samples with a time length of 10 s ago.
In the X-ITE Pain Database, −10 indicates samples with problems such as false start and restart of the stimuli, overlapping between heat or electrical stimulation, unbalanced phasic estimation, short pause, short tonic electrical stimulus, single heat stimulus in front, or additional stimulus. Furthermore, −11 indicates the samples when the subject speaks or interacts during the experiment (the beginning and after the first and second tonic stimuli of the experiment). The remaining annotations were presented in Table 1.
Several pre-processing steps were proposed on the X-ITE Pain Database to reduce the impact of the extremely imbalanced database problem: First, we investigated the intensities of facial expressions for most samples when expressing pain intensity, then we assigned all subjects into four categories based on how they expressed pain intensity. Second, we suggested splitting the database into three splits: training set, validation, and testing set. Randomly, subjects for each split from each category were selected based on 80% of data for training (100 subjects = 572,696 samples), 10% for validation (13 subjects = 75,537 samples), and 10% for testing (14 subjects = 79,485 samples). Each split contained subjects from all intensity categories. Third, the obtained splits from the database were processed: (a) all sequences of samples with labels −10, −11, and no pain samples sequence before and after these samples were excluded to simplify the problem and reduce the impact of imbalance in the database; (b) the obtained dataset was split into 6 subsets to evaluate the proposed methods (see the Subsets, which are the first six datasets in Table 2); (c) each obtained dataset was reduced by removing some no pain samples prior to pain intensity samples in a time series for each subject to evaluate the proposed methods; these datasets are called Reduced Subsets; see the Reduced Subsets, which were the last six datasets in Table 2.

3.2. Automatic Pain Intensity Monitoring Methods

Random Forest (RF) was used because it is an applicable method regarding classification and regression tasks. RF is parallelizable method, which means that the process can be split into multiple machines to run, and this leads to a faster computation time (faster to train and predict). In contrast, Boosting is a sequential ensemble method, which takes longer to compute. Further, RF is good with high dimensionality data, robust to outliers and non-linear data, good to handle imbalanced data, and it has also low bias and moderate variance. Alongside other studies [14,16,20,21,22], RFc and RFr were trained with 100 trees and a maximum depth of 10 nodes for classification and regression tasks with single modality experiments. The RFc method in [14,16,48] showed good results in predicting pain intensity and no pain from the time windows of samples that were cut out from the continuous recording of the main stimulation phase. LSTM is an effective method for better handling time series prediction compared to other time series methods because it has a memory cell that can maintain information in memory for long periods of time. It is more accurate on datasets using large sequences. LSTM classification architecture comprised a single LSTM layer activated by ReLU followed by a flatten layer, and then one dense layer activated by ReLU. The final output layer had n neurons (n = number of classes). The Softmax was used as the activation function in the output layer, and the Categorical Cross-Entropy (CCE) was the use as the loss function. The configurations of LSTM regression architecture were similar to LSTM classification architecture, except the final output layer with one neuron was activated using a sigmoid function. The pre-processed data (time series data = samples) were inserted into LSTMs one by one in sequence.
After observing the highly imbalanced database and imbalanced proposed datasets (see Table 1 and Table 2), LSTM again was used after increasing the weight of the training samples with more facial responses, called LSTM using sample weighting (LSTM-SW). The sample weighting method was based on duplicating some samples with high scores. The samples with prediction scores higher than 0.3 in training data when using the RFc with FAD modality were determined, and then these samples were replicated once. The duplicates are desirable, as some single images could appear multiple times per epoch because the LSTM model puts more weight on getting these samples (with observable pain reaction) correct and focuses less on samples without an observable pain reaction. The samples after increasing were trained on the LSTM for classification and regression. To ensure comparability of test results, samples were never duplicated in the test data.

3.3. Experiments

Three categories of experiments applied on 11 datasets from the X-ITE Pain Database: Uni-modality experiments using data from single modalities, Bi-modality experiments using data from two modalities, and Multi-modality experiments using data from multiple modalities. RF was used only with single modality (Uni-moality) experiments. Due to RF’s poor performance compared to LSTM and LSTM, we decided to continue two and all fused modalities experiments using LSTM and LSTM. For more details about those methods, see Section 3.2. For reference, a Trivial classifier was calculated within classification experiments, which always votes for the majority class of the dataset (no pain in our experiments). In order to be able to know which modality is best for monitoring continuous pain intensity, the suggested automatic methods were trained with the time series data from each single modality. In these experiments, the time series data (experimental data = FAD, Audio-D, ECG-D, EMG-D, and EDA-D) were used individually to predict pain intensity using RF, LSTM, and LSTM-SW for classification (discrete predictions) and regression (continuous predictions). All LSTM/LSTM-SW classification models were optimized using the loss function CCE, and LSTM/LSTM-SW regression models were optimized using the loss function BCE. The obtained models were trained for 2000 epochs with different lr ( 10 4   or   10 5   or   10 6 ) when setting up the batch size equal to 512 and using adam optimizer. Table 3 shows the architecture of LSTM/LSTM-SW within Uni-modality experiments.
The architectures A(c), B(c), C(c), and D(c) for classification and A(r), B(r) for regression all have input size of 10 × 252/288/36/12, the number of features was variant according to the used modality. The number 10 indicates timesteps (25 Hz time series were reduced to one Hz after applying temporal integration process), 252 indicates facial features (FAD), 288 indicates audio features (Audio-D), 12 indicates ECG features (ECG-D) or EDA features (EDA-D), and 36 indicates features (EMD-D). A(c) and C(c) classification architectures comprised a single LSTM layer with 4 units activated by ReLU followed by a flatten layer, and then one dense layer with 128 neurons activated by ReLU. The final output layer had 7 neurons in A(c) and 4 neurons in C(c). B(c) and D(c) classification architectures comprised a single LSTM layer with 8 units activated by ReLU and followed by flatten layer, and then one dense layer with 64 neurons activated by ReLU. The final dense output layer had 7 neurons in B(c) and 4 neurons in D(c). The Softmax was used as the activation function in the output layer, and the CCE was the use as the loss function. The configurations of A(r) regression architecture were similar to A(c) and C(c), and the configurations of B(r) regression architecture were similar to B(c) and D(c), except the final output layer with one neuron was activated using a sigmoid function.
Two Uni-modality architectures (FAD/EMG-D and EDA-D) were combined using LSTM/LSTM-SW by merging their final dense layers using a concatenate layer (see Table 4). A-Bi(c) and C-Bi(c) classification architectures of (FAD and EDA-D) Bi-modality or (EMG-D and EDA-D) Bi-modality, including A(c) and C(c), comprised a single LSTM layer with 4 units activated by ReLU and followed by a flatten layer, and then one dense layer with 128 neurons activated by ReLU. The output of dense layer (dense1) from X (EDA-D) modality architecture was concatenated with the output of dense layer (dense1) from Y (FAD/EMG-D) modality. The final layer (dense2) had 7 neurons in A-Bi(c) and 4 neurons in C-Bi(c). B-Bi(c) and D-Bi(c) classification architectures, including B(c) and D(c), comprised a single LSTM layer with 8 units activated by ReLU and followed by flatten layer, and then one dense layer with 64 neurons activated by ReLU. The output of dense layer (dense1) from X (EDA-D) modality architecture was concatenated with the output of dense layer (dense1) from Y (FAD/EMG-D) modality. The final layer (dense2) had 7 neurons in B-Bi(c) and 4 neurons in D-Bi(c). The configurations of A-Bi(r) regression architecture were similar to A-Bi(c) and C-Bi(c), and the configurations of B-Bi(r) regression architecture were similar to B-Bi(c) and D-Bi(c), except the output layer with one neuron.
Table 5 shows the overview of the Multi-modality experiments. All single modalities’ architectures using LSTM/LSTM-SW were combined by concatenating the outputs from the dense1 layers using concatenate layer. The Multi-modality experiments were similar to Bi-modality experiments, except the input were the time series data from all modalities (FAD, Audio-D, ECG-D, EMG-D, and EDA-D). A-Mu(c) and C-Mu(c) classification architectures of Multi-modality, including A(c) and C(c), comprised a single LSTM layer with 4 units activated by ReLU and followed by a flatten layer, and then one dense layer with 128 neurons activated by ReLU. The dense2 layer had 7 neurons in A-Mu(c) and 4 neurons in C-Mu(c). B-Mu(c) and D-Mu(c) classification architectures, including B(c) and D(c), comprised a single LSTM layer with 8 units activated by ReLU and followed by flatten layer, and then one dense layer with 64 neurons activated by ReLU. The dense2 layer has 7 neurons in B-Mu(c) and 4 neurons in D-Mu(c). The configurations of A-Mu(r) regression architecture were similar to A-Mu(c) and C-Mu(c), and the configurations of B-Mu(r) regression architecture were similar to B-Mu(c) and D-Mu(c), except for the final dense out-put layer with one neuron.
The learning rate (lr) within A(c), A(r), A-Bi(c), A-Bi(r), A-Mu(c), A-Mu(r), C(c), C(r), C-Bi(c), C-Bi(r), C-Mu(c), and C-Mu(r) using PD, HPD, and EPD is 10 5 ; using TD, HTD, and ETD it is 10 6 ; and using RPD, RHPD, and REPD it is 10 4 . The lr within B(c), B(r), B-Bi(c), B-Bi(r), B-Mu(c), B-Mu(r), D(c), D(r), D-Bi(c), D-Bi(r), D-Mu(c), and D-Mu(r) using RTD and RETD is 10 6 . In the output layers, 7 levels indicate no pain versus heat and electrical pain intensity recognition for one of two different (phasic/tonic), and 4 levels indicate no pain versus heat/electrical pain intensity recognition for one of two different (phasic/tonic).

4. Results

In this section, we introduce the comparison between the proposed methods (RF, LSTM, and LSTM-SW) with the best single modality (EDA), best two fused modalities (EMG and EDA), and fused all modalities (facial expressions, audio, ECG, EMG, and EDA). The experimental data was used (EDA-D, (EMG-D and EDA-D), (FAD, Audio-D, ECG-D, EMG-D, and EDA-D)). Section 4.1 shows the findings when evaluating the test set using regression measures (Mean Squared Error (MSE) and Intraclass Correlation Coefficient (ICC) [61]), whereas Section 4.2 shows the findings in terms of classification measures when classification outperformed the regression in terms of regression measures. The classification measures were Accuracy, Micro average precision (Micro avg. precision), Micro average recall (Micro avg. recall), and Micro average F1-score (Micro avg. F1-score).

4.1. Classification vs. Regression

After comparing between single modalities models when applying baseline methods (RFc and RFr), LSTM, and LSTM-SW with 11 datasets from the experimental data, we found that EDA-D Uni-modality modality models performed the best with most datasets regarding classification and regression. Table 6 and Table 7 show the comparison of the best Uni-modality, the best Bi-modality, and Multi-modality models in terms of MSE and ICC measures, respectively. EMG-D outperformed EDA-D only with the TD dataset and performed similarly to EDA-D with the HTD dataset regarding classification. In line with Uni-modality modality experiments, the best results were obtained when using EMG-D and EDA-D Bi-modality with the same methods and all datasets except the HTD dataset, and FAD and EDA-D Bi-modality were the best.
Regarding regression, both FAD and EDA-D Bi- and Multi-modality models performed similarly when using LSTM with the EPD dataset. The Multi-modality models were the best when using TD, ETD, and RETD; they got MSE of 0.08, 0.06, and 0.09, and the ICC of 0.30, 0.33, and 0.56, respectively. Further, both LSTM and LSTM-SW Uni-modality models outperformed the baseline Uni-modality model (RFc and RFr). Bi- and Multi-modality models outperformed Uni-modality models except with the REPD dataset, and EDA-D Uni-modality model when applying LSTM-SW was the best regarding regression (MSE of 0.03 and ICC of 0.88). The results did not improve after fusing EDA-D as the best Uni-modality with the other Uni-modality, possibly because the outlier rate was increased when using more than one modality. The EMG-D and EDA-D Bi-modality models performed the best with the PD, EPD, RHPD, RPD, and RTD datasets regarding regression. With HTD, Bi-modality models performed the best when applying LSTM-SW (MSE of 0.13 and ICC of 0.42) Classification performed better than regression with almost balanced datasets (HTD). Classification and regression performed similarly when using LSTM-SW and LSTM with HPD; they got MSE of 0.09 and 0.07, and ICC of 0.41 and 0.41, respectively. Figure 2 shows the best models after comparing between Uni-modality modality models (EDA-D modality), the best Bi-modality models (EMG-D and EDA-D Bi-modality), and Multi-modality models. Finally, almost all the Bi- and Multi-modality models performed the best regarding regression in terms of standard deviation (STD) and mean.

4.2. Classification

Classification models got the superior performance to regression models with the HTD dataset and performed similarly with the HPD dataset in terms of MSE and ICC measures. Thus, we applied RFc (EDA Uni-modality), LSTM, and LSTM-SW with HPD and HTD datasets regarding classification to introduce the best performance. Table 8 shows that LSTM and LSTM-SW Uni-modality models with HPD and HTD successfully predicted discrete pain intensity levels in sequences compared to Trivial and RFc in terms of Accuracy, Micro avg. precision, Micro avg. recall, and Micro avg. F1-Score.
The EMG-D and EDA-D Bi-modality and Multi-modality models when applying LSTM-SW with HPD and HTD were better than those models using LSTM in terms of Micro avg. recall and Micro avg. F1-Score. The t-test as inferential statistic used to determine if there is a significant difference between the means of two methods (RF and LSTM/LSTM-SW). Further, EMG-D and EDA-D Bi-modality models when applying LSTM-SW outperformed LSTM models with HTD in terms of Accuracy (49.8%) and Micro avg. precision (48.7%). EDA-D Uni-modality model, when applying LSTM-SW with HTD, performed the best in terms of Micro avg. recall (100%); however, the Bi-modality model when applying LSTM-SW with the same dataset performed excellent (97%). Additionally, the Multi-modality model, when applying LSTM-SW with HPD, performed the best in terms of Micro avg. recall (22.3%). In terms of F1-score, Bi-modality models, when applying LSTM-SW with HPD and HTD, performed the best, at about 26.3% and 63.3%.
After closely investigating the correctly predicted samples from classification results when applying RFc (EDA Uni-modality), LSTM, and LSTM-SW, we found that the models with HPD performed the best in recognizing no pain and the highest pain intensity, whereas models with HTD performed the best in recognizing intermediate pain (low and moderate); see Table 9. The reason is that models may have difficulty in recognizing intermediate pain intensity stimulation in large imbalanced datasets like HPD when HTD is less imbalanced (no pain (about 20%), intermediate pain (about 27%), and the highest pain (about 26%)).
Additionally, Bi- and Multi-modality models improved the performance of Uni-modality models when applying both LSTM and LSTM-SW with the imbalanced HPD. Finally, the EMG-D and EDA-D Bi-modality models, when applying LSTM-SW with HPD and HTD, performed the best based on calculating the average of the model performance using HPD and HTD, which are 37.5% and 48.2%, respectively.

5. Discussion

We conducted several experiments using three methods (RF, LSTM, and LSTM-SW) regarding classification and regression in order to gain insights into monitoring continuous pain intensity. Further, these experiments applied to compare the performance of proposed methods when using single modalities (using RF, LSTM, and LSTM-SW), two fused modalities, and all modalities (using LSTM and LSTM-SW) from the X-ITE Pain Database. The data from each modality were split into 11 datasets to simplify the imbalanced database problem, improve the results, and generalize the capability of the desired system. An additional reason for using these different datasets was to explore the generalization capability of a continuous pain intensity monitoring system. For evaluation, MSE and ICC were used to measure the performance of classification models versus regression models, and the best classification models outperformed the regression models were further evaluated using Accuracy, Micro avg. precision, Micro avg. recall, and Micro avg. F1-score. In this paper, we introduced the best results from single modality (EDA-D Uni-modality) experiments, two fused modalities (EMG-D and EDA-D Bi-modality) experiments, and all modalities (Multi-modality) experiments. The obtained results were alongside to prior works on the X-ITE Pain Database [14,16], the BioVid, and SenseEmotion databases [45,62,63]. EMG-D Uni-modality performed similarly to EDA-D Uni-modality with the almost balanced dataset (HTD); the changes in muscle activity (EMG) to heat tonic stimuli tend to be similar intense as the changes in the superficial muscles of the skin of the hand (EDA).
As we saw in the quantitative results in Section 3, the reduction strategy reduced the influence of outliers (unimportant samples) by reducing some no pain samples prior to pain intensity sequences in a time series for each subject; the Subsets performances improved after applying the reduction strategy (see Reduced Subsets results in Table 6 and Table 7 and Figure 2). It is important to reduce the noise in imbalanced data than to keep hard samples. Further, regression outperforms classification when using huge imbalanced datasets, whereas classification was the best with the almost balanced dataset (HTD); see Figure 2. Both classification and regression perform similarly with Heat Phasic Dataset (HPD) in terms of ICC measure. Trivial failed to recognize pain intensity with HTD and HPD, whereas the proposed methods are significantly better in terms of Micro avg. precision, Micro avg. recall, and Micro avg. F1-score; see Table 8. Accuracy measure does not a proper measure with imbalanced datasets, because it does not distinguish between the numbers of correctly classified examples of different classes. After investigating the HPD and HTD datasets further, we assumed that classification performed well when the distribution of pain intensity samples was high in the imbalanced datasets (the majority of samples experienced no pain), which was 26.7% with HTD, followed by HPD (7.71%) based on the mean value; see Table 9. It shows also LSTM-SW is the best compared to LSTM when fusing two modalities of EMG and EDA (EMG-D and EDA-D Bi-modality) regarding classification with HPD and HTD datasets. Additionally, LSTM-SW increased the performance compared to LSTM of several datasets; see Table 6 and Table 7. The success of LSTM-SW is based on downweighted samples in the training set with a less facial response by using the sample weighting method; these samples might negatively affect the model performance.
Models using the best two and all fused modalities (EMG-D and EDA-D Bi-modality and Multi-modality) improved the continuous monitoring of pain intensity significantly compared to those using the best single modality (EDA-D Uni-modality) in 10 out of 11 datasets. Only the performance of the EDA-D Uni-modality model did not improve when using the REPD dataset, which might be due to the conflicts between modalities in these datasets, especially because of the outliers. In line with [14,16], Multi-modality models outperformed EDA-D Uni-modality models and EMG-D and EDA-D Bi-modality models with imbalanced tonic datasets (TD, ETD, and RETD). The possible reason is that the responses to tonic stimuli are more intense than the response to phasic stimuli for each modality. Further, the response to electrical tonic stimuli tends to start earlier and more rapidly than the response to heat for each modality. Thus, all modalities could significantly benefit from this Multi-modality model when the responses in all modalities are more intense. With EPD, both EMG-D and EDA-D Bi-modality and Multi-modality models performed the same. The EMG-D and EDA-D Bi-modality models outperformed Multi-modality models when using LSTM and LSTM-SW on six datasets: PD, HPD, HTD, RPD, RHPD, and RTD; see the acronyms description in Table 2, probably because the outliers in the worst modalities (audio and ECG). The ECG is sensitive to miscellaneous mixed noises. The audio signal includes many label noises, possibly due to the vocalizations responses when some subjects were stimulated with different pain intensities or when some subjects talk or produce other vocal responses during no pain is experienced. A reliable and trustworthy pain assessment is necessary for adequate pain management, changing analgesic dose, and additional interventions if required. Additionally, good care requires more than one pain intensity measure (self-report, external observations, or physiological tests. Therefore, this work was introduced as a baseline study for future research regarding continuous pain intensity monitoring systems using single or multiple modalities with the X-ITE pain database. Further, we recommend using our data pre-processing strategy and the proposed methods with other pain databases; considering the reliable results in this paper, they will probably perform well.

6. Conclusions

As seen in Figure 3, the proposed system confirms that it is possible to monitor continuous pain intensity using machine learning models with fused modalities (facial expressions, audio, and physiological (ECG, EMG, and EDA)). The difficulties for any model are the huge imbalance in the database and the outliers. The models failed to recognize minority classes and deal with noise when using the database. Due to the distribution of the classes being variants, we split the database into six datasets in terms of four qualities of pain stimuli (phasic (short) and tonic (long) variants of each heat and electrical stimuli). Further, we reduced the noise in the modalities data by removing some no pain samples prior to pain intensity samples in a time series for each subject in each of the six datasets. There are plenty of no pain labels, which are inconsistent with the modalities’ responses. After applying different models with all datasets, we concluded that regression is better than classification with imbalanced datasets, which is the case with real data. Regression reduces the effect of confounding variables by isolating the effect of each variable by allowing for the role of each independent variable to learn without worrying about other variables in the model. Further, LSTM and LSTM-SW successfully recognize pain intensity compared to RF (used with single modalities). We also confirmed that EDA is the best single modality, and EMG and EDA are the best two fused modalities. Two and all fused modalities improved results further compared to a single modality with almost all datasets. The fused modalities that use all modalities were the best with imbalanced tonic datasets.
EMG and EDA fused modality performed the best with imbalanced phasic datasets and balanced dataset (HTD); this result shows that EMG and EDA are very good options for cost-effective pain monitoring; there is no need to use all modalities. However, there is some limitation of using this system: (1) The database does not contain any of a vulnerable group; however, this system can help to predict pain, particularly with vulnerable patients, but it has not yet been implemented for them. (2) Although applying the proposed reduction strategy to the imbalanced datasets helps to improve the performance, there are still plenty of outliers that limit further performance improvement. (3) A larger dataset with more pain intensities is necessary for more reliable automatic monitoring of continuous pain intensity.

Author Contributions

Methodology, software, validation, investigation, and visualization: E.O.; formal analysis, writing—original draft preparation: E.O.; project leading and funding: A.A.-H.; conceptualization, analyzing, writing—review and editing: E.O., P.W., F.S., M.-A.F. and A.A.-H. All authors have read and agreed to the published version of the manuscript.

Funding

This NIT-research is supported by German Academic Exchange Service (DAAD), the Federal Ministry of Education and Research of Germany (BMBF; Robo-Lab no. 03ZZ04X02B) and DFG (vital parameters AL 638/14-1).

Institutional Review Board Statement

The study was conducted according to the guidelines of the Declaration of Helsinki and approved by the Ethics Committee of the Ulm University (protocol code: 372/16, date of approval: 5 January 2017).

Informed Consent Statement

Written informed consent was obtained from the patient(s) to publish this paper.

Data Availability Statement

Data are available from the authors upon request (to Sascha Gruss [email protected] or Steffen Walter [email protected]) for researchers of academic institutes who meet the criteria for access to the confidential data.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

ECGElectrocardiogram
EMGElectromyogram
AUsAction Units
Audio-DAudio Descriptor
two modalitiesBi-modality
DDPdelivered duty paid
ECG-DECG Descriptor
EDAEDA Descriptor
EPDElectrical Phasic Dataset
ETDElectrical Tonic Dataset
EDAElectrodermal Activity
EMGEMG Descriptor
FACSFacial Action Coding System
FADFacial Activity Descriptor
fpsframes per second
HRVHeart Rate Variability
HPDHeat Phasic Dataset
HTDHeat Tonic Dataset
ICCIntraclass Correlation Coefficient
HNRlogarithmic Harmonics to Noise Ratio
LSTMLong-Short Term Memory
LLDlow-level descriptor
LSTM-SWLSTM using a sample weighting
MSEMean Squared Error
MFCCsMel Frequency Cepstral Coefficients
Micro avg. F1-score Micro average F1-score
Micro avg. precision Micro average precision
Micro avg. recallMicro average recall
all modalities Multi-modality
PDPhasic Dataset
RFRandom Forest
RFcRandom Forest classifier
RFrRandom Forest regression
REPDReduced Electrical Phasic Dataset
RETDReduced Electrical Tonic Dataset
RHPDReduced Heat Phasic Dataset
RHTDReduced Heat Tonic Dataset
RPDReduced Phasic Dataset
RTDReduced Tonic Dataset
RMSRoot-Mean Square
SCLSkin Conductance Level
SCRSkin Conductance Response
STDStandard Deviation
SHSSubharmonic-Summation
SVRSupport Vector Regression
TDTonic Dataset
single modalityUni-modality

References

  1. Williams, A.C.C. Facial Expression of Pain: An Evolutionary Account. Behav. Brain Sci. 2002, 25, 439–455. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Kunz, M.; Lautenbacher, S.; LeBlanc, N.; Rainville, P. Are both the sensory and the affective dimensions of pain encoded in the face? Pain 2012, 153, 350–358. [Google Scholar] [CrossRef] [PubMed]
  3. Thiam, P.; Kestler, H.; Schwenker, F. Multimodal Deep Denoising Convolutional Autoencoders for Pain Intensity Classification based on Physiological Signals. In Proceedings of the 9th International Conference on Pattern Recognition Applications and Methods, Valletta, Malta, 22–24 February 2020; pp. 289–296. [Google Scholar]
  4. Gruss, S.; Geiger, M.; Werner, P.; Wilhelm, O.; Traue, H.C.; Al-Hamadi, A.; Walter, S. Multi-Modal Signals for Analyzing Pain Responses to Thermal and Electrical Stimuli. J. Vis. Exp. 2019, 146, e59057. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Ekman, P. Facial Expression and Emotion. Am. Psychol. 1993, 48, 384–392. [Google Scholar] [CrossRef]
  6. Oosterman, J.M.; Zwakhalen, S.; Sampson, E.L.; Kunz, M. The Use of Facial Expressions for Pain Assessment Purposes in Dementia: A Narrative Review. Neurodegener. Dis. Manag. 2016, 6, 119–131. [Google Scholar] [CrossRef]
  7. Corbett, A.; Achterberg, W.; Husebo, B.; Lobbezoo, F.; de Vet, H.; Kunz, M.; Strand, L.; Constantinou, M.; Tudose, C.; Kappesser, J.; et al. An International Road Map to Improve Pain Assessment in People with Impaired Cognition: The Development of the Pain Assessment in Impaired Cognition (PAIC) Meta-tool. BMC Neurol. 2014, 14, 229. [Google Scholar] [CrossRef] [Green Version]
  8. Snoek, K.G.; Timmers, M.; Albertyn, R.; van Dijk, M. Pain Indicators for Persisting Pain in Hospitalized Infants in A South African Setting: An Explorative Study. J. Pain Palliat. Care Pharmacother. 2015, 29, 125–132. [Google Scholar] [CrossRef]
  9. Walter, S.; Gruss, S.; Ehleiter, H.; Tan, J.; Traue, H.C.; Werner, P.; Al-Hamadi, A.; Crawcour, S.; Andrade, A.O.; da Silva, G.M. The BioVid Heat Pain Database: Data for the Advancement and Systematic Validation of an Automated Pain Recognition System. In Proceedings of the IEEE International Conference on Cybernetics (CYBCO), Lausanne, Switzerland, 13–15 June 2013. [Google Scholar]
  10. Pouromran, F.; Radhakrishnan, S.; Kamarthi, S. Exploration of Physiological Sensors, Features, and Machine Learning Models for Pain Intensity Estimation. PLoS ONE 2021, 16, e0254108. [Google Scholar] [CrossRef]
  11. Chu, Y.; Zhao, X.; Han, J.; Su, Y. Physiological Signal-Based Method for Measurement of Pain Intensity. Front. Neurosci. 2017, 11, 279. [Google Scholar] [CrossRef]
  12. Lopez-Martinez, D.; Picard, R. Continuous Pain Intensity Estimation from Autonomic Signals with Recurrent Neural Networks. In Proceedings of the Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Honolulu, HI, USA, 17–21 July 2018; pp. 5624–5627. [Google Scholar]
  13. Thiam, P.; Kessler, V.; Amirian, M.; Bellmann, P.; Layher, G.; Zhang, Y.; Velana, M.; Gruss, S.; Walter, S.; Traue, H.C.; et al. Multi-modal Pain Intensity Recognition based on the SenseEmotion Database. IEEE Trans. Affect. Comput. 2019, 12, 743–760. [Google Scholar] [CrossRef]
  14. Werner, P.; Al-Hamadi, A.; Gruss, S.; Walter, S. Twofold-Multimodal Pain Recognition with the X-ITE Pain Database. In Proceedings of the 8th International Conference on Affective Computing and Intelligent Interaction Workshops and Demos (ACIIW), Cambridge, UK, 3–6 September 2019. [Google Scholar]
  15. Walter, S.; Gruss, S.; Limbrecht-Ecklundt, K.; Traue, H.C. Automatic Pain Quantification using Autonomic Parameters. Psychol. Neurosci. 2014, 7, 363–380. [Google Scholar] [CrossRef] [Green Version]
  16. Walter, S.; Al-Hamadi, A.; Gruss, S.; Frisch, S.; Traue, H.C.; Werner, P. Multimodale Erkennung von Schmerzintensität und-modalität mit maschinellen Lernverfahren. Schmerz 2020, 34, 400–409. [Google Scholar] [CrossRef]
  17. Susam, B.T.; Riek, N.T.; Akcakaya, M.; Xu, X.; de Sa, V.R.; Nezamfar, H.; Diaz, D.; Craig, K.D.; Goodwin, M.S.; Huang, J.S. Automated Pain Assessment in Children using Electrodermal Activity and Video Data Fusion via Machine Learning. IEEE Trans. Biomed. Eng. 2022, 69, 422–431. [Google Scholar] [CrossRef]
  18. Subramaniam, S.D.; Dass, B. Automated Nociceptive Pain Assessment using Physiological Signals and a Hybrid Deep Learning Network. IEEE Sens. J. 2021, 21, 3335–3343. [Google Scholar] [CrossRef]
  19. Odhner, M.; Wegman, D.; Freeland, N.; Steinmetz, A.; Ingersoll, G.L. Assessing Pain Control in Nonverbal Critically ill Adults. Dimens. Crit. Care Nurs. 2003, 22, 260–267. [Google Scholar] [CrossRef]
  20. Othman, E.; Werner, P.; Saxen, F.; Al-Hamadi, A.; Gruss, S.; Walter, S. Classification Networks for Continuous Automatic Pain Intensity Monitoring in Video using Facial Expression on the X-ITE Pain Database. J. Vis. Commun. Image Represent. 2022, 21, 3273. [Google Scholar]
  21. Othman, E.; Werner, P.; Saxen, F.; Al-Hamadi, A.; Walter, S. Regression Networks for Automatic Pain Intensity Recognition in Video using Facial Expression on the X-ITE Pain Database. In Proceedings of the 25th International Conference on Image Processing, Computer Vision & Pattern Recognition (IPCV’21), Las Vegas, NV, USA, 26–29 July 2021. [Google Scholar]
  22. Othman, E.; Werner, P.; Saxen, F.; Al-Hamadi, A.; Gruss, S.; Walter, S. Facial Expression and Electrodermal Activity Analysis for Continuous Pain Intensity Monitoringon the X-ITE Pain Database. IEEE Access 2022. submitted. [Google Scholar]
  23. Craig, K.D. The Facial Expression of Pain Better than a Thousand Words? APS J. 1992, 1, 153–162. [Google Scholar] [CrossRef]
  24. Prkachin, K.M. The Consistency of Facial Expressions of Pain: A Comparison Across Modalities. Pain 1992, 51, 297–306. [Google Scholar] [CrossRef]
  25. Schiavenato, M. Facial Expression and Pain Assessment in the Pediatric Patient: The Primal Face of Pain. J. Spec. Pediatric Nurs. 2008, 13, 89–97. [Google Scholar] [CrossRef]
  26. Tavakolian, M. Efficient Spatiotemporal Representation Learning for Pain Intensity Estimation from Facial Expressions Doctor of Philosophy. Ph.D. Thesis, University of Oulu, Oulu, Finland, 2021. [Google Scholar]
  27. Feldt, K.S. The Checklist of Nonverbal Pain Indicators (CNPI). Pain Manag. Nurs. 2000, 1, 13–21. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  28. Waters, S.J.; Riordan, P.A.; Keefe, F.J.; Lefebvre, J.C. Pain Behavior in Rheumatoid Arthritis Patients: Identification of Pain Behavior Subgroups. J. Pain Symptom. Manag. 2008, 36, 69–78. [Google Scholar] [CrossRef] [PubMed]
  29. Naranjo-Hernández, D.; Reina-Tosina, J.; Roa, L.M. Sensor Technologies to Manage the Physiological Traits of Chronic Pain: A Review. Sensors 2020, 20, 365. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  30. Brown, L. Physiologic Responses to Cutaneous Pain in Neonates. Neonatal Netw. 1987, 6, 18–22. [Google Scholar] [PubMed]
  31. Greisen, J.; Juhl, C.B.; Grøfte, T.; Vilstrup, H.; Jensen, T.S.; Schmitz, O. Acute Pain Induces Insulin Resistance in Humans. Anesthesiology 2001, 95, 578–584. [Google Scholar] [CrossRef] [Green Version]
  32. Stevens, B.J.; Johnston, C.C. Physiological Responses of Premature Infants to a Painful Stimulus. Nurs. Res. 1994, 43, 226–231. [Google Scholar] [CrossRef] [PubMed]
  33. Moscato, S.; Cortelli, P.; Chiari, L. Physiological responses to pain in cancer patients: A systematic review. Comput. Methods Programs Biomed. 2022, 2017, 106682. [Google Scholar] [CrossRef]
  34. Saccò, M.; Meschi, M.; Regolisti, G.; Detrenis, S.; Bianchi, L.; Bertorelli, M.; Pioli, S.; Magnano, A.; Spagnoli, F.; Giuri, P.G.; et al. The Relationship between Blood Pressure and Pain. J. Clin. Hypertens. 2013, 15, 600–605. [Google Scholar] [CrossRef]
  35. Littlewort, G.C.; Bartletta, M.S.; Lee, K. Automatic Coding of Facial Expressions Displayed during Posed and Genuine Pain. Image Vis. Comput. 2009, 27, 1797–1803. [Google Scholar] [CrossRef]
  36. Lucey, P.; Cohn, F.J.; Prkachind, F.K.; Solomon, E.P.; Chewf, S.; Matthews, I. Painful Monitoring: Automatic Pain Monitoring using the UNBC-McMaster Shoulder Pain Expression Archive Database. Image Vis. Comput. 2012, 30, 197–205. [Google Scholar] [CrossRef]
  37. Borsook, D.; Moulton, E.A.; Schmidt, K.F.; Becerra, L.R. Neuroimaging Revolutionizes Therapeutic Approaches to Chronic Pain. Mol. Pain 2007, 3, 25. [Google Scholar] [CrossRef] [Green Version]
  38. Aslaksen, P.M.; Myrbakk, I.N.; Høifødt, R.S.; Flaten, M.A. The Effect of Experimenter Gender on Autonomic and Subjective Responses to Pain Stimuli. Pain 2006, 129, 260–268. [Google Scholar] [CrossRef]
  39. Koenig, J.; Jarczok, M.N.; Ellis, R.J.; Hillecke, T.K.; Thayer, J.F. Heart Rate Variability and Experimentally Induced Pain in Healthy Adults: A systematic Review. Eur. J. Pain 2014, 18, 301–314. [Google Scholar] [CrossRef]
  40. Werner, P.; Al-Hamadi, A.; Niese, R.; Walter, S.; Gruss, S.; Traue, H.C. Automatic Pain Recognition from Video and Biomedical Signals. In Proceedings of the 22nd International Conference on Pattern Recognition, Stockholm, Sweden, 24–28 August 2014. [Google Scholar]
  41. Storm, H. Changes in Skin Conductance as a Tool to Monitor Nociceptive Stimulation and Pain. Curr. Opin. Anaesthesiol. 2008, 12, 796–804. [Google Scholar] [CrossRef]
  42. Ledowski, T.; Bromilow, J.; Paech, M.J.; Storm, H.; Hacking, R.; Schug, S.A. Monitoring of Skin Conductance to Assess Postoperative Pain Intensity. Br. J. Anaesth. 2006, 97, 862–865. [Google Scholar] [CrossRef] [Green Version]
  43. Loggia, M.L.; Juneau, M.; Bushnell, C.M. Autonomic Responses to Heat Pain: Heart Rate, Skin Conductance, and their Relation to Verbal Ratings and Stimulus Intensity . Pain 2011, 152, 592–598. [Google Scholar] [CrossRef]
  44. Werner, P.; Al-Hamadi, A.; Limbrecht-Ecklundt, K.; Walter, S.; Gruss, S.; Traue, H.C. Automatic Pain Assessment with Facial Activity Descriptors. IEEE Trans. Affect. Comput. 2017, 8, 286–299. [Google Scholar] [CrossRef]
  45. Kächele, M.; Thiam, P.; Amirian, M.; Werner, P.; Walter, S.; Schwenker, F.; Palm, G. Multimodal Data Fusion for Person-Independent, Continuous Estimation of Pain Intensity. In Proceedings of the Engineering Applications of Neural Networks: 16th International Conference, Rhodes, Greece, 25–28 September 2015. [Google Scholar]
  46. Erekat, D.; Hammal, Z.; Siddiqui, M.; Dibeklioğlu, H. Enforcing Multilabel Consistency for Automatic Spatio-Temporal Assessment of Shoulder Pain Intensity. In Proceedings of the International Conference on Multimodal Interaction (ICMI’20 Companion), Utrecht, The Netherlands, 25–29 October 2020; pp. 156–164. [Google Scholar]
  47. Breiman, L. Random Forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [Green Version]
  48. Othman, E.; Werner, P.; Saxen, F.; Al-Hamadi, A.; Gruss, S.; Walter, S. Automatic vs. Human Recognition of Pain Intensity from Facial Expression on the X-ITE Pain Database. Sensors 2021, 21, 3273. [Google Scholar] [CrossRef]
  49. Othman, E.; Werner, P.; Saxen, F.; Al-Hamadi, A.; Walter, S. Cross-Database Evaluation of Pain Recognition from Facial Video. In Proceedings of the 11th International Symposium on Image and Signal Processing and Analysis (ISPA), Dubrovnik, Croatia, 23–25 September 2019. [Google Scholar]
  50. Thiam, P.; Kessler, V.; Schwenker, F. Hierarchical Combination of Video Features for Personalised Pain Level Recognition. In Proceedings of the European Symposium on Artificial Neural Networks (ESANN), Bruges, Belgium, 26–28 April 2017; pp. 465–470. [Google Scholar]
  51. Thiam, P.; Schwenker, F. Combining Deep and Hand-Crafted Features for Audio-Based Pain Intensity Classification. In Proceedings of the IAPR Workshop on Multimodal Pattern Recognition of Social Signals in Human-Computer Interaction, Bejing, China, 20 August 2018; pp. 49–58. [Google Scholar]
  52. Tsai, F.-S.; Hsu, Y.-L.; Chen, W.-C.; Weng, Y.-M.; Ng, C.-J.; Lee, C.-C. Toward Development and Evaluation of Pain Level-Rating Scale for Emergency Triage based on Vocal Characteristics and Facial Expressions. In Proceedings of the 17th Annual Conference of the International Speech Communication Association, San Francisco, CA, USA, 8–12 September 2016. [Google Scholar]
  53. Salekin, M.S.; Zamzmi, G.; Hausmann, J.; Goldgof, D.; Kasturi, R.; Kneusel, M.; Ashmeade, T.; Ho, T.; Suna, Y. Multimodal Neonatal Procedural and Postoperative Pain Assessment Dataset. Data Brief 2021, 35, 106796. [Google Scholar] [CrossRef]
  54. Thiam, P.; Hihn, H.; Braun, D.A.; Kestler, H.A.; Schwenker, F. Multi-Modal Pain Intensity Assessment Based on Physiological Signals: A Deep Learning Perspective. Front. Physiol. 2021, 1, 720464. [Google Scholar] [CrossRef]
  55. Hinduja, S.; Canavan, S.; Kaur, G. Multimodal Fusion of Physiological Signals and Facial Action Units for Pain Recognition. In Proceedings of the 15th IEEE International Conference on Automatic Face and Gesture Recognition (FG 2020), Buenos Aires, Argentina, 16–20 November 2020. [Google Scholar]
  56. Hochreiter, S.; Schmidhuber, J. Long Short-Term Memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  57. Baltrušaitis, T.; Robinson, P.; Morency, L.-P. OpenFace: An Open Source Facial Behavior Analysis Toolkit. In Proceedings of the IEEE Winter Conference on Applications of Computer Vision (WACV), Lake Placid, NY, USA, 7–10 March 2016. [Google Scholar]
  58. Eyben, F.; Wöllmer, M.; Schuller, B. OpenSMILE—The Munich Versatile and Fast Open-Source Audio Feature Extractor. In Proceedings of the 18th ACM International Conference on Multimedia (MM ‘10), New York, NY, USA, 25–29 October 2010; pp. 1459–1462. [Google Scholar]
  59. Hamilton, P.S.; Tompkins, W.J. Quantitative Investigation of QRS Detection Rules Using the MIT/BIH Arrhythmia Database. IEEE Trans. Biomed. Eng. 1986, 33, 1157–1165. [Google Scholar] [CrossRef] [PubMed]
  60. Othman, E.; Saxen, F.; Bershadskyy, D.; Werner, P.; Al-Hamadi, A.; Weimann, J. Predicting Group Contribution Behaviour in a Public Goods Game from Face-to-Face Communication. Sensors 2019, 19, 2786. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  61. Shrout, P.E.; Fleiss, J.L. Intraclass Correlations: Uses in Assessing Rater Reliability. Psychol. Bull. 1979, 86, 420–428. [Google Scholar] [CrossRef]
  62. Thiam, P.; Schwenker, F. Multi-modal Data Fusion for Pain Intensity Assessment and Classification. In Proceedings of the 7th International Conference on Image Processing Theory, Tools and Applications (IPTA), Montreal, QC, Canada, 28 November–1 December 2017. [Google Scholar]
  63. Kächele, M.; Amirian, M.; Thiam, P.; Werner, P.; Walter, S.; Palm, G.; Schwenker, F. Adaptive Confidence Learning for The Personalization of Pain Intensity Estimation Systems. Evol. Syst. 2017, 8, 71–83. [Google Scholar] [CrossRef]
Figure 1. The general pipeline of the proposed automatic system for continuous monitoring of pain intensity.
Figure 1. The general pipeline of the proposed automatic system for continuous monitoring of pain intensity.
Sensors 22 04992 g001
Figure 2. The best results of applying Uni-, Bi, and Multi-modality using RF, LSTM, and LSTM-SW regarding classification and regression in terms of MSE and ICC measures. Numbers with the same background color were used to easily highlight the improvement in the performance of each dataset before and after the reduction strategy was applied.
Figure 2. The best results of applying Uni-, Bi, and Multi-modality using RF, LSTM, and LSTM-SW regarding classification and regression in terms of MSE and ICC measures. Numbers with the same background color were used to easily highlight the improvement in the performance of each dataset before and after the reduction strategy was applied.
Sensors 22 04992 g002
Figure 3. A performance example of the proposed automated system for continuous monitoring pain intensity from test set, including results of the best models with RPD (see (a)) and RTD (see (b)) compared to Ground−Truth. RPD: Reduced Phasic Dataset. RTD: Reduced Tonic Dataset.
Figure 3. A performance example of the proposed automated system for continuous monitoring pain intensity from test set, including results of the best models with RPD (see (a)) and RTD (see (b)) compared to Ground−Truth. RPD: Reduced Phasic Dataset. RTD: Reduced Tonic Dataset.
Sensors 22 04992 g003
Table 1. List abbreviations of pain stimuli type, modalities, intensities, and numerical class labels with the percentage samples distribution. This table is also presented in our recent study [20,21,22].
Table 1. List abbreviations of pain stimuli type, modalities, intensities, and numerical class labels with the percentage samples distribution. This table is also presented in our recent study [20,21,22].
TypeModalityIntensities
SevereModerateLowNo Pain (77%)
PhasicHPH3 = 3 (2%)PH2 = 2 (2.1%)PH1 = 1 (2.1%)BL = 0
EPE3 = −3 (2.6%)PE2 = −2 (2.6%)PE1 = −1 (2.6%)
TonicHTH3 = 6 (1%)TH2 = 5 (1%)TH1 = 4 (1%)BL = 0
ETE3 = −6 (1%)TE2 = −5 (1%)TE1 = −4 (1%)
E: Electrical pain stimulus, H: Heat pain stimulus, BL: Baseline; −10 and −11 labels not used in the experiments: −10 (0.5%) and −11(2.5%).
Table 2. List of datasets with samples’ distribution based on labels.
Table 2. List of datasets with samples’ distribution based on labels.
Subsets (Experimental Data)No PainPain Intensities
PDPhasic DatasetExclude tonic samples and no pain samples before these samples and also after samples with −10, −11 labeled.77.7%22.23%
HPDHeat Phasic DatasetExclude electrical samples from PD and no-pain samples before these frames.87.5%21.5%
EPDElectrical Phasic DatasetExclude heat samples from PD and no pain frames before these frames.86.1%13.9%
TDTonic DatasetExclude phasic samples and no pain samples before these samples and also after samples with −10, −11 labeled.70.3%29.7%
HTDHeat Tonic DatasetExclude electrical samples from TD and no pain frames before these frames.20.0%80.0%
ETDElectrical Tonic DatasetExclude heat samples from TD and no pain frames before these frames.82.0%18.0%
Reduced Subsets (Experimental Data)No PainPain Intensities
RPDReduced Phasic DatasetReduce the no pain frames in PD to about 50%.50.0%50.0%
RHPDReduced Heat Phasic DatasetReduce the no pain frames in HPD to about 50%.50.1%49.9%
REPDReduced Electrical Phasic DatasetReduce the no pain frames in EPD to about 50%.50.0%50.0%
RTDReduced Tonic DatasetReduce the no pain frames in TD to about 38%.38.1%61.9%
RETDReduced Electrical Tonic DatasetReduce the no pain frames in ETD to about 49%.49.0%51.0%
Table 3. A summary of the LSTM/LSTM-SW architectures’ configurations using data from Uni-modality. This table is also presented in our recent study [20,21,22].
Table 3. A summary of the LSTM/LSTM-SW architectures’ configurations using data from Uni-modality. This table is also presented in our recent study [20,21,22].
Layer TypeAttributeClassificationRegression
A(c)B(c)C(c)D(c)A(r)B(r)
InputSize:10 × 25210 × 25210 × 25210 × 25210 × 25210 × 252
Timestep:101010101010
Features:252252252252252252
LSTMActivation:ReLUReLUReLUReLUReLUReLU
No. of units:484848
Dropoutwith p:0.50.50.50.50.50.5
FlattenOutput:804080408040
Dense1Activation:ReLUReLUReLUReLUReLUReLU
No. of units:128641286412864
Dense2Activation:SoftmaxSoftmaxSoftmaxSoftmaxLinear/Sigmoid
No. of units:774411
OutputContinuous----
Discrete --
7 levels7 levels4 levels4 levels
Table 4. A summary of the LSTM/LSTM-SW architectures’ configurations using data from FAD/EMG-D and EDA-D Bi-modality. A(c), B(c), C(c), D(c), A(r), and B(r) are LSTM/LSTM-SW architectures in Table 3.
Table 4. A summary of the LSTM/LSTM-SW architectures’ configurations using data from FAD/EMG-D and EDA-D Bi-modality. A(c), B(c), C(c), D(c), A(r), and B(r) are LSTM/LSTM-SW architectures in Table 3.
Layer TypeAttributeArchitectures Configurations (Bi-Modality)
ClassificationRegression
A-Bi(c)B-Bi(c)C-Bi(c)D-Bi(c)A-Bi(r)B-Bi(r)
Concatenate (after dense1)Modality XA(c)B(c)C(c)D(c)A(r)B(r)
+++++++
Modality YA(c)B(c)C(c)D(c)A(r)B(r)
Dense2Activation:SoftmaxSigmoid
No. of units:774411
OutputContinuous----
Discrete--
7 levels7 levels4 levels4 levels
Table 5. A summary of the LSTM/LSTM-SW architectures’ configurations using Multi-modality. A(c), B(c), C(c), D(c), A(r), and B(r) are LSTM/ LSTM-SW architectures using Uni-modality (see Table 3).
Table 5. A summary of the LSTM/LSTM-SW architectures’ configurations using Multi-modality. A(c), B(c), C(c), D(c), A(r), and B(r) are LSTM/ LSTM-SW architectures using Uni-modality (see Table 3).
Layer TypeAttributeArchitectures Configurations (Multi-Modality)
ClassificationRegression
A-Mu(c)B-Mu(c)C-Mu(c)D-Mu(c)A-Mu(r)B-Mu(r)
Concatenate (after dense1)Modality 1A(c)B(c)C(c) D(c) A(r) B(r)
++ +++ ++
Modality 2A(c)B(c)C(c) D(c)A(r)B(r)
++ +++ ++
Modality 3A(c)B(c)C(c)D(c) A(r)B(r)
++ ++ + ++
Modality 4A(c)B(c)C(c)D(c) A(r) B(r)
++ ++ + ++
Modality 5A(c)B(c)C(c)D(c)A(r)B(r)
Dense2Activation: SoftmaxSoftmaxSoftmaxSoftmaxSigmoidSigmoid
No. of units:774411
OutputContinuous----
Discrete--
7 levels7 levels4 levels4 levels
Table 6. Comparison of the best Uni-, Bi-, and Multi-modality models regarding classification and regression tasks with MSE measure. Meas.: Measure. The cells with light grey background indicate the models using LSTM, and cells with pink background indicate the models using LSTM-SW. The bold font indicates the best results.
Table 6. Comparison of the best Uni-, Bi-, and Multi-modality models regarding classification and regression tasks with MSE measure. Meas.: Measure. The cells with light grey background indicate the models using LSTM, and cells with pink background indicate the models using LSTM-SW. The bold font indicates the best results.
Meas. TaskClassificationRegression
DatasetUni-ModalityBi-ModalityMulti-ModalityUni-ModalityBi-ModalityMulti-Modality
MSESubsetsPD0.09
EDA-D
0.08
EMG-D
EDA-D
0.080.06
EDA-D
0.06
EMG-D
EDA-D
0.06
HPD0.10
EDA-D
0.09
EMG-D
EDA-D
0.100.08
EDA-D
0.07
EMG-D
EDA-D
0.09
EPD0.06
EDA-D
0.06
EMG-D
EDA-D
0.050.05
EDA-D
0.04
EMG-D
EDA-D
0.04
TD0.11
EDA-D
0.11
EMG-D
EDA-D
0.110.09
EDA-D
0.10
EMG-D
EDA-D
0.08
HTD0.15
EDA-D (LSTM-SW)
EMG-D (LSTM)
0.13
EMG-D
EDA-D
0.150.11
EDA-D
0.10
EMG-D
EDA-D
0.10
ETD0.11
(RFc)
0.08
EMG-D
EDA-D
0.080.07
EDA-D
0.06
EMG-D
EDA-D
0.06
STD0.030.020.030.020.020.02
Mean0.100.090.100.080.070.07
Reduced SubsetsRPD0.05
EDA-D
(both LSTM)
0.05
EMG-D
EDA-D
0.050.04
EDA-D
0.04
EMG-D
EDA-D
0.04
RHPD0.07
EDA-D
0.07
EMG-D
EDA-D
0.080.05
EDA-D
(both LSTM)
0.05
EMG-D
EDA-D
0.08
REPD0.05
EDA-D
(both LSTM)
0.05
EMG-D
EDA-D
(both LSTM)
0.05
(both LSTM)
0.03
EDA-D
0.04
EMG-D
EDA-D
0.06
RTD0.19
EDA-D
0.19
EMG-D
EDA-D
0.190.11
EDA-D
0.11
EMG-D
EDA-D
0.04
RETD0.16
EDA-D
0.15
EMG-D
EDA-D
0.160.10
EDA-D
0.09
EMG-D
EDA-D
(both LSTM)
0.09
STD0.070.060.070.040.030.02
Mean0.100.100.110.070.070.05
Table 7. Comparison of the best Uni-, Bi-, and Multi-modality models regarding classification and regression tasks with ICC measure. Meas.: Measure. The cells with light grey background indicate the models using LSTM, and cells with pink background indicate the models using LSTM-SW. The bold font indicates the best results.
Table 7. Comparison of the best Uni-, Bi-, and Multi-modality models regarding classification and regression tasks with ICC measure. Meas.: Measure. The cells with light grey background indicate the models using LSTM, and cells with pink background indicate the models using LSTM-SW. The bold font indicates the best results.
Meas.TaskClassificationRegression
DatasetUni-ModalityBi-ModalityMulti-ModalityUni-ModalityBi-ModalityMulti-Modality
ICCSubsetsPD0.40
EDA-D
0.45
EMG-D
EDA-D
0.460.43
EDA-D
0.51
EMG-D
EDA-D
0.49
HPD0.30
EDA-D
0.41
EMG-D
EDA-D
0.390.32
EDA-D
0.41
EMG-D
EDA-D
0.40
EPD0.50
EDA-D
0.53
EMG-D
EDA-D
0.570.53
EDA-D
0.58
EMG-D
EDA-D
0.58
TD0.15
EDA-D
0.18
EMG-D
EDA-D
0.230.17
EDA-D
0.26
EMG-D
EDA-D
0.30
HTD0.33
EDA-D (LSTM-SW)
EMG-D (LSTM)
0.42
EMG-D
EDA-D
0.350.30
EDA-D
0.32
EMG-D
EDA-D
0.38
ETD0.14
(RFc)
0.22
EMG-D
EDA-D
0.260.21
EDA-D
0.31
EMG-D
EDA-D
0.33
STD0.140.140.130.130.130.10
Mean0.300.390.380.330.400.41
Reduced SubsetsRPD0.83
EDA-D
(both LSTM)
0.83
EMG-D
EDA-D
0.820.84
EDA-D
0.85
EMG-D
EDA-D
0.82
RHPD0.76
EDA-D
0.79
EMG-D
EDA-D
0.740.81
EDA-D
(both LSTM)
0.83
EMG-D
EDA-D
0.73
REPD0.84
EDA-D
(both LSTM)
0.85
EMG-D
EDA-D
(both LSTM)
0.81
(both LSTM)
0.88
EDA-D
0.87
EMG-D
EDA-D
0.80
RTD0.31
EDA-D
0.32
EMG-D
EDA-D
0.280.24
(EDA-D)
0.33
EMG-D
EDA-D
0.29
RETD0.47
EDA-D
0.52
EMG-D
EDA-D
0.440.49
EDA-D
0.52
EMG-D
EDA-D
(both LSTM)
0.56
STD0.240.230.240.280.240.22
Mean0.640.660.620.650.750.62
Table 8. Comparison of the best Uni-, Bi-, and Multi-modality models with HTD regarding classification task. The Bi-modality models. * p < 0.05 when using paired t-test between RFc and LSTMs (LSTM and LSTM-SW). The bold font indicates the best results.
Table 8. Comparison of the best Uni-, Bi-, and Multi-modality models with HTD regarding classification task. The Bi-modality models. * p < 0.05 when using paired t-test between RFc and LSTMs (LSTM and LSTM-SW). The bold font indicates the best results.
MeasureDatasetsHPDHTD
ModelTriv.RFcLSTMLSTM-SWTriv.RFcLSTMLSTM-SW
Accuracy %EDA-D
(Uni-modality)
78.578.179.8 *79 *2041.048.447.7
FAD and EDA-D
(Bi-modality)
78.5-80.5 *80.2 *20-47.4 *49.8 *
Multi-modality78.5-79.377.620-41.642.2
Micro avg. precision%EDA-D
(Uni-modality)
024.636.6 *32.2 *042.748.247.7
FAD and EDA-D
(Bi-modality)
0-42.840.1 *0-47.248.7
Multi-modality0-34.9 *29.20-41.842.0
Micro avg. recall%EDA-D
(Uni-modality)
03.49.9 *10.9 *071.094.6 *100 *
FAD and EDA-D
(Bi-modality)
0-16.3 *21.4 *0-92.9 *97 *
Multi-modality0-19.8 *22.3 *0-90.8 *99.9 *
Micro avg. F1-Score%EDA-D
(Uni-modality)
05.915.2 *15.5 *052.962.3 *62.5 *
FAD and EDA-D
(Bi-modality)
0-22.3 *26.3 *0-60.7 *63.3 *
Multi-modality0-23.6 *24 *0-5657.9 *
Table 9. Recall% result of 4-Class continuous pain intensity recognition tasks of HPD and HTD on testing set. Uni-modality refers to EDA-D Uni-modality, Bi-modality refers to EMG and EDA-D Bi-modality. The bold font indicates the best results.
Table 9. Recall% result of 4-Class continuous pain intensity recognition tasks of HPD and HTD on testing set. Uni-modality refers to EDA-D Uni-modality, Bi-modality refers to EMG and EDA-D Bi-modality. The bold font indicates the best results.
ModelDataset
HPDHPD
BLPH1PH2PH3MeanBLTH1TH2TH3Mean
EDA-D
(Uni-modality)
Trivial1000002510000025
RFc98.71.526.327.134.427.34754.540.8
LSTM99.35.25.415.231.39.47335.167.846.3
LSTM-SW983.11.723.331.50.307239.86845
FAD and EDA-D
(Bi-modality)
Trivial1000002510000025
RFc----------
LSTM98.76.55.229.53520.345.456.162.546.1
LSTM-SW97.47.47.237.937.518.645.862.765.548.2
Multi-modalityTrivial1000002510000025
RFc----------
LSTM96.86.24.636.23611.256.328.963.540
LSTM-SW94.15.96.339.536.52.740.455.261.840
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Othman, E.; Werner, P.; Saxen, F.; Fiedler, M.-A.; Al-Hamadi, A. An Automatic System for Continuous Pain Intensity Monitoring Based on Analyzing Data from Uni-, Bi-, and Multi-Modality. Sensors 2022, 22, 4992. https://doi.org/10.3390/s22134992

AMA Style

Othman E, Werner P, Saxen F, Fiedler M-A, Al-Hamadi A. An Automatic System for Continuous Pain Intensity Monitoring Based on Analyzing Data from Uni-, Bi-, and Multi-Modality. Sensors. 2022; 22(13):4992. https://doi.org/10.3390/s22134992

Chicago/Turabian Style

Othman, Ehsan, Philipp Werner, Frerk Saxen, Marc-André Fiedler, and Ayoub Al-Hamadi. 2022. "An Automatic System for Continuous Pain Intensity Monitoring Based on Analyzing Data from Uni-, Bi-, and Multi-Modality" Sensors 22, no. 13: 4992. https://doi.org/10.3390/s22134992

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop