Next Article in Journal
A Novel Hybrid Gradient-Based Optimizer and Grey Wolf Optimizer Feature Selection Method for Human Activity Recognition Using Smartphone Sensors
Next Article in Special Issue
The Problem of Fairness in Synthetic Healthcare Data
Previous Article in Journal
Distributional Replication
Previous Article in Special Issue
Toward Learning Trustworthily from Data Combining Privacy, Fairness, and Explainability: An Application to Face Recognition
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Occlusion-Based Explanations in Deep Recurrent Models for Biomedical Signals

1
Computer Science Department, University of Pisa, 56127 Pisa, Italy
2
KDDLab, Computer Science Department, University of Pisa, 56127 Pisa, Italy
*
Author to whom correspondence should be addressed.
Entropy 2021, 23(8), 1064; https://doi.org/10.3390/e23081064
Submission received: 30 June 2021 / Revised: 4 August 2021 / Accepted: 12 August 2021 / Published: 17 August 2021
(This article belongs to the Special Issue Representation Learning: Theory, Applications and Ethical Issues)

Abstract

:
The biomedical field is characterized by an ever-increasing production of sequential data, which often come in the form of biosignals capturing the time-evolution of physiological processes, such as blood pressure and brain activity. This has motivated a large body of research dealing with the development of machine learning techniques for the predictive analysis of such biosignals. Unfortunately, in high-stakes decision making, such as clinical diagnosis, the opacity of machine learning models becomes a crucial aspect to be addressed in order to increase the trust and adoption of AI technology. In this paper, we propose a model agnostic explanation method, based on occlusion, that enables the learning of the input’s influence on the model predictions. We specifically target problems involving the predictive analysis of time-series data and the models that are typically used to deal with data of such nature, i.e., recurrent neural networks. Our approach is able to provide two different kinds of explanations: one suitable for technical experts, who need to verify the quality and correctness of machine learning models, and one suited to physicians, who need to understand the rationale underlying the prediction to make aware decisions. A wide experimentation on different physiological data demonstrates the effectiveness of our approach both in classification and regression tasks.

1. Introduction

The increasing amount of data generated in each field of human activity, paired with the increasing availability of computing power, has contributed to the success of Machine Learning models. Deep Learning systems, in particular, have gained a lot of traction in the last 10 years thanks to their ability to build an internal representation at different levels of abstraction [1]. This feature, along with the high accuracy exhibited in a variety of different settings, largely contributed to their adoption.
In the biomedical domain, Deep Learning has been applied to a variety of different tasks. One area of active study is related to the processing of one-dimensional physiological signals, with the majority of contributions focusing on classification [2]. Applying machine learning techniques also in a regression setting is of particular interest in this field as it enables new non-invasive monitoring techniques for several physiological signals, such as arterial blood pressure (ABP). Research has been conducted to estimate APB from several other signals, such as Photoplethysmogram (PPG) [3] or Electrocardiogram (ECG) and heart rate [4].
Given their inherently black-box nature, Deep Learning systems pose key challenges in the biomedical field where transparency is a critical feature. To trust a model, a clinician needs to know why such model is generating the predictions he/she is seeing. The same is true for patients who have the right to know the reasons behind a decision or a diagnosis. This need for transparency and interpretability has fostered a research effort targeting the development of models and techniques to gain insight and possibly an understanding of the models’ predictions and their inner workings [5,6,7,8,9]. This large body of research literature, however, is mostly limited to models for static data types, including flat vectorial information or images. On the other hand, a large share of the data produced in the life sciences is of sequential nature, these being time-series of physiological measurements, such as blood pressure, heart rate, electrodermal activity, or genomic/proteomic chunks.
In this paper, we attempt to fill this gap by specifically targeting the explainability within the context of recurrent neural networks for biomedical signals represented as time-varying sequential data. Within this context, we propose a model agnostic technique (based on systematic occlusion study) to gain granular knowledge about input influence on the predictions of the model. We do so while providing a multi-faceted access to interpretability, considering both the point of view of the machine learning practitioner and the life-science expert, providing targeted explanations for the two reference populations. Our approach is especially designed for explaining the black-box regressors, but we also discuss how it can be adapted for explaining the classification of time series. We evaluated our method on three different datasets of physiological signals in both regression and classification tasks. The remaining of the paper is organized as follows. Section 2 discusses related works. Section 3 formalizes the problem faced and introduces basic concepts for the explanation method, which is described in Section 4. Experimental results are presented in Section 5 and  Section 6. Section 7 concludes the paper.

2. Related Works

Interpretability is a multi-faceted problem, and even though it has recently received much attention and different explanation approaches have been proposed [5,6,7,8], a singular shared formalization is still lacking [10]. Explanation methods can be categorized as model-agnostic or model-specific, depending on whether they take into consideration the knowledge of the internal structure of the black box or not.
According to the type of explanations provided by a methodology, we can further differentiate between local and global methods: the former ones generate explanations for specific data instances, while the latter for the logic of the black box as a whole [8].
Some local explanation methods leverage gradient-based methods in order to identify relevant features [11,12,13]. Layer-wise relevance propagation (LRP) [14], instead, makes explicit use of the network activations. The core idea is to find a relevance score for each input dimension starting from the magnitude of the output. The backpropagation procedure implemented by LRP is subject to a conservation property: the relevance score received by a neuron must be redistributed to the lower layers in the same amount. Several different rules were proposed to favour a positive contribution or to generate sparser saliency heatmaps. The Integrated Gradients method [15] combines the sensitivity property of LRP and guarantees the implementation invariance property: if two models are functionally equivalent then the attributions are identical for both. LIME [16] and SHAP [7] are two well-known local methods. The first one generates a simpler interpretable model that approximates the behaviour of the black box in the specific neighbourhood of the instance to be explained. SHAP [7] is a framework that defines a class of additive feature attribution methods and uses a game theoretic approach to assign an importance score to each feature involved in a particular prediction. LRP [14], DeepLIFT [13], and LIME [16] can be considered particular instances of this class of methods.
For models that use attention [17], it is possible to inspect and visualize the learned weights to gain insights on the assigned importance for a given input instance. This approach has been widely applied for model inspection on different types of data and fields, including the biomedical one. RETAIN [18] is an RNN-based model for the analysis of electronic health record (EHR) data. It employs an attention mechanism that allegedly mimics the modus operandi of a clinician: higher weight is given to recent clinical events in the EHR to generate a prediction. The timeline [9] predicts the next category of a medical visit given past EHRs. First, it calculates a low-dimensional embedding of the medical codes of a given EHR; then, a self-attention mechanism generates a context vector. This context vector is then multiplied by a coefficient obtained from a specifically designed function, which takes into account the specific diseases and the time interval. The resulting visit representation vector is the input of a classifier. Given the presence of the multiplier coefficients, it is possible to know how much a specific event contributed to the prediction of the next visit. In [19], the authors show that time steps closer to therapy was associated with higher attention weights and were more influential on the prediction. An adaptation of Class Activation Mapping [15] to 1D time series is described in [20] and applied to Atrial Fibrillation Classification.
Models can also be explained by generating or querying prototypical instances that are representatives of specific output classes. PatchX [21] uses patches to segment the input time series. It extracts local patterns and classifies each of them according to the occurrence of the pattern in a given class. The classification outcome for a complete time series depends on the classes associated with each pattern within it. Other prototype-based approaches leverage the latent representation learned by autoencoders to generate explanations as in [22,23], but in this case, there is a trade-off between prototype quality and classification accuracy.
In [24], the explanations and prototypes are extracted using an information theoretic approach. The authors take the user’s understanding into consideration, which is modelled as a function of the input x of the systems: u ( · ) : R n R : x u : = u ( x ) and can be seen as a summary of that specific input. Similarly, an explanation e : = e ( x ) is a quantity presented to the users to help in the understanding of a specific prediction y ^ . By considering the data points as independent and identically distributed (i.i.d.) realizations of a random variable, the conditional Mutual Information I ( e ; y ^ | u ) represents the amount by which the explanation reduces the uncertainty about the prediction.
Our brief literature survey highlights that most of the interpretability methods are tailored to specific settings and sometimes learning architectures. Model agnostic techniques exist but are applied almost exclusively to classification problems and rarely to regression. Additionally, the availability of approaches for sequential data is substantially lower and limited to classification tasks and, sometimes, to forecasting scenarios [8]. The sequence generation setting is left with few approaches, such as [20], adapted from different tasks that need access to the internals of the models. The method proposed in this paper attempts to overcome such limitations by introducing a model agnostic method that can generate explanations in sequential data processing tasks comprising both regression and classification tasks.

3. Problem Statement

In this paper, we address the problem of explaining the behaviour of a black box model b in the prediction of a time series y given a multivariate time series X = { x 1 , x 2 , , x n } .
A prediction dataset X , Y , thus, consists of a set X = { X 1 , X 2 , , X s } R s × h × n of multivariate time series, where we have a target univariate time series Y R s × h assigned to each multivariate one. A multivariate time series X consists of n univariate time series, each one with h time points x = { t 1 , t 2 , , t h } R h . For instance, a single univariate time series can model an ECG signal. In the following, we also use the term signal to indicate a single univariate time series. We name a local subsection of a signal a sub-signal.
Definition 1
(Sub-signal). Given a signal x R h , a sub-signal x of x with length w < h is a sequence of w contiguous data points of x, i.e., x = { t p , , t p + w 1 } for 1 p h w + 1 .
Given a black box, time series predictor b and a multivariate time series X s.t. b ( X ) = y , our aim is to provide an explanation for the decision b ( X ) = y . We use the notation b ( X ) = Y as a shorthand for { b ( X ) | X X } = Y . We assume that b can be queried at will.

4. The mime Method

We approach the above explanation problem proposing mime (Masking Inputs for Model agnostic local Explanation), a method aiming at understanding why a recurrent neural network outputs a specific prediction and how it reacts to engineered changes in the input signal by using a methodology rooted on occlusions. By occlusion, we denote the alteration of a part of the input signals with a given value. This kind of technique has been applied to analyse the robustness of image classifiers, where important features of the image are masked to observe changes in the predicted class [25].
mime produces an explanation targeted at two different types of users: physicians and technical experts. Physicians receive information about the importance of a particular input signal for the final prediction and information about some particular parts of the input signals influencing the prediction. This information is supported by visualizations. Technical experts instead can use mime to analyse the robustness of the prediction model against some input perturbation.
The different components of our explanation are obtained by using the occlusion mechanism. The occlusion approach proposed in this work does not require prior knowledge concerning the data structure and distribution, and it only requires having access to input signals and model predictions. For each of the sequential input time series of the model, we generate an occluded version by substituting the original signal values with a user-defined value. The alteration can be chosen to last for the whole signal or for a fixed time-span. In the latter case, a windowed approach is employed to systematically analyse the effect of occluding different parts of each input signal.
In the following (Figure 1), we provide a step-by-step description of the proposed methodology, which includes: (i) The determination of the importance of each input signal; (ii) Analysis of the impact of the input signals perturbation; (iii) The extraction of the most influential sub-signals.

4.1. Occlusion Approach

Let X R s × h × n be a tensor representing samples of multivariate time series composed of n signals of the length h. Each signal x is represented by a vector x R h . We use 1 and 0 to denote vectors whose components are, respectively, all ones and all zeroes. The altered signal x ^ is obtained according to the type of modification required. In the case of a full length occlusion, we have: x ^ = o v 1 with o v being the occluding value and 1 R h .
In order to modify x with a localized alteration of duration d starting after p timesteps, we define two binary masking vectors m 1 and m 2 as:
m 1 = ( 1 a , 0 , 1 b ) m 2 = ¬ m 1 1 a R p , 0 R d , 1 b R h ( p + d )
where ¬ is bit-wise negation. By means of the above masks, we get the x ^ vector as:
x ^ = ( x m 1 ) + o v m 2
The localized alteration provides the basic elements to define an occlusion approach based on a window w covering a specific temporal range.
Given a multivariate time series X, we define an occlusion window w with a duration of d timesteps, and we derive the number of possible segments of a signal that we can occlude, i.e., q = h d + c , where c = 0 if the signal duration is divisible by d; otherwise, c = 1 .
Signal occlusion is performed on each segment i with i [ 1 q ] . For each x, we alter only a single segment per time. The alteration can be performed on any of the signals x j X with j [ 1 , n ] , one at time or by considering any subset of signals in X. Algorithm 1 reports the occlusion procedure for a single signal.
By generating the occlusions, we collect the model outputs for both the unaltered input samples Y = b ( X ) and under the occluded inputs X ^ , i.e., Y o c c = b ( X ^ ) . Then, we consider the discrepancies between the two output signals measured in terms of mean absolute error (MAE) between Y and Y o c c . Thus, higher values of MAE denote higher importance of the occluded signal parts. This approach allows us to investigate several aspects of the models trained for different tasks in the biomedical domain and to extract and analyse explanations. We discuss these aspects in the following sections.
Algorithm 1 Occlusion.
1:
procedureocclude( x , w s i z e , w i d x )
2:
     l e n length(x)
3:
     x o c c copy(x)
4:
     o v 0         ▹ user-defined occlusion value.
5:
     s t a r t w s i z e · w i d x
6:
     e n d s t a r t + w s i z e
7:
    if  e n d > l e n  then
8:
         e n d l e n
9:
    end if
10:
    for  i s t a r t , e n d  do
11:
         x o c c [ i ] o v         ▹ window occlusion
12:
    end for
13:
    return  x o c c
14:
end procedure

4.2. Input Signal Importance

The first step of mime aims at determining the importance of each input signal for the prediction task. A large number of approaches have been developed to investigate feature importance in machine learning models for interpretability purposes. Most of them are specifically designed to deal with classification tasks, while others (such as SHAP [7]) rely on assumptions that are not always valid, such as the independence of the input features. As an example, in our setting, two input signals, such as cardiac and respiratory data, cannot be considered independent.
In our approach, for each input signal x X , we evaluate the importance of x by applying the black box b on both the data with the entire signal x occluded and the original data without any occlusion. The MAE resulting from the comparison of the two predictions quantifies the importance of the signal x. Occluding the entire signal means considering a window with a size equal to the signal length, i.e., setting w s i z e = h and w i d x = 1 in Algorithm 1.

4.3. Estimating Duration of Induced Perturbation

Occluding parts of the input signals results in an alteration in the network outputs. The predicted signals under input occlusion manifest a perturbation that, as the empirical analysis will show, is clearly visible when plotting the two generated outputs. Following up on this intuition, we developed a procedure to quantify the duration of the induced alteration.
The rationale of our duration estimation procedure follows the approach discussed previously for the signal importance assessment. For any segment occluded in the input signals, we quantify the deviation of the occluded prediction from the unaltered one by computing their MAE over a window of d timesteps. In particular, given the two predicted signals y and y o c c , we apply the procedure described in Algorithm 2. First, we segment the two signals in q = h d + c sub-signals (with c = 0 if h is divisible by d, c = 1 otherwise), obtaining two lists of sub-signals s and s o c c , respectively, (lines 4–5, Algorithm 2). Then, we compute the MAE for any pair of aligned sub-signals, i.e., i [ 1 v ] . M A E ( s i , s o c c i ) (lines 6–9). Perturbation duration is quantified by counting the number of sub-signals for which the MAE is above a threshold T M A E (lines 10–15), whose value is application-dependent.
Algorithm 2 Perturbation duration.
1:
procedurePertDuration( y , y o c c , T M A E )
2:
     w s i z e d         ▹ user-defined size
3:
     m a e l empty list
4:
     s Segment( y , w s i z e )
5:
     s o c c Segment( y o c c , w s i z e )
6:
    for all  s i s  do
7:
         ϵ mae( s i , s o c c i )
8:
        Append ϵ to m a e l
9:
    end for
10:
     w c 0
11:
    for all  ϵ m a e l  do
12:
        if  ϵ > T M A E  then
13:
            w c w c + 1
14:
        end if
15:
    end for
16:
    return  w c         ▹ n. windows with MAE > T M A E
17:
end procedure

4.4. Determining Influential Sub-Signals

The windowed occlusion procedure can also serve to identify the most relevant or influential input sub-signals for the model. This is, again, obtained by contrasting original predictions with the model outputs under occlusion, measuring the mean discrepancy between the two. Algorithm 3 describes the details of our approach. In particular, it computes, for each input signal x X , the importance of each sub-signal of x. To this end, the input signal x is segmented in q sub-signals s 1 , , s q (line 4), and for each s i , an occluded version of the signal x is computed (line 6). Then, the importance of the sub-signal s i is measured by computing the derived MAE comparing the model prediction y on the unaltered signal and y o c c on the occluded signal (lines 8–10). Once the MAE is computed for each sub-signal, the algorithm produces a heatmap that provides a visual inspection that highlights the importance (measured by MAE) of each sub-signal (see Figure 3 as an example). Finally, the method extracts the top-k sub-signals with the highest MAE.
Next, the top-k sub-signals of each signal are used to provide the physicians with a set of important sub-signals of each category of the input signal. To this end, given the whole set of multivariate time series X , mime selects from each multivariate X X the single univariate signals x j and extracts the top-k sub-signals with the highest MAE, which we denote by T K j X (Algorithm 3).
Finally, mime derives the set I by computing the union of these top sub-signals obtained for each of the j-th signals, i.e., I = X X T K j X . Finally, it extracts the most important ones from such set, again relying on the MAE values.
Algorithm 3 Top-K influential Sub-signals.
 1:
procedureTopkSub-Signals( X , x , m o d e l , w s i z e , k )
 2:
     m a e _ s i g n a l empty list
 3:
     s u b s i g n a l s empty list
 4:
     s Segment( x , w s i z e )
 5:
    for  i 1 , | s |  do
 6:
         x o c c occlude( x , w s i z e , i )
 7:
         X o c c ( X { x } ) { x o c c }
 8:
         y o c c predict( X o c c , m o d e l )
 9:
         y predict( X , m o d e l )
10:
         ϵ mae( y , y o c c )
11:
        Append ( ϵ , s i ) to m a e _ s i g n a l
12:
    end for
13:
     m a e _ s i g n a l ReverseSort( m a e _ s i g n a l )
14:
    for  j 1 , k  do
15:
        Append m a e _ s i g n a l . g e t ( i ) [ 1 ] to s u b s i g n a l s
16:
    end for
17:
    return  s u b s i g n a l s
18:
end procedure

4.5. Self Organizing Maps Clustering of Influential Sub-Signals

The set I of influential sub-signals, extracted using the procedure described in the previous section, is then used as input for a Self Organizing Map (SOM) [26]. SOMs are the most popular family of neural-based approaches to topographic mapping. They leverage soft-competition among neighbouring neurons arranged on low-dimensional lattices to enforce the principle of topographic organization. Soft-competition ensures that nearby neurons respond to similar inputs, while lattice organization provides a straightforward means to visualize high-dimensional data onto simple topographic structures. Thanks to these characteristics, they have found wide application as an effective computational methodology for adaptive data exploration [27].
In this work, SOMs are used as a visualization tool targeted to domain experts. Thanks to the SOM ability to cluster signals by their morphological similarity and mapping them to a specific neuron, or more generally, to a neighbourhood of neurons, it is possible to obtain a synthetic and organized view of those signals. Exploiting the ability to project auxiliary information such as the MAE linked to each sub-signal in I, it is possible to identify prototypical portions of signals associated with the highest error. This process allows us to provide physicians with an intuitive tool to identify and visualize the “critical” parts of the signals. For the sake of our analysis, we can use all sub-signals in the original set I, or alternatively, we can operate on a subset G, obtained by selecting the most informative n (i.e., the ones with the highest error) elements from I.
After the training phase, we query the SOM to obtain the best matching unit (BMU) s i I and link the BMU to the MAE associated with s i . The set of all sub-signals mapped to the BMU with coordinates ( u , v ) is denoted by S u , v = [ s 1 , s 2 , , s z ] . We build a matrix E R u × v with the same dimensions of the map where each element E [ u , v ] is:
E [ u , v ] = z = 1 | S u , v | M A E ( s z ) | S u , v | s z S u , v
By projecting the matrix E on top of the original SOM map, we can easily identify neurons that react to sub-signals with a larger MAE thanks to colour intensity. Sub-signals associated with each BMU can be plotted in isolation or can be linked back to the original input signals they were extracted from, highlighting critical portions of the original time series.

4.6. Explaining Time Series Classification

As described above, mime is designed to explain regression tasks. However, it can be easily adapted for providing explanations in time series classification tasks. In this case, each multivariate time series is assigned to a label, i.e., the target Y R s . In order to adapt our approach to these tasks, we propose determining the signal influence (Section 4.2) and the most influential sub-signals (Section 4.4) by computing the MAE discrepancy between the model losses for the occluded and original signals rather than between the model outputs. Moreover, when selecting the influential sub-signals, we will look into those that lead the model to change its classification prediction. That is why we adapt the approach to return the sub-signals that have the highest MAE and y o c c y . Clearly, since the prediction is a class label here and there is no temporal information associated with the target, we cannot provide the analysis on the impact of the perturbation in terms of duration of the induced alteration.
All in all, the approach needs to be customized based on whether the predictive task is a regression or a classification problem. In a regression setting, the only actionable choice is the selection of the discrepancy function. For the sake of this work, we measure occluded-unoccluded output discrepancy using MAE. For classification problems, in Algorithm 3, we need to compute the MAE discrepancy between the model losses for the occluded and original inputs rather than between the model outputs. Moreover, when selecting the influential sub-signals, we are interested in those that cause the system to change its classification prediction. For this reason, we append the tuple ( ϵ , s i ) to the list of the candidate-important sub-signals (line 11) if and only if y o c c y .

5. Experimental Setup

We tested the approach on both classification and regression tasks using several models trained on three different datasets of physiological signals. In this section, we detail the dataset employed and the models used in the experimental assessment.

5.1. Datasets

The first set of signals is from the Cuff-Less Blood Pressure Estimation Data Set (CBPEDS) [28] available in the UCI ML repository [29]. CBPEDS contains a subset of the physiological signals available in MIMIC II Waveform Database [30] that are useful to create systems for non-invasive blood pressure estimation. MIMIC II is part of PhysioBank [31]. Three different types of synchronized patients recordings are available: electrocardiograms (ECG), photoplethysmograph from fingertip (PPG) and invasive arterial blood pressure (ABP).
The second dataset is the Combined measurement of ECG, Breathing and Seismocardiograms Database [32] (CEBSDB), which was constructed to compare RR time series of ECG and seismocardiograms (SCG). Signals were collected by asking 20 presumed healthy volunteers to be very still in a supine position on a comfortable conventional single bed and awake. The subjects were monitored in a basal state for 5 min, for 50 min while listening to classical music, and for another 5 min after the music ended. From this dataset, we used all the available recordings with exception of “ECG lead I”.
To test the approach on a classification task, we used a dataset obtained from the PTB Diagnostic ECG Database (PTBDB) [33]. A set of ECG beats were extracted from the original 549 full-length recordings. The nine diagnostic classes (eight for unhealthy heart conditions, one for healthy) in the original dataset were condensed into two classes: one for healthy beats and the other for pathologic conditions. We remand to [34] for details regarding preprocessing and beat extraction.
A summary of the main characteristics of these datasets is available in Table 1. Details on datasets preprocessing are reported in Appendix A.

5.2. Models

A total of 9 different models were trained, 3 for each dataset. Given the temporal nature of the physiological signals under analysis, Recurrent Neural Networks models were used. We trained 2 RNN models together with a third non-recurrent one to be used as a baseline competitor. Models were implemented using Keras [35] with Tensorflow 2.0 [36] backend.
Using signals from CBPEDS, we trained the models for the task of estimating the full-length ABP signal using ECG and PPG signals as inputs. On this regression setting, we selected the following models:
  • a convolutional autoencoder (AUT) [37] composed of a total of 26 layers: 15 for the encoder and 10 for the decoder;
  • a Gated Recurrent Units network (GRU) [38] composed of 5 layers, with a single output;
  • a convolutional GRU (CNN-GRU) [39] network of 5 layers and a single output.
A similar regression task was designed with signals from CEBSDB. With the ECG and Breathing signals as input, we predict the whole SCG signal. Given the similarity of the two regression tasks, the six models share most of the architectural choices. Some hyperparameters were tuned to adapt the models to the specific task (details in Appendix B).
We also trained 3 additional models in a binary classification setting using the ECG signals from the PTBDB dataset:
  • a fully connected feed forward neural network (MLP) [40] composed of 5 layers;
  • a Gated Recurrent Units network (GRU) [38] composed of 4 layers, with a single output;
  • a convolutional GRU (CNN-GRU) [39] network of 4 layers and a single output.
Differently from the regression setting, in this case, we used a fully connected network (MLP) as a baseline. This choice is motivated by the fact that it exhibited predictive performances comparable with those of the recurrent models. For all models, the dataset was split into 3 parts: 70% of the data has been used for the training, 10% for validation and 20% for the test set. Networks trained on CEBSDB and CBPEDS used the Mean Absolute Error (MAE) as the loss function, while Binary Cross Entropy was used for models trained on PTBDB.
In the following, we denote models trained on each dataset with the subscripts α , β and δ , for the CBPEDS, CEBSDB and PTBDB, respectively. Table 2 summarizes model performances in the unoccluded case.

6. Experiments

In the following sections, we describe the results of the experiments performed using the mime explainer. First, we report results for signal importance assessment using both whole length occlusion and the windowed approach. Next, we describe the analysis pertaining to the duration of induced perturbations. Following, we detail experiments to extract the most influential sub-signals and the associated SOM-based visualizations. Lastly, we provide examples of the Signal Occlusion Contribution Visualization targeting the clinical experts.

6.1. Signal Importance

Experiments to quantify signal importance for models trained on regression tasks (CBPEDS and CEBSDB) were performed by occluding segments of the input signal with zero values or with the mean value of the dataset for the whole duration. The effects have been evaluated on the validation and test sets from both datasets.
Table 3 reports the results for models trained on the CBPEDS dataset. We include the MAE with the unaltered input as a reference. Different models with different inductive biases learn different representations, and in doing so, they assign different levels of importance to the input signals. The table highlights (in boldtype) that the GRU α model relies more on the PPG signal, as occluding it results in a larger MAE. We have similar results for the AUT α model, while the CNN - GRU α model, instead, has a larger MAE when the ECG signal is occluded. The type of occlusion seems to play a secondary role, probably related to samples distribution, as results on the validation set indicates. The most important input signals remain the same for all three models, with MAE score variations according to the occlusion type.
For the CEBSDB dataset, the signal importance assessment in Table 4 reveals a strong reliance of all the three networks on the ECG input signal to correctly generate the SCG output signal. This behaviour is evident when analysing the errors in Table 4: the MAE associated with ECG occlusion is always higher, with the only exception of the GRU β model on the test set. The occlusion value has a strong impact on the autoencoder model, while the effects are of smaller magnitude than for the other models. Figure 2 shows a graphical example of the different outputs of the GRU β model (trained to predict the SCG ) when different input signals are occluded.

6.2. Windowed Occlusion

In this section, we report the results obtained by occluding the input signals with zero values for a fixed window of time for all window indexes. This approach has been applied in both classification and regression models. In the former case, we report the average MAE error obtained across all the windows, and in the latter case, the mean accuracy obtained by considering the occluded prediction.
Table 5 shows the results on CBPEDS datasets. In general, larger mean MAE values are associated with the occlusion of the most meaningful sub-signals, and the error increases with the window size. In predicting the arterial blood pressure, the GRU α model exhibits larger errors when ECG is occluded. The autoencoder is the worst performer of the three models when the PPG signal is occluded, while the CNN - GRU α model is the most robust among the tested networks.
Table 6 reports results on the CEBSDB dataset. In this regression task, the AUT β model is the most susceptible model when the ECG signal is occluded, while the CNN - GRU β , as in the ABP estimation task, is less influenced by the occlusion. The GRU β model confirms its larger reliance on the breathing signal compared to the other networks, as the associated MAE shows.
The accuracy results obtained with the three different models for the classification task on PTBDB are reported in Table 7. Here, the ECG is the only input signal, and we experimented with different occlusion values. Several window sizes were tested with duration of 25, 50, 75, 100 and 125 time steps. The choice of the occlusion value (zero or mean signal value on the dataset) has a negligible impact on the accuracy (from 1% to 4%). Interestingly, all models worsen their prediction when the occlusion is zero, especially at lower window sizes. Increasing the occlusion duration results in a larger accuracy loss for all models, independently from the value used. The feedforward network used as a baseline is the less susceptible model followed by the CNN - GRU δ model. The pure GRU model has, in general, the largest accuracy loss.

6.3. Induced Perturbation Duration

Table 8 provides the results for the experiments quantifying the duration of the perturbation caused by different occlusion types in CBPEDS. The GRU α model shows the largest sensibility to the alteration of ECG input and takes more timesteps to undo the induced error, which can last up to 250 timesteps (2 s) even for small occlusion durations, confirming the importance of this signal for this specific model. CNN - GRU α seems to recover faster than the GRU α model. Moreover, the duration of the perturbation is similar for ECG and PPG occlusions. The best model at dealing with the perturbation duration is AUT α . Its mean duration is the lowest in the Table, and when ECG is occluded, its effect lasts for zero timesteps. This does not mean, however, that the induced perturbation is zero: it rather indicates that the induced error is less than the chosen tolerance for the MAE.
The results for the CEBSDB task are reported in Table 9: in this setting, the autoencoder needs a larger time to recover from induced perturbation. Occluding the breathing signal causes no perturbation for both AUT β and the CNN - GRU β , while the effect is low for the GRU β model. Compared with the ABP estimation task, perturbation durations are, in general, lower, with the exception of the autoencoder model. This effect may be due to the nature of the predictive task: the SCG signal has higher variability than ABP, which is probably causing models to recover faster from alterations.

6.4. Visualizing Sub-Signal Occlusion Effects

The increasing availability of medical datasets motivates the need for tools to make sense of this large amount of information [41]; one of the fastest and most effective ways to convey key aspects of data under analysis is by visualization. The proposed explanation is targeted at experts in the medical domain. By expert in the medical domain, we mean a clinician or doctor, that is, a person who has no professional computer science background but rather a medical one.
We get our visualization by overlapping two different kinds of plots. The first one is the plot of the input signal we are considering, which in the case of CBPEDS, is either an ECG curve or a PPG curve. The second one is a windowed heatmap used as a background for the first plot. The heatmap is generated by occluding the signal under analysis for a specific user-defined window of time, with the approach described in Section 4.2.
For each occluded window index, we plot the associated MAE error with a proportionally intense background colour. Figure 3 shows an example of our visualization of the occlusion contribution for an ECG signal from the CBPEDS dataset with a window occlusion size of 50 timesteps. For the ECG signal analysed, it is clear that an occlusion in the first window of the signal results in a higher error. Moreover, the section of the signal around the 800th time step (indicated with a red triangle) is also associated with a high MAE. By observing this visualization, clinicians can get an insight into which portion of the input signals are influential for the output prediction of the model and assess whether the highlighted sub-signals are critical morphological features employed for classical diagnosis methods. Another example of such a visualization for a different window size is reported in Appendix C.
In order to assess the interpretation provided by mime, we compared it with the Integrated Gradients (IG) method [15]. Figure 4 compares the most influential sub-signals of an ECG signal identified by mime and IG in PTB dataset. The comparison points out how both methods are concordant in identifying the same important window as the key subsequence in the analysed signal.
We also provide a quantitative evaluation of the concordance between the mime and IG interpretations. To this end, we compute a score measuring when the two methods select the same sub-signal or sub-signals that are temporally close as the most influential ones. Given that, mime returns an importance score for each window of duration d over the signal x (as explained in Section 4.6). We define an importance score, based on Integrated Gradients, to compare our results with the IG method. In particular, given a window of duration d, we calculate this score as the sum of the IG values I G j of each timestep j, i.e., I G s c o r e = j = 1 d I G j . We assign an index to each window; thus, we can derive from each signal which window index corresponds to the highest importance score in both methods. We name them i n d e x I G and i n d e x M I M E . Then, we compute how many windows identified by mime and IG perfectly match or differ by no more than 1 window index, i.e., | i n d e x I G i n d e x M I M E | < = 1 . A preliminary investigation conducted on MLP δ found that mime and IG have a concordance score of 68.20% for the signals in PTB dataset. We leave a more in-depth quantitative characterization of the relationships between the two approaches as future work.

6.5. Most Influential Sub-Signals

In this section, we describe the SOM-based analysis performed on the most influential samples extracted from the various datasets and according to the different models. The maps were trained using the MiniSOM python library [42]. For each recurrent model, we trained several SOMs using the top 5000 sub-signals extracted from the corresponding training dataset as input. All maps have dimensions ( 12 , 17 ) . We used a Gaussian neighbourhood function with σ = 2.05 and hexagonal topology. SOMs were trained with a learning rate l r = 0.7 for a total of 10 5 steps.
After the training phase, we have tested the SOM with the top 2000 sub-signals extracted from the test portion of each dataset to build the E matrix with E R 12 × 17 . We normalize E to have values in the [ 0 , 1 ] range and project this information on the SOM as a heatmap. Figure 5 and Figure 6 show two examples of visualizations obtained from the SOMs trained on ECG signals. In the figures, we report also a close-up on the prototypical signals associated with the most active neuron in the map. Signals associated with the highlighted neurons show large MAE errors and share morphological characteristics.
Self Organizing Maps obtained from the training dataset can be shared with users of the predictor to help them in assessing the behaviour of the model on novel data. By repeating the most influential sub-signal extraction phase on a production dataset, the SOM can be used to generate an updated visualization. Such visualization will provide a useful global overview of problematic sub-signals of new time series data.

7. Conclusions

In this work, we presented an interpretability approach for sequential data based on input occlusion. The approach is model-agnostic and only requires access to model inputs and outputs. Using the proposed methodology, we studied several recurrent neural networks trained on both regression and classification tasks and analysed the importance assigned by the models to each input signal.
Our results highlight how different models rely on different input signals to generate their predictions and show larger errors when that input is occluded. The perturbation induced by occlusion lasts longer when the occluded input are those resulting from the signal importance analysis. In regression tasks, recurrent models are more robust compared to the convolutional autoencoder baselines, with the CNN-GRUs suffering less from input alteration compared to the pure GRU models. The increased robustness is probably due to the convolutional layer providing “look-ahead” capabilities to the recurrent layer.
The simple feedforward network used as a baseline in the classification task is more robust with respect to the two recurrent models. As in the regression setting, the CNN-GRU performed better compared to the vanilla GRU and exhibited a minor loss of classification accuracy.
Moreover, leveraging the occlusion approach, we designed two different visualizations aimed at clinicians. The first one gives a detailed view of the error associated with the occlusion of portions of a single input signal.
The second one is based on Self Organizing Maps and is used to visually inspect and discover critical sub-signals associated with high prediction errors.
Interesting future work directions are the development of a data-driven algorithm to select the optimal occlusion window size and increasing the human–machine interaction degree. The latter would allow the proposed approach to be used in “what if?" scenarios, enabling faster comparisons of explanations generated from user-specified parameters.

Author Contributions

Conceptualization, M.R., A.M. and D.B.; methodology, M.R., A.M. and D.B; software, M.R.; formal analysis, M.R.; investigation, M.R.; writing—original draft preparation, M.R.; writing—review and editing, M.R., A.M. and D.B.; visualization, M.R.; supervision, A.M. and D.B.; project administration, A.M. and D.B. All authors have read and agreed to the published version of the manuscript.

Funding

This work is partially supported by the University of Pisa, under the PRA fund on “Emerging Trends in Data Science" (grant n. PRA_2018_43) and the European Community H2020 programme under the funding schemes: G.A. 952215 TAILOR, INFRAIA-1-2014-2015 Res. Infr. G.A. 871042 SoBigData++, G.A. 952026 HumanE AI Net, G.A. 834756 XAI.

Data Availability Statement

Not Applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Dataset Preprocessing

Appendix A.1. CBPEDS

The Cuff-Less Blood Pressure Estimation Data Set is available from the UCI ML repository. Each entry in the dataset is a multivariate time series containing the following signals:
  • ECG signal: electrocardiogram from channel II;
  • PPG signal: photoplethysmograph from fingertip;
  • ABP signal: invasive arterial blood pressure.
Each signal is collected at a sampling frequency of 125 Hz. All three signals are synchronized for each patient.
Preprocessing steps:
  • Smoothing all signals with a simple averaging filter;
  • Removing signal blocks with irregular and unacceptable human blood pressure values;
  • Removing signal blocks with severe discontinuities, which was not resolved by smoothing filter in step 1;
  • The calculation of PPG signal autocorrelation, which indicates the degree of similarity between successive pulses, and removing blocks with high alteration.
  • Removing samples with irregular readings appearing in at least one of the three signals;
  • Signals duration normalization. Each signal was divided into segments of 10 s.

Appendix A.2. CEBSDB

The The combined measurement of ECG, Breathing and Seismocardiograms Database is available from Physiobank.
Each entry in the dataset is a multivariate time series containing the following signals:
  • ECG signal: electrocardiogram from channel I;
  • ECG signal: electrocardiogram from channel II;
  • Breathing signal: breathing signal from thoracic piezoresistive band;
  • SCG signal: seismocardiogram obtained using a triaxial accelerometer.
Each channel was sampled at 5 kHz. All four signals are synchronized for each patient.
Preprocessing steps:
  • removing ECG obtained from channel I from all samples to maintain consistency with the CBPEDS dataset;
  • downsampling all signals to have a duration of 1250 time steps.

Appendix B. Trained Models

Appendix B.1. Models for CBPEDS

Adam with a learning rate of 3 × 10 3 was used as the optimizer, and the batch size was set to 512 for all architectures. The best models were selected according to the best validation score of three training runs.

Appendix B.1.1. GRU α Model

The network is composed of five layers:
  • Input: all the 1250 timesteps of both PPG and ECG are passed as input to this layer;
  • Dense: a fully connected layer of 256 neurons with linear activation functions;
  • Batch-Norm: the batch normalization layer. Keras default parameters were used;
  • Recurrent layer: the recurrent part of the network is based on a GRU of 256 neurons with hyperbolic tangent activations and sigmoid as recurrent activation functions;
  • Output: the fully-connected layer composed of a single neuron with a linear activation function.

Appendix B.1.2. CNN-GRU α Model

The network involves five layers:
  • Input: ECG and PPG signals are passed in their full length as in the GRU α case;
  • 1-D Conv: one-dimensional convolutional layer composed of a kernel of 6 convolutional filters with a length of 128 time steps. This layer has linear activation function and stride equal to 1;
  • Batch-Norm: the batch normalization layer. Keras default parameters were used;
  • Recurrent: the recurrent part of the network is composed of a GRU of 256 neurons with hyperbolic tangent activations and sigmoid as recurrent activation functions;
  • Output: the fully-connected layer composed of a single neuron with a linear activation function.

Appendix B.1.3. AUT α Model

A deep convolutional autoencoder composed of a total of 26 layers. The encoder part of the network comprises 15 layers, and the decoder part counts 10 layers. The encoder is composed of a sequence of four blocks, each composed of a sequence of four layers, except for the last one that is missing the max-pooling layer:
  • 1-D Conv: one-dimensional convolution with a filter bank of 9 filters and ReLU activations. Filter lengths are different in each block. For block one to four, lengths are, respectively, 256, 128, 64 and 16;
  • Batch-Norm: the batch normalization layer. Keras default parameters were used;
  • Dropout: random dropout of 15%;
  • Max Pooling: max pooling is applied with different pool sizes and strides parameters in each block. From block one to three pool sizes are, respectively, 2,5 and 5. The same is valid for stride parameters.
The decoder is composed of three blocks performing transposed convolution to output a signal with the same length of the input. Each block comprises three parts:
  • Up-sampling layer: repeats input before convolution operation;
  • 1-D Conv: analogous to the ones presented in the encoder but in reverse order. Bank of 9 filters with different lengths, respectively, 32, 64 and 1 for blocks from 1 to 3;
  • Batch-Norm: identical to the encoder’s one;
  • Dropout: identical to the encoder’s one.

Appendix B.2. Models for CEBSDB

Adam with a learning rate of 3 × 10 5 was used as the optimizer, and the batch size was set to 512 for all architectures. The best models were selected according to the best validation score of three training runs.

Appendix B.2.1. GRU β Model

The sequence of the network layer is the same as the GRU model for the CBPEDS, with the difference being that the Dense layer and the GRU layer are composed of 1250 and 64 units, respectively.

Appendix B.2.2. CNN-GRU β Model

The architecture of the convolutional GRU is identical to the one used for the CBPEDS. We remand to the previous section for the details.

Appendix B.2.3. AUT β Model

The architecture of the autoencoder is identical to the one used for the CBPEDS. We remand to the previous section for the details.

Appendix B.3. Models for PTBDB

Appendix B.3.1. GRU δ Model

This model was trained for 180 epochs using Adam with a learning rate of 3 × 10 3 and a batch size of 256. It is composed of the following layers:
  • Input: 187 ECG timesteps;
  • Recurrent layer: the recurrent part of the network is based on a GRU of 200 neurons with hyperbolic tangent activations and sigmoid as recurrent activation functions;
  • Dense: the fully connected layer of 100 neurons with ReLU activation functions;
  • Output: a single neuron with the sigmoid activation function.

Appendix B.3.2. CNN-GRU δ Model

This model was trained for 220 epochs using Adam with a learning rate of 3 × 10 3 and a batch size of 512. It is composed of the following layers:
  • Input: 187 ECG time steps;
  • 1-D Conv: one-dimentional convolutional layer with a filter bank of 64 kernel of size 30 and stride 1;
  • Recurrent layer: the recurrent part of the network is composed by a GRU of 32 neurons with hyperbolic tangent activations and sigmoid as recurrent activation functions;
  • Output: a single neuron with the sigmoid activation function.

Appendix B.3.3. MLP δ Model

This model was trained for 100 epochs using Adam with a learning rate of 3 × 10 4 and a batch size of 512. It is composed of the following layers:
  • Input: 187 ECG timesteps;
  • Dropout: random dropout of 10%;
  • Dense: fully connected layer of 20 units with ReLU activations;
  • Dropout: random dropout of 20%;
  • Dense: fully connected layer of 20 units with ReLU activations;
  • Dropout: random dropout of 20%;
  • Dense: fully connected layer of 20 units with ReLU activations;
  • Dropout: random dropout of 30%;
  • Output: fully-connected layer composed of a single neuron with a sigmoid activation function.

Appendix C. Visualizing Sub-Signal Occlusion Effects

Here, we show another example of the proposed visualization that highlights sub-signals contributions to the error (Figure A1). The plot shows an ECG signals from CBPEDS occluded with a window size of 25 time steps.
Figure A1. A visualization of sub-signal occlusion contributions for an ECG signal from CBPEDS. The occlusion window size is equal to 25. Red triangles mark the portions of the signal that contribute more to MAE increase.
Figure A1. A visualization of sub-signal occlusion contributions for an ECG signal from CBPEDS. The occlusion window size is equal to 25. Red triangles mark the portions of the signal that contribute more to MAE increase.
Entropy 23 01064 g0a1

References

  1. Bacciu, D.; Lisboa, P.J.; Martín, J.D.; Stoean, R.; Vellido Alcacena, A. Bioinformatics and medicine in the era of deep learning. Proceedings fo the ESANN 2018: 26th European Symposium on Artificial Neural Networks, Bruges, Belgium, 25–27 April 2018; pp. 345–354. [Google Scholar]
  2. Ganapathy, N.; Swaminathan, R.; Deserno, T.M. Deep Learning on 1-D Biosignals: A Taxonomy-Based Survey. Yearb. Med Inform. 2018, 27, 98–109. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Zhang, Y.; Feng, Z. A SVM Method for Continuous Blood Pressure Estimation from a PPG Signal. In Proceedings of the 9th International Conference on Machine Learning and Computing, Singapore, 24–26 February 2017; pp. 128–132. [Google Scholar]
  4. Wang, R.; Jia, W.; Mao, Z.H.; Sclabassi, R.J.; Sun, M. Cuff-Free Blood Pressure Estimation Using Pulse Transit Time and Heart Rate. In Proceedings of the 2014 12th international conference on signal processing (ICSP), Hangzhou, China, 19–23 October 2014; pp. 115–118. [Google Scholar]
  5. Adadi, A.; Berrada, M. Peeking inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI). IEEE Access 2018, 6, 52138–52160. [Google Scholar] [CrossRef]
  6. Miller, T. Explanation in Artificial Intelligence: Insights from the Social Sciences. Artif. Intell. 2019, 267, 1–38. [Google Scholar] [CrossRef]
  7. Lundberg, S.M.; Lee, S. A Unified Approach to Interpreting Model Predictions. In Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA; 2017; pp. 4765–4774. [Google Scholar]
  8. Guidotti, R.; Monreale, A.; Ruggieri, S.; Turini, F.; Giannotti, F.; Pedreschi, D. A Survey of Methods for Explaining Black Box Models. Acm Comput. Surv. (CSUR) 2018, 51, 93. [Google Scholar] [CrossRef] [Green Version]
  9. Bai, T.; Zhang, S.; Egleston, B.L.; Vucetic, S. Interpretable Representation Learning for Healthcare via Capturing Disease Progression through Time. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, London, UK, 19–23 August 2018; pp. 43–51. [Google Scholar]
  10. Lipton, Z.C. The Mythos of Model Interpretability: In Machine Learning, the Concept of Interpretability is Both Important and Slippery. Queue 2018, 16, 31–57. [Google Scholar] [CrossRef]
  11. Springenberg, J.T.; Dosovitskiy, A.; Brox, T.; Riedmiller, M. Striving for Simplicity: The All Convolutional Net. arXiv 2015, arXiv:412.6806. [Google Scholar]
  12. Zeiler, M.D.; Fergus, R. Visualizing and Understanding Convolutional Networks. In Proceedings of the Computer Vision—ECCV 2014, Zurich, Switzerland, 6–12 September 2014; Springer: Cham, Switzerland, 2014; pp. 818–833. [Google Scholar]
  13. Shrikumar, A.; Greenside, P.; Kundaje, A. Learning Important Features Through Propagating Activation Differences. In Proceedings of the International Conference on Machine Learning, Sydney, Australia, 6–11 August 2017; pp. 3145–3153. [Google Scholar]
  14. Bach, S.; Binder, A.; Montavon, G.; Klauschen, F.; Müller, K.R.; Samek, W. On Pixel-Wise Explanations for Non-Linear Classifier Decisions by Layer-Wise Relevance Propagation. PLoS ONE 2015, 10, e0130140. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Sundararajan, M.; Taly, A.; Yan, Q. Axiomatic Attribution for Deep Networks. In Proceedings of the International Conference on Machine Learning, Sydney, Australia, 6–11 August 2017; pp. 3319–3328. [Google Scholar]
  16. Ribeiro, M.T.; Singh, S.; Guestrin, C. “Why Should I Trust You?”: Explaining the Predictions of Any Classifier. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 1135–1144. [Google Scholar]
  17. Bahdanau, D.; Cho, K.; Bengio, Y. Neural Machine Translation by Jointly Learning to Align and Translate. In Proceedings of the 3rd International Conference on Learning Representations (ICLR 2015), San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  18. Choi, E.; Bahadori, M.T.; Kulas, J.A.; Schuetz, A.; Stewart, W.F.; Sun, J. RETAIN: An Interpretable Predictive Model for Healthcare Using Reverse Time Attention Mechanism. arXiv 2017, arXiv:1608.05745. [Google Scholar]
  19. Girkar, U.M.; Uchimido, R.; Lehman, L.H.; Szolovits, P.; Celi, L.A.; Weng, W. Predicting Blood Pressure Response to Fluid Bolus Therapy Using Attention-Based Neural Networks for Clinical Interpretability. arXiv 2018, arXiv:1812.00699. [Google Scholar]
  20. Beer, T.; Eini-Porat, B.; Goodfellow, S.; Eytan, D.; Shalit, U. Using deep networks for scientific discovery in physiological signals. In Proceedings of the Machine Learning for Healthcare Conference (MLHC 2020), Virtual Event, Durham, NC, USA, 7–8 August 2020. [Google Scholar]
  21. Mercier, D.; Dengel, A.; Ahmed, S. PatchX: Explaining Deep Models by Intelligible Pattern Patches for Time-series Classification. arXiv 2021, arXiv:2102.05917. [Google Scholar]
  22. Gee, A.H.; García-Olano, D.; Ghosh, J.; Paydarfar, D. Explaining Deep Classification of Time-Series Data with Learned Prototypes. In Proceedings of the 4th International Workshop on Knowledge Discovery in Healthcare Data Co-Located with the 28th International Joint Conference on Artificial Intelligence, KDH@IJCAI 2019, Macao, China, 10 August 2019. [Google Scholar]
  23. Li, O.; Liu, H.; Chen, C.; Rudin, C. Deep Learning for Case-Based Reasoning Through Prototypes: A Neural Network That Explains Its Predictions. In Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence, (AAAI-18), the 30th innovative Applications of Artificial Intelligence (IAAI-18), and the 8th AAAI Symposium on Educational Advances in Artificial Intelligence (EAAI-18), New Orleans, LA, USA, 2–7 February 2018. [Google Scholar]
  24. Jung, A.; Nardelli, P.H.J. An Information-Theoretic Approach to Personalized Explainable Machine Learning. IEEE Signal Process. Lett. 2020, 27, 825–829. [Google Scholar] [CrossRef]
  25. Zhou, C.; Yuan, J. Bi-Box Regression for Pedestrian Detection and Occlusion Estimation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018. [Google Scholar]
  26. Kohonen, T. Self-Organizing Maps, 3rd ed.; Springer Series in Information Sciences; Springer: Berlin/Heidelberg, Germany, 2001. [Google Scholar]
  27. Bacciu, D.; Bertoncini, G.; Morelli, D. Topographic Mapping for Quality Inspection and Intelligent Filtering of Smart-Bracelet Data. Neural Comput. Appl. 2021. [Google Scholar] [CrossRef]
  28. Kachuee, M.; Kiani, M.M.; Mohammadzade, H.; Shabany, M. Cuff-Less High-Accuracy Calibration-Free Blood Pressure Estimation Using Pulse Transit Time. In Proceedings of the IEEE International Symposium on Circuits and Systems (ISCAS), Lisbon, Portugal, 24–27 May 2015; pp. 1006–1009. [Google Scholar]
  29. Dua, D.; Graff, C. UCI Machine Learning Repository. Available online: https://ergodicity.net/2013/07/ (accessed on 17 May 2021).
  30. Saeed, M.; Villarroel, M.; Reisner, A.T.; Clifford, G.; Lehman, L.W.; Moody, G.; Heldt, T.; Kyaw, T.H.; Moody, B.; Mark, R.G. Multiparameter Intelligent Monitoring in Intensive Care II (MIMIC-II): A Public-Access Intensive Care Unit Database. Crit. Care Med. 2011, 39, 952–960. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  31. Goldberger Ary, L.; Amaral Luis, A.N.; Glass, L.; Hausdorff Jeffrey, M.; Ivanov Plamen, C.; Mark Roger, G.; Mietus Joseph, E.; Moody George, B.; Peng, C.-K.; Stanley, H.E. PhysioBank, PhysioToolkit, and PhysioNet. Circulation 2000, 101, e215–e220. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  32. García-González, M.A.; Argelagós-Palau, A.; Fernández-Chimeno, M.; Ramos-Castro, J. A Comparison of Heartbeat Detectors for the Seismocardiogram. In Proceedings of the Computing in Cardiology, Zaragoza, Spain, 22–25 September 2013; pp. 461–464. [Google Scholar]
  33. Bousseljot, R.; Kreiseler, D.; Schnabel, A. Nutzung Der EKG-Signaldatenbank CARDIODAT Der PTB Über Das Internet. Biomed. Tech. Eng. 2009, 317–318. [Google Scholar] [CrossRef]
  34. Kachuee, M.; Fazeli, S.; Sarrafzadeh, M. ECG Heartbeat Classification: A Deep Transferable Representation. In Proceedings of the 2018 IEEE International Conference on Healthcare Informatics (ICHI), New York, NY, USA, 4–7 June 2018; pp. 443–444. [Google Scholar] [CrossRef] [Green Version]
  35. Chollet, F. Keras. Available online: https://github.com/keras-team/keras (accessed on 15 March 2015).
  36. Abadi, M.; Agarwal, A.; Barham, P.; Brevdo, E.; Chen, Z.; Citro, C.; Corrado, G.S.; Davis, A.; Dean, J.; Devin, M.; et al. TensorFlow: Large-Scale Machine Learning on Heterogeneous Systems. arXiv 2016, arXiv:1603.04467. [Google Scholar]
  37. Hinton, G.E.; Zemel, R.S. Autoencoders, Minimum Description Length and Helmholtz Free Energy. In Proceedings of the 7th Neural Information Processing Systems (NIPS), Denver, CO, USA, 29 November–2 December 1993; pp. 3–10. [Google Scholar]
  38. Cho, K.; van Merriënboer, B.; Gulcehre, C.; Bahdanau, D.; Bougares, F.; Schwenk, H.; Bengio, Y. Learning Phrase Representations Using RNN Encoder—Decoder for Statistical Machine Translation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), Doha, Qatar, 25–29 October 2014; pp. 1724–1734. [Google Scholar]
  39. Cun, Y.L.; Jackel, L.D.; Boser, B.; Denker, J.S.; Graf, H.P.; Guyon, I.; Henderson, D.; Howard, R.E.; Hubbard, W. Handwritten Digit Recognition: Applications of Neural Network Chips and Automatic Learning. IEEE Commun. Mag. 1989, 27, 41–46. [Google Scholar] [CrossRef]
  40. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; MIT Press: Cambridge, MA, USA, 2016. [Google Scholar]
  41. Vellido, A. The Importance of Interpretability and Visualization in Machine Learning for Applications in Medicine and Health Care. Neural Comput. Appl. 2020, 32, 18069–18083. [Google Scholar] [CrossRef] [Green Version]
  42. Vettigli, G. MiniSom: Minimalistic and NumPy-Based Implementation of the Self Organizing Map. Available online: https://github.com/JustGlowing/minisom (accessed on 8 May 2018).
Figure 1. MIME overview. The original input X is occluded, generating X o c c . A black box model b generates predictions y and y o c c using both altered and unaltered inputs. The two predictions are compared using an error measure (e.g., MAE).
Figure 1. MIME overview. The original input X is occluded, generating X o c c . A black box model b generates predictions y and y o c c using both altered and unaltered inputs. The two predictions are compared using an error measure (e.g., MAE).
Entropy 23 01064 g001
Figure 2. GRU β predictions for the SCG with occluded input signals. With ECG occluded, the output prediction is a signal oscillating around zero values.
Figure 2. GRU β predictions for the SCG with occluded input signals. With ECG occluded, the output prediction is a signal oscillating around zero values.
Entropy 23 01064 g002
Figure 3. A visualization of sub-signal occlusion contributions for an ECG signal from CBPEDS. The occlusion window size is equal to 50. Red triangles mark the portions of the signal that contribute more to MAE increase.
Figure 3. A visualization of sub-signal occlusion contributions for an ECG signal from CBPEDS. The occlusion window size is equal to 50. Red triangles mark the portions of the signal that contribute more to MAE increase.
Entropy 23 01064 g003
Figure 4. A comparison of importance assigned to sub-signals of an ECG from PTB by mime (left) and Integrated Gradients (right). The window with the highest score is marked by the red triangle.
Figure 4. A comparison of importance assigned to sub-signals of an ECG from PTB by mime (left) and Integrated Gradients (right). The window with the highest score is marked by the red triangle.
Entropy 23 01064 g004
Figure 5. The SOM map with MAE-based colouring for a GRU α model tested on ECG sub-signals.
Figure 5. The SOM map with MAE-based colouring for a GRU α model tested on ECG sub-signals.
Entropy 23 01064 g005
Figure 6. SOM maps with MAE-based colouring for a GRU δ model tested on ECG sub-signals.
Figure 6. SOM maps with MAE-based colouring for a GRU δ model tested on ECG sub-signals.
Entropy 23 01064 g006
Table 1. Summary of datasets.
Table 1. Summary of datasets.
CBPEDSCEBSDBPTBDB
# of timeseries10,158151214,552
# of variables331
Length (time points)12501250187
Sampling freq.125 Hz5 kHz125 Hz
# of classes--2
# of normal time-series--4046
# of abnormal time-series--10,506
Table 2. Selected model performances on each dataset.
Table 2. Selected model performances on each dataset.
CBPEDS (MAE)CBPEDS (MAE)PTBDB (Accuracy)
Valid. SetTest SetValid. SetTest SetValid. SetTest Set
GRU α 14.90917.636----
CNN - GRU α 14.81417.826----
AUT α 11.74413.581----
GRU β --1.4941.662--
CNN - GRU β --1.4511.779--
AUT β --1.3152.073--
GRU δ ----99.31%99.18%
CNN - GRU δ ----96.46%96.29%
MLP δ ----92.58%99.18%
Table 3. CBPEDS Signal importance results.
Table 3. CBPEDS Signal importance results.
Occlusion TypeValidation Set MAETest Set MAE
GRU α CNN - GRU α AUT α GRU α CNN - GRU α AUT α
No occlusion14.9014.8111.7717.6317.8213.58
ECG zeroed17.4217.1713.7018.9619.6315.68
PPG zeroed17.3214.5214.2120.0118.6516.18
ECG mean16.9717.0713.7418.8319.7115.57
PPG mean18.0514.5714.1119.8018.2316.22
Table 4. CEBSDB signal importance results.
Table 4. CEBSDB signal importance results.
Occlusion TypeValidation Set MAETest Set MAE
GRU β CNN - GRU β AUT β GRU β CNN - GRU β AUT β
No occlusion1.4941.4511.3151.6621.7792.073
ECG zeroed1.8621.8642.4161.7731.7862.130
Breathing zeroed1.5321.4511.3191.8971.7791.949
ECG mean1.8681.8631.8681.7731.7862.135
Breathing mean1.5331.4261.3191.8991.6461.954
Table 5. CBPEDS mean MAE results for different window sizes. Occlusion with zero values. Lowest values are in bold, highlighting models less affected by occlusion error.
Table 5. CBPEDS mean MAE results for different window sizes. Occlusion with zero values. Lowest values are in bold, highlighting models less affected by occlusion error.
W size Signal Test Set MAE ¯
GRU α CNN - GRU α AUT α
25ECG 0.619 ± 0.143 0.406 ± 0.076 0 . 405 ± 0 . 054
25PPG 0.522 ± 0.119 0 . 448 ± 0 . 094 0.856 ± 0.112
25Both 0.810 ± 0.170 0 . 741 ± 0 . 139 0.959 ± 0.111
75ECG 1.088 ± 0.254 0 . 760 ± 0 . 153 0.810 ± 0.131
75PPG 1.088 ± 0.239 0 . 901 ± 0 . 185 1.232 ± 0.168
75Both 1.657 ± 0.344 1 . 328 ± 0 . 232 1.681 ± 0.224
125ECG 1.425 ± 0.232 1 . 080 ± 0 . 150 1.180 ± 0.140
125PPG 1.544 ± 0.227 1 . 299 ± 0 . 174 1.664 ± 0.190
125Both 2.507 ± 0.337 1 . 973 ± 0 . 213 2.362 ± 0.203
Table 6. CEBSDB Mean MAE results for different window sizes. Occlusion with zero values. Lowest values are in bold, highlighting models less affected by occlusion error.
Table 6. CEBSDB Mean MAE results for different window sizes. Occlusion with zero values. Lowest values are in bold, highlighting models less affected by occlusion error.
W size Signal Test Set MAE ¯
GRU β CNN - GRU β AUT β
25ECG 0 . 061 ± 0 . 034 0.062 ± 0.025 0.112 ± 0.016
25Breath 0 . 068 ± 0 . 057 0.071 ± 0.032 0.0682 ± 0.010
25Both 0.093 ± 0.059 0 . 091 ± 0 . 036 0.147 ± 0.020
75ECG 0 . 129 ± 0 . 057 0.137 ± 0.056 0.236 ± 0.042
75Breath 0.127 ± 0.094 0.11 ± 0.053 0 . 0887 ± 0 . 017
75Both 0.183 ± 0.098 0 . 161 ± 0 . 067 0.268 ± 0.047
125ECG 0 . 194 ± 0 . 078 0.206 ± 0.073 0.339 ± 0.041
125Breath 0.187 ± 0.126 0.147 ± 0.068 0 . 111 ± 0 . 0183
125Both 0.263 ± 0.118 0 . 231 ± 0 . 082 0.363 ± 0.046
Table 7. PTBDB models mean accuracy decrease for different window sizes. Lowest decreases in bold.
Table 7. PTBDB models mean accuracy decrease for different window sizes. Lowest decreases in bold.
Test Set Accuracy Decrease (%)
W size Occ Value GRU δ CNN - GRU δ MLP δ
25zero13.1814.2910.86
25mean10.1812.298.0
50zero20.1819.2916.0
50mean18.1817.2914.0
75zero26.1823.2919.0
75mean26.1819.1917.0
100zero29.1826.2920.0
100mean25.1826.2921.0
125zero28.1825.2921.0
125mean28.1826.2921.0
Table 8. CBPEDS perturbation duration for the different models occluded with zero value. Lowest durations are highlighted in bold.
Table 8. CBPEDS perturbation duration for the different models occluded with zero value. Lowest durations are highlighted in bold.
W size SignalTest Set Mean Duration (ts)
GRU α CNN - GRU α AUT α
25ECG 188.00 ± 40.69 138.00 ± 25.61 0 . 00 ± 0 . 00
25PPG 161.50 ± 33.25 136.00 ± 25.57 15 . 50 ± 12 . 13
25Both 206.00 ± 47.05 171.00 ± 35.83 17 . 00 ± 11 . 66
75ECG 91.18 ± 20.90 195.59 ± 43.08 0 . 00 ± 0 . 00
75PPG 244.12 ± 62.15 202.94 ± 45.28 20 . 59 ± 9 . 53
75Both 91.18 ± 20.90 213.24 ± 49.35 23 . 53 ± 5 . 88
125ECG 310.00 ± 68.19 257.50 ± 44.79 0 . 00 ± 0 . 00
125PPG 300.00 ± 64.23 257.50 ± 44.79 25 . 00 ± 0 . 00
125Both 327.50 ± 76.20 280.00 ± 55.68 25 . 00 ± 0 . 00
Table 9. CEBSDB Perturbation duration for the different models occluded with zero value. Lowest durations are highlighted in bold.
Table 9. CEBSDB Perturbation duration for the different models occluded with zero value. Lowest durations are highlighted in bold.
W size Signal Test Set Mean Duration (ts)
GRU β CNN - GRU β AUT β
25ECG 0 . 00 ± 0 . 00 14.50 ± 12.34 0 . 00 ± 0 . 00
25Breath 0 . 00 ± 0 . 00 0 . 00 ± 0 . 00 0 . 00 ± 0 . 00
25Both 18.00 ± 12.29 25.50 ± 3.50 1 . 50 ± 5 . 94
75ECG 38 . 24 ± 17 . 40 58.82 ± 14.71 91.18 ± 28.36
75Breath 2.94 ± 11.76 0 . 00 ± 0 . 00 0 . 00 ± 0 . 00
75Both 69 . 12 ± 18 . 25 75.00 ± 8.57 98.53 ± 26.39
125ECG 87 . 50 ± 16 . 77 117.50 ± 11.46 155.00 ± 18.71
125Breath 10.00 ± 30.00 0 . 00 ± 0 . 00 0 . 00 ± 0 . 00
125Both 122 . 50 ± 17 . 50 127.50 ± 7.50 192.50 ± 46.17
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Resta, M.; Monreale, A.; Bacciu, D. Occlusion-Based Explanations in Deep Recurrent Models for Biomedical Signals. Entropy 2021, 23, 1064. https://doi.org/10.3390/e23081064

AMA Style

Resta M, Monreale A, Bacciu D. Occlusion-Based Explanations in Deep Recurrent Models for Biomedical Signals. Entropy. 2021; 23(8):1064. https://doi.org/10.3390/e23081064

Chicago/Turabian Style

Resta, Michele, Anna Monreale, and Davide Bacciu. 2021. "Occlusion-Based Explanations in Deep Recurrent Models for Biomedical Signals" Entropy 23, no. 8: 1064. https://doi.org/10.3390/e23081064

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop