Next Article in Journal
Removal of Intra-Array Statics in Seismic Arrays Due to Variable Topography and Positioning Errors
Previous Article in Journal
Simultaneous Coded Plane-Wave Imaging Using an Advanced Ultrasound Forward Model
Previous Article in Special Issue
Modeling Health Data Using Machine Learning Techniques Applied to Financial Management Predictions
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Model Interpretation Considering Both Time and Frequency Axes Given Time Series Data

1
Department of Computer Science & Engineering, Hanyang University, Ansan 15588, Republic of Korea
2
Major in Bio Artificial Intelligence, Department of Applied Artificial Intelligence, Hanyang University, Ansan 15588, Republic of Korea
*
Author to whom correspondence should be addressed.
Appl. Sci. 2022, 12(24), 12807; https://doi.org/10.3390/app122412807
Submission received: 21 October 2022 / Revised: 10 December 2022 / Accepted: 11 December 2022 / Published: 13 December 2022
(This article belongs to the Special Issue Big Data Intelligence in Healthcare Applications)

Abstract

:
Recently, deep learning-based models have emerged in the medical domain. Although those models achieve high performance, it is difficult to directly apply them in practice. Specifically, most models are not considered reliable yet, while they are not interpretable. Therefore, researchers attempt to interpret their own deep learning applications. However, the interpretation is task specific or only appropriate for image data such as computed tomography (CT) scans and magnetic resonance imaging (MRI). Currently, few works focus on the model interpretation given time series data such as electroencephalography (EEG) and electrocardiography (ECG) using LIME. Because the explanation generated by LIME is from the permutation of the divided input data, the performance of interpretation is highly dependent on the split method. In the medical domain, for the time series data, existing interpretations consider only the time axis, whereas physicians take account of the frequency too. In this work, we propose the model interpretation using LIME considering both time and frequency axes. Our key idea is that we divide the input signal using graph-based image clustering after transforming it using short-time Fourier transform, which is utilized to capture the change of frequency content over time. In our experiments, we utilize real-world data, which is EEG signals recorded from patients during polysomnographic (PSG) studies, as well as prove that ours captures a significantly more critical explanation than the state-of-the-art. In addition, we show that the representation obtained by ours reflects the physician’s standard such as K-complexes and delta waves, which are considered strong evidence of the second sleep stage and a clue of the third sleep stage. We expect that our work can be applied to establish computer-aided diagnosis as well as to measure the reliability of deep learning models taking the time series into them.

1. Introduction

Recently, deep learning-based models have been tremendously developed by many researchers in the medical domain for image and signal datasets. Specifically, for image data such as computed tomography (CT) scan, magnetic resonance imaging (MRI), deep learning models based on U-Net [1], which is proposed to segment biomedical images, achieve high performance. The studies include organ segmentation, brain tumor detection and skin cancer classification [2,3,4,5,6]. In addition, deep learning models are applied to one-dimensional medical data as well. The works utilize convolutional neural networks (CNNs) and long short-term memory networks (LSTMs) for arrhythmia detection, mortality prediction, sleep stage detection, bruxism detection and insomnia detection [7,8,9,10,11,12,13,14]. However, those methods are difficult to directly apply in practice. The reason why is that the deep learning-based model is hard to interpret. The lack of interpretability means the lack of reliability. Although, some works attempt to interpret the prediction results, the interpretation highly depends on their specific task [15,16,17].
Since the interpretation methods for deep learning have been proposed, such as Grad-CAM [18] the gradient-based visualization, KernelSHAP [19] the Shapley value-based method from game theory and LIME, which exploits surrogate model [20], the interpretations are also applied to the medical domain [21,22,23]. However, Grad-CAM-based visualization is only appropriate for image data. Besides, the exact Shapley value is difficult to earn, because the Shapley value is calculated from all combinations of the features [19]. Therefore, for one-dimensional data, LIME is utilized in the state-of-the-art interpretations, which are named NEVES [24] and LIMESegment [25].
Given a target model that is to be explained and the input data, LIME exploits a surrogate model that is considered easy to interpret, such as linear model or decision tree [20]. The explanation is obtained from training the surrogate model using perturbed samples and associated logits predicted by the target model. The perturbed samples are a randomly permuted subset of the input. For example, given an image of skin cancer, the image can be divided into lesions and normal skin of the image. Then, the perturbed samples can be the image of lesions, the image of normal skin and the image of both. Because the characteristics of LIME, the explanation depends on how to split the input instance. Former works exploit LIME, split the input time series at the same interval and fill it with interpolation [24] or divide from the change point, which is the boundary of the most dissimilar between window slides [25]. They are, however, limited in that they consider the signal changing along the time axis whereas physicians consider the characteristics of signals along the time and frequency axes in practice [26,27,28,29].
In this work, we propose the method to interpret models based on deep learning given time series medical data. Specifically, we consider the medical signals, which are that not only the change point along the time axis but also the frequency plays an important role in diagnoses such as electroencephalography (EEG), electromyography (EMG) and electrocardiography (ECG) [30,31,32]. Given the input signal, our suggestion considers both time and frequency axes. We transform the input signal using short-time Fourier transform (STFT) and divide it into aspects of both axes. The divided signal is randomly permuted and inverted by inverse STFT (ISTFT). Finally, our method produces the explanation of the input as depicted in Figure 1. In the experiments, we show that our method figures out a significantly more critical explanation than the state-of-the-art and it generates a comparative explanation as a physician’s decision.
Our contributions are as follows:
  • We propose the novel approach to interpret a deep learning classifier given medical signals such as EEG considering both aspects of time and frequency while physicians consider both of them to diagnose in practice.
  • In the experiments, we confirm the suggestion from this work with the real-world dataset, which is EEG signals recorded from patients during polysomnogrphic studies.
  • We show that our suggestion captures the probable explanations such as K-complexes and delta waves, which are considered strong evidence of the second sleep stage and the third sleep stage, respectively.
The rest of our paper is organized as follows. In Section 2, we review existing interpretations for deep learning models and their limitations in the general and medical domains. We define our problem and introduce our method in Section 3. We compare the performance between ours and the state-of-the-art in Section 4. Finally, we conclude our research in Section 6. Moreover, we attach additional experimental results in Appendix A.

2. Related Work

Although deep learning model interpretation has been paid attention by many researchers [18,19,33], those works are limited to the task for image data. Some works consider the medical domain, but most studies consider only image data such as computerized tomography (CT) scan and skin lesions [15,16,17], although only a few works attempt to interpret models given the time series data [24,25]. In this section, we review existing deep learning model interpretation methods in a general domain and in a specific domain, which is the medical domain, as well as the model interpretation given time series data.

2.1. Model Interpretation in General Domain

Researchers attempt to interpret the decision of the model by calculating the importance score [19,33] and the heat map [18]. DeepLIFT [33] and KernelSHAP [19] compute the difference of scores between the original class and target class. However, both DeepLIFT and KernelSHAP are limited in the following ways. DeepLIFT has a limitation that users should find the reference input, which plays an important role in score calculation. In other words, the quality of the explanation depends on the experimenter’s domain knowledge. KernelSHAP acquires the Shapley value, which is proposed to measure the feature importance in game theory [19]. Although authors introduce an approximate method to derive the Shapley value, their interpretation process still has a huge time cost [19]. In addition, Grad-CAM [18] creates a heat map, which is related with the class based on the feature map from the hidden layers. The heat map acts as the highlight of the most connected part of the input image. However, Grad-CAM has the limitation that, if there are multiple objects that are related with the class, it does not capture them.

2.2. Model Interpretation in Medical Domain

In the medical domain, researchers acquire the attention network and gradient to interpret segmentation models given images of magnetic resonance imaging (MRI), CT and skin lesions. SAUNet exploits the attention network to extract the global context of MRI segments [15]. Factorizer [16] solves non-negative matrix factorization as a self-attention role to segment brain tumors. The attention of both SAUNet and Factorizer is utilized to interpret the model. Moreover, gradient-based saliency maps are employed to interpret a skin cancer classifier [17]. Although they claim that their interpretations assure the decision of their model is similar to a human’s, their interpretation methods are limited to the specific task, specifically the images.
Given the time series data in the medical domain such as electrocardiogram (ECG), NEVES [24] and LIMESegment [25], we find the most important subset of the input using LIME [20]. NEVES divides the signal by the same interval, while LIMESegment separates it by the change point, which is the most different point between window slides. Although they attempt to interpret the model given medical time series data, they are limited to only consider the aspect of time axis, whereas the frequency is also the important role to diagnose in practice.

3. Model Interpretation for Signal Classifier

As we mentioned in Section 2, model interpretation works have been proposed by many researchers. In the medical domain, most researchers focus on a model interpretation given image data, but few works deal with it given time series data. Specifically, in the medical domain, for diagnosis given time series data such as electroencephalogram (EEG) and electromyography (EMG), physicians consider not only periodic patterns, but also frequency bands [27,28,29], whereas existing methods take account of periodic patterns [24,25]. Therefore, we aim to develop an improved model interpretation given time series data. We simply define our problem as follows.
Problem definition: Given a black box classifier f, which is inaccessible model parameters, and to classify an input data x, which is a time series, our goal is to provide an interpretable representation ξ ( x ) of the input.
In the rest of this section, we introduce our explanation method and the building blocks for our approach.

3.1. Prerequisites

LIME is an explanation method of machine learning models that is a black box classifier [20]. Given a model f to be explained and an input instance x, LIME defines explanation ξ ( x ) as follows using explainable model g, which has trainable parameters such as a linear model or decision tree:
ξ ( x ) = arg min g L ( f , g , π x ) + Ω ( g )
where π x represent the distance between the input instance and the perturbed sample, as well as Ω ( g ) is the complexity of the explainable model g such as the number of non-zero weights for a linear model. For the complexity Ω ( g ) , we choose a constant so that LIME claims that Ω ( g ) can be a hyper-parameter [20].
Once the complexity is chosen, the input x is divided by discrete components that are interpretable. For example, given an image of a dog in the grass and a complexity of 2, it can be split into a set Z of two images such as the dog and the grass using a super-pixel such as Felzenszwalb [34]. Then, a trainable linear model g is obtained by minimizing the loss as below:
L ( f , g , π x ) = z , z Z π x ( z ) f ( z ) g ( z ) 2
where z and z represent the perturbed sample and the discrete component of the input x named perturbed sample, respectively, e.g., a perturbed sample z can be one of the image of the dog or the image of the grass. At that time, the positive model parameter of g describes the importance of the component because entries of z are 0 or 1 where those entries represent “absence” and “presence” of the component, respectively. Consequently, selection of the split method plays an important role. For time series data, former works apply LIME as well [24,25,35]. Unlikely for images, they not only focus on how to split the input, but also focus on how to generate perturbed regions. They acquire random split [24,35] or finding the most probable change points comparing the similarity between window slides [25]. In addition, they attempt to fill the regions to generate perturbed samples using constant value, linear interpolation or both short-time Fourier transform (STFT) and inverse STFT (ISTFT).

3.2. Perturbed Sample Generation for Time Series Data

Our approach is based on the given time series data being decomposed into two-dimensional information, which are periodic patterns on the time axis and frequency characteristics on the frequency axis. Moreover, as we mentioned earlier in Section 3.1, the selection of both split method and filling method are important to generate the most probable perturbed samples to exploit LIME for time series data. Our split method and filling method are described in the following sections.

3.2.1. Separating Representative Component of Signal Data Using STFT and Super-Pixel

In practice, observed signals in medical centers are non-stationary, e.g., recorded EEG for diagnosis seizure. In other words, biomedical signals change frequency over time. Moreover, because physicians consider both periodic patterns and frequency bands simultaneously to diagnose a patient [27,28,29], we exploit STFT to transform the input signal.
STFT is considered to capture the change of frequency content over time, whereas the conventional discrete time Fourier transform (DTFT) averages frequency contents over time [36]. STFT given the input signal is defined as follows:
X [ k , m ] = n = m m + ( N 1 ) x [ n ] w [ n m ] e j ω ( n m )
where x [ n ] is the observed data point at time n, w [ · ] is the window function given the window size and ω represents the angular frequency as well as N is the number of samples. Therefore, transformed X is represented as a 2-dimensional matrix that has the axes of the frequency and the time segment. STFT describes the magnitudes of the frequencies for the m-th segment. For example, we transform the input x using STFT and visualize it as the spectrogram at the left diagram in Figure 2. As we can see in the second-row, the high magnitude of the low frequency is represented well around the 19th time segment.
Since our main purpose in this subsection is to divide the input signal into representative components, our next step is to segment the spectrogram. In this work, we utilize Felzenszwalb [34], which is proposed to segment an image into sub-regions. The intuition of Felzenszwalb is that pixels in an image and the differences between pixel values are considered as nodes v i V and weights e i j E for nodes v i and v j in an undirected graph G = ( V , E ) . v i is already in a region S, but neighbor node v j is not yet. It compares the weights e i j to the maximum weight M a x ( w ) in the region S. After that, it decides whether those pixels are in a same region or not using threshold τ . The threshold is added to the maximum weight M a x ( w ) . We apply it for the spectgrogram as shown in Figure 2. Consequently, we separate clusters, which are divided by associated frequency and time pixels as we depicted in the third-row of the left diagram.

3.2.2. Restoration the Signal Data from Separated Representations

We generate a perturbed sample from the segmentation mentioned in Section 3.2.1. Because of the characteristics of Felzenszwalb, it divides around 40 segments. We select the masking ratio for the the predicted labels by the model f to be explained as evenly distributed. More concretely, for an input instance x, we generate 100 masked samples for each masking ratio in range from 0.1 to 0.9 and choose the ratio, which maximizes the entropy of the prediction by the model f. In Figure 2, in the right diagram, we show the masking and sampling process. The upper three boxes are masked samples. Based on the masking regions of the spectrograms, we zero out the real and imaginary value obtained by STFT. Finally, we derive ISTFT as the following formula:
x [ n ] = 1 N W m = n ( N 1 ) n k = 0 N 1 X [ k , m ] e j ω ( n m )
where W is the summation of the window function for all segments and other terms are same as Equation (3). In Figure 2, in the lower boxes in the right diagram, we depict the results of ISTFT. The inverse signals that are highlighted in red show that our approach masks the signal, not only the aspect of time but also the aspect of frequency.

3.3. Data Description

Our target model, which is to be explained, is trained using SleepEDF [37]. The research consists of the cassette study and the telemetry study, which are able to discover the effect of nothing from healthy patients and an insomnia medication temazepam from patients who have mild difficulty falling asleep, respectively. Because our purpose focuses on model interpretation, not the effect of a medicine, we consider only the dataset from the cassette study. The dataset is obtained from around 40 healthy patients of varied ages and sexes. The original dataset includes various channels, but we use only one channel of EEG in this work. Each record is around 8 h. We split the record for each epoch, which is 30 s. Moreover, the dataset is annotated into six sleep stages, which are wakefulness, non-REM 1–4 and REM according to R&K rule [38], but we merge non-REM 3 and 4 as stated in the American Academy of Sleep Medicine (AASM) standard [26]. Therefore, the target model classifies five sleep stages given a 30 s EEG signal. Finally, we obtain 6955, 2581, 15951, 5051 and 7042 records of wakefulness, non-REM 1–3 and REM, respectively. We split the data into 75% and 25% to train and interpret the target model, respectively.

4. Experiments

In this section, we describe our experimental environment and results as well as we discuss our results. Every method including ours is implemented using PyTorch 1.11 in Python 3.8. All experiments reported in this section are performed on machines with Intel(R) Core(TM) i9-7900X CPU @ 3.30 GHz and 128 GB main memory running Ubuntu 18 OS. We also utilize a single GPU and NVIDIA GeForce 1080 Ti equipped with 11 GB of memory.

4.1. Implementation Methods

We implement a target model, which is to be explained, and the interpretation methods in the following list.
  • Target model: Since our main goal is to interpret a deep learning classifier, we train a deep learning model, which is to be explained, by the proposed method. The target model is proposed for automatic sleep stage classification [39]. We train the model using the SleepEDF dataset, which is described in Section 3.3. The model consists of convolutional neural networks (CNNs) and long short-term memory networks (LSTMs). CNNs and LSTMs in the model are designed to capture features from a given epoch of an electroencephalography (EEG) signal and from sequential epochs of EEG signals, respectively. Because only CNN layers, which are called FeatureNet, achieve comparative performance with the full model in [39], our target model is FeatureNet.
  • LIMESegment [25]: We compare our method with LIMESegment. They also exploit LIME to interpret a classifier for time series. They focus on detecting the change point given the time series data, which are considered the boundary of segments. Once the change points of the input signal are determined comparing the similarity of window slides, it generates perturbed samples. To create the perturbed samples, the intuition is that the segment in the input signal is filtered based on the threshold frequency. The threshold is determined by the highest frequency value over time with minimal variance. After that, anything lower than the threshold is filtered out. We exploit the imeplementation given by the authors from their GitHub repository.
  • Ours: We also implement our interpretation method, which is described in Section 3.

4.2. Qualitative Results

Since we aim to provide the interpretable representation, we measure how much the representation describes the decision of the target model given the input well. Therefore, we remove the most representative region searched by our work as well as LIMESegment, the baseline. Then, we compute the difference of the logit. The experiment is based on our assumption as follows. A modified input that is removed from the most representative region is given to the target model. Then, we assume that the logit will be significantly changed.
We remove 25% of the most representative region according to each method and calculate the difference of the logit between the modified input and associated original input. For example, let a signal of wakefulness given to the target model produce the logit 0.85. If we remove 25% of the most representative region in the signal and it is given to the target model, the logit will be decreased to some value such as 0.35. Then, the difference of the logit is 0.5. In other words, the representative region provides probable decision criteria of the target model given the signal. We list the mean of the difference for the sleep stages, which are wakefulness, non-REM 1–3 and REM in Table 1. In the table, the five sleep stages are annotated as “W”, “N1”, “N2”, “N3” and “REM”. In the table, the mean of the difference of logits shows that our work finds a more representative region than the baseline.
Subsequently, we exploit t-test for each class, which compares the means of two groups and whether they are significantly different or not. We set the null hypothesis and the alternative hypothesis for the test as below.
  • Null hypothesis: The representations obtained by ours and LIMESegment are same.
  • Alternative hypothesis: The representation obtained by ours is more critical for the model decision than the baseline.
Because the standard deviations are not same between ours and the baseline, we apply Welch’s t-test and set alpha 10%. Finally, we obtain p-values for each sleep stage less than 1 × 10 5 . Therefore, we reject the null hypothesis and statistically prove that our method is significantly critical to describe the model decision.

4.3. Case Study

According to the AASM manual [26], the K-complexes are shown during the sleep stage of non-REM 2 and they are considered strong evidence for the sleep stage. A K-complex is a wave starting from a high-voltage peak and followed by slow and large negative complex peaks around 350–550 ms. Moreover, there is an important clue for the sleep stage of non-REM 3. If the delta wave exists for more than 50% in an epoch, it is considered N3 [26]. The delta wave is the frequency between from 0.4 to 5 Hz. In this section, we show samples of representations obtained by ours and the baseline. Then, we discuss how the samples are representative, such as the rule stated in the AASM manual.
In Figure 3a, we depict the representations obtained by us and LIMESegment for N2. The representative regions are filled with light green. The 1st and 4th rows show the most representative regions for each method. Note that ours is selected from the spectrogram transformed by STFT while the baseline is raw signals. The most representative region for the class is shown as the blue lines in the 2nd and 5th rows, whereas without the representation is shown as the red lines in 3rd and 6th rows. The title of each signal explains the predicted class from the target model given the signal. As shown in the figure, the proposed method captures the K-complexes well, as pictured in the blue boxes. The baseline captures the K-complexes as the most opposite representation in the red boxes. Therefore, the only representation signal obtained by ours is still classified as the original class N2 as shown by the blue lines. However, the representation from the baseline is not classified as the original class, rather it is classified as N1 the blue line at the upper sample. Moreover, without the representative region, both signals from ours are misclassified as REM, but from the baseline is still classified as N2, the red line at the lower sample. In addition, when we retain only the most opposite representation by ours, the remaining signals are classified as different sleep stages as REM and N1, while the baseline ones are classified as the original sleep stage.
In addition, we also draw the representative regions from both methods in Figure 3b. The regions are filled light green as well. As we can see in the figure, our method captures delta waves well in the blue boxes at the 2nd and 5th rows, whereas the baseline does only part of it. Therefore, only representative signals from ours are classified as the original sleep stage N3, but the baseline ones are misclassified as N2. Without the representation, the signals made by ours are misclassified as N1 and N2, while the signals from the baseline are still classified as N3.

5. Discussion

In this section, we analyze the results produced in the experiments. Moreover, we introduce a the possible scenario of developing applications using our method in practice. In addition, we discuss the limitations of our method and the future direction.

5.1. Analysis of Results

Our method is confirmed to outperform the baseline as described in Section 4.2. It is proved using statistical analysis. Moreover, in Section 4.3, it shows that the explanations obtained by ours for sleep stage N2 and N3 represent plausible reasons of diagnosis by the target model as physicians, as stated in the AASM standard [26]. Specifically, our method captures not only the explanation of the K-complexes for sleep stage N2, but also the explanation of delta waves for sleep stage N3. We analyze that because our method considers more fine-grained components of the input signal, and unlike the baseline considers coarse components, it is available to capture more representative regions.

5.2. Possible Application of Proposed Method

Moreover, our method can be adopted to develop an application to support physicians who want to utilize deep learning-based computer-aided diagnosis (CAD). For example, a physician can reduce the time cost of diagnosis using the CAD, and the reliability of the CAD is provided using ours, such as the explanations as depicted as the blue lines in Figure 3.

5.3. Limitations and Future Directions

Although our method considers both time and frequency axes together, it is limited in that the suggestion in this work depends on a clustering method for a transformed signal such as Felzenszwalb [34], which is exploited in this work. Moreover, the clustering method could not take account of the important bands of frequency. For example, according to the AASAM manual [26], humans consider the frequency at most beta waves, which is between 14–30 Hz. However, the clustering method does not consider the bands. Therefore, our method captures the over-explained sample from 0–100 Hz as shown in the region filled with light green in Figure 3. In the future, we plan to elaborate our method to consider the frequency more precisely and apply the model given the signal, not only EEG, but also different ones such as ECG or EMG.

6. Conclusions

In this work, we introduced the method to generate perturbed samples considering time and frequency axes to interpret the model given the signal data acquiring LIME. In the experiments, we provided our model interpretation, which captures significantly more important representation compared to the state-of-the art in a statistical manner. Moreover, in the case study, we showed that the representation obtained by ours reflects the physician’s standards, such as AASM. We expect that our method promotes the model based on deep learning and can support computer-aided diagnosis in medical centers.

Author Contributions

Conceptualization, W.L. and Y.K.; methodology, W.L.; software, W.L., G.K. and J.Y.; validation, W.L., G.K. and J.Y.; formal Analysis, W.L.; investigation, W.L., G.K. and J.Y.; data curation, W.L., G.K. and J.Y.; writing-original draft preparation, W.L., G.K. and J.Y.; writing-review and editing, all authors; visualization, G.K. and J.Y.; supervision, W.L. and Y.K.; project administration, W.L. and Y.K.; funding acquisition Y.K. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Research Foundation of Korea (NRF) grant funded by the Korea government (MSIT) (No. 2020R1G1A1011471) and the Artificial Intelligence Convergence Research Center (Hanyang University ERICA), under Grant 2020-0-01343.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Restrictions apply to the availability of these data. Data was obtained from “Sleep-EDF Database Expanded” and are available https://www.physionet.org/content/sleep-edfx/1.0.0/ (accessed on 1 July 2022).

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

Appendix A. Additional Representations

In this section, we attach additional representations obtained by our work and LIMESegment. In Figure A1, our representations show that it captures K-complexes well, which is the hard evidence to annotate as N2, while the baseline is not. In addition, we depict how ours catches delta waves, which is the important clue for N3, whereas the baseline only does partly in Figure A2.
Figure A1. Additional representations obtained by ours and LIMESegment for N2.
Figure A1. Additional representations obtained by ours and LIMESegment for N2.
Applsci 12 12807 g0a1
Figure A2. Additional representations obtained by ours and LIMESegment for N3.
Figure A2. Additional representations obtained by ours and LIMESegment for N3.
Applsci 12 12807 g0a2

References

  1. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  2. Zhou, Z.; Rahman Siddiquee, M.M.; Tajbakhsh, N.; Liang, J. Unet++: A nested u-net architecture for medical image segmentation. In Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support; Springer: Granada, Spain, 2018; pp. 3–11. [Google Scholar]
  3. Zhou, Y.; Huang, W.; Dong, P.; Xia, Y.; Wang, S. D-UNet: A dimension-fusion U shape network for chronic stroke lesion segmentation. IEEE/ACM Trans. Comput. Biol. Bioinform. 2019, 18, 940–950. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Li, X.; Chen, H.; Qi, X.; Dou, Q.; Fu, C.W.; Heng, P.A. H-DenseUNet: Hybrid densely connected UNet for liver and tumor segmentation from CT volumes. IEEE Trans. Med. Imaging 2018, 37, 2663–2674. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Isensee, F.; Kickingereder, P.; Wick, W.; Bendszus, M.; Maier-Hein, K.H. Brain tumor segmentation and radiomics survival prediction: Contribution to the brats 2017 challenge. In Proceedings of the International MICCAI Brainlesion Workshop, Quebec City, QC, Canada, 14 September 2017; pp. 287–297. [Google Scholar]
  6. Kayalibay, B.; Jensen, G.; van der Smagt, P. CNN-based segmentation of medical imaging data. arXiv 2017, arXiv:1701.03056. [Google Scholar]
  7. Razavian, N.; Marcus, J.; Sontag, D. Multi-task prediction of disease onsets from longitudinal laboratory tests. In Proceedings of the Machine Learning for Healthcare Conference. PMLR, Los Angeles, CA, USA, 19–20 August 2016; pp. 73–100. [Google Scholar]
  8. Kachuee, M.; Fazeli, S.; Sarrafzadeh, M. Ecg heartbeat classification: A deep transferable representation. In Proceedings of the 2018 IEEE International Conference on Healthcare Informatics (ICHI), New York, NY, USA, 4–7 June 2018; pp. 443–444. [Google Scholar]
  9. Yamaç, M.; Duman, M.; Adalıoğlu, İ.; Kiranyaz, S.; Gabbouj, M. A Personalized Zero-Shot ECG Arrhythmia Monitoring System: From Sparse Representation Based Domain Adaption to Energy Efficient Abnormal Beat Detection for Practical ECG Surveillance. arXiv 2022, arXiv:2207.07089. [Google Scholar]
  10. Li, R.; Zhang, X.; Dai, H.; Zhou, B.; Wang, Z. Interpretability analysis of heartbeat classification based on heartbeat activity’s global sequence features and BiLSTM-attention neural network. IEEE Access 2019, 7, 109870–109883. [Google Scholar] [CrossRef]
  11. Bin Heyat, M.B.; Akhtar, F.; Khan, A.; Noor, A.; Benjdira, B.; Qamar, Y.; Abbas, S.J.; Lai, D. A novel hybrid machine learning classification for the detection of bruxism patients using physiological signals. Appl. Sci. 2020, 10, 7410. [Google Scholar] [CrossRef]
  12. Heyat, M.B.B.; Lai, D.; Khan, F.I.; Zhang, Y. Sleep bruxism detection using decision tree method by the combination of C4-P4 and C4-A1 channels of scalp EEG. IEEE Access 2019, 7, 102542–102553. [Google Scholar] [CrossRef]
  13. Heyat, M.B.; Akhtar, F.; Khan, M.H.; Ullah, N.; Gul, I.; Khan, H.; Lai, D. Detection, treatment planning, and genetic predisposition of bruxism: A systematic mapping process and network visualization technique. CNS Neurol. Disord.-Drug Targets (Former. Curr. Drug Targets-CNS Neurol. Disord.) 2021, 20, 755–775. [Google Scholar]
  14. Bin Heyat, M.B.; Akhtar, F.; Ansari, M.; Khan, A.; Alkahtani, F.; Khan, H.; Lai, D. Progress in detection of insomnia sleep disorder: A comprehensive review. Curr. Drug Targets 2021, 22, 672–684. [Google Scholar]
  15. Sun, J.; Darbehani, F.; Zaidi, M.; Wang, B. Saunet: Shape attentive u-net for interpretable medical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Lima, Peru, 4–8 October 2020; pp. 797–806. [Google Scholar]
  16. Ashtari, P.; Sima, D.; De Lathauwer, L.; Sappey-Marinierd, D.; Maes, F.; Van Huffel, S. Factorizer: A Scalable Interpretable Approach to Context Modeling for Medical Image Segmentation. arXiv 2022, arXiv:2202.12295. [Google Scholar] [CrossRef]
  17. Margeloiu, A.; Simidjievski, N.; Jamnik, M.; Weller, A. Improving interpretability in medical imaging diagnosis using adversarial training. arXiv 2020, arXiv:2012.01166. [Google Scholar]
  18. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-cam: Visual explanations from deep networks via gradient-based localization. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 618–626. [Google Scholar]
  19. Lundberg, S.M.; Lee, S.I. A unified approach to interpreting model predictions. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Volume 30. [Google Scholar]
  20. Ribeiro, M.T.; Singh, S.; Guestrin, C. “Why should i trust you? ” Explaining the predictions of any classifier. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 1135–1144. [Google Scholar]
  21. He, T.; Guo, J.; Chen, N.; Xu, X.; Wang, Z.; Fu, K.; Liu, L.; Yi, Z. MediMLP: Using Grad-CAM to extract crucial variables for lung cancer postoperative complication prediction. IEEE J. Biomed. Health Inform. 2019, 24, 1762–1771. [Google Scholar] [CrossRef] [PubMed]
  22. Lu, S.; Chen, R.; Wei, W.; Belovsky, M.; Lu, X. Understanding Heart Failure Patients EHR Clinical Features via SHAP Interpretation of Tree-Based Machine Learning Model Predictions. In AMIA Annual Symposium Proceedings; American Medical Informatics Association: San Diego, CA, USA, 2021; Volume 2021, p. 813. [Google Scholar]
  23. Magesh, P.R.; Myloth, R.D.; Tom, R.J. An explainable machine learning model for early detection of Parkinson’s disease using LIME on DaTSCAN imagery. Comput. Biol. Med. 2020, 126, 104041. [Google Scholar] [CrossRef] [PubMed]
  24. Neves, I.; Folgado, D.; Santos, S.; Barandas, M.; Campagner, A.; Ronzio, L.; Cabitza, F.; Gamboa, H. Interpretable heartbeat classification using local model-agnostic explanations on ECGs. Comput. Biol. Med. 2021, 133, 104393. [Google Scholar] [CrossRef] [PubMed]
  25. Sivill, T.; Flach, P. LIMESegment: Meaningful, Realistic Time Series Explanations. In Proceedings of the International Conference on Artificial Intelligence and Statistics. PMLR, Virtual, 28–30 March 2022; pp. 3418–3433. [Google Scholar]
  26. Berry, R.B.; Brooks, R.; Gamaldo, C.E.; Harding, S.M.; Marcus, C.; Vaughn, B.V. The AASM manual for the scoring of sleep and associated events. Rules Terminol. Tech. Specif. Darien Ill. Am. Acad. Sleep Med. 2012, 176, 2012. [Google Scholar]
  27. Krahn, L.E.; Silber, M.H.; Morgenthaler, T.I. Atlas of Sleep Medicine; CRC Press: Boca Raton, FL, USA, 2015. [Google Scholar]
  28. Moser, D.; Anderer, P.; Gruber, G.; Parapatics, S.; Loretz, E.; Boeck, M.; Kloesch, G.; Heller, E.; Schmidt, A.; Danker-Hopfe, H.; et al. Sleep classification according to AASM and Rechtschaffen & Kales: Effects on sleep scoring parameters. Sleep 2009, 32, 139–149. [Google Scholar] [PubMed]
  29. Fournier, E.; Arzel, M.; Sternberg, D.; Vicart, S.; Laforet, P.; Eymard, B.; Willer, J.C.; Tabti, N.; Fontaine, B. Electromyography guides toward subgroups of mutations in muscle channelopathies. Ann. Neurol. Off. J. Am. Neurol. Assoc. Child Neurol. Soc. 2004, 56, 650–661. [Google Scholar] [CrossRef]
  30. Li, Q.; Gao, J.; Huang, Q.; Wu, Y.; Xu, B. Distinguishing Epileptiform Discharges From Normal Electroencephalograms Using Scale-Dependent Lyapunov Exponent. Front. Bioeng. Biotechnol. 2020, 8, 1006. [Google Scholar] [CrossRef]
  31. Li, Q.; Gao, J.; Zhang, Z.; Huang, Q.; Wu, Y.; Xu, B. Distinguishing epileptiform discharges from normal electroencephalograms using adaptive fractal and network analysis: A clinical perspective. Front. Physiol. 2020, 11, 828. [Google Scholar] [CrossRef]
  32. Gao, J.; Sultan, H.; Hu, J.; Tung, W.W. Denoising nonlinear time series by adaptive filtering and wavelet shrinkage: A comparison. IEEE Signal Process. Lett. 2009, 17, 237–240. [Google Scholar]
  33. Shrikumar, A.; Greenside, P.; Kundaje, A. Learning important features through propagating activation differences. In Proceedings of the International Conference on Machine Learning. PMLR, Sydney, Australia, 6–11 August 2017; pp. 3145–3153. [Google Scholar]
  34. Felzenszwalb, P.F.; Huttenlocher, D.P. Efficient graph-based image segmentation. Int. J. Comput. Vis. 2004, 59, 167–181. [Google Scholar] [CrossRef]
  35. Guillemé, M.; Masson, V.; Rozé, L.; Termier, A. Agnostic local explanation for time series classification. In Proceedings of the 2019 IEEE 31st International Conference on Tools with Artificial Intelligence (ICTAI), Portland, OR, USA, 4–6 November 2019; pp. 432–439. [Google Scholar]
  36. Oppenheim, A.V. Discrete-Time Signal Processing; Pearson Education India: Noida, India, 1999. [Google Scholar]
  37. Kemp, B.; Zwinderman, A.H.; Tuk, B.; Kamphuisen, H.A.; Oberye, J.J. Analysis of a sleep-dependent neuronal feedback loop: The slow-wave microcontinuity of the EEG. IEEE Trans. Biomed. Eng. 2000, 47, 1185–1194. [Google Scholar] [CrossRef] [PubMed]
  38. Wolpert, E.A. A Manual of Standardized Terminology, Techniques and Scoring System for Sleep Stages of Human Subjects. Arch. Gen. Psychiatry 1969, 20, 246–247. [Google Scholar] [CrossRef]
  39. Supratak, A.; Dong, H.; Wu, C.; Guo, Y. DeepSleepNet: A model for automatic sleep stage scoring based on raw single-channel EEG. IEEE Trans. Neural Syst. Rehabil. Eng. 2017, 25, 1998–2008. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Given the target model f, which is to be explained, and the input signal x, which is an EEG signal, the explanation ξ ( x ) is obtained using the surrogate model g, which is easily interpretable as a linear model. The model explanation ξ ( x ) depends on the perturbed samples z.
Figure 1. Given the target model f, which is to be explained, and the input signal x, which is an EEG signal, the explanation ξ ( x ) is obtained using the surrogate model g, which is easily interpretable as a linear model. The model explanation ξ ( x ) depends on the perturbed samples z.
Applsci 12 12807 g001
Figure 2. The process of generating perturbed samples consists of two sub processes: (1) As shown in the left diagram, given one-dimensional EEG signal x, the input is transformed using STFT. After that, it is divided using image segmentation such as Felzenszwalb [34]. (2) The divided samples are randomly permuted to generate perturbed samples as depicted at the upper side in the right diagram. Then, the perturbed samples are converted into one-dimensional signal again using ISTFT.
Figure 2. The process of generating perturbed samples consists of two sub processes: (1) As shown in the left diagram, given one-dimensional EEG signal x, the input is transformed using STFT. After that, it is divided using image segmentation such as Felzenszwalb [34]. (2) The divided samples are randomly permuted to generate perturbed samples as depicted at the upper side in the right diagram. Then, the perturbed samples are converted into one-dimensional signal again using ISTFT.
Applsci 12 12807 g002
Figure 3. We depict how the representations are critical for the decision of the target model. The blue box and the red box indicate the most probable positive explanation of the target model and the most probable negative explanation of the target model given the signal respectively. They can be acquired to support physicians to interpret the target model. (a) Original sleep stage of the EEG samples is N2. (b) Original sleep stage of the EEG samples is N3.
Figure 3. We depict how the representations are critical for the decision of the target model. The blue box and the red box indicate the most probable positive explanation of the target model and the most probable negative explanation of the target model given the signal respectively. They can be acquired to support physicians to interpret the target model. (a) Original sleep stage of the EEG samples is N2. (b) Original sleep stage of the EEG samples is N3.
Applsci 12 12807 g003
Table 1. The difference of the logit between the input, which is removed from the most representative region and associated original input. The entry describes the mean of the the difference with regard to the sleep stage. Each column is for wakefulness, non-REM 1–3 and REM.
Table 1. The difference of the logit between the input, which is removed from the most representative region and associated original input. The entry describes the mean of the the difference with regard to the sleep stage. Each column is for wakefulness, non-REM 1–3 and REM.
Methods\Sleep StagesWN1N2N3REM
LIMESegment0.070.230.040.240.05
Ours0.460.760.510.990.65
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Lee, W.; Kim, G.; Yu, J.; Kim, Y. Model Interpretation Considering Both Time and Frequency Axes Given Time Series Data. Appl. Sci. 2022, 12, 12807. https://doi.org/10.3390/app122412807

AMA Style

Lee W, Kim G, Yu J, Kim Y. Model Interpretation Considering Both Time and Frequency Axes Given Time Series Data. Applied Sciences. 2022; 12(24):12807. https://doi.org/10.3390/app122412807

Chicago/Turabian Style

Lee, Woonghee, Gayeon Kim, Jeonghyeon Yu, and Younghoon Kim. 2022. "Model Interpretation Considering Both Time and Frequency Axes Given Time Series Data" Applied Sciences 12, no. 24: 12807. https://doi.org/10.3390/app122412807

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop