Next Article in Journal
Analysis of the Influence of Moisture Variation on the Behavior of Tropical Soils of Carajás Railway
Next Article in Special Issue
Distributional Representation of Cyclic Alternating Patterns for A-Phase Classification in Sleep EEG
Previous Article in Journal
Electrochemical Analysis of MnO2 (α, β, and γ)-Based Electrode for High-Performance Supercapacitor Application
Previous Article in Special Issue
Transposed Convolution as Alternative Preprocessor for Brain-Computer Interface Using Electroencephalogram
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

DABaCLT: A Data Augmentation Bias-Aware Contrastive Learning Framework for Time Series Representation

School of Artificial Intelligence, Beijing University of Posts and Telecommunications, Beijing 100876, China
*
Authors to whom correspondence should be addressed.
Appl. Sci. 2023, 13(13), 7908; https://doi.org/10.3390/app13137908
Submission received: 18 May 2023 / Revised: 26 June 2023 / Accepted: 4 July 2023 / Published: 6 July 2023
(This article belongs to the Special Issue Deep Learning for Electroencephalography(EEG) Data Analysis)

Abstract

:
Contrastive learning, as an unsupervised technique, has emerged as a prominent method in time series representation learning tasks, serving as a viable solution to the scarcity of annotated data. However, the application of data augmentation methods during training can distort the distribution of raw data. This discrepancy between the representations learned from augmented data in contrastive learning and those obtained from supervised learning results in an incomplete understanding of the information contained in the real data from the trained encoder. We refer to this as the data augmentation bias (DAB), representing the disparity between the two sets of learned representations. To mitigate the influence of DAB, we propose a DAB-aware contrastive learning framework for time series representation (DABaCLT). This framework leverages a raw features stream (RFS) to extract features from raw data, which are then combined with augmented data to create positive and negative pairs for DAB-aware contrastive learning. Additionally, we introduce a DAB-minimizing loss function (DABMinLoss) within the contrasting module to minimize the DAB of the extracted temporal and contextual features. Our proposed method is evaluated on three time series classification tasks, including sleep staging classification (SSC) and epilepsy seizure prediction (ESP) based on EEG and human activity recognition (HAR) based on sensors signals. The experimental results demonstrate that our DABaCLT achieves strong performance in self-supervised time series representation, 0.19% to 22.95% accuracy improvement for SSC, 2.96% to 5.05% for HAR, 1.00% to 2.46% for ESP, and achieves comparable performance to the supervised approach. The source code for our framework is open-source.

1. Introduction

Time series data, collected from diverse sources such as personal devices, clinical centers, and financial institutions, have a wide range of applications such as sleep stage classification, human activity recognition, and epilepsy seizure prediction [1,2,3,4,5]. Representation learning is crucial for reducing dimensionality and constructing effective representations of time series data, enabling tasks such as classification, forecasting, and clustering. Consequently, representation learning methods for time series have gained significant attention [6,7,8,9]. However, challenges arise in designing supervised representation methods due to labeling limitations and difficulties in capturing context. These challenges necessitate improved generalization ability and the availability of more labeled data. In this context, Ref. [10] compares heuristics and event detection algorithms to determine optimal window sizes for time representation, highlighting the reliable estimation provided by the false nearest neighbors method. Additionally, Ref. [11] refines feature representations for time series using instance-level augmentation, dropout, and iterative bilinear temporal–spectral fusion. Moreover, Ref. [12] proposes a novel framework for multivariate time series representation learning using the transformer encoder architecture and unsupervised pre-training.
Unlabeled time series data are frequently encountered in practical situations, leading to increased attention on unsupervised and self-supervised time series representation learning [11]. Contrastive learning, initially proposed for representation learning in computer vision and natural language processing, has also found applications in unsupervised time series representation learning [13,14,15]. For example, Ref. [16] presents a universal representation learning framework for time series using hierarchical contrasting to learn scale-invariant representations within augmented context views. Furthermore, Ref. [17] proposes a contrastive learning framework for learning domain-invariant semantics in multivariate time series through nearest-neighbor contrastive learning. This approach constructs positive and negative pairs via data augmentation and aims to maximize the similarity of positive pairs while minimizing the similarity of negative pairs. Additionally, Ref. [9] combines adversarial training with self-supervised learning for emotion recognition, optimizing classifiers using diverse simulated time series samples. Another contrasting approach introduced by [18] utilizes expert features to encourage desirable properties in learned representations.
Various augmentation methods are specifically designed to construct positive and negative pairs. For instance, Ref. [19] considers sampled subseries as augmentation pairs, while Ref. [11] adopts a dropout strategy to randomly mask parts of the original data. Ref. [20] transforms time series into the frequency domain, and Ref. [7] selects various transformations, such as scaling and permutation, to learn transformation-invariant representations. Contrastive learning methods typically employ two variants of the same augmentation [21], aiming to improve robustness and convergence efficiency.
However, most contrastive methods overlook the representation biases introduced by augmentation techniques. Most of the previous frameworks only train encoders using augmented views [22], leading to a misunderstanding of the underlying data distribution when real data are absent. This gap between the representations learned from augmented data in contrastive learning and those learned from supervised learning is referred to as the data augmentation bias (DAB). The DAB represents the disparity between representations learned from augmented data in contrastive learning and those obtained from supervised learning. Consequently, when fine-tuning the classifier for downstream tasks, the encoder fails to rectify the misconceptions embedded within its representations, and the learned representations may not capture the underlying features of the raw data. The existence and impact of the DAB problem are discussed in the fourth section.
To address the influence of the DAB, we propose the DAB-aware contrastive learning framework (DABaCLT) that aims to recognize and minimize the impact of DAB on downstream tasks. The framework consists of two streams: the DAB features stream (DABFS) and the raw features stream (RFS). The DABFS extracts augmented temporal and contextual features, while the RFS helps identify and mitigate the DAB issue present in the learned representations from DABFS. The DABFS incorporates a transformation inconsistency augmentation (TIA) module with a tempo-scale strategy for data augmentation. In the RFS, features from raw data are also extracted and used in the contrasting module alongside augmented data to create negative and positive pairs for DAB awareness. To minimize the DAB in the learned representations, we design DAB-minimizing temporal and contextual contrasting modules (DABa-TC and DABa-CC) that operate in higher dimensions and leverage the assistance of RFS to extract temporal and contextual features. This work is the first to address the DAB problem and propose a method to minimize it in learned representations.
In summary, the main contributions of this work are the following:
  • We introduce the concept of the data augmentation bias problem (DAB) and present the DABaCLT framework to recognize and minimize DAB during the learning of time series representations. The RFS is developed as a criterion for perceiving DAB and is effectively combined with the DABFS.
  • We propose DAB-minimizing contrastive loss (DABMinLoss) in the contrasting module to reduce the DAB in extracted temporal and contextual features. The complete DABMinLoss consists of five parts implemented in the DABa-TC and DABa-CC modules.
  • Extensive experiments demonstrate the superiority of DABaCLT over previous works. Compared to self-supervised learning, DABaCLT achieves significant improvements ranging from 0.19% to 22.95% in sleep staging classification, 2.96% to 5.05% in human activity recognition, and 1.00% to 2.46% in epilepsy seizure prediction. The results also compete favorably with supervised methods.

2. Materials and Methods

2.1. Data Description

Three real-world time series tasks were utilized for method evaluation, including sleep stage classification (SSC), human activity recognition (HAR), and epilepsy seizure prediction (ESP). We compared our performances with state-of-the-art approaches on three publicly available datasets. Table 1 summarizes the details of each dataset, e.g., the number of training samples (Train), validation samples (Valid), and testing samples (Test), the length of the sample (Length), the number of sensor channels (Channel), and the number of classes (Class).
Sleep staging classification is a task that involves classifying EEG signals into one of five sleep stages: wake (W), non-rapid eye movement (N1, N2, N3), and rapid eye movement (REM). For this task, we utilized the Sleep-EDF dataset [23]. The Sleep-EDF dataset contains 42,308 epochs in 39 sleep cassette files collected from 20 subjects aged 25–34. Each subject in the Sleep-EDF database contains 2 day–night PSG recordings, except subjects 13, 36, and 52, whose 1 recording was lost due to device failure. The duration of each epoch is 30 s, and it has been labeled as wake, REM, N1, N2, N3, N4, MOVEMENT, UNKNOWN by experts according to the R&K standard. Specifically, we focused on a single EEG channel (Fpz-Cz) sampled at a rate of 100 Hz, following the methodology established in previous studies [21,24].
Human activity recognition is concerned with the classification of sensor readings into one of six activities: walking, walking upstairs, walking downstairs, standing, sitting, and lying down. To address this task, we utilized the UCI HAR dataset [25]. Human activity recognition contains 8828 epochs from 30 participants aged 19–48 years. It is concerned with the classification of sensor readings into one of six activities: walking, walking upstairs, walking downstairs, standing, sitting, and lying down. Each sample selects data in a fixed time window of 2.56 s, which is 128 sampling points.
Epilepsy seizure prediction involves predicting the occurrence of epileptic seizures using EEG recordings. For this task, we leveraged the Epileptic Seizure Recognition (ESR) dataset [26], which consists of EEG recordings obtained from 500 subjects. Each subject’s brain activity was recorded for a duration of 23.6 s. As the dataset originally included multiple seizure-related classes, we merged four out of the five classes into a single class as they did not include epileptic seizures. Consequently, the task was transformed into a binary classification problem.

2.2. The Proposed Method

2.2.1. Transformation Inconsistency Augmentation

For contrastive learning, the construction of positive and negative pairs is mainly based on the design of the augmentation method. The original time series is augmented by two random operations. Usually, the operations are selected from the same augmentation algorithm, such as dropout [11], timestamp masking, and cropping [16], to preserve the global temporal information and maintain the original properties for time series. Different augmentations can improve the robustness of the learned representations and reduce the training time for contrastive loss convergence. In this paper, the TIA module is designed in DABFS to learn invariant representations. To obtain more discriminating representations, a tempo-scale strategy is adopted. The scale augmentation lightly scales up the magnitude of the time series with random variations. For temporal augmentation, a permutation-and-jitter strategy is applied, which randomly splits the time series into a set of segments, randomly shuffles segments, and adds random variations. The max segment number N s is a hyper-parameter. The augmentation operations do not change the dimensions of the time series. The number of segments and parameters of random variations are selected based on previous research on augmentation.
Given an original time series x R S , the augmentation can be described as x t = A t ( x ) and x s = A s ( x ) R S ,where S represents the number of samples. The tempo-scale strategy results in different similarities between the transformed and the original time series. As shown in Figure 1, the transformed time series x t and x s are fed into a basic encoder module to extract their high-dimensional space representation, z t and z s . The encoder adopts a 3-block convolutional architecture as in previous studies. The complexity of the encoder is controlled by hyper-parameters to ensure that representation and generalization capabilities are balanced. The encoding operation is defined as Z = F e n c ( x ) , Z = { z i R d , i T } , where d is the feature length of each step and T is the time steps of each time series representation. Meanwhile, the original time series x is also encoded as z u for following bias awareness. The z s , z t , and z u are fed into DAB-aware Temporal Contrasting module.

2.2.2. DAB-minimize Contrasting Learning

The DAB-minimize contrasting module algorithm, presented in Algorithm 1, comprises two submodules: DABa-TC and DABa-CC. These modules aim to perceive and minimize the data augmentation bias (DAB) by optimizing the prediction performance and vector similarity with the assistance of the raw features stream (RFS). The loss functions are designed for both the min and max options.
Algorithm 1 DAB-aware Contrasting module algorithm.
  • Input:  { z n t } n = 1 N , { z n s } n = 1 N , { z n u } n = 1 N , K
  • define:
  • F t r a n s : the transformer encoder
  • T k : a set of linear layers
  • P r o j e c t o r : a non-linear projection head
  • T: the time steps of each time series representation
  • B: the batch size
  • d i s ( y , y ) : the difference between prediction output y and ground truth y
  • sim c ( u , v ) : the cosine similarity between u and v
  • sim d ( u , v ) : the dot similarity between u and v
  • Initialize:
  • randomly select: t from 1 to ( T K )
  • Z n , t t = { z n , i t , i t } , Z n , t s = { z n , i s , i t }
  • for all n { 1 , N } do
  •     c n , t t = F t r a n s ( Z n , t t ) ,   c n , t s = F t r a n s ( Z n , t s ) ,   c n , t u = F t r a n s ( Z n , t u )
  •     p n t = P r o j e c t o r ( c n , t t ) ,   p n s = P r o j e c t o r ( c n , t s ) ,   p n u = P r o j e c t o r ( c n , t u )
  •    for 1 k K do
  •       α n , k t = T k ( c n , t t ) ,   α n , k s = T k ( c n , t s )
  •      min d i s ( α n , k t , z n , t + k u ) + d i s ( α n , k t , z n , t + k s ) , min d i s ( α n , k s , z n , t + k u ) + d i s ( α n , k s , z n , t + k t )
  •      max m n ( d i s ( α n , k t , z m u ) + d i s ( α n , k t , z m s ) ) , max m n ( d i s ( α n , k s , z m u ) + d i s ( α n , k s , z m t ) )
  •    end for
  •    max s i m ( p n t , p n s ) , s i m ( p n t , p n u ) , s i m ( p n s , p n u )
  •    min m n ( s i m c ( p n t , p m s ) + s i m c ( p n s , p m t ) ) , min m n ( s i m d ( p n t , p m u ) + s i m d ( p n s , p m u ) )
  • end for

DAB-Aware Temporal Contrasting

In contrast to image and natural language data, temporal feature awareness is crucial for time series representation. To enhance the temporal aspect of the final representation, the DABa-TC module was designed as shown in the Figure 2. This module incorporates two attention blocks as a temporal encoding model, inspired by the transformer encoder F t r a n s , known for its superior capability in extracting temporal features. The time steps Z t s = z i s , i t and Z t t = z i t , i t are temporally represented as c t s and c t t R m , respectively, where m denotes the dimension of the hidden layers in the transformer.
The encoded c t is then used to predict the future time steps z t , K = { z t + k , 1 k K } , where t is randomly selected from 1 to T K , and K is a hyper-parameter. Inspired by TS-TCC, the cross-view prediction is adopted. The temporal representation c t s is used to predict the scale future time steps z t , K t , and vice versa. Meanwhile, to minimize the DAB, both the c t s and c t t are used to predict raw future time steps z t , K u . Specifically, c t is mapped to have the same dimension as z t by a set of linear layers T k .
DABa-TC aims to optimize the prediction performance for the positive pairs, minimize the performance for negative pairs, and minimize the DAB of extracted temporal features. For the negative pairs, z generated by other samples within the same minibatch are defined as z n . Thus, the loss functions L T C s and L T C t can be formed as follows:
L T C s = 1 K k = 1 K l o g e x p ( T k ( c t s ) T z t + k t + T k ( c t s ) T z t + k u ) n e x p ( T k ( c t s ) T z n t + T k ( c t s ) T z n u )
L T C t = 1 K k = 1 K l o g e x p ( T k ( c t t ) T z t + k s + T k ( c t t ) T z t + k u ) n e x p ( T k ( c t t ) T z n s + T k ( c t t ) T z n u )

DAB-Aware Contextual Contrasting

As shown in the Figure 2. To generate contextual information, a non-linear projection head [13] is applied to the temporal representations c t t , contextual representations c t s , and raw representations c t u , resulting in projected representations p t = P r o j ( c t ) R m 4 . Figure 3 illustrates the formation of positive and negative pairs. Positive pairs consist of the representations obtained from the same sample, but are transformed using temporal and scale augmentation. Negative pairs, on the other hand, are formed by taking the representations transformed from different samples.
In the TIA module, 2 N augmented samples are generated from N time series samples. Traditionally, contrastive learning methods only utilize the 2 N augmented samples for encoding, resulting in one positive pair and 2 ( N 1 ) negative pairs for each sample. In contrast, DABaCLT takes into account the raw N samples and feeds them into the encoder concurrently with the augmented samples. Similarity-based techniques have been applied in various time series classification tasks and other applications [27,28].
In the contrasting module, assessing the similarity between the temporal- and scale-augmented samples is crucial for learning discriminative representations. To maximize the similarity of positive pairs and minimize the similarity of negative pairs, the loss function L C C s w is defined as follows:
L C C st = i = 1 N b l o g e x p ( sim ( ( p i s , p i t ) / τ ) k i , a { s , t } e x p ( sim ( ( p i s , p k a ) / τ )
The sim ( u , v ) = u T v / u v denotes the dot product between L 2 -normalized u and v (i.e., cosine similarity), N b is the sample size of a minibatch, and τ is the temperature parameter to help the model learn from hard negatives. Furthermore, evaluating the similarity between raw and augmented samples makes the representation closer to the original data distribution, thereby minimizing the DAB of extracted contextual features. This similarity is optimized by L C C t and L C C t as follows:
L C C s = i = 1 N b l o g e x p ( sim ( ( p i s , p i u ) / τ ) k i e x p ( sim ( ( p i s , p k u ) / τ )
L C C t = i = 1 N b l o g e x p ( sim ( ( p i t , p i u ) / τ ) k i e x p ( sim ( ( p i t , p k u ) / τ )
For Equations (4) and (5), sim ( u , v ) = u T v denotes the dot product (i.e., dot similarity). The DABMinLoss is the combination of the L T C and L C C as follows:
L = λ 1 · L T C s + λ 2 · L T C t + λ 3 · L C C st + λ 4 · ( L C C t + L C C s )
where λ 1 , λ 2 , λ 3 , and λ 4 are fixed scalar hyper-parameters denoting the relative weight of each loss. DABMinLoss minimizes the DAB of extracted temporal and contextual features. Therefore, the learned representation has lower DAB, which can achieve better performance in the downstream task.

3. Results

Comparison with Baseline Approaches

The performances of DABaCLT are compared with the baseline approaches: (1) Supervised: supervised training of both encoder and classifier models; (2) SimCLR [13]; (3) TS-TCC [21]; and (4) TS2Vec [16]. The publicly available or re-implemented codes were used for the baseline models. Brief descriptions of the models are as follows:
simCLR learns representations by maximizing the agreement between different augmented views of the same data example through contrastive loss in latent space. Two different data augmentation operators are sampled from the same augmentation family.
TS-TCC encourages the consistency of different data augmentations to learn trans-formation-invariant representations. It learns robust representations from time series data through a difficult cross-view prediction task.
TS2Vec uses the hierarchical contrast method to capture multi-scale context information in the instance time dimension. It can learn the context representations of any subsequence at different semantic levels.
For fair comparisons, we further trained a linear classifier on the learned representations. Two matrices were used to evaluate the performance of the sleep stage models, namely, the prediction accuracy (ACC) and the macro-averaged F1 score (MF1). We implemented the TS-TCC, SimCLR, and supervised methods through a common code with the same encoder architecture and parameters. The original encoder architecture of TS2Vec was used because the TS2Vec has a specially designed encoder architecture for contrastive learning. To evaluate the performance and robustness of the proposed DAB-aware contrastive learning framework (DABaCLT) for classification tasks, we adopted a rigorous evaluation strategy using 10-fold cross-validation. The available labeled time series data were divided into 10 equal-sized subsets. The training and validation procedure was then repeated 10 times, with each fold serving as the validation set once and the remaining nine folds used for training. The more detailed experimental setup is shown in Appendix A.
Table 2 shows that with only a linear classifier, DABaCLT performs best on two of three datasets, while achieving comparable performance to the supervised approach on the third dataset. Compared with self-supervised learning, the average accuracy of DABaCLT achieves a 0.19% to 22.95% improvement for SSC, 2.96% to 5.05% for HAR, and 1% to 2.46% for ESP. The results demonstrate that DABaCLT adequately learns temporal and contextual information in time series, and further perceives and minimizes the DAB between the augmented and raw representations. As shown in Figure A1, the performance of DABaCLT can be observed in the confusion matrix for each category. More detailed results with precision, recall, sensitivity, and specificity are presented in the table in the Appendix B.

4. Discussion

4.1. Comparisons of Time Series Augmentation Methods

To further prove the effectiveness of our TIA (tempo-scale strategy) module, we compared our method with the other four augmentation strategies: slice, scaling, dropout, and permutation. The classification accuracy comparisons of different augmentations on the HAR dataset are illustrated in Figure 4A. The proposed tempo-scale strategy has the best performance in average accuracy, which demonstrates that the discriminating representations are learned using tempo-scale transformation augmentation. Further, considering the random transformation with the same augmentation method generates two similar views, our TIA generates views with more information entropy. Previous studies consider the dropout strategy to be the best augmentation. However, when the DAB is aware, more discriminative features are constructed, and the model can achieve better performance. Though the augmentation process introduces the DAB, it can be perceived and minimized by the DABa-TC and DABa-CC modules, which means the TIA is more stable for the DABaCLT framework.

4.2. Impact of Raw Features Stream

RFS is the essential part of DABaCLT, which includes the raw time series x u , encoded representation z u , and the loss functions L T C in the DABa-TC and L C C DABa-CC modules. We argue that the DAB exists in different levels of features and can be minimized by different raw features (UF). To evaluate whether the RFS does assist the DAB awareness in learned representations, we conducted ablation experiments on the HAR dataset. We compared four situations: (1) drop total RFS, (2) calculate L T C s and L T C t without z u (drop UF in DABa-TC), (3) drop L C C s and L C C t (drop UF in DABa-CC), and (4) complete DABaCLT (no drops). Figure 4B shows the classification performance when dropping different raw features. When dropping the total RFS, the framework can still learn temporal and contextual information. However, the classification performance is not as good as using z u or calculating L C C s and L C C s . Moreover, the complete DABaCLT achieves the best average accuracy, which means the RFS is greatly meritorious in terms of DAB awareness and minimization.

4.3. Impact of Parameter Selection

The sensitivity analysis was performed on the HAR dataset to study the influence of parameters λ 1 , λ 2 , λ 3 , and λ 4 on the contrasting learning process. Figure 5 shows the evaluation results. When evaluating λ i , the others are fixed to 1. It is observed that the model is not sensitive to λ 1 , and it is easier for λ 2 , λ 3 , and λ 4 to influence the representation. Commonly, the model achieves the best performance when 0.1 < λ s < 10 .

4.4. The Visualization of DAB Problem

The distribution of raw signals and augmented signals from the three datasets is shown in Figure 6. It can be observed that the HAR and ESR datasets exhibit a more chaotic distribution after applying data augmentation methods. However, there is no significant change in the distribution of the Sleep-EDF data. The spatial distribution of different categories in the Sleep-EDF dataset appears to be more uniform compared to the other two datasets. This can be mainly attributed to the sample length of the data. For the HAR and ESR datasets, the number of sampling points in their individual samples is 128 and 178, respectively, making the tempo-scale augmentation method significantly alter the data distribution. On the other hand, the Sleep-EDF dataset has a sample length of 3000, resulting in a less-pronounced disruption of its distribution by the tempo-scale augmentation method. The accuracy results further support this observation, as DABaCLT demonstrates improved prediction accuracy for the HAR and Epilepsy datasets compared to supervised methods, while showing no improvement for SSC.

5. Conclusions

This paper introduces the concept of the DAB and presents a DAB-aware contrastive learning framework for time series representations. It incorporates DABFS and RFS, and utilizes DAB-aware temporal and contextual contrasting modules to minimize DAB through DAB-aware contrastive loss. The experimental results demonstrate the effectiveness of the framework, with improved accuracy compared to self-supervised learning. The tempo-scale strategy and RFS play important roles in enhancing the discriminative power of the learned representations and addressing the data augmentation bias. The DAB problem proposed in this article is a common phenomenon in contrastive learning. The ideas of RFS and DABmin Loss proposed in this article provide a solution for more contrastive learning methods. However, the current research has not yet addressed the impact of data length on the model, and the complexity of the model needs further optimization. The influence of sample length and the quantification of DAB will be explored in future research.

Author Contributions

Conceptualization, Y.Z.; methodology, Y.Z. and Y.L.; software, Y.Z.; validation, Y.Z., Y.L. and H.S.; formal analysis, Y.Z. and Y.L.; investigation, Y.Z. and H.S.; resources, Y.Z. and Y.L.; data curation, Y.L.; writing—original draft preparation, Y.Z.; writing—review and editing, Y.Z.; visualization, Y.Z.; supervision, L.Z. and L.L.; project administration, L.Z. and L.L.; funding acquisition, L.Z. and L.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the National Natural Science Foundation of China (Grants No. 61971056, 62176024); the National Key R&D Program of China (2022ZD01161); Beijing Municipal Science & Technology Commission (Grant No. Z181100001018035); project LB20201A010010; and the Engineering Research Center of Information Networks, Ministry of Education.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable. This study did not involve humans, all the data have been obtained from the open access database.

Data Availability Statement

Publicly available datasets were analyzed in this study. The Sleep-EDF can be found at https://www.physionet.org/content/sleep-edfx/1.0.0/ (accessed on 1 November 2022), the UCI HAR dataset can be found at https://archive.ics.uci.edu/ml/datasets/Epileptic+Seizure+Recognition (accessed on 1 November 2022), and the Epilepsy dataset can be found at https://archive.ics.uci.edu/ml/datasets/Epileptic+Seizure+Recognition (accessed on 1 November 2022).

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Experimental Setup

For the DABaCLT, the Adam optimizer with the learning rate of 3 × 10 4 was applied. The weight decay of Adam was set to 3 × 10 4 , the betas (b1, b2) were used as (0.9, 0.999), respectively, and the epsilon value was 1 × 10 8 . The pre-training and fine-tune epochs are 40.

Appendix B. Detailed Results

Table A1. The detailed results of DABaCLT for epilepsy seizure prediction.
Table A1. The detailed results of DABaCLT for epilepsy seizure prediction.
PrecisionSensitivitySpecificity
ESP98.49 ± 0.3399.29 ± 0.3193.98 ± 1.39
Table A2. The detailed results of DABaCLT for human activity recognition.
Table A2. The detailed results of DABaCLT for human activity recognition.
ClassWalkingWalking UpstairsWalking Downstairs
recall96.57 ± 2.9294.48 ± 5.9197.86 ± 2.39
classstandingsittinglying down
recall78.82 ± 2.6487.41 ± 2.2997.58 ± 0.39
Table A3. The detailed results of DABaCLT for sleep staging classification.
Table A3. The detailed results of DABaCLT for sleep staging classification.
ClassWN1N2N3REM
recall94.34 ± 1.7528.84 ± 5.5181.80 ± 1.9387.02 ± 1.5682.56 ± 2.79
Figure A1. Confusion matrix result of DABaCLT for (A) SSC, (B) ESP, and (C) HAR.
Figure A1. Confusion matrix result of DABaCLT for (A) SSC, (B) ESP, and (C) HAR.
Applsci 13 07908 g0a1

References

  1. Torres, J.F.; Hadjout, D.; Sebaa, A.; Martínez-Álvarez, F.; Troncoso, A. Deep Learning for Time Series Forecasting: A Survey. Big Data 2021, 9, 3–21. [Google Scholar] [CrossRef] [PubMed]
  2. Xu, J.; Ding, X.; Ke, H.; Xu, C.; Zhang, H. Student Behavior Prediction of Mental Health Based on Two-Stream Informer Network. Appl. Sci. 2023, 13, 2371. [Google Scholar] [CrossRef]
  3. Sharma, S.; Khare, S.K.; Bajaj, V.; Ansari, I.A. Improving the separability of drowsiness and alert EEG signals using analytic form of wavelet transform. Appl. Acoust. 2021, 181, 108164. [Google Scholar] [CrossRef]
  4. Sun, Z.; Ke, Q.; Rahmani, H.; Bennamoun, M.; Wang, G.; Liu, J. Human action recognition from various data modalities: A review. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 3200–3225. [Google Scholar]
  5. Jana, R.; Mukherjee, I. Deep learning based efficient epileptic seizure prediction with EEG channel optimization. Biomed. Signal Process. Control 2021, 68, 102767. [Google Scholar]
  6. Khare, S.K.; Nishad, A.; Upadhyay, A.; Bajaj, V. Classification Of Emotions From Eeg Signals Using Time-order Representation Based On the S-transform And Convolutional Neural Network. Electron. Lett. 2020, 56, 1359–1361. [Google Scholar] [CrossRef]
  7. Mohsenvand, M.N.; Izadi, M.R.; Maes, P. Contrastive Representation Learning for Electroencephalogram Classification. In Proceedings of the Machine Learning for Health NeurIPS Workshop; Alsentzer, E., McDermott, M.B.A., Falck, F., Sarkar, S.K., Roy, S., Hyland, S.L., Eds.; 2020; Volume 136, pp. 238–253. Available online: proceedings.mlr.press/v136/mohsenvand20a.html (accessed on 13 May 2023).
  8. Jiang, X.; Zhao, J.; Du, B.; Yuan, Z. Self-supervised Contrastive Learning for EEG-based Sleep Staging. In Proceedings of the 2021 International Joint Conference on Neural Networks (IJCNN), Shenzhen, China, 18–22 July 2021; pp. 1–8. [Google Scholar] [CrossRef]
  9. Zhang, Z.; Zhong, S.h.; Liu, Y. GANSER: A Self-supervised Data Augmentation Framework for EEG-based Emotion Recognition. IEEE Trans. Affect. Comput. 2022. [Google Scholar] [CrossRef]
  10. Levasseur, G.; Bersini, H. Time Series Representation for Real-World Applications of Deep Neural Networks. In Proceedings of the 2022 International Joint Conference on Neural Networks (IJCNN), Padua, Italy, 18–23 July 2022; pp. 1–8. [Google Scholar] [CrossRef]
  11. Yang, L.; Hong, S. Unsupervised Time-Series Representation Learning with Iterative Bilinear Temporal-Spectral Fusion. In Proceedings of the 39th International Conference on Machine Learning, Baltimore, MD, USA, 17–23 July 2022; Volume 162, pp. 25038–25054. [Google Scholar]
  12. Zerveas, G.; Jayaraman, S.; Patel, D.; Bhamidipaty, A.; Eickhoff, C. A transformer-based framework for multivariate time series representation learning. In Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining, Virtual Event, Singapore, 14 –18 August 2021; pp. 2114–2124. [Google Scholar]
  13. Chen, T.; Kornblith, S.; Norouzi, M.; Hinton, G. A simple framework for contrastive learning of visual representations. In Proceedings of the International Conference on Machine Learning, Virtual Event, 13–18 July 2020; pp. 1597–1607. [Google Scholar]
  14. Rethmeier, N.; Augenstein, I. A Primer on Contrastive Pretraining in Language Processing: Methods, Lessons Learned and Perspectives. ACM Comput. Surv. 2022, 55, 1–17. [Google Scholar] [CrossRef]
  15. Sun, L.; Yolwas, N.; Jiang, L. A Method Improves Speech Recognition with Contrastive Learning in Low-Resource Languages. Appl. Sci. 2023, 13, 4836. [Google Scholar] [CrossRef]
  16. Yue, Z.; Wang, Y.; Duan, J.; Yang, T.; Huang, C.; Tong, Y.; Xu, B. TS2Vec: Towards Universal Representation of Time Series. arXiv 2021. [Google Scholar] [CrossRef]
  17. Ozyurt, Y.; Feuerriegel, S.; Zhang, C. Contrastive Learning for Unsupervised Domain Adaptation of Time Series. arXiv 2022, arXiv:2206.06243. [Google Scholar]
  18. Nonnenmacher, M.T.; Oldenburg, L.; Steinwart, I.; Reeb, D. Utilizing Expert Features for Contrastive Learning of Time-Series Representations. In Proceedings of the 39th International Conference on Machine Learning, Baltimore, MD, USA, 17–23 July 2022; 162, pp. 16969–16989. [Google Scholar]
  19. Franceschi, J.Y.; Dieuleveut, A.; Jaggi, M. Unsupervised Scalable Representation Learning for Multivariate Time Series. In Proceedings of the Advances in Neural Information Processing Systems, Vancouver, BC, Canada, 8–14 December 2019; Wallach, H., Larochelle, H., Beygelzimer, A., d’Alché-Buc, F., Fox, E., Garnett, R., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2019; Volume 32. [Google Scholar]
  20. Zhang, X.; Zhao, Z.; Tsiligkaridis, T.; Zitnik, M. Self-supervised contrastive pre-training for time series via time-frequency consistency. arXiv 2022, arXiv:2206.08496. [Google Scholar]
  21. Eldele, E.; Ragab, M.; Chen, Z.; Wu, M.; Kwoh, C.K.; Li, X.; Guan, C. Time-Series Representation Learning via Temporal and Contextual Contrasting. In Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence, IJCAI-21, Montreal, QC, Cananda, 19–27 August 2021; pp. 2352–2359. [Google Scholar]
  22. Deldari, S.; Xue, H.; Saeed, A.; He, J.; Smith, D.V.; Salim, F.D. Beyond Just Vision: A Review on Self-Supervised Representation Learning on Multimodal and Temporal Data. arXiv 2022, arXiv:2206.02353. [Google Scholar]
  23. Goldberger, A.; Amaral, L.; Glass, L.; Hausdorff, J.; Ivanov, P.C.; Mark, R.; Mietus, J.; Moody, G.; Peng, C.; Stanley, H. PhysioBank, PhysioToolkit, and PhysioNet: Components of a new research resource for complex physiologic signals. Circulation 2000, 101, E215–E220. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  24. Yubo, Z.; Yingying, L.; Bing, Z.; Lin, Z.; Lei, L. MMASleepNet: A multimodal attention network based on electrophysiological signals for automatic sleep staging. Front. Neurosci. 2022, 16, 973761. [Google Scholar] [CrossRef] [PubMed]
  25. Anguita, D.; Ghio, A.; Oneto, L.; Parra Perez, X.; Reyes Ortiz, J.L. A public domain dataset for human activity recognition using smartphones. In Proceedings of the 21th International European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning, Bruges, Belgium, 5–7 October 2013; pp. 437–442. [Google Scholar]
  26. Andrzejak, R.G.; Lehnertz, K.; Mormann, F.; Rieke, C.; David, P.; Elger, C.E. Indications of nonlinear deterministic and finite-dimensional structures in time series of brain electrical activity: Dependence on recording region and brain state. Phys. Rev. E 2001, 64, 061907. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  27. Zha, D.; Lai, K.H.; Zhou, K.; Hu, X. Towards similarity-aware time-series classification. In Proceedings of the 2022 SIAM International Conference on Data Mining (SDM), SIAM, Alexandria, VA, USA, 28–30 April 2022; pp. 199–207. [Google Scholar]
  28. Jastrzebska, A.; Nápoles, G.; Salgueiro, Y.; Vanhoof, K. Evaluating time series similarity using concept-based models. Knowl.-Based Syst. 2022, 238, 107811. [Google Scholar] [CrossRef]
Figure 1. Illustration of the DAB problem: (A) Supervised representation learning, where representations are learned using labeled data as the ground truth. (B) Traditional self-supervised representation learning. Due to the changes in data distribution caused by data augmentation methods, the representations of real data obtained from a pre-trained encoder deviate from the ground truth, resulting in the DAB problem. (C) The proposed self-supervised representation learning approach to mitigate the DAB problem.
Figure 1. Illustration of the DAB problem: (A) Supervised representation learning, where representations are learned using labeled data as the ground truth. (B) Traditional self-supervised representation learning. Due to the changes in data distribution caused by data augmentation methods, the representations of real data obtained from a pre-trained encoder deviate from the ground truth, resulting in the DAB problem. (C) The proposed self-supervised representation learning approach to mitigate the DAB problem.
Applsci 13 07908 g001
Figure 2. The architecture of DABaCLT. The encoder, attention, and projector blocks in the raw features stream (RFS) share weights with the corresponding blocks in the DAB features stream (DABFS). The DABa-TC and DABa-CC modules learn temporal and contextual features and minimize the DAB with DABMinLoss, which includes L T C s , L T C t , L C C s , L C C t , and L C C st .
Figure 2. The architecture of DABaCLT. The encoder, attention, and projector blocks in the raw features stream (RFS) share weights with the corresponding blocks in the DAB features stream (DABFS). The DABa-TC and DABa-CC modules learn temporal and contextual features and minimize the DAB with DABMinLoss, which includes L T C s , L T C t , L C C s , L C C t , and L C C st .
Applsci 13 07908 g002
Figure 3. The definition of constructing positive pairs and negative pairs. The color represents augmentation type (scale, temporal, or no augmentation), and the shape represents the different samples.
Figure 3. The definition of constructing positive pairs and negative pairs. The color represents augmentation type (scale, temporal, or no augmentation), and the shape represents the different samples.
Applsci 13 07908 g003
Figure 4. Classification performance on HAR dataset with (A) different augmentation strategy, # represents that the two views are transformed with the same augmentation methods, and (B) dropping different raw features.
Figure 4. Classification performance on HAR dataset with (A) different augmentation strategy, # represents that the two views are transformed with the same augmentation methods, and (B) dropping different raw features.
Applsci 13 07908 g004
Figure 5. Classification performance on HAR dataset of different λ 1 , λ 2 , λ 3 , and λ 4 with different values.
Figure 5. Classification performance on HAR dataset of different λ 1 , λ 2 , λ 3 , and λ 4 with different values.
Applsci 13 07908 g005
Figure 6. Spatial distribution visualization of signals with T-SNE method. Spatial distribution of (A1) the raw, (A2) scale-augmented, and (A3) tempo-augmented signals in HAR dataset; spatial distribution of (B1) the raw, (B2) scale-augmented and (B3) tempo-augmented signals in ESR dataset; spatial distribution of (C1) the raw, (C2) scale-augmented, and (C3) tempo-augmented signals in SleepEDF dataset.
Figure 6. Spatial distribution visualization of signals with T-SNE method. Spatial distribution of (A1) the raw, (A2) scale-augmented, and (A3) tempo-augmented signals in HAR dataset; spatial distribution of (B1) the raw, (B2) scale-augmented and (B3) tempo-augmented signals in ESR dataset; spatial distribution of (C1) the raw, (C2) scale-augmented, and (C3) tempo-augmented signals in SleepEDF dataset.
Applsci 13 07908 g006
Table 1. Description of datasets used in our experiments. The train and valid data are randomly divided 10 times with 10-fold validation.
Table 1. Description of datasets used in our experiments. The train and valid data are randomly divided 10 times with 10-fold validation.
Dataset#Train#Valid#TestLength#Channel#Class
Sleep-EDF31,95235516805300015
HAR6616736294712896
ESR8280920230017812
Table 2. Comparison between DABaCLT model against baselines using linear classifier evaluation experiments.
Table 2. Comparison between DABaCLT model against baselines using linear classifier evaluation experiments.
MethodSSCHARESP
Avg. AccMF1Avg. AccMF1Avg. AccMF1
supervised81.14 ± 0.3674.07 ± 0.3391.47 ± 0.7591.09 ± 0.8396.63 ± 0.3194.56 ± 0.54
simCLR56.58 ± 0.4448.81 ± 0.1386.99 ± 0.1886.05 ± 0.2096.00 ± 0.2093.33 ± 0.35
TS-TCC79.34 ± 0.3372.92 ± 0.4989.08 ± 0.1988.45 ± 0.1995.59 ± 0.1392.62 ± 0.21
TS2Vec73.50 ± 1.4865.93 ± 0.2587.91 ± 0.0286.55 ± 0.2297.05 ± 0.3694.63 ± 0.21
DABaCLT (ours)79.53 ± 0.1173.25 ± 0.1592.04 ± 0.2491.52 ± 0.2698.05 ± 0.1096.93 ± 0.18
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zheng, Y.; Luo, Y.; Shao, H.; Zhang, L.; Li, L. DABaCLT: A Data Augmentation Bias-Aware Contrastive Learning Framework for Time Series Representation. Appl. Sci. 2023, 13, 7908. https://doi.org/10.3390/app13137908

AMA Style

Zheng Y, Luo Y, Shao H, Zhang L, Li L. DABaCLT: A Data Augmentation Bias-Aware Contrastive Learning Framework for Time Series Representation. Applied Sciences. 2023; 13(13):7908. https://doi.org/10.3390/app13137908

Chicago/Turabian Style

Zheng, Yubo, Yingying Luo, Hengyi Shao, Lin Zhang, and Lei Li. 2023. "DABaCLT: A Data Augmentation Bias-Aware Contrastive Learning Framework for Time Series Representation" Applied Sciences 13, no. 13: 7908. https://doi.org/10.3390/app13137908

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop