Next Article in Journal
Fingerprint Classification through Standard and Weighted Extreme Learning Machines
Next Article in Special Issue
Missing Value Imputation in Stature Estimation by Learning Algorithms Using Anthropometric Data: A Comparative Study
Previous Article in Journal
Understanding Digital Radio Frequency Memory Performance in Countermeasure Design
Previous Article in Special Issue
FedOpt: Towards Communication Efficiency and Privacy Preservation in Federated Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Time-Series Classification Based on Fusion Features of Sequence and Visualization

1
The Xinjiang Technical Institute of Physics & Chemistry, CAS, Urumqi 830011, China
2
University of Chinese Academy of Sciences, Beijing 100049, China
3
Xinjiang Laboratory of Minority Speech and Language Information Processing, Urumqi 830011, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2020, 10(12), 4124; https://doi.org/10.3390/app10124124
Submission received: 16 May 2020 / Revised: 4 June 2020 / Accepted: 11 June 2020 / Published: 15 June 2020
(This article belongs to the Special Issue Advances in Deep Learning Ⅱ)

Abstract

:
For the task of time-series data classification (TSC), some methods directly classify raw time-series (TS) data. However, certain sequence features are not evident in the time domain and the human brain can extract visual features based on visualization to classify data. Therefore, some researchers have converted TS data to image data and used image processing methods for TSC. While human perceptionconsists of a combination of human senses from different aspects, existing methods only use sequence features or visualization features. Therefore, this paper proposes a framework for TSC based on fusion features (TSC-FF) of sequence features extracted from raw TS and visualization features extracted from Area Graphs converted from TS. Deep learning methods have been proven to be useful tools for automatically learning features from data; therefore, we use long short-term memory with an attention mechanism (LSTM-A) to learn sequence features and a convolutional neural network with an attention mechanism (CNN-A) for visualization features, in order to imitate the human brain. In addition, we use the simplest visualization method of Area Graph for visualization features extraction, avoiding loss of information and additional computational cost. This article aims to prove that using deep neural networks to learn features from different aspects and fusing them can replace complex, artificially constructed features, as well as remove the bias due to manually designed features, in order to avoid the limitations of domain knowledge. Experiments on several open data sets show that the framework achieves promising results, compared with other methods.

1. Introduction

Time-series (TS) data is a set of values sequentially ordered in time, which is seen frequently in real-life, such as financial data, trajectory data, weather data, and so on. With the development and application of the Internet of Things (IoT), the data collected by various sensors is also TS data. Research on TS data is diverse, such as compression, storage and query, anomaly detection, prediction, and so on [1]. This paper focuses on the time-series classification (TSC) task, the purpose of which is to classify concrete TS data into pre-determined categories which have similar characteristics.
For the task of TSC, our predecessors have done a lot of research and produced many methods. These methods are mainly based on distance functions, such as dynamic time warping (DTW) [2]; features, such as the shapelet transform (ST) [3]; and ensemble methods, such as the hierarchical vote COTE (HIVE-COTE) [4], combining the former two types of methods.
With the development and application of deep learning, a large number of methods for this task have also been proposed. Researchers typically use various transformation methods or artificially constructed features to input into deep neural networks (DNN) and train a model for TSC; yet, some methods directly use DNN, such as long short-term memory (LSTM) [5], to process raw TS data and learn an end-to-end classification model.
Human beings can use visual observation to distinguish the differences between TS records, according to key features, and classify them. Visualization is also very important in scientific research. In the field of natural language processing (NLP), Chinese character representations have been improved by using the pictographic meanings of Chinese characters [6]. In the field of TS research, features extracted by visualization have been used for the task of prediction [7]. Among deep learning methods for TSC, a type of method converts TS data into image data, then use some image processing methods to classify the TS data. Representative transformation methods include Gramian fields [8,9], recurrence plots (RP) [10,11], Markov transition fields (MTF) [12], and so on.
Usually, a human’s perception of things is a combination of human senses from different aspects, while the deep learning methods for TSC proposed so far have either used only sequential features or visualization features. In the field of few-shot learning, scholars have used DNNs to extract data features from different perspectives, in order to improve the performance of deep learning models on small training data sets [13,14].
Enlightened by these advances, we propose a framework, TSC-FF, based on deep learning to learn features from different aspects, which enhances the feature space and replaces complex, artificially constructed features, in order to remove the bias due to manually designed features, as well as to avoid the limitations of domain knowledge. Specifically, in order to enable the network to learn the most discriminant useful features for the classification task, we use well-trained LSTM-A neural networks and CNN-A neural networks to extract the features of TS data. LSTM-A is used to extract sequence features, while CNN-A is used to extract visualization features after the TS data is visualized.
Our approach differs from the existing visual transformation methods, such as Gramian fields [8,9], RP [10,11], and MTF [12]. In order to imitate human vision (i.e., what we see is what we get), we take the most direct visualization method, transforming the TS data into an Area Graph. Area Graphs are a type of Line Graph, but with the area below the line filled in with a certain colour or texture. Compared with the existing visualization methods, Area Graph is simpler and avoids the loss of some information in complex conversions and additional computational cost. In addition, an attention mechanism is used to enhance the features of key points, which is generated from contributing regions in the data for the specific labels.
In this paper, experiments are carried out on the University of California at Riverside (UCR) collection [15]. The experimental results show that the proposed method can achieve satisfactory results. Our research in this paper significantly differs from previous work in the following aspects:
  • The framework imitates the mechanism of the human brain, in that the human brain’s cognition is a combination of the human body’s multiple senses. Thus, we use a DNN to learn features from different aspects to enhance the feature space. Then, based on the fusion of features, we carry out the TSC task and gain promising results.
  • The framework uses well-trained LSTM-A and CNN-A to extract sequence features and visualization features, as well as combining an attention mechanism to extract the key features that contribute to classification. In particular, an innovational category trend attention (CT-Attention) is gained from data belonging to the same category in an innovative way.
  • The framework transforms TS data into Area Graphs. Compared with the existing visualization methods (such as RP), this conversion method is simpler and avoids the loss of some information in complex conversions as well as additional computational cost.
The rest of this paper is organized as follows: In Section 2, we present the related work. Section 3 describes the proposed method in detail, including the structure of the network and how to calculate the attention to find out the contributing region in the raw data for the specific labels. Section 4 presents the evaluation results. Finally, we present conclusions, discussions, and suggestions for future research in Section 5.

2. Related Work

2.1. Related Work on TSC

There are a variety of methods for the task of TSC. These methods can be divided into four main categories: distance-based methods, feature-based methods, ensemble methods, and deep learning methods.
Distance-based methods. This type of method uses a variety of distance functions [16] to measure the similarity between TS records for classification. The most widely used distance function methods are DTW and its variants, such as the one nearest neighbor (1NN) classifier with DTW (1NN-DTW) [2]. In [17], the authors proposed a framework named Proximity Forest (PF), which uses Proximity Trees with 11 distance measures for the TSC task. The main drawback of this kind of method is the huge computational cost involved.
Feature-based methods. The basis of this type of method is a variety of features learned from TS data, through which we can distinguish the differences between data and classify them. The methods in this class include ST [18,19], bag of symbolic Fourier approximation (SFA) symbols (BOSS) [20], time-series forest (TSF) [21], and TS classification based on a bag-of-features representation (TSBF) [22]. Word ExtrAction for time SEries cLassification (WEASEL) [23] uses a novel discriminative feature generation and a feature selection method based on bag-of-patterns (BOP). The Shapelet Transform Classification (STC) uses a novel way to find shapelet and increases accuracy for multi-class problems [24]. Random Interval Spectral Ensemble (RISE) [25] combines tree structure with multiple features for TSC. Some methods input hand-engineered features using some domain knowledge into a DNN discriminative classifier. The disadvantages of these methods lie in the complexity and weak generality of building features, which obviously limits their versatility. Besides hand-engineered features, some methods use a DNN to extract the features of TS for classification: In [26,27], the authors added a deconvolutional operation into a convolutional neural network (CNN)-based model to reconstructing a multivariate time-series. Deep Belief Networks (DBNs) [28] and Recurrent Neural Network auto-encoders [29] have also been used to model the latent features in an unsupervised manner. Other studies [30,31] have used self-predicting modeling to ensure the effectiveness of feature learning. Inspired by computer vision processing methods, some scholars have converted TS data into image data and used image processing methods to extract features for TSC. Typical image transform methods include Gramian Angular Field (GAF) [8,9], RP [10,11], Markov Transition Fields (MTF) [12], and so on. However, these methods only use a single kind of feature.
Ensemble methods. This kind of method combines several effective methods, in order to obtain the most appropriate classification results from the classification results obtained through different mechanisms. It includes three typical methods: Elastic Ensemble (EE) [32], the collection of transformation ensembles (Flat-COTE) [33], and HIVE-COTE [4]. Among them, EE integrates 13 classification methods based on distance measurement; Flat-COTE includes 35 classification methods, where several feature-based classification methods are added in Flat-COTE; while HIVE-COTE is based on Flat-COTE, improving the mechanism of getting the final classification result from each sub-classification result. Time Series Combination of Heterogeneous and Integrated Embedding Forest (TS-CHIEF) [34] rivals HIVE-COTE in accuracy but is faster than HIVE-COTE. Random Convolutional Kernel Transform (ROCKET) [35] claims to achieve optimal performance with less computational cost. At the cost of a slight decrease in accuracy, Canonical Time-series Characteristics (Catch22) [36] generates a diverse and interpretable feature set with a greatly reduced number according to the properties of the TS data. The basis of these methods is still distance-based or feature-based; although the drawbacks of both types of methods are alleviated by ensembling, their flaws still exist. In addition, as a variety of classification models and features are ensembled, the model is very complex, which limits its practical application.
Deep learning methods. A variety of deep learning models have been proposed for TSC. This type of method trains a DNN model to form a mapping from data to categories. These models are discriminative, where they input raw TS data and output a probability distribution over the class variables in a data set. The models include Multi-scale CNN (MVCNN) [37], fully convolutional networks (FCN) [27], and deep residual networks (ResNet) [38], as well as many hybrid models such as attention-based LSTM-CNNs [39], multivariate LSTM-FCNs [40], and LSTM-FCNs [41]. It is worth noting that InceptionTime [42] claims higher accuracy and being faster than HIVE-COTE.

2.2. Related Work on Feature Extraction through DNN from Different Aspects

Deep learning has made amazing progress in many fields. However, the impact of changing the structure of the network on classification accuracy has been getting smaller and smaller and, so, researchers have begun to focus on expanding data sets, as the scale of the existing common data sets is inadequate compared to the current level of deep learning development. However, expanding data sets is not a simple task, as manual marking is needed, in general, because artificial participation leads to errors. These errors inevitably become factors that affect the training effect. Another method is to use more effective features. However, artificial features can only achieve good results in specific tasks.
A DNN imitates the mechanisms of the human brain [43] and is able to automatically learn the characteristics of data without heavy pre-processing for further processing. However, a human’s understanding of things is often a combination of multiple human senses; therefore, some scholars have used DNN to extract features from different perspectives for some tasks. In [13], the authors trained two VAEs to extract visual and semantic features of images, respectively, for generalized zero- and few-shot learning. In [14], the authors used a pre-trained saliency model to segment the foreground and background, and trained two feature extractors—one for extracting foreground features and the other for extracting background features—to improve image classification. In the field of NLP, the authors used historical Chinese scripts to enrich the pictographic evidence in characters and design CNN structures tailored to Chinese character image processing. The proposed glyph-based models gained outstanding results in multiple Chinese NLP tasks [6]. In the field of TS data mining, TS have been converted to bar images and a CNN was used for TS prediction, the results of which was also promising [7].

3. Model Design

The proposed framework consists of two parts, LSTM-A and CNN-A (as shown in Figure 1). LSTM-A is pre-trained with the loss function A U X 1 on raw TS data and CNN-A is pre-trained with the loss function A U X 2 on Area Graphs converted from the raw TS data. After LSTM-A and CNN-A are well-trained, raw test TS data are input to the LSTM-A to extract sequence features and corresponding Area Graphs to CNN-A for visualization features. Fusion features are obtained through a fully-connected layer concatenating the sequence features and visualization features. The final classification operation is realized by a softmax layer. An Attention mechanism is used to make the model focus on key sub-sequences and sub-regions which contain more discriminative information for TSC.

3.1. Data Notation

We use D = x i , y i i = 1 N to represent a time-series database, where N represents the number of records, x i = x 1 , x 2 x m represents the i th record, x j represents the ordered observation value of the record in the whole time-series m, and y i 1 , 2 C represents the category of the i th record, where C Z + is the number of classes. Generally speaking, the goal of TSC is to learn a mapping function from x i to y i , as shown in the following formula:
y i = f x i .
where f · is the mapping function we aim to learn. The following two sections describe how to extract sequence features and visualization features through LSTM-A and CNN-A, respectively.

3.2. LSTM-A

An RNN is a type of neural network used to process sequence data. There are many variants of RNNs, some representative ones including LSTM, Gated Recurrent Unit (GRU) [44], Bi-directional RNN [45], and so on. Greff et al. [46] compared popular RNN variants, showing that the RNN variants have almost the same performance and that LSTM is superior to other simplified LSTMs (such as GRU). In [47], the authors tested more than 10,000 RNN structures and found that, in certain tasks or situations, some RNN variants work better than LSTM; but only in special cases. Comparing GRU and LSTM, on the one hand, GRU has fewer parameters, so its training is faster and requires less data to generalize. On the other hand, if given enough training data, the great expressive power of LSTM may produce better results than GRU. In addition, in the field of TSC research, many studies have chosen LSTM [39,40,41] to do TSC and achieved promising results. So, we chose LSTM to learn the temporal dependencies of TS data. However, the temporal dependencies of long input sequences cannot be reasonably learned and, so, we added attention mechanisms to learn these long-term dependencies [48]. Finally, we used the LSTM-A model (Figure 2) to learn sequence features.
Given a time-series data set D, the LSTM-A model processes the records as follows:
z = t a n h ( x t , h t 1 ) ,
z f = s i g m o i d ( x t , h t 1 ) ,
z i = s i g m o i d ( x t , h t 1 ) ,
z o = s i g m o i d ( x t , h t 1 ) ,
c t = c t 1 z f z i z ,
h t = z o t a n h c t .
where h t 1 and c t 1 are the outputs of the last LSTM cell; x t is the current input; stands for the splicing operation; z i , z f , and z o are the input, forget, and output gates obtained by different parameters, respectively; ⊙ and ⊕ represent matrix multiplication and matrix addition, respectively; and h t and c t are the outputs of the current LSTM cell.
The T-Attention block is used to enhance the performance of sequence feature learning for very long input records. Its calculation formula is as follows:
M t = t a n h w h h t ,
a t = s o f t m a x w t M t ,
F s = F C a t h t ,
P s = s o f t m a x F s ,
A U X 1 = H P s , Y = i y i l o g p i s .
where h t is the output of the t th hidden unit of the last LSTM layer, a t is the t th attention weight, w h and w t are weighted matrices, a t and h t are merged, and ⊗ represents the matrix merge operation. Then, we use a Fully-connected layer to transform it to obtain sequence features F s . To prevent overfitting and gradients from disappearing, we use dropout and Batch Normalization continuously after the Fully-connected layer. Finally, a softmax layer is used to obtain the classification results, where P s is the output prediction probability sequence of LSTM-A. In the pre-training phase of LSTM-A, we use cross-entropy as the loss function A U X 1 , Y is the true probability sequence, and p i s and y i are the predicted probability and true probability of the record belonging to category i, respectively.

3.3. CNN-A

TS data is a series of sequential data. It is difficult for human beings to classify them only by ordered numbers, as vision is an important sense for human beings. By visualizing TS data, humans can easily classify them. Deep learning is a technology that imitates the human brain and, so, we consider visualizing time-series data and extracting features to imitate human vision for TSC. In addition, through data visualization, human vision can notice the parts with the largest differences and classify them. This is similar to the attention mechanism in deep learning. Therefore, we added an attention mechanism to CNN to extract key visualization features.
We used CNN-A to imitate human vision (the network structure is shown in Figure 3). For univariate TS data D, we convert each TS record x i into a black-and-white Area Graph (examples shown in Figure 4a,b) as input. In order to avoid the influence of unnecessary information such as co-ordinates, we removed the co-ordinates and other information from the Area Graphs, and only retained the graph part showing the data fluctuations.
Differing from the existing visualization transformation methods (e.g., Gramian fields [8,9], RP [10,11], and MTF [12]), the Area Graph is similar to human vision, directly reflecting information such as the fluctuation of TS data. The most important point is that this method is a very simple transformation method without additional calculation, avoiding the loss of some information in the calculation process. Such image data can extract visualization features through the CNN-A model.
In human vision, information such as peak value and mean value is more likely to attract attention. TS data belonging to the same category have similar peaks and averages after visualization, while data of different categories has large differences in such information. So, we used the CT-Attention block to extract these features and improve TSC. This block extracts the mean and maximum information of the category through a pixel average pooling layer based on Area Graphs of TS records belonging to same category, where the output is the CT Attention. The extracted CT attention is combined separately with the features (average and max features) extracted from each time-series record and, then, through a global average pooling layer, a Dropout layer, and a Batch Normalization layer, the final visualization features are obtained. Finally, classification is carried out through the softmax layer.
I = t r a n s x i x i D ,
C = f W I + B ,
H = P C ,
C T = A v g P i x x , y x , y I , y I = y , y 1 , 2 c ,
F v = G A P C T H ,
P v = s o f t m a x F v ,
A U X 2 = H P v , Y = i y i l o g p i v .
The process of learning visualization features by CNN-A can be expressed by Equation (4), where t r a n s means converting the TS record x to an Area Graph I, C denotes the result of a convolution (where * indicates the standard dot product) applied on every I, W and B are filter parameters, and f represents the combination of Batch Normalization and a final activation operation such as Rectified Linear Unit (Relu). P stands for max and average pooling, and H represents the pooling results. C T is the category-related attention obtained by A v g pooling on every pixel ( x , y ) of Area Graphs belonging to the same category, and y I is the category that image I belongs to (as shown in Figure 5). H and C T are merged by ⊙ and, after G A P processing, the final visualization features F v are obtained. In the pre-training phase of CNN-A, we use cross-entropy as the loss function A U X 2 . P v is the output prediction probability sequence of CNN-A and Y is the true probability sequence, and p i v and y i are the predicted probability and true probability of the record belonging to category i, respectively.

4. Experiment

In this section, we analyze the framework from different aspects. The TS data used in our experiments was the UCR collection [15], which is the largest public TS data classification archive. The data sets in UCR involve multiple different domains and consist of clusters of different sizes, shapes, and densities. We used 112 of the data sets and excluded 15 data sets with unequal length and one (Fungi) which had a single record per category in the training data set. Through experiments, we first compared the performance of the integrated model with the selected baseline model and evaluated the factors that affected the performance of the model. Secondly, we evaluated two sub-models, including the convergence status and the effectiveness of the learned features. We also evaluated whether the attention mechanism helped to learn features better. Our source code sharing website is in Supplementary Materials.

4.1. Experiment Setting and Model Configuration

All experiments were conducted using Python (Keras 2.2.4 [49] and Tensorflow 1.12.0 [50] as the backend). The Python packages used for data loading, visualization, and pre-processing included Numpy 1.16.5, Pandas 0.24.2, Scikit 0.20.4, and MatPlotLib 2.2.4. The hardware information was as follows:
  • CPU: Intel(R) Xeon(R) Gold 6140 CPU @ 2.30GHz (2 CPUs, 18 cores per CPU).
  • GPU: NVIDIA Corporation GK210GL [Tesla K80] (8 GPUs).
  • RAM: 256GB and 128GB video RAM for GPUs.
  • OS: Ubuntu 18.04.3 LTS.
Table 1 shows the hyper-parameters of LSTM-A and CNN-A. Relu was selected as the activation function of the activation layer. As the basis of the proposed model is the features extracted by pre-trained LSTM-A and CNN-A from TS data, the training of the two sub-models had a great impact on feature extraction. Therefore, we adopted multiple methods to improve the training of the two sub-models in the experiments. First, we trained the two sub-models using the loss function of ”categorical cross-entropy” with the Adam optimizer [51]. The initial learning rate was 1 e 3 , reducing by a factor of 0.5 to the final learning rate of 1 e 4 every 50/100 epochs (depending on epoch sizes; for epoch sizes smaller than 100, this was set to 50, otherwise 100) of no improvement in the validation score. Second, we used the early stop method to prevent the model from overfitting. When the change range of validation accuracy was less than 0.0003% every 50 epochs, the training was stopped. Finally, we replaced the Fully-connected layer with a global average pooling layer before the softmax layer in the CNN-A model, which greatly reduced the amount of parameters. In addition, we did not perform additional pre-processing on the data (e.g., regularization), as the UCR data sets have already been z-normalized. The batch and epoch sizes were chosen from {5, 20, 100, 200} and {50, 250, 1000, 2000}, respectively. The pixel size of the Area Graphs converted from TS data was chosen from {30 × 30, 150 × 150}.
In the field of TSC research, much work based on LSTM, CNN [27,52], and hybrid LSTM-CNN [39,40,41] has been carried out, achieving promising results. The architectures of these models are used for reference. In the experiments, we also tried models with different LSTM layers or CNN layers. We found that reducing the number of layers damaged the performance of model, while increasing the number of layers not only damaged the performance of model as well, but also increased the model training time.

4.2. Integrated Model Evaluation

Among the existing four types of methods, we used 15 methods for comparison, among which we reproduced 1NN_DTW, ST and FCN, supplementing the results on the 27 data sets of the additional data sets in UCR, and the results of the remaining methods come from [15]. The chosen methods were as follows:
  • Distance-based methods: 1NN_DTW [2] and PF [17];
  • Feature-based methods: BOSS [20], ST [18,19], STC [24], WEASEL [23], TSF [21] and RISE [25];
  • Ensemble methods: HIVE-COTE [4], TS-CHIEF [34], ROCKET [35] and Catch22 [36];
  • Deep learning methods: ResNet [38], FCN [38] and InceptionTime [42].
For the detailed structure and parameters of each method, please refer to their respective references. All results were obtained by taking the average value from multiple experiments. All models were evaluated using classification accuracy and mean-per-class-error (MPCE), which is defined as the average error of each class for all data sets and mathematically represented in Equation (5). The Average Arithmetic Rank (AVG Rank) is the mean of the classification accuracy ranking over the 112 data sets. Table 2 lists the comparison results. In order to facilitate the display, the names of some methods have been abbreviated again, including 1NN_DTW (labelled DTW), WEASEL (WS), HIVE-COTE (HCT), TS-CHIEF (CHI), ROCKET (RK), Catch22 (C2), ResNet (RN), InceptionTime (IcT).
P C E k = 1 a c c u r a c y n u m b e r o f c l a s s e s ,
M P C E = 1 N k = 1 N P C E k .
Compared with the baseline methods on 112 data sets in UCR, the proposed mthod TSC-FF won or tied on 20 data sets. Judging from the AVG Rank and MPSE, its overall ranking was fifth, after HIVE-COTE, TS-CHIEF, ROCKET and InceptionTime. By comparison, we found that TSC-FF was far better than six feature-based methods. Therefore, even though the deep learning-based method TSC-FF proposed in this paper did not achieve the best performance, it showed the greatest potential of the deep learning methods considered for feature learning, proving that DNN can automatically complete feature extraction, not requiring complicated feature engineering. With the development of DNN and the accumulation of data, deep learning methods can even achieve better performance.
Although TSC-FF could not defeat the deep learning method InceptionTime, which achieved the lowest MPSE as a whole, by comparing the results of FordA, FordB, and Wafer, we found that our framework achieved higher accuracy on these data sets. This shows that, when a DNN is given enough data to learn features, using pre-trained sub-models to learn features from different aspects and then fusing features can improve DNN performance. Of course, the amount of data and the number of categories also has an effect on DNN performance, which will be explored in later experiments. On the data set ElectricDevices, the featured-based method BOSS had the highest accuracy. The reason for this is the artificially constructed features used by BOSS are better than those learned by the DNN. After all, there still exists a certain gap between DNNs and the human brain. On some data sets, the features learned by DNN are not as effective as those designed by humans. However, the results of BOSS on other data sets proved the weak generality of artificially constructed features.
Among all methods, HIVE-COTE achieved the best performance. Comparing the two distance-based methods, the accuracy of PF is much improved compared to DTW. The difference between the two is that PF uses multiple distance functions. Among the feature-based methods, STC achieved the best performance. What makes STC special is that it uses some innovative methods to find more effective shapelet features. In the deep learning methods, the best performance is achieved by InceptionTime. InceptionTime combines several DNNs with the same structure but different parameters, and it can be considered that multiple DNNs learn different features, thus InceptionTime improved the performance compared to other deep learning methods.
Those achieving excellent performance methods either combine multiple distance functions and features, or find and select more effective features, or use multiple DNNs to learn various features from limited data. The common point of these methods is to use more effective features (or more features to cover the most effective features) to improve performance in the case of limited data. Similarly, the proposed method TSC-FF purposefully selected and used the fusion features of multiple features learned from different aspects, so it achieved promising results.
In the Wilcoxon Sign Rank Test (Table 3), the p-values of TSC-FF with 1NN_DTW, TSF, RISE and Catch22 were smaller than 0.05, while with other methods were larger than 0.05, indicating that TSC-FF had almost equally good results as those achieving excellent performance methods. The reason why TSC-FF did not achieve the best overall result is that TSC-FF is a method based on deep learning. TSC-FF does not suffer from high computational complexity, due to employing multiple different classifiers, compared with the ensemble method HIVE-COTE, but the sub models used by TSC-FF are more affected by overfitting due to the training set being too small, considering its difference in structure compared to InceptionTime, FCN and ResNet. Some data sets in UCR had a small training set and a large number of categories, such that the sub-models were easily over-fitted, resulting in insufficient extracted features and affecting the classification performance.
We used the DNN to learn the characteristics of different classes. Whether there are enough records for each class to learn is one of the problems that must be considered. Table 4 shows the performance of all methods with different amounts of records in each category. The methods based on deep learning achieved the best performance. In particular, InceptionTime, which used multiple DNNs, achieved the highest average accuracy in most cases with different data volumes in each category. This shows that DNNs can learn more effective features than artificial features and replace the work of constructing features by human beings.
In our proposed framework, TS data was converted into Area Graphs to extract visualization features. However, some existing methods convert TS data into image data through other different methods to extract features, so we also carried out a comparison with these methods. We chose two such methods: The first method uses Tiled Convolutional Neural Networks (tiled CNNs) on Gramian Angular Summation/Difference Fields (GASF/GADF) and MTF to extract features for classification, abbreviated as GASF-GADF-MTF. The second method uses a Support Vector Machine (SVM) on Texture Features extracted from RP for classification, abbreviated as TFRP. Table 5 shows the comparison results. Through the comparison, we can see that the proposed framework had better performance than the selected comparison methods. The reason for this may be that the visualization method in this paper is simple, avoiding the loss of some information in the conversion process. In addition, the combination of different aspects of features (i.e., sequence features and visualization features) makes up for the deficiencies of single-category features.

4.3. Sub-Model Evaluation

Figure 6 depicts the accuracy and loss curves of two sub-models trained on the data set StarLightCurves. From the figure, we can see that, although the accuracy rate and loss fluctuated, the overall accuracy rate gradually increased and the loss gradually decreased, finally reaching a stable state, and that the results on the training set were slightly better than the results of the verification set, which shows that the model has good generalization ability.
Figure 4 shows an example of transformed Area Graphs and features from the ’BeetleFly’ data set. In the data set BeetleFly, the records are divided into two categories. We randomly selected one record from the validation set of each category and show the corresponding Area Graphs (Figure 4a,b), sequence features (Figure 4c,d), and visualization features (Figure 4e,f) extracted by LSTM-A and CNN-A, respectively. Using the area in the red box in Figure 4a,b for comparison, the numbers in this area in the original record were quite different, which may be the main trend used to distinguish record categories. Correspondingly, in the learned sequence features, the corresponding positions were the wave crest and wave trough; that is, they were quite different. In the visualization features, we can see that most of the features of the corresponding positions were similar, as the numbers around the red box in the original record are similar, and only one part of the visual features corresponding to the value of the red box position is different, which proves that CNN-A has learned the features of similar parts, as well as the features of the different parts, of the record.
We also compared the accuracy of the pre-trained model with that of the integrated model. Table 6 shows the accuracy comparison of the two sub-models and the integrated model on 112 data sets. The bold values in TSC-FF denote model wins or ties with sub-models, while those in LSTM-A and CNN-A denote model wins or ties with LSTM and CNN, respectively. Among them, the integrated model TSC-FF achieved a higher accuracy on 72 data sets than the two sub-models. This shows that the combination of the two features generally improved the classification performance. By comparing the accuracy of the sub-models with and without the attention mechanism, we can see that LSTM-A achieved higher accuracy on 95 data sets than without the attention mechanism, and the same for CNN-A was 103. This indicates that, for the two sub-models, when using the attention mechanism, the model can learn more effective features. In addition, the difference in performance between CNN-A and CNN also illustrates the effectiveness of the CT-Attention method proposed in this paper.
Finally, we explored the influence of the length of TS records on the feature learning of sub-models. From Table 7, we can see that, when the record length was small (<80), the average test accuracy obtained by LSTM-A was higher than that of CNN-A and, as the record length increased, CNN-A performed better than LSTM-A. This shows that the sequence features learned by LSTM-A on short TS records were more effective than the visual features learned by CNN-A. This is because, as the record length increases, the LSTM-A may "forget" some features, while using CNN-A to learn visualization features is not affected by this. In addition, when we visualize long data, some small fluctuations will not be clearly displayed on the picture. Therefore, CNN-A can learn some more distinguishable features, such as the main trends of the data.

5. Conclusions and Future Work

In this paper, we used a DNN to learn features from different aspects to enhance the feature space. Specifically, we used LSTM-A to learn sequence features from raw TS data and CNN-A to learn visual features from Area Graphs converted from TS data. Then, we classified TS data based on the fused features. Through various forms of comparison and analysis, we found that the well-trained LSTM-A and CNN-A learned features that could effectively distinguish the TS data and promote each other. With the proposed framework, we have proved that, in the task of TSC, using deep learning methods can achieve similar performance as the complex ensemble methods, and the features extracted by deep learning are more effective and general than artificially constructed features.
We proposed the use of CNN-based methods to extract visualization features, which is feasible for univariate TS data; however, on multivariate TS data, how to perform visualization to extract visual features and how to handle the correlation between multi-dimensional data provides difficulties which need to be solved for the framework proposed in this paper to expand on multivariate TS data.

Supplementary Materials

Our source code has been uploaded to https://github.com/wangbaoquan520/TSC-FF.

Author Contributions

Conceptualization, B.W.; Data curation, B.M.; Formal analysis, B.W.; Funding acquisition, T.J., X.Z.; Investigation, B.W.; Methodology, B.W.; Project administration, H.J.; Resources, B.W.; Software, B.W.; Supervision, F.Z.; Validation, B.W.; Visualization, F.Z.; Writing—original draft, B.W.; Writing—review & editing, B.M., Y.W. and B.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by The Xinjiang Science and Technology Major Project, Grant No.2016A03007-2, The Youth Innovation Promotion Association of Chinese Academy of Sciences, Grant No.Y9290802, The West Light Foundation of The Chinese Academy of Sciences, Grant No.2019-XBQNXZ-A-004, The Tianshan Excellent Young Scholars of Xinjiang, Grant No.2018Q032.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Esling, P.; Agon, C. Time-series data mining. ACM Comput. Surv. (CSUR) 2012, 45, 12. [Google Scholar] [CrossRef] [Green Version]
  2. Berndt, D.J.; Clifford, J. Using Dynamic Time Warping to Find Patterns in Time Series; KDD Workshop: Seattle, WA, USA, 1994; Volume 10, pp. 359–370. [Google Scholar]
  3. Ye, L.; Keogh, E. Time series shapelets: a new primitive for data mining. In Proceedings of the 15th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Paris, France, 28 June–1 July 2009; pp. 947–956. [Google Scholar]
  4. Lines, J.; Taylor, S.; Bagnall, A. Hive-cote: The hierarchical vote collective of transformation-based ensembles for time series classification. In Proceedings of the 2016 IEEE 16th International Conference on Data Mining (ICDM), Barcelona, Spain, 12–15 December 2016; pp. 1041–1046. [Google Scholar]
  5. Graves, A. Supervised sequence labelling. In Supervised Sequence Labelling with Recurrent Neural Networks; Springer: New York, NY, USA, 2012; pp. 5–13. [Google Scholar]
  6. Wu, W.; Meng, Y.; Han, Q.; Li, M.; Li, X.; Mei, J.; Nie, P.; Sun, X.; Li, J. Glyce: Glyph-vectors for Chinese Character Representations. arXiv 2019, arXiv:1901.10125. [Google Scholar]
  7. Berat Sezer, O.; Murat Ozbayoglu, A. Financial Trading Model with Stock Bar Chart Image Time Series with Deep Convolutional Neural Networks. arXiv 2019, arXiv:1903.04610. [Google Scholar]
  8. Wang, Z.; Oates, T. Encoding time series as images for visual inspection and classification using tiled convolutional neural networks. In Proceedings of the Workshops at the Twenty-Ninth AAAI Conference on Artificial Intelligence, Austin, TX, USA, 25–30 January 2015. [Google Scholar]
  9. Wang, Z.; Oates, T. Imaging time-series to improve classification and imputation. In Proceedings of the Twenty-Fourth International Joint Conference on Artificial Intelligence, Buenos Aires, Argentina, 25–31 July 2015. [Google Scholar]
  10. Tripathy, R.; Acharya, U.R. Use of features from RR-time series and EEG signals for automated classification of sleep stages in deep neural network framework. Biocybern. Biomed. Eng. 2018, 38, 890–902. [Google Scholar] [CrossRef]
  11. Hatami, N.; Gavet, Y.; Debayle, J. Classification of time-series images using deep convolutional neural networks. In Tenth International Conference on Machine Vision (ICMV 2017); International Society for Optics and Photonics: Bellingham, WA, USA, 2018; Volume 10696, p. 106960Y. [Google Scholar]
  12. Wang, Z.; Oates, T. Spatially encoding temporal correlations to classify temporal data using convolutional neural networks. arXiv 2015, arXiv:1509.07481. [Google Scholar]
  13. Schonfeld, E.; Ebrahimi, S.; Sinha, S.; Darrell, T.; Akata, Z. Generalized zero-and few-shot learning via aligned variational autoencoders. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 8247–8255. [Google Scholar]
  14. Zhang, H.; Zhang, J.; Koniusz, P. Few-shot Learning via Saliency-guided Hallucination of Samples. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 2770–2779. [Google Scholar]
  15. Dau, H.A.; Keogh, E.; Kamgar, K.; Yeh, C.C.M.; Zhu, Y.; Gharghabi, S.; Ratanamahatana, C.A.; Yanping, B.H.; Begum, N.; Bagnall, A.; et al. The UCR Time Series Classification Archive. Available online: https://www.cs.ucr.edu/~eamonn/time_series_data_2018/ (accessed on 31 October 2018).
  16. Ding, H.; Trajcevski, G.; Scheuermann, P.; Wang, X.; Keogh, E. Querying and mining of time series data: experimental comparison of representations and distance measures. Proc. VLDB Endow. 2008, 1, 1542–1552. [Google Scholar] [CrossRef] [Green Version]
  17. Lucas, B.; Shifaz, A.; Pelletier, C.; O’Neill, L.; Zaidi, N.; Goethals, B.; Petitjean, F.; Webb, G.I. Proximity forest: An effective and scalable distance-based classifier for time series. Data Min. Knowl. Discov. 2019, 33, 607–635. [Google Scholar] [CrossRef] [Green Version]
  18. Hills, J.; Lines, J.; Baranauskas, E.; Mapp, J.; Bagnall, A. Classification of time series by shapelet transformation. Data Min. Knowl. Discov. 2014, 28, 851–881. [Google Scholar] [CrossRef] [Green Version]
  19. Grabocka, J.; Schilling, N.; Wistuba, M.; Schmidt-Thieme, L. Learning time-series shapelets. In Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, New York, NY, USA, 24–27 August 2014; pp. 392–401. [Google Scholar]
  20. Schäfer, P. The BOSS is concerned with time series classification in the presence of noise. Data Min. Knowl. Discov. 2015, 29, 1505–1530. [Google Scholar] [CrossRef]
  21. Deng, H.; Runger, G.; Tuv, E.; Vladimir, M. A time series forest for classification and feature extraction. Inf. Sci. 2013, 239, 142–153. [Google Scholar] [CrossRef] [Green Version]
  22. Baydogan, M.G.; Runger, G.; Tuv, E. A bag-of-features framework to classify time series. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 2796–2802. [Google Scholar] [CrossRef]
  23. Schäfer, P.; Leser, U. Fast and accurate time series classification with weasel. In Proceedings of the 2017 ACM on Conference on Information and Knowledge Management, Singapore, 6–10 November 2017; pp. 637–646. [Google Scholar]
  24. Bostrom, A.; Bagnall, A. Binary shapelet transform for multiclass time series classification. In Transactions on Large-Scale Data-and Knowledge-Centered Systems XXXII; Springer: New York, NY, USA, 2017; pp. 24–46. [Google Scholar]
  25. Flynn, M.; Large, J.; Bagnall, T. The Contract Random Interval Spectral Ensemble (c-RISE): The Effect of Contracting a Classifier on Accuracy. In International Conference on Hybrid Artificial Intelligence Systems; Springer: New York, NY, USA, 2019; pp. 381–392. [Google Scholar]
  26. Wang, Z.; Song, W.; Liu, L.; Zhang, F.; Xue, J.; Ye, Y.; Fan, M.; Xu, M. Representation learning with deconvolution for multivariate time series classification and visualization. arXiv 2016, arXiv:1610.07258. [Google Scholar]
  27. Mittelman, R. Time-series modeling with undecimated fully convolutional neural networks. arXiv 2015, arXiv:1508.00317. [Google Scholar]
  28. Wang, S.; Hua, G.; Hao, G.; Xie, C. A cycle deep belief network model for multivariate time series classification. Math. Probl. Eng. 2017, 2017. [Google Scholar] [CrossRef] [Green Version]
  29. Mehdiyev, N.; Lahann, J.; Emrich, A.; Enke, D.; Fettke, P.; Loos, P. Time series classification using deep learning for process planning: a case from the process industry. Procedia Comput. Sci. 2017, 114, 242–249. [Google Scholar] [CrossRef]
  30. Aswolinskiy, W.; Reinhart, R.F.; Steil, J. Time series classification in reservoir-and model-space. Neural Process. Lett. 2018, 48, 789–809. [Google Scholar] [CrossRef]
  31. Bianchia, F.M.; Scardapaneb, S.; Løksea, S.; Jenssena, R. Reservoir computing approaches for representation and classification of multivariate time series. arXiv 2018, arXiv:1803.07870. [Google Scholar]
  32. Lines, J.; Bagnall, A. Time series classification with ensembles of elastic distance measures. Data Min. Knowl. Discov. 2015, 29, 565–592. [Google Scholar] [CrossRef]
  33. Bagnall, A.; Lines, J.; Hills, J.; Bostrom, A. Time-series classification with COTE: the collective of transformation-based ensembles. IEEE Trans. Knowl. Data Eng. 2015, 27, 2522–2535. [Google Scholar] [CrossRef]
  34. Shifaz, A.; Pelletier, C.; Petitjean, F.; Webb, G.I. Ts-chief: A scalable and accurate forest algorithm for time series classification. Data Min. Knowl. Discov. 2020, 1–34. [Google Scholar] [CrossRef] [Green Version]
  35. Dempster, A.; Petitjean, F.; Webb, G.I. ROCKET: Exceptionally fast and accurate time series classification using random convolutional kernels. arXiv 2019, arXiv:1910.13051. [Google Scholar]
  36. Lubba, C.H.; Sethi, S.S.; Knaute, P.; Schultz, S.R.; Fulcher, B.D.; Jones, N.S. catch22: CAnonical Time-series CHaracteristics. Data Min. Knowl. Discov. 2019, 33, 1821–1852. [Google Scholar] [CrossRef] [Green Version]
  37. Cui, Z.; Chen, W.; Chen, Y. Multi-scale convolutional neural networks for time series classification. arXiv 2016, arXiv:1603.06995. [Google Scholar]
  38. Wang, Z.; Yan, W.; Oates, T. Time series classification from scratch with deep neural networks: A strong baseline. In Proceedings of the 2017 international joint conference on neural networks (IJCNN), Anchorage, AK, USA, 14–19 May 2017; pp. 1578–1585. [Google Scholar]
  39. Du, Q.; Gu, W.; Zhang, L.; Huang, S.L. Attention-based LSTM-CNNs For Time-series Classification. In Proceedings of the 16th ACM Conference on Embedded Networked Sensor Systems, Shenzhen, China, 4–7 November 2018; pp. 410–411. [Google Scholar]
  40. Karim, F.; Majumdar, S.; Darabi, H.; Harford, S. Multivariate lstm-fcns for time series classification. Neural Netw. 2019, 116, 237–245. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  41. Karim, F.; Majumdar, S.; Darabi, H.; Chen, S. LSTM fully convolutional networks for time series classification. IEEE Access 2017, 6, 1662–1669. [Google Scholar] [CrossRef]
  42. Fawaz, H.I.; Lucas, B.; Forestier, G.; Pelletier, C.; Schmidt, D.F.; Weber, J.; Webb, G.I.; Idoumghar, L.; Muller, P.A.; Petitjean, F. InceptionTime: Finding AlexNet for Time Series Classification. arXiv 2019, arXiv:1909.04939. [Google Scholar]
  43. Kim, P. Matlab deep learning. Mach. Learn. Neural Netw. Artif. Intell. 2017, 130. [Google Scholar] [CrossRef]
  44. Cho, K.; Van Merriënboer, B.; Gulcehre, C.; Bahdanau, D.; Bougares, F.; Schwenk, H.; Bengio, Y. Learning phrase representations using RNN encoder-decoder for statistical machine translation. arXiv 2014, arXiv:1406.1078. [Google Scholar]
  45. Schuster, M.; Paliwal, K.K. Bidirectional recurrent neural networks. IEEE Trans. Signal Process. 1997, 45, 2673–2681. [Google Scholar] [CrossRef] [Green Version]
  46. Greff, K.; Srivastava, R.K.; Koutník, J.; Steunebrink, B.R.; Schmidhuber, J. LSTM: A search space odyssey. IEEE Trans. Neural Netw. Learn. Syst. 2016, 28, 2222–2232. [Google Scholar] [CrossRef] [Green Version]
  47. Jozefowicz, R.; Zaremba, W.; Sutskever, I. An empirical exploration of recurrent network architectures. In Proceedings of the International Conference on Machine Learning, Lille, France, 7–9 July 2015; pp. 2342–2350. [Google Scholar]
  48. Bahdanau, D.; Cho, K.; Bengio, Y. Neural machine translation by jointly learning to align and translate. arXiv 2014, arXiv:1409.0473. [Google Scholar]
  49. Documentation, K. Available online: https://keras.io/api/ (accessed on 14 November 2017).
  50. Abadi, M.; Barham, P.; Chen, J.; Chen, Z.; Davis, A.; Dean, J.; Devin, M.; Ghemawat, S.; Irving, G.; Isard, M.; et al. Tensorflow: A system for large-scale machine learning. In Proceedings of the 12th {USENIX} Symposium on Operating Systems Design and Implementation ({OSDI} 16), Savannah, GA, USA, 2–4 November 2016; pp. 265–283. [Google Scholar]
  51. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  52. Liu, C.L.; Hsaio, W.H.; Tu, Y.C. Time series classification with multivariate convolutional neural network. IEEE Trans. Ind. Electron. 2018, 66, 4788–4797. [Google Scholar] [CrossRef]
Figure 1. Overview of the framework. A U X 1 and A U X 2 are loss functions in the pre-training stage.
Figure 1. Overview of the framework. A U X 1 and A U X 2 are loss functions in the pre-training stage.
Applsci 10 04124 g001
Figure 2. Architecture of the LSTM-A model. T-Attention means temporal attention block, the permute layers are used to permute the dimensions of the input matrix, and BN stands for batch normalization.
Figure 2. Architecture of the LSTM-A model. T-Attention means temporal attention block, the permute layers are used to permute the dimensions of the input matrix, and BN stands for batch normalization.
Applsci 10 04124 g002
Figure 3. Architecture of the CNN-A model. BN stands for batch normalization, the GAP layer is the global average polling layer, and CT-Attention is the category trend attention.
Figure 3. Architecture of the CNN-A model. BN stands for batch normalization, the GAP layer is the global average polling layer, and CT-Attention is the category trend attention.
Applsci 10 04124 g003
Figure 4. Records and features for categories of the data set BeetleFly.
Figure 4. Records and features for categories of the data set BeetleFly.
Applsci 10 04124 g004
Figure 5. Depiction of pixel Average pooling.
Figure 5. Depiction of pixel Average pooling.
Applsci 10 04124 g005
Figure 6. Accuracy and loss curves of training two sub-models on the data set StarLightCurves.
Figure 6. Accuracy and loss curves of training two sub-models on the data set StarLightCurves.
Applsci 10 04124 g006
Table 1. Model Configuration.
Table 1. Model Configuration.
NameConfiguration (the Parameter Value Was Selected from · , According to the Optimal Accuracy on the Validation Set in Different Experiments).
TSC-FF
Fully-connected Layer{64, 128, 256}
LSTM-A
LSTM layer1{64, 128, 256}
LSTM layer2{64, 128, 256}
LSTM layer3{64, 128, 256}
Dropout{0.2, 0.5, 0.8}
Batch Normalizationmomentum = 0.99, epsilon = 0.001
Fully-connected Layer{64, 128, 256}
CNN-A
Conv2D1filter number {32, 64, 128}, kernel size {5 × 5, 8 × 8, 10 × 10}, strides {1, 2, 4}
Conv2D2filter number {32, 64, 128}, kernel size {5 × 5, 8 × 8, 10 × 10}, strides {1, 2, 4}
Conv2D3filter number {32, 64, 128}, kernel size {5 × 5, 8 × 8, 10 × 10}, strides {1, 2, 4}
Dropout{0.2, 0.5, 0.8}
Batch Normalizationmomentum = 0.99, epsilon = 0.001
Avg. Poolpool size {5 × 5, 8 × 8}
Max Poolpool size {5 × 5, 8 × 8}
Fully-connected Layer{64, 128, 256}
Table 2. Accuracy comprison on 112 data sets. Bold values denote the model with the best performance.
Table 2. Accuracy comprison on 112 data sets. Bold values denote the model with the best performance.
DTWPFBOSSSTSTCWSTSFRISEHCTCHIRKC2RNFCNIcTTSC-FF
ACSF10.56000.63830.76830.78260.83830.81800.63500.76000.85000.80700.80700.77770.82400.88000.82670.8552
Adiac0.58570.72220.74900.78260.79320.79880.71190.75800.79620.77970.77200.68470.81540.85700.82230.7875
ArroHea0.77710.88360.86880.73710.80670.84840.79680.82820.87600.88110.85900.75030.85870.88000.88040.8025
Beef0.66670.59440.61220.90000.73560.74000.68890.74220.73560.63220.76000.47330.67670.75000.68220.8444
BeetleFl0.75000.86000.94330.90000.93330.88670.83330.87170.96330.95830.88500.84000.85330.95000.89330.8889
BirChick0.85000.90330.98330.80000.87000.86500.81500.86830.94000.96330.88170.89330.94500.95000.95170.9222
BME0.79330.99910.86580.82670.92980.94780.96240.78600.98220.99640.99730.90490.99910.75330.99640.8921
Car0.73330.80560.84830.91670.85830.83440.76610.75330.86890.87890.91170.74610.90830.91700.90110.9500
CBF0.59220.99360.99890.97440.98530.97980.97190.94900.99830.99840.99590.95370.98821.00000.99611.0000
Chinatow0.84060.94800.87710.91450.96300.95730.95300.88850.96280.96180.96690.93450.97010.98260.96490.9861
ChlCon0.68310.63110.65820.69970.73520.75490.72310.76480.73390.66080.79610.59800.84100.84300.86360.7246
CCECGTor0.71090.93770.91480.95430.97780.98500.95820.94740.99370.95340.86410.80300.76790.81300.83280.9511
Coffee0.92860.99170.98570.96430.98930.98930.98690.98450.99290.99051.00000.97980.99641.00000.99881.0000
Compu0.66800.71430.80050.73600.79910.77850.64880.77890.81110.75390.84290.78030.86040.84800.86560.7200
CricketX0.63080.80040.76240.77180.79210.77570.69270.70620.81620.83040.83900.60890.80810.81500.85320.8077
CricketY0.55900.79980.75040.77950.77790.77960.68590.70910.80980.81700.84500.59050.81020.79200.86000.8000
CricketZ0.55640.80280.76930.78720.80720.78990.70590.72160.83390.83820.85320.62820.81320.81300.86110.8282
Crop0.64880.75310.68570.70890.73670.72380.74560.73000.76820.76210.75170.65310.76380.75370.79320.7622
DiaSizRed0.93460.95680.94520.92480.85930.90810.94170.93210.91430.94590.95800.92470.30620.93000.95070.9471
DiPhOuAG0.73190.80220.82060.77540.79620.79280.80940.82160.82400.82810.81150.78300.77600.83500.76570.7661
DiPhOuCor0.67630.82340.81170.76980.82730.81920.80580.81120.82360.81930.82440.81210.80920.81200.81560.7120
DiPhTW0.58990.69210.67150.66190.68990.67870.69110.69450.69620.69180.70120.68110.66710.79000.66590.6875
Earthqu0.67630.74960.74600.74100.74200.74750.74750.74820.74750.74820.74840.73880.71700.80100.73210.8160
ECG2000.83000.87300.87830.83000.83900.85900.86000.85100.85870.85500.89900.78870.88370.90000.89570.8916
ECG50000.91960.93950.94010.94380.94180.94590.94340.93670.94560.94850.94740.93620.93700.94100.94210.9481
ECGFiveD0.68640.88280.99230.98370.97790.99350.95200.97290.99380.99440.99600.81590.95100.98500.99590.9991
ElecDev0.57090.84240.79740.74700.88180.87170.79310.82400.87970.86500.89320.87210.88840.72300.89020.7413
EOGHoSi0.54700.82440.70670.77610.75990.74730.70600.62490.79570.85370.81380.67300.86470.72980.88360.8358
EOGVeSi0.58010.77760.66030.75150.71320.69750.67260.60350.76300.80950.78110.61490.75170.72540.81450.7722
EthLevel0.26800.32410.50870.62560.85570.70510.66930.66110.84900.60560.62530.39740.85250.64200.87550.8442
FacAll0.87280.97710.97010.77870.95370.97310.95000.96490.97970.98270.98800.81080.98190.92900.98330.8705
FaFour0.62500.94550.99550.85230.65640.98110.90680.87690.97310.99960.93140.67950.92500.93200.93860.9091
FacUCR0.84340.95610.95130.90590.91010.95620.90400.89200.96130.97290.97130.70870.96430.94800.97690.9593
FiftyW0.69230.82590.70590.70550.73660.77700.72170.66660.77160.84270.82510.59780.72400.67900.82680.8041
Fish0.89710.93390.96970.98860.94990.95090.83050.85900.97940.98150.97410.77260.96990.97100.97260.9886
FordA0.71740.85000.92140.97120.93420.96870.81580.94000.94410.94740.94240.90920.93150.90600.95911.0000
FordB0.66540.83940.90740.80740.91980.93710.79070.91730.92800.91950.92300.87010.91280.88300.94091.0000
FrRegTra0.90460.94240.98810.99510.99930.99060.99710.95230.99910.99850.99440.99820.99670.99650.99580.9989
FrSmaTra0.71470.82340.96160.96890.98860.90060.96140.87870.98370.99550.98720.95980.94950.69020.94890.9668
GunPoint0.99330.99130.99641.00000.98640.99310.95530.98090.99821.00000.99200.94310.99091.00000.99511.0000
GPAgSpa0.95050.99690.99490.98050.96600.98130.97770.98630.99660.99960.99350.94390.99440.94240.98391.0000
GPMaVeFe0.94680.99940.99960.99150.98650.99390.99600.99110.99990.99940.99990.99350.99010.94940.99830.9957
GPOlVeYo0.85601.00000.99920.99450.97830.98601.00000.99981.00001.00000.98960.96421.00000.89521.00000.9264
Ham0.47620.78350.83750.68570.81080.82130.79940.81970.84000.80510.85520.69400.80730.76200.85020.7128
HandOut0.7378N/A0.91480.93240.92140.94540.90660.88100.9166N/AN/A0.86840.90300.7760N/A0.9069
Haptics0.28570.45830.46740.52270.54170.44920.46560.47880.53940.52330.53420.45870.49650.55100.53660.5213
Herring0.50000.57450.59580.67190.63280.60210.60420.59840.61200.59740.62500.55570.59690.70300.62500.6491
HouTwe0.85710.93700.95600.92040.97510.81060.83780.92970.97870.97030.96270.94620.95550.89080.95350.9613
InlSkate0.45270.56100.50560.37270.43930.67460.37280.39040.51520.57190.43790.47240.41010.41100.53440.4616
IERegTra1.00001.00000.99680.99570.99340.97181.00000.99291.00001.00000.98700.92881.00001.00001.00001.0000
IESmaTra1.00001.00000.98490.87360.88770.90670.99590.97221.00001.00000.93430.83720.39630.35741.00000.5150
InWinSou0.24340.60690.51180.62680.62860.61950.60350.63640.64030.63220.65670.55920.49140.40200.62730.6269
IPoDem0.91350.95600.87090.94750.95380.94680.95950.94450.95820.96240.96160.87750.95710.97000.96030.9633
LarKitApp0.76800.81010.83580.85870.92990.79270.63630.73580.92010.86020.93000.88630.95390.89600.95240.8723
Lightn20.67210.84860.81910.73770.65850.62730.76450.68200.77320.76890.77650.74480.80050.80300.81640.8155
Lightn70.57530.79220.67120.72600.74340.71280.72050.69770.75750.79360.79770.64610.81000.86300.82150.8115
Mallat0.89720.97200.94980.96420.90800.96640.93570.95410.96720.97670.95720.90610.96900.98000.96250.9542
Meat0.66670.98720.98060.85000.96780.97670.98390.98670.98610.98440.98890.94280.99390.96700.98390.9333
MedIma0.66320.77140.71580.66970.70980.70880.74620.66670.74040.79910.80510.75680.79220.79200.79630.7578
MiPhOC0.69070.65890.65630.79380.66750.66040.65950.69980.69780.69440.71080.68810.59650.76800.59460.7663
MiPhAG0.55840.82410.80950.64290.83150.82830.79950.80550.81290.80580.83450.77270.82360.79500.83400.7000
MiPhTW0.50000.54910.53230.51950.57860.55430.56860.58510.58380.57320.58980.55740.53120.61200.52660.5745
MSRegTra0.75180.96380.91790.93810.96020.96330.92300.93770.96590.97140.96570.92150.97010.86680.96630.9690
MSSmaTra0.80430.92810.87170.90610.93680.92450.84020.89830.94470.94730.93070.87520.91620.91050.91340.9386
MotStra0.71570.91500.84420.89700.90890.90480.85550.87800.93650.93010.90550.84850.90310.95000.88060.8943
NonECG10.69720.87890.84160.94960.93590.92960.88480.90070.92780.91670.90220.84810.94420.96100.87440.8818
NonECG20.82650.85850.90270.95110.95010.93620.91560.91840.95070.97360.93870.87720.95010.95500.95260.9424
OliveOil0.86670.82910.87560.90000.87890.91330.89330.89330.88330.82530.84470.74560.86220.83300.86150.8667
OSULeaf0.88020.32040.96910.96690.95630.85180.64330.65430.97480.36980.27510.72420.97470.98800.33460.9699
PhOutCor0.74010.27550.81740.76340.83370.82170.80570.81250.82650.96010.19550.79190.84770.82600.92210.8096
Phoneme0.25630.71570.25460.32070.35650.25950.19370.34660.37120.96890.90850.30020.34570.34500.93320.3058
PigAir0.2260N/A0.94180.96560.97740.53780.28010.18250.9577N/AN/A0.27690.43700.2067N/A0.5370
PigArt0.6538N/A0.96760.95840.94570.93780.30720.79460.9668N/AN/A0.82360.95370.8798N/A0.9719
PigCVP0.42310.50050.95950.92910.89970.89790.17820.68250.95300.96140.88170.47610.45400.25000.90530.5253
Plane1.00001.00000.99811.00000.99900.99490.99590.99651.00001.00001.00000.98831.00001.00000.99681.0000
PowCons0.76110.98740.89000.89060.94060.91940.99310.95800.99240.97940.95610.88630.88610.83890.98610.9601
PPOAG0.81790.84020.82850.88320.84550.84490.84500.85720.85590.84630.85240.85840.81740.84900.82210.8538
PPOCor0.81950.86590.86550.84390.89530.87630.84890.87370.88520.87550.89930.83370.90560.90000.90630.8761
ProPhTW0.75120.79070.76860.80490.80760.80150.80160.81320.81610.81110.80360.78630.78940.81000.78070.8043
RefDev0.40800.67240.78280.58130.77720.73970.61160.65180.79090.72680.73000.70880.78680.53300.75940.5697
Rock0.62000.77470.80270.85250.86730.85470.76000.78200.85530.83200.80470.70530.41600.38000.62730.6133
ScreTyp0.42670.57200.58480.52000.71790.59590.47180.60590.72420.59420.60900.60070.75530.66700.70560.5392
SeHaGen0.74500.96310.88770.84640.93230.78140.94740.87000.96920.93890.92310.87060.80890.74330.88460.9521
SeHaMov0.66000.89090.66490.70940.78830.46030.86410.72400.88900.88500.65260.68250.42100.42000.55100.5916
SeHaSub0.70220.93810.83510.83690.91330.79350.91600.82410.95060.93310.91190.84870.55470.57110.76390.6121
ShapSim0.53890.78931.00000.95560.99960.99740.51370.76761.00001.00000.99810.99370.72720.86700.92350.8062
ShapAll0.83500.89010.90850.84170.86530.91600.80430.85010.93130.94110.92320.82930.93310.89800.93820.8148
SmKApp0.65330.73810.74710.79200.81280.80520.78820.80600.82830.83800.81830.81700.79660.80300.77060.8146
SmoSub0.78670.99800.40730.78000.93580.85530.98730.84840.98620.99730.97490.85530.99290.84670.98490.9833
SonSur10.73040.92010.89770.84360.80090.90930.86370.86700.82630.88970.95810.88340.96040.96800.95420.7630
SonSur20.85830.89900.87940.93390.93700.93530.87430.91250.93670.90110.93500.90230.96890.96200.95130.9016
StaCurv0.96220.97960.97660.97850.97860.98040.97040.97440.97960.98110.98100.97020.97200.96700.97810.9799
Strberry0.96220.96050.97050.96220.97170.97860.96750.97300.97510.97400.97870.92290.97490.96900.97550.9709
SweLeaf0.88480.95310.92020.92800.93400.95760.89790.92290.94940.96190.96320.88050.95860.96600.96980.8779
Symbols0.97090.96700.96320.88240.90140.95320.87790.91270.96850.97080.96860.94800.94670.96200.96950.8000
SynCont0.56670.99830.96660.98330.99190.98700.99160.67790.99420.99900.99780.96700.99440.99000.99581.0000
ToeSeg10.78510.83550.92490.96490.95340.94300.66710.88040.95960.95980.93290.81270.95420.96900.95320.9793
ToeSeg20.68460.88590.96150.90770.94510.92850.80260.91180.96820.96260.93260.83510.95310.91500.96360.9318
Trace1.00001.00001.00001.00001.00001.00000.99200.98301.00001.00001.00000.99971.00001.00001.00001.0000
TwLeECG0.99470.98180.98470.99740.99910.99750.87060.91070.99630.99010.99850.85390.99941.00000.99480.9288
TwoPatt0.99751.00000.99170.95500.99030.98140.99380.43900.99951.00001.00000.84860.99980.89701.00001.0000
UMD0.84720.95440.96640.95140.93680.93220.83330.54120.96740.98330.98290.86940.95250.94580.97990.8885
UWavAll0.67700.97310.94490.80290.95800.95780.96240.91740.96660.97070.97730.82640.86900.82600.95120.9705
UWavX0.58400.83140.75260.73030.82020.81760.80000.63390.83350.84680.85710.76850.79040.75400.83360.8460
UWavY0.57710.76720.66210.74850.74490.72580.72190.66820.75470.78800.78360.70440.67590.72500.77060.7835
UWavZ0.84920.76740.69550.94220.77250.75480.73340.66350.77510.79120.79580.70640.75140.72900.77320.7918
Wafer0.97780.99610.99891.00001.00000.99990.99660.99540.99990.99890.99860.99730.99890.99700.99861.0000
Wine0.51850.85620.89260.79630.88580.93020.86230.87100.89200.89810.91420.70000.85620.88900.88640.6333
WordSyn0.68030.77820.65840.57050.62300.71270.64790.59160.69320.79370.76440.54400.61330.58000.75180.6605
Worms0.58440.69310.72550.74030.73330.77840.61210.68660.71650.76840.72030.72510.75890.66900.77970.6522
WormsTC0.63640.76930.80780.83120.78530.80040.69350.78530.78960.78610.78960.79220.76800.72900.80350.7946
Yoga0.82000.88740.91020.81770.88000.89240.86580.83720.91240.87260.91380.80380.87720.84500.91230.9007
AVG Rank13.38398.00928.86619.06257.11617.758910.410710.04464.43754.62395.018311.60717.34827.87505.33956.5982
MPSE0.07460.04380.03900.04250.03490.03750.04810.04620.02930.03210.03550.05100.03980.03720.03200.0380
Table 3. Wilcoxson signed-rank test comparision of each model; the significance level (alpha) of 0.05 was selected for all statistical tests, bold values indicate smaller than 0.05.
Table 3. Wilcoxson signed-rank test comparision of each model; the significance level (alpha) of 0.05 was selected for all statistical tests, bold values indicate smaller than 0.05.
IcTFCNRNC2RKCHIHCTRISETSFWSSTCSTBOSSPFDTW
FCN 1.5 × 10 2
RN 2.3 × 10 1 2.0 × 10 1
C2 6.6 × 10 6 6.2 × 10 2 1.4 × 10 3
RK 7.8 × 10 1 2.9 × 10 2 4.1 × 10 1 2.6 × 10 5
CHI 6.3 × 10 1 6.7 × 10 3 1.1 × 10 1 3.2 × 10 6 4.5 × 10 1
HCT 4.5 × 10 1 1.7 × 10 3 6.0 × 10 2 1.4 × 10 7 2.8 × 10 1 8.1 × 10 1
RISE 4.8 × 10 4 3.9 × 10 1 2.6 × 10 2 2.8 × 10 1 1.2 × 10 3 2.3 × 10 4 1.6 × 10 5
TSF 1.8 × 10 3 4.6 × 10 1 4.6 × 10 2 2.8 × 10 1 3.9 × 10 3 6.0 × 10 4 7.6 × 10 5 9.1 × 10 1
WS 1.8 × 10 1 1.8 × 10 1 8.9 × 10 1 6.0 × 10 4 3.1 × 10 1 7.4 × 10 2 3.3 × 10 2 1.7 × 10 2 4.3 × 10 2
STC 3.6 × 10 1 8.6 × 10 2 8.2 × 10 1 8.2 × 10 5 5.7 × 10 1 1.7 × 10 1 7.5 × 10 2 3.8 × 10 3 1.3 × 10 2 6.8 × 10 1
ST 4.2 × 10 2 5.6 × 10 1 3.8 × 10 1 7.9 × 10 3 7.5 × 10 2 1.6 × 10 2 4.8 × 10 3 1.0 × 10 1 1.8 × 10 1 4.1 × 10 1 2.1 × 10 1
BOSS 2.0 × 10 1 2.7 × 10 1 8.6 × 10 1 2.3 × 10 3 2.9 × 10 1 8.5 × 10 2 3.2 × 10 2 4.0 × 10 2 5.9 × 10 2 8.7 × 10 1 5.9 × 10 1 5.8 × 10 1
PF 8.7 × 10 2 5.3 × 10 1 5.9 × 10 1 1.1 × 10 2 1.5 × 10 1 3.7 × 10 2 1.4 × 10 2 1.1 × 10 1 1.6 × 10 1 6.1 × 10 1 3.5 × 10 1 8.8 × 10 1 6.9 × 10 1
DTW 1.4 × 10 10 2.2 × 10 5 1.3 × 10 7 1.7 × 10 3 5.1 × 10 10 7.0 × 10 11 5.4 × 10 13 6.8 × 10 5 1.3 × 10 4 5.0 × 10 9 3.0 × 10 10 9.8 × 10 8 9.1 × 10 8 7.6 × 10 7
TSC-FF 4.6 × 10 1 1.4 × 10 1 7.2 × 10 1 7.8 × 10 4 6.3 × 10 1 2.8 × 10 1 1.4 × 10 1 1.8 × 10 2 2.6 × 10 2 7.3 × 10 1 9.3 × 10 1 3.0 × 10 1 6.0 × 10 1 3.7 × 10 1 2.7 × 10 8
Table 4. Average accuracy of all seven methods with different amounts of records in each category. Bold values denote the model with the best performance.
Table 4. Average accuracy of all seven methods with different amounts of records in each category. Bold values denote the model with the best performance.
Records in Each Category(0,20](20,50](50,100](100,200](200,500](500,1000](1000,+)
Distance-based methods
DTW0.72080.71860.74330.64970.81560.74010.6512
PF0.85930.83110.87450.74720.90030.27550.8439
Feature-based methods
BOSS0.87030.81480.84630.75430.89170.81740.8754
ST0.85740.83770.84170.75850.86600.76340.8419
STC0.86230.84970.86760.81100.90590.83370.9119
WS0.87210.83740.83500.77340.90370.82170.9126
TSF0.77040.80770.86890.74360.89270.80570.7999
RISE0.79980.78530.85770.74830.83150.81250.8938
Ensemble methods
HCT0.89770.86870.89030.82320.90670.82650.9173
CHI0.91470.85430.88440.78620.90210.96010.9106
RK0.90010.84840.86790.80740.90850.19550.9195
C20.77210.76380.83470.73750.85320.79190.8838
Deep learning methods
RN0.80600.87280.80810.79570.90560.84770.9109
FCN0.79360.83580.76600.77100.87410.82600.8373
Ict0.90090.85970.84390.81370.91270.92210.9301
TSC-FF0.83330.86400.83240.80270.87870.80960.9138
Table 5. Accuracy comparison of visual feature-based methods on 20 data sets in UCR. Bold values denote model with the best performance.
Table 5. Accuracy comparison of visual feature-based methods on 20 data sets in UCR. Bold values denote model with the best performance.
GASF-GADF-MTFTFRPTSC-FF
Adiac0.62700.79540.7875
Beef0.76700.63330.8444
CBF0.9910N/A1.0000
Coffee1.00000.96431.0000
ECGFiveDays0.91000.89550.9991
FaceAll0.76300.71010.8705
FaceFour0.93200.78410.9091
50words0.69900.56260.8041
fish0.88600.88000.9886
Gun Point0.92000.98001.0000
Lighting20.88600.85251.0000
Lighting70.74000.68490.8115
OliveOil0.80000.86670.8667
OSULeaf0.64200.92980.9695
SwedishLeaf0.93500.95040.8779
SyntheticControl0.9930N/A1.0000
Trace1.0000N/A1.0000
Two Patterns0.9090N/A1.0000
wafer1.00000.99981.0000
yoga0.80400.85870.9007
Average Rank2.05002.31251.2500
MPSE2.6500 × 10 2 3.1100 × 10 2 9.8900 × 10 3
Table 6. Accuracy comprison of TSC-FF with sub-models on 112 data sets.
Table 6. Accuracy comprison of TSC-FF with sub-models on 112 data sets.
TSC-FFLSTM-ALSTMCNN-ACNN
ACSF10.85520.82780.78000.86600.8400
Adiac0.78750.76840.61990.73990.6513
ArrowHead0.80250.85860.75800.81850.6822
Beef0.84440.87040.82220.78520.7963
BeetleFly0.88890.83330.61110.81210.5023
BirdChicken0.92220.94440.84230.73420.7932
BME0.89210.84000.83330.92000.9133
Car0.95000.96670.88520.86670.8222
CBF1.00000.99890.85931.00000.6716
Chinatown0.98610.94780.95360.96520.9217
ChlCon0.72460.68920.63990.71270.6306
CinCECGtorso0.95110.87190.83000.97780.8911
Coffee1.00001.00001.00001.00001.0000
Computers0.72000.56000.54000.72000.6022
CricketX0.80770.79230.73560.76940.7279
CricketY0.80000.75560.64000.74100.5761
CricketZ0.82820.79260.54990.76670.5627
Crop0.76220.76360.75890.77120.7640
DiaSizeRed0.94710.88000.56730.92390.4982
DislPhaOutCor0.76610.74600.76210.76610.5968
DisPhaOutAgeGro0.71200.69600.40000.70400.4240
DisPhaTW0.68750.69600.49600.64800.4800
Earthquakes0.81600.80800.72000.73600.7440
ECG2000.89160.87240.82450.88690.8046
ECG50000.94810.94270.72300.94860.9442
ECGFiveDays0.99910.73560.80520.99980.7935
ElectricDevices0.74130.74200.73810.69380.6757
EOGHorSig0.83580.78180.75690.80110.7707
EOGVerSig0.77220.75410.73480.75690.7486
EthLevel0.84420.80850.78640.81820.7485
FaceAll0.87050.85670.73440.88240.7879
FaceFour0.90910.86080.27850.88610.2785
FacesUCR0.95930.97860.86180.83900.8119
FiftyWords0.80410.78220.43180.79220.4320
Fish0.98860.94710.80190.91340.8338
FordA1.00001.00001.00001.00001.0000
FordB1.00001.00001.00001.00001.0000
FreRegTra0.99890.99400.99610.99790.9965
FreSmaTra0.96680.96600.95470.96420.9323
GunPoint1.00000.95560.68890.98520.5333
GPAgSpa1.00001.00001.00001.00001.0000
GPMalVerFem1.00001.00001.00001.00000.9968
GPOldVerYou0.92640.92440.91990.90110.8507
Ham0.71280.64890.61700.70210.4787
HandOutlines0.90690.86790.36340.87390.6366
Haptics0.52130.38630.39710.46210.2986
Herring0.64910.63160.61400.50880.3860
HouTwe0.96130.94120.88240.96640.9496
InlineSkate0.46160.37470.44040.46940.4040
InsEPGRegTra1.00001.00001.00001.00001.0000
InsEPGSmaTra0.51500.47390.46260.52380.4610
InsWinSou0.62690.57520.51960.60660.3889
ItaPowDem0.96330.95140.95680.95360.4989
LarKitApp0.87230.82490.83980.85670.8205
Lightning20.81550.81260.79550.81750.7246
Lightning70.81150.80770.72310.79230.7769
Mallat0.95420.90950.83130.89480.8564
Meat0.93330.91480.91480.94070.8519
MedicalImages0.75780.74420.74560.69880.5146
MidPhaOutCor0.76630.77780.55560.72030.5556
MidPhaOutAgeGro0.70000.59420.50000.52170.5725
MidPhalanxTW0.57450.56520.47250.50720.4812
MixShaRegTra0.96900.97110.96910.86130.8566
MixShaSmaTra0.93860.91850.88180.92090.9177
MoteStrain0.89430.87570.33570.85350.5444
NonInvFatECGTho10.88180.91430.77660.84150.7538
NonInvFatECGTho20.94240.93940.83050.91000.8236
OliveOil0.86670.88520.87040.84070.7704
OSULeaf0.96990.87470.83460.97120.8502
PhaOutlCor0.80960.81870.60880.72930.6088
Phoneme0.30580.29140.24920.27280.2400
PigAirPre0.53700.55290.41830.56730.5106
PigArtPre0.97190.97120.92790.97600.9663
PigCVP0.52530.49330.43650.54440.5198
Plane1.00000.95740.84890.98940.7957
PowerCons0.96010.91110.90560.88330.8944
ProPhaOutCor0.85380.72800.68580.78540.6858
ProPhaOutAgeGro0.87610.78910.83260.83700.7707
ProPhalanxTW0.80430.75330.74890.79890.7315
RefrigeDevices0.56970.43620.48070.54600.3412
Rock0.61330.42130.37560.42650.3961
ScreenType0.53920.46200.42140.48870.3917
SHGenCh20.95210.81330.80000.96510.9517
SHMovCh20.59160.42440.39560.60440.4156
SHSubCh20.61210.55560.52890.61670.5644
ShapeletSim0.80620.81230.79380.79380.7338
ShapesAll0.81480.81480.74070.76480.7870
SmaKitApp0.81460.80200.60240.81670.6205
SmoSub0.98330.99330.98670.54550.5195
SonAIBORobSur10.76300.77590.77590.72590.7241
SonAIBORobSur20.90160.87340.80460.90160.8254
StarlightCurves0.97990.95410.96110.94930.8562
Strawberry0.97090.95770.93470.92590.8577
SwedishLeaf0.87790.89860.87720.85050.8270
Symbols0.80000.80780.78880.75980.6536
SyntheticControl1.00000.98520.99261.00000.3296
ToeSeg10.97930.92680.89510.97240.9463
ToeSeg20.93180.88630.91620.94470.8966
Trace1.00000.98890.98891.00000.7000
TwoLeadECG0.92880.78930.50340.97170.5034
TwoPatterns1.00001.00000.99940.99890.2478
UMD0.88850.95830.94750.86670.6478
UWaveGesLibAll0.97050.95070.74150.96560.9693
UWaveGesLibX0.84600.81770.78130.79860.7821
UWaveGesLibY0.78350.78170.66650.70090.7287
UWaveGesLibZ0.79180.75810.70560.71730.6440
Wafer1.00001.00001.00001.00001.0000
Wine0.63330.50000.60000.63330.6000
WordSynonyms0.66050.63480.62540.71450.6164
Worms0.65220.52170.59420.44930.5217
WormsTwoClass0.79460.75220.73620.73770.7348
Yoga0.90070.89330.79480.89300.5359
Table 7. Average test accuracy grouped by data set length.
Table 7. Average test accuracy grouped by data set length.
Length<8081–250251–450451–700701–1000>1000
LSTM-A0.79150.88400.84130.86580.70600.7143
CNN-A0.75770.89570.84960.81860.72050.7472

Share and Cite

MDPI and ACS Style

Wang, B.; Jiang, T.; Zhou, X.; Ma, B.; Zhao, F.; Wang, Y. Time-Series Classification Based on Fusion Features of Sequence and Visualization. Appl. Sci. 2020, 10, 4124. https://doi.org/10.3390/app10124124

AMA Style

Wang B, Jiang T, Zhou X, Ma B, Zhao F, Wang Y. Time-Series Classification Based on Fusion Features of Sequence and Visualization. Applied Sciences. 2020; 10(12):4124. https://doi.org/10.3390/app10124124

Chicago/Turabian Style

Wang, Baoquan, Tonghai Jiang, Xi Zhou, Bo Ma, Fan Zhao, and Yi Wang. 2020. "Time-Series Classification Based on Fusion Features of Sequence and Visualization" Applied Sciences 10, no. 12: 4124. https://doi.org/10.3390/app10124124

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop