Next Article in Journal
An Efficient Numerical Scheme for a Time-Fractional Black–Scholes Partial Differential Equation Derived from the Fractal Market Hypothesis
Previous Article in Journal
A Fractional-Order Creep-Damage Model for Carbonaceous Shale Describing Coupled Damage Caused by Rainfall and Blasting
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Analyzing Monofractal Short and Very Short Time Series: A Comparison of Detrended Fluctuation Analysis and Convolutional Neural Networks as Classifiers

by
Juan L. López
1,*,† and
José A. Vásquez-Coronel
2,*,†
1
Centro de Innovación en Ingeniería Aplicada, Universidad Católica del Maule, Av. San Miguel 3605, Talca 3460000, Chile
2
Department of Computer Science and Industries, Universidad Católica del Maule, Av. San Miguel 3605, Talca 3460000, Chile
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Fractal Fract. 2024, 8(8), 460; https://doi.org/10.3390/fractalfract8080460
Submission received: 25 June 2024 / Revised: 22 July 2024 / Accepted: 3 August 2024 / Published: 6 August 2024

Abstract

Time series data are a crucial information source for various natural and societal processes. Short time series can exhibit long-range correlations that reveal significant features not easily discernible in longer ones. Such short time series find utility in AI applications for training models to recognize patterns, make predictions, and perform classification tasks. However, traditional methods like DFA fail as classifiers for monofractal short time series, especially when the series are very short. In this study, we evaluate the performance of the traditional DFA method against the CNN-SVM approach of neural networks as classifiers for different monofractal models. We examine their performance as a function of the decreasing length of synthetic samples. The results demonstrate that CNN-SVM achieves superior classification rates compared to DFA. The overall accuracy rate of CNN-SVM ranges between 64 % and 98 % , whereas DFA’s accuracy rate ranges between 16 % and 64 % .

1. Introduction

Nature and society encompass numerous processes that exhibit fractal or multifractal behavior [1,2,3]. These processes yield valuable information through time series obtained from measurements or observations. However, these time series may be affected by non-stationary uncertainties arising from experiments or observations. Separating these uncertainties from the intrinsic fluctuations and correlations of the studied system is a highly intricate task, and several methods have been proposed to achieve this objective [4]. One such method, detrended fluctuation analysis (DFA), has proven effective in detecting long-range correlations in data with trends. It was introduced initially by [5]. Subsequently, in [6] the authors generalized this method to analyze multifractal time series (MFDFA), which has further been extended to multidimensional series [7], and to investigate power-law correlations between simultaneously recorded time series [8,9,10]. MFDFA has been favorably compared to other methods [11,12] and has found applications in various fields. For example, it has been utilized in geophysics [13], physiology [14,15,16], financial markets [17,18], and notably, the study of currency exchange rates [19,20]. In these cases, due to the non-linear characteristics, different techniques have been used to improve the characterization and interpretation of fluctuation and correlation for the interconnection between different physiological systems, including wavelet analysis (WT) [21], fractal dimension analysis [22,23], spectral analysis [24,25,26], detrended fluctuation analysis (DFA) [5,27], and multifractal detrended fluctuation analysis (MFDFA) [6]. Related to spectral analysis, if the time series is stationary, we can apply standard spectral analysis techniques and calculate the power spectrum C ( f ) of the time series as a function of the frequency f to determine its self-affine scaling behavior [28]. DFA and MFDFA perform well for time series with a length of at least 2 16 elements, but it is crucial to assess the methods performance on shorter time series [29]. Two main reasons drive the need for this evaluation: first, many relevant records are short in length, and second, there are processes where long records exist that exhibit changing behavior over time, prompting the study of short segments to gain significant insights.
Short time series are important in forecasting, anomaly detection, and trend analysis [30,31,32]. In [32], the authors describe a study of historical rainfall data from the Sindh River Basin, India, which were analyzed for monthly, seasonal, and annual trends. Likewise, Ref. [30] presents a short-term series analysis for the forecast of electricity demand in Singapore, where a multiplicative decomposition model and a seasonal ARIMA model are proposed to accurately predict short-term demand.
Short time series can reveal patterns and trends that can inform decision-making, for example, to identify short-term trends in stock prices or other market data [33]. In healthcare, short time series can be used to monitor patient health metrics, which can help doctors identify and respond to potential health issues before they become serious [34,35].
In machine learning applications, short time series can be useful to train models to recognize patterns and make predictions [36,37], which can be especially important in applications where data are scarce or expensive to collect. In [37], the authors implemented and compared several forecasting techniques and showed how these methods improve the root mean square error score, while [36] focused on the application of machine learning techniques, particularly recurrent neural networks (RNN), to predict pore water pressure time series data.
In fractal analysis, short time series can reveal important features of the fractal structure that might not be apparent in longer time series or estimate a measure of the degree of self-similarity or complexity of the time series [33,38,39,40]. Related to the length of short time series, making the distinction between short-term and very short-term depends on the context and the specific characteristics of the data. For example, in time-series analysis, very short-term variations typically refer to changes that occur over relatively brief time intervals, such as seconds or hours (market volatility, seismic activity, or heart rate variability). Now, by understanding the monofractal nature of a very short time series, better forecasting models can be developed. These models can be used for very-short-term predictions where rapid and accurate forecasting is essential [41,42]. Despite these advantages, analyzing monofractal very short time series also comes with challenges. The very short length of the series may limit the robustness of the fractal analysis, and care must be taken to ensure that the identified patterns are statistically significant.
As stated before, short time series can also be useful in machine learning applications to train models to recognize patterns, make predictions and perform classification tasks [43], while ref. [44] presents an overview of state-of-the-art AI-based electrocardiogram signal processing methods. Focusing on classification tasks, these are essential for extracting meaningful insights, making informed decisions, and automating tasks. Currently, with scientific and technological advancements, various methods and tools have emerged aimed at performing classification tasks. Classification processes using AI offer several advantages over traditional techniques, including higher accuracy, adaptability and learning, scalability, handling complex and non-linear relationships, and efficient automation, to name a few [45,46,47,48]. The choice of approach depends on the specific problem, the available data, the interpretability requirements, and other factors.
There are several AI methods used for the classification of time series that have shown promising results in natural language processing tasks and that are also applicable to time series analysis. In particular, deep learning has been used extensively in time series analysis, particularly for forecasting and anomaly detection [49,50,51,52]. Some deep learning models used for time series analysis include recurrent neural networks (RNNs) [53], long short-term memory (LSTM) networks [54], and convolutional neural networks (CNNs) [55,56,57,58,59,60]. Ref. [56] provides a comprehensive overview of deep learning techniques used for time series classification tasks. In addition, ref. [60] introduces a sequence-to-sequence (seq2seq) model, which utilizes RNNs for tasks like machine translation, discusses the encoder-decoder architecture, and demonstrates the effectiveness of RNNs in capturing sequential dependencies.
RNNs and LSTMs are particularly useful for modeling time series data because they are designed to handle sequences of varying lengths, while CNNs have also been used for time series analysis, particularly to analyze the structure and patterns within the data, and identify features that are relevant for prediction and classification tasks [61,62,63]. In [61], the authors introduce the Long Short-Term Memory (LSTM) architecture, a type of RNN that overcomes the vanishing gradient problem. It discusses the memory cell and the forget gate mechanism, which enable LSTMs to remember and forget information over long sequences, while in [63], the application of hidden Markov models to improve classification performance capturing the underlying structure and dependencies in time series was discussed.
To summarize the above, one source of information for many processes of interest in nature and society is time series. Short time series can reveal important features that might not be apparent in longer time series. They can be useful in AI applications for training models in classification tasks, forecasting, and anomaly detection. Additionally, analyzing monofractal short time series can make it easier to identify patterns and structures, be more efficient and effective in modeling and understanding the dynamics of the system of interest, and assist in the early detection of abnormalities or diseases.
Based on the highlighted antecedents, the contributions of this work are the following:
The significance of classifying monofractal time series lies in the insights it provides for classification, forecasting, and anomaly detection. While most studies focus on analyzing long-term series [29], short- and very-short-term series receive less attention. However, this research aims to expand the investigation into the performance and application of DFA and neural networks as classification methods for synthetic monofractal short-term series.
We compare two distinct approaches for classifying monofractal short and very short time series of different lengths: DFA and CNN-SVM.
Our findings show that CNN-SVM achieves higher classification rates than DFA, and both methods performance declines as the length of short and very short time series decreases.
The aim of this research, as a first approximation, focuses on monofractal models and their comparison with the expected outcomes based on analytical predictions when DFA and convolutional neural networks are applied as classifiers in synthetic short time series. The understanding of monofractal time series is of great importance as it can significantly enhance our ability to create better models, make more accurate predictions, and facilitate improved decision-making in a wide range of domains. This can be critical for understanding and predicting system behavior, especially in short time frames. In some cases, monofractal models may not adequately describe the system and multifractal models could be necessary. With this study, we aim to open new opportunities for the use of neural networks to identify monofractality and its changes in short and very short time series, where DFA and MFDFA fail, which could be of great help for the development of fast and accurate forecasting models.

2. Detrended Fluctuation Analysis (DFA)

Detrended fluctuation analysis (DFA) has emerged as a crucial method for investigating scaling characteristics and identifying long-range correlations within non-stationary and noisy time series [5,64]. The DFA technique, detailed in [5], can be outlined in five steps as follows:
  • Take a finite time series w ( i ) of length M, with a minor portion of w ( i ) elements being zero and compute a new time series W ( j ) , where j = 1 , , M :
    W ( j ) = i = 0 j w ( i ) < w > .
  • Divide the new series W ( j ) into M s segments of size s from the beginning and repeat the process starting from the end to obtain 2 M s segments.
  • Calculate the local trend of order m, denoted as P ν m , for all segments ν and all sizes s, and compute the variance as:
    F 2 ( ν , s ) = 1 s i = 1 s W ( ν 1 ) s + i P ν m ( i ) 2 .
  • Compute the second-order fluctuations as an average over all segments of a given size s:
    F 2 ( s ) = 1 2 M s ν = 1 2 M s F 2 ( ν , s ) 1 / 2 ,
  • For a range of sizes, s min < s < s max , observe the relationship:
    F 2 ( s ) s H .
    where H is the output of the DFA algorithm, known as the Hurst exponent.

3. Machine Learning Classification Approaches

3.1. Support Vector Machines

Support vector machines (SVM) are supervised learning algorithms originally designed for binary classification tasks [65]. However, thanks to their flexibility and capability, they have been successfully extended to multiclass schemes [66]. The most common approach applied to this type of problem is to split the task into multiple binary classifiers [67]. Then, each result is combined to obtain the final classification. Numerically, given N training samples ( x 1 , y 1 ) , , ( x N , y N ) , defined for k classes, with x i R m and y i { 1 , , k } , with y i being the class label of x i , the quadratic programming problem of the r-th SVM is as follows:
minimize : J ( w r , ζ r ) = 1 2 ( w r ) T w r + C j = 1 N ζ j r subject   to : y i [ ( w r ) T ϕ ( x i ) + d r ] 1 ζ i , r = 1 , , k ζ i 0 , i = 1 , , N
where w r R N is the weight vector, C is the penalty parameter, ζ j r is an error term between the misclassified point and the separation hyperplane, proportional to the C value, ϕ ( · ) is an operator that projects the input data to a higher dimensional space, and d is a scalar. The index r of the SVM schema indicates that it will be trained with input attributes corresponding to class r, assigning positive labels, while the remaining examples will receive negative labels.
This multiclass technique, known as one-versus-all, transforms the original problem into k binary classifiers, as detailed in Equation (5). After constructing the k dual classifiers, each associated with the decision function f r ( x ) = ( w r ) T ϕ ( x ) + d r , the maximal value of the k grouped classifiers predicts the class y of a sample x , that is,
y = a r g max r f r ( x ) .
In addition to this multiple classification strategy, there are other popular variants, such as the one-versus-one approach and hierarchical support vector machines [68]. The SVM classifier, originally designed to process linearly separable data, can be inefficient for non-linear tasks. To solve this problem, the literature introduced classifier systems based on the kernel function. In practice, the most common kernel types are the linear, Gaussian, polynomial, and radial basis functions [68].

3.2. Convolutional Neural Network

The convolutional neural network, a variant of deep learning, can extract relevant features of data without requiring human activity. Initially designed for image processing, it was subsequently extended to various disciplines due to its ability to capture high-level information. Although there are multiple variants, they all share three similar basic components, known as the convolutional, pooling, and fully connected layers. These elements, detailed in [69], can be summarized as follows:
Convolutional layer. A convolutional layer is the main building block of a CNN. It includes a set of filters or kernels, parameters that learn local informative features of the image, controlled by the padding and the stride. The padding controls the edges of the image, while the stride determines the number of pixels considered in the filter at each step during the convolution. The kernel operator, a matrix of weights, is applied to local regions of the input to generate output features, multiplying and summing values at each movement of the convolutional block (sliding window). Mathematically, the latent output representation of the s-th feature map of the current layer, denoted as F s , is as follows:
F s = g l L X l W s + b s ,
where W s and b s are the filters and the bias of the s-th feature map of the current layer, respectively, X l is the l-th feature map of the total feature maps ( L ) of the previous layer, ⊗ denotes the convolution operation, and g ( · ) is an activation function, normally non-linear.
The common activation functions include sigmoid, tanh, and ReLU. The first convolutional layer captures simple features, such as lines, while subsequent layers extract more complex features, such as shapes and specific objects [70,71].
Pooling layer. This layer reduces the resolution of the feature map in each channel through a pooling operator, maintaining the most relevant spatial features of the current convolutional layer. The common types of pooling layers include max pooling, average pooling, and global pooling. The mathematical expression corresponding to this layer is as follows:
P s = h ( F s ) ,
where P s is the output of the current pooling layer associated with the s-th feature map ( F s ) of the convolutional layer, and h ( · ) is the spatial reduction function.
Fully connected layers. These consist of a block of fully connected layers, trained with previously extracted features. At this stage, an optimization problem, which integrates a cost function, is the main support for the optimal configuration of the model parameters. The cost function quantifies the error between the model prediction y ^ i and the true label y i . To illustrate, consider the cross-entropy cost function. The associated minimization problem is defined as follows:
m i n i m i z e : J ^ ( θ ) = 1 N k = 1 N [ y k log ( y ^ k ) + ( 1 y k ) log ( 1 y ^ k ) ] ,
where J ^ ( · ) is the objective function, N is the number of samples, and θ is the set of weights and biases. Other cost functions incorporated into its architecture are the mean square error, hinge, and Huber. In addition, to obtain a more stable and robust performance, different regularization strategies are incorporated in Equation (9). These techniques include the L p -norm with p 0 , Dropout, and DropConnect. Mainly, they help to mitigate the problem of overfitting and deal with outlier data. In this classification or prediction stage, the multilayer perceptron network is the classical algorithm used. However, other machine learning models can be incorporated. Among the most popular are logistic regression, linear discriminant analysis, decision tree classifiers, SVM, and non-iterative methods.

4. Materials and Methods

4.1. Monofractal Time Series

A monofractal time series exhibits self-similarity at all time scales, where the scaling properties of the time series are described by a single fractal dimension which captures how the patterns and structures within the time series change as you change the scale [4,29,72]. This means that when you zoom in or out on different parts of the time series, you see similar patterns and structures as you would observe in the entire series. Monofractal time series are commonly found in many natural systems. They are often used in various fields of research, including economics, finance, and physics, to model, analyze, and compare the fractal properties of more complex time series [73,74,75,76]. In addition, monofractal time series appear in various systems like weather patterns, river flows, and physiological processes within the human body [77,78,79,80]. However, not all time series are monofractal; some time series may exhibit multi-fractal behavior, which means that their scaling properties are described by multiple fractal dimensions [6,29].

Monofractal Synthetic Data

Our primary goal is to understand the minimum series length that each model can reliably analyze and determine the precision achievable for these lengths. We concentrate exclusively on monofractal models as an initial step and compare the findings with their corresponding theoretical forecasts. Monofractal models are particularly intriguing for evaluating the methods performance due to their straightforward functional form for the Hurst exponent, which remains constant. We studied distinct monofractal models in a 0.1 < H < 0.9 range with step 0.1. The models studied correspond to white noise with a Hurst exponent (H) of 0.5 , signifying the absence of long-range correlations, and models with H values higher than 0.5 and less than 0.5 , indicating the presence of long-range correlations and anti-correlations in the data, respectively.
Artificial signals were created to evaluate the efficacy of various “neural network techniques” in categorizing monofractal time series, considering the signal’s length. For each of the synthetic models under investigation, we use series of various lengths, namely, 2 k , where k takes on the values of 10, 9, 8, and 7. To ensure a robust analysis, we produced multiple independent realizations for each case according to [81,82]. Specifically, we used 10 realizations for k = 20 from which short time series were obtained.
It is vital to emphasize that these investigations yield estimations of the potential shortest lengths and analysis precision. Nevertheless, they do not furnish definitive forecasts because real-world time series data are significantly more intricate and complex compared to the artificially generated signals employed in our study.

4.2. Environment

All the experiments in this study were performed in a controlled environment, using a laptop with the following characteristics: Server Supermicro 540A-TR 4U, equipped with the Windows 11 Operating System, 1CPU Xeon 6338 N 2 P 32 C/64 T, 2.2 G 24 M 11.2 Gt, RAM 8 × 32 GB DDR4-2Rx8 ECC REG DIMM, 1 Micron 7450 PRO 960 GB NVMe PCIe 4.0 M.2 × HDD 3.5” 4 TB, SATA 6 Cg/s 7.2 KRPM, 256 MB, 1 × GPU NVIDIA PNYQuadro TRXA4000 16 GB GDDR6 PCIe 4.0, KIT RAIL. Both the model configuration and the results analysis were executed using the official Matlab documentation.

4.3. Performance Metrics

We applied four different evaluation metrics to discuss the performance of the proposed multiple classification system on monofractal synthetic data. These metrics include accuracy (Acc), sensitivity (Sen), specificity (Spe), and positive predictive value (PPV). The following formulas summarize these machine learning statistics, defined in terms of true positive (TP), false positive (FP), true negative (TN), and false negative rates (FN):
Acc = T P + T N T P + T N + F P + F N
Sen = T P T P + F N
Spe = T N F P + T N
PPV = T P T P + F P
The Acc index is the relation between correct and incorrect predictions. It is an easy measure to interpret in classification tasks. The PPV provides the ratio of correct predictions to the total number of predictions. The higher the PPV, the better the model distinguishes between positive and negative classes. The Sen highlights its ability to provide the radius of correct predictions with respect to the actual number of cases. The Spe indicates its ability to correctly identify negative instances.
The following pseudo-code summarizes the proposed monofractal synthetic signal classification approach (Algorithm 1).
Algorithm 1 Monofractal synthetic signal classification process
1:
Datasets defined according to data length: 128, 256, 512 and 1024
2:
Convolutional layers of interest: 3, 4, 5, 6 and 7
3:
Classifiers used: MLP and SVM.
4:
Evaluation metrics: Acc, Sen, Spe, and PPV
5:
for  L = { 128 , 256 , 512 , 1024 }  do
6:
    Generation of monofractal synthetic signals of L length
7:
    Split the dataset into training and testing
8:
    for  C L = { 3 , 4 , 5 , 6 , 7 }  do
9:
        CNN model trained with C L convolutional layers
10:
        Obtain the feature map X c of the convolutional layer C L
11:
        for  M = { MLP , SVM }  do
12:
           Train M with the corresponding X c characteristics map.
13:
           Predict test labels
14:
           for  E M = { Acc , Sen , Spe , PPV }  do
15:
               Calculate the evaluation metric E M
16:
               Report the evaluation metric E M
17:
           end for
18:
        end for
19:
    end for
20:
end for

5. Results and Discussion

5.1. Performance of DFA on Monofractal Synthetic Data

Within this section, the performance of DFA as a classification method is evaluated using monofractal synthetic signals across varying series lengths. The primary aim here is to gain insights into the reliability of DFA when classifying the shortest synthetic series from each monofractal model. Specifically, we seek to determine the precision level achievable for such short lengths.
It is essential to emphasize that these studies provide estimations regarding the shortest possible lengths and the precision of the analysis. However, it is important to note that real-time series are significantly more complex than their synthetic counterparts. In the context of this study, all synthetic cases underwent detrending using a second-order polynomial in the third step of the DFA method.
The main assumption of the DFA method is that F 2 ( s ) behaves like s H within a specific range of s, allowing us to extract H using a log-log scale line fit. This principle has been confirmed through extensive mono-fractal series studies, as indicated by [7,29]. Our primary objective is to evaluate DFA’s ability to classify monofractal models while considering the diminishing length of the short series.
To ensure statistical reliability, we computed the average F 2 ( s ) , denoted as F 2 ( s ) , across all independent realizations of a particular model. Figure 1 shows the behavior of F 2 ( s ) for three representative H values: H = 0.5 for white noise, H = 0.7 for long-range correlations, and H = 0.3 for anti-correlations across the four lengths studied.
To simplify the analysis and reduce the reliance on H and s correlation, we focused on s > 20 . In this defined s range, each series realization displayed the power-law behavior outlined as F 2 ( s ) s H . We determined the Hurst exponents through linear least squares fits for each realization, categorizing them into a “Predicted Class” based on the extracted Hurst exponent (H) and comparing them with the “True Class” of the simulated monofractal models. The outcomes for all monofractal scenarios are illustrated in Figure 2, encompassing both the shorter and longer series lengths examined in this study. This visual representation shows that the DFA model faces greater difficulties in distinguishing samples of length 128 compared with signals of size 1024. This weakness is noticeable since the error spread outside the main diagonal of the H-index is wider for shorter signals.
The performance assessment of DFA as a classifier is summarized in Table 1. Specifically, the table presents the proportion of samples correctly classified under each H-index. Moreover, for the overall accuracy, the best performance was observed for length L = 1024 , achieving corresponding accuracy percentages of 16.29%, 34.92%, 48.73%, and 63.64% for lengths 128, 256, 512, and 1024, respectively.

5.2. Performance of CNN-SVM on Monofractal Synthetic Data

5.2.1. Training of the Deep Learning Model

The synthetic data were organized according to the length of the monofractal synthetic signal (128, 256, 512, and 1024), each containing 53,820 samples. Subsequently, we evaluated the performance of the CNN-SVM classifier following the classical five-fold cross-validation scheme. This training and testing methodology avoids the use of fixed parts of the dataset, an important condition for disseminating unbiased and precise results. Each designed dataset was divided into five folds of uniform size. At each iteration, one fold was reserved as test data and trained with the other four parts. The overall results were obtained as the average value of the five iterations.
The proposed CNN model includes seven depth levels, focused on deep feature extraction. The output of the convolutional layers was normalized by applying the batch-normalized technique, followed by the ReLu function. These powerful tools significantly improve the performance of the network, making the training faster and more stable. Table 2 summarizes the structure of the designed CNN, including the layer type, filter size, padding, stride, and activation function. For feature extraction, the crossentropyex cost function was incorporated into its architecture, adapted to the nine H-index categories. During the training phase, we used the stochastic gradient descent with momentum (SGDM) algorithm as the optimizer, with a learning rate of 0.01 and a momentum factor of 0.0009. We also employed an L 2 -regularizer of 0.0001, a batch size of 64, and 4 training epochs. After training the CNN, the SVM replaces the fully connected layers to classify the monofractal synthetic data. In this process, the SVM was trained with CNN features and evaluated with independent data, organized according to the cross-validation scheme. Hinge was the cost function, and SGD was the training algorithm, with a learning rate of 0.01. The linear kernel and the constant C = 1 were other settings.
To evaluate the potential of the deep features extracted by the CNN network, we integrated the coded features of the last five depth levels with the SVM classifier. Precisely, the SVM was trained using the output of convolutional layers 3, 4, 5, 6, and 7. These generated models were named CNN-SVM3, CNN-SVM4, CNN-SVM5, CNN-SVM6, and CNN-SVM7. The network topology, detailed in Table 2, was invariant for all the databases designed in this study. To speed up the training, we adopted an empirical or manual approach for hyperparameter tuning. In all the experiments designed, the models showed good results, assigning the same setting of values.
In our study, we addressed the problem of overfitting and underfitting using various machine learning strategies, such as regularization techniques ( L 2 -norm and batch normalization), evaluation of the CNN model with different convolutional layers, optimal hyperparameter search (learning rate, batch size, among others), and incorporation of the five-fold cross-validation scheme. The L 2 -norm regulated the complexity of the model and improved its generalization capability, while batch normalization stabilized the training and accelerated the convergence of the optimization approach. To avoid underfitting, we evaluated the classification system by adding convolutional layers progressively, thus capturing complex patterns of the monofractal signals across various feature maps. We also validated the experiments thoroughly using the five-fold cross-validation scheme, reporting unbiased evaluation metrics, as shown in Figure 3.

5.2.2. Analysis of Classification Metrics of Monofractal Synthetic Data

Short-duration monofractals quantify the degree and distribution of irregularities in a signal. With the synthetic data designed, we start by discussing the overall accuracy of the CNN-SVM model, as shown in Table 3. The deep learning system evaluated its performance by focusing on the depth of the feature map, with signals of lengths 128, 256, 512, and 1024. For the short synthetic series of size 128, the models from CNN-SVM3 to CNN-SVM7 obtained an overall accuracy of 64%, 67.06%, 67.89%, 68.12%, and 68.01%, respectively. Meanwhile, their accuracy rate with the larger length series was 97.66%, 97.98%, 97.83%, 97.85%, and 98.22%. The other results correspond to lengths of 256 and 512. The summarized table shows that network depth was a key factor in the generalization capacity of the model. Likewise, as expected, the CNN-SVM demonstrated better accuracy rates as the length of the synthetic signal increased. In general, both CNN-SVM6 and CNN-SVM7 tended to provide better performance on the multiple classification task. Although the overall accuracy of CNN-SVM ranged between 64% and 68% on shorter signals, these results remain relevant compared with the DFA scheme.
The bar chart in Figure 3 shows the performance of the CNN-SVM model on different training and test sets derived from the cross-validation scheme. Each bar represents the overall accuracy of the CNN-SVM models on each of the five folds, using signals of sizes 128 and 1024. Through this visualization, we conclude that the deep learning approach is not sensitive to data partitioning in the classification system. The stability of the performance supports the objectivity of the overall accuracy in Table 3, ensuring robust and reliable results.
Although, as a general rule, accuracy is a very good indicator of performance, a more thorough classification analysis requires careful examination of other evaluation metrics, such as Acc, PPV, Sen, and Spe. Following these evaluation criteria, in Table 4, we present the performance of CNN-SVM6, while in Table 5, we show the prediction statistics of CNN-SVM7. These tables organize the model information according to the length of the monofractal series, indicating in each row the value of the metric associated with one of the nine categories. The values listed are the average of the results obtained in the five folds of the cross-validation scheme. Indeed, the classifiers obtained better results with data of size 1024. In Table 4, for the signals with length 128, the CNN-SVM6 model achieved an average Acc of 92.9% in the nine categories. The average Sen for all categories was 68.1%, indicating that the model correctly identified 68.1%% of the total true positive cases. Likewise, the average Spe reached 96%, suggesting that the model was able to correctly identify 96% of the total true negative cases. In addition, an average PPV of 68% was observed, indicating that 68% of all positive predictions made by the model were correct. Similar reasoning can be applied to the results obtained by the CNN-SVM7 network. This confirms that the proposed CNN-SVM model provides an optimal classification process for monofractal synthetic signals.
The confusion matrix is a fundamental tool to summarize the correct and incorrect predictions in classification tasks, introduced in our study to deepen our previous analysis. Figure 4 and Figure 5 overview the performance of the CNN-SVM6 and CNN-SVM7 models, respectively, for data with sizes 128 and 1024. Each matrix was selected from the cross-validation scheme, considering the fold with the highest accuracy rate. These representations facilitate a more precise distinction between the indices TP, FP, TN, and FN. For the CNN-SVM7 approach with data of length 1024 in Figure 5, it is observed that, for the H 0.1 category, 0.51% of the signals were misclassified as H 0.2 . Regarding the category H 0.2 , 0.85% of the signals were misclassified as H 0.1 and 0.34% as H 0.3 . Similarly, 1.53% and 1.02% of the signals in the H 0.3 category were incorrectly labeled as H 0.2 and H 0.4 , respectively. For the H 0.4 index, 0.66% of the signals were incorrectly classified as H 0.3 , and 1.49% as H 0.5 series. A similar criterion can be followed for the analysis of the categories H 0.5 , H 0.6 , H 0.7 , H 0.8 , and H 0.9 . The discussion of the other confusion matrices follows a similar evaluation scheme. In addition, when the signal size increases, the H-index classification with the CNN-SVM6 and CNN-SVM7 models tends to concentrate on the main diagonal. Comparing Figure 2 with Figure 4 and Figure 5, the performance of the proposed CNN-SVM is remarkable versus the traditional DFA scheme.
The results presented in this study show the successful performance of the CNN-SVM method over the traditional DFA approach in the classification of short-length synthetic data. However, there are some limitations that we should consider in the analysis of future results. Below, we summarize some disadvantages of this research and make recommendations for future work.
The limited amount of monofractal data may limit the training and generalization capabilities of the DFA and CNN-SVM models. Likewise, due to their short length, they are more likely to be susceptible to noise, representing an additional challenge in capturing complex temporal patterns. Another inherent disadvantage of these classifiers is the processing time, especially when the complexity of the time series increases.
Our research focuses on the classification of short-length synthetic data, which may restrict the model’s ability to generalize to prediction tasks that require longer series. In addition, monofractal analysis may not be suitable for numerous application problems that demand multifractal models. This implies that the data may not fully reflect the complexity of the task, sometimes skewing the model’s performance.
The study compares the DFA and CNN-SVM schemes; nevertheless, it would be desirable to extend the evaluation of the classification system to other advanced machine learning models currently available.
In future work, we intend to link this research with applications that may be of interest to humanity. For example, through synthetic monofractal signals, we can detect types of heart disease and categorize the degree of the disease using deep learning techniques or machine learning algorithms. This implementation will be the basis for the analysis of more complex models, which could improve the classification rate by carefully tuning certain parameters.

6. Conclusions

Monofractal synthetic data have variable statistical properties that model a variety of natural phenomena, from fluid turbulence to cardiovascular signal analysis. The short-term series excel in this field, in part due to the lack of appropriate processing methods. In this study, we analyzed the performance of two machine learning approaches in the classification of short-duration monofractal data. Specifically, the traditional DFA scheme and a CNN-SVM neural network were considered. Both models were trained and validated using four datasets, defined according to the signal size. Compared with the DFA algorithm, the proposed CNN-SVM excelled as the best model in the classification task. The results exhibited remarkable overall accuracy in the four designed scenarios, with values that ranged between 64% and 98%. In particular, the classification rate of 64% corresponded to signals of size 128, while for signals of length 1024, it reached an impressive rate of 98%. In contrast, the DFA showed limited performance, with a maximum accuracy of 64% on longer signals and only 16% on shorter signals. Even the CNN-SVM performance with three convolutional layers obtained better evaluation statistics. These results underline the superiority of the proposed CNN-SVM method in the classification of synthetic data of short lengths, offering promising perspectives for practical applications in fields such as medicine and engineering.

Author Contributions

J.L.L.: Conceptualization, methodology, software, experimental execution and validation, research, writing—original draft preparation, writing—review, editing, visualization, supervision, project administration, and funding acquisition; J.A.V.-C.: Conceptualization, methodology, software, experimental execution and validation, research, and writing—original draft preparation. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by project grants “Fondecyt 11230276” (J. L. López) from the National Agency for Research and Development (ANID) of the Chilean government.

Data Availability Statement

The developed codes for this research are available at https://github.com/jlophys, which can be downloaded freely. Any questions regarding the codes can be directed to the corresponding author.

Acknowledgments

The authors acknowledge Centro de Innovación en Ingeniría Aplicada (CIIA) for permitting the use of their facilities as well as Luis Morán for their technical assistance and Viviana Torres for administrative support.

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

Abbreviations

The following abbreviations are used in this manuscript:
CNNConvolutional Neural Network
SVMSupport Vector Machines
ReLURectified Linear Unit
PPVPositive Predictive Value
TPTrue Positive
FPFalse Positive
TNTrue Negative
FNFalse Negative
MFDFAMonofractal Detrended Fluctuation Analysis
DFADetrended Fluctuation Analysis
WTWavelet Analysis
LSTMLong Short-term Memory
RNNsRecurrent Neural Networks
seq2seqSequence-to-sequence

References

  1. Mandelbrot, B.B.; Mandelbrot, B.B. Intermittent turbulence in self-similar cascades: Divergence of high moments and dimension of the carrier. Multifractals 1/ƒ Noise Wild-Self-Affin. Phys. (1963–1976) 1999, 317–357. [Google Scholar] [CrossRef]
  2. Bunde, A.; Havlin, S. Fractals in Science; Springer: Berlin/Heidelberg, Germany, 2013. [Google Scholar]
  3. Ivanov, P.C.; Amaral, L.A.N.; Goldberger, A.L.; Havlin, S.; Rosenblum, M.G.; Struzik, Z.R.; Stanley, H.E. Multifractality in human heartbeat dynamics. Nature 1999, 399, 461–465. [Google Scholar] [CrossRef]
  4. Kantelhardt, J.W. Fractal and multifractal time series. arXiv 2008, arXiv:0804.0747. [Google Scholar]
  5. Peng, C.K.; Buldyrev, S.V.; Havlin, S.; Simons, M.; Stanley, H.E.; Goldberger, A.L. Mosaic organization of DNA nucleotides. Phys. Rev. E 1994, 49, 1685–1689. [Google Scholar] [CrossRef] [PubMed]
  6. Kantelhardt, J.W.; Zschiegner, S.A.; Koscielny-Bunde, E.; Havlin, S.; Bunde, A.; Stanley, H. Multifractal detrended fluctuation analysis of nonstationary time series. Phys. Stat. Mech. Its Appl. 2002, 316, 87–114. [Google Scholar] [CrossRef]
  7. Gu, G.F.; Zhou, W.X. Detrended fluctuation analysis for fractals and multifractals in higher dimensions. Phys. Rev. E 2006, 74, 061104. [Google Scholar] [CrossRef]
  8. Podobnik, B.; Stanley, H.E. Detrended cross-correlation analysis: A new method for analyzing two nonstationary time series. Phys. Rev. Lett. 2008, 100, 084102. [Google Scholar] [CrossRef] [PubMed]
  9. Podobnik, B.; Horvatic, D.; Petersen, A.M.; Stanley, H.E. Cross-correlations between volume change and price change. Proc. Natl. Acad. Sci. USA 2009, 106, 22079–22084. [Google Scholar] [CrossRef]
  10. Zhou, W.X. Multifractal detrended cross-correlation analysis for two nonstationary signals. Phys. Rev. E 2008, 77, 066211. [Google Scholar] [CrossRef]
  11. Oświecimka, P.; Kwapień, J.; Drożdż, S. Wavelet versus detrended fluctuation analysis of multifractal structures. Phys. Rev. E 2006, 74, 016103. [Google Scholar] [CrossRef]
  12. Kantelhardt, J.W.; Rybski, D.; Zschiegner, S.A.; Braun, P.; Koscielny-Bunde, E.; Livina, V.; Havlin, S.; Bunde, A. Multifractality of river runoff and precipitation: Comparison of fluctuation analysis and wavelet methods. Phys. Stat. Mech. Its Appl. 2003, 330, 240–245. [Google Scholar] [CrossRef]
  13. Kavasseri, R.G.; Nagarajan, R. A multifractal description of wind speed records. Chaos Solitons Fractals 2005, 24, 165–173. [Google Scholar] [CrossRef]
  14. Makowiec, D.; Gała, R.; Dudkowska, A.; Rynkiewicz, A.; Zwierz, M. Long-range dependencies in heart rate signals—Revisited. Phys. Stat. Mech. Its Appl. 2006, 369, 632–644. [Google Scholar] [CrossRef]
  15. Dutta, S.; Ghosh, D.; Samanta, S.; Dey, S. Multifractal parameters as an indication of different physiological and pathological states of the human brain. Phys. Stat. Mech. Its Appl. 2014, 396, 155–163. [Google Scholar] [CrossRef]
  16. Gospodinova, E.; Lebamovski, P.; Georgieva-Tsaneva, G.; Negreva, M. Evaluation of the Methods for Nonlinear Analysis of Heart Rate Variability. Fractal Fract. 2023, 7, 388. [Google Scholar] [CrossRef]
  17. Oświe, P.; Kwapień, J.; Drożdż, S. Multifractality in the stock market: Price increments versus waiting times. Phys. Stat. Mech. Its Appl. 2005, 347, 626–638. [Google Scholar]
  18. Wang, L.; Lee, R.S. Stock Index Return Volatility Forecast via Excitatory and Inhibitory Neuronal Synapse Unit with Modified MF-ADCCA. Fractal Fract. 2023, 7, 292. [Google Scholar] [CrossRef]
  19. Wang, Y.; Wu, C.; Pan, Z. Multifractal detrending moving average analysis on the US Dollar exchange rates. Phys. Stat. Mech. Its Appl. 2011, 390, 3512–3523. [Google Scholar] [CrossRef]
  20. Oh, G.; Eom, C.; Havlin, S.; Jung, W.S.; Wang, F.; Stanley, H.E.; Kim, S. A multifractal analysis of Asian foreign exchange markets. Eur. Phys. J. B 2012, 85, 1–6. [Google Scholar] [CrossRef]
  21. Ebrahimi, F.; Setarehdan, S.K.; Ayala-Moyeda, J.; Nazeran, H. Automatic sleep staging using empirical mode decomposition, discrete wavelet transform, time-domain, and nonlinear dynamics features of heart rate variability signals. Comput. Methods Programs Biomed. 2013, 112, 47–57. [Google Scholar] [CrossRef]
  22. Gough, N.A.J. Fractal analysis of foetal heart rate variability. Physiol. Meas. 1993, 14, 309–315. [Google Scholar] [CrossRef]
  23. Shah, E.; Reddy, N.; Rothschild, B. Fractal analysis of acceleration signals from patients with CPPD, rheumatoid arthritis, and spondyloarthroparthy of the finger joint. Comput. Methods Programs Biomed. 2005, 77, 233–239. [Google Scholar] [CrossRef]
  24. Shiomi, T.; Guilleminault, C.; Sasanabe, R.; Hirota, I.; Maekawa, M.; Kobayashi, T. Augmented very low frequency component of heart rate variability during obstructive sleep apnea. Sleep 1996, 19, 370–377. [Google Scholar] [CrossRef]
  25. Hadase, M.; Azuma, A.; Zen, K.; Asada, S.; Kawasaki, T.; Kamitani, T.; Kawasaki, S.; Sugihara, H.; Matsubara, H. Very Low Frequency Power of Heart Rate Variability is a Powerful Predictor of Clinical Prognosis in Patients With Congestive Heart Failure. Circ. J. Off. J. Jpn. Circ. Soc. 2004, 68, 343–347. [Google Scholar] [CrossRef]
  26. Usui, H.; Nishida, Y. The very low-frequency band of heart rate variability represents the slow recovery component after a mental stress task. PLoS ONE 2017, 12, e0182611. [Google Scholar] [CrossRef]
  27. Koscielny-Bunde, E.; Bunde, A.; Havlin, S.; Goldreich, Y. Analysis of daily temperature fluctuations. Phys. Stat. Mech. Its Appl. 1996, 231, 393–396. [Google Scholar] [CrossRef]
  28. Otsuka, K.; Nakajima, M.S.Y.K.; Yamanaka, T. Heart rate variability including 1/f fluctuations versus conventional autonomic functions. J. Ambul. Monit. 1995, 8, 91–100. [Google Scholar]
  29. López, J.L.; Contreras, J.G. Performance of multifractal detrended fluctuation analysis on short time series. Phys. Rev. E 2013, 87, 022918. [Google Scholar] [CrossRef]
  30. Deng, J.; Jirutitijaroen, P. Short-term load forecasting using time series analysis: A case study for Singapore. In Proceedings of the 2010 IEEE Conference on Cybernetics and Intelligent Systems, Singapore, 28–30 June 2010; pp. 231–236. [Google Scholar]
  31. Braei, M.; Wagner, S. Anomaly detection in univariate time-series: A survey on the state-of-the-art. arXiv 2020, arXiv:2004.00433. [Google Scholar]
  32. Gajbhiye, S.; Meshram, C.; Mirabbasi, R.; Sharma, S. Trend analysis of rainfall time series for Sindh river basin in India. Theor. Appl. Climatol. 2016, 125, 593–608. [Google Scholar] [CrossRef]
  33. Wu, Y.; Shang, P.; Li, Y. Modified generalized multiscale sample entropy and surrogate data analysis for financial time series. Nonlinear Dyn. 2018, 92, 1335–1350. [Google Scholar] [CrossRef]
  34. Accardo, A.; Affinito, M.; Carrozzi, M.; Bouquet, F. Use of the fractal dimension for the analysis of electroencephalographic time series. Biol. Cybern. 1997, 77, 339–350. [Google Scholar] [PubMed]
  35. Dlask, M.; Kukal, J.; Poplová, M.; Sovka, P.; Cifra, M. Short-time fractal analysis of biological autoluminescence. PLoS ONE 2019, 14, e0214427. [Google Scholar] [CrossRef] [PubMed]
  36. Wei, X.; Zhang, L.; Yang, H.Q.; Zhang, L.; Yao, Y.P. Machine learning for pore-water pressure time-series prediction: Application of recurrent neural networks. Geosci. Front. 2021, 12, 453–467. [Google Scholar] [CrossRef]
  37. Chaurasia, V.; Pal, S. Application of machine learning time series analysis for prediction COVID-19 pandemic. Res. Biomed. Eng. 2020, 38, 35–47. [Google Scholar] [CrossRef]
  38. Delignieres, D.; Ramdani, S.; Lemoine, L.; Torre, K.; Fortes, M.; Ninot, G. Fractal analyses for short time series: A re-assessment of classical methods. J. Math. Psychol. 2006, 50, 525–544. [Google Scholar] [CrossRef]
  39. Li, R.; Wang, J.; Yu, H.; Deng, B.; Wei, X.; Chen, Y. Fractal analysis of the short time series in a visibility graph method. Phys. Stat. Mech. Its Appl. 2016, 450, 531–540. [Google Scholar] [CrossRef]
  40. Gao, J.; Hu, J.; Tung, W.W.; Zheng, Y. Multiscale analysis of economic time series by scale-dependent Lyapunov exponent. Quant. Financ. 2013, 13, 265–274. [Google Scholar] [CrossRef]
  41. Kleiger, R.E.; Stein, P.K.; Bosner, M.S.; Rottman, J.N. Time domain measurements of heart rate variability. Cardiol. Clin. 1992, 10, 487–498. [Google Scholar] [CrossRef]
  42. Look AHEAD Research Group. Long-term effects of a lifestyle intervention on weight and cardiovascular risk factors in individuals with type 2 diabetes mellitus: Four-year results of the Look AHEAD trial. Arch. Intern. Med. 2010, 170, 1566–1575. [Google Scholar]
  43. Sharma, S.; Guleria, K. Deep learning models for image classification: Comparison and applications. In Proceedings of the 2022 2nd International Conference on Advance Computing and Innovative Technologies in Engineering (ICACITE), Greater Noida, India, 28–29 April 2022; pp. 1733–1738. [Google Scholar]
  44. Ardeti, V.A.; Kolluru, V.R.; Varghese, G.T.; Patjoshi, R.K. An Overview on State-of-the-Art Electrocardiogram Signal Processing Methods: Traditional to AI-Based Approaches. Expert Syst. Appl. 2023, 217, 119561. [Google Scholar] [CrossRef]
  45. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 770–778. [Google Scholar]
  46. Miotto, R.; Wang, F.; Wang, S.; Jiang, X.; Dudley, J.T. Deep learning for healthcare: Review, opportunities and challenges. Briefings Bioinform. 2018, 19, 1236–1246. [Google Scholar] [CrossRef] [PubMed]
  47. Gal, Y.; Ghahramani, Z. A theoretically grounded application of dropout in recurrent neural networks. Adv. Neural Inf. Process. Syst. 2016, 29. [Google Scholar] [CrossRef]
  48. Li, M.; Andersen, D.G.; Park, J.W.; Smola, A.J.; Ahmed, A.; Josifovski, V.; Long, J.; Shekita, E.J.; Su, B.Y. Scaling distributed machine learning with the parameter server. In Proceedings of the 11th USENIX Symposium on Operating Systems Design and Implementation (OSDI 14), Broomfield, CO, USA, 6–8 October 2014; pp. 583–598. [Google Scholar]
  49. Canizo, M.; Triguero, I.; Conde, A.; Onieva, E. Multi-head CNN–RNN for multi-time series anomaly detection: An industrial case study. Neurocomputing 2019, 363, 246–260. [Google Scholar] [CrossRef]
  50. Lutsiv, N.; Maksymyuk, T.; Beshley, M.; Lavriv, O.; Andrushchak, V.; Sachenko, A.; Vokorokos, L.; Gazda, J. Deep Semisupervised Learning-Based Network Anomaly Detection in Heterogeneous Information Systems. Comput. Mater. Contin. 2022, 70, 413–431. [Google Scholar] [CrossRef]
  51. Hu, M.; Ji, Z.; Yan, K.; Guo, Y.; Feng, X.; Gong, J.; Zhao, X.; Dong, L. Detecting anomalies in time series data via a meta-feature based approach. IEEE Access 2018, 6, 27760–27776. [Google Scholar] [CrossRef]
  52. Demertzis, K.; Iliadis, L.; Tziritas, N.; Kikiras, P. Anomaly detection via blockchained deep learning smart contracts in industry 4.0. Neural Comput. Appl. 2020, 32, 17361–17378. [Google Scholar] [CrossRef]
  53. Uribarri, G.; Mindlin, G.B. Dynamical time series embeddings in recurrent neural networks. Chaos Solitons Fractals 2022, 154, 111612. [Google Scholar] [CrossRef]
  54. Seabe, P.L.; Moutsinga, C.R.B.; Pindza, E. Forecasting cryptocurrency prices using LSTM, GRU, and bi-directional LSTM: A deep learning approach. Fractal Fract. 2023, 7, 203. [Google Scholar] [CrossRef]
  55. Lim, B.; Zohren, S. Time-series forecasting with deep learning: A survey. PHilosophical Trans. R. Soc. A 2021, 379, 20200209. [Google Scholar] [CrossRef]
  56. Ismail Fawaz, H.; Forestier, G.; Weber, J.; Idoumghar, L.; Muller, P.A. Deep learning for time series classification: A review. Data Min. Knowl. Discov. 2019, 33, 917–963. [Google Scholar] [CrossRef]
  57. Chalapathy, R.; Chawla, S. Deep learning for anomaly detection: A survey. arXiv 2019, arXiv:1901.03407. [Google Scholar]
  58. Wang, Z.; Yan, W.; Oates, T. Time series classification from scratch with deep neural networks: A strong baseline. In Proceedings of the 2017 International Joint Conference on Neural Networks (IJCNN), Anchorage, AK, USA, 14–19 May 2017; pp. 1578–1585. [Google Scholar]
  59. Cui, Z.; Chen, W.; Chen, Y. Multi-scale convolutional neural networks for time series classification. arXiv 2016, arXiv:1603.06995. [Google Scholar]
  60. Sutskever, I.; Vinyals, O.; Le, Q.V. Sequence to sequence learning with neural networks. Adv. Neural Inf. Process. Syst. 2014, 27, page. [Google Scholar]
  61. Gers, F.A.; Schmidhuber, J.; Cummins, F. Learning to forget: Continual prediction with LSTM. Neural Comput. 2000, 12, 2451–2471. [Google Scholar] [CrossRef] [PubMed]
  62. Wang, S.; Xiang, J.; Zhong, Y.; Zhou, Y. Convolutional neural network-based hidden Markov models for rolling element bearing fault identification. Knowl. Based Syst. 2018, 144, 65–76. [Google Scholar] [CrossRef]
  63. Esmael, B.; Arnaout, A.; Fruhwirth, R.K.; Thonhauser, G. Improving time series classification using Hidden Markov Models. In Proceedings of the 2012 12th International Conference on Hybrid Intelligent Systems (HIS), Pune, India, 4–7 December 2012; pp. 502–507. [Google Scholar]
  64. Hu, K.; Ivanov, P.C.; Chen, Z.; Carpena, P.; Eugene Stanley, H. Effect of trends on detrended fluctuation analysis. Phys. Rev. E 2001, 64, 011114. [Google Scholar] [CrossRef] [PubMed]
  65. Cortes, C.; Vapnik, V. Support-vector networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
  66. Lorena, A.C.; De Carvalho, A.C.; Gama, J.M. A review on the combination of binary classifiers in multiclass problems. Artif. Intell. Rev. 2008, 30, 19–37. [Google Scholar] [CrossRef]
  67. Hsu, C.W.; Lin, C.J. A comparison of methods for multiclass support vector machines. IEEE Trans. Neural Netw. 2002, 13, 415–425. [Google Scholar]
  68. Jing, C.; Hou, J. SVM and PCA based fault classification approaches for complicated industrial process. Neurocomputing 2015, 167, 636–642. [Google Scholar] [CrossRef]
  69. Khanday, N.Y.; Sofi, S.A. Deep insight: Convolutional neural network and its applications for COVID-19 prognosis. Biomed. Signal Process. Control 2021, 69, 102814. [Google Scholar] [CrossRef] [PubMed]
  70. Ribeiro, M.; Lazzaretti, A.E.; Lopes, H.S. A study of deep convolutional auto-encoders for anomaly detection in videos. Pattern Recognit. Lett. 2018, 105, 13–22. [Google Scholar] [CrossRef]
  71. López, J.L.; Vásquez-Coronel, J.A. Congestive Heart Failure Category Classification Using Neural Networks in Short-Term Series. Appl. Sci. 2023, 13, 13211. [Google Scholar] [CrossRef]
  72. Lopes, R.; Betrouni, N. Fractal and multifractal analysis: A review. Med. Image Anal. 2009, 13, 634–649. [Google Scholar] [CrossRef] [PubMed]
  73. Shi, W.; Shang, P.; Wang, J.; Lin, A. Multiscale multifractal detrended cross-correlation analysis of financial time series. Phys. Stat. Mech. Its Appl. 2014, 403, 35–44. [Google Scholar] [CrossRef]
  74. Xu, Y.; Qian, C.; Pan, L.; Wang, B.; Lou, C. Comparing monofractal and multifractal analysis of corrosion damage evolution in reinforcing bars. PloS ONE 2012, 7, e29956. [Google Scholar] [CrossRef]
  75. Pastén, D.; Muñoz, V.; Cisternas, A.; Rogan, J.; Valdivia, J.A. Monofractal and multifractal analysis of the spatial distribution of earthquakes in the central zone of Chile. Phys. Rev. E 2011, 84, 066123. [Google Scholar] [CrossRef]
  76. Saâdaoui, F. Skewed multifractal scaling of stock markets during the COVID-19 pandemic. Chaos Solitons Fractals 2023, 170, 113372. [Google Scholar] [CrossRef]
  77. Huang, Z.W.; Liu, C.Q.; Shi, K.; Zhang, B. Monofractal and multifractal scaling analysis of pH time series from Dongting lake inlet and outlet. Fractals 2010, 18, 309–317. [Google Scholar] [CrossRef]
  78. Shi, K.; Liu, C.Q.; Ai, N.S. Monofractal and multifractal approaches in investigating temporal variation of air pollution indexes. Fractals 2009, 17, 513–521. [Google Scholar] [CrossRef]
  79. Curto-Risso, P.; Medina, A.; Hernández, A.C.; Guzman-Vargas, L.; Angulo-Brown, F. Monofractal and multifractal analysis of simulated heat release fluctuations in a spark ignition heat engine. Phys. Stat. Mech. Its Appl. 2010, 389, 5662–5670. [Google Scholar] [CrossRef]
  80. Lopez, J.L.; Hernández, S.; Urrutia, A.; López-Cortés, X.A.; Araya, H.; Morales-Salinas, L. Effect of missing data on short time series and their application in the characterization of surface temperature by detrended fluctuation analysis. Comput. Geosci. 2021, 153, 104794. [Google Scholar] [CrossRef]
  81. Abry, P.; Sellan, F. The wavelet-based synthesis for fractional Brownian motion proposed by F. Sellan and Y. Meyer: Remarks and fast implementation. Appl. Comput. Harmon. Anal. 1996, 3, 377–383. [Google Scholar] [CrossRef]
  82. Bardet, J.M.; Lang, G.; Oppenheim, G.; Philippe, A.; Taqqu, M.S. Generators of long-range dependent processes: A survey. Theory Appl. Long-Range Depend. 2003, 1, 579–623. [Google Scholar]
Figure 1. Representative average fluctuations at order 2 as a function of s for series of four different lengths and three monofractal models.
Figure 1. Representative average fluctuations at order 2 as a function of s for series of four different lengths and three monofractal models.
Fractalfract 08 00460 g001
Figure 2. Confusion matrix for DFA performance on synthetic short-length series for different monofractal models.
Figure 2. Confusion matrix for DFA performance on synthetic short-length series for different monofractal models.
Fractalfract 08 00460 g002
Figure 3. Overall accuracy of CNN-SVM for each fold in the cross-validation scheme.
Figure 3. Overall accuracy of CNN-SVM for each fold in the cross-validation scheme.
Fractalfract 08 00460 g003
Figure 4. Confusion matrix of CNN-SVM6 on synthetic short-length series of sizes 128 and 1024.
Figure 4. Confusion matrix of CNN-SVM6 on synthetic short-length series of sizes 128 and 1024.
Fractalfract 08 00460 g004
Figure 5. Confusion matrix of CNN-SVM7 on synthetic short-length series of sizes 128 and 1024.
Figure 5. Confusion matrix of CNN-SVM7 on synthetic short-length series of sizes 128 and 1024.
Fractalfract 08 00460 g005
Table 1. Overall accuracy of DFA for the classification of short-time synthetic signals.
Table 1. Overall accuracy of DFA for the classification of short-time synthetic signals.
H-IndexSignal Length: 128Signal Length: 256Signal Length: 512Signal Length: 1024
Acc (%)Acc (%)Acc (%)Acc (%)
H 0.1 28.233.136.844.3
H 0.2 25.252.668.280.2
H 0.3 20.249.264.881.0
H 0.4 17.239.456.274.1
H 0.5 14.033.950.967.4
H 0.6 12.230.344.761.2
H 0.7 10.825.641.256.3
H 0.8 9.925.840.156.0
H 0.9 8.924.535.852.2
Table 2. Structure of the CNN architecture for monofractal feature extraction.
Table 2. Structure of the CNN architecture for monofractal feature extraction.
Depth LevelLayer TypeFilter SizeStridePaddingActivation
1Convolutional 7 × 7 × 32 1sameReLU
Max Pooling 2 × 2 2
2Convolutional 5 × 5 × 64 1sameReLU
Max Pooling 2 × 2 2
3Convolutional 5 × 5 × 64 1sameReLU
Max Pooling 2 × 2 2
4Convolutional 5 × 5 × 64 1sameReLU
Max Pooling 2 × 2 2
5Convolutional 3 × 3 × 32 1sameReLU
Max Pooling 2 × 2 2
6Convolutional 3 × 3 × 16 1sameReLU
Max Pooling 2 × 2 2
7Convolutional 3 × 3 × 32 1sameReLU
Table 3. Overall accuracy of CNN-SVM for the classification of short-time synthetic signals.
Table 3. Overall accuracy of CNN-SVM for the classification of short-time synthetic signals.
DatasetSignal LengthOverall Accuracy (%)
CNN-SVM3CNN-SVM4CNN-SVM5CNN-SVM6CNN-SVM7
Short-time synthetic signals12864.0067.0667.8968.1268.01
25673.6478.6781.5481.8881.67
51290.3190.4891.6092.3592.44
102497.6697.9897.8397.8598.22
Table 4. Evaluation metrics of CNN-SVM6 for the classification of short-time synthetic signals.
Table 4. Evaluation metrics of CNN-SVM6 for the classification of short-time synthetic signals.
H-IndexSignal Length: 128Signal Length: 256Signal Length: 512Signal Length: 1024
AccPPVSenSpeAccPPVSenSpeAccPPVSenSpeAccPPVSenSpe
H 0.1 95.277.280.097.197.287.587.698.498.994.895.299.399.798.498.699.8
H 0.2 91.159.959.695.094.775.976.596.997.990.590.898.899.497.597.199.7
H 0.3 91.360.462.094.994.976.778.197.097.991.090.798.999.497.597.499.7
H 0.4 91.059.659.794.994.776.375.997.197.889.890.298.799.497.197.399.6
H 0.5 90.758.657.194.994.675.975.397.097.689.689.098.799.497.197.199.6
H 0.6 91.261.158.595.395.178.177.297.397.789.889.798.799.497.497.299.7
H 0.7 92.867.267.895.996.282.982.897.998.392.692.499.199.698.098.299.8
H 0.8 94.977.477.397.297.488.488.598.698.995.095.499.499.798.598.699.8
H 0.9 98.090.891.398.998.995.295.199.499.597.997.899.799.899.299.199.9
Table 5. Evaluation metrics of CNN-SVM7 for the classification of short-time synthetic signals.
Table 5. Evaluation metrics of CNN-SVM7 for the classification of short-time synthetic signals.
H-IndexSignal Length: 128Signal Length: 256Signal Length: 512Signal Length: 1024
AccPPVSenSpeAccPPVSenSpeAccPPVSenSpeAccPPVSenSpe
H 0.1 95.177.179.997.097.387.688.098.498.994.995.299.499.798.699.199.8
H 0.2 91.159.958.995.194.876.476.397.197.990.990.998.999.598.197.799.8
H 0.3 91.360.262.894.894.976.478.596.998.191.291.498.999.698.198.199.8
H 0.4 91.159.958.695.194.776.175.897.097.890.190.298.899.597.797.999.7
H 0.5 90.758.458.194.894.575.674.996.997.689.289.298.799.597.897.699.7
H 0.6 91.361.458.595.494.977.676.997.297.890.089.998.899.597.797.999.7
H 0.7 92.867.467.995.996.182.781.997.998.493.392.499.299.698.498.399.8
H 0.8 94.876.976.697.197.388.187.698.598.994.994.999.499.798.698.399.8
H 0.9 97.990.090.898.898.994.795.199.399.597.497.799.799.898.999.199.9
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

López, J.L.; Vásquez-Coronel, J.A. Analyzing Monofractal Short and Very Short Time Series: A Comparison of Detrended Fluctuation Analysis and Convolutional Neural Networks as Classifiers. Fractal Fract. 2024, 8, 460. https://doi.org/10.3390/fractalfract8080460

AMA Style

López JL, Vásquez-Coronel JA. Analyzing Monofractal Short and Very Short Time Series: A Comparison of Detrended Fluctuation Analysis and Convolutional Neural Networks as Classifiers. Fractal and Fractional. 2024; 8(8):460. https://doi.org/10.3390/fractalfract8080460

Chicago/Turabian Style

López, Juan L., and José A. Vásquez-Coronel. 2024. "Analyzing Monofractal Short and Very Short Time Series: A Comparison of Detrended Fluctuation Analysis and Convolutional Neural Networks as Classifiers" Fractal and Fractional 8, no. 8: 460. https://doi.org/10.3390/fractalfract8080460

APA Style

López, J. L., & Vásquez-Coronel, J. A. (2024). Analyzing Monofractal Short and Very Short Time Series: A Comparison of Detrended Fluctuation Analysis and Convolutional Neural Networks as Classifiers. Fractal and Fractional, 8(8), 460. https://doi.org/10.3390/fractalfract8080460

Article Metrics

Back to TopTop