Next Article in Journal
Federated and Centralized Machine Learning for Cell Segmentation: A Comparative Analysis
Next Article in Special Issue
Evaluating the Performance of DenseNet in ECG Report Automation
Previous Article in Journal
HistoMoCo: Momentum Contrastive Learning Pre-Training on Unlabeled Histopathological Images for Oral Squamous Cell Carcinoma Detection
Previous Article in Special Issue
Hypergraph Neural Network for Multimodal Depression Recognition
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Automated Selection of Time Series Forecasting Models for Financial Accounting Data: Synthetic Data Application

by
Rokas Štrimaitis
1,*,
Simona Ramanauskaitė
2 and
Pavel Stefanovič
1
1
Department of Information Systems, Vilnius Gediminas Technical University, LT-10223 Vilnius, Lithuania
2
Department of Information Technology, Vilnius Gediminas Technical University, LT-10223 Vilnius, Lithuania
*
Author to whom correspondence should be addressed.
Electronics 2025, 14(7), 1253; https://doi.org/10.3390/electronics14071253
Submission received: 28 January 2025 / Revised: 2 March 2025 / Accepted: 14 March 2025 / Published: 22 March 2025
(This article belongs to the Special Issue Digital Intelligence Technology and Applications)

Abstract

:
The continuous forecasting of anticipated trends in company accounting helps to prepare for possible challenges and investment possibilities. The forecasting of performance indicators for each partner and metric in a company requires a high volume of resources, as each forecasting model requires some supervised adjustment. To solve the challenge of manual work in this task, AutoML solutions can be used. In this study, we propose an automated forecasting model selection option. The time series data are summarized by 21 statistical features and different experiments are performed to find the best-suited forecasting method. Different classifier models are tested with the dataset, estimating the impact of data sampling and synthetic data generation. The results indicate that the undersampling approach is more suitable, as it helps in balancing the classes. Random forest methods usually show the best performance, achieving about 74% accuracy. The usage of a synthetic data-based dataset for model training reduced the accuracy by almost 20%, while the integration of synthetic and real time series data allowed us to achieve balance between both classes. This confirms that synthetic time series data generation might increase the accuracy of forecasting method selection, but it should be used in combination with real data.

1. Introduction

Data analytics and forecasting have become an important part of business management [1]. Artificial intelligence (AI) and machine learning (ML) methods are applied to analyze available data, and based on this, to predict upcoming results, possible risks, etc. In the current state, general AI is already capable of performing some common data analysis for financial tasks [2], but more specific models are used for certain data analysis and prediction tasks. This allows for more accurate model results to be obtained. However, building such an AI model requires experts to work to adjust the model or to conduct systematic experiments to analyze all possible model parameters and feature sets to find the most suitable configuration. In both cases, building an accurate AI model requires a good quality dataset and time [1,2]. The task of forecasting company financial accounting data for each client does not provide any of these resources. The following requirements and limitations apply to this task:
  • Companies have multiple clients with different collaboration patterns. Building one model to fit all clients is not feasible, considering the data variety of clients’ collaboration patterns.
  • The financial accounting data for each client have a set of monitored features (number of sales/purchases, amount, delays, discounts, etc.). The development of separate AI models for each client and each relevant financial accounting data feature would lead to an enormous number of models and consequently model building efforts.
  • New clients are coming and going; therefore, in the case of one AI model, it should be retrained or fine-tuned to reflect the collaboration pattern of the client. Otherwise, a new AI model or models have to be developed for this client specifically.
  • Forecasting tasks are very “live”, as data are changing constantly. This requires new model training after each data update.
  • The landscape of data forecasting models and their parameters is wide. The selection of a suitable method and the grid search of its parameters might be too costly and slow.
Automated Machine Learning (AutoML) is used to overcome these limitations and challenges. These are methods and processes designed to automate the application of machine learning in real-world situations [3]. The automation can include steps such as collecting and processing the dataset, selecting the model and tuning its parameters, and evaluating the results.
The main objective of this research is to automate the forecasting of financial accounting data for each company client and the monitored financial accounting data fields. This reflects a real situation when companies want to forecast different features for the selected clients and an AutoML solution is needed. They do not have time to manually build separate forecasting models when the forecast is needed immediately to make decisions on further steps. Therefore, we aim to reduce the data forecasting model selection time by proposing a model data forecasting method, which would allow companies to fully automate data forecasting, keeping the needed forecast accuracy.
This paper is oriented on two main research questions:
  • How does the utilization of synthetic data affect the data forecasting prediction method selection accuracy?
  • How accurately can the best forecasting method be predicted based on time series descriptive data?
To answer the research questions, in Section 2 we review the related work. Then, Section 3 summarizes the forecasting method prediction experiments with the existing dataset. The dataset, classification model selection, and the obtained results are presented there. Section 4 describes the dataset augmentation solution we propose and investigates the effect of different dataset balancing options. Section 5 and Section 6 summarize the research.

2. Related Work

Data forecasting involves predicting the future values of a specific variable based on historical data [4]. Typically, this forecasting is performed for a single variable, considering only its past values and excluding additional variables. Such data forecasting is commonly referred to as time series forecasting. The choice and application of existing data forecasting methods for financial accounting data can vary significantly, so understanding the different types and their unique characteristics is crucial.
One of the earliest approaches to time series forecasting was through stochastic methods [5]. Stochastic models make assumptions about the underlying process and associated probability distributions [6]. These models perform well when data changes are not overly complex, and the initial conditions meet the stationary assumption. Examples of stochastic time series forecasting models include VAR, ARIMA, and ARIMAX models [7], with variations proposed for seasonal data (such as SARIMAX [8]).
However, the requirement for data to be stationary limits the applicability of these models. Exponential smoothing is a solution for non-stationary data [9]. One such model is the Damped Local Trend (DLT) model, which supports linear, log-linear, and logistic deterministic global trends, along with a regression component.
For handling non-stationary data, artificial intelligence methods come into play [10]. Common AI models in this context include Support Vector Machine (SVM), Naive Bayes, Decision Tree, Multilayer Perceptron (MLP), Recurrent Neural Network (RNN), and Long Short-Term Memory (LSTM) [11]. Among these, LSTM is widely used for general-purpose time series forecasting due to its ability to address the vanishing gradient problem [12].
Despite the effectiveness of these models, manually building a data forecast for every client a company collaborates with can be impractical due to the workload on data analysts. Automation is essential to transition toward adaptive data forecasting models or automated model development.
To assist data analysts in building accurate data forecasts and prediction models, additional solutions are available. Algorithm selection and hyperparameter optimization fall under the umbrella of Automated Machine Learning (AutoML) [13]. Existing AutoML tools (such as auto-sklearn [14], TPOT [15], ML-Plan [16], Auto-WEKA [17], and others) test various models with different hyperparameters to find the best fit. As the performance and application requirements dictate, AutoML is gaining traction.
Research papers explore AutoML applications in diverse fields, including healthcare [18], video analytics [19], manufacturing [20], and telecommunications [21]. In finance and accounting, Agrapetidou et al. [22] demonstrated that AutoML can enhance financial data analysts’ productivity by automating bank failure forecasts. However, further research is needed to fully unlock AutoML’s potential and reduce the cost of building forecast models, especially in scenarios requiring numerous individual models.
For higher forecasting automation, zero-shot forecasting models have arisen during the last few years [23,24]. The ForecastPFN solution [23] is based on building a large time series forecasting model based on synthetic data. The trained model is then used for the forecast of the input data to achieve the prediction. Some solutions [24] adapt large language models for time series forecasting by encoding numerical values into numerical digits. Such solutions indicate promising results and a very high level of automation, while at the same time, their limitation is the need for the large model and dataset it requires.
In the area of time series data augmentation and synthetic data generation, multiple initiatives exist. In some cases, the time series are used for anomaly detection [25] and fault diagnosis [26]. These situations require additional data for the rare cases but keep a similar pattern of data. In this case, noise adding is more effective, rather than the generation of synthetic data. While different time series application areas exist, the data augmentation of time series always tries to take into account the existing data and extend it, rather than generate new patterns [27].
Generating a new time series is common as well, but is usually closely related to a specific area. For example, in medicine [28] and edge computing [29] synthetic time series data can be generated to reflect real signals. Such a situation requires the real signals to be mimicked and they are usually based on some theoretical model or the identification of the existing time series properties. Generating new time series data for unknown patterns might be complicated. Recently, some tools were introduced for time series data generation [30], but the main metric for a generated time series is usually human evaluation and quality estimation.
To combine the benefits of traditional forecasting methods and the requirement for limited resource usage, a two-shot approach could be utilized. In the first step, the most suitable forecasting model would be selected, while the second step would produce the output of the selected model.

3. Model for Company Financial Accounting Data Forecasting Method Selection

3.1. Company Financial Accounting Data

For the automated data forecasting model selection, accounting data from four companies were collected. All four companies were located in Lithuania; however, their activity and size varied, and foreign customers existed among their clients. The accounting data were selected from the year when the companies started using the accounting system and had a full year of accounting records available. The data from the accounting system was aggregated by month (to reflect the traditional accounting period), reflecting the culminative values of 23 features, indicating the count of invoices, products, the sum of purchase, selling costs, delays in payment, discounts, etc. For further experiments, only the sum of purchase costs for each month were used, forming a data file for each collaborating company and presenting records for each month. This field was selected as the main one, illustrating the possible income in the company. In addition, it was the most common non-zero-value-containing field. The data field was used as the time series data and the main idea was to forecast its values using univariate forecasting methods for the next year ahead.
The summarized numbers of the initial dataset used are presented in Table 1. They illustrate that company #2 is more oriented on short, non-recurrent orders as it has the biggest number of clients, while having the smallest relative number of non-zero records (only 8% of months in the history of collaboration have non-zero values). Meanwhile, about half of the records of all other companies have non-zero values, but the variation in the constant collaboration is very high (the standard deviation is more than 0.25).
The data are not very systematic, as none of the companies have many strongly connected clients which purchase something every month. From the dataset, company #2 had two clients which purchased something in all 72 months in the analyzed period. Companies #1 and #4 had only one company which purchased something every month, while company #3 had no such companies.

3.2. Experimental Environment

The research experiments were conducted using the KNIME 5.4.3 platform and JupyterLab environment for Python 3 programming, using various Python libraries and packages. Python was used to build the dataset for the forecasting method selection: the statsmodels package was used to analyze the time series and forecast with the SARIMA model, the Orbits package was used for the DLT forecasting model, and the Keras library from Tensorflow library was used to implement the LSTM model. The KNIME platform was used to model different situations and investigate the suitability of the classifiers for the selection of the prediction model.

3.3. Dataset for Forecasting Method Selection

The purpose of this research was to predict which data forecasting model would work best for specific time series data from company accounting data. The original dataset was not suitable for use as an input for the model because it had to be presented as one record with fixed features. Therefore, from the obtained initial dataset, a new dataset was built. To build the dataset, we took the 1055 data files representing each client of the companies. Records with at least 12 non-zero values in the clients’ time series were selected as part of the next dataset. This left only 424 files/records and was performed to eliminate mostly inactive collaboration cases, where mostly constant zero values were present.
The time series were summarized into 21 features (quantity of monthly records, min. value, max. value, mean value, standard deviation, lower quartile, median, upper quartile, annual correlation coefficient, augmented Dickey–Fuller test value, p-value, critical values for augmented Dickey–Fuller test statistic at 5% and 10%, augmented Dickey–Fuller test value after first order differencing, p-value after first order differencing, critical values for augmented Dickey–Fuller test statistic at 5% and 10% after first order differencing, augmented Dickey–Fuller test value after first order differencing and third-degree root transformation, p-value after first order differencing and third-degree root transformation, and critical values for augmented Dickey–Fuller test statistic at 5% and 10% after first order differencing and third-degree root transformation), reflecting the statistical features of the time series.
At the same time, the normalized time series data were used to find the best forecasting model. For each of the files, three time series forecasting methods were used using the AutoML option, where, in total, 659 combinations of their parameters were applied to achieve the least RMSE value for the next 12 months [31]. As a result, the LSTM model was the most suitable (with the lowest RMSE value) in 319 cases out of 424. SARIMA showed the smallest error in 105 cases. While DLT was never better than other methods, it sometimes demonstrated the same lowest accuracy as the SARIMA method (mostly in cases of constant zero values). In this way, we achieved a new dataset composed of 424 records. The process of building the dataset is presented in Figure 1.
The number of records obtained in the dataset was relatively small to try to achieve a very high accuracy. To estimate what the accuracy of the forecasting method prediction could be, the initial experiments were executed.

3.4. Forecasting Method Selection Performance

Business recommendation systems are characterized by a large number of clients. For each of them, a recommendation model needs to be developed, and therefore also a classification model. The classification model cannot be complex without overloading the servers/system processes. Therefore, in this paper, traditional machine learning algorithms were chosen: k-NN, Naïve Bayes, SVM, MLP, PNN, random forest, and gradient boosted trees. For the same reasons, the parameterization options of the algorithms were constrained. The analyzed parameters are listed in Table 2.
The research was carried out in two cases: with an imbalanced dataset (424 records) and with a dataset balanced towards the smaller SARIMA class (210 records). The oversampling option was not analyzed, as it would generate synthetic data and would not reflect a real situation where only real data are used.
The flowchart for the research on the imbalanced dataset can be seen in Figure 2. The dataset consisted of 424 records, of which 105 were SARIMA and 319 were LSTM. The data were min–max normalized and split by 20%/80% to train and test the classifiers. Five-fold cross validation was applied to maximize the accuracy of the results. The whole process was repeated five times.
In the case of undersampling, a subset of LSTM was randomly selected so that the size of the two classes was the same: 105 SARIMA and 105 LSTM. The methodology was the same as in the imbalanced case and the research was repeated five times, each time selecting a different subset of the LSTM class dataset (see Figure 3).
The parameters that were optimized for each model can be seen in Table 3. It indicates that in the case of the imbalanced dataset, all the classifiers learned to mostly predict the LSTM class (except for the Naïve Bayes, which learned to predict only the SARIMA class). In the case of the undersampled dataset, gradient boosted trees and MLP showed the best results.
Taking into account that there were only two classes, the classification accuracy could have been higher. In the case of imbalanced data, the average accuracy was higher than in the case of undersampling (reached 78% in comparison to the maximum of 65% accuracy). However, the high differences in recall between each class made the classifiers unreliable. Therefore, attempts have been made to improve it by adding synthetic data to the dataset.

4. Synthetic Data Generation for Dataset Extension in Forecasting Method Selection

4.1. Synthetic Data Generation Method

The real dataset is rather small and unbalanced, which makes it difficult to train classification models. Real client data are characterized by an abundance of zero: only 28 clients have a minimum data value that is not zero, and at least 25% of the data are zero for 333 out of 425 clients. Most of the real client data shows poor seasonality too. For these two reasons, it is decided to generate synthetic client data with non-zero values and add records with annual seasonality.
The augmentation of the real time series is not the best option in this case, as it would mimic the existing patterns but would not introduce a new one. This way, the augmented dataset usage would potentially lead to overfitting of the model and would not work for newly added time series patterns. Therefore, the approach to generate at least some common pattern data is selected.
Traditionally, adding synthetic data to the dataset might incorporate bias or distortions that are not common in real data. Therefore, the data augmentation and generation must be performed with high attention and validation. Existing research [32] indicates that the usage of time series data augmentation can significantly increase forecast accuracy; however, it is very dependent on using data augmentation methods, dataset size, and situation specifics.
In this research, the forecasting method selection uses preprocessed time series data, not the time series itself. Based on the estimated features of the time series, the classification model estimates under what properties of the time series one or another time series forecasting model would potentially be the most suitable. To avoid the generation of the classification model training data, we generate time series, which are used to model different situations and to estimate which method is the most accurate for its data forecasting. In this way, the classification model dataset is real, obtained by multiple modelling activities, and not generated. Only the data used to obtain the dataset is synthetic.
Usually, a time series signal can be expressed in three components: seasonality, trend, and noise. Taking into account the fact that the interaction between them is additive, the signal can be expressed as:
Y t = f S t , T t , N t = S t + T t + N t ,
where Y t is the time series signal at time t, S t —seasonality, T t —trend, and N t —noise.
Different seasonality and trend cases were generated for the synthetic dataset, while in all cases, the same red noise [33] function—noise serially correlated in time, in other words, noise with memory—was used to define the noise. The noise was used to reflect possible variations in the time series data, which could rise in real life (accounting delays, change in supply chain, economics, etc.).
The red noise parameters were chosen during all synthetic data generations as follows: variation was randomly chosen between 0.01 and 0.3 and multiplied by the trend base, and correlation was randomly chosen between 0.7 and 0.99. A relatively small noise variation was chosen because a larger variation would obscure the signal and make it unpredictable. A high correlation was chosen to better match real data. The exact ranges of the parameters were determined by manual experimental tests to reflect various but still similar to realistic conditions.
For the dataset extension with synthetic data, 900 different time series were generated. For the signal generation, the mokseries library was used [34]. To reflect different patterns of the time series, several strategies were used:
  • A total of 60 synthetic time series were generated without seasonality and with an exponential trend and red noise, including different a and b parameter values to reflect the exponential trend by the function
Y t = a t + b + N t ,
  • A total of 300 synthetic time series were generated without seasonality and with a linear trend and red noise, including different a and b parameter values to reflect the linear trend by the function
Y t = a × t + b + N t ,
  • A total of 300 synthetic time series were generated with sinusoidal yearly seasonality, a linear trend, and red noise, including different k parameter values to reflect the amplitude of seasonality and different parameter a and b values to reflect the linear trend by the function
Y t = k · sin t + a · t + b + N t ,
  • A total of 240 synthetic time series were generated with yearly periodic seasonality, a flat trend, and red noise, including different parameter b values to reflect the trend by the function
Y t = S [ 1 12 ] t + b + N t ,
The selected functions for synthetic data generation allowed different time series patterns to be reflected, while variations in each of the time series cases provided a better distribution of the extended dataset.

4.2. Description of Synthetic Time Series Generation Situations

Each of the four strategies for synthetic time series generation are detailed below, explaining how the parameter values were selected for the dataset extension.

4.2.1. Time Series Without Seasonality and with an Exponential Trend and Red Noise

To reflect the exponential trend, both increasing and decreasing trends were modelled. The list of parameters and their values is summarized in Table 4.
For an exponentially increasing trend, the trend base value was set to 180 because it was the average of the minimum values of the real client data. Analogously, in the case of an exponentially decreasing trend, 6263 was the average of the maximum values of the real client data. The average of the standard deviations of the real data was 1124 and was intended to diversify the data (a random number between −100 and 0 is used in the case of an exponentially increasing trend, so that the values are not less than 0 and do not go to infinity so quickly).
The trend factor interval (1.01, 1.1) was determined by manual testing and comparing it with real company data to see what trends were visible there. Values greater than 1.1 approached infinity too quickly and did not match the true data at all. Similarly, for exponential decreases, trend factor values less than 0.7 quickly approached 0, resulting in the generation of only zeros. Values between 0.95 and 1.01 had practically no effect on the trend.
Synthetic monthly data for 4, 6, and 9 years were generated due to the real client data characteristics. Despite the fact that the six listed strategies for time series generation did not reflect all possible cases, the inclusion of 10 different time series with different random parameters for each of them allowed some variations to be reflected while keeping the identified pattern of the time series.

4.2.2. Time Series Without Seasonality and with a Linear Trend and Red Noise

This time series generation strategy had three factors, which were changed to generate the time series (see Table 5). These were selected to reflect increasing and decreasing trends as in the previous strategy. In this strategy, 50 time series for each case were generated, taking into account that the noise effect to the linear trend might be more notable.
The trend base value was set to 962 because it was the average of the averages of the real client data. The range of the trend coefficient did not exceed 10 and −10, because the generated values would become too large and not representative of the real data. A random number between −200 and 200 was intended to expand the variety of data, while the lengths of the time series were selected to match the previously used strategy for better comparison.

4.2.3. Time Series with Sinusoidal Yearly Seasonality, a Linear Trend, and Red Noise

The generation of this synthetic time series was identical to the previous case (see Table 5), but for this strategy, a sinusoidal function was added with an amplitude which in all cases was equal to the trend base multiplied by a random number between 0.01 and 0.3. The sinusoid was adjusted to reflect the yearly seasonality. Therefore, the final function for the signal generation was the following:
Y t = k · sin π 6 t + a · t + b + N t ,
As a result, 300 time series were generated with the help of this strategy, where half of them had an increasing trend, while other were decreasing. In addition, the time series represented 4, 6, and 9 year periods for each case.

4.2.4. Time Series with Yearly Periodical Seasonality, a Flat Trend, and Red Noise

To avoid seasonality being purely sinusoidal, an additional strategy was used. In this strategy, the months of a year were divided into four intervals. During the interval, the value could decrease or increase by a random value from 0 to 374 (one-third of the average of the standard deviation of the real data). Considering the possible variations in the three months to increase or decrease, there were eight possible variations for the intervals:
  • [decrease; decrease; decrease];
  • [decrease; decrease; increase];
  • [decrease; increase; decrease];
  • [decrease; increase; increase];
  • [increase; decrease; decrease];
  • [increase; decrease; increase];
  • [increase; increase; decrease];
  • [increase; increase; increase].
These eight variations were used to build the first three intervals, while the last interval was intended to return to the initial value of the year.

4.3. Description of the Dataset for Forecasting Method Selection

With the help of synthetic data, the real data were supplemented by two times more synthetic data. The average RMSE, achieved by the best forecasting model, decreased by 1% in the normalized time series data (see Table 6). Using the t-test, the achieved improvement was significant (p < 0.0001). This could mean that the synthetic data, used for dataset augmentation did not reflect the real data. However, in this study, the purpose was not to mimic real data, but to extend it with more different patterns. The results indicated that the usage of clear patterns, even with added noise, allowed more accurate time series prediction.
Analyzing the distribution of cases where the LSTM or SARIMA models achieved lower RMSE values, the number of LSTM cases in the synthetic dataset increased proportionally to the real dataset (composed of about 70–75% of the dataset). Meanwhile, the number of cases where both models achieved the same results drastically decreased. Therefore, for further analysis, the SARIMA model was selected as being preferred in cases of the same results. This was in order to increase the number of SARIMA classes as well as to reflect the dataset class distribution.

4.4. Synthetic Dataset Usage Impact Estimation

The impact of synthetic data on data classification was analyzed in two cases: training classifiers using only synthetic data and training classifiers using real data augmented with synthetic data. The same classifiers as in Section 3.3 were investigated: k-NN, Naïve Bayes, Support Vector Machine, Multilayer Perceptron, probabilistic neural network, random forest, and gradient boosted trees.
In the first case, the possibility of training the classifiers on synthetic data only and then testing them on real data only was investigated. The dataset was imbalanced—the LSTM class had twice as many synthetic records and three times as many real records as the SARIMA class; therefore, the data were balanced according to the class with fewer records—SARIMA. Random undersampling was applied for balancing.
Tree-Based Parzen Estimator was used to optimize the parameters of the MLP, random forest, and gradient boosted tree classifiers, while the hill climbing technique was applied to the remaining classifiers. To make the balancing less influential on the result, the whole process was repeated five times, and the average of the results was calculated. The flowchart of the research can be seen in Figure 4.
The results showed similar results as in the case of the not-augmented dataset, and even worse results, as the average accuracy was similar but the variation between class accuracy varied a lot. The random forest classifier achieved the best result—60.5%. The worst results were shown by Naïve Bayes, which failed to distinguish the class data at all and always guessed the LSTM class. It can be noted that all classifiers have learned to prioritize the LSTM class records. The results can be seen in Table 7.
In the second case, to see how the accuracy of the classifiers changed, the classifiers were trained and tested on a real dataset augmented with synthetic data, i.e., the real dataset was merged with a synthetic dataset and the resulting dataset was submitted with a ratio of 80–20% for the training and testing of classifiers. The impact of balancing the dataset was simultaneously examined. The dataset was either unbalanced or balanced towards the SARIMA class with the fewest records (random undersampling) or balanced towards the LSTM class with the most records (oversampling using the Synthetic Minority Oversampling Technique (SMOTE)). For each data balancing case, cross validation with five subsets and stratified distribution was applied to validate the results. Finally, each experiment was repeated five times, and the average accuracy was evaluated.
All three cases of the real and synthetic data balancing strategies are illustrated in Figure 5. The balancing of the oversampled dataset was performed after the splitting of the dataset without affecting the testing of the data and thus distorting the results. In the case of an undersampled dataset, the balancing of the dataset was performed before the cross validation with the same options.
Analyzing the results between the sampling techniques, the imbalanced and oversampled options provided the highest average accuracy; however, they also showed the biggest variations between each class’s accuracy (see Table 8). Meanwhile, the undersampled option led to well-balanced accuracy between both classes in most cases, while the accuracy of the models increased by almost 10% in comparison to the real dataset only.
After multiple testing cases, the MLP classifier with the undersampled dataset and the parameters set to 95 iterations, one layer, 67 neurons, the “relu” activation function, and the “adam” solver should be used to achieve the highest and more balanced prediction of both classes. In Figure 6, a case where the average accuracy is 70% (71% for LSTM and 69% for SARIMA) is illustrated (see Figure 6a). The analysis of the real (b) and synthetic (c) data among this case illustrates that the classification accuracy is a little bit higher for the real data (72%) rather than the synthetic (69%).

5. Discussion

This research concentrated on the prediction of the best time series forecasting data method. As a test case, multiple companies’ accounting data were selected and aggregated by month to predict the next 12-month potential purchase costs for each partner. This allowed us to form a dataset where 21 statistical time series measures were used as features for the best forecasting model definition.
The initial experiments (presented in chapter three) illustrated the need for a choice between more accurate average accuracy and a balance between each class prediction accuracy. This was affected by the imbalanced dataset and the limited number of records in the dataset. For example, the highest average accuracy was achieved using an imbalanced dataset and a random forest method—78.5%. In this case, the recall for the LSTM method was 91% and for SARIMA was 49%. This did not seem fair and should not have been accepted as a trusted model. Meanwhile, for the undersampled dataset, the highest accuracy was achieved by random forest and reached only 74% on average, but both classes had a very similar recall.
The Naïve Bayes method was probably the worst model in terms of the balance between classes. In all cases, it had the lowest average accuracy as well as always preferring one class (LSTM) against the second one (SARIMA). This could be explained by the nature of the probabilistic model and the variety of the features and results, where the Naïve Bayes method did not wave enough cases to match all possible cases for both classes.
This could be explained by the fact that random forest is better for matching to complex feature interactions, the prioritization of the features, weighted voting, and adjusting class weights to balance the classes, dynamically ignoring irrelevant features. Naïve Bayes does not have these features and treats all features as equally relevant and also does not work in this case, as all the features in this dataset are continuous and do not necessarily match Gaussian distribution.
A clear difference between the used classification models (except Naïve Bayes) was not very visible—some models worked better in one situation, some in another. This did not allow us to estimate what had the biggest effect on the average classification accuracy, as multiple factors might have affected it: a non-linear, complex dependency between features; a class imbalance or lack of feature value diversification for certain classes; and a limited number of records in the dataset. Random forest usually shows the best accuracy, because it is able to balance all of the properties, despite not being the fastest one.
While undersampling helps to balance the class accuracy, an attempt to increase the accuracy by generating synthetic data was performed (presented in Section 4). Different time series generation strategies were applied to generate more data for forecasting using different methods. By augmenting the dataset with the synthetically generated time series data, the dataset increased from 424 to 1324 records, adding 900 based on synthetic data. The purpose of the synthetic data was to add some time series with different patterns, as the real accounting data had a high level of variation and a lack of seasonality.
The ideal situation would be if we could train the model with synthetic data and accurately predict the best forecasting method for real data. This approach was tested (Chapter 4.4) and indicated that it did not provide good results—the average accuracy was close to 55%, but it was not balanced between each class, even if the dataset was undersampled. Such results might have been affected by the fact that real data did not match the patterns reflected in the synthetic dataset. However, to specify the specific patterns and reflect them, synthetic data were not an option, as they would require additional resources, while the purpose of the automated forecasting method selection was mainly to reduce the forecasting process resource usage.
Mixing real and synthetic data for training and testing was an option, as it would increase the dataset size and reflect patterns from both the real and synthetic time series. This approach for dataset balancing using oversampling with SMOTE was tested, with undersampling to the smaller class and usage of the imbalanced cases also tested. The same tendency was noticed as with real data only—undersampling in this case worked better than other strategies in the case of average accuracy and a balance between the accuracy of each class. The increase in dataset size through synthetic data generation contributed to an improvement in forecasting method selection. In comparison to real undersampled datasets, the average accuracy for real + synthetic dataset usage increased by approximately 8%. This was a statistically significant difference (using t-test for the seven cases, p = 0.0444).
All the results indicated that the synthetic time series generation for forecasting method selection had potential for increasing the accuracy. In this study, only several different patterns of synthetic data were generated and were allowed to generate twice as much data as we had in the real dataset. Generating a wider range of patterns and a higher volume of variations for each of them could lead to even higher accuracy. The increase in classification accuracy might have been affected by the increase in the training set as well as the introduction of more patterns in the time series. The fact that oversampled data did not perform much better than imbalanced data, but were a bit better than the undersampled dataset models, gave us reason to believe the variety of different patterns introduced was more important than the increase in the training dataset. The tested classification model’s accuracy varied for the analyzed cases, as they were different in their nature, bias variance, complexity, hyperparameter sensitivity, etc. As the accounting data were diverse, the classification model selection was also very dependent on the data. Therefore, modelling different datasets and model variations gave us new knowledge towards the search for the most suitable AutoML solution for financial accounting time series preprocessing and prediction.
This research was performed on company accounting data, aggregating the data for each month. The data normalization and the statistical measurement used as features for model prediction reduced the chance that the same approach would not work for another data aggregation period; however, it was not tested in this research and should be considered. Despite this, the research results indicated their potential to be applied in cases where a high volume of time series exist and it is not realistic to work on them manually, while the forecasting results could lead to better performance or planning. The company management case is the most common example—if a company wants to predict the different metrics (orders, purchases, debts, complains, costs, etc.) of its collaboration with all of its clients (for which the numbers can reach hundreds or even thousands of clients), automation is necessary. Proper forecasting of different metrics could be used as a tool for decision support or even undesired case prevention by applying some automated actions in advance, such as automatically estimating the potential best time series data forecasting method whose performance is 74% better than using a random one or modelling all possible cases (which could lead to high resource consumption). Based on the prediction, the time series forecasting count can be performed automatically too and leads to further steps oriented on notifications and even certain decision-making actions.

6. Conclusions and Future Work

Training time series data forecasting method selection models only on synthetic data is not a suitable option for selecting the best model for forecasting time series data. The experiments illustrate that fully synthetic data usage for training model for testing with real data cannot increase the model accuracy, and even decreases accuracy by almost 20%, even if the number of records increases multiple times. The need for real data is a must in the scope of the analyzed case.
The undersampling strategy for dataset balancing is the best option for forecasting the method prediction accuracy. Undersampling showed the best results for all our experiments with accounting data for time series model prediction: real data were used for model training and testing; synthetic data for training and real data for testing; and a mix of real and synthetic data for both training and testing. This indicates the importance of class balance and the limitations of data augmentation methods for training datasets.
The combination of real and synthetic data can improve the accuracy of the classifier. The experiment results indicate that the change is statistically significant (p < 0.05), which proves the potential of synthetic data generation for the best time series model prediction.
Data augmentation for the best time series forecasting model selection should take into account both synthetic data generation and undersampling to balance the data and guarantee that class prioritization will be avoided. Experiments with data sampling strategies show that the highest average accuracy can be achieved with an imbalanced dataset, but the accuracy of the classes becomes unbalanced as well. While undersampling reduces the average accuracy, it also guarantees both classes will be equally reflected in the model.
Synthetic time series generation to extend a dataset for best time series data forecasting model selection is an effective way to increase the model accuracy. The experiment results illustrate that it is possible to achieve up to 74% of the accuracy and maintain the balance between classes in comparison to a not-augmented dataset, where the maximum accuracy reached 65% only. In some cases, this might be enough, while other areas would require a higher accuracy. Considering the limited number of data augmentation patterns, the generation of synthetic data seems very promising and leads to a hope that the accuracy could be increased even more by extending the dataset with new data records.
For higher applicability, further research could concentrate not only on the forecast method prediction, but the best parameter estimation as well. Different strategies should be tested to estimate how effective the model and its parameter prediction is in comparison to the AutoML approach, including the forecasting error and resource utilization.

Author Contributions

Conceptualization, R.Š. and S.R.; methodology, S.R.; software, R.Š.; validation, S.R. and P.S.; formal analysis, S.R.; investigation, R.Š.; resources, R.Š.; data curation, R.Š.; writing—original draft preparation, R.Š.; writing—review and editing, S.R. and P.S.; visualization, R.Š.; supervision, S.R. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The used datasets can be downloaded at Kaggle [35].

Acknowledgments

The authors would like to thank LTD STEKAS for collaborating in gathering the company data and providing an opportunity to adopt the research idea in the project “Extension of Accounting and Business Management Systems using artificial intelligence methods” (No. 01.2.1-LVPA-K-856-01-0083).

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AutoMLAutomated Machine Learning
AIArtificial intelligence
MLMachine learning
VARVector Auto-Regression
ARIMAAuto-Regressive Integrated Moving Average model
SARIMASeasonal ARIMA
ARIMAXARIMA with eXogenous inputs
SARIMAXSeasonal ARIMA with eXogenous inputs
DLTDamped Local Trend
SVMSupport Vector Machine
k-NNk-Nearest Neighbours
MLPMultilayer Perceptron
PNNProbabilistic neural network
RNNRecurrent Neural Network
LSTMLong Short-Term Memory
SMOTE Synthetic Minority Oversampling Technique

References

  1. Ajayi-Nifise, A.O.; Odeyemi, O.; Mhlongo, N.Z.; Ibeh, C.V.; Elufioye, O.A.; Awonuga, K.F. The future of accounting: Predictions on automation and AI integration. World J. Adv. Res. Rev. 2024, 21, 399–407. [Google Scholar] [CrossRef]
  2. Cao, S.S.; Jiang, W.; Lei, L.G. Applied AI for finance and accounting: Alternative data and opportunities. Pac.-Basin Financ. J. 2024, 84, 102307. [Google Scholar] [CrossRef]
  3. Salehin, I.; Islam, M.S.; Saha, P.; Noman, S.M.; Tuni, A.; Hasan, M.M.; Baten, M.A. AutoML: A systematic review on automated machine learning with neural architecture search. J. Inf. Intell. 2024, 2, 52–81. [Google Scholar] [CrossRef]
  4. Hendikawati, P.; Subanar; Abdurakhman; Tarno. A Survey of Time Series Forecasting from Stochastic Method to Soft Computing. J. Phys. Conf. Ser. 2020, 1613, 012019. [Google Scholar] [CrossRef]
  5. De Gooijer, J.; Hyndman, R. 25 Years of Time Series Forecasting. Int. J. Forecast. 2006, 22, 443–473. [Google Scholar] [CrossRef]
  6. Petropoulos, F.; Apiletti, D.; Assimakopoulos, V.; Babai, M.Z.; Barrow, D.K.; Taieb, S.B.; Bergmeir, C.; Bessa, R.J.; Bijak, J.; Boylan, J.E.; et al. Forecasting: Theory and Practice. Int. J. Forecast. 2022, 38, 705–871. [Google Scholar] [CrossRef]
  7. Adu, W.K.; Appiahene, P.; Afrifa, S. VAR, ARIMAX and ARIMA Models for Nowcasting Unemployment Rate in Ghana Using Google Trends. J. Electr. Syst. Inf. Technol. 2023, 10, 12. [Google Scholar] [CrossRef]
  8. Xie, M.; Sandels, C.; Zhu, K.; Nordström, L. A Seasonal ARIMA Model with Exogenous Variables for Elspot Electricity Prices in Sweden. In Proceedings of the 10th International Conference on the European Energy Market (EEM), Stockholm, Sweden, 27–31 May 2013; pp. 1–4. [Google Scholar] [CrossRef]
  9. Dharmawan, P.A.S.; Indradewi, I.G.A.A.D. Double Exponential Smoothing Brown Method Towards Sales Forecasting System with a Linear and Non-stationary Data Trend. J. Phys. Conf. Ser. 2021, 1810, 012026. [Google Scholar] [CrossRef]
  10. Torre, K.; Delignières, D.; Lemoine, L. Detection of Long-range Dependence and Estimation of Fractal Exponents Through ARFIMA Modelling. Br. J. Math. Stat. Psychol. 2007, 60, 85–106. [Google Scholar] [CrossRef] [PubMed]
  11. Li, W.; Liao, J. A Comparative Study on Trend Forecasting Approach for Stock Price Time Series. In Proceedings of the 11th IEEE International Conference on Anti-counterfeiting, Security, and Identification (ASID), Xiamen, China, 27–29 October 2017; pp. 74–78. [Google Scholar] [CrossRef]
  12. Elsworth, S.; Guttel, S. Time Series Forecasting Using LSTM Networks: A Symbolic Approach. arXiv 2020, arXiv:2003.05672. [Google Scholar]
  13. Gijsbers, P.; Vanschoren, J. GAMA: Genetic Automated Machine learning Assistant. J. Open Source Softw. 2019, 4, 1132. [Google Scholar] [CrossRef]
  14. Feurer, M.; Klein, A.; Eggensperger, K.; Springenberg, J.T.; Blum, M.; Hutter, F. Auto-sklearn: Efficient and Robust Automated Machine Learning. In Automated Machine Learning, 1st ed.; Hutter, F., Kotthoff, L., Vanschoren, J., Eds.; The Springer Series on Challenges in Machine Learning (SSCML); Springer: Cham, Switzerland, 2019; pp. 113–134. [Google Scholar] [CrossRef]
  15. Olson, R.S.; Urbanowicz, R.J.; Andrews, P.C.; Lavender, N.A.; Kidd, L.C.; Moore, J.H. Automating Biomedical Data Science Through Tree-Based Pipeline Optimization. In Proceedings of the 19th European Conference on the Applications of Evolutionary Computation (EvoApplications), Porto, Portugal, 1 April 2016; Lecture Notes in Computer Science (LNCS). Squillero, G., Burelli, P., Eds.; Springer: Cham, Switzerland, 2016; Volume 9597, pp. 123–137. [Google Scholar] [CrossRef]
  16. Mohr, F.; Wever, M.; Hüllermeier, E. ML-Plan: Automated Machine Learning via Hierarchical Planning. Mach. Learn. 2018, 107, 1495–1515. [Google Scholar] [CrossRef]
  17. Thornton, C.; Hutter, F.; Hoos, H.H.; Leyton-Brown, K. Auto-WEKA: Combined Selection and Hyperparameter Optimization of Classification Algorithms. In Proceedings of the 19th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD), Chicago, IL, USA, 11–14 August 2013; pp. 847–855. [Google Scholar] [CrossRef]
  18. Anwar, T. COVID-19 Diagnosis using AutoML from 3D CT scans. In Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops (ICCVW), Montreal, BC, Canada, 11–17 October 2021; pp. 503–507. [Google Scholar] [CrossRef]
  19. Galanopoulos, A.; Ayala-Romero, J.A.; Leith, D.J.; Iosifidis, G. AutoML for Video Analytics with Edge Computing. In Proceedings of the IEEE Conference on Computer Communications (INFOCOM), Vancouver, BC, Canada, 10–13 May 2021; pp. 1–10. [Google Scholar] [CrossRef]
  20. Koppe, T.; Schatz, J. Cloud-based ML Technologies for Visual Inspection: A Case Study in Manufacturing. In Proceedings of the 54th Hawaii International Conference on System Sciences (HICSS), Kauai, HI, USA, 5 January 2021; pp. 1020–1030. [Google Scholar] [CrossRef]
  21. He, Y.; Lin, J.; Liu, Z.; Wang, H.; Li, L.J.; Han, S. AMC: AutoML for Model Compression and Acceleration on Mobile Devices. In Proceedings of the 15th European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; Lecture Notes in Computer Science (LNCS). Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y., Eds.; Springer: Cham, Switzerland, 2018; Volume 11211, pp. 815–851. [Google Scholar] [CrossRef]
  22. Agrapetidou, A.; Charonyktakis, P.; Gogas, P.; Papadimitriou, T.; Tsamardinos, I. An AutoML Application to Forecasting Bank Failures. Appl. Econ. Lett. 2020, 28, 5–9. [Google Scholar] [CrossRef]
  23. Dooley, S.; Khurana, G.S.; Mohapatra, C.; Naidu, S.V.; White, C. ForecastPFN: Synthetically-Trained Zero-Shot Forecasting. In Proceedings of the Thirty-Seventh Conference on Neural Information Processing Systems (NeurIPS), New Orleans, LA, USA, 10–16 December 2023; pp. 2403–2426. [Google Scholar]
  24. Gruver, N.; Finzi, M.; Qiu, S.; Wilson, A.G. Large Language Models Are Zero-Shot Time Series Forecasters. In Proceedings of the Thirty-Seventh Conference on Neural Information Processing Systems (NeurIPS), New Orleans, LA, USA, 10–16 December 2023; pp. 19622–19635. [Google Scholar]
  25. Kim, W.H.; Kim, G.W.; Ahn, J.; Chung, K. Time-Series Data Augmentation for Improving Multi-Class Classification Performance. Int. J. Adv. Sci. Eng. Inf. Technol. 2024, 14, 887–893. [Google Scholar] [CrossRef]
  26. Kulevome, D.K.B.; Wang, H.; Cobbinah, B.M.; Mawuli, E.S.; Kumar, R. Effective time-series data augmentation with analytic wavelets for bearing fault diagnosis. Expert Syst. Appl. 2024, 249, 123536. [Google Scholar] [CrossRef]
  27. Annaki, I.; Rahmoune, M.; Bourhaleb, M. Overview of Data Augmentation Techniques in Time Series Analysis. Int. J. Adv. Comput. Sci. Appl. 2024, 15, 1201–1211. [Google Scholar] [CrossRef]
  28. Vetter, J.; Macke, J.H.; Gao, R. Generating realistic neurophysiological time series with denoising diffusion probabilistic models. Patterns 2024, 5, 101047. [Google Scholar] [CrossRef] [PubMed]
  29. Yousuf, M.F.; Mahmud, M.S. Generating synthetic time-series data on edge devices using generative adversarial networks. In Proceedings of the International Conference on Computing, Networking and Communications (ICNC), Big Island, HI, USA, 19–22 February 2024; pp. 441–445. [Google Scholar] [CrossRef]
  30. Nikitin, A.; Iannucci, L.; Kaski, S. TSGM: A flexible framework for generative modeling of synthetic time series. Adv. Neural Inf. Process. Syst. 2025, 37, 129042–129061. [Google Scholar]
  31. Štrimaitis, R.; Ramanauskaitė, S.; Stefanovič, P. Analysis of Automated Forecasting Model Possibilities for Company Financial Accounting Data. In Proceedings of the 3rd International Conference on Electrical, Computer, Communications and Mechatronics Engineering (ICECCME), Tenerife, Canary Islands, Spain, 19 July 2023; pp. 1–6. [Google Scholar] [CrossRef]
  32. Semenoglou, A.A.; Spiliotis, E.; Assimakopoulos, V. Data augmentation for univariate time series forecasting with neural networks. Pattern Recognit. 2023, 134, 109132. [Google Scholar] [CrossRef]
  33. Bunde, A. The Different Types of Noise and How They Effect Data Analysis. Chem. Ing. Tech. 2023, 95, 1758–1767. [Google Scholar] [CrossRef]
  34. Mockseries. Available online: https://mockseries.catheu.tech/ (accessed on 24 September 2024).
  35. Kaggle. Available online: https://www.kaggle.com/datasets/rokastrimaitis/true-and-synth-metadata-classification-dataset (accessed on 19 January 2025).
Figure 1. Principal schema of forecast method prediction dataset construction.
Figure 1. Principal schema of forecast method prediction dataset construction.
Electronics 14 01253 g001
Figure 2. The model with the imbalanced dataset.
Figure 2. The model with the imbalanced dataset.
Electronics 14 01253 g002
Figure 3. The model with undersampling balancing.
Figure 3. The model with undersampling balancing.
Electronics 14 01253 g003
Figure 4. The model for training on synthetic data and testing on real data.
Figure 4. The model for training on synthetic data and testing on real data.
Electronics 14 01253 g004
Figure 5. The model with different balancing strategies for the augmented dataset.
Figure 5. The model with different balancing strategies for the augmented dataset.
Electronics 14 01253 g005
Figure 6. Confusion matrices of MLP application for undersampled dataset, illustrating one-fold data (a) and its division into real (b) and synthetic (c) data only.
Figure 6. Confusion matrices of MLP application for undersampled dataset, illustrating one-fold data (a) and its division into real (b) and synthetic (c) data only.
Electronics 14 01253 g006
Table 1. Summary of company data used as initial dataset.
Table 1. Summary of company data used as initial dataset.
Company IDNo. of ClientsRecord FromRecord UntilNo. of Non-Zero RecordsPercentage of Non-Zero Months
#11962012 January2020 December868241% (stdev = 0.25)
#26982015 January2020 December38848% (stdev= 0.15)
#31402017 January2020 December393058% (stdev = 0.29)
#4212015 January2020 December83455% (stdev = 0.33)
Table 2. Parameters for the forecasting method prediction.
Table 2. Parameters for the forecasting method prediction.
Classification MethodParameterParameter Values
StartEndStep
k-NNk3101
SVMc0.011000.1
Naïve BayesMax. number of norm. values201001
MLPIterations2020000
Layers151
Neurons81002
PNNTheta plus0.410.1
Theda minus0.10.40.1
random forestTrees1020010
Tree depth5501
gradient boosted treesTrees1020010
Tree depth5501
Table 3. Evaluation of forecasting method classifiers with original dataset.
Table 3. Evaluation of forecasting method classifiers with original dataset.
Classifiers Average Acc.Best Case Metrics
ClassesRecallPrecisionF1Parameters
Imbalancedk-NN0.755LSTM0.8300.7720.800k = 3
SARIMA0.2570.3330.290
SVM0.752LSTM1.0000.7520.858c = any value
SARIMA0.0000.0000.000
Naïve Bayes0.366LSTM0.2060.8080.32956 max number norm. val.
SARIMA0.8470.2600.398
MLP0.776LSTM0.9150.8000.853138 iter., 3 layers,
60 neurons
SARIMA0.3040.5420.390
PNN0.774LSTM0.9560.7760.856theta plus 0.573, theta minus 0.105
SARIMA0.1610.5480.250
random forest0.778LSTM0.9680.7860.86754 trees,
17 tree depth
SARIMA0.2000.6770.308
gradient boosted trees0.767LSTM0.7740.8170.79591 trees,
50 tree depth
SARIMA0.4760.4090.440
Undersampledk-NN0.585LSTM0.6950.6030.646k = 4
SARIMA0.5420.6400.587
SVM0.584LSTM0.3140.7330.440c = 0.456
SARIMA0.8850.5630.688
Naïve Bayes0.537LSTM0.6760.5540.60972 max number norm. val.
SARIMA0.4570.5850.513
MLP0.630LSTM0.6470.6470.64747 iter., 3 layer,
88 neurons
SARIMA0.6470.6470.647
PNN0.611LSTM0.6190.6560.637theta plus 0.847, theta minus 0.392
SARIMA0.6760.6390.657
random forest0.615LSTM0.6470.6530.650101 trees,
30 tree depth
SARIMA0.6570.6500.654
gradient boosted trees0.650LSTM0.7040.6720.688199 trees,
6 tree depth
SARIMA0.6570.6900.673
Table 4. List of parameters and their values for time series generation with no seasonality and with an exponential trend and red noise.
Table 4. List of parameters and their values for time series generation with no seasonality and with an exponential trend and red noise.
Quantity of Synthetic Client DataTrend Factor (a)Trend Base (b)The Length of One Client Dataset
10random (1.01, 1.1)180 + random (−100, 0)108 (9 years × 12 months)
1072 (6 years × 12 months)
1048 (4 years × 12 months)
10random (0.7, 0.95)6263 + random (−1124, 1124)108 (9 years × 12 months)
1072 (6 years × 12 months)
1048 (4 years × 12 months)
Table 5. List of parameters and their values for time series generation with no seasonality and with a linear trend and red noise.
Table 5. List of parameters and their values for time series generation with no seasonality and with a linear trend and red noise.
Quantity of Synthetic Client DataTrend Coefficient (a)Trend Base (b)The Length of One Client Dataset
50random (0.1, 10)962 + random (−200, 200)108 (9 years × 12 months)
5072 (6 years × 12 months)
5048 (4 years × 12 months)
50random (−10, −0.1)108 (9 years × 12 months)
5072 (6 years × 12 months)
5048 (4 years × 12 months)
Table 6. Comparison of real and synthetic dataset properties.
Table 6. Comparison of real and synthetic dataset properties.
Dataset PropertiesReal DataSynthetic Data
Number of records424900
RMSE for best modelAverage8.67.6
Standard deviation5.03.1
Best modelLSTM319607
SARIMA29288
Equal765
Final class distributionClass LSTM319607
Class SARIMA105293
Table 7. Classifier evaluation (dataset balanced to the smallest class).
Table 7. Classifier evaluation (dataset balanced to the smallest class).
Classifiers Average Acc.Best Case Metrics
ClassesRecallPrecisionF1Parameters
k-NN0.570LSTM0.7710.5580.648k = 7
SARIMA0.3900.6300.482
SVM0.561LSTM0.8470.5560.671c = 87.027
SARIMA0.3230.6800.438
Naïve Bayes0.500LSTM1.0000.5000.666any value of the max no. of norm values
SARIMA0.0000.0000.000
MLP0.592LSTM0.8000.5790.67222 iter., 4 layers,
17 neurons
SARIMA0.4190.6760.517
PNN0.556LSTM0.8760.5500.676theta plus 0.920, theta minus 0.372
SARIMA0.2850.6970.405
random forest0.605LSTM0.8000.6080.691199 trees,
17 tree depth
SARIMA0.4850.7080.576
gradient boosted trees0.575LSTM0.6850.5710.623173 trees,
5 tree depth
SARIMA0.4850.6070.539
Table 8. Accuracy and best model parameters for forecasting method prediction for different sampling techniques in mixed real and synthetically augmented datasets.
Table 8. Accuracy and best model parameters for forecasting method prediction for different sampling techniques in mixed real and synthetically augmented datasets.
Classifiers Average Acc.Best Case Metrics
ClassesRecallPrecisionF1Parameters
Oversampledk-NN0.705LSTM0.7140.8400.772k = 9, knn smote = 44
SARIMA0.6830.5070.582
SVM0.745LSTM0.9840.7220.834c = 0.049, knn smote 50
SARIMA0.1880.7960.205
Naïve Bayes0.708LSTM0.9980.7050.82785 max number norm. val., knn smote 36
SARIMA0.0300.9230.058
MLP0.744LSTM0.7680.8520.808143 iter., 5 layers,
40 neurons, knn smote 5
SARIMA0.6900.5620.620
PNN0.772LSTM0.8900.8020.844theta plus 0.817, theta minus 0.362, knn smote 24
SARIMA0.4890.6580.561
random forest0.758LSTM0.8360.8360.83528 trees, 6 tree depth,
knn smote 27
SARIMA0.6170.6200.619
gradient boosted trees0.761LSTM0.8520.8270.83982 trees, 5 tree depth,
knn smote 10
SARIMA0.5850.6290.606
Imbalancedk-NN0.762LSTM0.8760.8000.836k = 5
SARIMA0.4890.6310.551
SVM0.765LSTM0.8870.7980.840c = 19.982
SARIMA0.4790.6470.551
Naïve Bayes0.706LSTM0.9980.7040.82662 max number norm. val.
SARIMA0.0250.9090.048
MLP0.783LSTM0.8930.8120.850171 iter., 4 layers,
48 neurons
SARIMA0.5200.6760.588
PNN0.777LSTM0.9480.7790.855theta plus 0.661,
theta minus 0.107
SARIMA0.3740.7560.500
random forest0.785LSTM0.9140.8070.857139 trees,
27 tree depth
SARIMA0.4920.7120.582
gradient boosted trees0.769LSTM0.8880.8020.84354 trees,
6 tree depth
SARIMA0.4890.6540.560
Undersampledk-NN0.683LSTM0.7010.7040.702k = 9
SARIMA0.7060.7020.704
SVM0.642LSTM0.6850.6530.669c = 57.519
SARIMA0.6350.6690.652
Naïve Bayes0.513LSTM0.9970.5110.67665 max number norm. val.
SARIMA0.0470.9500.090
MLP0.731LSTM0.7410.7350.73895 iter., 1 layer,
67 neurons
SARIMA0.7330.7390.736
PNN0.704LSTM0.7730.7040.737theta plus 0.427,
theta minus 0.138
SARIMA0.6750.7490.710
random forest0.740LSTM0.7760.7330.754187 trees,
22 tree depth
SARIMA0.7180.7620.739
gradient boosted trees0.717LSTM0.7110.7360.72321 trees,
6 tree depth
SARIMA0.7460.7200.733
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Štrimaitis, R.; Ramanauskaitė, S.; Stefanovič, P. Automated Selection of Time Series Forecasting Models for Financial Accounting Data: Synthetic Data Application. Electronics 2025, 14, 1253. https://doi.org/10.3390/electronics14071253

AMA Style

Štrimaitis R, Ramanauskaitė S, Stefanovič P. Automated Selection of Time Series Forecasting Models for Financial Accounting Data: Synthetic Data Application. Electronics. 2025; 14(7):1253. https://doi.org/10.3390/electronics14071253

Chicago/Turabian Style

Štrimaitis, Rokas, Simona Ramanauskaitė, and Pavel Stefanovič. 2025. "Automated Selection of Time Series Forecasting Models for Financial Accounting Data: Synthetic Data Application" Electronics 14, no. 7: 1253. https://doi.org/10.3390/electronics14071253

APA Style

Štrimaitis, R., Ramanauskaitė, S., & Stefanovič, P. (2025). Automated Selection of Time Series Forecasting Models for Financial Accounting Data: Synthetic Data Application. Electronics, 14(7), 1253. https://doi.org/10.3390/electronics14071253

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop