Towards Time-Series Feature Engineering in Automated Machine Learning for Multi-Step-Ahead Forecasting †

: Feature engineering is an essential step in the pipelines used for many machine learning tasks, including time-series forecasting. Although existing AutoML approaches partly automate feature engineering, they do not support specialised approaches for applications on time-series data such as multi-step forecasting. Multi-step forecasting is the task of predicting a sequence of values in a time-series. Two kinds of approaches are commonly used for multi-step forecasting. A typical approach is to apply one model to predict the value for the next time step. Then the model uses this predicted value as an input to forecast the value for the next time step. Another approach is to use multi-output models to make the predictions for multiple time steps of each time-series directly. In this work, we demonstrate how automated machine learning can be enhanced with feature engineering techniques for multi-step time-series forecasting. Speciﬁcally, we combine a state-of-the-art automated machine learning system, auto-sklearn, with tsfresh, a library for feature extraction from time-series. In addition to optimising machine learning pipelines, we propose to optimise the size of the window over which time-series data are used for predicting future time-steps. This is an essential hyperparameter in time-series forecasting. We propose and compare (i) auto-sklearn with automated window size selection, (ii) auto-sklearn with tsfresh features, and (iii) auto-sklearn with automated window size selection and tsfresh features. We evaluate these approaches with statistical techniques, machine learning techniques and state-of-the-art automated machine learning techniques, on a diverse set of benchmarks for multi-step time-series forecasting, covering 20 synthetic and real-world problems. Our empirical results indicate a signiﬁcant potential for improving the accuracy of multi-step time-series forecasting by using automated machine learning in combination with automatically optimised feature extraction techniques.


Introduction
Time-series (TS) data, such as electrocardiograms, music, exchange rates, and energy consumption, is everywhere in our daily life and business world.Multi-step-ahead forecasting is an important task in time-series modeling with many industrial applications, such as crude oil price forecasting [1] and flood forecasting [2].Both ML models [3,4] and statistical models [5,6] are used for this purpose.Creating an ML pipeline for such TS data analysis is, however, difficult for many domain experts with limited machine learning expertise, due to the complexity of the data sets and the ML models.
To reduce the complexity of creating machine learning pipelines, automated machine learning (AutoML) research has recently focused on developing algorithms that can automatically design ML pipelines optimised for a given data set without human input [7,8].The components in a pipeline may include a data cleaning component, feature engineering component, ML model component, and ensemble construction component.While there has been work proposed in the past that used AutoML for TS analysis [9], the current AutoML systems do not support specialised techniques used for designing machine learning pipelines for TS data, such as TS feature engineering.As demonstrated by Christ et al. [10], including such specialized techniques for extracting TS features or feature importance filtering could significantly improve the accuracy of ML models.Therefore, our goal in this paper is to study if extending AutoML systems by including such techniques can improve the quality of automatically generated machine learning pipelines for multi-step forecasting.
This paper presents a study of AutoML for time-series multi-step TS forecasting implemented through (i) multi-output modeling, and (ii) recursive modeling.We combine the state-of-the-art AutoML system of auto-sklearn with tsfresh, a well-known library for feature extraction from TS.We implement three AutoML variants and state-of-the-art baseline models, including auto-sklearn, SVM, GBM, N-BEATS, and Auto-Keras.More specifically, our contributions are as follows: • We adapt the auto-sklearn AutoML system to the task of forecasting and introduce three AutoML forecasting variants for multi-step-ahead TS forecasting.

•
We demonstrate the importance of feature selection and window size selection in forecasting problems and further show that by incorporating such approaches, our TS AutoML techniques outperform available AutoML methods.

•
We evaluate our methods on 20 benchmarking data sets from 20 different categories and against the baselines.We found that our proposed AutoML method outperformed the traditional ML baseline on 14 out of 20 data sets and N-BEATS on 15 out of 20 data sets.
The remainder of this paper is structured as follows: Section 2 covers related work on TS forecasting, TS feature engineering and AutoML.Section 3 introduces the problem statement of AutoML for multi-step TS forecasting.In Section 4, the methodology of our newly proposed models is explained.Section 5 presents the results of the empirical performance comparison of different ML models for multi-step TS forecasting.A summary of our work and directions for future work are given in Section 6.

Related Work
Multi-step TS Forecasting: different ML methods and statistical methods have been used for both single-step and multi-step TS analysis in the past few decades [11].These include artificial neural networks [3], support vector machines [4], gradient boosting machine [12], random forest [13], auto-regressive moving average models [5], and exponential smoothing models [6].To use these models for multi-step forecasting, two major approaches are used: direct and recursive strategies [14].The first of these uses multi-output regression models to predict multiple TS steps into the future directly or use multiple models (one for each time step) to make multi-step forecasting.Multi-output models show good performance and require less computational resources than training multiple models to realise multi-step forecasting [15].The second approach uses a single model recursively, using the predicted values as an additional input to forecast the next step.In this case, the error in the prediction may be accumulated [14].Recursive strategies only require one model, which saves significant computational time [14].Other methods based on these two approaches are also used, such as the DirRec strategy [16], which combines the recursive strategies and direct strategies.
TS feature engineering: Feature engineering is an essential component in ML pipelines.Feature extraction methods have also been used for TS analysis tasks [17,18].There are a number of TS feature extraction libraries that are widely used for TS analysis, including tsfresh [19], Catch22 [20], and hctsa [21].Catch22 extracts a selected list of the 22 most useful features of the 4791 features of hctsa from a TS.Extracting features with Catch22 is more computationally efficient than hctsa, with only a 7% reduction in accuracy on average.In the following, we use tsfresh, which extracts more than 700 TS features in parallel and has previously shown strong performance [19], since it does not require huge computational resources like hctsa, or suffers from an accuracy reduction like Catch22.
AutoML: AutoML systems have been used in many domains, such as image classification [22], language processing [23] and energy consumption forecasting [9].However, TS features are not well-studied in AutoML systems.In our earlier work [9], we studied AutoML for short-term load forecasting demonstrating the competitive performance of Au-toML systems.However, we did not investigate the use of TS features.In another work [24], we studied AutoML with TS features for single-step TS forecasting.Our experimental results indicate that AutoML with TS features can further improve the accuracy of AutoML systems.In this work, we focus on studying how to improve the performance of AutoML systems on multi-step forecasting tasks by using TS features.Many AutoML systems have been recently developed, including Auto-sklearn [25], AutoGluon [8] and Auto-Keras [26] .Auto-sklearn is used in our experiments, since it supports various algorithms (e.g., SVMs) that are not available in the other systems, and it is easy to extend.Auto-Keras is used to create a deep learning baseline in our experiments.

Problem Statement
Given a univariate TS x = [x 1 , . . ., x i ] composed of i observations.We are interested in predicting the next k values x = [x i+1 , . . ., x i+k ], where k > 1 denotes the forecasting window.Usually not all the data points show the same influence on the predictions of [x i+1 , . . ., x i+k ].The more recent data points tend to be more important.Specifically, given a TS segment [x i−w+1 , . . ., x i ], we are interested in forecasting of [x i+1 , . . ., x i+k ]; the window size w indicates how much historical data are used to make the prediction.In our previous work [24], we found that the window size w plays an essential role in single-step TS forecasting; specifically, if an ML model gets too much or too little information, this may reduce the model's performance.Here, we extend the automated window size selection technique to multi-step forecasting.Besides this, we use tsfresh to automatically extract features from TS data within these windows.

AutoML for Multi-Step TS Forecasting
We define the Combined Algorithm Selection and Hyperparameter optimisation (CASH) problem [27] for multi-output TS forecasting as the following joint optimisation problem: Given a TS data set x = [x 1 , . . ., x n ] that is split into x train and x valid , we are interested in building an optimised model using x train by minimising loss on x valid .Formally, we define the automated TS forecasting problem as: Let A = {A (1) , . . ., A (k) } be a set of algorithms with associated hyperparameter spaces Λ (1) , . . ., Λ (k) .Let w = {w (1) , . . ., w (l) } be the set of the possible window sizes.Furthermore, let x train be a training set, and x valid be a validation set.Finally, let L(A (i) λ , w (j) , x train , x valid ) denote the loss that algorithm A (i) achieves on x valid when trained on x train with hyperparameters λ ∈ Λ (i) and window size w (j) .Then the automated TS forecasting problem is to find the window size, algorithm, and hyperparameter setting that minimises this loss, i.e., to determine:

Methodology
This section presents the two multi-step forecasting techniques and the AutoML technique enhanced with TS features we use later in our experiments.

Auto-Sklearn with TS Feature Engineering
In this work, we study how we can extend auto-sklearn [25] to perform automatic feature extraction on TS data.Originally, the pipelines constructed by auto-sklearn include a preprocessor, feature preprocessor and ML components.The ensemble construction used in auto-sklearn uses a greedy algorithm to build the ensembles.The workflow of auto-sklearn is illustrated in Figure 1.Auto-sklearn has a powerful feature preprocessor component.However, it does not support any specialised TS feature extractors.In our work, we use our newly proposed TS feature extractors in the search space instead of the feature extractor in auto-sklearn.Automated feature extraction in this case considers both the selection of the window size and extraction of relevant features.Therefore, we propose three variants of auto-sklearn that are specially designed for TS forecasting tasks by replacing the feature extractors of auto-sklearn with one of the following.

1.
Auto-sklearn with automated window size selection (W): the first variant of autosklearn for TS forecasting optimises the window size w.The TS x = [x i−w+1 , . . ., x i ] are used to train a model that predicts [x i+1 , . . ., Auto-sklearn with tsfresh features (T): the second variant of auto-sklearn extracts tsfresh TS features from the TS segment x = [x i−w+1 , . . ., In this case, the window size w is predefined and fixed.The TS features g(x) = g([x i−w+1 , . . . ,x i ]) are calculated using the TS feature extractor g.Feature importance is calculated using the Benjamini-Hochberg procedure [29] to select the important features.The Benjamini-Hochberg procedure selects important features for each step in the TS separately.We then use the union of all the important features to predict [x i+1 , . . ., x i+k ].

3.
Auto-sklearn with automated window size selection and tsfresh features (WT): this approach combines the two previously mentioned approaches.Both window size w and the TS extractor g are optimised in this variant.

Experimental Results
Our key empirical results are based on aggregate performance over 20 data sets and 8 models.More detailed descriptions of the data sets and models are described in the section.

Data Sets
The open-source datasets from CompEngine [30] are used in our experiments.Com-pEngine is a TS data engine containing 197 types of data sets comprising 29,514 in total.These data sets include both real-world and synthetic data sets.We chose ten real-world and ten synthetic data sets from different categories.These are comprised of the following 20 categories: Audio: Animal sounds, Human speech, Music; Ecology: Zooplankton growth; Economics: Macroeconomics, Microeconomics; Finance: Crude oil prices, Exchange rate, Gas prices; Medical: Electrocardiography ECG; Flow: Driven pendulum with dissipation, Duffing-van der Pol Oscillator, Driven van der Pol oscillator, Duffing two-well oscillator, Diffusionless Lorenz Attractor; Stochastic Process: Autoregressive with noise, Correlated noise, Moving average process, Nonstationary autoregressive, Random walk.
Since there are usually more than one data set in each category, we choose the first one of each category.We split every data set into 67% training and 33% test set, based on temporal order, since the data sets are TS.

Experimental Setup
All the experiments were executed on 8 cores of an Intel Xeon E5-2683 CPU (2.10 GHz) with 10 GB RAM.In the experiments, version 0.8.0 of auto-sklearn and version 0.16.1 of tsfresh were used.To evaluate the quality of an ML pipeline, we used quantified error/accuracy.RMSE was used as a performance metric in the optimisation.The maximum evaluation time for one ML pipeline was set to 20 min wall-clock time.The time budget for every AutoML optimisation on each data set was set to 3 h wall-clock time.In these experiments, we used hold-out validation (training:validation = 67:33), the default validation technique in auto-sklearn.The split was carried out only on the training data, such that the optimisation process never sees the test data.However, we did not shuffle the data set in order to preserve the temporal structure of the TS data.All remaining choices were left at their default settings.Since experiments are very time-consuming, we used bootstrapping to create distributions of performance results in order to investigate their variability.Every experiment was run 25 times.We then randomly sampled 5 out of the 25 results and selected the model with the lowest RMSE on the training set out of these five models and reported the RMSE on the test set.We repeated this process 100 times per model and data set.The distributions we showed are based on these 100 values.
We compared the AutoML methods, including Auto-Keras, auto-sklearn and our proposed variants, with traditional ML baselines and N-BEATS.Both recursive and multioutput techniques are used in the ML baselines (GBM and SVM).All other models use the multi-output approach.

•
Gradient Boosting Machine (GBM): Gradient Boosting Machine is a classical ML model used for TS analysis tasks that has shown promising performance in the M3, M4 competitions [31].For hyperparameter optimisation, we performed a random search on GBM with 30 iterations and window size w = 100 [32].In this case, the search space is the same as the search space of GBM in auto-sklearn.In this experiment, we did not split the training set into the training set and validation sets.• Support vector machine (SVM): SVM is another classical ML model that has been used for TS forecasting (e.g., [13]).Similar to GBM experiments, we use 30 iterations of random search and window size w = 100.The search space is the same as the search space of SVM in auto-sklearn.• N-BEATS: N-BEATS [33] uses fully-connected layers with residual links to improve 3% over the winner of the M4 competition, which demonstrates state-of-the-art performance.We used the default hyperparameter settings in the implementation provided by Oreshkin et al. [33] and the bootstrapping approach mentioned in Section 5 to create distributions of results.The number of epochs was set to 500.
• Auto-Keras: Auto-Keras [26] is a neural architecture search system that uses Bayesian optimisation to search for high-performance neural network architectures.Some neural network units available in its search space (e.g., LSTM, GRU), have been used for TS forecasting (see, e.g., [34,35]).Vanilla Auto-Keras does not support multi-output models.To deal with multi-step forecasting tasks, we designed our new search space using three types of blocks available in Auto-Keras: Input block, RNN Block and Regression Head (see Figure 2).The RNN Block is the critical component in our networks.We use RNN as a baseline, as it has been recently studied in the literature on TS forecasting (see, e.g., [34,36]).Several hyperparameters need to be considered for this block, including bidirectionality, the number of layers and layer type (LSTM or GRU).Auto-Keras cannot choose the window size w automatically.We manually preprocessed the data with window size w = 100.We used Bayesian optimisation for architecture search.The number of epochs was set to 100, and we left the remaining settings of Auto-Keras at their default values.

•
Vanilla auto-sklearn (VA): We manually preprocessed the data with window size w = 100 and then fed it to the auto-sklearn.The time budget for the optimisation was set to 3 h.

Auto-sklearn with automated window size selection (W):
For the W variant, we did not need to manually preprocess the data, since the window size w is selected automatically.The window size ranges from 50 to 200.• Auto-sklearn with tsfresh features (T): In the T variant, the TS feature extractor tsfresh was used as an internal component of auto-sklearn.Auto-sklearn used these TS features as input data to search over ML pipelines.The window size was set to w = 100.• Auto-sklearn with automated window size selection and tsfresh features (WT): In WT, we set the window size w to range from 50 to 200.The TS features were extracted from these input data.
Tables 1 and 2 compare the performance achieved by different methods in terms of RMSE on the test set.Table 1, shows the results for traditional ML baseline models, while Table 2 presents the results for AutoML techniques and N-BEATS.To present the results in these tables, we calculated the statistical significance of the results by the non-parametric Mann-Whitney U-test [37] with a standard significance level set to 0.05.The bold-faced entries show the lowest mean RMSE achieved on a given data set, and the * means the RMSE is statistically best.To determine the answer to this question, we compared the recursive and multi-output versions of GBM and SVM algorithms.Among the baselines we consider, N-BEATS as described in the original work is not a recursive model.Therefore, we do not consider it for this analysis.Looking at Table 1, we generally observe that GBM-multioutput performs better than GBM-recursive on 12 out of 20 data sets, while SVM-multioutput outperforms SVM-recursive on 17 out of 20 data sets in terms of RMSE.As we have observed that multi-output models tend to perform better, which is in line with the results from [14].Therefore, we only use the multi-output technique in our next experiments.

Q2: To what extent can AutoML techniques (Auto-Keras, auto-sklearn, and our variants) beat the traditional baselines (GBM, SVM)?
Looking at Tables 1 and 2, one can compare the performance achieved by different methods in terms of RMSE on the test set.We observe that Auto-Keras beats all the traditional ML baseline models (GBM-recursive, GBM-multioutput, SVM-recursive, and SVM-multioutput) on 4 out of 20 data sets.Vanilla auto-sklearn outperforms all the traditional ML baselines on 8 out of 20 data sets.Our three variants W, T, WT show lower error than all the traditional ML baselines on 10, 5, and 5 out of 20 data sets, respectively.The best AutoML (W) outperforms the best traditional ML baseline (SVM-multioutput) on 14 out of 20 data sets.

Q3: To what extent can AutoML techniques beat N-BEATS?
Looking at Table 2, we observe that the best AutoML (W) outperforms N-BEATS on 14 out of 20 data sets.For other AutoML techniques we observe that Auto-Keras, VA, T, and WT beat N-BEATS on 5, 12, 11, and 10 out of 20 data sets, respectively.AutoML methods that are based on standard machine learning are beating this neural networks based model.

Conclusions
In this paper, extend AutoML for multi-step TS forecasting with TS features.We found that AutoML can achieve significantly higher accuracy than the traditional ML baselines on 14 out of 20 data sets in terms of RMSE.Although N-BEATS performs better than Auto-Keras and vanilla auto-sklearn on many data sets, our AutoML TS variants still managed to beat it on 14 out of 20 data sets.We found that the multi-output technique tends to perform better with the same budget than the recursive technique in the multi-step TS forecasting tasks.Overall, these results clearly demonstrate that the use of AutoML techniques and multi-output strategies for multi-step TS forecasting is promising.
Interesting avenues for future work include AutoML for online learning and TS classification.Most AutoML systems focus on a stable data set.Characteristics of TS data might change over time and consequently, the best configuration of data sets may vary over time.We see potential value in extending AutoML on evolving data streams.Furthermore, in TS classification typically classification with a fixed-sized sliding window has been studied.However, the best window size might not be easily determined.Our automated window size selection technique may help to improve the performance of the classification tasks.

Figure 1 .
Figure 1.Workflow of auto-sklearn.Auto-sklearn uses Bayesian optimisation to search the space of ML pipelines, including a preprocessor, feature preprocessor and ML components.

Figure 2 .
Figure 2. Workflow of our customised search space of Auto-Keras for TS forecasting.The input data flows through the RNN Block.Hyperparameters, such as the number of layers and learning rate, will be optimised during the search.The Regression Head then generates output based on the information from the RNN Block.

Table 2 .
RMSE on test set acquired from different AutoML methods including vanilla auto-sklearn (VA), our proposed variants (W, T, and WT), Auto-Keras and the state-of-the-art method N-BEATS.