Next Article in Journal
Synthetic Rainfall Modeling Using a Modified Hybrid Gamma-GP Distribution
Previous Article in Journal
Elastic Time-Lapse FWI for Anisotropic Media: A Pyrenees Case Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Robust Hybrid Forecasting Framework for the M3 and M4 Competitions: Combining ARIMA and Ata Models with Performance-Based Model Selection

by
Tuğçe Ekiz Yılmaz
* and
Güçkan Yapar
Department of Statistics, Faculty of Science, Dokuz Eylül University, Buca, 35390 Izmir, Türkiye
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(17), 9552; https://doi.org/10.3390/app15179552
Submission received: 24 July 2025 / Revised: 23 August 2025 / Accepted: 27 August 2025 / Published: 30 August 2025
(This article belongs to the Section Computing and Artificial Intelligence)

Abstract

Featured Application

The proposed hybrid forecasting model can be applied in business forecasting systems, retail demand planning, and energy load prediction where adaptive model selection is needed under variable time series frequencies.

Abstract

This study proposes a hybrid forecasting framework that integrates the Auto-Regressive Integrated Moving Average (ARIMA) model with multiple variations of the Ata model, using a performance-based model selection strategy to enhance forecasting accuracy on the M3 and M4 competition datasets. For each time series, seven versions of the Ata model are generated by adjusting level and trend parameters, and the version with the lowest in-sample symmetric mean absolute percentage error (sMAPE) is selected. To improve robustness and prevent overfitting, the median-performing Ata model is also included. These selected models’ forecasts are then combined with ARIMA outputs through optimized weighting schemes tailored to the characteristics of each series. Given the varying frequencies (e.g., yearly, quarterly, monthly, weekly, daily, and hourly) and diverse lengths of time series, a grid search algorithm is employed to determine the best hybrid combination for each frequency group. The model is applied in a series-specific manner, allowing it to adapt to different seasonal, trend, and irregular patterns. Extensive empirical results demonstrate that the hybrid model outperforms its individual components and traditional benchmarks across all frequency categories. It ranked first in the M3 competition and achieved second place in the M4 competition based on the official error metric, the sMAPE and Overall Weighted Average (OWA), respectively. The results highlight the framework’s adaptability and scalability for complex, heterogeneous time series environments.

1. Introduction

Time series forecasting is very important in fields like economics, energy, retail, and environmental sciences, where making decisions and plans depends on making accurate predictions. It is especially important to make robust forecasting models for competitiongrade datasets like M3 and M4, which have a wide range of real-world series with different patterns and frequencies.
Traditional statistical methods such as Auto-Regressive Integrated Moving Average (ARIMA) have long been used because they are easy to understand and have a strong theoretical basis. More recently, the Ata method has gained attention for its simplicity, computational efficiency, and adaptability to both seasonal and non-seasonal patterns. However, no single forecasting method performs optimally across all types of time series data. This shows how important it is to use adaptive methods that consider the specific features of each series.
This study proposes a hybrid forecasting framework that integrates the best variants of the Ata methods and ARIMA, employing a data-driven optimization process based on grid search and in-sample sMAPE minimization for model selection. By evaluating multiple variations of the Ata model for each series and selecting those with the best in-sample performance, we aim to construct exclusive forecasts that are both accurate and reliable. The proposed method is tested on these datasets, and its performance is evaluated using the official accuracy measures used by the competitions. This research contributes to the literature by demonstrating how statistical modeling and adaptive model selection can be effectively integrated to improve forecasting accuracy across different time series.

2. Literature Review

Time series forecasting is now a very important field that combines statistics, machine learning, and decision sciences. This is because there is a growing need for accurate predictions in many areas, including economics, healthcare, and environmental science. As problems in the real world become more complicated and there are more data available, the need for strong and flexible forecasting methods has grown even more. Over the years, the development of forecasting methods has been influenced by both theoretical progress and large-scale real-world tests, especially through international benchmark competitions. The M3 and M4 forecasting competitions have been especially important because they have led to new ways of doing things and more thorough comparisons. Researchers have been able to test, compare, and improve their methods in these competitions, which has sped up progress in the field.
The M3 competition, which was organized by Makridakis and Hibon [1], was significant because it examined 24 different forecasting methods on 3003 time series. One of the most striking findings was that relatively simple statistical models—such as exponential smoothing and ARIMA—performed on par with, and sometimes even outperformed, more complex and sophisticated approaches. This went against the common belief that making a model more complex always makes it more accurate at predicting. The M3 competition showed that complexity for its own sake is not always useful and that classical statistical methods are still useful today. Because of this, the M3 competition resulted in a lot of research into hybrid and ensemble methods that try to combine the ease of use and understanding of traditional models with the power and flexibility of more advanced ones. This change has contributed to the creation of forecasting frameworks that are useful for researchers. As illustrated in Figure 1, the M3 dataset spans multiple frequencies and domains, providing a rich ground for testing and comparing forecasting methods.
Building on the foundation laid by M3, the M4 competition [2,3] expanded the scope substantially, including 100,000 time series and 61 forecasting methods, comprising a wide array of statistical, machine learning, and hybrid approaches. The unprecedented scale and diversity of the M4 dataset allowed for a more comprehensive assessment of forecasting methods across different domains and time series characteristics. As illustrated in Figure 2, the dataset covers multiple frequencies—yearly 23,000 (23%), quarterly 24,000 (24%), monthly 48,000 (48%), weekly 359 (0.36%), daily 4,227 (4.23%), and hourly 414 (0.41%)—and spans diverse domains (micro, industry, macro, finance, demographic, and other categories), thereby capturing heterogeneous structures for model evaluation. The results of M4 confirmed the idea that no single method consistently dominates across all types of time series. Instead, ensemble and combination methods—particularly those that integrate statistical and machine learning models—tended to perform the most robust and reliable performance. This finding has important real-world implications. It suggests that being flexible and adaptable is essential for making accurate predictions in diverse settings. The M4 competition also showed that how important it is for forecasting research to be reproducible and open, establishing new standards for future studies.
One of the earliest and most influential works in this line of research was Zhang [4], who introduced a hybrid ARIMA and Artificial Neural Network (ANN) model. This pioneering study established the foundation for combining linear time series components modeled by ARIMA with the nonlinear learning capacity of neural networks, achieving better performance than either method separately. Building on this foundation, Atesongun and Gulsen [5] extended the ARIMA–ANN framework with a structured hybrid forecasting model, confirming its efficiency across multiple real-world datasets. Similarly, Tsoku et al. [6] proposed a hybrid ARIMA–ANN model to forecast South African crude oil prices, showing that the integration of Box–Jenkins methodology with neural networks enhances predictive accuracy in volatile financial markets. In a related contribution, Alshawarbeh et al. [7] applied the ARIMA–ANN hybrid approach to high-frequency datasets, underscoring its robustness in handling rapidly changing series and reducing forecast errors in complex environments. More recently, Smyl [8], the overall winner of the M4 competition, proposed a hybrid method of exponential smoothing (ETS) and Recurrent Neural Networks (RNNs). This study provided compelling evidence that hybrid models, when carefully designed, can significantly outperform both classical statistical methods and standalone neural networks, setting a new benchmark for large-scale forecasting competitions. Together, these works demonstrate the enduring importance of hybrid statistical–machine learning approaches as benchmarks and their adaptability across diverse forecasting contexts.
Recent research has emphasized hybrid forecasting frameworks that combine traditional statistical models with regression-based or machine learning approaches to enhance predictive accuracy. For example, Iftikhar et al. [9] modeled and forecasted carbon dioxide emissions by integrating regression techniques with time series models. Their study demonstrated that a hybrid structure can capture both linear and nonlinear dynamics, offering more reliable forecasts in environmental applications. Similarly, Sherly et al. [10] proposed a hybrid ARIMA–Prophet model, showing that the integration of Prophet’s trend–seasonality decomposition with ARIMA’s statistical rigor yields superior accuracy compared to either method alone. Ampountolas [11] conducted a comparative analysis of machine learning, hybrid, and deep learning forecasting methods in European financial markets and Bitcoin, highlighting the competitive advantage of hybrid designs in complex financial domains. These studies highlight that hybrid designs are not restricted to financial or technological data but can be successfully applied in sustainability and policy-relevant domains as well.
The integration of deep learning into hybrid forecasting has further advanced the field. Qin et al. [12] proposed a BiLSTM–ARIMA hybrid optimized with a whale-driven metaheuristic algorithm for financial forecasting. Their results showed significant improvements in capturing long-term dependencies and market volatility. Likewise, Çınarer [13] presented a hybrid deep learning and stacking ensemble model for global temperature forecasting, effectively addressing the challenges of nonlinear climate dynamics. Dong and Zhou [14] introduced a novel hybrid forecasting framework that integrates the ARIMA model with Convolutional Neural Networks (CNN) and Long Short-Term Memory (LSTM) networks, combined with Complete Ensemble Empirical Mode Decomposition with Adaptive Noise (CEEMDAN) and Sample Entropy (SE) techniques for financial forecasting. Their approach demonstrated superior performance by leveraging signal decomposition, statistical modeling, and deep neural networks in a unified structure. These contributions underline that integrating deep learning with statistical backbones allows hybrid models to operate effectively across both financial and environmental domains.
Hybrid forecasting models have also proven valuable in areas where interpretability and domain-specific insights are crucial. Iacobescu and Susnea [15] developed an ARIMA–ANN framework for crime risk forecasting, incorporating socioeconomic and environmental indicators. Their approach not only improved accuracy but also enhanced interpretability, enabling policymakers to connect predictions with underlying risk factors. In the domain of financial markets, Liagkouras and Metaxiotis [16] combined LSTM networks with sentiment analysis for stock market forecasting. This hybrid captured both quantitative and qualitative signals, highlighting the value of integrating textual data sources with numerical series. Finally, Liu et al. [17] introduced a multi-scale hybrid forecasting framework that combined traditional statistical models with deep learning to handle heterogeneous time series across varying temporal resolutions. Their results emphasized scalability and adaptability, making the framework applicable to multi-domain forecasting challenges.
Model selection has become a major topic in modern forecasting research because time series data are naturally diverse. Because time series structures, frequencies, and outside factors can be so different, flexible methods are required that can choose the best model for each series based on its own unique features. For instance, Karim et al. [18] showed how useful it is to add external data sources like Google Trends to models that forecast macroeconomic indicators in Australia. Their research showed that using relevant outside information along with dynamic model selection can greatly improve the accuracy of predictions, especially in environments that change quickly. In a similar manner, Hananya and Katz [19] proposed dynamic model selection frameworks that automatically select the best machine learning model in real time, based on changing series characteristics. This is another step forward in the field of adaptive forecasting. All of these studies show that more and more people are realizing that choosing a model is not a one-time thing but an ongoing process that needs to change with new data and outside factors. Moreover, recent advances in artificial intelligence have also begun to influence model selection processes. Wei et al. [20] came up with a new way to automate the recommendation of forecasting models by using large language models (LLMs). Their framework makes the model selection process easier by using the natural language processing capabilities of LLMs. This points to a future where artificial intelligence (AI) not only builds models but also helps people choose and use them in smart ways. This new development opens new ways to combine expert knowledge and data-driven insights. It could also make advanced forecasting tools available to more people.
As these developments occurred, the Ata method has become an important method in statistical forecasting. The Ata method was first introduced as an alternative to Holt’s linear trend method [21]. Since then, it has been recognized and tested in competitive settings [22,23]. Its ability to handle various types of time series with minimal parameter tuning makes it particularly appealing in practice. Furthermore, the method’s simplicity, interpretability, and accuracy make it a valuable candidate for hybrid and ensemble forecasting frameworks. Its rising popularity in both academic and real-world settings shows its usefulness.
Based on these findings, this study suggests a hybrid forecasting framework, namely, the Ata-best model, based on model selection and examines the performance of the Ata and ARIMA methods using the M3 and M4 datasets. The study’s goal is to improve forecasting accuracy at the individual series level by systematically finding the best-performing model for each time series based on the in-sample symmetric mean absolute percentage error (sMAPE). This method not only uses the real-world results from major forecasting competitions, but it also incorporates new developments in model selection and hybridization. This makes it a better way to forecast time series in both environmental and general settings. Consequently, the study aims to help improve forecasting by showing how adaptive, data-driven model selection can help achieve both accuracy and interpretability.
To highlight the comparative strengths of the two statistical approaches employed in this study, Table 1 provides a structured summary of the ARIMA and Ata methods. The table illustrates their key characteristics, including modeling assumptions, adaptability, computational efficiency, and practical limitations. By presenting advantages and disadvantages side by side, it becomes clear that while ARIMA offers a well-established framework with strong theoretical foundation, the Ata method introduces a flexible and adaptive alternative that is particularly suitable for heterogeneous or unstable series. This comparative view not only clarifies the methodological reasons behind selecting these two models for further investigation, but also emphasizes their complementary potential in hybrid forecasting frameworks.

3. Methodology

In this study, a model selection-based hybrid approach is utilized for time series forecasting. For each time series, both the Ata and ARIMA models are applied. For the ARIMA model, we used the official forecast values declared in the M3 and M4 competitions. For the Ata method, we choose the two versions of the Ata model: the one model with the lowest in-sample sMAPE (lowest Ata) and the one model with the median in-sample sMAPE (median Ata). A grid search algorithm was used to find these models because of the difference in the dataset’s frequency and the varying structural characteristics of the time series, as well as to determine how to weight the contributions of the ARIMA- and Ata-based models in the hybrid structure. The goal of this strategy is to improve the overall accuracy of the forecasts by finding the best method, i.e., “Ata-best” for each series based on its own unique features. In this section, the mathematical foundations of the employed methods and the construction of the hybrid framework are presented in detail. Below, the fundamental formulations of the ARIMA and Ata methods used in this study are provided.
Before introducing the individual forecasting methods, the overall structure of the hybrid framework is pictured in Figure 3. As shown, the M3 and M4 datasets are modeled using multiple variations of the Ata method and the official ARIMA forecasts from both competitions. A hybrid framework is then constructed through weight optimization via grid search, producing the final hybrid forecasts. Additionally, model selection procedures, such as in-sample sMAPE evaluation and alternative aggregation strategies (Ata-lowest and Ata-median), are incorporated to ensure robustness in performance evaluation.

3.1. ARIMA

The ARIMA model, first formalized by Box and Jenkins [24], is a fundamental statistical method for time series forecasting. When it comes to modeling and forecasting linear relationships in univariate time series data, the ARIMA model is considered the most traditional and widely used method. The model is called ARIMA (p, d, q), where p is the autoregressive (AR) component’s order, d is the number of differences needed to reach stationarity, and q is the moving average (MA) component’s order. The general ARIMA model is expressed as follows:
ϕ ( B ) ( 1 B ) d y t = θ ( B ) ε t
where y t is the value at time t, B is the backshift operator ( B y t = y t 1 ), ( 1 B ) d is the differencing operator applied d times, ϕ ( B ) and θ ( B ) are polynomials of orders p and q, and ε t is white-noise error.
The ARIMA model captures linear relationships in stationary univariate time series by combining autoregressive, differencing, and moving average components. Choosing a model usually entails determining the appropriate orders for each component and testing for stationarity. Despite being popular due to its simplicity and efficiency with linear patterns, ARIMA has limitations when it comes to handling nonlinear dynamics, structural breaks, or external regressors. In such complex scenarios, these limitations often lead to the adoption of more sophisticated or hybrid forecasting techniques.

3.2. Ata Method

The Ata method is an innovative approach to use statistical analysis to make forecasts about the future. It is based on traditional forecasting techniques for time series, the most famous of which is Holt’s linear trend method. The Modified Holt’s Linear Trend Method [21] was the first name given to Ata. Later, it was made official as a new way of modeling that is different from traditional models such as ARIMA and Exponential Smoothing [22]. The main benefit of the Ata method over traditional Exponential Smoothing is that it can easily adjust the parameters like trend and level component when the model assumptions are right. This makes it more accurate when predicting time series with complicated trend structures. Notably, the Ata method is based on the ideas behind classical exponential smoothing, but it stands out because it uses adaptive smoothing weights that change over time. This lets the model react quickly to sudden changes in the data. Ata can also be used in both additive and multiplicative forms, and a dampening factor can be added to the trend component to control how long the trend lasts.
The formal mathematical representation of the Ata model is provided below for both additive and multiplicative forms. Here, S t denotes the level component, T t represents trends, and X ^ t ( h ) is the h-step-ahead forecast at time t. The additive form is denoted as A t a a d d ( p , q , ϕ ) , and the multiplicative form is denoted as A t a m u l t ( p , q , ϕ ) , where ϕ = 1 for non-dampened data.
  • Additive Ata Model: A t a a d d ( p , q , ϕ )
    S t = p t X t + t p t ( S t 1 + T t 1 ) if t > p
    S t = X t if t p
    T t = q t ( S t S t 1 ) + t q t T t 1 if t > q
    T t = X t X t 1 if t q
    T t = 0 if t = 1
    X ^ t ( h ) = S t + h T t
  • Multiplicative Ata Model: A t a m u l t ( p , q , ϕ )   
    S t = p t X t + t p t ( S t 1 T t 1 ) if t > p
    S t = X t if t p
    T t = q t S t S t 1 + t q t T t 1 if t > q
    T t = X t X t 1 if t q
    T t = 1 if t = 1
    X ^ t ( h ) = S t T t h
    where t > p q , p { 1 , 2 , , n } , q { 0 , 1 , , p } , and h = 1 , 2 , .

4. Data and Preprocessing

In this study, we used the M3 and M4 competition datasets, two of the most wellknown and widely used time series datasets in the forecasting literature. Because of their size and diversity, these datasets are frequently used as benchmarks to assess forecasting techniques. Both datasets include a wide range of real-world time series, with differences in length and frequency as well as in the underlying domains and structural features. This variety offers a comprehensive framework for evaluating the generalizability, flexibility, and robustness of forecasting models in various application domains.
There are 3003 time series in the M3 dataset, all of which have been methodically categorized by frequency and domain. Micro, macro, industry, finance, demographics, and other are the six domains in which the series are specifically dispersed. The M3 dataset also includes four primary frequency types: yearly, quarterly, monthly, and other frequencies. From short-term to long-term horizons, from economic indicators to demographic trends, this dual categorization ensures that the dataset captures a broad range of real-world temporal dynamics and forecasting challenges. Likewise, with 100,000 time series, the M4 dataset provides an even bigger and more varying collection. The M4 dataset, like M3, contains series from a variety of domains, including demographics, industry, finance, and macro. The frequency spectrum is further broadened by M4, which offers series at hourly, daily, weekly, monthly, quarterly, and yearly intervals. This thorough framework makes it possible to assess forecasting models over a far larger range of application domains and temporal resolutions.
The primary objective of this research was to establish a hybrid statistical modeling framework that could combine and adaptively choose the best forecasting models for each individual time series. In order to achieve this, we concentrated on two popular statistical forecasting techniques: the Ata method, a more modern and adaptable modeling technique intended for automated and effective forecasting, and the well-known ARIMA model, which has long been a benchmark method in time series literature. A crucial part of our methodology was the implementation of a performance-driven model selection approach for the Ata method, where grid search and in-sample sMAPE minimization were employed. For each time series, we evaluated seven distinct Ata model variations which consist of without trend parameter, trend parameter with additive model, trend parameter with multiplicative model, level-fixed additive model (i.e., choosing optimum level and trend parameter with additive model), level-fixed multiplicative model (i.e., choosing optimum level and trend parameter with multiplicative model), a simple average of the first and second models, and a simple average of the first and third models, respectively. The selection was based on the in-sample sMAPE, and for each series, we identified two representative Ata models: the one with the lowest sMAPE Ata model (“Ata-lowest”) and the one with the median sMAPE Ata model (“Ata-median”), capturing both optimal and robust model performance. This dual selection strategy allowed us to capture both the optimal and a more robust, central tendency of Ata model performance.
For the ARIMA component, although we first experimented with auto.ARIMA based forecasts, we finally decided to use the official ARIMA forecasts published as part of the M3 and M4 competition results. Our hybrid framework’s dependability and comparability were guaranteed by empirical comparisons that consistently showed that the benchmark ARIMA forecasts performed better than our own implementations in terms of out-of-sample accuracy.
To integrate the strengths of both the Ata and ARIMA models, namely, obtaining the “Ata-best” model, we employed a grid search algorithm to determine the optimal combination weights for each time series. In order to minimize the forecasting error, this weight optimization was performed separately for each series due to the significant heterogeneity in series frequency, length, and domain. Our hybrid model was able to maximize predictive accuracy by flexibly adapting to the distinct features of each time series thanks to this customized approach. All computations were performed in R (version 4.4.3) through the RStudio environment (version 2024.12.1+563). The Ata method was implemented using the ATAforecasting package (version 0.0.60). The code was structured to facilitate reproducibility and enable efficient handling of a wide range of time series.
Hence, this study’s methodological pipeline was identified by its thorough statistical foundations, systematic methodical data-driven model selection, and repeatable validation processes. Our hybrid approach takes advantage of the scale and diversity of the M3 and M4 datasets to offer a strong foundation to enhance time series forecasting.

5. Discussion

This study discusses the reason for using ARIMA and Ata methodologies within the forecasting framework, while also considering their individual strengths and limitations. ARIMA has long been considered a benchmark statistical model for time series forecasting because it has an established theoretical basis and can capture linear relationships in a straightforward manner using autoregressive, differencing, and moving average components. Its interpretability and analytical rigor make it a standard reference point in empirical forecasting studies. However, ARIMA’s dependence on linearity and stationarity limits its performance when nonlinear dynamics, structural breaks, or irregular seasonal behaviors are present. This makes it less flexible in complicated real-world situations. The Ata method, on the other hand, is a new and modern way to generate forecasts that builds on the ideas of exponential smoothing by using adaptive updating mechanisms. Its structure allows parameters to evolve dynamically over time, enabling the model to respond flexibly to shifts in level and trend. Moreover, the Ata method’s capacity to incorporate additive and multiplicative forms across diverse series enhances its robustness, while its algorithmic simplicity makes it computationally efficient and easily scalable to large datasets, such as those encountered in forecasting competitions. These qualities represent Ata as a versatile and easily adaptable method that bridges the gap between simplicity and flexibility, thereby complementing the more rigid structure of ARIMA. Taken together, the comparative use of these two approaches not only ensures methodological robustness but also highlights the potential for hybrid frameworks that integrate the interpretability of ARIMA with the adaptability of Ata, ultimately broadening the scope of effective forecasting strategies.
Table 2 summarizes the optimized hybrid weights assigned to the ARIMA, Atalowest, and Ata-median models for different frequency categories in the M3 dataset. These weights were obtained using a grid search procedure combined with in-sample sMAPE minimization. The grid search systematically explored candidate weight combinations, and the final values correspond to those that minimized the error for each frequency type. A higher weight indicates a stronger contribution of the corresponding model to the hybrid forecast. The results highlight that the Ata-median model dominates for yearly series, while the Ata-lowest model gains importance in other categories, and ARIMA plays a larger role in monthly and quarterly data.
The sMAPE evaluation results, presented in Table A1, Table A2, Table A3 and Table A4, highlight key findings across all frequency categories—namely, yearly, quarterly, monthly, and other—together with the overall performance of the proposed hybrid model (Ata-best). These findings provide compelling evidence of the benefits of model combination strategies that are customized using performance-based weighting. Combining complementary patterns captured by various Ata variations, the hybrid model produces more accurate forecasts, outperforming both the lowest and median Ata variants. Since the sMAPE was adopted as the official error measure of the M3 competition and, in addition, it is a scale-independent metric, it was chosen as the primary criterion when analyzing the M3 dataset in this study. The scale-independence property is particularly important because the M3 dataset contains time series with highly diverse magnitudes, ranging from small-scale microeconomic indicators to large-scale macroeconomic aggregates. By using the sMAPE, all series can be evaluated on a comparable basis, ensuring a fair assessment of model accuracy without being biased toward the absolute scale of the data.
To maintain the coherence of the main text and avoid overloading it with high-volume numerical tables, the detailed frequency-level results have been moved to the Appendix A. This ensures both readability and completeness, allowing interested readers to examine the extended evidence without interrupting the flow of the discussion.
Importantly, the Ata-best model outperforms the ARIMA model, which was known as B-J Automatic in the original M3 competition tables. Given the Box–Jenkins methodology’s historical success and wide application to time series forecasting, this finding is especially noteworthy. The hybrid Ata-based approach shows that it is not only competitive but also robust in a variety of time series structures by outperforming B–J Automatic in the entire dataset and in all frequency subsets. These results also demonstrate how adaptable the suggested hybrid structure is, as it can dynamically adapt to underlying patterns in the data without depending entirely on strict statistical assumptions. Moreover, the results indicate that the Ata model’s potential is further increased when utilized as part of a weighted ensemble that takes into account the complementary strengths of various configurations and models, even though the model by itself already provides a competitive performance due to its adaptive structure.
In conclusion, the proposed hybrid approach “Ata-best” performs better in terms of forecasting accuracy across all frequency tables than both the traditional ARIMA benchmark and all Ata models. In the case of the M3 competition, the suggested Ata-best model outperforms the ARIMA model and previous Ata models, as well as the Theta model, which was the M3 competition’s overall winner, in almost every forecasting horizons. According to Table 3, the Ata-best model achieved a lower sMAPE than the Theta method, the official winner of the M3 competition; therefore, we can state that Ata-best can be considered the top-performing model for the M3 dataset. This outstanding performance not only demonstrates the Ata method’s inherent strength but also the effectiveness and adaptability of statistical hybrid modeling when applied through performance-driven model combination. To further clarify the contribution of each individual component, an ablation study was conducted (Table A7), showing that the superiority of Ata-best arises from systematic improvements over Ata-lowest, Ata-median, ARIMA, and the winning Theta model across different forecasting horizons. These findings offer strong evidence that, when properly planned, adjusted, and tailored to the structural properties of the data, hybrid statistical approaches can obtain substantial improvements in challenging forecasting scenarios such as the M3 dataset.
The best hybrid weights for each frequency type in the M4 dataset—ARIMA, Atalowest, and Ata-median—are shown in Table 4. The weight distributions differ across frequencies, just like in the M3 results, suggesting that the contribution of each model component should be modified according to the time series’ characteristics. However, Table 5, Table A5 and Table A6’s comprehensive evaluation metrics offer the most important insights. As in the case of the M3 dataset, the detailed frequency-level results for M4 are also provided in the Appendix B to preserve the readability of the main text.
Table A5, which presents performance based on the sMAPE metric, makes it clear that the suggested hybrid model Ata-best outperforms the conventional ARIMA model and the individual Ata variants (Ata-lowest and Ata-median) in terms of forecasting accuracy. Out of all the models that take part in the M4 competition, the Ata-best model comes in third place (rank 3 * ) overall, outperforming Ata-lowest (rank 21), Ata-median (rank 14), and ARIMA (reported in the official table of the M4 study as The M4 Team (ARIMA) with rank 29) according to the sMAPE ranking. Here, the star ( * ) indicates that, without altering the original table reported in the official M4 paper [3], we have highlighted the improvement by showing that our proposed model rises to the third position. This notation is also consistently applied in Table 5 and Table A6, where it conveys the same meaning. This consistent benefit across many frequency ranges (from hourly to yearly) demonstrates the hybrid approach’s endurance in terms of scale-invariant performance and directional accuracy.
Table A6, which displays performance based on the MASE (mean absolute scaled error) metric, provides additional evidence of improvement. The Ata-best model ranks third as regards the MASE metric and again produces lower MASE values than its counterparts i.e., Ata models and ARIMA. This implies that the hybrid model improves performance in scale-sensitive error measurement contexts in addition to improving the sMAPE, which is frequently used in the M4 competition. Importantly, the MASE, like the sMAPE, is a scaleindependent error metric, which makes it particularly suitable for comparing forecasting accuracy across heterogeneous time series of different magnitudes. Its inclusion in this study ensures that the evaluation is not biased by series scale and strengthens the robustness of the comparative analysis.
According to Table 5, the Ata-best model achieves an outstanding second place among all competing methods, which takes into account the OWA metric, the official evaluation criterion used to determine final rankings in the M4 competition. This further reinforces its ability to generalize well across both high-frequency (e.g., hourly, daily) and lowfrequency (e.g., yearly, quarterly) time series, which are often characterized by different noise structures and trend dynamics. It is the OWA metric that integrates both into a single performance indicator and ultimately establishes the leaderboard rankings, even though sMAPE and MASE are also reported for transparency and interpretability. This result further supports the suggested hybrid model’s ability to generalize across these categories. Notably, the Ata-best model outperforms the ARIMA model, which is known as a classical benchmark in statistical forecasting, on every evaluation metric. For transparency, Table A8 provides an ablation study isolating the roles of ARIMA, Ata variants, and the winning Smyl model, confirming that the Ata-best configuration reliably secures the second rank overall. The results indicate that, even with ARIMA’s historical significance and extensive use in many ensemble techniques, a well-designed Ata-based hybrid can be a more precise and adaptable substitute, particularly in a major forecasting competition like M4.
In the official M4 paper [3], the reported runtimes highlight that several participants who submitted Ata-based models, including Selamlar, Taylan, Yapar et al., Yilmaz, and Çetin listed in Table 5, Table A5 and Table A6, implemented their approaches through a standalone .exe platform. According to these records, their computations ranged from approximately 37.2 min to 393.5 min, depending on the specific configuration and dataset frequency. By comparison, some of the benchmark models reported in the same study required substantially longer runtimes, with the winning hybrid ETS–RNN method proposed by Smyl [8] taking 8056.0 min and the ARIMA benchmark requiring 3030.9 min. In contrast, the present study extends this line of research by incorporating both the ARIMA benchmark and multiple variants of the Ata method within a unified framework, executed entirely in the RStudio environment. Unlike single-model implementations, our methodology systematically evaluates seven distinct variations of the Ata model for each individual series, selecting the most suitable versions (“Ata-lowest” and “Ata-median”) based on in-sample sMAPE. This model selection process, combined with the subsequent gridsearch-based weight optimization for hybridization with ARIMA, substantially increases the computational burden. As a result, the complete experimental pipeline required approximately 2880 min for the M4 dataset. This comparison underscores that, although the hybrid design involves a more comprehensive and computationally demanding workflow than individual Ata-based runs, its runtime remains considerably more efficient than some state-of-the-art benchmarks (e.g., Smyl’s ETS–RNN or ARIMA benchmark). Thus, the proposed hybrid framework achieves a favorable balance between predictive accuracy and computational feasibility.

6. Conclusions

This study proposed a novel hybrid framework for time series forecasting that integrates the classical ARIMA model with multiple variations of the Ata method through a performance-based model selection and grid search optimization weighting strategy. By combining the statistical robustness of ARIMA with the adaptive flexibility of the Ata models, the approach aims to capture both linear dependencies and more complex structural patterns inherent in time series data.
The effectiveness of the proposed framework was rigorously evaluated on two largescale and widely recognized forecasting benchmarks: the M3 and M4 competition datasets. These datasets encompass a broad spectrum of time series with diverse frequencies and levels of complexity, providing a comprehensive testing environment. Results from the M3 dataset demonstrate that the proposed Ata-best model achieved a superior forecasting performance compared to both the traditional ARIMA benchmark and existing Ata variations. Importantly, according to Table 3, the Ata-best model also outperformed the Theta method, which had been recognized as the overall winner of the M3 competition, thereby underscoring the competitive edge and accuracy of the proposed hybrid approach.
In the case of the M4 dataset, the generalizability of the method was further confirmed. The Ata-best model consistently ranked among the top-performing approaches across different frequencies, ultimately taking the second position overall based on the official OWA metric, according to Table 5. Furthermore, the model exhibited stable improvements over ARIMA, Ata variants, and other benchmark methods, demonstrating its robustness in handling both stationary and non-stationary components of large-scale real-world time series.
These findings collectively emphasize several significant contributions. First, the study demonstrates that hybridized statistical models, when thoroughly adjusted and weighted, can outperform both individual traditional methods and benchmarks that have won competitions. Second, the results provide strong evidence that the Ata method, particularly when embedded in a hybrid structure, constitutes a highly competitive alternative for practical forecasting tasks. Finally, the suggested framework is not only accurate but also scalable and adaptable, making it useful for a wide range of real-world forecasting circumstances, from finance and economics to energy demand and environmental applications.
In conclusion, this research demonstrates that the integration of ARIMA with the Ata family of models through performance-driven combination results in substantial accuracy improvements in difficult forecasting situations. The success of the Ata-best model on both the M3 and M4 datasets provides strong evidence that hybrid approaches are a promising direction for improving statistical forecasting methods.
In future studies, the proposed hybrid framework could be extended in several promising directions. One avenue is the integration of machine learning and deep learning components, which may enhance the ability to capture complex nonlinear dynamics and long-range dependencies that are difficult for traditional models alone to address. Additionally, expanding the framework to multivariate forecasting tasks, for instance, using the M5 forecasting competition, would allow researchers to incorporate multiple interdependent series, thereby capturing richer temporal interactions. Beyond this, future work could also focus on incorporating exogenous variables (e.g., macroeconomic indicators or weather conditions), which may significantly improve forecasting accuracy in applied domains such as energy demand, financial markets, and retail planning. Collectively, these directions would strengthen the generalizability, interpretability, and practical applicability of the hybrid forecasting approach beyond competition datasets and into real-world decision-making contexts.

Author Contributions

T.E.Y.: conceptualization, methodology, software, validation, writing—original draft preparation, writing—review and editing, visualization; G.Y.: conceptualization, methodology, validation, supervision. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by both The Scientific and Technological Research Council of Türkiye (TÜBİTAK) for 2211/A National PhD scholarship program and by The Council of Higher Education (YÖK) 100/2000 PhD Scholarship Program within the scope of Bio-statistics and Bioinformatics.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets analyzed during the current study are publicly available. The M3 dataset can be accessed at https://forecasters.org/resources/time-series-data/m3-competition/ (accessed on 12 January 2025), and the M4 dataset is available at https://forecasters.org/resources/time-series-data/m4-competition/ (accessed on 1 April 2025).

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ARIMAAuto-Regressive Integrated Moving Average
SARIMASeasonal Auto-Regressive Integrated Moving Average
ETSExponential Smoothing
ANNArtificial Neural Network
RNNRecurrent Neural Network
CNNConvolutional Neural Network
LSTMLong Short-Term Memory
CEEMDANComplete Ensemble Empirical Mode Decomposition with Adaptive Noise
SESample Entropy
sMAPESymmetric Mean Absolute Percentage Error
MASEMean Absolute Scaled Error
OWAOverall Weighted Average
LLMsLarge Language Models
AIArtificial Intelligence
ARAuto-Regressive
MAMoving Average

Appendix A. Additional Results: M3 Frequency-Specific Tables

Appendix A provides the detailed results for the M3 competition dataset. While the main text presents only the overall sMAPE outcomes for clarity, the frequency-level results are reported here. Table A1, Table A2, Table A3 and Table A4 contain the sMAPE values for the yearly, quarterly, monthly, and other frequency levels, respectively. These supplementary tables allow readers to observe how the proposed hybrid model performs across different temporal resolutions, thereby providing deeper insights into the model’s robustness and generalizability.
Table A1. M3 competition yearly sMAPE results.
Table A1. M3 competition yearly sMAPE results.
MethodForecasting HorizonsAverages
12 3 4 5 6 1–41–6
Naive28.513.217.819.923.024.914.8517.88
Single8.513.317.619.822.824.814.8217.82
Holt8.313.719.022.025.227.315.7719.27
B-J automatic8.613.017.520.022.824.514.7817.73
ForecastPro8.312.216.819.322.224.114.1517.14
Theta8.012.216.719.221.723.614.0216.90
RBF8.212.116.418.320.822.713.7516.42
ForcX8.612.416.118.221.022.713.8016.48
ETS9.313.618.320.823.425.815.4818.53
Ata(p, 0)9.113.517.619.922.825.115.0418.00
Ata(p, 1)8.312.216.818.621.523.313.9516.78
Ata–comb8.412.316.518.321.022.713.8716.54
Ata–lowest8.012.116.418.921.924.013.8516.88
Ata–median8.111.415.217.119.621.112.9415.40
Ata–best7.911.215.117.119.521.112.8215.31
Table A2. M3 competition quarterly sMAPE results.
Table A2. M3 competition quarterly sMAPE results.
MethodForecasting HorizonsAverages
1 2 3 4 5 6 8 1–4 1–6 1–8
Naive25.47.48.19.210.412.413.77.558.829.95
Single5.37.27.89.210.212.013.47.388.639.72
Holt5.06.98.310.411.513.115.67.679.2110.67
B-J automatic5.57.48.49.910.912.514.27.799.1010.26
ForecastPro4.96.87.99.610.511.913.97.288.579.77
Theta5.06.77.48.89.410.912.07.008.048.96
RBF5.77.48.39.39.911.412.67.698.679.57
ForcX4.86.77.79.210.011.613.67.128.359.54
ETS5.06.67.99.710.912.114.27.328.719.94
Ata(p, 0)5.27.17.89.710.111.813.57.458.629.71
Ata(p, 1)5.36.87.69.19.911.012.47.198.289.24
Ata–comb5.16.87.59.09.610.912.37.108.139.07
Ata–lowest5.06.77.69.29.711.212.87.108.229.26
Ata–median5.06.87.79.19.711.012.67.168.229.23
Ata–best4.96.67.59.09.410.812.46.988.039.03
Table A3. M3 competition monthly sMAPE results.
Table A3. M3 competition monthly sMAPE results.
Forecasting HorizonsAverages
Method 1 2 3 4 5 6 8 12 15 18 1–4 1–6 1–8 1–12 1–15 1–18
Naive215.013.515.717.014.914.415.616.019.320.715.3015.0815.2615.5516.1616.89
Single13.012.114.015.113.513.113.814.518.319.413.5313.4413.6013.8314.5115.32
Holt12.211.613.414.613.613.313.714.818.820.212.9513.1113.3313.7714.5115.36
B-J automatic12.311.712.814.312.712.313.014.117.819.312.7812.7012.8613.1913.9514.80
ForecastPro11.510.711.712.911.812.012.613.216.418.311.7211.7812.0212.4313.0713.85
Theta11.210.711.812.412.212.713.216.218.218.211.5411.7512.0912.4813.0913.83
RBF13.712.313.714.312.312.513.514.117.317.813.4913.1413.3613.6414.1914.76
ForcX11.611.212.614.012.412.012.813.917.818.712.3212.2812.4412.8113.5814.44
ETS11.510.612.313.412.313.213.214.117.618.911.9312.0512.4312.9613.6414.45
Ata(p, 0)11.510.812.613.812.612.512.913.917.318.912.2012.3312.7812.9813.6714.49
Ata(p, 1)11.010.912.213.412.813.013.815.418.920.911.8612.1612.6013.8714.3915.33
Ata–comb11.110.712.113.112.011.912.413.116.317.411.7411.8112.0412.4813.0713.75
Ata–lowest11.610.812.212.912.712.713.013.916.918.411.9012.1612.5713.0213.6114.33
Ata–median11.810.712.313.012.312.312.813.716.517.911.9412.0612.4412.8613.4114.09
Ata–best11.610.512.113.011.611.812.413.316.417.811.7911.7712.0712.4413.0613.76
Table A4. M3 competition other sMAPE results.
Table A4. M3 competition other sMAPE results.
Method Forecasting HorizonsAverages
1 2 3 4 5 6 8 1–4 1–6 1–8
Naive22.23.65.46.37.87.69.24.385.496.30
Single2.13.65.46.37.87.69.24.365.486.29
Holt1.92.93.94.75.85.67.23.324.134.81
B-J automatic1.83.04.54.96.16.17.53.524.385.06
ForecastPro1.93.04.04.45.45.46.73.314.004.60
Theta1.82.73.84.55.65.26.13.203.934.41
RBF2.73.85.25.86.96.37.34.385.125.60
ForcX2.13.14.14.45.65.46.53.424.104.64
ETS2.03.04.04.45.45.16.33.373.994.51
Ata(p, 0)2.13.55.46.37.87.59.14.345.456.26
Ata(p, 1)1.92.94.14.86.05.77.13.464.264.87
Ata–comb3.04.55.16.35.86.88.33.624.424.94
Ata- lowest1.82.73.84.35.44.95.93.123.804.27
Ata-median1.92.84.24.85.95.36.23.434.154.62
Ata–best1.82.63.84.35.44.95.83.143.804.24

Appendix B. Additional Results: M4 Frequency-Specific Tables

Appendix B reports the detailed evaluation results for the M4 competition dataset. In the main body of the paper, the overall performance based on the OWA metric is emphasized. To maintain readability, the frequency-level error measures are provided here in Table A5 and Table A6. Specifically, Table A5 presents the sMAPE values, and Table A6 provides the MASE results across the six frequencies of the M4 dataset (yearly, quarterly, monthly, weekly, daily, and hourly). These tables offer a more detailed view of forecasting performance, complementing the main findings with transparent, frequency-specific evidence.
Table A5. M4 competition sMAPE results.
Table A5. M4 competition sMAPE results.
TeamMethod TypeYearlyQuarterlyMonthlyWeeklyDailyHourlyTotalRank
SmylH13.1769.67912.1267.8173.1709.32811.3741
Montero-Manso, et al.C (S & ML)13.5289.73312.6397.6253.09711.50611.7203
Pawlikowski, et al.C (S)13.9439.79612.7476.9192.4529.61111.8455
Jaganathan & PrakashC (S & ML)13.7129.80912.4876.8143.0379.93411.6952
Fiorucci & LouzadaC (S)13.6739.81612.7378.6272.98515.56311.8364
Petropoulos & SvetunkovC (S)13.6699.80012.8886.7262.99513.16711.8876
ShaubC (S)13.67910.37812.8397.8183.22213.46612.0209
Legaki & KoutsouriS13.36610.15513.0029.1483.04117.56711.9868
Doornik, et al.C(S)13.91010.00012.7806.7283.0538.91311.9247
Selamlar (252)S13.9309.96013.1318.2013.00313.11112.10812
Pedregal, et al.C (S)13.82110.09313.1518.9893.0269.76512.11413
Taylan (255)S13.93010.29212.9368.5403.09512.85112.09811
Spiliotis & AssimakopoulosS13.80410.12813.1428.9903.02717.75612.14815
RoubinchteinC (S)14.44510.17212.9118.4353.27012.87112.18317
IbrahimS13.67710.08913.3219.0893.07118.09312.19818
Yapar, et al. (009)S13.98110.01613.0477.5403.00412.85112.09310
Tartu M4 seminarC (S & ML)14.09611.10913.2908.5132.85213.85112.49623
WaheebC (S)14.78310.05912.7707.0762.99712.04712.14614
Darin & StellwagenS14.66310.15513.0586.5823.07711.68312.27919
Dantas & Cyrino OliveiraC (S)14.74610.25413.4628.8733.24516.94112.55325
The M4 Team (Theta)S14.59310.31113.0029.0933.05318.13212.30920
The M4 Team (Com)S14.84810.17513.4348.9442.98022.05312.55527
The M4 Team (Arima)S15.16810.43113.4438.6533.19312.04512.66129
The M4 Team (Damped)S15.19810.23713.4738.8663.06419.26512.66130
The M4 Team (ETS)S15.35610.29113.5258.7273.04617.30712.72531
Yilmaz (256)S13.93310.20713.0858.3043.02213.39912.14816
The M4 Team (Holt)S16.35410.90714.8129.7083.06629.24913.77543
Çetin (253)S16.52910.67113.4098.2133.05612.77113.01135
The M4 Team (SES)S16.39610.60013.6189.0123.04518.09413.08737
Ata–lowestS15.23110.14512.9088.3932.86413.45912.34121
Ata–medianS13.71210.17313.1468.3933.05712.03712.11514
Ata–bestC (S)13.4749.85412.6147.4152.96011.84611.7193 *
* The star indicates the re-ranked position (third for sMAPE metric) highlighted without altering the official M4 table.
Table A6. M4 competition MASE results.
Table A6. M4 competition MASE results.
TeamMethod TypeYearlyQuarterlyMonthlyWeeklyDailyHourlyTotalRank
SmylH2.9801.1180.8842.3563.4460.8931.5361
Montero-Manso, et al.C (S & ML)3.0601.1110.8932.1083.3440.8191.5513
Pawlikowski, et al.C (S)3.1301.1250.9052.1582.6420.8731.5472
Jaganathan & PrakashC (S & ML)3.1261.1350.8952.3503.2580.9761.5716
Fiorucci & LouzadaC (S)3.0461.1220.9072.3683.1941.2031.5544
Petropoulos & SvetunkovC (S)3.0821.1180.9132.1333.2291.4581.5655
ShaubC (S)3.0381.1980.9292.9473.4791.3721.5957
Legaki & KoutsouriS3.0091.1980.9662.6013.2542.5571.6018
Doornik, et al.C (S)3.2621.1630.9312.3023.2840.8011.62711
Selamlar (252)S3.1241.1550.9622.4993.2462.3661.6139
Pedregal, et al.C (S)3.1851.1640.9432.4883.2321.0491.61410
Taylan (255)S3.1171.2310.9622.5783.2772.2381.63113
Spiliotis & AssimakopoulosS3.1841.1780.9592.4883.2321.8081.62812
RoubinchteinC (S)3.2441.1590.9212.2903.6321.1291.63315
IbrahimS3.0751.1850.9772.5833.8942.3881.64416
Yapar, et al. (009)S3.1151.1661.0982.5783.2252.2381.67820
Tartu M4 seminarC (S & ML)3.0911.2501.0022.3753.0251.0581.63314
WaheebC (S)3.4001.1601.0292.1803.3210.8611.70627
Darin & StellwagenS3.4061.1680.9242.1074.1280.8561.69325
Dantas & Cyrino OliveiraC (S)3.2941.1700.9522.5343.4361.5981.65717
The M4 Team (Theta)S3.3821.2320.9702.6373.2622.4551.69626
The M4 Team (Com)S3.2801.1730.9662.4323.2034.5821.63818
The M4 Team (Arima)S3.4021.1650.9302.5563.4100.9431.66619
The M4 Team (Damped)S3.3791.1730.9722.4043.3262.9561.68323
The M4 Team (ETS)S3.4441.1610.9482.5273.2531.8241.68021
Yilmaz (256)S3.1241.2031.1822.5283.2393.6051.98544
The M4 Team (Holt)S3.5501.1981.0092.4203.2239.3561.77234
Çetin (253)S4.0001.3501.0092.6023.2992.1631.88640
The M4 Team (SES)S3.9811.3401.0192.6853.2812.3851.88539
Ata–lowestS3.4101.1820.9652.5033.2731.7651.68624
Ata–medianS3.0811.2050.9682.5613.3031.5881.61811
Ata–bestC (S)3.0041.1330.9112.2263.2641.2091.5513 *
* The star indicates the re-ranked position (third for MASE metric) highlighted without altering the official M4 table.

Appendix C. Additional Results: Ablation Study on M3 Dataset

This appendix provides additional evidence regarding the role of individual model components in the proposed hybrid framework. Table A7 reports the comparative sMAPE results of B-J automatic (ARIMA), Theta, Ata-lowest, Ata-median, and Ata-best for the M3 competition overall dataset. The aim of this ablation study is to isolate the contribution of each component and demonstrate how the proposed hybrid strategy (Ata-best) achieves improvements over its constituents. As shown in Table A7, while both Ata-lowest and Ata-median offer competitive performance, the hybrid combination consistently yields superior results, surpassing even the M3 winning method Theta (overall sMAPE = 13.00) with a lower overall sMAPE of 12.79 achieved by Ata-best.
Table A7. Ablation study on selected models for the M3 competition overall dataset.
Table A7. Ablation study on selected models for the M3 competition overall dataset.
Forecasting HorizonsAveragesRank
Method 1 2 3 4 5 6 8 12 15 18 1–4 1–6 1–8 1–12 1–15 1–18
B-J automatic (ARIMA)9.210.412.213.914.014.613.014.117.819.311.4212.3912.5212.7813.3313.995
Theta (Winner)8.49.611.312.513.213.912.013.216.218.210.4411.4711.6111.9412.4113.002
Ata–lowest8.69.611.512.813.514.312.413.916.918.410.6011.7011.9212.3212.8013.394
Ata–median8.79.411.312.412.813.512.213.716.517.910.4611.3611.6112.0312.5013.083
Ata–best8.59.211.112.412.413.212.013.316.417.810.3011.1311.3311.7112.2012.791

Appendix D. Additional Results: Ablation Study on M4 Dataset

In order to complement the findings on the M3 dataset, a similar ablation analysis was conducted on the M4 competition dataset using the OWA metric. Table A8 presents the comparative results of the M4 winning model (Smyl, OWA = 0.821), the benchmark The M4 Team (ARIMA) (OWA = 0.902), and the Ata variants. The results confirm that the proposed hybrid model (Ata-best, OWA = 0.837) ranks second overall, outperforming both the ARIMA benchmark and the other Ata configurations (Ata-lowest, OWA = 0.895; Atamedian, OWA = 0.869). This demonstrates that the hybrid strategy retains its robustness and scalability across different datasets and error measures, further validating the effectiveness of the framework.
Table A8. Ablation study on selected models for the M4 competition overall dataset (OWA metric).
Table A8. Ablation study on selected models for the M4 competition overall dataset (OWA metric).
MethodYearlyQuarterlyMonthlyWeeklyDailyHourlyTotalRank
Smyl (Winner)0.7780.8470.8360.8511.0460.4400.8211
The M4 Team (ARIMA)0.8920.8980.9030.9321.0440.5240.90223
Ata–lowest0.8950.8920.9010.9090.9700.7350.89519
Ata–median0.8070.9010.9110.9191.0060.6590.86911
Ata–best0.7900.8610.8660.8050.9840.5750.8372

References

  1. Makridakis, S.; Hibon, M. The M3-Competition: Results, conclusions and implications. Int. J. Forecast. 2000, 16, 451–476. [Google Scholar] [CrossRef]
  2. Makridakis, S.; Spiliotis, E.; Assimakopoulos, V. The M4 Competition: Results, findings, conclusion and way forward. Int. J. Forecast. 2018, 34, 802–808. [Google Scholar] [CrossRef]
  3. Makridakis, S.; Spiliotis, E.; Assimakopoulos, V. The M4 Competition: 100,000 time series and 61 forecasting methods. Int. J. Forecast. 2020, 36, 54–74. [Google Scholar] [CrossRef]
  4. Zhang, G.P. Time series forecasting using a hybrid ARIMA and neural network model. Neurocomputing 2003, 50, 159–175. [Google Scholar] [CrossRef]
  5. Atesongun, A.; Gulsen, M. A hybrid forecasting structure based on arima and artificial neural network models. Appl. Sci. 2024, 14, 7122. [Google Scholar] [CrossRef]
  6. Tsoku, J.T.; Metsileng, D.; Botlhoko, T. A Hybrid of Box-Jenkins ARIMA Model and Neural Networks for Forecasting South African Crude Oil Prices. Int. J. Financial Stud. 2024, 12, 118. [Google Scholar] [CrossRef]
  7. Alshawarbeh, E.; Abdulrahman, A.T.; Hussam, E. Statistical modeling of high frequency datasets using the ARIMA-ANN hybrid. Mathematics 2023, 11, 4594. [Google Scholar] [CrossRef]
  8. Smyl, S. A hybrid method of exponential smoothing and recurrent neural networks for time series forecasting. Int. J. Forecast. 2020, 36, 75–85. [Google Scholar] [CrossRef]
  9. Iftikhar, H.; Khan, M.; Z˙ ywiołek, J.; Khan, M.; López-Gonzales, J.L. Modeling and forecasting carbon dioxide emission in Pakistan using a hybrid combination of regression and time series models. Heliyon 2024, 10, e33148. [Google Scholar] [CrossRef]
  10. Sherly, A.; Christo, M.S.; Elizabeth, J.V. A Hybrid Approach to Time Series Forecasting: Integrating ARIMA and Prophet for Improved Accuracy. Results Eng. 2025, 27, 105703. [Google Scholar] [CrossRef]
  11. Ampountolas, A. Comparative Analysis of Machine Learning, Hybrid, and Deep Learning Forecasting Models: Evidence from European Financial Markets and Bitcoins. Forecasting 2023, 5, 472–486. [Google Scholar] [CrossRef]
  12. Qin, P.; Ye, B.; Li, Y.; Cai, Z.; Gao, Z.; Qi, H.; Ding, Y. Hybrid BiLSTM-ARIMA Architecture with Whale-Driven Optimization for Financial Time Series Forecasting. Algorithms 2025, 18, 517. [Google Scholar] [CrossRef]
  13. Çınarer, G. Hybrid Deep Learning and Stacking Ensemble Model for Time Series-Based Global Temperature Forecasting. Electronics 2025, 14, 3213. [Google Scholar] [CrossRef]
  14. Dong, Z.; Zhou, Y. A Novel Hybrid Model for Financial Forecasting Based on CEEMDAN-SE and ARIMA-CNN-LSTM. Mathematics 2024, 12, 2434. [Google Scholar] [CrossRef]
  15. Iacobescu, P.; Susnea, I. Hybrid ARIMA-ANN for Crime Risk Forecasting: Enhancing Interpretability and Predictive Accuracy Through Socioeconomic and Environmental Indicators. Algorithms 2025, 18, 470. [Google Scholar] [CrossRef]
  16. Liagkouras, K.; Metaxiotis, K. A Hybrid Long Short-Term Memory with a Sentiment Analysis System for Stock Market Forecasting. Electronics 2025, 14, 2753. [Google Scholar] [CrossRef]
  17. Liu, Z.; Zhang, Z.; Zhang, W. A Hybrid Framework Integrating Traditional Models and Deep Learning for Multi-Scale Time Series Forecasting. Entropy 2025, 27, 695. [Google Scholar] [CrossRef]
  18. Karim, A.A.; Pardede, E.; Mann, S. A model selection approach for time series forecasting: Incorporating google trends data in Australian macro indicators. Entropy 2023, 25, 1144. [Google Scholar] [CrossRef]
  19. Hananya, R.; Katz, G. Dynamic selection of machine learning models for time-series data. Inf. Sci. 2024, 665, 120360. [Google Scholar] [CrossRef]
  20. Wei, W.; Yang, T.; Chen, H.; Rossi, R.A.; Zhao, Y.; Dernoncourt, F.; Eldardiry, H. Efficient Model Selection for Time Series Forecasting via LLMs. arXiv 2025, arXiv:2504.02119. [Google Scholar] [CrossRef]
  21. Yapar, G.; Capar, S.; Selamlar, H.T.; Yavuz, I. Modified Holt’s linear trend method. Hacet. J. Math. Stat. 2018, 47, 1394–1403. [Google Scholar]
  22. Yapar, G.; Taylan Selamlar, H.; Capar, S.; Yavuz, ˙I. ATA Method. Hacet. J. Math. Stat. 2019, 48, 1838–1844. [Google Scholar] [CrossRef]
  23. Çapar, S.; Selamlar, H.T.; Yavuz, ˙I.; Taylan, A.S.; Yapar, G. Ata method’s performance in the M4 competition. Hacet. J. Math. Stat. 2023, 52, 268–276. [Google Scholar]
  24. Box, G.E.P.; Jenkins, G.M. Time Series Analysis: Forecasting and Control; John Wiley and Sons: New York, NY, USA, 1970. [Google Scholar]
Figure 1. Distribution of time series in the M3 dataset across frequencies (a) and domains (b). Pie slices show category counts; legend shows full category names and counts.
Figure 1. Distribution of time series in the M3 dataset across frequencies (a) and domains (b). Pie slices show category counts; legend shows full category names and counts.
Applsci 15 09552 g001
Figure 2. Composition of time series in the M4 dataset across frequencies (a) and domains (b). Pie slices show category initials; legend shows full category names and counts.
Figure 2. Composition of time series in the M4 dataset across frequencies (a) and domains (b). Pie slices show category initials; legend shows full category names and counts.
Applsci 15 09552 g002
Figure 3. Hybrid framework flow diagram for M3 & M4 forecasting using Ata Models and ARIMA based on model selection process.
Figure 3. Hybrid framework flow diagram for M3 & M4 forecasting using Ata Models and ARIMA based on model selection process.
Applsci 15 09552 g003
Table 1. Comparison of ARIMA and Ata method in time series forecasting.
Table 1. Comparison of ARIMA and Ata method in time series forecasting.
AspectARIMAAta Method
Modeling approachLinear stochastic model combining autoregressive (AR), differencing (I), and moving average (MA) components.Modern, innovative smoothing-based approach with adaptive updating of level, trend and any other components.
Stationarity requirementRequires (covariance) stationarity; differencing used to remove trends/seasonality; SARIMA (seasonal ARIMA) handles a single regular seasonality.No strict stationarity requirement; directly adapts to evolving level/trend, and can accommodate multiplicative effects via functional form.
ComplexityModel identification requires selecting ( p , d , q ) with diagnostics; risk of over/underfitting in short or noisy series.Relatively simple updating structure; few tuning choices; parameters are set adaptively.
InterpretabilityParameters have clear statistical meaning; transparent for low orders, but interpretation can be harder for higher-order models.Transparent level/trend decomposition and update rules.
AdaptabilityLimited adaptability to nonlinear dynamics, structural breaks, regime changes, or multiple/unstable seasonalities without extensions.Highly adaptive to shifts in level/trend or any other parameters; effective under unstable seasonality or intermittent patterns via adaptive smoothing.
Forecasting accuracyStrong baseline on linear, stationary series; accuracy degrades with structural breaks or nonlinearities.Demonstrated competitive accuracy (e.g., M3, M4, and M6 forecasting competitions), especially for short, irregular, or heterogeneous series.
Computational efficiencyModerate; iterative estimation and diagnostic checking raise cost for large-scale model selection.High; lightweight computations enable scaling to thousands of series.
Application areasWell-suited for stationary and short-memory time series; extensively used in economics, finance, and engineering.Versatile and broadly applicable across diverse domains (e.g., economics, healthcare, environmental forecasting, and intermittent demand); also highly adaptable for hybrid and ensemble frameworks.
LimitationsStruggles with nonlinearities, structural breaks, multiple or unstable seasonalities; endogenous-only unless extended.Relatively new but increasingly recognized; performance benefits from adaptive form and flexible selection procedure.
Table 2. Optimal hybrid weights by frequency for M3 dataset.
Table 2. Optimal hybrid weights by frequency for M3 dataset.
Frequency/ModelARIMAAta-LowestAta-Median
Yearly0.050.150.80
Quarterly0.150.350.50
Monthly0.300.400.30
Other0.000.800.20
Table shows optimized weights for each model by frequency type using grid search algorithms.
Table 3. M3 competition overall dataset sMAPE results.
Table 3. M3 competition overall dataset sMAPE results.
Forecasting HorizonsAverages
Method 1 2 3 4 5 6 8 12 15 18 1–4 1–6 1–8 1–12 1–15 1–18
Naive210.511.313.615.115.115.914.516.019.320.712.6213.5713.7614.2414.8115.47
Single9.510.612.714.114.315.013.314.518.319.411.7312.7112.8413.1313.6714.32
Holt9.010.212.613.714.515.813.014.818.820.211.6712.9313.1113.4213.9514.60
B-J automatic9.210.412.213.914.014.613.014.117.819.311.4212.3912.5212.7813.3313.99
ForecastPro8.69.611.412.913.314.212.613.216.418.310.6411.6711.8412.1212.5813.18
Theta8.49.611.312.513.213.912.013.216.218.210.4411.4711.6111.9412.4113.00
RBF9.910.512.413.413.214.112.814.117.317.811.5612.2612.4012.7613.2413.74
ForcX8.79.811.613.113.213.812.613.917.818.710.8211.7211.8812.2112.8013.48
ETS8.89.812.013.513.914.713.014.117.618.911.0412.1312.3212.6613.1413.77
Ata(p, 0)8.910.012.113.713.914.712.813.917.318.911.1612.2112.3412.6413.1313.77
Ata(p, 1)8.49.711.512.913.614.212.915.418.920.910.6411.7211.9412.6613.3214.09
Ata–comb8.59.611.412.813.013.612.013.116.317.410.5611.4711.5811.9412.4012.94
Ata–lowest8.69.611.512.813.514.312.413.916.918.410.6011.7011.9212.3212.8013.39
Ata–median8.79.411.312.412.813.512.213.716.517.910.4611.3611.6112.0312.5013.08
Ata–best8.59.211.112.412.413.212.013.316.417.810.3011.1311.3311.7112.2012.79
Table 4. Optimal hybrid weights by frequency for M4 competition.
Table 4. Optimal hybrid weights by frequency for M4 competition.
Frequency/ModelARIMAAta-LowestAta-Median
Yearly0.150.150.70
Quarterly0.300.400.30
Monthly0.400.350.25
Weekly0.500.000.50
Daily0.500.250.25
Hourly0.400.100.50
Table 5. M4 competition OWA results.
Table 5. M4 competition OWA results.
TeamMethod TypeYearlyQuarterlyMonthlyWeeklyDailyHourlyTotalRank
SmylH0.7780.8470.8360.8511.0460.4400.8211
Montero-Manso, et al.C (S & ML)0.7990.8470.8580.7961.0190.4840.8382
Pawlikowski, et al.C (S)0.8200.8550.8670.7660.8060.4440.8413
Jaganathan & PrakashC (S & ML)0.8130.8590.8540.7950.9960.4740.8424
Fiorucci & LouzadaC (S)0.8020.8550.8680.8970.9770.6740.8435
Petropoulos & SvetunkovC (S)0.8060.8530.8760.7510.9840.6630.8486
ShaubC (S)0.8010.9080.8820.9571.0600.6530.8607
Legaki & KoutsouriS0.7880.8980.9050.9680.9961.0120.8618
Doornik, et al.C (S)0.8360.8780.8810.7821.0020.4100.8659
Selamlar (252)S0.8190.8730.9080.8980.9880.8510.86810
Pedregal, et al.C (S)0.8240.8830.8990.9390.9900.4850.86911
Taylan (255)S0.8180.9160.9010.9301.0080.8170.87212
Spiliotis & AssimakopoulosS0.8230.8890.9070.9390.9900.8600.87413
RoubinchteinC(S)0.8500.8850.8810.8731.0910.5860.87614
IbrahimS0.8050.8900.9210.9611.0980.9910.88015
Yapar, et al. (009)S0.8200.8800.9690.9300.9850.8170.88516
Tartu M4 seminarC (S & ML)0.8200.9600.9320.8920.9300.5980.88817
WaheebC (S)0.8800.8800.9270.7790.9990.5070.89418
Darin & StellwagenS0.8770.8870.8870.7391.1350.4960.89519
Dantas & Cyrino OliveiraC (S)0.8660.8920.9140.9411.0570.7940.89620
The M4 Team (Theta)S0.8720.9170.9070.9710.9991.0060.89721
The M4 Team (Com)S0.8670.8900.9200.9260.9781.5560.89822
The M4 Team (Arima)S0.8920.8980.9030.9321.0440.5240.90223
The M4 Team (Damped)S0.8900.8930.9240.9170.9971.1410.90725
The M4 Team (ETS)S0.9030.8910.9150.9310.9960.8520.90826
Yilmaz (256)S0.8190.9021.0090.9080.99013.6850.96736
The M4 Team (Holt)S0.9470.9320.9880.9660.9952.7490.97137
Çetin (253)S1.0090.9770.9390.9171.0050.7990.97338
The M4 Team (SES)S1.0030.9700.9510.9751.0000.9900.97539
Ata–lowestS0.8950.8920.9010.9090.9700.7350.89519
Ata–medianS0.8070.9010.9110.9191.0060.6590.86911
Ata–bestC (S)0.7900.8610.8660.8050.9840.5750.8372 *
* The star indicates the re-ranked position (second overall) highlighted without altering the official M4 table.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ekiz Yılmaz, T.; Yapar, G. A Robust Hybrid Forecasting Framework for the M3 and M4 Competitions: Combining ARIMA and Ata Models with Performance-Based Model Selection. Appl. Sci. 2025, 15, 9552. https://doi.org/10.3390/app15179552

AMA Style

Ekiz Yılmaz T, Yapar G. A Robust Hybrid Forecasting Framework for the M3 and M4 Competitions: Combining ARIMA and Ata Models with Performance-Based Model Selection. Applied Sciences. 2025; 15(17):9552. https://doi.org/10.3390/app15179552

Chicago/Turabian Style

Ekiz Yılmaz, Tuğçe, and Güçkan Yapar. 2025. "A Robust Hybrid Forecasting Framework for the M3 and M4 Competitions: Combining ARIMA and Ata Models with Performance-Based Model Selection" Applied Sciences 15, no. 17: 9552. https://doi.org/10.3390/app15179552

APA Style

Ekiz Yılmaz, T., & Yapar, G. (2025). A Robust Hybrid Forecasting Framework for the M3 and M4 Competitions: Combining ARIMA and Ata Models with Performance-Based Model Selection. Applied Sciences, 15(17), 9552. https://doi.org/10.3390/app15179552

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop