Previous Article in Journal
Analysis of Economic and Growth Synchronization Between China and the USA Using a Markov-Switching–VAR Model: A Trend and Cycle Approach
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Proceeding Paper

Multivariate Forecasting Evaluation: Nixtla-TimeGPT †

by
S M Ahasanul Karim
1,*,‡,§,
Bahram Zarrin
2,§ and
Niels Buus Lassen
3,*
1
Microsoft Research Hub, Microsoft Development Center Copenhagen, 2800 Copenhagen, Denmark
2
Microsoft Dynamics 365 AI ERP, Microsoft Development Center Copenhagen, 2800 Copenhagen, Denmark
3
Department of Digitalization, Copenhagen Business School, 2000 Copenhagen, Denmark
*
Authors to whom correspondence should be addressed.
Presented at the 11th International Conference on Time Series and Forecasting, Canaria, Spain, 16–18 July 2025.
Current address: Department of Digitalisation, Copenhagen Business School, 2000 Frederiksberg, Denmark.
§
These authors contributed equally to this work.
Comput. Sci. Math. Forum 2025, 11(1), 29; https://doi.org/10.3390/cmsf2025011029
Published: 26 August 2025

Abstract

Generative models are being used in all domains. While primarily built for processing texts and images, their reach has been further extended towards data-driven forecasting. Whereas there are many statistical, machine learning and deep learning models for predictive forecasting, generative models are special because they do not need to be trained beforehand, saving time and computational power. Also, multivariate forecasting with the existing models is difficult when the future horizons are unknown for the regressors because they add mode uncertainties in the forecasting process. Thus, this study experiments with TimeGPT(Zeroshot) by Nixtla where it tries to identify if the generative model can outperform other models like ARIMA, Prophet, NeuralProphet, Linear Regression, XGBoost, Random Forest, LSTM, and RNN. To determine this, the research created synthetic datasets and synthetic signals to assess the individual model performances and regressor performances for 12 models. The results then used the findings to assess the performance of TimeGPT in comparison to the best fitting models in different real-world scenarios. The results showed that TimeGPT outperforms multivariate forecasting for weekly granularities by automatically selecting important regressors whereas its performance for daily and monthly granularities is still weak.

1. Introduction

Generative AI has taken the world by storm. It has shown its ability to undertake a wide variety of tasks like question answering, rational thinking, summarising, sentiment analysis, image analysis, and so forth. Whereas the majority of users are using these for texts and images, recent developments have further extended their use towards forecasting. Models like TTM by IBM [1], Lag-llama by Facebook [2], TimesFM by Google [3], and TimeGPT by Nixtla [4], which are pre-trained with many time series, can forecast within a very short runtime and computation overhead while also achieving great accuracy. However, these models are often benchmarked using univariate forecasts rather than multivariate forecasts. Multivariate forecasting is important because it can take external events into account and provide great context for predicting shocks or structural breaks. But it is a challenge to predict with better accuracy with a multivariate approach, especially when the future horizon is unknown for the exogenous regressor because it adds more uncertainty into the equation, and the errors of the predicted horizon for the regressors often drive more error for predicting the target. Generative AI can be a promising development for dealing with this issue.
Ekambaram et al. analysed exogenous variables with generative models like TTM, Moment, GPT4TS, and Time-LLM in a wide variety of datasets, but TimeGPT and other state-of-the-art models were not considered [1]. At this date, this is possibly the only available research where performance assessment has been carried out with generative models for multivariate forecasting. For this paper, we have chosen to work with the generative model TimeGPT by Nixtla (nixtlats library version “0.5.2” on python) because it has a proper encoder–decoder structure and it is pre-trained with the largest collection of time series ranging up to 100 billion data points from multiple sources like finance, economics, demographics, healthcare, weather, IoT, sensor data, energy, web traffic, sales, transport, and banking, which makes it most suitable among other generative models for forecasting. Garza et al. showed that univariate TimeGPT can surpass the accuracy of models like ZeroModel, HistoricAverage, SeasonalNaive, Theta, DOTheta, ETS, CES, ADIDA, IMAPA, CrostonClassic, LGBM, LSTM, DeepAR, and TFT [4].
However, it is questionable if TimeGPT can outperform the best-fitting non-generative model with the right external variables and make robust predictions based on different scenarios. Thus, this paper attempts to determine if a generative AI model like TimeGPT(Zeroshot) can leverage external events better to forecast results for daily, weekly, and monthly granularity in terms of mean absolute error in comparison toother models like ARIMA, Seasonal ARIMA, Prophet, Linear Regression, XGBoost, LightGBM, Random Forest, LSTM, RNN, and Neural Prophet using different Python libraries (pmdarima “1.8.3”, statsforecast “1.7.5”, prophet “1.1.5”, mlforecast “0.13.2”, pycaret “3.3.1”, nixtlats “0.5.2”, neuralprophet “0.9.0” and neuralforecast “0.1.0”). The mean absolute error metric is chosen because it delivers a correct estimation of the average error while forecasting multiple time series altogether. It is always positive; hence, after averaging, any negative errors do not cancel positive errors. Also, they do not explode in magnitude for outliers or simply for a bad fit. It also gives an exact idea of how much we are underestimating or overestimating while forecasting. The best models were tested in real-world demand forecasting scenarios for assessment and comparison of collective forecasting accuracies and statistical significance.

2. Materials and Methods

First, we will create simulated scenarios and events that we control to discover model performances for different patterns and interpret useful insights from them with controlled experiments. We will find the best-fitting model for each scenario and the best type of regressors to compare TimeGPT’s(nixtlats “0.5.2”) performance with them. For the subsequent case studies, we will collect the data, preprocess it, and evaluate our synthetic findings by multivariate forecasting. The overall procedure of this research is portrayed in Figure 1.

2.1. Dataset Creation and Evaluation

In the simulated dataset creation stage, we will create a synthetic dataset of time series with diverse characteristics and resample it in daily, weekly, and monthly formats. We will also create simulated external variables for testing. Then, we will move to the model fitting stage and fit the models from the mentioned models with different libraries by splitting the synthetic data into training and test sets. We will test the models both with and without external variables to understand the effects of the external variables and assess the impact of each external signal by excluding it from the training set. Then, in the performance evaluation stage, we will evaluate each model’s performance and external variable effects by MAE. Then, we will use the results from the simulated experiments to find out the best performing model and the best external signals to be used to make forecasts for real-world use cases and compare it with a generative AI model to assess their competitive performance.
To simulate a set of time series with different characteristics, we designed a set of 216 time series with a combination of stable, upwards, downwards trends, the presence or absence of seasonality, cycles and noises (non-linearity), additive or multiplicative modes, and from 0 up to 5 structural breaks in random time points using the following equations:
(i) Trend Components:
T ( t ) = 1 + t N 1 for t [ 0 , N 1 ]
T ( t ) = 2 t N 1 for t [ 0 , N 1 ]
T ( t ) = 1 for all t
(ii) Seasonal Components:
S ( t ) = 0.5 cos 2 π t / ( 365 7 )
(iii) Cycle Components:
C ( t ) = 0.3 sin 2 π t / ( 365 5 )
(iv) Noise Components:
N ( t ) = Noise , N ( t ) N ( 0 , 20 )
(v) Modes:
a d d i t i v e : Y ( t ) = T ( t ) + S ( t ) + C ( t ) + N ( t )
m u l t i p l i c a t i v e : Y ( t ) = T ( t ) ( 2 + S ( t ) ) ( 2 + C ( t ) ) + ( 1 + N ( t ) / 100 )
(vi) Structural Breaks: 0 to 5 breaks in random time points.
a d d i t i v e : Y ( t ) = Y ( t ) + constant for t t break
m u l t i p l i c a t i v e : Y ( t ) = Y ( t ) constant for t t break
Running all these attributes in a loop, we generated a 5-year time series dataset with different trends, cycles, seasonality, noises, modes, and breaks, resulting in 216 time series with diverse characteristics. Each time series was labelled with a unique identifier. The dataset was stored in daily, monthly, and weekly granularity for further characteristic testing of model performances. Adding positive constants in the break points pushed some downward trends upwards and created seasonalities of their own in some situations. Also, changing granularities has completely changed structural aspects for some time series. Therefore, the metadata containing details about individual series characteristics was manually adjusted later. The dataset was then forecasted with all the models by an 80:20 training/test split.

2.2. Simulated External Regressor Creation

For the dataset created for assessment, we considered 6 external regressors to understand their effect on prediction accuracy for the remaining models. The first two signals are static and inspired by the work of [5], we designed exogenous features for the NBEATSx model, which means that these signals remain the same throughout each of the 216 time series. The 3rd, 4th, and 5th signals are dynamic and series-specific, exogenous, and calculated from the target variable and its lagged values to mimic events that have a delayed effect. The last signal is random noise-generated. Below are the equations described for the external variables and their correlations.
trend ( t ) = 200 × a 0 + a 1 t 1 + a 2 t 2 + + a d t d
seasonal ( t ) = k = 1 K 2000 × sin 2 π k t P + 2000 × cos 2 π k t P
exogenous ( y ) = α × y + | ϵ |
exogenous lag 3 ( t ) = exogenous ( t + 3 )
exogenous lag 5 ( t ) = exogenous ( t + 5 )
noise ( t ) = 1 N i = 1 N ϵ i ( t )
These variables were used to assess the competitive performance of TimeGPT with respect to the other models. The average correlation score of the external variables with the target is portrayed in Table 1. The trend, regardless of being static, shows great correlation towards the dependent variable. This can be because most series pushed upwards due to rapid structural breaks. Also, the mathematically modeled y dependent exogenous feature still showed a lower correlation due to the added noise term ϵ .

2.3. Case Studies

Case Study 1: This case involves forecasting the UK’s monthly electricity demand in each 30 min interval over the span of 35 months. This dataset is gathered by National Grid ESO, an electricity system operator for Great Britain, which is updated twice per hour, 48 times each day [6]. It contains daily data over 14 years which were aggregated for each 30 min time interval into 48 time series. All the electricity variables are estimated in Megawatts (MW). In the dataset, most of the features are related to supply more than demand, and some are dependent on demand rather than demand depending on them. Thus, for external features, items such as the following were further added from other external sources: disaster-affected days, average temperature (tavg), precipitation (prcp), hours of sunshine (tsun), electricity on gdp, oil prices, CPI, and weather. The data was converted into a monthly granularity to determine the monthly electricity load. After the training/test split, the forecast horizon was 35 months. The patterns of the main series were gradually decreasing over time, with regular seasonal fluctuations that multiplicatively shrunk the demand over time as shown in Figure 2.
Case Study 2: In this case, we will forecast docked bike demand for each hour at each station weekly. Divvy is a popular bike-sharing service in Chicago. It allows users to dock and rent bikes from biking stations across the city. The available dataset in Kaggle contains several datasets, but among those only “Divvy_Trips_1719.csv” and “Chicago_Weather_2017_2022[1].csv” were found suitable, representing trip details and weather details, respectively. There was another dataset of bus and rail boarding called “CTA Bus Rail Daily Totals.csv” but it was found to be missing years of data; therefore, it was omitted and added from another external source [7]. Weather features such as temperature and dew were then aggregated by weekly means. Other external variables like bus rail passengers, disaster-affected days, holidays and working days, and oil prices, were further added from external sources. Among the 14,614 unique_ids, the unique_ids that had 156 instances of data were segmented as a smooth forecasting sample, which left us with a total of 4447 time series. After the training/test split, the forecast horizon was 32 weeks, and they collectively showed a stable trend, and seasonal characteristics as shown in Figure 2.
Case Study 3: In this case, we will forecast demand for each perishable good at each store daily. The dataset is from Brick-and-Mortar grocery stores in Ecuador, which is a perfect dataset for daily demand planning, as a good number of goods are perishable and over-forecasting or under-forecasting too much is an issue for thousands of stores across many locations [8]. Here, store ID and item ID were joined to make a unique ID, which leaves us with a combination of 174,685 stores and products. From the item information, we segment out the perishable items which are crucial for daily demand forecasting and contain 4.5 years of regular daily data, imputing the missing values with zero for unit sales. This leaves us with 1654 unique_ids to be forecasted. External events like local, regional, and national holidays, workdays, payday, oil price, and promotion were further added. After the training/test split, the forecast horizon was 365 days, the common pattern being stable trend with rapid breaks, sudden spikes and drops as shown in Figure 2.
We will later find out the best-performing models in these different kinds of scenarios and granularities with our synthetic dataset and fit them to these cases to compare their multivariate performance with TimeGPT.

3. Results

The results derived from fitting all the models without any exogenous regressors for the synthetic dataset for daily, monthly, and weekly granularities are given in Table 2, where the top three models are in bold with the best model in asterics and the worst model is underlined. Libraries like Nixtla and PyCaret are used for further comparative forecast performance assessment, repeating some models adopting different hyperparameter tuning and grid search approaches. For the machine learning models, feature engineering from [9] was implemented, as the study claimed that it had outperformed the statistical models in demand forecasting scenarios. For comparison among various models, the different MAEs were scaled between 0 and 1. If e max is the maximum amount of error among a set of errors { e 1 , e 2 , , e n } for different models for a series, then the scaling factor S i can be expressed as
S i = e i e max , Where e i = { e 1 , e 2 , , e n } .
The results show that, despite being computationally very expensive, AutoLSTM and AutoRNN with 20 were not performing accordingly in collective forecasting scenarios; therefore, they were omitted from further testing. From the experiments on the synthetic dataset, we discovered that almost all the models generally tend to forecast better for weekly granularities. This indicates the number of data points over-representing or under-representing the characteristics, and the forecast horizons being too long or too short for daily and monthly granularities, respectively. This could also be due to the fact that aggregating the data by their weekly mean smoothens and handles the outliers to such an extent that it is neither too abstract nor too over-detailed, thus easier for the algorithms to grasp. We also learned that a machine learning algorithm, such as linear regression is a badly performing model in these kind of forecasting scenarios in both Nixtla’s mlforecast and PyCaret, because of their sensitivity to outliers and multicollinearity, whereas statistical models like ARIMA, Prophet, and TimeGPT tend to perform well overall in all scenarios, and Prophet performs better when structural breaks are present from Figure 3.
The results in Table 3 show that some models are performing better with all the external variables included, while some models are not. TimeGPT, except for the daily scenario, is performing better in monthly and weekly cases. We can also see that, in one case, linear regression exceeded the error value for maximum float64 capacity, and linear regression and Neural Prophet models generally performed badly; therefore, they were further omitted from the study. ARIMA models from Nixtla were also omitted because they were found to be forecasting negative demand.
Removing each external feature added the following positive and negative importance to that variables as shown in Table 4 and Figure 4. The external variables were further evaluated by keeping this all-feature-inclusive MAE as baseline and removing each external feature from the model in turns, where feature importance was calculated as (MAE without a distinct feature—MAE with all features included) so that the positive values indicate the positive impact.
When we tried the same variables with TimeGPT, and tried to calculate the importance for each external variable, we found all feature importances to be 0. Interestingly, it did not show any distinct feature importances, and all the scores were 0 or almost 0 until 7 decimal places for all these experiments. This could either be because of the LLM hallucination or TimeGPT inherently only choosing automatically the feature that is best for the data by automated interpretations. Table 4 further backs up the argument by showing that TimeGPT has performed better by adding all external features used together, whereas for other models, it has been somewhat the opposite, except for the daily granularity. It has a surge in error for daily granularities, but it is to be noted that since linear regression was not present this time with external variables, the averaging metrics were manipulated. This can also be a reason for TimeGPT not having enough daily data in its training corpus. The results mimic those from [4]. However, TimeGPT can also be fine-tuned further, but it is not free and is expensive in usage credit units.
Considering the results and model performances in different scenarios, we chose to fit the best state-of-the-art models and regressors with the three cases to compare them with TimeGPT. Table 5 shows that all the models performed worse than their univariate results when we picked every correlated external event with their best chosen models for multivariate setting, keeping the future values of the extra regressors unknown. We decided to fit TimeGPT with all the external features, as from our synthetic experiment, it seemingly autodetects or ignores the relevant external effects for forecasting. We found TimeGPT performing better than the best state-of-the-art model in the weekly docked bike demand forecasting case, just like it did on Table 4.

4. Discussion

With the results, we can conclude that whether external features would add more accuracy to a model or not is a matter of relativity, and there are no two similar ways. It is more difficult for scenarios where thousands of time series are involved, and the exogenous effects are not uniform among all of them. Also, forecasting exogenous features for the target horizon can make the future values of those regressors more error-prone, which is a threat to the main target forecast. When using multiple regressors as exogenous, the features can have plenty of different kinds of intercorrelations with each other apart from the target variable, which also poses a challenge of overfitting. A generative pre-trained model like TimeGPT can be a solution in this case, but the performance of TimeGPT is still to be improved to surpass the best-fitting model in all aspects.
This paper assumes that TimeGPT is trained with more weekly time series than daily and monthly ones, which can explain its specific efficiency in weekly granularities for multivariate forecasting. However, further tests need to be performed to explain the reasons behind its peak performance. Future works can also focus on evaluating error metrics other than MAE with more robust analysis and analysing more complex correlations.

5. Conclusions

With the case studies, we can conclude that multivariate forecasting is a challenge. It involves many barriers in finding and fitting the best model and incorporating the right exogenous variable. This research attempted to explore these challenges and tested whether they can be solved by using generative AI models. The main research interest was to find the best models and external variables based on different scenarios to see if TimeGPT could surpass their forecasting accuracy. TimeGPT could overcome that challenge for weekly granularities, but it could not perform better for the other cases. This research further suggests exploring results from fine-tuned TimeGPT instead of zero-shot approaches and explaining why TimeGPT performs as it does. More in-depth research should also be carried out on what makes TimeGPT perform better, and more models further need to be compared, like Autotheta, AutoCES, N-Beats, N-Hits, XLSTM, LLama-3, TTM, and TimesFM. Also, in this research, forecasting performance was only evaluated in MAE; thus, other error metrics can also bring about different results. Further research in this area with generative models will lead us to explore more futuristic forecasting.

Author Contributions

Conceptualization, S.M.A.K.; Methodology, S.M.A.K.; Software, S.M.A.K.; Validation, S.M.A.K., B.Z., and N.B.L.; Formal Analysis, S.M.A.K., B.Z., and N.B.L.; Investigation, S.M.A.K., B.Z., and N.B.L.; Resources, S.M.A.K. and B.Z. Data Curation, S.M.A.K.; Writing—Original Draft Preparation, S.M.A.K.; Writing—Review and Editing, N.B.L. and S.M.A.K.; Visualization, S.M.A.K.; Supervision, B.Z. and N.B.L.; Project Administration, S.M.A.K.; Funding Acquisition, B.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research received necessary computational resources and funding support from Microsoft Development Center Copenhagen and the Department of Digitalisation, Copenhagen Business School.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The publicly available Data for UK electricity consumption can be found at: https://www.kaggle.com/datasets/albertovidalrod/electricity-consumption-uk-20092022 (accessed on 1 July 2024). The bike sharing data can be found at: https://www.kaggle.com/datasets/leonidasliao/divvy-station-dock-capacity-time-series-forecast (accessed on 1 July 2024). The additional public transport data (bus, rail onboarding) can be found at: https://www.transitchicago.com/ridership/ (accessed on 1 July 2024). The Corporación Favorita grocery sales forecasting data can be found at: https://www.kaggle.com/competitions/favorita-grocery-sales-forecasting/overview (accessed on 1 July 2024).

Acknowledgments

A huge thanks to the Microsoft Demand Planning Team for their incredible support and collaboration throughout this research project. Their insights, expertise, and willingness to share knowledge made a huge difference in shaping our work. We truly appreciate the time and effort they put into discussions, feedback, and guidance—it has been an amazing learning experience. We are also grateful to the broader Microsoft team for fostering an environment that encourages curiosity, innovation, and problem-solving. This project would not have been the same without their support.

Conflicts of Interest

Author Bahram Zarrin was employed by the company Microsoft. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest. The authors declare that this study received funding and resource support from Microsoft Development Center Copenhagen and Department of Digitalization, Copenhagen Business School. The funders were not involved in the study design, collection, analysis, interpretation of data, the writing of this article or the decision to submit it for publication.

References

  1. Ekambaram, V.; Jati, A.; Dayama, P.; Mukherjee, S.; Nguyen, N.H.; Gifford, W.M.; Reddy, C.; Kalagnanam, J. Tiny Time Mixers (TTMs): Fast Pre-trained Models for Enhanced Zero/Few-Shot Forecasting of Multivariate Time Series. arXiv 2024, arXiv:2401.03955. [Google Scholar]
  2. Rasul, K.; Ashok, A.; Williams, A.R.; Ghonia, H.; Bhagwatkar, R.; Khorasani, A.; Bayazi, M.J.D.; Adamopoulos, G.; Riachi, R.; Hassen, N.; et al. Lag-Llama: Towards Foundation Models for Probabilistic Time Series Forecasting. arXiv 2024, arXiv:2310.08278. [Google Scholar] [CrossRef]
  3. Das, A.; Kong, W.; Sen, R.; Zhou, Y. A decoder-only foundation model for time-series forecasting. arXiv 2024, arXiv:2310.10688. [Google Scholar]
  4. Garza, A.; Challu, C.; Mergenthaler-Canseco, M. TimeGPT-1. arXiv 2024, arXiv:2310.03589. [Google Scholar]
  5. Olivares, K.G.; Challu, C.; Marcjasz, G.; Weron, R.; Dubrawski, A. Neural basis expansion analysis with exogenous variables: Forecasting electricity prices with NBEATSx. Int. J. Forecast. 2023, 39, 884–900. [Google Scholar] [CrossRef]
  6. Kaggle. Electricity Consumption UK 2009–2024. 2024. Available online: https://www.kaggle.com/datasets/albertovidalrod/electricity-consumption-uk-20092022 (accessed on 1 July 2024).
  7. Kaggle. Divvy Station Dock Capacity Time Series Forecast. 2022. Available online: https://www.kaggle.com/datasets/leonidasliao/divvy-station-dock-capacity-time-series-forecast (accessed on 1 July 2024).
  8. Kaggle. Corporación Favorita Grocery Sales Forecasting. 2018. Available online: https://www.kaggle.com/competitions/favorita-grocery-sales-forecasting/overview (accessed on 1 July 2024).
  9. Ghiletki, D. Machine Learning-Based Demand Forecasting for Supply Chain Management. Master’s Thesis, Department of Technology, Management and Economics, DTU, Copenhagen, Denmark, 2023. [Google Scholar]
Figure 1. Overview of the process flow for multivariate forecasting research.
Figure 1. Overview of the process flow for multivariate forecasting research.
Csmf 11 00029 g001
Figure 2. Comparison of demand across different sectors (electricity, docked bikes, and groceries).
Figure 2. Comparison of demand across different sectors (electricity, docked bikes, and groceries).
Csmf 11 00029 g002
Figure 3. Model performances on samples of the synthetic dataset.
Figure 3. Model performances on samples of the synthetic dataset.
Csmf 11 00029 g003
Figure 4. Individual external feature importance for daily, weekly, and monthly granularities in MAE.
Figure 4. Individual external feature importance for daily, weekly, and monthly granularities in MAE.
Csmf 11 00029 g004
Table 1. Average correlation (linear) with target for simulated external features.
Table 1. Average correlation (linear) with target for simulated external features.
X FeatureDailyWeeklyMonthly
Static trend0.510.570.63
Static seasonal−0.07−0.07−0.04
Exogenous0.310.360.43
Exogenous_lag_30.270.350.4
Exogenous_lag_50.270.350.44
Noise000
Table 2. Scaled MAEs for different models across time frequencies for univariate forecasting.
Table 2. Scaled MAEs for different models across time frequencies for univariate forecasting.
ModelDailyWeeklyMonthly
AutoLSTM (neuralforecast)0.0096520.0021370.08955
AutoRNN (neuralforecast)0.0098920.0022770.090511
Neuralprophet0.0094290.0026620.104538
TimeGPT(zeroshot)0.0060970.001962 *0.090954
LightGBM (Nixtla)0.006530.0074870.158012
Linear Regression (Nixtla)0.3649130.1237130.320041
Random Forest (Nixtla)0.0112170.0032260.124173
XGBoost (Nixtla)0.0063770.0047070.134841
LightGBM (PyCaret)0.0088080.002260.101377
Linear Regression (PyCaret)0.7284140.9060340.783822
XGBoost (PyCaret)0.0086870.0021920.100657
Random Forest (PyCaret)0.0087480.0022430.100578
ARIMA (Pmdarima)0.0059020.0020240.088495
ARIMA (Nixtla)0.005812 *0.002340.077143 *
MSTL ARIMA (Nixtla)0.0058340.0022980.081126
Prophet0.0072150.0019650.098457
Table 3. Scaled MAE for different time series characteristics.
Table 3. Scaled MAE for different time series characteristics.
TitleScenarioAutoLSTMAutoRNNNeuralProphetTimeGPTLGBM(Nixtla)RF(Nixtla)Xgb(Nixtla)Lgbm(PyC)Xgb(PyC)Arima(Pmd)MSTL Arima(Nixtla)ProphetArima(Nixtla)
trend- upward, linearWeekly0.030.070.010.001.000.110.280.000.000.020.050.000.02
Monthly0.100.120.060.010.590.500.630.000.000.010.120.020.00
Daily0.010.010.000.050.660.500.520.000.000.000.050.000.05
trend- upward with breaks, nonlinearWeekly0.040.011.000.050.020.070.050.430.390.030.070.330.04
Monthly0.020.010.180.020.570.631.000.070.090.110.010.070.01
Daily0.880.871.000.060.120.060.020.400.410.060.060.780.06
trend- upward with breaks, linear, AdditiveWeekly0.030.050.170.011.000.000.440.020.010.000.000.010.01
Monthly0.020.050.330.011.000.140.920.020.010.020.010.090.01
Daily0.280.340.420.011.000.010.840.100.070.010.010.000.01
trend- upward with breaks, linearWeekly0.580.590.670.440.690.560.570.560.570.360.320.470.32
Monthly0.340.340.390.330.770.640.460.380.380.300.260.370.19
Daily0.530.560.600.280.380.440.410.460.460.230.420.330.25
trend- upward with breaksWeekly0.110.140.890.031.000.010.100.350.130.020.010.220.01
Monthly0.010.010.130.010.960.211.000.010.010.000.010.020.00
Daily0.880.890.920.000.050.020.110.660.640.010.010.680.01
trend- stable with breaks, nonlinearWeekly0.210.250.940.100.060.030.100.690.670.000.020.320.00
Monthly0.040.050.120.011.000.490.590.080.080.070.010.060.00
Daily0.880.890.920.000.050.020.110.660.640.010.010.680.01
trend- stable with breaks, linear, MultiplicativeWeekly0.980.991.000.960.960.960.960.980.980.960.960.940.96
Monthly0.991.001.000.960.760.680.660.980.980.920.950.970.96
Daily1.001.001.000.950.950.890.940.970.970.950.950.910.95
trend- stable with breaks, linearWeekly0.230.280.950.190.290.210.290.400.400.130.140.340.12
Monthly0.110.150.300.090.790.520.690.130.140.080.080.130.07
Daily0.740.720.410.120.280.190.200.190.190.070.080.240.06
trend- downward, linearWeekly0.090.060.340.030.740.280.470.120.030.060.020.020.02
Monthly0.260.350.330.080.840.160.720.060.030.100.050.070.00
Daily0.300.250.360.060.930.070.900.110.080.090.060.110.06
trend- downward with breaks, nonlinearWeekly0.400.390.480.500.760.031.000.610.620.390.150.520.15
Monthly0.030.030.360.421.000.870.630.510.510.380.120.480.04
Daily0.430.430.270.350.960.490.180.380.380.041.000.360.04
trend- downward with breaks, linearWeekly0.350.380.830.320.640.390.610.610.600.280.240.430.24
Monthly0.290.320.510.290.680.570.460.350.350.260.230.350.20
Daily0.600.490.700.220.390.450.300.400.420.210.320.350.21
trend- downward with breaksWeekly0.220.261.000.020.060.050.190.320.330.040.010.420.03
Monthly0.010.010.120.000.321.000.610.020.010.010.020.010.01
Daily0.300.231.000.210.290.680.210.310.280.190.190.170.22
trend- downwardWeekly0.560.570.690.561.000.480.710.410.340.610.460.480.59
Monthly0.540.790.560.461.000.540.950.500.610.510.490.600.50
Daily0.780.880.810.810.790.980.811.000.980.870.790.760.89
multiple seasonalities, upward, linear, AdditiveWeekly0.290.300.580.240.540.291.000.260.490.390.210.160.23
Monthly0.350.480.680.220.850.240.600.180.180.290.270.260.24
Daily0.430.580.570.151.000.400.910.250.240.210.150.280.15
multiple seasonalities, stable, linear, MultiplicativeWeekly0.790.771.000.820.750.830.810.870.910.830.810.760.81
Monthly0.250.240.330.271.000.260.160.290.300.280.260.290.27
Daily0.750.761.000.820.790.440.690.870.870.820.820.670.82
multiple seasonalities, downward, linear, MultiplicativeWeekly0.910.901.000.950.770.890.300.940.960.930.940.920.94
Monthly0.410.410.450.421.000.200.160.420.410.400.420.420.42
Daily0.760.771.000.910.840.910.770.920.900.950.920.860.92
cycles- upward, nonlinear, AdditiveWeekly0.260.460.230.130.880.460.920.360.370.140.140.210.04
Monthly0.300.350.370.150.840.630.650.340.310.220.250.120.09
Daily0.250.350.980.020.181.000.090.170.340.200.020.150.02
cycles- upward, linear, MultiplicativeWeekly0.070.041.000.110.740.030.150.140.130.160.020.150.03
Monthly0.200.261.000.150.130.140.190.380.480.220.060.300.02
Daily0.250.350.980.020.181.000.090.170.340.200.020.150.02
cycles- upward, linear, AdditiveWeekly0.000.050.350.011.000.220.310.080.010.000.010.010.00
Monthly0.150.180.510.021.000.210.960.180.280.200.040.290.03
Daily1.000.970.310.060.430.490.340.310.230.030.040.170.04
cycles- upward with breaks, linear, MultiplicativeWeekly0.270.201.000.090.160.410.050.130.020.100.210.290.12
Monthly0.090.110.310.050.230.690.550.050.100.080.070.120.07
Daily0.150.150.790.141.000.150.960.160.160.140.140.500.15
cycles- upward with breaks, linear, AdditiveWeekly0.370.500.660.380.590.380.570.470.440.400.400.310.39
Monthly0.340.480.570.270.490.640.340.350.370.360.320.270.26
Daily0.660.610.440.300.760.220.690.230.240.240.270.280.27
cycles- stable, linear, AdditiveWeekly0.270.310.990.260.710.530.510.290.350.270.250.300.27
Monthly0.440.511.000.230.370.300.380.350.380.340.300.370.28
Daily0.860.920.610.220.320.390.350.340.320.220.220.460.22
cycles- stable with breaks, linear, AdditiveWeekly0.210.340.920.170.730.370.440.270.230.150.070.380.03
Monthly0.340.450.900.220.530.460.340.320.280.430.200.420.08
Daily0.520.510.760.020.100.270.070.160.080.060.020.250.02
cycles- downward with breaks, linear, MultiplicativeWeekly0.200.251.000.100.190.520.390.290.250.120.110.470.14
Monthly0.070.070.270.030.510.880.620.060.080.300.100.060.08
Daily0.530.550.890.280.170.300.290.200.210.120.241.000.26
cycle- upward, linear, AdditiveWeekly0.861.000.460.910.400.910.940.820.980.770.070.700.20
Monthly0.440.450.370.730.720.490.590.590.7210.130.60.13
Daily0.530.550.890.280.170.30.290.20.210.120.2410.26
Table 4. MAE for different models with and without all external variables.
Table 4. MAE for different models with and without all external variables.
Model NameDaily MAEWeekly MAEMonthly MAE
Without X With X Without X With X Without X With X
LightGBM (Nixtla)0.0068650.0133440.0635230.0633840.0635230.07267
Linear Regression (Nixtla)0.3527630.0815120.0671390.0956010.0671390.171675
Random Forest (Nixtla)0.0116330.0071330.0628870.0161480.0628870.057416
XGBoost (Nixtla)0.0066710.0065970.0627940.047390.0627940.060364
LightGBM (PyCaret)0.0096630.0096850.0686350.0004820.0686350.068837
Linear Regression (PyCaret)0.731683inf0.363750.835490.363750.685998
Random Forest (PyCaret)0.0095920.0094910.0682170.0004860.0682170.068322
XGBoost (PyCaret)0.0095150.0096980.0683990.0004860.0683990.068336
TimeGPT0.006450.0641040.0669950.0004890.0669950.064504
ARIMA (pmdarima)0.0062280.00660.0694040.0005190.0694040.069404
Neuralprophet0.0102630.0102550.0702150.0006760.0702150.070183
Prophet0.0073070.0080820.0710830.0004720.0710830.067343
Table 5. Case study results.
Table 5. Case study results.
Case StudyBest Model UsedGranularityForecast
Horizon
MAE
Univariate
MAE with
All Corr.
External Features
MAE TimeGPT
Electricity Demand ForecastingXGBoostMonthly35 months47,731.78 (MW)91,854.41 (MW)73,488.35 (MW)
Docked Bike Demand ForecastingARIMAWeekly32 weeks1.1121.74450.8898
Perishable Goods Demand ForecastingProphetDaily365 days6.566.72942.098
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Karim, S.M.A.; Zarrin, B.; Lassen, N.B. Multivariate Forecasting Evaluation: Nixtla-TimeGPT. Comput. Sci. Math. Forum 2025, 11, 29. https://doi.org/10.3390/cmsf2025011029

AMA Style

Karim SMA, Zarrin B, Lassen NB. Multivariate Forecasting Evaluation: Nixtla-TimeGPT. Computer Sciences & Mathematics Forum. 2025; 11(1):29. https://doi.org/10.3390/cmsf2025011029

Chicago/Turabian Style

Karim, S M Ahasanul, Bahram Zarrin, and Niels Buus Lassen. 2025. "Multivariate Forecasting Evaluation: Nixtla-TimeGPT" Computer Sciences & Mathematics Forum 11, no. 1: 29. https://doi.org/10.3390/cmsf2025011029

APA Style

Karim, S. M. A., Zarrin, B., & Lassen, N. B. (2025). Multivariate Forecasting Evaluation: Nixtla-TimeGPT. Computer Sciences & Mathematics Forum, 11(1), 29. https://doi.org/10.3390/cmsf2025011029

Article Metrics

Back to TopTop