Next Article in Journal
High-Order Difference Scheme for Time-Fractional Quasilinear Parabolic Equations
Next Article in Special Issue
A Regularized Backbone-Level Cross-Modal Interaction Framework for Stable Temporal Reasoning in Video-Language Models
Previous Article in Journal
An ALE Framework with an HLLC-2D Riemann Solver for Reactive Gas–Particle Flows
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Diffusion Weighted Ensemble Framework for Robust Short-Horizon Global SST Forecasting from Multivariate GODAS Data

by
Gwangun Yu
,
GilHan Choi
,
Moonseung Choi
,
Sun-hong Min
and
Yonggang Kim
*
Department of Software, Kongju National University, Cheonan 31080, Republic of Korea
*
Author to whom correspondence should be addressed.
Mathematics 2026, 14(4), 740; https://doi.org/10.3390/math14040740
Submission received: 23 January 2026 / Revised: 20 February 2026 / Accepted: 20 February 2026 / Published: 22 February 2026

Abstract

Accurate time series forecasting of sea surface temperature (SST) is essential for understanding the ocean climate system and large-scale ocean circulation, yet it remains challenging due to regime-dependent variability and correlated errors across heterogeneous prediction models. This study addresses these challenges by formulating SST ensemble time series forecasting aggregation as a stochastic, sample-adaptive weighting problem. We propose a diffusion-conditioned ensemble framework in which heterogeneous base forecasters generate out-of-sample SST predictions that are combined through a noise-conditioned weighting network. The proposed framework produces convex, sample-specific mixture weights without requiring iterative reverse-time sampling. The approach is evaluated on short-horizon global SST forecasting using the Global Ocean Data Assimilation System (GODAS) reanalysis as a representative multivariate dataset. Under a controlled experimental protocol with fixed input windows and one-step-ahead prediction, the proposed method is compared against individual deep learning forecasters and conventional global pooling strategies, including uniform averaging and validation-optimized convex weighting. The results show that adaptive, diffusion-weighted aggregation yields consistent improvements in error metrics over the best single-model baseline and static pooling rules, with more pronounced gains in several mid- to high-latitude regimes. These findings indicate that stochastic, condition-dependent weighting provides an effective and computationally practical framework for enhancing the robustness of multivariate time series forecasting, with direct applicability to global SST prediction from large-scale geophysical reanalysis data.

1. Introduction

Sea surface temperature (SST) is an important factor to examine in the ocean climate system, as it directly influences air–sea heat exchange, large-scale ocean circulation, and atmosphere–ocean interactions [1]. Fluctuations in SST play a critical role in climate variability, seasonal forecasting, and marine ecosystem dynamics. Consequently, accurate prediction of SST is essential for understanding and monitoring oceanic and climatic processes across multiple temporal scales. Global ocean reanalysis products have been widely used to analyze and model SST variability [2,3]. Among these, the Global Ocean Data Assimilation System (GODAS), developed by the National Oceanic and Atmospheric Administration (NOAA) and the National Centers for Environmental Prediction (NCEP), provides spatially and temporally consistent near-surface ocean temperature fields by assimilating heterogeneous observational data into numerical ocean circulation models. GODAS reanalysis data offer long-term, global-scale SST records and have become a vital data source for ocean climate studies. However, forecasting future SST from global ocean data remains challenging, as SST variability reflects the combined effects of atmospheric forcing, ocean dynamics, and unresolved sub-grid scale processes [2].
Traditional SST prediction approaches have mainly relied on physics-based numerical models or statistical methods. Physics-based models are physically interpretable but computationally expensive and sensitive to uncertainties in initial conditions and parameterization. Statistical approaches are computationally efficient but often assume linearity or stationarity, which limits their ability to represent nonlinear dynamics and long-range temporal dependencies. These limitations have motivated increasing interest in data-driven methods for SST prediction [4]. Recent advances in deep learning (DL) have significantly improved time series forecasting performance. Transformer-based architectures, such as iTransformer, have further improved multivariate relations and time-series dependency modeling through attention mechanisms, while PatchTST enhances long-horizon forecasting by learning informative representations from segmented temporal patches. In contrast, linear decomposition-based models, such as DLinear, provide efficient modeling of trend and seasonal components with low data complexity. Despite these advances, no single model architecture consistently achieves optimal performance across all temporal scales, regions, and forecasting horizons. SST prediction performance is inherently affected by model uncertainty and data variability. Different models tend to excel under varying temporal regimes, seasonal conditions, and degrees of variability. Relying on a single forecasting model can lead to unstable or suboptimal predictions, particularly under complex oceanic conditions. In addition to model uncertainty, SST predictability is also shaped by horizon-dependent time-series characteristics. At short forecast horizons, predictive skill is often dominated by persistence and strong local autocorrelation, and short-range predictability can degrade abruptly in a seasonally dependent manner [5,6]. As the forecast horizon increases, local persistence becomes less informative, and predictability is increasingly related to seasonal structure and slowly varying, low-frequency components associated with basin-scale modes of SST variability [7]. Moreover, spatiotemporally correlated and nonstationary error behavior has been reported in SST forecast products, implying that model performance can vary across regions even at the same lead time [8]. As a result, ensemble learning has been recognized as an effective strategy to enhance robustness and generalization by integrating the complementary predictive behaviors of multiple models [9,10].
Basic ensemble approaches for SST prediction depend on simple averaging or weighted combinations of deterministic model outputs. While effective to some extent, such approaches do not explicitly model predictive uncertainty or account for the complex, multivariate nature of SST forecasting errors. To address these limitations, probabilistic ensemble frameworks that can represent uncertainty in a principled manner are required [11,12]. Diffusion-weighted generative models provide a robust probabilistic framework for ensemble forecasting by leveraging noise-conditioned representations to estimate optimal ensemble weights without expensive iterative generation. Unlike conventional ensemble methods, the diffusion approach can learn the underlying uncertainty structure of model outputs and generate refined forecasts that accurately represent complex temporal variability [13,14,15].
In this study, we propose a diffusion-weighted ensemble framework for SST prediction using GODAS reanalysis data. We first generated base forecasts using heterogeneous DL models, including LSTM, iTransformer, PatchTST, and DLinear, each capturing distinct temporal characteristics of SST variability. The proposed diffusion-weighted ensemble model integrates these diverse predictions by modeling correlated error patterns and refining them into a unified forecast. This approach allows the ensemble to adaptively exploit the unique strengths of individual models while reducing prediction variance and improving robustness under region dependent variability within the short-horizon forecasting setting.
  • We construct an ensemble composed of diverse model architectures, including LSTM, iTransformer, PatchTST, and DLinear, each capturing complementary temporal characteristics of SST variability, such as long-range dependencies, patchwise temporal patterns, and trend–seasonal components.
  • We apply the proposed diffusion-weighted ensemble framework to the multivariate SST GODAS reanalysis data, leveraging consistent global-scale oceanic variables for robust spatiotemporal forecasting.
  • We demonstrate that the diffusion-weighted ensemble can effectively refine and combine model predictions by learning their joint uncertainty structure, leading to improved robustness across different forecasting horizons and temporal regimes compared to single models and conventional ensemble approaches.

2. Related Work

2.1. Traditional SST Prediction

Traditional SST prediction has been primarily based on physics-driven numerical ocean models and coupled atmosphere-ocean forecasting systems. Barreto et al. developed and evaluated an operational multigrid ocean forecasting system, demonstrating that physically consistent numerical models can provide reliable SST forecasts over regional domains when properly configured [16]. The main advantage of such systems lies in their physical interpretability and dynamical consistency; however, their performance is sensitive to model resolution and atmospheric forcing errors, leading to persistent regional biases. To mitigate systematic errors in numerical SST forecasts, data-driven bias correction has been explored. Storto et al. proposed a neural network-based surface heat flux correction method embedded in a Nucleus for European Modelling of the Ocean (NEMO) ocean model, showing a significant reduction in SST bias compared to the original configuration [17]. While this hybrid approach improves forecast accuracy without modifying the dynamical core, it remains dependent on the quality and representativeness of the training data and does not explicitly address forecast uncertainty.
In parallel, several studies have concentrated on the refinement and validation of operational forecasting systems. Kong et al. presented validation results for an operational marine forecasting system, highlighting that forecast skill varies substantially across regions and seasons [18]. Although such system-level improvements enhance overall SST performance, the deterministic nature of the forecasts limits their usefulness for probabilistic risk assessment.
The limitations of traditional SST prediction become more evident for extreme events. de Boisséson and Balmaseda investigated the seasonal predictability of marine heatwave occurrence and duration using the European Centre for Medium Range Weather Forecast (ECMWF) seasonal forecast system, showing that while some predictability exists, extreme SST events are often underestimated [19]. Similarly, Koul et al. analyzed the seasonal prediction of marine heatwaves in the Arabian Sea and reported limited reliability for event duration forecasts. These studies indicate that physics-based systems struggle to provide well-calibrated uncertainty information for extremes, motivating the need for probabilistic post-processing frameworks [20].

2.2. DL-Based SST Prediction

DL approaches have been increasingly applied to SST prediction due to their ability to learn nonlinear spatiotemporal dependencies directly from data. Hao et al. investigated ConvLSTM and ST-ConvLSTM models for SST prediction in the South China Sea, systematically analyzing the impact of traditional methods; nevertheless, performance degraded for longer lead times and under dynamically complex conditions [21]. To enhance temporal dependency modeling, Xu et al. proposed a DL framework for short-term global SST prediction using reanalysis data [22]. The study demonstrated that DL models can outperform conventional statistical baselines for global scale forecasting. Nevertheless, the model focused on deterministic point predictions and did not quantify predictive uncertainty, limiting its applicability to risk-aware forecasting.
Attention mechanisms have been introduced to address the limitations of recurrent architectures. Zrira et al. proposed an attention-based BiLSTM model for SST time series forecasting, demonstrating improved long-range temporal feature extraction and higher accuracy than standard LSTM models [23]. Despite these advancements, the method remains primarily time-series-based and does not fully exploit spatial field structures. To enhance spatial modeling, Shi et al. integrated a deformable attention transformer into a ConvLSTM framework, enabling the model to adaptively capture spatial heterogeneity in SST fields [24]. This approach improves prediction accuracy in regions with significant mesoscale variability but introduces increased model complexity and computational cost.
A growing body of work explores multivariate learning approaches. Fu et al. proposed a hybrid model combining LSTM and Transformer architectures, which integrates sequential learning with global attention, demonstrating improved SST prediction across various coastal regions [25]. Yang et al. further indicated that incorporating various physically related variables into a multifactor DL framework enhances robustness under diverse oceanic conditions [26]. However, most multivariate DL studies still emphasize deterministic accuracy metrics and lack probabilistic evaluation.

2.3. Ensemble Techniques for SST Prediction

Ensemble techniques have been widely adopted to enhance SST prediction robustness by integrating the complementary strengths of diverse forecasting models. Dai et al. proposed a stacked generalization ensemble that combines multiple DL predictors through a meta-learning stage, demonstrating that aggregating heterogeneous models can consistently outperform individual predictors in SST forecasting [27]. The key advantage of this approach lies in its ability to exploit complementary error characteristics across models; however, it requires training and maintaining multiple base learners, as well as an additional meta-learner, resulting in increased computational and operational complexity. From a modeling perspective, ensemble performance strongly depends on the diversity of base learners. Qian et al. showed that combining predictors that capture different physical and dynamical aspects of the ocean, such as integrating SST with sea surface height anomalies, geostrophic velocities, and wind stress, leads to improved forecast skill in dynamically complex regions [28]. While this multivariate ensemble framework enhances deterministic accuracy, predictive uncertainty is still inferred indirectly from ensemble spread rather than being explicitly modeled.
Ensemble learning has also been applied to SST-related extreme event prediction. Bonino et al. employed machine-learning-based ensemble approaches to predict SST variability and marine heatwave occurrence across multiple Mediterranean subregions, highlighting that ensemble aggregation improves robustness under anomalous conditions [29]. Nevertheless, their framework focused on deterministic event prediction, and probabilistic uncertainty was not formally quantified or calibrated. A different ensemble perspective leverages diversity across climate model simulations rather than observational predictors. Boschetti et al.trained ensemble machine learning models exclusively on climate model outputs and demonstrated that machine learning can function as an interpolator among ensemble members, enhancing SST predictability assessment [30]. Although effective in utilizing model diversity, this approach remains dependent on the fidelity of the underlying climate simulations and does not directly address forecast uncertainty calibration.

3. Dataset

In this study, we use ocean temperature and related oceanographic variables obtained from the GODAS. GODAS integrates in situ and satellite observations with a numerical ocean circulation model through data assimilation, providing dynamically consistent estimates of the ocean state [31]. Moreover, NCEP GODAS data are provided by the NOAA Physical Sciences Laboratory (PSL), Boulder, CO, USA, on their website at https://psl.noaa.gov. Note that GODAS does not provide a variable explicitly labeled as SST, as shown in Figure 1, which represents the temperature value in GODAS. The dataset includes ocean potential temperature fields at different vertical levels. Hereafter, SST refers to the near-surface potential temperature from GODAS unless otherwise stated; specifically, it denotes the ocean temperature at approximately 5–10 m depth. This near-surface temperature is used as a proxy for SST in reanalysis-based ocean climate and prediction studies and is suitable for analyzing large-scale and seasonal SST variability.

3.1. Rationale for Selecting GODAS

To contextualize our selection of GODAS, we briefly compare it with two different global ocean reanalysis products, Ocean Reanalysis System 5 (ORAS5) and Global Ocean Physics Reanalysis (GLORYS) 12. ORAS5 provides a global ocean–sea ice ensemble reanalysis and analysis framework, where multiple ensemble members are generated to represent uncertainty in the estimated coupled ocean–ice state [32]. GLORYS12 is a high-resolution global ocean and sea-ice reanalysis designed to represent mesoscale variability with greater detail, particularly during the altimetry era [33]. Both products are, therefore, well suited for studies that prioritize uncertainty quantification (ORAS5) or eddy-resolving ocean dynamics (GLORYS12). In this study, however, our goal is to benchmark heterogeneous deep forecasters and a diffusion-weighted ensemble under a controlled short-horizon protocol at the global scale, where the primary requirement is a stable, long-record, and globally consistent near-surface thermal field. From this perspective, GODAS provides an appropriate and practical baseline that supports reproducible preprocessing and equitable model comparison.

3.2. Data Structure

The GODAS dataset covers the global ocean domain on a regular latitude–longitude grid. Each grid point represents a fixed geographic location for which various oceanographic variables are available as time series. In this study, land grid points where ocean variables are undefined are removed during preprocessing. The data are organized into a spatio-temporal structure in which each sample corresponds to a specific grid point and time index. For each latitude–longitude location, multivariate time series are constructed by stacking the SST and additional oceanographic variables along the temporal dimension. Table 1 summarizes the oceanographic variables provided by the GODAS dataset. Monthly mean data are used throughout this study.

3.3. Data Preprocessing

Several preprocessing steps are applied to the GODAS data prior to model training. Although GODAS provides a wide range of oceanographic variables, not all of them are utilized in this study. Variables are selected to ensure spatial and temporal consistency across the global ocean, which is essential for learning coherent spatio-temporal dependencies in multivariate forecasting models. In particular, variables that exhibit persistent missing values over large contiguous ocean regions or extended temporal intervals are excluded. Such irregular spatial and temporal coverage can introduce structured sparsity, distort local spatio-temporal correlations, and bias the learning of shared representations across grid points and time steps. When incorporated into multivariate time series models, these inconsistencies may lead to unstable training dynamics and degrade generalization by forcing the model to learn from uneven or discontinuous spatio-temporal data. Therefore, we restrict the input set to variables that are spatially continuous and temporally stable over the study period to avoid irregular sampling effects. The final set of input variables used in this study is summarized in Table 2.
To reduce the influence of the mean seasonal cycle and to focus the learning process on interannual and subseasonal variability, anomaly fields are computed for selected variables. For sea surface temperature, surface heat flux, and sea surface height, anomalies X ( u , v , t ) are calculated as follows:
X ( u , v , t ) = X ( u , v , t ) X c train ( u , v ) .
Here, ( u , v ) index the spatial grid coordinates (latitude and longitude) and t denotes the time index, where X ( u , v , t ) denotes the raw variable at grid point ( u , v ) and time t, and X c train ( u , v ) represents the monthly climatological mean for calendar month c { 1 , , 12 } , computed at the same grid point. Other variables, including zonal and meridional ocean currents and upper ocean layer depth variables, are used in their original form.
All input variables are standardized using statistics. Standardization is performed as follows:
X ˜ ( u , v , t ) = X ( u , v , t ) μ train σ train ,
where μ train and σ train denote the mean and standard deviation calculated over the training period. This procedure prevents information leakage from the validation and test periods and improves numerical stability during model training.
For variables that contain missing values, such as ocean mixed layer depth and isothermal layer depth, missing values are retained during preprocessing and set to zero after standardization. Binary masks indicating valid observations are maintained to distinguish missing values during model training.

4. Methodology

This section describes the experimental protocol and the proposed ensemble framework for SST prediction. To ensure a fair and controlled comparison across learning paradigms, we fix the input window and forecasting horizon to seq _ len = 12 and pred _ len = 1 for all models. All approaches share the same preprocessing pipeline and the same train/validation/test split described in Section 3. In addition, all normalization statistics (e.g., mean and standard deviation for z-score standardization) are computed using the training split only and then reused for the validation and test splits to prevent any information leakage.

4.1. Base Forecasting Models

We employ a set of heterogeneous forecasters to capture complementary temporal characteristics of SST variability. Let M DL = { DLinear , iTransformer , PatchTST , LSTM } denote the deep-learning forecasters. To provide additional reference baselines under the short-horizon setting ( pred _ len = 1 ), we also include two classical machine-learning regressors: Random Forest (RF) and LinearSVR [34,35]. RF is a strong nonlinear tree-ensemble baseline that captures feature interactions with minimal modeling assumptions, while LinearSVR offers scalable learning with a linear kernel on high-dimensional inputs. We define the full set of trained models as M base = M DL { LinearSVR , RF } . For ensemble construction, however, we restrict candidates to the deep-learning set M ens = M DL to focus subset enumeration and adaptive aggregation on models that explicitly learn temporal structure. This restriction ensures that ensemble aggregation operates on forecasts derived from sequence-aware representations rather than flattened lag features. Such sequence-aware models produce forecasts whose errors reflect regime-dependent temporal dynamics, which is essential for learning meaningful, condition-dependent ensemble weights.

4.2. Input Representation for Classical Baselines

Unlike sequence models, classical regressors require fixed-length vectors. For LinearSVR and RF, we convert each sample’s multivariate history into a fixed-dimensional feature vector that matches the temporal context of the deep models. Specifically, the past seq _ len = 12 time steps are concatenated into z R 12 d , where d is the number of input variables. The flattening order across time and variables is kept identical across all experiments to ensure reproducibility. The resulting vectors are standardized using training-set statistics; LinearSVR uses standardized inputs by default, and RF is also trained with the same standardized inputs for consistency.

4.3. Hyperparameter Optimization and Model Selection

For each architecture m M base , we optimize hyperparameters using Optuna [36]. Each Optuna trial is trained on the training split and evaluated on the validation split using a fixed loss metric (e.g., MSE or RMSE). For deep models, we apply early stopping and retain the checkpoint that achieves the lowest validation loss. For LinearSVR and RF, we refit the estimator using the selected hyperparameters on the training split and then obtain validation predictions. After selecting the best configuration for each model, we fix the corresponding forecasters { f m } m M base and generate out-of-sample predictions on both the validation and test splits. These predictions are used for (i) single-model comparisons and (ii) deep-learning ensemble construction.

4.4. Validation-Based Ensemble Subset Selection

Figure 2 illustrates the overall workflow. Importantly, ensemble fitting and configuration selection are conducted only within the validation split via a meta-train/meta-validation protocol, and the test split is used exactly once for final reporting.
Let P val = { y ^ n ( m ) m M ens , n D val } denote the collection of out-of-sample base predictions on the validation split. Similarly, let P test = { y ^ n ( m ) m M ens , n D test } denote the corresponding predictions on the test split. To exploit complementary error patterns among forecasters, we evaluate all non-empty subset ensembles S M ens . Let K = | S | be the number of models in subset S. Let y n denote the ground-truth target for sample n. For each sample n, we collect base predictions into a vector
x 0 , n = y ^ n ( m 1 ) , y ^ n ( m 2 ) , , y ^ n ( m K ) R K , { m 1 , , m K } = S ,
where y ^ n ( m i ) denotes the prediction of base model m i S for sample n, and y n denotes the ground-truth target. When K = 1 , the ensemble reduces to the corresponding single model. When K 2 , we compare multiple aggregation rules, including uniform averaging, validation-optimized convex weighting, Bayesian model averaging (BMA), quantile regression forests (QRF), and the proposed diffusion-weighted ensemble. To reduce overfitting during ensemble selection and to avoid any leakage from the test split, we adopt a two-stage protocol on the validation split. Concretely, we first fit any ensemble parameters on a meta-train subset, then select the best subset and aggregation rule using a disjoint meta-validation subset. The selected configuration is subsequently refit using the full validation split, and the test split is used exactly once for final reporting. We denote these two disjoint subsets as D meta - train and D meta - validation (both drawn from the validation split), where ensemble parameters are fitted on D meta - train and configurations are selected based on performance on D meta - validation .

4.5. Aggregation Rules for Sample-Adaptive Ensemble Forecasting

Ensemble performance depends not only on which base forecasters are included but also on how their outputs are combined. Because different forecasters can exhibit correlated errors and regime-dependent strengths (e.g., seasonal vs. interannual variability), there is no single universally optimal pooling strategy. Therefore, we compare five aggregation rules with increasing modeling capacity and adaptivity to (i) establish a transparent baseline, (ii) test whether a global, validation-fitted combination improves over naive averaging, (iii) benchmark against evidence-weighted mixtures commonly used in geophysical forecasting, (iv) assess a nonlinear nonparametric combiner that can capture heteroscedastic and asymmetric regimes, and (v) evaluate whether sample-adaptive weighting yields additional gains. Concretely, we consider: (1) a uniform mean as a strong, assumption-light reference; (2) validation-optimized convex weights (linear pool) as a classical forecast-combination method that learns a global mixture while mitigating overfitting via nonnegativity and sum-to-one constraints; (3) Bayesian model averaging (BMA) as an evidence-weighted global mixture baseline; (4) quantile regression forests (QRF) as a nonlinear, distribution-aware combiner; and (5) the proposed diffusion-weighted rule as a sample-adaptive mechanism that can vary mixture weights across different oceanic conditions. This controlled comparison allows us to attribute improvements to the aggregation strategy itself rather than to changes in base models, and it clarifies the value of sample-adaptive weighting beyond both conventional global pooling and recent geophysical ensemble baselines.
Given a subset S of K base forecasters, for each sample n we define the stacked base prediction vector x 0 , n = [ y ^ n ( 1 ) , , y ^ n ( K ) ] R K and denote by y n the corresponding ground-truth target (each sample can be interpreted as a specific spatiotemporal point, e.g., a grid cell at a given time). The uniform mean ensemble is defined as
y ^ mean = 1 K i = 1 K y ^ ( i ) .
As a stronger global pooling baseline, we also learn nonnegative convex weights on the meta-train split:
w = arg min w R K n D meta - train w x 0 , n y n 2 s . t . w 0 , 1 w = 1 ,
and predict by y ^ cw , n = ( w ) x 0 , n . While convex weighting provides a simple yet competitive global combination rule, it cannot adapt to sample-specific conditions. To strengthen the positioning of our diffusion-weighted aggregation against recent ensemble practices in geophysical forecasting, we additionally include two widely used alternatives: Bayesian model averaging (BMA) and quantile regression forests (QRF). BMA provides an evidence-weighted global mixture by assigning higher weights to base forecasters that better explain the meta-train targets. Concretely, let mse i = 1 | D meta - train | n D meta - train y ^ n ( i ) y n 2 denote the meta-train mean squared error of model i. Under a Gaussian-error approximation, we compute AIC-like weights
w ˜ i exp 1 2 | D meta - train | log ( mse i ) + 2 , w i = w ˜ i j = 1 K w ˜ j ,
and form the BMA prediction as y ^ bma , n = i = 1 K w i y ^ n ( i ) . This yields a principled evidence-weighted pooling rule that remains global (sample-invariant) yet goes beyond uniform or convex weighting.
In contrast, QRF serves as a nonlinear, nonparametric combiner that can better accommodate regime-dependent and potentially asymmetric error behavior. Specifically, we train a random-forest regressor g ϕ on the meta-train split to map the base prediction vector x 0 , n to the target y n . At inference, we collect per-tree predictions { g ϕ , t ( x 0 , n ) } t = 1 T and use their empirical quantile as a robust point forecast:
y ^ qrf = Quantile q g ϕ , t ( x 0 , n ) b = 1 B .
Here, b indexes the trees in the random forest, and B is the total number of trees. Unless stated otherwise, we use q = 0.5 (the median). Together, BMA and QRF provide complementary and widely recognized baselines—evidence-weighted global pooling and nonlinear, distribution-aware combining, respectively—which help contextualize the gains from sample-adaptive diffusion-based weighting. Building on these global and nonparametric baselines, we propose a diffusion-weighted ensemble, inspired by diffusion probabilistic models [37], that uses only the forward diffusion process to construct noise-conditioned inputs while omitting iterative reverse-time sampling. Specifically, we perturb the base prediction vector via a forward noising process:
x τ = α ¯ τ x 0 + 1 α ¯ τ ϵ , ϵ N ( 0 , I K ) ,
where τ Unif { 0 , , T 1 } , T is the number of diffusion steps, β τ ( 0 , 1 ) is a predefined noise schedule, α τ = 1 β τ , and for τ 1 we define α ¯ τ = s = 1 τ α s (with α ¯ 0 = 1 by convention). In practice, β τ is set to a monotone increasing schedule over τ = 0 , , T 1 . To assess sensitivity to this design choice, we consider five schedule families (linear, cosine, quadratic, sigmoid, and exponential) while keeping ( β start , β end , T ) fixed. All schedules are mapped to the same range [ β start , β end ] for a fair comparison. Given the noised input x τ , we learn a noise-conditioned weighting network f θ ( · ) that takes the concatenated input [ x τ ; e ( τ ) ] and outputs K logits, which are converted into mixture weights via
w τ = softmax f θ ( [ x τ ; e ( τ ) ] ) R K , i = 1 K w τ , i = 1 ,
where e ( τ ) R d e denotes a sinusoidal embedding of the diffusion step τ . The final forecast is computed as a convex combination of the clean base predictions:
y ^ diff = w τ x 0 ,
i.e., the weights are conditioned on the noised input x τ but applied to the clean predictions x 0 . We train θ on the meta-train split while freezing all base models by minimizing the expected mean squared error under randomly sampled ( τ , ϵ ) :
L ( θ ) = E ( x 0 , y ) D meta - train , τ , ϵ w τ x 0 y 2 ,
where the expectation is approximated by mini-batch sampling during training. At inference, we perform Monte-Carlo averaging by drawing M pairs ( τ , ϵ ) , averaging the resulting weights w ¯ = 1 M j = 1 M w ( j ) , and applying the averaged weights to x 0 to obtain y ^ diff = w ¯ x 0 . Unless stated otherwise, we use an exponential noise schedule
β τ = β start β end β start τ T 1 , τ = 0 , , T 1 ,
with β start = 10 4 , β end = 0.02 , and T = 50 . To address potential sensitivity to the choice of β τ , we additionally evaluate four alternative schedules (linear, cosine, quadratic, and sigmoid) under the same ( β start , β end , T ) setting. The overall architecture of the proposed diffusion-weighted ensemble is illustrated in Figure 3. Across all 11 ensemble subsets ( K { 2 ,   3 ,   4 } ), the differences are small: the exponential schedule yields the lowest average test RMSE, whereas the cosine schedule attains the best single-subset RMSE (Table A5, Table A6 and Table A7). For clarity and reproducibility, we adopt the exponential schedule as the default in the main experiments and report schedule sensitivity in the Appendix A. The weighting network f θ is implemented as a four-hidden-layer MLP with sinusoidal timestep embeddings (embedding dimension d e = 32 , hidden dimension 64) and is trained using AdamW (learning rate 10 3 , weight decay 10 6 ) for 20 epochs with gradient clipping at 1.0; at inference, we use M = 8 Monte-Carlo draws. We select the MLP depth as 4 hidden layers based on an ablation study over depths { 2 ,   4 ,   8 ,   16 } conducted across all 11 ensemble subsets ( K { 2 ,   3 ,   4 } ). Appendix Table A3 summarizes the aggregate behavior of each depth by reporting the mean ± std and the minimum RMSE over the 11 subsets, which provides a robustness-oriented rationale for the default choice. Appendix Table A4 complements this summary by listing the full subset-by-depth RMSE matrix (val/test) to document all runs and to show that depth 4 most frequently attains the lowest test RMSE across subsets. Accordingly, depth 4 achieves the lowest average RMSE and yields the overall best RMSE among all tested depths, while overly deep networks (e.g., depth 16) slightly degrade the mean performance.

4.6. Relation to Standard Diffusion Models

Standard diffusion generative models learn a reverse-time process to synthesize samples [37]. In contrast, our approach uses only forward noising and noise-conditioning to compute ensemble weights, avoiding iterative reverse sampling. Therefore, advances that accelerate reverse sampling (e.g., EDM, DPM-Solver, Consistency Models, and flow-based reformulations) are orthogonal to our objective [38,39,40,41]. Based on the methodology described above, we now evaluate the proposed ensemble strategies on the held-out test set. We compare the performance of individual base forecasters and different ensemble aggregation rules, with particular emphasis on the proposed diffusion-weighted ensemble. All reported results are obtained using the fixed experimental protocol and evaluation metrics described in this section.

5. Results

We evaluate forecasting performance using root-mean-square error (RMSE), mean absolute error (MAE), and the coefficient of determination ( R 2 ). All metrics are reported on the held-out test split unless stated otherwise. For interpretability, we report RMSE both in standardized anomaly units (after z-score normalization) and in physical units (°C) by converting predictions back using the training-set statistics. Although the relative improvements in RMSE and MAE may appear numerically modest, they are meaningful in our short-horizon setting. For one-step-ahead SST anomaly forecasting, persistence and strong local autocorrelation already constitute a strong baseline, leaving limited headroom for further error reduction. In such a regime, consistent improvements in aggregate metrics suggest that the proposed method reduces residual error components beyond persistence, potentially including regime-dependent mismatches and correlated model errors. Moreover, because evaluation is conducted over a global domain with a large number of spatial samples, small average error reductions are less likely to arise from isolated cases and instead indicate broadly consistent changes in the overall error distribution. These effects can correspond to more substantial localized gains in dynamically complex regions, which is consistent with our latitude-band analysis showing more pronounced improvements in several mid- to high-latitude bands.

5.1. Single-Model Performance

Table 3 compares the performance of individual forecasting models, including four DL forecasters (DLinear, iTransformer, PatchTST, LSTM) and two classical machine-learning baselines (LinearSVR and RF). Overall, the best single model achieves the lowest RMSE and MAE and the highest R 2 , indicating stronger agreement with the observed SST anomaly variability. We observe that the classical baselines remain competitive under the short-horizon setting ( pred _ len = 1 ), and that flattened lag features can still provide strong predictive signals for near-term SST anomalies.
Among individual forecasters, LSTM achieves the strongest overall accuracy, yielding the lowest test RMSE (0.3612 °C) and MAE (0.2627 °C) and the highest R 2 (0.6874) in Table 3. This suggests that, under the short-horizon setting ( pred _ len = 1 ), recurrent modeling of local temporal dynamics remains effective for SST anomaly prediction. Notably, the classical baselines (LinearSVR and RF) remain competitive despite using flattened lag features, indicating that near-term SST anomalies contain strong linear and low-order nonlinear signals that can be exploited without explicit sequence modeling.

5.2. Ensemble Performance by Aggregation Rule

We compare three aggregation rules: uniform mean, validation-optimized convex weights, and the proposed diffusion-weighted ensemble. As shown in Table 4, uniform averaging provides a competitive but not consistently superior baseline, whereas validation-optimized convex weighting and the proposed diffusion-weighted ensemble improve upon the best single model.
Relative to the best single model (LSTM), the diffusion-weighted ensemble reduces the test RMSE from 0.3612 °C to 0.3586 °C, corresponding to a relative improvement of approximately 0.70%. A similar improvement is observed in MAE (from 0.2627 °C to 0.2612 °C; ∼0.57%), while R 2 increases from 0.6874 to 0.6918.
Among global (sample-invariant) pooling rules, uniform mean provides a stable reference but cannot account for correlated forecast errors, and validation-optimized convex weighting improves upon averaging by learning a constrained linear pool. To better contextualize diffusion-based weighting relative to ensemble practices commonly used in geophysical forecasting, we additionally evaluate Bayesian model averaging (BMA) and quantile regression forests (QRF). In our setting, BMA performs comparably to the best single forecaster, suggesting that evidence-weighted yet still global mixing offers limited gains when base-model errors are strongly correlated or when the evidence weights concentrate on a single dominant forecaster. Meanwhile, QRF underperforms, indicating that a generic nonlinear combiner does not automatically translate to improved accuracy under our controlled protocol and fixed feature set.
Overall, the proposed diffusion-weighted aggregation achieves the best performance across all metrics by producing sample-adaptive mixture weights, enabling the ensemble to vary its weighting across different oceanic conditions while consistently outperforming both conventional global pooling and recent geophysical ensemble baselines.

5.3. Latitude-Band Error Analysis

To examine where the proposed ensemble improves the most, we compute test RMSE within 10° latitude bands and report band-wise errors for LSTM and the diffusion-weighted ensemble. This analysis highlights regional differences in predictability and helps interpret whether improvements are concentrated in specific latitude regimes.
Figure 4 and Table 5 demonstrate that the performance of the diffusion-weighted ensemble, configured with an exponential noise schedule, is latitude-dependent. Notably, clearer gains are observed in mid-to-high latitude bands (e.g., 35–65°), with the largest improvement at 40° (Imp ≈ 2.61%), whereas some low-to-mid latitude bands exhibit marginal or slightly negative changes (e.g., −10° shows ≈−1.03%). A plausible explanation is that mid-to-high latitudes are characterized by stronger variance and regime shifts, which amplify inter-model disagreement and nonstationary error correlations; under such conditions, diffusion-conditioned reweighting can better exploit the complementary strengths of heterogeneous forecasters. In contrast, tropical and low-latitude regimes are often dominated by smoother variability, leading to higher concordance among base predictors and limiting the potential gains from adaptive reweighting. Critically, these results should be interpreted together with absolute error magnitudes: low-latitude regions typically exhibit smaller baseline RMSEs than higher latitudes; so, even small absolute differences can appear as noticeable percentage changes due to the smaller denominators. Thus, slightly negative values in a few low-latitude bands do not necessarily indicate a substantive degradation, but rather reflect the sensitivity of relative metrics in low-variance regimes. Overall, the proposed method is most impactful in complex, high-variance regions, while remaining competitive in smoother regimes.

5.4. Best Subset Selection: Validation vs. Test

To study whether all four models are necessary, we evaluate all non-empty subsets of M ens under each aggregation rule using the meta-train/meta-validation protocol described in Section 4.4. We consider three aggregation rules: uniform mean, validation-optimized convex weights, and the proposed diffusion-weighted ensemble. We report (i) the configuration selected by the lowest validation RMSE (the proper selection criterion) and (ii) the configuration that achieves the lowest test RMSE (reported only as a post hoc reference). Table 6 summarizes both cases.
The Val-best row in Table 6 represents the only configuration that is valid under our experimental protocol because it is selected using validation RMSE without accessing any test labels. In contrast, the Test-best (post hoc) row is reported solely as a reference point; this test-oracle choice is not a valid selection method and must not be interpreted as a deployable configuration because it implicitly uses test outcomes for model selection. Importantly, the validation-selected diffusion configuration remains highly competitive on the test set (0.3597 °C), with only a negligible gap compared to the post hoc best-test configuration (0.3592 °C). This small difference suggests that validation-driven subset selection provides a reliable proxy for generalization and that the proposed diffusion-weighted ensemble remains effective even when the ensemble is formed from a reduced subset of base forecasters.

6. Discussion

In this study, we investigated short-horizon sea surface temperature (SST) forecasting from GODAS reanalysis using a diffusion-weighted ensemble framework. We constructed a multivariate global dataset from GODAS and defined SST as the near-surface (uppermost-level) potential temperature, with anomalies computed using training-period monthly climatology to reduce the dominant seasonal cycle. Under a controlled protocol with a fixed input window ( seq _ len = 12 ) and one-step forecasting horizon ( pred _ len = 1 ), we trained heterogeneous base forecasters (DLinear, iTransformer, PatchTST, and LSTM) and generated out-of-sample predictions for validation and test splits to enable leak-free ensemble construction.
To combine base forecasts, we compared three aggregation rules with increasing modeling capacity: uniform mean, validation-optimized convex weighting (linear pool), and the proposed diffusion-weighted ensemble that enables one-shot weight estimation without reverse sampling. Ensemble configuration selection was performed strictly within validation using a meta-train/meta-validation split, and the test set was used only once for final reporting. Overall, the results demonstrate that combining heterogeneous deep-learning forecasters improves robustness compared to the best single model, and that the diffusion-weighted ensemble consistently outperforms simpler global pooling strategies.
The latitude band analysis suggests that the benefits of sample-adaptive weighting are not uniform across regions; improvements are more pronounced in several mid-to-high latitude bands, while some low-latitude bands show marginal or negative changes. Finally, subset enumeration indicates that competitive performance can be achieved with reduced ensembles (e.g., three-model subsets), implying that strong performance does not necessarily require all candidate models when selection is conducted properly on validation.
Despite these promising results, several limitations remain. First, our experiments focus on a single short-horizon setting ( pred _ len = 1 ) and monthly GODAS data; extending the framework to longer lead times and higher temporal resolution (e.g., weekly or daily products) is required to fully characterize the advantages of diffusion-based adaptive aggregation under increased uncertainty. Second, GODAS is a reanalysis product and therefore reflects model and assimilation biases; evaluating the proposed ensemble on independent observational datasets and alternative reanalyses would strengthen the generality of the conclusions. Third, while we analyzed latitude-dependent errors, additional diagnostics (e.g., seasonality, regional ocean basins, and extreme-event regimes such as marine heatwaves) would provide deeper insight into when and why sample-adaptive weighting is most beneficial.
Our current experiments are limited to one-step forecasting. Nonetheless, the proposed diffusion-weighted aggregation is, in principle, compatible with multi-step forecasting settings. (i) In a recursive rollout, one-step base forecasts are iteratively propagated; the ensemble can recompute diffusion-conditioned weights at each step using the available base prediction vector for that step, while acknowledging the well-known risk of error accumulation and compounding errors. (ii) In a direct multi-step setting, where an H-step output is predicted in a single forward pass, the weighting network could be conditioned on the lead time (e.g., via a horizon embedding) or could output an H × K set of weights, allowing the mixture to vary across lead times if error correlations differ. A systematic assessment of these extensions (including training base forecasters for multi-horizon targets and re-tuning diffusion/noising-related hyperparameters) is beyond the scope of this study and is left for future work.
Future work will extend the proposed framework in three directions. (i) We will evaluate multi-step forecasting horizons and probabilistic calibration, including uncertainty-aware metrics, to quantify whether diffusion-weighted aggregation improves reliability as lead time increases. Moreover, the increased sample size at daily or sub-seasonal resolution may support training higher-capacity forecasters; however, any performance gain must be verified empirically under the same leak-free selection protocol. (ii) We will incorporate additional physically relevant predictors and alternative anomaly definitions to assess the sensitivity of model performance to preprocessing choices. (iii) We will explore lightweight and computationally efficient variants of the weighting network to enable deployment-oriented inference while preserving the benefits of sample-adaptive mixture weights.

7. Conclusions

Collectively, this work demonstrates that diffusion-weighted, sample-adaptive aggregation provides a practical and effective pathway to improving the robustness of SST forecasting from multivariate global reanalysis data.

Author Contributions

Conceptualization, Y.K.; methodology, G.Y.; investigation, G.Y.; formal analysis, G.Y. and G.C.; validation, M.C. and S.-h.M.; writing—original draft preparation, G.Y. and G.C.; writing—review and editing, Y.K.; and supervision, Y.K. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Regional Innovation System & Education(RISE) program through the Chungnam RISE center, funded by the Ministry of Education(MOE) and the Chungcheongnam-do, Republic of Korea (2025-RISE-12-003).

Data Availability Statement

The data used in this study are derived from public domain resources. The Global Ocean Data Assimilation System (GODAS) reanalysis data are publicly available from the NOAA National Centers for Environmental Information (NCEI) at https://psl.noaa.gov.

Acknowledgments

This work was supported by the Regional Innovation System & Education(RISE) program through the Chungnam RISE center, funded by the Ministry of Education(MOE) and the Chungcheongnam-do, Republic of Korea (2025-RISE-12-003).

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A. Complete Results

This appendix provides full validation/test results to ensure transparency and reproducibility. All ensemble configurations are selected using validation only (meta-train/meta-validation protocol; Section 4.4), and the test split is used once for final reporting. We additionally report a test-oracle configuration solely as a post hoc reference to illustrate the upper bound.

Appendix A.1. Single-Model Baselines

Table 3 compares all single models on validation and test splits using RMSE, MAE, and R 2 . For the deep-learning models, we report both standardized metrics (z-score space) and physical-unit errors in °C (after inverse standardization). Classical baselines (LinearSVR, RF) are evaluated in standardized space. Note that RMSE/MAE in °C are obtained by inverse-transforming standardized predictions using training-set statistics, whereas z-space metrics directly reflect errors in normalized anomaly units.
Table A1. Single-model performance on the validation and test splits.
Table A1. Single-model performance on the validation and test splits.
ModelVal RMSE (z)Test RMSE (z)Test MAE (z)Test R 2 Test RMSE (°C)Test MAE (°C)
DLinear0.63310.64780.47460.66220.3750.275
iTransformer0.63030.64720.47570.66280.3750.276
PatchTST0.62540.64490.47150.66520.3740.273
LSTM0.61560.62310.45330.68740.3610.263
LinearSVR0.62480.64210.46840.66480.3720.271
RF0.61880.63980.46160.66720.3700.267
For deep-learning models, we additionally report errors in physical units (SST in °C) after inverse standardization. For classical baselines (LinearSVR, RF), only standardized metrics are available.
For brevity, we use the following model abbreviations: D (DLinear), I (iTransformer), P (PatchTST), and L (LSTM). For ensemble rules, single denotes the base model, mean the uniform mean ensemble, cw the validation-optimized convex weighting, and diff the diffusion-weighted ensemble.

Appendix A.2. Ablation on the Depth of the Diffusion Weighting Network

To justify the architecture choice for the diffusion weighting network f θ , we ablate the number of hidden layers (MLP depth) in { 2 ,   4 ,   8 ,   16 } while keeping all other hyperparameters and the experimental protocol fixed. For each depth, we report the configuration selected by the lowest meta-validation RMSE and its corresponding test performance. As shown in Table A2, depth =   4 achieves the best meta-validation RMSE and the strongest (or tied-best) test accuracy; thus, we use depth =   4 as the default setting throughout this paper. Table A3 further reports aggregate statistics (mean ± std and minimum RMSE over the 11 subsets) to quantify how consistently each depth performs on average, whereas Table A4 provides the complete subset-by-depth RMSE matrix (val/test) as a full experimental record, enabling verification of which depth wins for each subset and demonstrating that the ablation was exhaustively evaluated rather than selectively reported.
Table A2. Effect of MLP depth (number of hidden layers) in the diffusion weighting network f θ . For each depth, we report the best configuration selected on meta-validation RMSE and its test performance. Best values are highlighted in bold.
Table A2. Effect of MLP depth (number of hidden layers) in the diffusion weighting network f θ . For each depth, we report the best configuration selected on meta-validation RMSE and its test performance. Best values are highlighted in bold.
MLP DepthVal-Best SubsetVal RMSE (°C)Test RMSE (°C)Test MAE (°C)Test R 2
2I + P + L0.35480.35970.26210.6899
4I + P + L0.35460.35900.26130.6911
8D + I + P + L0.35470.35920.26140.6909
16I + P + L0.35480.35990.26210.6896
Table A3. Aggregate performance across all 11 ensemble subsets ( K { 2 ,   3 ,   4 } ). We report mean ± std and minimum RMSE (°C) over the 11 subsets for each MLP depth. Lower is better.
Table A3. Aggregate performance across all 11 ensemble subsets ( K { 2 ,   3 ,   4 } ). We report mean ± std and minimum RMSE (°C) over the 11 subsets for each MLP depth. Lower is better.
MLP DepthMeta-Val RMSETest RMSE
Mean ± StdMinMean ± StdMin
20.3571 ± 0.00290.35480.3634 ± 0.00540.3592
40.3571 ± 0.00290.35460.3633 ± 0.00550.3590
80.3571 ± 0.00280.35470.3635 ± 0.00540.3592
160.3572 ± 0.00290.35480.3637 ± 0.00530.3596
Table A4. Full results of the diffusion-weighting ablation across all 11 ensemble subsets ( K { 2 ,   3 ,   4 } ) and all MLP depths. Each cell reports val/test RMSE (°C), and the best test RMSE per row is bolded.
Table A4. Full results of the diffusion-weighting ablation across all 11 ensemble subsets ( K { 2 ,   3 ,   4 } ) and all MLP depths. Each cell reports val/test RMSE (°C), and the best test RMSE per row is bolded.
KEnsemble SubsetDepth = 2Depth = 4Depth = 8Depth = 16
2D + L0.3559/0.35920.3559/0.35930.3559/0.35930.3559/0.3596
2D + P0.3609/0.37170.3609/0.37170.3607/0.37140.3611/0.3713
2D + I0.3625/0.37030.3624/0.37030.3625/0.37060.3626/0.3713
2P + L0.3551/0.35990.3550/0.35980.3551/0.36010.3551/0.3603
2I + L0.3552/0.35960.3551/0.35950.3552/0.35950.3554/0.3600
2I + P0.3595/0.36940.3595/0.36950.3594/0.36970.3597/0.3696
3D + P + L0.3551/0.35980.3550/0.35930.3551/0.35990.3551/0.3600
3D + I + L0.3551/0.35920.3551/0.35920.3552/0.35940.3553/0.3597
3D + I + P0.3595/0.36910.3596/0.36940.3593/0.36940.3597/0.3693
3I + P + L0.3548/0.35970.3546/0.35900.3548/0.35980.3548/0.3599
4D + I + P + L0.3548/0.35950.3546/0.35910.3547/0.35920.3548/0.3600

Appendix A.3. Sensitivity to the Noise Schedule

In practice, β τ is set to a monotone increasing schedule over τ = 0 , , T 1 . To assess sensitivity to this design choice, we consider five schedule families (linear, cosine, quadratic, sigmoid, and exponential) while keeping ( β start , β end , T ) fixed. Following our implementation, all schedules are mapped to the same range [ β start , β end ] for a fair comparison. Table A5 summarizes the aggregate performance across all 11 ensemble subsets ( K { 2 ,   3 ,   4 } ), while Table A6 reports the best-case subset under each schedule. For completeness, Table A7 provides the full subset-by-schedule RMSE matrix. In addition, the mathematical definitions of the different scheduling functions are provided as follows. Let τ { 0 , , T 1 } and define the normalized diffusion time
u = τ T 1 .
We consider five monotone schedule families for β τ . For a fair comparison, all schedules share the same endpoints ( β start , β end ) and the same number of steps T.
Linear : β τ = β start + ( β end β start ) u ,
Quadratic : β τ = β start + ( β end β start ) u 2 ,
Exponential : β τ = β start β end β start u ,
Sigmoid : g ( u ) = sigmoid κ ( u 1 2 ) , β τ = β start + ( β end β start ) g ( u ) g ( 0 ) g ( 1 ) g ( 0 ) ,
Cosine : α ¯ ( u ) = cos 2 π 2 · u + ϵ 1 + ϵ , α ¯ τ = α ¯ τ T 1 .
For the cosine family, we first derive an unnormalized β ˜ τ from α ¯ τ using
β ˜ τ = 1 α ¯ τ α ¯ τ 1 ( τ 1 ) , β ˜ 0 = β ˜ 1 .
We then linearly rescale β ˜ τ to match the common endpoint range:
β τ = β start + ( β end β start ) · β ˜ τ min τ β ˜ τ max τ β ˜ τ min τ β ˜ τ + δ ,
where δ is a small constant for numerical stability. Unless stated otherwise, we use T = 50 , β start = 10 4 , and β end = 0.02 .
Table A5. Schedule sensitivity across all 11 ensemble subsets ( K { 2 ,   3 ,   4 } ) with MLP depth = 4 . We report test RMSE (°C) mean ± std and the min/max over subsets. Lower is better.
Table A5. Schedule sensitivity across all 11 ensemble subsets ( K { 2 ,   3 ,   4 } ) with MLP depth = 4 . We report test RMSE (°C) mean ± std and the min/max over subsets. Lower is better.
ScheduleMean ± StdMinMax
Exponential0.362844 ± 0.0055770.3584430.371630
Cosine0.362955 ± 0.0056540.3582980.371515
Quadratic0.363018 ± 0.0055170.3587910.371663
Sigmoid0.363220 ± 0.0055380.3587640.371615
Linear0.363314 ± 0.0053620.3591250.371676
Table A6. Best-case subset under each schedule (MLP depth = 4 ), selected by the lowest test RMSE (°C).
Table A6. Best-case subset under each schedule (MLP depth = 4 ), selected by the lowest test RMSE (°C).
ScheduleBest Subset (K)Test RMSE (°C)Test MAE (°C)
CosineD + I + P + L (4)0.3582980.261126
ExponentialD + I + L (3)0.3584430.261240
SigmoidD + I + L (3)0.3587640.261607
QuadraticI + P + L (3)0.3587910.261433
LinearD + I + L (3)0.3591250.261746
Table A7. Full RMSE (°C) matrix for schedule sensitivity (MLP depth = 4 ). Each column highlights the best subset for that schedule.
Table A7. Full RMSE (°C) matrix for schedule sensitivity (MLP depth = 4 ). Each column highlights the best subset for that schedule.
SubsetLinearCosineQuadraticSigmoidExponential
D + L0.3593260.3590680.3589820.3592330.359073
D + P0.3716760.3715150.3716630.3716150.371630
D + I0.3702890.3700740.3699750.3703470.369631
P + L0.3598410.3599190.3598030.3597620.359864
I + L0.3596510.3589990.3590890.3593640.358720
I + P0.3706180.3701900.3702900.3706260.370122
D + P + L0.3592540.3591680.3589470.3591230.358608
D + I + L0.3591250.3587910.3589630.3587640.358443
D + I + P0.3700060.3694840.3694840.3696520.369384
I + P + L0.3594720.3590430.3587910.3590530.358725
D + I + P + L0.3592270.3582980.3588300.3587810.358653

Appendix A.4. Best Ensemble Configurations

We report (i) the best ensemble configuration selected using validation RMSE (leak-free selection) and its corresponding test performance (Table A8), and (ii) the oracle best configuration by test RMSE for reference (Table A9). All configurations are selected from the candidate set that includes the following aggregation rules: uniform mean (mean), validation-optimized convex weights (cw), Bayesian model averaging (bma), quantile regression forests (qrf), and the proposed diffusion-weighted ensemble (diff). The “validation-selected” configuration is the only deployable choice under a proper generalization protocol. The “oracle” configuration is included for comparison only and must not be interpreted as a valid selection rule.
Table A8. Best ensemble configuration selected by validation RMSE (standardized).
Table A8. Best ensemble configuration selected by validation RMSE (standardized).
SubsetKRuleVal RMSE (z)Test RMSE (z)Test R 2 Test RMSE (°C)
D + I + P + L4diff0.61150.61880.69180.3586
This follows the leak-free protocol: we select the subset+aggregation rule using validation only, then report the corresponding test performance once.
Table A9. Oracle best ensemble configuration by test RMSE (standardized).
Table A9. Oracle best ensemble configuration by test RMSE (standardized).
SubsetKRuleVal RMSE (z)Test RMSE (z)Test R 2 Test RMSE (°C)
D + I + L3diff0.61260.61850.69210.3584
Shown for reference only. Selecting a configuration by test performance is not permitted for proper generalization evaluation.

Appendix A.5. All Subset and Aggregation Combinations

Table A10 and Table A11 enumerate all non-empty subsets of { D , I , P , L } and all aggregation rules (uniform mean/convex weights/Bayesian model averaging/quantile regression forests/diffusion-weighted). This complete listing enables direct comparison between subset choice and aggregation strategy under the same evaluation pipeline.
Table A10. Complete results for all subset and aggregation-rule combinations (deep-learning ensembles only) (Part 1/2).
Table A10. Complete results for all subset and aggregation-rule combinations (deep-learning ensembles only) (Part 1/2).
SubsetKRuleVal RMSE (z)Val MAE (z)Val R 2 Test RMSE (z)Test MAE (z)Test R 2 Test RMSE (°C)
D1single0.63310.45100.64690.64780.47460.66220.375
L1single0.61560.44060.66620.62310.45330.68740.361
P1single0.62540.44750.65540.64490.47150.66520.374
I1single0.63030.45050.65000.64720.47570.66280.375
D + L2cw0.61410.43820.66780.62050.45180.69000.360
D + L2diff0.61440.43860.66740.61980.45150.69070.359
D + L2mean0.61660.43910.66500.62380.45530.68670.362
D + L2bma0.61510.43910.66670.62240.45330.68830.361
D + L2qrf0.61650.43950.66510.62460.45560.68590.362
D + P2cw0.62300.44480.65810.64070.46870.66960.371
D + P2diff0.62370.44540.65730.64280.47000.66720.372
D + P2mean0.62360.44490.65740.64040.46870.66990.371
D + P2bma0.62500.44600.65580.64460.47150.66510.374
D + P2qrf0.62420.44530.65670.64220.46990.66780.372
D + I2cw0.62550.44590.65530.64070.47010.66950.371
D + I2diff0.62610.44640.65460.64170.47060.66830.372
D + I2mean0.62540.44550.65550.64040.46960.66990.371
D + I2bma0.63000.44930.65020.64710.47570.66290.375
D + I2qrf0.62670.44690.65390.64290.47160.66700.373
P + L2cw0.61280.43740.66930.62150.45220.68900.360
P + L2diff0.61310.43760.66890.62120.45230.68940.360
P + L2mean0.61360.43790.66830.62440.45490.68620.362
P + L2bma0.61690.44020.66470.62380.45330.68690.362
P + L2qrf0.61520.43890.66660.62500.45610.68550.362
I + L2cw0.61300.43660.66900.62110.45230.68950.360
I + L2diff0.61360.43710.66830.62200.45250.68840.360
I + L2mean0.61450.43750.66730.62400.45530.68660.362
I + L2bma0.61570.44060.66620.62300.45330.68750.361
I + L2qrf0.61520.43810.66670.62410.45610.68650.362
I + P2cw0.62060.44280.66070.63770.46680.67270.370
I + P2diff0.62060.44280.66070.63770.46690.67270.370
I + P2mean0.62070.44280.66070.63770.46680.67260.370
I + P2bma0.62520.44600.65560.64470.47150.66490.374
I + P2qrf0.62260.44400.65860.64070.46830.66960.371
Table A11. Complete results for all subset and aggregation-rule combinations (deep-learning ensembles only) (Part 2/2).
Table A11. Complete results for all subset and aggregation-rule combinations (deep-learning ensembles only) (Part 2/2).
SubsetKRuleVal RMSE (z)Val MAE (z)Val R 2 Test RMSE (z)Test MAE (z)Test R 2 Test RMSE (°C)
D + P + L3cw0.61270.43720.66930.62120.45200.68940.360
D + P + L3diff0.61240.43700.66950.62060.45210.69020.360
D + P + L3mean0.61570.43860.66600.62650.45730.68400.363
D + P + L3bma0.61580.44080.66610.62360.45330.68710.362
D + P + L3qrf0.61500.43870.66690.62330.45480.68750.361
D + I + L3cw0.61300.43660.66900.62100.45230.68950.360
D + I + L3diff0.61110.43570.67110.61840.45130.69210.358
D + I + L3mean0.61640.43840.66540.62610.45750.68450.363
D + I + L3bma0.61570.44060.66620.62300.45330.68750.361
D + I + L3qrf0.61510.43850.66670.62390.45560.68660.362
D + I + P3cw0.62050.44240.66090.63720.46640.67320.369
D + I + P3diff0.62060.44250.66070.63730.46640.67310.369
D + I + P3mean0.62110.44260.66020.63710.46650.67330.369
D + I + P3bma0.62990.44920.65030.64700.47570.66310.375
D + I + P3qrf0.62270.44410.65840.64050.46820.66990.371
I + P + L3cw0.61220.43600.66980.62090.45210.68960.360
I + P + L3diff0.61180.43580.67020.61940.45170.69130.359
I + P + L3mean0.61380.43700.66820.62540.45640.68510.362
I + P + L3bma0.61570.44060.66620.62300.45330.68750.361
I + P + L3qrf0.61510.43840.66670.62390.45560.68660.362
D + I + P + L4cw0.61220.43600.66980.62090.45210.68960.360
D + I + P + L4diff0.61120.43570.67100.62060.45190.68990.360
D + I + P + L4mean0.61550.43800.66620.62730.45830.68330.364
D + I + P + L4bma0.61570.44060.66620.62300.45330.68750.361
D + I + P + L4qrf0.61410.43710.66790.62500.45400.68540.362

References

  1. Deser, C.; Alexander, M.A.; Xie, S.P.; Phillips, A.S. Sea surface temperature variability: Patterns and mechanisms. Annu. Rev. Mar. Sci. 2010, 2, 115–143. [Google Scholar] [CrossRef]
  2. Storto, A.; Alvera-Azcárate, A.; Balmaseda, M.A.; Barth, A.; Chevallier, M.; Counillon, F.; Domingues, C.M.; Drevillon, M.; Drillet, Y.; Forget, G.; et al. Ocean reanalyses: Recent advances and unsolved challenges. Front. Mar. Sci. 2019, 6, 418. [Google Scholar] [CrossRef]
  3. Sun, D.; Li, F.; Jing, Z.; Hu, S.; Zhang, B. Frequent marine heatwaves hidden below the surface of the global ocean. Nat. Geosci. 2023, 16, 1099–1104. [Google Scholar] [CrossRef]
  4. Ham, Y.G.; Kim, J.H.; Luo, J.J. Deep learning for multi-year ENSO forecasts. Nature 2019, 573, 568–572. [Google Scholar] [CrossRef]
  5. Tian, F.; Ren, H.L.; Liu, M.; Su, B.; Wang, R. Intensity and timing of persistence barriers of global sea surface temperature anomalies. Geosci. Lett. 2023, 10, 16. [Google Scholar] [CrossRef]
  6. Cluett, A.; Jacox, M.; Amaya, D.; Alexander, M.; Scott, J. Atmospheric precursors of skillful SST prediction in the Northeast Pacific. J. Clim. 2024, 37, 5337–5353. [Google Scholar] [CrossRef]
  7. Sengupta, A.; Waliser, D.E.; DeFlorio, M.J.; Guan, B.; Delle Monache, L.; Ralph, F.M. Role of evolving sea surface temperature modes of variability in improving seasonal precipitation forecasts. Commun. Earth Environ. 2025, 6, 256. [Google Scholar] [CrossRef] [PubMed]
  8. Yang, Z.; Liu, J.; Yang, C.Y.; Hu, Y. Correcting Nonstationary Sea Surface Temperature Bias in NCEP CFSv2 Using Ensemble-Based Neural Networks. J. Atmos. Ocean. Technol. 2023, 40, 885–896. [Google Scholar] [CrossRef]
  9. Montero-Manso, P.; Hyndman, R.J. Principles and algorithms for forecasting groups of time series: Locality and globality. Int. J. Forecast. 2021, 37, 1632–1653. [Google Scholar] [CrossRef]
  10. Wen, Q.; Zhou, T.; Zhang, C.; Chen, W.; Ma, Z.; Yan, J.; Sun, L. Transformers in time series: A survey. arXiv 2022, arXiv:2202.07125. [Google Scholar]
  11. De León Pérez, D.; Salazar-Galán, S.; Francés, F. Beyond Deterministic Forecasts: A Scoping Review of Probabilistic Uncertainty Quantification in Short-to-Seasonal Hydrological Prediction. Water 2025, 17, 2932. [Google Scholar] [CrossRef]
  12. Ehmimed, N.; Chkouri, M.Y.; Touhafi, A. Reliable and Adaptive Probabilistic Forecasting for Event-Driven Water-Quality Time Series Using a Gated Hybrid–Mixture Density Network. Sensors 2025, 25, 7560. [Google Scholar] [CrossRef]
  13. Li, L.; Carver, R.; Lopez-Gomez, I.; Sha, F.; Anderson, J. Generative emulation of weather forecast ensembles with diffusion models. Sci. Adv. 2024, 10, eadk4489. [Google Scholar] [CrossRef]
  14. Chen, F.; Gao, L. Learning Residual Distributions with Diffusion Models for Probabilistic Wind Power Forecasting. Energies 2025, 18, 4226. [Google Scholar] [CrossRef]
  15. Su, C.; Cai, Z.; Tian, Y.; Chang, Z.; Zheng, Z.; Song, Y. Diffusion models for time series forecasting: A survey. arXiv 2025, arXiv:2507.14507. [Google Scholar] [CrossRef]
  16. Barreto, F.T.; Curbani, F.E.; Zielinsky, G.M.; da Silva, M.B.; Lacerda, K.C.; Rodrigues, D.F. Development of a multigrid operational forecast system for the oceanic region off Rio de Janeiro State. Ocean Model. 2023, 184, 102206. [Google Scholar]
  17. Storto, A.; Frolov, S.; Slivinski, L.; Yang, C. Correction of sea surface biases in the NEMO ocean general circulation model using neural networks. Geosci. Model Dev. 2025, 18, 4789–4804. [Google Scholar] [CrossRef]
  18. Kong, W.; Lam, C.c.; Lau, D.s.; Chow, C.k.; Chong, S.n.; Chan, P.w.; Leung, N.c. Model validation and applications of wave and current forecasts from the Hong Kong Observatory’s Operational Marine Forecasting System. Ocean Model. 2024, 190, 102393. [Google Scholar] [CrossRef]
  19. de Boisséson, E.; Balmaseda, M.A. Predictability of marine heatwaves: Assessment based on the ECMWF seasonal forecast system. Ocean Sci. 2024, 20, 265–278. [Google Scholar] [CrossRef]
  20. Koul, V.; Brune, S.; Akimova, A.; Düsterhus, A.; Pieper, P.; Hövel, L.; Parekh, A.; Schrum, C.; Baehr, J. Seasonal prediction of Arabian Sea marine heatwaves. Geophys. Res. Lett. 2023, 50, e2023GL103975. [Google Scholar] [CrossRef]
  21. Hao, P.; Li, S.; Song, J.; Gao, Y. Prediction of sea surface temperature in the South China Sea based on deep learning. Remote. Sens. 2023, 15, 1656. [Google Scholar] [CrossRef]
  22. Xu, T.; Zhou, Z.; Li, Y.; Wang, C.; Liu, Y.; Rong, T. Short-term prediction of global sea surface temperature using deep learning networks. J. Mar. Sci. Eng. 2023, 11, 1352. [Google Scholar] [CrossRef]
  23. Zrira, N.; Kamal-Idrissi, A.; Farssi, R.; Khan, H.A. Time series prediction of sea surface temperature based on BiLSTM model with attention mechanism. J. Sea Res. 2024, 198, 102472. [Google Scholar] [CrossRef]
  24. Shi, B.; Ge, C.; Lin, H.; Xu, Y.; Tan, Q.; Peng, Y.; He, H. Sea surface temperature prediction using convlstm-based model with deformable attention. Remote. Sens. 2024, 16, 4126. [Google Scholar] [CrossRef]
  25. Fu, Y.; Song, J.; Guo, J.; Fu, Y.; Cai, Y. Prediction and analysis of sea surface temperature based on LSTM-transformer model. Reg. Stud. Mar. Sci. 2024, 78, 103726. [Google Scholar] [CrossRef]
  26. Yang, Y.; Lam, K.M.; Dong, J.; Ju, Y. Multi-Factor Deep Learning Model for Sea Surface Temperature Forecasting. Remote Sens. 2025, 17, 752. [Google Scholar] [CrossRef]
  27. Dai, H.; Lei, F.; Wei, G.; Zhang, X.; Lin, R.; Zhang, W.; Shang, S. Sea surface temperature prediction by stacked generalization ensemble of deep learning. Deep Sea Res. Part I Oceanogr. Res. Pap. 2024, 209, 104343. [Google Scholar] [CrossRef]
  28. Qian, J.; Wang, Q.; Liang, P.; Peng, S.; Wang, H.; Wu, Y. Deep learning–based ensemble forecast and predictability analysis of the Kuroshio intrusion into the south China sea. J. Phys. Oceanogr. 2024, 54, 1503–1517. [Google Scholar] [CrossRef]
  29. Bonino, G.; Galimberti, G.; Masina, S.; McAdam, R.; Clementi, E. Machine learning methods to predict sea surface temperature and marine heatwave occurrence: A case study of the Mediterranean Sea. Ocean Sci. 2024, 20, 417–432. [Google Scholar] [CrossRef]
  30. Boschetti, F.; Feng, M.; Hartog, J.R.; Hobday, A.J.; Zhang, X. Sea surface temperature predictability assessment with an ensemble machine learning method using climate model simulations. Deep Sea Res. Part II Top. Stud. Oceanogr. 2023, 210, 105308. [Google Scholar] [CrossRef]
  31. Behringer, D.W.; Ji, M.; Leetmaa, A. An improved coupled model for ENSO prediction and implications for ocean initialization. Part I: The ocean data assimilation system. Mon. Weather. Rev. 1998, 126, 1013–1021. [Google Scholar] [CrossRef]
  32. Zuo, H.; Balmaseda, M.A.; Tietsche, S.; Mogensen, K.; Mayer, M. The ECMWF operational ensemble reanalysis–analysis system for ocean and sea ice: A description of the system and assessment. Ocean Sci. 2019, 15, 779–808. [Google Scholar] [CrossRef]
  33. Jean-Michel, L.; Eric, G.; Romain, B.B.; Gilles, G.; Angélique, M.; Marie, D.; Clément, B.; Mathieu, H.; Olivier, L.G.; Charly, R.; et al. The Copernicus global 1/12 oceanic and sea ice GLORYS12 reanalysis. Front. Earth Sci. 2021, 9, 698876. [Google Scholar] [CrossRef]
  34. Breiman, L. Random Forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  35. Smola, A.J.; Schölkopf, B. A tutorial on support vector regression. Stat. Comput. 2004, 14, 199–222. [Google Scholar] [CrossRef]
  36. Akiba, T.; Sano, S.; Yanase, T.; Ohta, T.; Koyama, M. Optuna: A Next-generation Hyperparameter Optimization Framework. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining (KDD ’19), Anchorage, AK, USA, 4–8 August 2019; pp. 2623–2631. [Google Scholar] [CrossRef]
  37. Ho, J.; Jain, A.; Abbeel, P. Denoising Diffusion Probabilistic Models. arXiv 2020, arXiv:2006.11239. [Google Scholar] [CrossRef]
  38. Karras, T.; Aittala, M.; Aila, T.; Laine, S. Elucidating the Design Space of Diffusion-Based Generative Models. In Proceedings of the Advances in Neural Information Processing Systems (NeurIPS), New Orleans, LO, USA, 28 November–9 December 2022. [Google Scholar]
  39. Lu, C.; Zhou, Y.; Bao, F.; Chen, J.; Li, C.; Zhu, J. DPM-Solver: A Fast ODE Solver for Diffusion Probabilistic Model Sampling in Around 10 Steps. In Proceedings of the Advances in Neural Information Processing Systems (NeurIPS), New Orleans, LO, USA, 28 November–9 December 2022. [Google Scholar]
  40. Song, Y.; Dhariwal, P.; Chen, M.; Sutskever, I. Consistency Models. In Proceedings of the International Conference on Machine Learning (ICML), Honolulu, HI, USA, 23–29 July 2023. [Google Scholar]
  41. Liu, X.; Gong, C.; Liu, Q. Flow Straight and Fast: Learning to Generate and Transfer Data with Rectified Flow. arXiv 2022, arXiv:2209.03003. [Google Scholar] [CrossRef]
Figure 1. Global map of near-surface potential temperature from GODAS.
Figure 1. Global map of near-surface potential temperature from GODAS.
Mathematics 14 00740 g001
Figure 2. Overview of the deep-learning ensemble pipeline employing noise-conditioned weighting without reverse sampling. Base forecasters (DLinear, iTransformer, PatchTST, LSTM) generate out-of-sample predictions on the validation split ( P val ). Ensemble parameters and the best subset/aggregation rule are selected using a meta-train/meta-validation protocol within validation only, and the selected ensemble is then applied once to the test predictions ( P test ) for final evaluation.
Figure 2. Overview of the deep-learning ensemble pipeline employing noise-conditioned weighting without reverse sampling. Base forecasters (DLinear, iTransformer, PatchTST, LSTM) generate out-of-sample predictions on the validation split ( P val ). Ensemble parameters and the best subset/aggregation rule are selected using a meta-train/meta-validation protocol within validation only, and the selected ensemble is then applied once to the test predictions ( P test ) for final evaluation.
Mathematics 14 00740 g002
Figure 3. Architecture of the proposed diffusion-weighted ensemble. Unlike standard diffusion models, the network directly predicts mixing weights conditioned on noise levels, bypassing iterative reverse sampling. Given the clean base prediction vector x 0 , we apply forward noising to obtain x τ and use a noise-conditioned weighting network f θ ( [ x τ ; e ( τ ) ] ) to produce convex weights via softmax. At inference, Monte-Carlo averaging over multiple ( τ , ϵ ) draws yields the final weights w ¯ , which are applied to x 0 to compute the final forecast y ^ diff = w ¯ x 0 .
Figure 3. Architecture of the proposed diffusion-weighted ensemble. Unlike standard diffusion models, the network directly predicts mixing weights conditioned on noise levels, bypassing iterative reverse sampling. Given the clean base prediction vector x 0 , we apply forward noising to obtain x τ and use a noise-conditioned weighting network f θ ( [ x τ ; e ( τ ) ] ) to produce convex weights via softmax. At inference, Monte-Carlo averaging over multiple ( τ , ϵ ) draws yields the final weights w ¯ , which are applied to x 0 to compute the final forecast y ^ diff = w ¯ x 0 .
Mathematics 14 00740 g003
Figure 4. Test RMSE by latitude band (10° bins). The Diffusion-weighted ensemble is compared against the best single model (LSTM) to reveal latitude-dependent performance differences.
Figure 4. Test RMSE by latitude band (10° bins). The Diffusion-weighted ensemble is compared against the best single model (LSTM) to reveal latitude-dependent performance differences.
Mathematics 14 00740 g004
Table 1. All oceanographic variables are provided by the GODAS reanalysis dataset.
Table 1. All oceanographic variables are provided by the GODAS reanalysis dataset.
CategoryVariableDescription
TemperaturePotential temperatureOcean potential temperature is provided at multiple vertical levels
SalinitySalinityOcean salinity field provided at multiple
vertical levels
Ocean currentsU of currentZonal (east-west) component of ocean
current velocity
V of currentMeridional (north–south) component of ocean
current velocity
Vertical motionGeometric vertical velocityVertical velocity component of ocean flow
Sea levelSea surface height relative to geoidSea surface height referenced to the geoid
Vertical structureOcean mixed layer depth below sea surfaceDepth of the surface ocean mixed layer
Ocean isothermal layer depth below sea surfaceDepth of the isothermal layer below the sea surface
Surface forcingTotal downward heat flux at surfaceNet downward heat flux at the ocean surface
Zonal momentum fluxZonal component of surface momentum flux
Meridional momentum fluxMeridional component of surface momentum flux
Salt fluxSurface salt flux at the ocean–atmosphere interface
Table 2. Oceanographic variables selected from the GODAS dataset and used in this study.
Table 2. Oceanographic variables selected from the GODAS dataset and used in this study.
CategoryVariableDescription
TemperaturePotential temperature (uppermost level)Near-surface ocean potential temperature used as a proxy for sea surface temperature
Ocean currentsU of currentZonal (east-west) component of ocean current velocity
V of currentMeridional (north-south) component of ocean
current velocity
Vertical structureOcean mixed layer depth below sea surfaceDepth of the surface ocean mixed layer
Ocean isothermal layer depth below sea surfaceDepth of the isothermal layer below the sea surface
Sea levelSea surface height relative to geoidSea surface height referenced to the geoid
Surface forcingTotal downward heat flux at the surfaceNet downward heat flux at the ocean surface
Table 3. Test performance of individual forecasting models. RMSE is reported in standardized units (std) and in physical units (°C). Best values (lowest error/highest R 2 ) are highlighted in bold.
Table 3. Test performance of individual forecasting models. RMSE is reported in standardized units (std) and in physical units (°C). Best values (lowest error/highest R 2 ) are highlighted in bold.
ModelRMSE (std)RMSE (°C)MAE (°C) R 2
DLinear0.64780.37550.27510.6622
iTransformer0.64720.37510.27310.6628
PatchTST0.64490.37380.27330.6652
LSTM0.62310.36120.26270.6874
LinearSVR0.64200.37200.26830.6648
RF0.64000.37100.26200.6672
Table 4. Test performance of ensemble methods. For each aggregation rule, we report the test performance of the best-performing subset selected on the meta-validation split. Best values are highlighted in bold.
Table 4. Test performance of ensemble methods. For each aggregation rule, we report the test performance of the best-performing subset selected on the meta-validation split. Best values are highlighted in bold.
MethodRMSE (std)RMSE (°C)MAE (°C) R 2
Single (LSTM)0.62310.36120.26270.6874
Ensemble—Mean0.62730.36360.26560.6833
Ensemble—Convex Weight0.62090.35990.26200.6896
Ensemble—BMA0.62310.36120.26270.6874
Ensemble—QRF0.63210.36640.27040.6770
Ensemble—Diffusion-weighted0.61880.35860.26120.6918
Table 5. Band-wise test RMSE by latitude. Imp ( % ) denotes the relative difference in the Diffusion-weighted ensemble over LSTM: ( RMSE LSTM RMSE Diff ) / RMSE LSTM × 100 . we use half-open intervals [a,b).
Table 5. Band-wise test RMSE by latitude. Imp ( % ) denotes the relative difference in the Diffusion-weighted ensemble over LSTM: ( RMSE LSTM RMSE Diff ) / RMSE LSTM × 100 . we use half-open intervals [a,b).
Lat. CenterRangeNRMSE (LSTM)RMSE (Diff)Imp (%)
−70[−75,−65)34820.23620.23301.3396
−60[−65,−55)62520.26880.26640.8983
−50[−55,−45)63250.27980.27880.3412
−40[−45,−35)58140.37110.3718−0.1836
−30[−35,−25)51130.35990.3618−0.5415
−20[−25,−15)49000.31730.3192−0.5941
−10[−15,−5)46540.33220.3356−1.0308
0[−5,5)47390.40280.40030.6308
10[5,15)47520.34750.34680.2088
20[15,25)41360.32910.3292−0.0230
30[25,35)36610.43010.42561.0400
40[35,45)31430.59530.57972.6149
50[45,55)25940.43520.43160.8316
60[55,65)18610.41050.40251.9405
Table 6. Best ensemble configurations. The left row is selected by the lowest validation RMSE (proper selection). The right row shows the post hoc best test RMSE (for reference) and is not a valid selection method. Here, D = DLinear, I = iTransformer, P = PatchTST, and L = LSTM.
Table 6. Best ensemble configurations. The left row is selected by the lowest validation RMSE (proper selection). The right row shows the post hoc best test RMSE (for reference) and is not a valid selection method. Here, D = DLinear, I = iTransformer, P = PatchTST, and L = LSTM.
SelectionSubset + AggregationVal RMSE (°C)Test RMSE (°C)Test MAE (°C)Test R 2
Val-bestI + P + L; Diffusion0.35480.35970.26210.6899
Test-best (post hoc)D + L; Diffusion0.35590.35920.26160.6909
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yu, G.; Choi, G.; Choi, M.; Min, S.-h.; Kim, Y. A Diffusion Weighted Ensemble Framework for Robust Short-Horizon Global SST Forecasting from Multivariate GODAS Data. Mathematics 2026, 14, 740. https://doi.org/10.3390/math14040740

AMA Style

Yu G, Choi G, Choi M, Min S-h, Kim Y. A Diffusion Weighted Ensemble Framework for Robust Short-Horizon Global SST Forecasting from Multivariate GODAS Data. Mathematics. 2026; 14(4):740. https://doi.org/10.3390/math14040740

Chicago/Turabian Style

Yu, Gwangun, GilHan Choi, Moonseung Choi, Sun-hong Min, and Yonggang Kim. 2026. "A Diffusion Weighted Ensemble Framework for Robust Short-Horizon Global SST Forecasting from Multivariate GODAS Data" Mathematics 14, no. 4: 740. https://doi.org/10.3390/math14040740

APA Style

Yu, G., Choi, G., Choi, M., Min, S.-h., & Kim, Y. (2026). A Diffusion Weighted Ensemble Framework for Robust Short-Horizon Global SST Forecasting from Multivariate GODAS Data. Mathematics, 14(4), 740. https://doi.org/10.3390/math14040740

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop