Next Article in Journal
Motion Error of V-Type Roller Linear Guides Considering Geometric Errors of the Guide Rails
Next Article in Special Issue
MMHFormer: Multi-Source and Multi-View Hierarchical Transformer for Traffic Flow Prediction
Previous Article in Journal
Taxonomic and Genomic Characterization of Brevibacillus sp. JNUCC 42 from Baengnokdam Crater Lake, Mt. Halla, and Its Cosmeceutical Potential
Previous Article in Special Issue
A Novel Hybrid Framework for Stock Price Prediction Integrating Adaptive Signal Decomposition and Multi-Scale Feature Extraction
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

KOSLM: A Kalman-Optimal Hybrid State-Space Memory Network for Long-Term Time Series Forecasting

School of Electronic Information and Artificial Intelligence, Shaanxi University of Science and Technology, Xi’an 710021, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(23), 12684; https://doi.org/10.3390/app152312684
Submission received: 23 October 2025 / Revised: 24 November 2025 / Accepted: 24 November 2025 / Published: 29 November 2025
(This article belongs to the Special Issue Advanced Methods for Time Series Forecasting)

Abstract

Long-term time series forecasting (LTSF) remains challenging, as models must capture long-range dependencies and remain robust to noise accumulation. Traditional recurrent models, such as Recurrent Neural Networks (RNNs) and Long Short-Term Memory (LSTM), often suffer from instability and information degradation over extended horizons. The state-of-the-art method xLSTMTime improves memory retention through exponential gating and enhanced memory-transition rules, but it still lacks principled guidance. To address these issues, we propose the Kalman-Optimal Selective Long-Term Memory (KOSLM) model, which embeds a Kalman-optimal selective mechanism driven by the innovation signal within a structured state-space reformulation of LSTM. KOSLM dynamically regulates information propagation and forgetting to minimize state estimation uncertainty, providing both theoretical interpretability and practical efficiency. Extensive experiments across energy, finance, traffic, healthcare, and meteorology datasets show that KOSLM reduces mean squared error (MSE) by 14.3–38.9% compared with state-of-the-art methods, with larger gains at longer horizons. The model is lightweight, scalable, and achieves up to 2.5× speedup over Mamba-2. Beyond benchmarks, KOSLM is further validated on real-world Secondary Surveillance Radar (SSR) tracking under noisy and irregular sampling, demonstrating robust and generalizable long-term forecasting performance.

1. Introduction

Long-term time series forecasting (LTSF) aims to predict future values over extended horizons based on historical observations. Accurate long-range forecasts are critical for applications such as energy scheduling, climate modeling, financial planning, traffic management, and healthcare resource allocation. LTSF presents unique challenges: capturing long-range dependencies, mitigating error accumulation, and adapting to non-stationary temporal dynamics [1].
Traditional recurrent architectures, including Recurrent Neural Networks (RNNs) [2] and Long Short-Term Memory (LSTM) networks [3,4], often struggle with long sequences due to vanishing or exploding gradients [5]. While LSTM gates alleviate some of these issues, their heuristic design lacks explicit structural constraints derived from a principled optimality criterion, potentially leading to suboptimal memory retention over extended horizons. The recently proposed xLSTMTime further enhances long-range modeling by introducing the sLSTM and mLSTM components. The sLSTM employs scalar memory with exponential gating to manage long-term dependencies, while the mLSTM component uses matrix memory and a covariance update rule to enhance storage capacity and relevant information retrieval capabilities. Despite these improvements, xLSTMTime remains constrained by heuristic recurrent design.
The Kalman filter (KF) [6] provides optimal state estimation under Gaussian noise [7]. Recent studies have explored integrating KF with LSTM to improve the accuracy of time-series forecasting. Representative approaches include the following: (i) Deep Kalman Filters [8], which parameterize the state-transition and observation functions of KF using LSTM; (ii) KalmanNet [9], which employs an LSTM to learn residual corrections to the Kalman gain under a known KF model; and (iii) uncertainty-aware LSTM–KF hybrids [10], which estimate the covariance or uncertainty structures of KF through recurrent dynamics. However, these approaches generally maintain a loose coupling between LSTM and KF—they do not embed Kalman-inspired feedback directly into the internal gating dynamics of LSTM.
Recently, selective state space models (SSMs) [11,12] have demonstrated efficient sequence modeling with linear-time complexity. By dynamically modulating SSM’s parameters based on the input, they can filter task-irrelevant patterns while retaining critical long-term information. This selective modulation property motivates revisiting LSTM gates from a state-space perspective, into which Kalman-inspired feedback can be injected, thereby endowing the gating mechanism with Kalman-optimal structural constraints.
In this work, we propose the Kalman-Optimal Selective Long-Term Memory (KOSLM) model, which establishes a context-aware feedback pathway that optimally balances memory retention and information updating, providing both theoretical interpretability and practical efficiency.
Our main contributions are as follows:
  • State-space reformulation of LSTM: We formalize LSTM networks as input- and state-dependent SSMs, where each gate dynamically parameterizes the state-transition and input matrices. This framework provides a principled explanation of LSTM’s long-term memory behavior.
  • Kalman-optimal selective gating: Inspired by the Kalman filter and selective SSMs, we introduce a Kalman-optimal selective mechanism in which the state-transition and input matrices are linearly modulated by a Kalman gain learned from the innovation term, establishing a feedback pathway that minimizes state estimation uncertainty.
  • Applications to real-world forecasting: KOSLM consistently outperforms state-of-the-art baselines across LTSF benchmarks in energy, finance, traffic, healthcare, and meteorology, achieving 14.3–38.9% lower mean squared error (MSE) and a maximum reduction in mean absolute error (MAE) of 25.2%; it also delivers up to 2.5× faster inference than Mamba-2. In real-world Secondary Surveillance Radar (SSR) tracking under noisy and irregular sampling, KOSLM demonstrates strong robustness and generalization.
By bridging heuristic LSTM gating with principled Kalman-optimal estimation, KOSLM provides a robust, interpretable, and scalable framework for long-term sequence modeling, offering both methodological novelty and practical forecasting utility.

2. Background and Theory

2.1. LSTM Networks

RNNs model sequential data through recursive temporal computation. However, traditional RNNs often suffer from vanishing and exploding gradients when capturing long-term dependencies. The LSTM network [3] addresses this issue by introducing gating mechanisms that regulate the flow of information over time. The network structure of the LSTM neuron is shown in Figure 1.
Each LSTM unit maintains a cell state C t that carries long-term information and a hidden state H t that provides short-term representations. The cell state is updated according to
C t = F t C t 1 + I t C ˜ t ,
where F t , I t , and O t denote the forget, input, and output gates, respectively. These gates are nonlinear functions of the current input x t and the previous hidden state H t 1 . The final output is obtained as follows:
H t = O t tanh ( C t ) .
This gating mechanism allows the model to selectively retain or discard information, mitigating gradient degradation. However, the gates are learned heuristically through data-driven optimization rather than derived from an explicit structural optimality constraint, making the LSTM sensitive to noise and unstable for long-term dependencies.

2.2. State Space Models

2.2.1. Selective State Space Models

Selective State Space Models (S6) are a recent class of sequence models for deep learning that are broadly related to RNNs and classical SSMs. They are inspired by a particular system (Equation (3)) that maps a one-dimensional function or sequence x t R y t R through an implicit latent state h t R N :
h t = A h t 1 + B x t ,
y t = M h t .
where A is the state transition matrix, defining how the past latent state h t 1 influences the current state; B is the input matrix, determining how x t modulates h t ; and M maps the implicit latent state h t to the output.
These models integrate the SSM described above into deep learning frameworks and introduce input-dependent selective mechanisms (see Appendix A for a detailed discussion), achieving Transformer-level modeling capability with linear computational complexity.
The theoretical connections among LSTM, KF, and SSM provide the foundation for constructing a unified Kalman-optimal selective memory framework.

2.2.2. Kalman Filter

The KF [6] is a classical instance of the state-space model, providing the minimum mean-square-error (MMSE) estimate of hidden system states under noisy observations. The system dynamics are expressed as
h t = A h t 1 + w t , w t N ( 0 , Q t ) ,
z t = M h t + v t , v t N ( 0 , R t ) ,
where M is the same as the M in the SSM introduced above, but in the context of the KF, it is commonly referred to as the observation matrix; w t and v t denote process and observation noise with covariances Q t and R t , respectively. Note that the observation z t can be regarded as the input x t in the SSMs.
At each time step, the KF performs two operations: prediction and update. The prediction step estimates the prior state h ^ t and prior error covariance P ^ t based on the posterior state h ^ t 1 and posterior error covariance P ^ t 1 from the previous time step:
h ^ t = A h ^ t 1 ,
P ^ t = A P ^ t 1 A T + Q t
The update step then refines this prediction using the observation z t :
K t = P ^ t M T ( M P ^ t M T + R t ) 1
h ^ t = h ^ t + K t ( z t M h ^ t )
P ^ t = ( I K t M ) P ^ t
Here, K t is the Kalman gain, which minimizes the posterior error covariance P ^ t . It determines how much new information (the innovation term  z t M h ^ t ) should be incorporated into the updated state ( h ^ t is the posterior state). This principle of optimal selective information integration forms the theoretical foundation for the innovation-driven gating design proposed later. A detailed justification for interpreting the Kalman gain as a prototype of dynamic selectivity is provided in Appendix B.1.

3. Proposed Method

3.1. Reformulating LSTM as a State-Space Model

Following the LSTM formulation in Equation (1), the LSTM cell can be equivalently reformulated as a time-varying SSM, where the cell state C t evolves under nonlinear, input- and state-dependent dynamics:
C t = A t C t 1 + B t z t ,
Here, z t serves as the observation in the KF (corresponding to the input x t in the SSM and LSTM), and the matrices A t and B t are determined by the forget gate F t and input gate I t , respectively. A detailed derivation of this LSTM-to-SSM reconstruction, including the mapping of gating mechanisms to state-space parameters, is provided in Appendix B.2.

3.2. Kalman-Optimal Selectivity via Innovation-Driven Gain

We introduce the innovation term from the KF:
Innov t = z t M t A t 1 C t 1 ,
which measures the discrepancy between the observation input z t and the predicted state based on the previous cell state C t 1 , serving as a real-time correction signal between model prediction and actual measurement.
In classical KF, this innovation drives the computation of the Kalman gain  K t (Equation (7)), which regulates the incorporation of new information and the retention of prior state during the update step. In KOSLM, rather than explicitly solving the Riccati recursion [13], we learn a functional mapping:
K t = ϕ ( Innov t ; θ ϕ ) ,
where ϕ ( · ) is a lightweight multi-layer perceptron (MLP) with sigmoid activation, with its width and depth controlled by two hyperparameters, MLP_d and MLP_layers (e.g., MLP_d = 64 , MLP_layers = 2 ), and parameterized by θ ϕ . The sigmoid activation ensures that the estimated gain K t ( 0 , 1 ) , maintaining physical interpretability as an adaptive weighting coefficient and preventing numerical instability.
The learned gain dynamically regulates the trade-off between prior memory and new information, yielding a learnable yet principled mechanism for Kalman-optimal selectivity. The bounded output range of K t further stabilizes the state-space update and constrains divergence during long-horizon inference. Appendix B.3 provides the theoretical derivation demonstrating that the innovation term serves as a sufficient statistic for learning the Kalman gain. Appendix B.4 further presents controlled experiments confirming that the learned gain accurately approximates the oracle Kalman gain across various ( Q , R ) regimes (Table A1). Controlled ablation studies (see Section 4.2.2 and Table 1) show that increasing the hidden layer size or depth yields only marginal improvements in predictive performance, indicating that KOSLM’s performance is largely insensitive to the specific configuration of ϕ . This confirms that the performance gains primarily stem from the Kalman-optimal structural design rather than the network capacity. These results jointly establish the theoretical and empirical foundation for embedding Kalman-optimal selectivity into deep learning models.
We then define the state-space evolution as follows:
A t = ( I K t M t ) A , B t = K t ,
where A is the base state transition matrix, serving as a learnable parameter of the model; A t and B t are dynamically modulated by the Kalman gain K t .

3.3. Structural Overview of KOSLM

The KOSLM cell preserves the computational efficiency of a standard LSTM while embedding a Kalman-inspired feedback loop. The KOSLM network architecture is illustrated in Figure 2; at each timestep, the operations are as follows:
  • Initialization: The learnable matrix A is initialized following S4D-Lin and for the real case is S4D-Real [14], which is based on the HIPPO theory [15]. These define the n-th element of A as 1 2 + n i and ( n + 1 ) respectively;
  • Compute z t and M t via the candidate gate C ˜ t and output gate O t ;
  • Compute the innovation: Innov t = z t M t A C t 1 ;
  • Estimate the Kalman gain: K t = ϕ ( Innov t ) ;
  • Update state transition matrices: A t = ( I K t M t ) A , B t = K t ;
  • Propagate the hidden state: C t = A t C t 1 + B t z t ;
  • Compute the output hidden representation: H t = M t C t .
To further clarify the conceptual and structural differences between KOSLM and existing Kalman-based neural architectures, Table 2 summarizes a direct comparison with the KalmanNet [9].

3.4. Theoretical Interpretation

Under linear–Gaussian assumptions and with a sufficiently expressive mapping ϕ , the learned gain K t converges to the oracle Kalman solution. Consequently, KOSLM inherits the nonlinear expressive power of LSTM while achieving the minimum-variance estimation property of the Kalman filter in its linear regime. This leads to improved stability and robustness, particularly in long-horizon or noisy sequence modeling.

3.5. Practical Advantages

KOSLM offers several practical benefits:
  • Robustness: The feedback structure mitigates error accumulation and improves performance under noise or distributional shifts.
  • Efficiency: With only 0.24 M parameters, KOSLM achieves up to 2.5× faster inference than Mamba-2, while maintaining competitive accuracy.
  • Versatility: The model generalizes across diverse domains, from energy demand forecasting to radar-based trajectory tracking.

4. Results

This section provides a comprehensive evaluation of the proposed KOSLM model through large-scale long-term forecasting benchmarks (Section 4.1), component ablation studies (Section 4.2), efficiency assessments (Section 4.3), and a real-world radar trajectory tracking case study (Section 4.4). The experimental analyses collectively aim to validate both the predictive accuracy and robustness of KOSLM across diverse and noisy temporal conditions.

4.1. Main Experiments on Benchmark Datasets

To examine the capability of KOSLM in modeling long-range dependencies and maintaining stability over extended forecasting horizons, we conduct systematic experiments on nine widely used real-world datasets covering domains such as traffic flow, electrical load, exchange rate, meteorology, and epidemiology. The datasets vary in frequency, dimensionality, and temporal regularity, providing a comprehensive benchmark for assessing model generalization.

4.1.1. Dataset Details

We summarize the datasets used in this study as follows. Weather [16] contains 21 meteorological variables (e.g., temperature and humidity) recorded every 10 min throughout 2020. ETT (Electricity Transformer Temperature) [17] includes four subsets: two hourly-level datasets (ETTh1, ETTh2) and two 15-minute-level datasets (ETTm1, ETTm2). Electricity [18], derived from the UCI Machine Learning Repository, records hourly power consumption (kWh) of 321 clients from 2012 to 2014. Exchange [19] comprises daily exchange rates among eight countries. Traffic [20] consists of hourly road occupancy rates measured by 862 sensors on San Francisco Bay Area freeways from January 2015 to December 2016. Illness (ILI) dataset [21] tracks the weekly number of influenza-like illness patients in the United States. Table 3 summarizes the statistical properties of all nine benchmark datasets. All datasets are divided into training, validation, and test subsets with a ratio of 7:1:2.

4.1.2. Implementation Details

All models are trained using the Adam optimizer without weight decay. The learning rate is selected from [ 1 × 10 3 , 1 × 10 2 ] via grid search. Batch size is set to 32 by default, adjustable up to 256 depending on GPU memory. Training is performed for 15 epochs, and the checkpoint with the lowest validation loss is used for testing. Experiments are repeated five times, and average results are reported (RMS values are provided in the Supplementary Materials). All models adopt a 2-layer architecture with hidden dimension 64. We use PyTorch’s default weight initialization, and no additional regularization (dropout or gradient clipping) is applied. All experiments are implemented in PyTorch 2.1.0 with Python 3.11 on NVIDIA RTX 4090 GPUs. Random seeds for Python, NumPy, and PyTorch are fixed to ensure reproducibility.
The gain function ϕ is implemented as a one-layer MLP with hidden dimension 64 and sigmoid activation, followed by a linear projection to the gain matrix K t . The same ϕ network is shared across all timesteps to ensure parameter consistency.

4.1.3. Experimental Setup of Main Experiments

All experiments follow the evaluation protocol established in xLSTMTime [22], adopting prediction horizons of T { 96 , 192 , 336 , 720 } for standard datasets and T { 24 , 36 , 48 , 60 } for the weekly-sampled ILI dataset. We compare the proposed KOSLM with nine recent state-of-the-art baselines that represent diverse architectural paradigms, spanning state-space, recurrent, attention-based, and linear modeling frameworks:
  • SSM-based: FiLM [23], S-Mamba [24];
  • LSTM-based: xLSTMTime [22];
  • Transformer-based: FEDformer [25], iTransformer [26], Crossformer [27];
  • MLP/TCN-based: DLinear [28], PatchTST [29], TimeMixer [30].
This comprehensive selection enables a fair and systematic comparison across diverse sequence modeling paradigms.

4.1.4. Overall Performance

Table 4 and Table 5 report the long-term multivariate forecasting results on nine real-world datasets, evaluated by MSE and MAE. Across nearly all datasets and prediction horizons, the proposed KOSLM achieves the best or near-best performance, highlighting its superior generalization and robustness under diverse temporal dynamics.
  • Consistent superiority across domains: KOSLM outperforms all competing baselines, particularly under complex and noisy datasets. In terms of average MSE and MAE reductions, KOSLM achieves relative improvements of 31.96% and 18.77% on Traffic, 13.37% and 25.20% on Electricity, 5.47% and 20.27% on Exchange, 11.26% and 18.43% on Weather, and 22.04% and 23.09% on ILI. Moreover, on the four ETT benchmarks, KOSLM achieves MSE improvements of up to 27.46% and MAE improvements of up to 25.2%, demonstrating strong adaptability to varying periodic and nonstationary patterns. These consistent gains verify that the Kalman-inspired selective updating mechanism effectively filters noise and dynamically adjusts to regime shifts, ensuring stable forecasting accuracy over long horizons.
  • Stable error distribution and reduced variance: The MSE–MAE gap of KOSLM remains narrower than that of other baselines, implying reduced large deviations and more concentrated prediction errors. This indicates more stable error behavior, which is crucial for long-horizon forecasting where cumulative drift often occurs. The innovation-driven Kalman gain estimation provides adaptive correction at each timestep, ensuring smooth and consistent prediction trajectories under uncertain dynamics.
  • Strong scalability and generalization: KOSLM achieves leading performance not only on large-scale datasets (Traffic, Electricity) but also on small, noisy datasets (ILI), confirming robust generalization across different temporal resolutions and noise levels. Its consistent advantage over Transformer-based (e.g., iTransformer, FEDFormer, PatchTST), recurrent (e.g., xLSTMTime), and state-space models (e.g., S-Mamba, FiLM) demonstrates that the proposed Kalman-optimal selective mechanism provides an effective inductive bias for modeling long-term dependencies.
  • Advantage over LSTM-based architectures: Compared with advanced LSTM-based models such as xLSTMTime, KOSLM achieves consistently better results across nearly all datasets and horizons. This verifies that replacing heuristic gates with Kalman-optimal selective gating enhances memory retention and update stability. While xLSTMTime alleviates gradient decay via hierarchical memory, KOSLM further refines state updates through innovation-driven gain estimation, thereby achieving a more principled and stable information flow.
To further demonstrate the long-horizon stability of the proposed model, we compare KOSLM with the recent S-Mamba [24], a state-of-the-art state space model that represents the latest advancement in efficient sequence modeling. Figure 3 presents the forecasting trajectories at the long horizon ( T = 720 ) across five representative datasets. KOSLM maintains accurate trend alignment and amplitude consistency with the ground truth, showing particularly superior convergence behavior on smoother and more stationary datasets such as Weather and Exchange, where the Kalman gain adaptation stabilizes long-term predictions. In contrast, S-Mamba exhibits slight temporal lag and amplitude attenuation under extended forecasting conditions. These results visually confirm the advantage of the proposed Kalman-based feedback selectivity in preserving long-term temporal fidelity.

4.2. Ablation Study

To further analyze the effectiveness of each proposed component, we conduct ablation experiments on four widely used long-term time-series forecasting datasets (ETTm1, ETTh1, Traffic, and Exchange) with a prediction length of L = 720 , following the data preprocessing described in Section 4.1.1. All model variants are trained and tested following the default implementation settings described in Section 4.1.2, and the results are averaged over five runs to ensure statistical reliability.

4.2.1. Structural Ablation

We first evaluate the contribution of the Kalman-based structure by comparing three variants: (i) Full (KOSLM): the complete model where K t = ϕ ( Innov ) and A t = ( I K t M t ) A , B t = K t ; (ii) No-Gain: ϕ receives the innovation but directly outputs ( A t , B t ) , removing the explicit computation of the Kalman gain K t , i.e., the pathway Innov K t ( A t , B t ) is replaced by Innov ( A t , B t ) ; (iii) No-Innov: the innovation statistic is removed. ϕ receives the standard network input (e.g., [ x t ; H t 1 ] ), while K t is still computed and mapped to ( A t , B t ) via the Kalman form, i.e., [ x t ; H t 1 ] K t ( A t , B t ) .
As shown in Table 6, removing either the gain computation or the innovation input consistently leads to higher errors across all datasets. For instance, on ETTm1, the MSE increases by 32.9% when the gain path is removed, confirming that both the innovation statistic and Kalman gain are essential for stable long-horizon prediction. This finding aligns with the Kalman filtering principle that the innovation serves as a sufficient statistic for state correction.

4.2.2. Capacity Ablation

To verify that the performance gains of KOSLM mainly stem from the Kalman-optimal structural design rather than the network capacity, we conduct a capacity ablation study. We evaluate the ϕ network under four progressively larger configurations: (i) Linear: A single linear layer with input dimension equal to the number of input features and output dimension 64, without any activation; (ii) SmallMLP: A two-layer MLP with 128 hidden units and sigmoid activation; (iii) MediumMLP: A three-layer MLP with 256 hidden units per layer and sigmoid activation; (iv) HighCap: A four-layer MLP with 512 hidden units per layer and sigmoid activation.
All models are trained under identical settings with five independent random seeds, and the results are reported as mean ± standard deviation. This allows us to evaluate both performance trends and statistical reliability. From Table 1, it is evident that increasing the ϕ network capacity beyond SmallMLP does not consistently improve MSE or MAE; in some cases, the error even increases slightly. The small MediumMLP improvements (ETTm1) or slight deterioration (ETTh1, Traffic, Exchange) indicate that performance gains are dominated by the Kalman-optimal structural design rather than network depth or over-parameterization. This confirms the effectiveness of the innovation-driven gain estimation in KOSLM for stable long-horizon prediction.

4.3. Efficiency Benchmark

To evaluate the computational efficiency of KOSLM, we benchmark it against representative sequence modeling baselines, including Transformer, Mamba-2, and LSTM, across four key metrics: runtime scalability, end-to-end throughput, memory footprint, and parameter count. All experiments are conducted using PyTorch 2.1.0 with torch.compile optimization on a single NVIDIA RTX 4090 GPU.

4.3.1. Runtime Scalability

We assess runtime performance on both a controlled synthetic setup and the ETTm1 dataset. As illustrated in Figure 4, KOSLM exhibits near-linear growth in inference time with increasing sequence length L. For sequences longer than L = 2 k , KOSLM surpasses optimized Transformer implementations (FlashAttention [31]) and achieves up to 2.5 × faster execution than the fused-kernel version of Mamba-2 [24] at L > 4 k . Compared with a standard PyTorch LSTM, KOSLM achieves approximately 1.3 1.9 × speedup across the tested sequence lengths, demonstrating robust scalability for long sequences.

4.3.2. Throughput Analysis

Figure 5 reports the end-to-end inference throughput measured in million tokens/s (M tokens/s). KOSLM maintains consistently high throughput across all tested sequence lengths, achieving up to 1.9 × higher throughput than LSTM and 2.7 × higher than Mamba-2 in the 4K–8K regime. The performance remains stable for even longer sequences, highlighting the model’s suitability for extended temporal dependencies.

4.3.3. Memory Footprint

Table 7 reports GPU memory usage during training for varying batch sizes with input length 2048 and models containing approximately 0.24 M parameters. KOSLM demonstrates favorable memory efficiency relative to similarly sized Mamba-2 implementations and remains comparable to a standard LSTM with roughly ten times more parameters. This efficiency results from KOSLM’s compact Kalman-optimal selective mechanism, which dynamically modulates state transitions with minimal parameter overhead.

4.3.4. Model Size

Table 8 summarizes the parameter counts. KOSLM contains only 0.24 M parameters, roughly 4.5% of the Transformer and 11% of the LSTM baselines, while achieving comparable or superior forecasting performance. The compact design emphasizes KOSLM’s suitability for deployment in resource-constrained environments.
This efficiency stems directly from the Kalman-optimal selective mechanism, which adaptively regulates the state update through innovation-driven gain modulation. Such a formulation not only reduces redundant computation but also provides a principled pathway for scaling to long sequences.

4.4. Real-World Application: Secondary Surveillance Radar (SSR) Target Trajectory Tracking

To further evaluate the robustness and practical generalizability of KOSLM under real-world noisy conditions, we conducted a real-world experiment on SSR target trajectory tracking. SSR is a ground-based air traffic surveillance system where aircraft respond to radar interrogations with encoded transponder signals, generating sparse range–azimuth sequences known as raw SSR plots. These observation sequences present three major challenges for sequential modeling:
  • High stochastic noise: Measurement noise leads to random fluctuations in the estimated positions;
  • Irregular sampling: Aircraft maneuvers and radar scan intervals result in uneven temporal spacing;
  • Correlated anomalies: Spurious echoes or missing detections introduce discontinuities in the trajectories.
These characteristics make SSR a natural yet challenging testbed for assessing model stability, noise resilience, and temporal consistency. KOSLM addresses these challenges by integrating context-aware selective state updates with learnable dynamics under the Kalman optimality principle, enabling robust filtering, smoothing, and extrapolation in partially observed environments.

4.4.1. Experimental Setup

To achieve a balance between realism and experimental controllability, we adopted a semi-physical simulation approach. Training data were derived from Automatic Dependent Surveillance–Broadcast (ADS-B) flight tracks obtained from the OpenSky Network [32]. Controlled Gaussian noise with an SNR of 33 dB was added to emulate SSR observation uncertainty. For evaluation, we deployed KOSLM on real SSR radar data collected from an operational ground-based system (peak transmit power 6 kW), capturing eight live air-traffic targets under operational conditions. Detailed data acquisition and preprocessing procedures are provided in Appendix C.
Unlike the main benchmarking experiments on standardized long-term time series forecasting (LTSF) datasets, this case study aims to demonstrate practical applicability and operational robustness. Due to irregular sampling, sparse observations, and high stochastic noise, conventional quantitative metrics such as MSE or MAE are not meaningful. Therefore, we focus on qualitative trajectory visualizations to illustrate model performance in realistic conditions.

4.4.2. Results and Analysis

Figure 6 presents representative trajectories of eight air-traffic targets, comparing classical KF algorithm, Transformer, Mamba, and KOSLM. The classical KF algorithm exhibits locally inaccurate and jagged trajectories due to its fixed linear dynamics, which cannot adapt to irregular sampling or abrupt maneuvers. The Transformer captures general trends but produces fragmented and temporally inconsistent tracks under sparse and noisy observations. Mamba improves noise robustness but still demonstrates local instability during complex maneuvers.
In contrast, KOSLM generates smooth, coherent, and temporally consistent trajectories that closely follow the true flight paths, highlighting its ability to handle high stochastic noise, irregular sampling, and correlated anomalies. This robustness stems from the innovation-driven Kalman gain and context-aware selective state updates, which allow KOSLM to adapt dynamically to non-stationary motion patterns.

5. Conclusions

This study addressed a fundamental limitation of recurrent architectures such as LSTM, whose heuristically designed gates lack structural constraints for optimality, leading to instability and information decay over long sequences. To overcome this issue, we proposed the Kalman-Optimal Selective Long-Term Memory Network (KOSLM), which reconstrues LSTM as a nonlinear, input- and state-dependent state-space model, and integrates an innovation-driven Kalman-optimal gain path for principled information selection. This formulation unifies LSTM gating, selective state-space modeling, and Kalman filtering into a single theoretically grounded recurrent framework. We note that the proposed KOSLM is lightweight, containing only 0.24 M parameters—significantly smaller than typical LSTM or Transformer baselines—while still achieving competitive or superior performance. This compact design ensures efficient implementation on standard GPU hardware without imposing constraints on memory or computation.
Extensive experiments demonstrate that KOSLM achieves state-of-the-art performance on long-term forecasting benchmarks, while ablation studies confirm the essential role of the innovation statistic and Kalman-form gain in stabilizing long-horizon modeling. Moreover, validation on real-world SSR trajectory tracking highlights its robustness under noisy and non-stationary conditions.
In summary, embedding Kalman-optimal principles into deep recurrent networks provides both theoretical insights and practical benefits for robust long-term sequence modeling. Future work will focus on extending KOSLM beyond Gaussian assumptions and applying it to multimodal and cross-domain time-series scenarios.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/app152312684/s1, Table S1: KOSLM run-to-run variability (RMS) over 5 runs.

Author Contributions

All authors contributed to the conception and design of the study. Material preparation, code implementation, data collection, and analysis were performed by X.T. and L.W. The first draft of the manuscript was written by L.W. and all authors commented on previous versions of the manuscript. X.T. and M.W. supervised the research and provided critical revisions and theoretical guidance. Y.Z. assisted in data verification and manuscript editing. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The datasets and source code used in this study are publicly available at https://github.com/wl822513/KOSLM (accessed on 23 October 2025).

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A. Selective State Space Models

This section focuses on S6, which provide a unified framework connecting the gating with time-varying state-space representations. Their associated selection mechanisms have become a central design principle in modern SSMs, such as Mamba [11]. These mechanisms improve modeling efficiency by dynamically identifying and retaining task-relevant information over long sequences, effectively compressing the latent state without compromising representational capacity.
  • Relation to Existing Concepts.
Selection mechanisms are conceptually related to several prior ideas, including gating, hypernetworks, and data-dependent parameterization:
  • Gating was originally introduced as a signal-control mechanism in recurrent neural networks (RNNs) such as LSTM [3] and GRU [33], where multiplicative gates regulate information updates across time steps. The concept has been generalized to other architectures, including gated convolutions and transformers [34,35], though often without an explicit interpretation in terms of temporal signal control.
  • Hypernetworks [36] generate network parameters dynamically through auxiliary networks, enabling the model’s internal dynamics to adapt based on input signals.
  • Data-dependent parameterization [37] represents a broader paradigm in which model parameters are directly conditioned on input data. Both gating and hypernetworks can be viewed as specific instances within this larger class.
While selection mechanisms share certain similarities with these paradigms, they form a distinct modeling category. Unlike classical gating or hypernetwork mechanisms that often operate locally or step-wise, selection mechanisms are explicitly designed to route, filter, or suppress sequence-level information in an input- or state-dependent manner, enabling stable long-horizon modeling. In selective SSMs, this is typically realized by parameterizing system matrices (e.g., Δ , A, B, C) as functions of the input at each time step.
  • From Implicit to Explicit Selection.
Early structured SSMs (S4) [38] encoded fixed inductive biases through learned structured dynamics, providing an implicit, input-independent form of selection via controlled signal propagation. Later models, including Mamba [11], introduced explicit selection, where the parameters of the state-space system (e.g., Δ , A, B, C) are conditioned directly on the current input, allowing the model to dynamically emphasize or suppress features at each step.
  • Semantic Clarification.
As discussed in S6 [11], while selection mechanisms can be loosely related to gating, hypernetworks, or general data dependence, these broader descriptions do not capture the defining characteristic of selection. The term selection is reserved for mechanisms that explicitly operate along the sequence dimension, enabling adaptive memory compression and long-range information control. From a dynamical systems perspective, such mechanisms can be interpreted through state-space discretization and signal propagation theory [39,40]. Importantly, classical RNN gating can be viewed as a local, step-wise precursor to these more general selection mechanisms.
  • Scope and Relevance.
The principle of selection underlies recent progress in linear-time sequence modeling. Building upon this paradigm, our proposed KOSLM extends selective SSMs by introducing an innovation-driven Kalman-optimal feedback pathway, transforming heuristic selection into a principled, uncertainty-aware mechanism for long-term sequence modeling.

Appendix B. Detailed Derivation

Appendix B.1. Kalman Gain as a Prototype for Dynamic Selectivity

In classical filtering theory, the Kalman gain K ( t ) acts as a dynamic weighting factor that determines how prior state estimates and new observations are combined to produce a new state [6]. It is formally derived as the solution to the Riccati differential equation, and its value depends on the evolving uncertainty in the internal system state (captured by the prior covariance P ( t ) ) and the noise characteristics of the observation process (captured by the measurement covariance R) [41]. This time-varying gain governs the extent to which incoming measurements correct the state estimate, ensuring minimum mean squared error under Gaussian noise assumptions.
This correction process is essentially a trade-off between the historical state information of the system and the incoming input information. In the context of deep learning, we interpret the observation as the current input sequence, and the prior state estimate as the latent representation of the system’s history. From this perspective, the Kalman gain plays a role analogous to a dynamic selection factor that balances the contribution of contextual knowledge (from the model state) and content information (from the input) in generating the updated representation. Therefore, the Kalman gain can be viewed as a principled prototype for content-aware and context-sensitive information selection. This perspective justifies the broader use of dynamic selection strategies in sequence models, especially when aiming to balance stability and adaptability in long-range modeling.

Appendix B.2. LSTM-to-SSM Reconstruction

To endow LSTM networks with structured modeling semantics, in this section we reconstruct them as a form of nonlinear time-varying state-space model ((see Equation (3a)). Specifically, we identify that the interactions among the forget gate, input gate, output gate, input signal, and cell state in LSTMs essentially constitute an input-dependent and state-aware state transition–observation process.
As illustrated in Figure 1, the gating mechanisms and the cell state C t in LSTM establish a pathway for memory propagation across time. The cell state C t is the core component that preserves long-term memory across the temporal dimension and models long-range dependencies. In terms of representational role, it is both semantically and structurally equivalent to the latent state h t in SSMs (see Equation (3a)). Therefore, we take C t as the central axis and analyze its information propagation process.
The forget gate F t regulates the degree to which short-term memory from cell state C t 1 is discarded. It is determined jointly by the current input x t and the previous hidden state H t 1 , thereby embodying an input- and state-dependent transition process:
A t F t = σ ( W f X t + b f ) ,
where X t = [ x t ; H t 1 ] denotes the concatenated input. The corresponding retained term is A t C t 1 .
The pathway for new information input in LSTM consists of two steps:
  • The candidate cell state C ˜ t can be interpreted as a differentiable nonlinear mapping from the joint input X t :
    z t C ˜ t = tanh ( W c X t + b c ) ,
    where the linear transformation W c X t + b c fuses the external input x t and the hidden state H t 1 into the observation space, with tanh ( · ) introducing nonlinearity to produce an intermediate representation z t . We treat this representation as the observation input at time t under the SSM (Here, the observation input denotes the externally observable signal that directly drives the state update process. It is written as z t in KF and as x t in SSMs, which are semantically equivalent).
  • The interaction between the input gate I t and the candidate state C ˜ t establishes a structured pathway for injecting new information into the state dynamics, analogous to the excitation of state evolution by external inputs in SSMs.
    B t z t I t C ˜ t = σ ( W i X t + b i ) tanh ( W c X t + b c ) ,
    where B t serves as the input matrix, structurally equivalent to the input gate I t .
Consequently, the update process of the LSTM cell state C t at time t (Equation (1)) can be rewritten in the form of a state-space transition:
C t = A t C t 1 + B t z t ,
where A t and B t are realized through the forget and input gates driven by the joint input X t , thereby endowing the model with the ability to selectively forget or memorize long-term information in an input-dependent and context-aware manner.
Finally, the output pathway of LSTM is given by
H t = O t tanh ( C t ) ,
which is formally equivalent to the observation equation of a classical SSM:
y t = M t C t ,
where M t is formed by the output gate O t together with the nonlinear transformation tanh ( · ) .
The above reconstruction shows that the gating mechanism in LSTM can be interpreted as a class of nonlinear, input- and state-dependent SSMs. This perspective establishes a valid unification of LSTM gating mechanisms and state-space models, and reveals that LSTM’s long-term memory capability originates from the structured realization of both state-space modeling and efficient information selection. Building on this understanding, we further introduce the Kalman gain into this state-space structure, imposing a structural constraint that minimizes the uncertainty of state estimation errors.

Appendix B.3. Kalman-Optimal Selective Mechanism

Recent studies on time series modeling have proposed a class of selective mechanisms based on SSMs, among which the most representative work is S6 [11]. Its core idea points out the following:
“One method of incorporating a selection mechanism into models is by letting their parameters that affect interactions along the sequence (e.g., the recurrent dynamics of an RNN or the convolution kernel of a CNN) be input-dependent.”
  • Inspired by this, we design a selective mechanism based on Kalman-optimal state estimation within the SSM (Equation (3)). Unlike the input-dependent selection mechanism represented by S6, we make the key parameters ( A t , B t ), which control how the model selectively propagates or forgets information along the sequence dimension, depend on the innovation term: the deviation between the observation input and the prior state prediction in the observation space, denoted as Innov in this paper. This method constructs a learnable selection path that integrates observational inputs and latent state feedback, with the optimization objective of minimizing the uncertainty in state estimation errors.
  • Learnable Gain from Innovation.
In the KF algorithm, the closed-form solution of the Kalman gain K t (Equation (7)) is obtained by minimizing the covariance of the state estimation error [6]. Here, the observation matrix M t and the observation noise covariance R t are considered known system priors, usually derived from physical modeling or domain knowledge. Therefore, the dynamics of K t entirely stem from the statistical properties of the prior estimation error e t = h t h ^ t . Under the linear observation model (4b), e t is mapped into the observation space, yielding the following representation of the innovation term:
Innov = z t M t h ^ t = M t h t h ^ t + v t ,
This equation shows that Innov is a linear mapping of e t , which, although perturbed by additive observation noise v t , still retains the full uncertainty information of e t . Thus, Innov can serve as a sufficient statistic for estimating K t , providing a theoretical basis for directly estimating K t from the innovation term. The estimation process is defined as
K t = ϕ ( Innov t ; θ ϕ ) ,
where ϕ ( · ) is a differentiable nonlinear function, parameterized by a neural network and trained via gradient descent. Theoretical soundness of this formulation is confirmed by a controlled experiment (Appendix B.4), which demonstrates that ϕ ( · ) can reliably recover the oracle Kalman gain from the innovation statistics with negligible estimation error.
  • Optimal Selectivity via Gain.
By substituting the KF prediction process (Equation (5)) into the state update equation (Equation (8)), the two processes can be expressed as a nonlinear time-varying SSM form (Equation (3a)):
h ^ t = A t h ^ t 1 + B t z t , A t = ( I K t M t ) A , B t = K t ,
Within this framework, parameters A t and B t are linearly modulated by K t driven by the innovation term, yielding a context-aware Kalman-optimal selective path that minimizes state estimation uncertainty and imposes structural optimality constraints on LSTM.

Appendix B.4. Empirical Validation of Innovation-Driven Kalman Gain Learning

To empirically validate the theoretical claim that the Kalman gain can be dynamically inferred from the innovation term, we design a controlled linear–Gaussian synthetic experiment where the optimal gain has a closed-form solution. This allows us to directly compare the learned gain against the ground-truth Kalman gain and quantify their alignment over time.
  • Experimental Design.
We consider a one-dimensional linear dynamical system as in Equation (4). The analytical Kalman gain K t is computed from the standard Riccati recursion and serves as the oracle reference. We design three model variants for comparison:
(i)
Oracle-KF: Standard Kalman filter using the known ( A , Q , R , M ) parameters to compute K t .
(ii)
Supervised- ϕ : A small MLP ϕ ( · ) takes the innovation  i n n o v t = z t M h ^ t as input and predicts K ^ t = ϕ ( i n n o v t ) , trained by minimizing L = | K ^ t K t | 2 over all timesteps. During training, the predicted prior state h ^ t is provided by the Oracle-KF to isolate the learning of the innovation-to-gain mapping. This tests whether the innovation term contains sufficient information to recover the oracle gain.
(iii)
End-to-End: The same MLP is trained to directly map from ( z t , h ^ t ) to K t without explicitly constructing the innovation term, serving as an ablation to assess the importance of innovation-driven modeling.
  • Setup and Metrics.
We simulate 10 , 000 trajectories of length T = 100 with fixed system parameters ( A = 0.9 , M = 1 ) . To test robustness under different noise regimes, we consider four parameter settings by varying process noise Q and observation noise R: (i) Q = 5 , R = 0.1 ; (ii) Q = 5 , R = 0.5 ; (iii) Q = 10 , R = 0.1 ; (iv) Q = 100 , R = 0.5 . The supervised models are trained for 200 epochs using Adam optimizer with a learning rate 10 3 . Performance is evaluated by
  • MSE of Gain:  MSE = 1 T t | K t K ^ t | 2 , measuring how well the learned gain matches the oracle trajectory.
  • State Estimation RMSE: Root-mean-square error between estimated and true state trajectories, where the state estimates are produced by running a Kalman update step with the learned K ^ t , verifying that accurate gain learning improves filtering quality.
  • Results and Analysis.
Figure A1 presents the learned Kalman gain trajectories across all considered noise regimes. Across every ( Q , R ) configuration, the Supervised- ϕ model closely matches the oracle Kalman gain K t , producing stable and accurate trajectories throughout both transient and steady phases. In contrast, the End-to-End model produces a noisier and slightly biased gain curve, indicating that omitting the explicit innovation term fails to accurately capture the gain dynamics. This qualitative observation is quantitatively confirmed in Table A1: Supervised- ϕ achieves up to four orders-of-magnitude lower gain MSE and reduces state-estimation RMSE by 4.5–32.8% under all ( Q , R ) noise settings. These results demonstrate that the innovation term is indeed sufficient for inferring the optimal Kalman gain and that explicitly modeling innovation-driven selectivity yields more stable and accurate state estimation. Together, the figure and table provide strong empirical support for our theoretical claim that innovation-guided gain learning constitutes a principled and robust mechanism for selective state updates.
Table A1. Quantitative Results of Kalman Gain Learning. Supervised- ϕ achieves orders-of-magnitude lower gain MSE and consistently better state estimation RMSE across all noise regimes compared to end-to-end learning, highlighting the critical role of explicitly modeling the innovation term in recovering accurate Kalman gain dynamics.
Table A1. Quantitative Results of Kalman Gain Learning. Supervised- ϕ achieves orders-of-magnitude lower gain MSE and consistently better state estimation RMSE across all noise regimes compared to end-to-end learning, highlighting the critical role of explicitly modeling the innovation term in recovering accurate Kalman gain dynamics.
(Q, R)Supervised- ϕ End-to-End
MSERMSEMSERMSE
(5, 0.1) 1.3 × 10 6 0.317 3.5 × 10 3 0.368
(5, 0.5) 2.9 × 10 5 0.684 1.8 × 10 3 0.694
(10, 0.1) 1.2 × 10 6 0.216 1.3 × 10 3 0.341
(100, 0.5) 2.2 × 10 7 0.310 6.8 × 10 4 0.747
Figure A1. Kalman gain learning across different noise regimes. Top-left: ( Q = 100 , R = 0.5 ) ; top-right: ( Q = 5 , R = 0.1 ) ; bottom-left: ( Q = 5 , R = 0.5 ) ; bottom-right: ( Q = 10 , R = 0.1 ) . Supervised- ϕ closely matches the oracle K t in all cases, while end-to-end trajectories remain noisy and biased. These results confirm that innovation-driven modeling provides stable and accurate gain learning across both low- and high-noise conditions.
Figure A1. Kalman gain learning across different noise regimes. Top-left: ( Q = 100 , R = 0.5 ) ; top-right: ( Q = 5 , R = 0.1 ) ; bottom-left: ( Q = 5 , R = 0.5 ) ; bottom-right: ( Q = 10 , R = 0.1 ) . Supervised- ϕ closely matches the oracle K t in all cases, while end-to-end trajectories remain noisy and biased. These results confirm that innovation-driven modeling provides stable and accurate gain learning across both low- and high-noise conditions.
Applsci 15 12684 g0a1

Appendix C. SSR Case Study: Data and Preprocessing Details

Appendix C.1. Data Acquisition

The semi-physical training dataset is constructed using ADS-B logs from the OpenSky Network, recorded on 15 June 2025, between 10:00 and 10:15 AM. The data cover a 300 km region surrounding Tokyo Haneda, Narita, and Incheon Airports. Each log provides updates every 5 s for 503 aircraft, including longitude, latitude, altitude, and timestamp information. A detailed statistical summary of the dataset is provided in Table A2.
To emulate realistic Secondary Surveillance Radar (SSR) observation noise, zero-mean Gaussian perturbations are applied to each measurement with a signal-to-noise ratio (SNR) of 33 dB, formulated as
X i = X i + N 0 , σ i 2 10 33 10 ,
where σ i 2 represents the empirical variance of the i-th feature. This noise configuration reflects typical radar tracking uncertainty under moderate atmospheric interference.
Table A2. Summary of the Semi-physical ADS-B Dataset Used for Training. The dataset consists of 5 s updates for 503 aircraft collected within a 15 min window around major East Asian airports. Each sample includes four state variables: longitude, latitude, altitude, and timestamp.
Table A2. Summary of the Semi-physical ADS-B Dataset Used for Training. The dataset consists of 5 s updates for 503 aircraft collected within a 15 min window around major East Asian airports. Each sample includes four state variables: longitude, latitude, altitude, and timestamp.
TargetsUpdate Frequency# FeaturesTotal SamplesTime Span
5035 s4166,11015 June 2025, 10:00–10:15

Appendix C.2. Normalization and Reverse Transformation

All input data are standardized to zero mean and unit variance:
y = x x ¯ σ , x ^ = σ y ^ + x ¯ .

Appendix C.3. Field Data Collection

For real-world evaluation, we use raw SSR plots collected by a field-deployed radar with a peak transmit power of 6 kW. These sequences contain irregular sampling, strong noise, and missing returns, providing a rigorous test of model robustness.

Appendix C.4. Processing Workflow

The overall pipeline—from semi-physical data generation to field testing—is summarized in Figure A2, where the upper branch denotes the training stage on noisy ADS-B data, and the lower branch represents inference on real SSR plots. This workflow ensures consistency between semi-physical training data and real-world SSR testing scenarios.
Figure A2. SSR trajectory modeling pipeline. The upper branch shows training using ADS-B data with noise simulation and normalization; the lower branch shows testing on real raw SSR plots using trained models, followed by trajectory prediction and comparison.
Figure A2. SSR trajectory modeling pipeline. The upper branch shows training using ADS-B data with noise simulation and normalization; the lower branch shows testing on real raw SSR plots using trained models, followed by trajectory prediction and comparison.
Applsci 15 12684 g0a2

References

  1. Cheng, M.; Liu, Z.; Tao, X.; Liu, Q.; Zhang, J.; Pan, T.; Zhang, S.; He, P.; Zhang, X.; Wang, D.; et al. A comprehensive survey of time series forecasting: Concepts, challenges, and future directions. TechRxiv 2025. [Google Scholar] [CrossRef] [PubMed]
  2. Salazar, C.; Banerjee, A.G. A distance correlation-based approach to characterize the effectiveness of recurrent neural networks for time series forecasting. Neurocomputing 2025, 629, 129641. [Google Scholar] [CrossRef]
  3. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  4. Benjamin, J.; Mathew, J. Enhancing continuous integration predictions: A hybrid LSTM-GRU deep learning framework with evolved DBSO algorithm. Computing 2025, 107, 9. [Google Scholar] [CrossRef]
  5. Pascanu, R.; Mikolov, T.; Bengio, Y. On the Difficulty of Training Recurrent Neural Networks. In Proceedings of the 30th International Conference on Machine Learning (ICML’13), Atlanta, GA, USA, 16–21 June 2013; pp. 1310–1318. [Google Scholar]
  6. Kalman, R.E. A New Approach to Linear Filtering and Prediction Problems. J. Basic Eng. 1960, 82, 35–45. [Google Scholar] [CrossRef]
  7. Yuan, X.; Li, J.; Kuruoglu, E.E. Robustness enhancement in neural networks with alpha-stable training noise. Digit. Signal Process. 2025, 156, 104778. [Google Scholar] [CrossRef]
  8. Krishnan, R.G.; Shalit, U.; Sontag, D. Deep kalman filters. arXiv 2015, arXiv:1511.05121. [Google Scholar] [CrossRef]
  9. Revach, G.; Shlezinger, N.; Ni, X.; Escoriza, A.L.; Van Sloun, R.J.; Eldar, Y.C. KalmanNet: Neural network aided Kalman filtering for partially known dynamics. IEEE Trans. Signal Process. 2022, 70, 1532–1547. [Google Scholar] [CrossRef]
  10. Dahan, Y.; Revach, G.; Dunik, J.; Shlezinger, N. Uncertainty quantification in deep learning based Kalman filters. In Proceedings of the ICASSP 2024—2024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Seoul, Republic of Korea, 14–19 April 2024; pp. 13121–13125. [Google Scholar]
  11. Gu, A.; Dao, T. Mamba: Linear-time sequence modeling with selective state spaces. arXiv 2023, arXiv:2312.00752. [Google Scholar] [CrossRef]
  12. Dao, T.; Gu, A. Transformers are ssms: Generalized models and efficient algorithms through structured state space duality. arXiv 2024, arXiv:2405.21060. [Google Scholar] [CrossRef]
  13. Aliyu, B.K.; Osheku, C.A.; Adetoro, L.M.; Funmilayo, A.A. Optimal Solution to Matrix Riccati Equation—For Kalman Filter Implementation. In MATLAB; Katsikis, V.N., Ed.; IntechOpen: Rijeka, Croatia, 2012; Chapter 4. [Google Scholar] [CrossRef][Green Version]
  14. Gu, A.; Goel, K.; Gupta, A.; Ré, C. On the parameterization and initialization of diagonal state space models. Adv. Neural Inf. Process. Syst. 2022, 35, 35971–35983. [Google Scholar]
  15. Gu, A.; Gulcehre, C.; Paine, T.; Hoffman, M.; Pascanu, R. Improving the gating mechanism of recurrent neural networks. In Proceedings of the International Conference on Machine Learning, PMLR, Virtual, 13–18 July 2020; pp. 3800–3809. [Google Scholar]
  16. Weather Dataset. 2024. Weather Data from the BGC Jena Climate Data Archive. Available online: https://www.hypermeteo.com/en/?gad_source=1&gad_campaignid=14712996685&gclid=EAIaIQobChMIvJ_HqOORkQMVSpCDBx3MFTosEAAYASAAEgLx7fD_BwE (accessed on 12 May 2024).
  17. ETT Datasets. 2024. Energy Transformer Temperature Dataset (ETT) for Long-Term Forecasting. Available online: https://ieee-dataport.org/documents/merged-ett-dataset-time-series-forecasting (accessed on 12 May 2024).
  18. Electricity Dataset. 2024. Electricity Load Diagrams Dataset from UCI Machine Learning Repository. Available online: https://ieee-dataport.org/documents/uci-dataset (accessed on 12 May 2024).
  19. Exchange Rate Dataset. 2024. Daily Exchange Rate Dataset for Long-Term Time Series Forecasting, Provided with the ETT Benchmark Collection. Available online: https://github.com/juyongjiang/TimeSeriesDatasets (accessed on 12 May 2024).
  20. Traffic Dataset. 2024. Traffic Flow Data from California Performance Measurement System (PeMS). Available online: https://www.smartmicro.com/traffic-sensor?utm_medium=advert&utm_source=google&utm_campaign=search-traffic&utm_content=ad&utm_term=traffic_sensor&gad_source=1&gad_campaignid=22435495111&gclid=EAIaIQobChMIlvLc8OORkQMVQJKDBx3FOSSQEAAYASAAEgITNPD_BwE (accessed on 12 May 2024).
  21. Illness (ILI) Dataset. 2024. Weekly Influenza-Like Illness (ILI) Dataset for Forecasting Tasks, Available Within the ETT Benchmark Suite. Available online: https://www.researchgate.net/figure/Influenza-like-illness-ILI-data-from-the-FluID-database-Weekly-scaled-rates-of-ILI_fig1_337718309 (accessed on 12 May 2024).
  22. Alharthi, M.; Mahmood, A. xlstmtime: Long-term time series forecasting with xlstm. AI 2024, 5, 1482–1495. [Google Scholar] [CrossRef]
  23. Zhou, T.; Ma, Z.; Wen, Q.; Sun, L.; Yao, T.; Yin, W.; Jin, R. Film: Frequency improved legendre memory model for long-term time series forecasting. Adv. Neural Inf. Process. Syst. 2022, 35, 12677–12690. [Google Scholar]
  24. Wang, Z.; Kong, F.; Feng, S.; Wang, M.; Yang, X.; Zhao, H.; Wang, D.; Zhang, Y. Is mamba effective for time series forecasting? Neurocomputing 2025, 619, 129178. [Google Scholar] [CrossRef]
  25. Zhou, T.; Ma, Z.; Wen, Q.; Wang, X.; Sun, L.; Jin, R. Fedformer: Frequency enhanced decomposed transformer for long-term series forecasting. In Proceedings of the International Conference on Machine Learning, PMLR, Baltimore, MD, USA, 17–23 July 2022; pp. 27268–27286. [Google Scholar]
  26. Liu, Y.; Hu, T.; Zhang, H.; Wu, H.; Wang, S.; Ma, L.; Long, M. itransformer: Inverted transformers are effective for time series forecasting. arXiv 2023, arXiv:2310.06625. [Google Scholar]
  27. Zhang, Y.; Yan, J. Crossformer: Transformer Utilizing Cross-Dimension Dependency for Multivariate Time Series Forecasting. In Proceedings of the Eleventh International Conference on Learning Representations (ICLR), Kigali, Rwanda, 1–5 May 2023. [Google Scholar]
  28. Su, J.; Shen, Y.; Wu, Z. DLinear: A Linear Complexity Approach for Long-Term Time Series Forecasting. arXiv 2022, arXiv:2205.13504. [Google Scholar]
  29. Huang, X.; Tang, J.; Shen, Y. Long time series of ocean wave prediction based on PatchTST model. Ocean. Eng. 2024, 301, 117572. [Google Scholar] [CrossRef]
  30. Wang, S.; Wu, H.; Shi, X.; Hu, T.; Luo, H.; Ma, L.; Zhang, J.Y.; Zhou, J. Timemixer: Decomposable multiscale mixing for time series forecasting. arXiv 2024, arXiv:2405.14616. [Google Scholar] [CrossRef]
  31. Dao, T. Flashattention-2: Faster attention with better parallelism and work partitioning. arXiv 2024, arXiv:2307.08691. [Google Scholar]
  32. Zhang, J.; Wei, L.; Yanbo, Z. Study of ADS-B data evaluation. Chin. J. Aeronaut. 2011, 24, 461–466. [Google Scholar] [CrossRef]
  33. Cho, K.; Van Merriënboer, B.; Gulcehre, C.; Bahdanau, D.; Bougares, F.; Schwenk, H.; Bengio, Y. Learning phrase representations using RNN encoder-decoder for statistical machine translation. arXiv 2014, arXiv:1406.1078. [Google Scholar] [CrossRef]
  34. Mehta, H.; Gupta, A.; Cutkosky, A.; Neyshabur, B. Long range language modeling via gated state spaces. arXiv 2022, arXiv:2206.13947. [Google Scholar] [CrossRef]
  35. Hua, W.; Dai, Z.; Liu, H.; Le, Q. Transformer quality in linear time. In Proceedings of the International Conference on Machine Learning, PMLR, Honolulu, HI, USA, 23–29 July 2022; pp. 9099–9117. [Google Scholar]
  36. Ha, D.; Dai, A.; Le, Q.V. Hypernetworks. arXiv 2016, arXiv:1609.09106. [Google Scholar] [PubMed]
  37. Poli, M. Data-dependent neural networks. arXiv 2023, arXiv:2305.13272. [Google Scholar]
  38. Gu, A.; Goel, K.; Ré, C. Combining recurrent, convolutional, and continuous-time models with structured state space. Adv. Neural Inf. Process. Syst. 2021, 34, 21984–21998. [Google Scholar]
  39. Funahashi, K.I.; Nakamura, Y. Approximation of dynamical systems by continuous time recurrent neural networks. Neural Netw. 1993, 6, 801–806. [Google Scholar] [CrossRef]
  40. Tallec, C.; Ollivier, Y. Can recurrent neural networks warp time? arXiv 2018, arXiv:1804.11188. [Google Scholar] [CrossRef]
  41. Tahirovic, A.; Redzovic, A. Optimal Estimation for Continuous-Time Nonlinear Systems Using State-Dependent Riccati Equation (SDRE). arXiv 2025, arXiv:2503.10442. [Google Scholar]
Figure 1. Structure of the LSTM network.
Figure 1. Structure of the LSTM network.
Applsci 15 12684 g001
Figure 2. Unfolded KOSLM network layer. The D-dimensional input x t is dynamically coupled with the N-dimensional hidden state H t 1 and mapped to the output H t through the higher-dimensional cell state C t R T . Compared to the heuristic forget and output gates of the classical LSTM, our mechanism introduces Kalman-optimal constraints by parameterizing SSM parameters ( A t , B t ) based on the innovation term.
Figure 2. Unfolded KOSLM network layer. The D-dimensional input x t is dynamically coupled with the N-dimensional hidden state H t 1 and mapped to the output H t through the higher-dimensional cell state C t R T . Compared to the heuristic forget and output gates of the classical LSTM, our mechanism introduces Kalman-optimal constraints by parameterizing SSM parameters ( A t , B t ) based on the innovation term.
Applsci 15 12684 g002
Figure 3. KOSLM vs. S-Mamba: Forecasting comparison on five representative datasets with both input and prediction horizons set to 720. The blue line denotes the ground truth, and the red line indicates model predictions. KOSLM demonstrates superior long-horizon stability and trend consistency compared to S-Mamba.
Figure 3. KOSLM vs. S-Mamba: Forecasting comparison on five representative datasets with both input and prediction horizons set to 720. The blue line denotes the ground truth, and the red line indicates model predictions. KOSLM demonstrates superior long-horizon stability and trend consistency compared to S-Mamba.
Applsci 15 12684 g003
Figure 4. Runtime benchmarks. Inference time as a function of sequence length L. KOSLM demonstrates near-linear scaling and maintains faster inference than Transformer and Mamba-2 for long sequences.
Figure 4. Runtime benchmarks. Inference time as a function of sequence length L. KOSLM demonstrates near-linear scaling and maintains faster inference than Transformer and Mamba-2 for long sequences.
Applsci 15 12684 g004
Figure 5. Throughput benchmarks. End-to-end inference throughput on RTX 4090. KOSLM maintains stable throughput as sequence length increases, outperforming LSTM and Mamba-2 for long sequences.
Figure 5. Throughput benchmarks. End-to-end inference throughput on RTX 4090. KOSLM maintains stable throughput as sequence length increases, outperforming LSTM and Mamba-2 for long sequences.
Applsci 15 12684 g005
Figure 6. SSR target tracking under real-world operational conditions. Trajectories of eight air-traffic targets tracked from SSR measurements. Red curves denote raw SSR observations, blue curves denote tracking by the conventional KF algorithm (on-site SSR interrogator output), and teal curves indicate deep model predictions (Transformer, Mamba, or KOSLM), followed by KF smoothing. KOSLM achieves smoother, more accurate, and stable tracking under strong noise and irregular sampling conditions.
Figure 6. SSR target tracking under real-world operational conditions. Trajectories of eight air-traffic targets tracked from SSR measurements. Red curves denote raw SSR observations, blue curves denote tracking by the conventional KF algorithm (on-site SSR interrogator output), and teal curves indicate deep model predictions (Transformer, Mamba, or KOSLM), followed by KF smoothing. KOSLM achieves smoother, more accurate, and stable tracking under strong noise and irregular sampling conditions.
Applsci 15 12684 g006
Table 1. Capacity ablation (corrected). Performance of KOSLM with varying ϕ network capacities on four long-term forecasting datasets ( L = 720 ). Averaged over five runs with standard deviations. Lower values (indicated by the down arrow “↓”) correspond to better performance.
Table 1. Capacity ablation (corrected). Performance of KOSLM with varying ϕ network capacities on four long-term forecasting datasets ( L = 720 ). Averaged over five runs with standard deviations. Lower values (indicated by the down arrow “↓”) correspond to better performance.
DatasetVariant#ParamsMSE ↓MAE ↓Δ MSE vs. SmallMLP
ETTm1Linear0.183M0.480 ± 0.0120.355 ± 0.008+46.3%
SmallMLP0.255M0.328 ± 0.0100.282 ± 0.007
MediumMLP0.389M0.335 ± 0.0110.284 ± 0.001+2.13%
HighCap0.898M0.342 ± 0.0150.288 ± 0.010+4.27%
ETTh1Linear0.183M0.440 ± 0.0100.330 ± 0.004+30.95%
SmallMLP0.255M0.336 ± 0.0080.291 ± 0.006
MediumMLP0.389M0.341 ± 0.0100.294 ± 0.002+1.49%
HighCap0.898M0.355 ± 0.0120.302 ± 0.008+5.65%
TrafficLinear0.183M0.290 ± 0.0110.232 ± 0.006+9.02%
SmallMLP0.255M0.266 ± 0.0100.212 ± 0.007
MediumMLP0.389M0.270 ± 0.0120.216 ± 0.002+1.50%
HighCap0.898M0.279 ± 0.0140.222 ± 0.005+4.89%
ExchangeLinear0.183M0.272 ± 0.0100.305 ± 0.005+4.62%
SmallMLP0.255M0.260 ± 0.0090.291 ± 0.009
MediumMLP0.389M0.264 ± 0.0100.294 ± 0.007+1.54%
HighCap0.898M0.285 ± 0.0120.310 ± 0.009+9.62%
Table 2. Core structural comparison between KalmanNet and KOSLM.
Table 2. Core structural comparison between KalmanNet and KOSLM.
AspectKalmanNetKOSLM (Ours)
Core ideaStrictly follows the classical KF architecture (Section 2.2.2); the neural network learns to correct the Kalman gain K t KF under partially known linear dynamics.Reinterprets LSTM gating as a Kalman-optimal state estimation problem; state estimation does not strictly adhere to the KF equations, but directly learns K t from the innovation term.
System dynamicsAssumes fixed parameters ( A , M , Q , R ) ; suitable for systems with known or partially known dynamics.Learns ( A t , B t , M t ) from data; A is a learnable parameter matrix; fully adaptive to nonlinear and nonstationary environments.
Form of gain K t = K t KF + f ( s t ; θ f ) —learns a residual correction to the classical Kalman gain. K t = ϕ ( Innov t ; θ ϕ ) —directly learns the gain function from the innovation.
Gain network input s t = [ h ^ t | t 1 , z t M h ^ t | t 1 ] —uses both predicted state and innovation as inputs. Innov t = z t M t A t 1 C t 1 —relies solely on the innovation signal for gain computation (see Appendices Appendix B.3 and Appendix B.4, which prove its sufficiency).
Output roleOutputs a residual correction Δ K t added to K t KF .Outputs K t and integrates gain estimation into the state transition: A t = ( I K t M t ) A , B t = K t , forming a unified recurrent–estimation pathway.
System dependencyRequires partial knowledge of ( A , H , Q , R ) for baseline Kalman computation.Fully data-driven; no analytical Kalman gain or explicit system parameters required.
Theoretical interpretationNeural residual learning to compensate model mismatch.Innovation-driven dynamic selectivity that enforces Kalman-optimal information update behavior.
Table 3. Details of benchmark datasets used in our experiments.
Table 3. Details of benchmark datasets used in our experiments.
DatasetFrequency# FeaturesTime StepsTime Span
ETTh11 h717,4202016–2017
ETTh21 h717,4202017–2018
ETTm115 min769,6802016–2017
ETTm215 min769,6802017–2018
Exchange1 day875881990–2010
Weather10 min2152,6962020
Electricity1 h32126,3042012–2014
ILI7 days79662002–2020
Traffic1 h86217,5442015–2016
Table 4. Long-term forecasting: Multivariate long-term forecasting results on the Traffic, Electricity, Exchange, Weather, and ILI datasets. The prediction length is set to T { 96 , 192 , 336 , 720 } for all datasets except ILI, which uses O { 24 , 36 , 48 , 60 } due to its weekly resolution. The best results are shown in bold red, and the second-best are underlined purple. All results are averaged over 5 runs. Lower is better.
Table 4. Long-term forecasting: Multivariate long-term forecasting results on the Traffic, Electricity, Exchange, Weather, and ILI datasets. The prediction length is set to T { 96 , 192 , 336 , 720 } for all datasets except ILI, which uses O { 24 , 36 , 48 , 60 } due to its weekly resolution. The best results are shown in bold red, and the second-best are underlined purple. All results are averaged over 5 runs. Lower is better.
ModelsKOSLMxLSTMTimeFiLMiTransformerFEDFormerS-MambaCrossformerDLinearPatchTSTTimeMixer
MetricMSEMAEMSEMAEMSEMAEMSEMAEMSEMAEMSEMAEMSEMAEMSEMAEMSEMAEMSEMAE
Traffic960.2900.2300.3580.2420.4160.2940.3950.2680.5870.3660.3820.2610.5220.2900.6500.3960.4620.2950.4620.285
1920.2600.2100.3780.2530.4080.2880.4170.2760.6040.3730.3960.2670.5300.2930.5980.3700.4660.2960.4730.296
3360.2270.1790.3920.2610.4250.2980.4330.2830.6210.3830.4170.2760.5580.3050.6050.3730.4820.3040.4980.297
7200.2900.2320.4340.2870.5200.3530.4670.3020.6260.3820.4600.3000.5890.3280.6450.3940.5140.3220.5060.313
+31.96%Avg0.2660.2120.3910.2610.4420.3080.4280.2820.6100.3760.4340.2870.5500.3040.6250.3830.4810.3040.4850.297
Electricity960.1360.1840.1280.2210.1540.2670.1480.2400.1930.3080.1390.2350.2190.3140.1970.2820.1810.2700.1530.247
1920.1310.1840.1500.2430.1640.2580.1620.2530.2010.3150.1590.2550.2310.3220.1960.2850.1880.2740.1660.256
3360.1200.1770.1660.2590.1880.2830.1780.2690.2140.3290.1760.2720.2460.3370.2090.3010.2040.2930.1850.277
7200.1580.2050.1850.2760.2360.3320.2250.3170.2460.3550.2040.2980.2800.3630.2450.3330.2460.3240.2250.310
+13.37%Avg0.1360.1870.1570.2500.1860.2850.1780.2700.2140.3270.1700.2650.2440.3340.2120.3000.2050.2900.1820.272
Exchange960.1350.2120.0860.2040.0860.2060.1480.2780.0860.2070.2560.3670.0880.2180.0880.2050.0950.207
1920.2790.2920.1880.2920.1770.2990.2710.3150.1820.3040.4700.5090.1760.3150.1760.2990.1510.293
3360.3410.3390.3560.4330.3310.4170.4600.4270.3320.4181.2680.8830.3130.4270.3010.3970.2640.361
7200.2820.3220.7270.6690.8470.6911.1950.6950.8670.7031.7671.0680.8390.6950.9010.7140.5860.602
+5.47%Avg0.2590.2910.3390.4000.3600.4030.5190.4290.3670.4080.9400.7070.3540.4140.3670.4040.2740.365
Weather960.1440.1710.1440.1870.1990.2620.1740.2140.2170.2960.1650.2100.1580.2300.1960.2550.1770.2180.1630.209
1920.2240.2360.1920.2360.2280.2880.2210.2540.2760.3360.2140.2520.2060.2770.2370.2960.2250.2590.2080.250
3360.2490.2480.2370.2720.2670.3230.2780.2960.3390.3800.2740.2970.2720.3350.2830.3350.2780.2970.2510.287
7200.1690.1750.3130.3260.3190.3610.3580.3470.4030.4280.3500.3450.3980.4180.3450.3810.3540.3480.3390.341
+11.26%Avg0.1970.2080.2220.2550.2530.3090.2580.2780.3090.3600.2510.2760.2590.3150.2650.3170.2590.2810.2400.271
ILI241.1600.5071.5140.6941.9700.8753.1541.2353.2281.2603.0411.1862.2151.0811.3190.7541.4530.827
361.2620.5611.5190.7221.9820.8592.5441.0832.6791.1503.4061.2321.9630.9631.5790.8701.6270.903
481.0980.5451.5000.7251.8680.8962.4891.1122.6221.0803.4591.2212.1301.0241.5530.8151.6440.914
601.1180.5831.4180.7152.0570.9292.6751.0342.8571.0783.6401.3052.3681.0961.4700.7881.6330.908
+22.04%Avg1.1600.5491.4880.7141.9690.8902.7151.1162.8471.1703.3871.2362.1691.0411.4800.8071.5890.888
Table 5. Long-term forecasting: Forecasting results on the ETT datasets with prediction lengths T { 96 , 192 , 336 , 720 } . The best results are shown in bold red, and the second-best are underlined purple. All results are averaged over 5 runs. Lower is better.
Table 5. Long-term forecasting: Forecasting results on the ETT datasets with prediction lengths T { 96 , 192 , 336 , 720 } . The best results are shown in bold red, and the second-best are underlined purple. All results are averaged over 5 runs. Lower is better.
ModelsKOSLMxLSTMTimeFiLMiTransformerFEDFormerS-MambaCrossformerDLinearPatchTSTTimeMixer
MetricMSEMAEMSEMAEMSEMAEMSEMAEMSEMAEMSEMAEMSEMAEMSEMAEMSEMAEMSEMAE
ETTm1960.2040.2210.2860.3350.3340.3680.3790.4190.3330.3680.4040.4260.3450.3720.3290.3670.3200.357
1920.2330.2370.3290.3610.3770.3910.4260.4410.3760.3900.4500.4510.3800.3890.3670.3850.3610.381
3360.4140.3390.3580.3790.4260.4200.4450.4590.4080.4130.5320.5150.4130.4130.3990.4100.3900.404
7200.4810.3680.4160.4110.4910.4590.5430.4900.4750.4480.6660.5890.4740.4530.4540.4390.4580.441
+4.03%Avg0.3330.2910.3470.3720.4070.4100.4480.4520.3980.4050.5130.4960.4030.4070.3870.4000.3820.395
ETTm2960.1000.2320.1640.2500.1650.2560.1800.2640.2030.2870.1790.2630.2870.3660.1930.2920.1750.2590.1750.258
1920.1320.2740.2180.2880.2220.2960.2500.3090.2690.3280.2500.3090.4140.4920.2840.3620.2410.3020.2370.299
3360.2440.3820.2710.3220.2770.3330.3110.3480.3250.3660.3120.3490.5970.5420.3690.4270.3050.3430.2980.340
7200.2680.4080.3610.3800.3710.3890.4120.4070.4210.4150.4110.4061.7301.0420.5540.5220.4020.4000.2750.323
+24.39%Avg0.1860.3240.2540.3100.2590.3190.2880.3320.3050.3490.2880.3320.7570.6100.3500.4010.2810.3260.2460.306
ETTh1960.2980.2770.3680.3950.3860.4050.3760.4190.3860.4050.4230.4480.3860.4000.4140.4190.3750.400
1920.3370.3060.4010.4160.4410.4360.4200.4480.4430.4370.4710.4740.4370.4320.4600.4450.4790.421
3360.3950.3340.4220.4370.4870.4580.4590.4650.4890.4680.5700.5460.4810.4590.5010.4660.4840.458
7200.4710.3640.4410.4650.5030.4910.5060.5070.5020.4890.6530.6210.5190.5160.5000.4880.4980.482
+8.09%Avg0.3750.3200.4080.4280.4540.4470.4400.4600.4550.4500.5290.5220.4560.4520.4690.4540.4590.440
ETTh2960.1940.3380.2730.3330.2970.3490.3580.3970.2960.3480.7450.5840.3330.3870.3020.3480.2890.341
1920.2380.3840.3400.3780.3800.4000.4290.4390.3760.3960.8770.6560.4770.4760.3880.4000.3720.392
3360.2580.3940.3730.4030.4280.4320.4960.4870.4240.4311.0430.7310.5940.5410.4260.4330.3860.414
7200.3140.4320.3980.4300.4270.4450.4630.4740.4260.4441.1040.7630.8310.6570.4310.4460.4120.434
+27.46%Avg0.2510.3870.3460.3860.3830.4070.4370.4490.3810.4050.9420.6840.5590.5150.3870.4070.3640.395
Table 6. Structural ablation. Comparison of KOSLM and its structural variants on four long-term forecasting datasets ( L = 720 ). Removing either the Kalman gain path or the innovation input leads to consistent performance degradation. All results are averaged over five runs with mean ± standard deviation. Lower values (indicated by the down arrow “↓”) correspond to better performance.
Table 6. Structural ablation. Comparison of KOSLM and its structural variants on four long-term forecasting datasets ( L = 720 ). Removing either the Kalman gain path or the innovation input leads to consistent performance degradation. All results are averaged over five runs with mean ± standard deviation. Lower values (indicated by the down arrow “↓”) correspond to better performance.
DatasetModelMSE ↓MAE ↓Δ MSE vs. Full
ETTm1Full0.498 ± 0.0110.416 ± 0.007
No-Gain0.662 ± 0.0170.553 ± 0.014+32.9%
No-Innov0.583 ± 0.0120.487 ± 0.008+17.1%
ETTh1Full0.510 ± 0.0120.378 ± 0.010
No-Gain0.605 ± 0.0220.448 ± 0.015+18.6%
No-Innov0.554 ± 0.0160.411 ± 0.010+8.6%
TrafficFull0.321 ± 0.0080.195 ± 0.016
No-Gain0.428 ± 0.0110.260 ± 0.008+33.4%
No-Innov0.375 ± 0.0090.228 ± 0.007+16.9%
ExchangeFull0.266 ± 0.0140.341 ± 0.023
No-Gain0.299 ± 0.0180.383 ± 0.023+12.3%
No-Innov0.294 ± 0.0150.377 ± 0.019+10.6%
Table 7. Memory footprint under different batch sizes, input length = 2048. KOSLM exhibits lower memory usage than Transformer and Mamba-2, while remaining comparable to LSTM despite its smaller size.
Table 7. Memory footprint under different batch sizes, input length = 2048. KOSLM exhibits lower memory usage than Transformer and Mamba-2, while remaining comparable to LSTM despite its smaller size.
Batch SizeTransformer (GB)Mamba-2 (GB)LSTM (GB)KOSLM (GB)
10.2230.0850.0810.061
20.3630.1580.1090.104
40.6400.2900.1660.188
81.1800.5610.2830.344
162.2561.1030.5180.668
324.4082.1880.9871.317
648.7124.3571.9262.615
12817.0008.6963.8035.211
Table 8. Model parameter sizes. KOSLM remains lightweight while providing competitive long-horizon forecasting performance.
Table 8. Model parameter sizes. KOSLM remains lightweight while providing competitive long-horizon forecasting performance.
ModelParameters (M)
LSTM2.17
Transformer5.33
Mamba-20.21
FiLM1.50
KOSLM0.24
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Tan, X.; Wang, L.; Wang, M.; Zhang, Y. KOSLM: A Kalman-Optimal Hybrid State-Space Memory Network for Long-Term Time Series Forecasting. Appl. Sci. 2025, 15, 12684. https://doi.org/10.3390/app152312684

AMA Style

Tan X, Wang L, Wang M, Zhang Y. KOSLM: A Kalman-Optimal Hybrid State-Space Memory Network for Long-Term Time Series Forecasting. Applied Sciences. 2025; 15(23):12684. https://doi.org/10.3390/app152312684

Chicago/Turabian Style

Tan, Xin, Lei Wang, Mingwei Wang, and Ying Zhang. 2025. "KOSLM: A Kalman-Optimal Hybrid State-Space Memory Network for Long-Term Time Series Forecasting" Applied Sciences 15, no. 23: 12684. https://doi.org/10.3390/app152312684

APA Style

Tan, X., Wang, L., Wang, M., & Zhang, Y. (2025). KOSLM: A Kalman-Optimal Hybrid State-Space Memory Network for Long-Term Time Series Forecasting. Applied Sciences, 15(23), 12684. https://doi.org/10.3390/app152312684

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop