Previous Article in Journal
AI-Powered Learning: Revolutionizing Education and Automated Code Evaluation
Previous Article in Special Issue
Deep Learning for Regular Raster Spatio-Temporal Prediction: An Overview
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Uncertainty-Aware QoS Forecasting with BR-LSTM for Esports Networks

Department of E-Sports Technology, Cheng Shiu University, Kaohsiung 833, Taiwan
Information 2025, 16(12), 1016; https://doi.org/10.3390/info16121016
Submission received: 19 October 2025 / Revised: 17 November 2025 / Accepted: 18 November 2025 / Published: 21 November 2025
(This article belongs to the Special Issue New Deep Learning Approach for Time Series Forecasting, 2nd Edition)

Abstract

Reliable forecasting of network QoS indicators such as latency, jitter, and packet loss is essential for managing real-time and risk-sensitive applications. This study addresses the challenge of uncertainty quantification in QoS prediction by proposing a Bayesian Regression-enhanced Long Short-Term Memory (BR-LSTM) framework. The method integrates Bayesian mean variance estimates into sequential LSTM learning to enable accurate point forecasts and well-calibrated confidence intervals. Experiments are conducted using a Mininet-based emulation platform that simulates dynamic esports network environments. The proposed model is benchmarked against ten probabilistic and deterministic baselines, including ARIMA, Gaussian Process Regression, Bayesian Neural Networks, and Monte Carlo Dropout LSTM. Results demonstrate that BR-LSTM achieves competitive accuracy while providing uncertainty intervals that improve decision confidence for Service-Level Agreement (SLA) management. The calibrated upper bound  ( μ + k σ )   can be compared directly against SLA thresholds to issue early warnings and prioritize rerouting, pacing, or bitrate adjustments when the bound approaches or exceeds policy limits, while calibration controls false alarms and prevents unnecessary interventions. The findings highlight the potential of uncertainty-aware forecasting for intelligent information systems in latency-critical networks.

Graphical Abstract

1. Introduction

Esports and cloud-gaming impose stringent end-to-end Quality of Service (QoS) requirements [1]. Millisecond-scale latency spikes, jitter bursts, and modest loss already disrupt responsiveness and perceived quality [2,3,4,5,6]. Operational QoS (RTT, throughput, access heterogeneity) varies over time and by path. Purely reactive control is therefore insufficient to maintain high-quality interactivity [7,8]. Operational reports further indicate that Quality of Experience (QoE) management is difficult to maintain with reactive mechanisms alone [9,10]. These observations motivate a predictive, risk-aware approach to QoS management for latency-critical interactive workloads.
Transport-layer design also shapes latency and loss. High-speed transports built on User Datagram Protocol (UDP), such as UDP-based Data Transfer (UDT), target high throughput with low delay in wide-area environments, whereas Reliable-UDP (RUDP) variants trade additional latency for stability [11,12]. Transport-layer choices involve latency–stability trade-offs. Transport configuration alone is therefore insufficient under rapidly fluctuating conditions, reinforcing the need for proactive prediction and control.
Recent studies have explored machine learning-based forecasting of QoS metrics, namely, latency, throughput, and loss, to trigger preemptive actions before degradation appears. Representative directions include Long Short-Term Memory (LSTM) sequence models for service-level or platform-level QoS forecasting [13], temporal transformer architectures that address non-stationarity and long-range dependencies [14], federated or hierarchical learning for distributed and privacy constrained settings [15], and Graph Neural Networks (GNNs) that encode topological and relational structure to remain robust under sparse or noisy observations [16]. Reputation-aware graph formulations further stabilize predictions in challenging regimes [17]. Taken together, these studies indicate that machine learning provides an effective basis for proactive QoS management under dynamic network conditions.
Despite recent progress, many studies still prioritize point prediction and offer limited guidance on predictive uncertainty and decision confidence. This is critical for latency-sensitive control loops such as preemptive switching, congestion pacing, and deadline-aware arbitration. Gaussian Processes (GP) provide probabilistic forecasts with posterior means and variances [18]. In deep learning, dropout is used as a Bayesian approximation, and uncertainty decomposition provides general mechanisms for estimating and interpreting predictive risk [19,20]. However, domain-tailored frameworks that jointly model temporal dependence and deliver calibrated risk bounds suitable for esports and cloud gaming operations remain comparatively scarce. This gap motivates the present study.
Compared with alternative uncertainty-aware temporal models, BR-LSTM injects per-step predictive summaries (μ, σ) at the input level and lets the sequence model propagate these uncertainty signals. Monte Carlo Dropout (MC-Dropout) LSTM represents epistemic uncertainty through weight stochasticity and typically requires multiple stochastic passes, while Gaussian Process Regression (GPR) encodes uncertainty in kernel posteriors with higher memory and scalability costs on multivariate sequences. By delivering calibrated μ + upper bounds and propagating uncertainty through time, BR-LSTM provides decision-useful intervals under regime shifts while remaining lightweight for edge deployment.
Different game genres respond to network volatility in distinct ways. Massively Multiplayer Online Role Playing Games (MMORPGs), such as World of Warcraft and Black Desert, reduce real-time interaction to mitigate the effects of latency and jitter [7]. In contrast, real-time strategy and racing games rely on dead reckoning and action prediction to maintain consistency [8,21]. Mobile multiplayer games that operate over cellular networks further amplify latency variability because of wireless channel fluctuations [22]. This issue is exacerbated in cloud gaming environments by significant variations in Round Trip Time (RTT) [23,24]. In wireless-mesh backbone networks, LSTM-based routing has proven effective for QoS-aware path selection, outperforming classical protocols such as Ad hoc On-demand Distance Vector (AODV) by achieving higher packet delivery rates and throughput under dynamic conditions [25]. Moysen et al. [14] addressed this challenge by using location-independent User Equipment (UE) metrics, such as Reference Signal Received Power (RSRP) and Reference Signal Received Quality (RSRQ), to predict throughput in Long Term Evolution (LTE) heterogeneous networks (HetNets) without relying on Global Positioning System (GPS) data. This approach improves scalability and cost efficiency in mobile QoS prediction.
To address this gap, this work introduces a BR-LSTM framework. A front-end Bayesian Regression (BR) produces per-sample analytic means and variances that are then consumed by an LSTM to capture nonlinear temporal dependence. This design enables analytic variance propagation through the sequence and yields accurate point forecasts together with calibrated prediction intervals. Evaluation is conducted in a Mininet-based emulation environment under diverse network conditions. Comparative baselines include Support Vector Regression (SVR) [26], Random Forest (RF) [27], Multilayer Perceptron (MLP) [28,29], standard LSTM [13], Bayesian Neural Network (BNN) [30], MC-Dropout LSTM [19], GPR [18,31], and Quantile Regression (QR) [32], together with graph and distributed learning settings for completeness [17,33,34]. Metrics comprise Mean Absolute Error (MAE), Root Mean Squared Error (RMSE), and the Coefficient of Determination (R2), together with risk-centric indicators such as Prediction Interval Coverage Probability (PICP), Mean Prediction Interval Width (MPIW), and False Alarm Rate (FAR). To anchor the comparison with foundational time-series models that provide uncertainty natively, the study includes Autoregressive Integrated Moving Average (ARIMA) [35] and a linear state–space (Kalman) [36] baseline. Their point-estimation results are reported in Table A1 using the same split and metrics as the learning-based models. Unlike MC-Dropout LSTM, which approximates posterior uncertainty via stochastic masking, and unlike GPR, which models uncertainty through a kernel on raw signals, the proposed BR-LSTM first derives analytic per-sample means and variances with Bayesian Ridge and then propagates  ( μ t , σ t )  through a sequential LSTM. Injecting uncertainty at the input layer preserves the interpretability of  μ  and  σ  and allows nonlinear temporal dependence to refine one-step-ahead risk bounds.
The main contributions of this study are summarized as follows:
  • This study introduces a hybrid QoS forecasting model, BR-LSTM, which combines BR-derived uncertainty estimates with a sequential LSTM to enable calibrated, risk-aware prediction under dynamic network conditions.
  • A comprehensive evaluation is performed against ten established baseline models, including ARIMA and a local level state–space model, within a unified Mininet-based simulation framework. The assessment covers point accuracy (MAE, RMSE, R2) and interval quality (PICP, MPIW, FAR).
  • After calibration to target 95% PICP on a validation set, BR-LSTM attains near target coverage for latency and loss and improved coverage for jitter, while maintaining a small model footprint (0.07 MB) and low inference cost, which supports real-time edge deployment in esports networks. The calibrated upper bound aligns with SLA thresholds and provides an actionable trigger for early warnings with controlled false-alarm rates.
The remainder of this paper is organized as follows. Section 2 reviews related work on QoS forecasting and uncertainty-aware models and summarizes the nomenclature. Section 3 details the proposed BR-LSTM methodology, including the problem formulation, the BR front end, LSTM sequence modeling, and interval calibration. Section 4 describes the Mininet-based emulation and dataset generation with tc netem, the training and evaluation protocol, and the comparative results across MAE/RMSE/R2, and risk metrics, including PICP, MPIW, FAR. Section 5 discusses the findings, emphasizing the coverage and sharpness trade-off, service level agreement (SLA) implications, and deployment considerations. Finally, Section 6 concludes the paper and outlines directions for future work.

2. Related Work

Interactive and cloud gaming are highly sensitive to end-to-end QoS fluctuations. Millisecond-scale impairments already degrade responsiveness and perceived quality. Field measurements show that access technology and mobility amplify temporal variability and path dependence. These effects make imperceptible latency difficult to sustain in mobile cloud gaming [37,38,39,40]. From an operational standpoint, reports from production ecosystems highlight the limits of purely reactive QoE and QoS controls during long-running sessions. Prior works link fine-grained QoS dynamics to perceptual outcomes and document frequent regime shifts under heterogeneous access and mobility. These findings motivate predictive, risk-aware control rather than post hoc mitigation.
This work builds on these findings by framing QoS forecasting as a calibrated risk estimation problem that outputs decision useful envelopes for latency-critical operations. One line of research focuses on time-series forecasting of QoS [29]. LSTM-based sequence models have been applied to forecast service level or platform level QoS attributes and have reported improvements over classical regressors when temporal dependence is strong [13,41].
Production pipelines that span telemetry to deployment show that performance forecasting can trigger preemptive traffic actions before degradation [23]. For real-time communications, packet-level modeling of concurrent Real-time Transport Protocol (RTP) flows has been used to obtain end-to-end QoS predictions that respect in-session constraints [42]. In edge-to-cloud scenarios, context-aware collaboration leverages auxiliary signals to mitigate geographic and uplink variability during QoS prediction [43]. In mobile settings, adding contextual signals likewise improves QoS predictability [44]. Complementary work in service ecosystems leverages multi-stage and multi-scale feature fusion or multi-modal integration to forecast QoS under diverse usage patterns [24,25]. Together, these studies indicate that learned temporal models are effective tools for forward-looking QoS management under dynamic conditions. Classical ARIMA and linear state–space models provide linear Gaussian baselines with native prediction intervals and serve as sanity checks. Both are included in the evaluation, as summarized in Table A1.
A second line of work exploits structural priors and distributed learning. GNN formulations encode topological and relational structures to improve robustness under sparse or noisy observations, and reputation-aware graph convolution has been used to enhance stability in challenging regimes [16,17]. Topology-aware recurrent designs, for example, the Topology-Aware QoS GRNN (TAQ-GRNN), integrate the network structure with gated recurrence to improve QoS prediction [45]. In systems that span multiple organizations or sites, federated and hierarchical learning supports privacy preserving and scalable QoS prediction [15]. Related efforts apply graph learning in cloud environments to anticipate SLA violations, showing that relational information can surface risk earlier than node-level predictors [46]. Under mobility and concept drift, distributed predictive schemes maintain accuracy with modest communication and energy overheads in automotive settings [47]. These results suggest that structural information and decentralized training are practical levers for robust QoS forecasting in the wild.
Despite these advances, much of the forecasting literature remains point estimate-centric and provides limited guidance on predictive uncertainty and the decision confidence required by latency-critical control loops. GP offers probabilistic forecasts with posterior means and variances but faces scalability challenges in large multivariate contexts [18]. Within deep learning, dropout is used as a Bayesian approximation, and uncertainty decomposition provides general mechanisms to estimate and interpret predictive risk, yet domain-tailored evaluations of calibrated intervals for esports and cloud gaming operations remain scarce [19,20]. This gap motivates models that jointly capture temporal dependence while yielding calibrated and decision-useful risk bounds, which is the focus of the present work.

3. Methodology

Although many models have been applied successfully to QoS prediction, most methods emphasize point estimation and overlook Uncertainty Quantification (UQ). This limitation is especially critical in dynamic and risk-sensitive applications such as esports and cloud gaming, where robust decision making depends not only on accurate predictions but also on reliable and well-calibrated confidence intervals. To address this gap, this paper presents a hybrid framework that integrates BR-LSTM networks. The proposed approach produces well-calibrated uncertainty intervals without sacrificing predictive accuracy.
This study focuses on three key QoS indicators in esports network environments: latency, jitter, and packet loss rate. To combine the time-series modeling strength of LSTMs with the uncertainty quantification capability of Bayesian methods, the BR-LSTM framework proceeds as follows. A BR stage first produces per-time step point estimates and uncertainty estimates for each QoS metric, denoted as  μ ^ t  and  σ ^ t . These estimates are concatenated to form an enhanced feature vector that serves as input to the LSTM. The intuition is that providing both the per-step mean and variance supplies the LSTM with a baseline trend for each metric, upon which it can model nonlinear temporal dependence. Because the BR stage conditions only on the time index, it may miss abrupt traffic surges or routing changes. The LSTM is therefore trained to predict the next step mean and uncertainty, i.e.,  μ ^ t + 1  and  σ ^ t + 1 , from these enriched input sequences [48]. The objective is to improve the accuracy of dynamic network modeling while simultaneously providing reliable uncertainty and risk quantification.
In the proposed framework, BR is first used to model the raw network features at each time step. With Gaussian priors, BR regularizes the parameters and infers their posterior distributions during training. This procedure yields two essential outputs for each QoS indicator, namely, the mean prediction and the associated prediction variance. These uncertainty measures capture potential fluctuations and anomalies in network behavior and serve as critical inputs for subsequent temporal modeling.
Next, the predicted means and variances for latency, jitter, and packet loss are concatenated to form a six-dimensional enhanced feature vector, which is then provided to the LSTM model. Using a sliding window of length L, the LSTM learns nonlinear temporal dependencies and long-term memory patterns among the QoS metrics and produces predictions for future time points. The final output comprises forecasts for all three QoS parameters, enabling forward-looking evaluation of esports network performance.
Placing BR before the LSTM provides real-time, input-level uncertainty estimates. This design enhances the interpretability and robustness of temporal modeling, especially in highly variable environments. Compared with feeding raw QoS data directly into an LSTM, the proposed architecture improves generalization and is well-suited for latency-sensitive and high-reliability applications, including esports platforms, streaming game servers, and cloud gaming infrastructures.

3.1. Problem Formulation

Let  y t = [ latenc y t ,   jitte r t ,   los s t ] R 3  denote the QoS vector at time t. Given a sliding window  X t = [ y t L + 1 , , y t ] , the objective is to learn the one-step-ahead predictive distribution  p ( y t + 1 X t ) N ( μ ^ t + 1 , diag ( σ ^ t + 1 2 ) ) , where  μ ^ t + 1 , σ ^ t + 1 R 3  are the per-metric mean and standard deviation for time t + 1 given history Xt, respectively.
For risk-aware control, the upper bound  μ ^ t + 1 + k σ ^ t + 1  is compared with service-level thresholds. The scalar k is chosen on a validation split to target a PICP of approximately 0.95. The notation and the priors used in the Bayesian stage are summarized in Table 1.

3.2. BR-LSTM

The framework combines a per-timestep BR with a sequential LSTM module, allowing uncertainty information to propagate over time. For each metric  ν latency , jitter , loss , a Bayesian linear model with a Gaussian prior is used as shown in (1).
y ( ν ) = X w ( ν ) + ϵ , ϵ N 0 , σ n 2 , w ( ν ) N 0 , λ 1 I
Given observations  ( X , y ( ν ) )  with  X R N × d  and  y ( ν ) R N , where N is the number of BR training samples and d is the input dimension, the posterior over weights is Gaussian, with mean  μ w ( ν )  and covariance  w ( ν ) . The predictive distribution for a new input  x *  is summarized in (2). All symbols used in (2)–(5) follow the notation summarized in Table 1.
μ ^ * , ( ν ) = ( x * ) μ w ( ν ) , ( σ ^ * , ( ν ) ) 2 = σ n 2 + ( x * ) w ( ν ) x *
For each time step t (taking  x * = x t ), the time-indexed BR summaries that feed the LSTM are given in (3).
μ ^ t * , ( ν ) = ( x t ) μ w ( ν ) , σ ^ t * , ( ν ) = σ n 2 + ( x t ) w ( ν ) x t
BR with a Gaussian likelihood and a zero-mean isotropic Gaussian prior on the weights is adopted. Conditioning on  X , y  yields the posterior  N μ w ( v ) , Σ w ( v ) , with  Σ w ( v ) = ( λ I + σ n 2 X X ) 1  and  μ w ( v ) = σ n 2 Σ w ( v ) X y ( v ) . The predictive distribution at x* is  N ( ( x * ) μ w ( v ) , σ n 2 + ( x * ) Σ w ( v ) x * ) . These summaries define  μ ^  and  σ ^  used in (2) and (3).
Stacking the three metrics forms a six-dimensional feature.  z t = [ μ t * , ( lat ) , σ t * , ( lat ) , μ t * , ( jit ) , σ t * , ( jit ) , μ t * , ( loss ) , σ t * , ( loss ) ] R 6 . The LSTM consumes  Z t = [ z t L + 1 , , z t ]  and outputs the next step mean and standard deviation as in (4).
μ ^ t + 1 , σ ^ t + 1 = LST M θ Z t , μ ^ t + 1 , σ ^ t + 1 R 3
Under the Gaussian prior-likelihood setting of Bayesian Ridge, the predictive variance decomposes as  V a r [ y * x * ] = σ n 2 + x * Σ w x * . The term  σ n 2  represents aleatoric noise, whereas  x * Σ w x *  captures epistemic uncertainty from the parameter posterior. The BR front end, therefore, supplies per-timestep  ( μ t , σ t )  signals that already contain both components. The sequence model consumes these inputs and maps them to  ( μ ^ t + 1 , σ ^ t + 1 ) ; the resulting  σ ^ t + 1  reflects propagated total predictive uncertainty rather than a separated decomposition. Interval calibration with a single factor k on a disjoint validation split corrects residual miscalibration before evaluation on the test split.
Let  μ ^ t + 1  and  σ ^ t + 1  denote model outputs. The loss combines point accuracy and interval coverage as in (5). Here, k denotes the validation-calibrated interval scaling factor, and  λ c o v  the coverage weight in the loss. Both are defined in Table 1.
L = MAE μ ^ , y + λ c o v PICP μ ^ ± k σ ^ 0.95
At inference, BR yields six scalars per step, and the LSTM produces  ( μ ^ t + 1 , σ ^ t + 1 ) R 3 × R 3 . The per-step cost is dominated by the LSTM forward pass O(LH + H2) with hidden size H; the BR stage is closed-form and lightweight. The overall footprint meets edge-deployment constraints with millisecond-level latency budgets.
Unless noted otherwise, experiments use a one-layer LSTM (hidden size 64) with six input features and a look-back window  L = 10 . Optimization uses Adam. The BR stage adopts Bayesian ridge (Gaussian) priors with default hyperparameters. All features are standardized, and the train/validation/test splits follow the evaluation protocol. Figures show one-step-ahead rolling forecasts on the held-out test split for latency, jitter, and packet loss. At time  t , the model outputs a point estimate  y ^ t  and an uncertainty interval  [ y ^ t k σ ^ t ,   y ^ t + k σ ^ t ] . The scale factor k is tuned once on the validation set to achieve 95% empirical coverage and is then fixed for testing. Using this global k avoids label leakage, preserves cross-model comparability, and yields intervals that combine model uncertainty with a fixed post hoc calibration. No dynamic adjustment of k is applied at test time, and the test split remains unseen during training and calibration. For service-level interpretation, horizontal lines mark the SLA thresholds for each metric. These plots complement the quantitative tables by indicating when the upper risk bound  y ^ t + k σ ^ t  crosses the SLA, which directly relates to FAR. Reported metrics include PICP, MPIW, and FAR.

3.3. Baseline Models/Comparison Methods

To contextualize the performance of the proposed BR-LSTM model, ten representative baseline methods were re-implemented using identical training and validation splits, standard scaler normalization, and the same 1000-round dataset generated via Mininet simulation. Unless otherwise specified, all deterministic models were trained using the default settings provided by scikit-learn version 1.6.1. The complete hyperparameters and training settings for every baseline are summarized in Table 2.
For each model, five evaluation metrics were computed on the standardized test set: MAE, RMSE, R2, PICP, and MPIW. Predictions were then transformed back to physical units by reversing the normalization. For BR-based models and the MC-Dropout LSTM, the prediction interval is defined as  μ ^ ± k σ ^ , with k chosen on the validation set to target 95% coverage. When comparing to SLA thresholds, the upper risk bound  μ ^ + k σ ^  is reported. For QR, the model uses its τ quantile output directly as the upper risk bound.
Models without native uncertainty are evaluated only with point estimation metrics (Table A1). For models that produce predictive uncertainty, including BR-LSTM, BNN, MC-Dropout LSTM, GPR, ARIMA, and the state–space model, interval metrics are reported in Table A2. Table 3 summarizes risk indicator availability, and Table 4 provides the ranking.
This unified experimental protocol standardizes data preprocessing and training conditions, thereby ensuring that any performance differences can be attributed solely to the architectural characteristics of each model. As a result, any observed performance improvements in BR-LSTM can be attributed to its hybrid design and uncertainty-aware formulation, rather than to favorable hyperparameter configurations.
The study reports point metrics (MAE, RMSE,  R 2 ) together with interval metrics (PICP, MPIW, FAR) relative to SLA thresholds. Because the held-out test series exhibit temporal dependence and occasional regime shifts, formal null-hypothesis tests on per-sample errors under i.i.d. assumptions may be anti-conservative. Accordingly, cross-model comparisons emphasize consistent effect directions and relative gaps across latency, jitter, and loss, while calibrated interval behavior is read alongside point accuracy. Serial-dependence-aware paired tests (e.g., moving-block bootstrap around paired error differences) are left to future work to quantify sampling uncertainty without assuming independence.

3.4. Computational Footprint and Relative Cost

A single forward pass is dominated by the LSTM term  O ( L H + H 2 ) , while the Bayesian Ridge front end remains closed form and lightweight. Under the reported configuration  ( L = 10 , H = 64 ) , the overall footprint fits millisecond-level latency budgets on commodity CPUs used for edge dashboards.
The model uses a lightweight configuration  ( L = 10 , H = 64 )  to balance accuracy, stability, and edge deployability while avoiding overtuning that could bias cross-model comparisons. In this setting, L controls the receptive field of temporal dependencies, whereas H governs capacity and the  O ( H 2 )  cost of the recurrent core. Under moderate changes to these settings, the claims in this study are framed around directional consistency and rank-order stability rather than exact point values. Interval behavior remains governed by the calibration procedure on the validation split, and the reported results, therefore, keep L and H fixed for fairness. Fuller ablations are left for future work.
To contextualize the models, a qualitative ranking by inference cost is provided in prose. Low-cost models include ARIMA and state space approaches with closed-form updates, RF with shallow depth, and a small MLP. Medium-cost models include SVR, whose inference depends on the number of support vectors, and the BR-LSTM variant with a single-layer LSTM. High-cost models include MC-Dropout LSTM, which requires multiple stochastic passes; BNNs, which rely on sampling; and GPR with large effective kernel sizes. This ranking reflects standard complexity terms and the reported architectural settings; no additional measurements are introduced.

4. Experiments

4.1. Experimental Settings

To rigorously evaluate the proposed BR-LSTM framework under realistic and dynamic network conditions representative of multiplayer esports environments, a customized multiplayer network emulation was developed by using Mininet. This emulation replicates competitive online gaming interactions and captures fluctuations in delay, jitter, and packet loss.
A star topology was employed, consisting of six player hosts (h1–h6) connected to a central game server (server1) via an OpenFlow-enabled switch (s1), as detailed in Table 5. Each player node communicates directly with the central server through the switch, using either TCP or UDP. Although the emulation supports both protocols, this study does not involve application-layer traffic generation. Instead, link conditions were configured directly using the tc netem utility to simulate network impairments. All links were instantiated using Mininet’s TCLink class, enabling fine-grained, dynamic control over key link-level parameters, including delay, jitter, and packet loss rate.
Network conditions were systematically varied in each simulation round to emulate realistic traffic dynamics and disturbances. Specifically, the link-level delay was modulated using a sinusoidal pattern to simulate cyclical network congestion.
The delay at round r was computed as follows:  D r = 30 + 20 × s i n 2 π r 200 + n o i s e , n o i s e U 20 , 20 , where 30 ms is the baseline delay, 20 ms is the modulation amplitude, 200 rounds define the period, and the additive noise is uniformly sampled from [−20,20] ms. Any delay value below 0 ms was clipped using Dr = max (Dr, 0 ms) to avoid non-physical negative delays. Jitter for each player was simulated as a sinusoidal signal with a mean of 20 ms, an amplitude of 10 ms, and a period of 100 rounds, superimposed with Gaussian noise ε~𝒩 (0, 5 ms). In addition, approximately 5% of rounds included jitter spikes ranging from 20 ms to 50 ms to emulate sporadic network turbulence. Packet loss was randomly injected in 20% of rounds, with the loss rate uniformly sampled between 10% and 80%.
To emulate severe degradation scenarios, every 30 rounds, one randomly selected player experienced either an additional 300 ms delay or elevated packet loss between 50% and 80%. Furthermore, in 10% of all rounds, extreme network anomalies were simulated by randomly assigning a player host either a complete disconnection (100% packet loss) or a delay spike of 500 ms.
Measurement metrics were derived directly from the network conditions configured at each simulation round. Crucially, the network QoS parameters were explicitly set and controlled for each player and the server via their respective virtual network interfaces (eth0), using the Linux Traffic Control (tc) utility and the netem (Network Emulator) queuing discipline.
It is important to note that this simulation focused exclusively on the controlled configuration and logging of QoS parameters. No application-layer traffic (e.g., generated by iperf) was transmitted between hosts, nor were active measurement tools (such as ping) used to estimate latency or packet loss. The primary objective was to generate a controlled dataset that reflects diverse and fluctuating network conditions, following predefined probabilistic rules, for the purpose of training and evaluating QoS prediction models. The dataset aims to compare methods and examine coverage-width behavior under diverse perturbations rather than to exhaustively capture all gameplay regimes. A lightweight configuration  ( L = 10 , H = 64 )  and conservative training settings reduce overfitting risk and help preserve fairness across baselines.
Average RTT was approximated as twice the mean one-way delay, under the assumption of symmetric paths typical in LAN-style esports environments. Jitter and packet loss rates were likewise derived directly from the configured tc parameters, thereby offering accurate and consistent ground truth values without the variability inherent in active measurement tools.
The 1000-round simulation yielded 1000 aggregated records, each summarizing latency, jitter, and packet loss across all six players. This dataset serves as the foundation for both training and rigorous evaluation of QoS forecasting models. Table 6 presents descriptive statistics of the core QoS features, including average latency (ms), jitter (ms), and packet loss rate (%). The dataset spans a wide range of conditions, with latency between 1 and 299 ms, jitter between 6 and 47 ms, and packet loss rates between 0 and 49%. These variations are consistent with the stochastic and periodic configurations applied in the simulation and allow for comprehensive model evaluation under both stable and highly variable network scenarios.
This simulation framework offers a comprehensive and reproducible environment for evaluating the accuracy and risk estimation capabilities of various models under network dynamics that approximate real-time esports scenarios. The pseudocode presented in Figure 1 illustrates the unified pipeline used for training, validating, and comparing multiple predictive models for risk-aware QoS forecasting. This pipeline incorporates both conventional regression models and uncertainty-aware architectures, including BR-LSTM and BNN.

4.2. Experimental Details

This study focuses on predictive modeling and risk analysis of three key QoS indicators, namely, latency, jitter, and packet loss. The experimental dataset was generated with Mininet-based network simulations that emulate realistic and dynamic conditions typical of multiplayer online gaming environments. Bayesian ridge regression was first applied independently to each indicator to estimate the predictive mean  μ ^  and standard deviation  σ ^ , thereby introducing uncertainty information into the features. These statistics provide the inputs for subsequent modeling.
The BR-LSTM model accepts a six-dimensional input at each time step that contains the predicted means and standard deviations for latency, jitter, and packet loss. This format allows the model to represent both expected values and their associated uncertainties over time. Other baselines, including SVR, RF, MLP, QR, BNN, GPR, and standard LSTM variants, as well as ARIMA and a linear state–space Kalman model, were implemented using either the enhanced features or the original QoS metrics.
Feature representations were arranged in flattened or sequential form according to each model’s input requirements. A concise taxonomy grouped by learning paradigm and temporal capability is provided in Table 7. The availability of probabilistic risk indicators for each algorithm, namely,  μ ^ ± k σ ^  or  q τ  (τ = 0.95), is summarized in Table 3.
Sequential prediction used a sliding window of length ten to forecast the QoS state at time t + 1. To prevent data leakage, feature standardization parameters were fit on the training split only and then applied to the test data.
The dataset was partitioned into 80% training and 20% testing by random shuffling with a fixed seed to ensure statistical diversity and reproducibility. This partitioning and preprocessing protocol was applied consistently across all model training and evaluation procedures. Per-sample inference latency was measured on the CPU with a single thread and a batch size of one. Each model was warmed up and then timed over repeated runs to record the mean and the 95th percentile (p95) latency. The qualitative “Inference cost” labels in Table 8 are derived from these measurements using fixed thresholds: Low ≤ 1 ms; 1–5 ms = Medium; >5 ms = High. For stochastic predictors, the latency reflects the evaluated configuration. MC-Dropout LSTM and BNN use S = 20 stochastic passes for predictive averaging in this study (see Table 2). Table 8 reports the single-pass latency; executing S passes on the CPU increases the per-sample wall time approximately S-fold.
Model configurations and hyperparameters were selected based on preliminary experiments. Bayesian ridge regression used default values for the prior precision and the noise variance. The BR-LSTM model used an input size of six, 64 hidden units, a single hidden layer, a batch size of 16, and a learning rate of 0.001, and was trained for 30 epochs.
The same number of training epochs was applied to the standard LSTM (input size three) and to the MC-Dropout LSTM (input size three, dropout rate 0.3). The BNN accepted a flattened 30-dimensional input vector derived from a 10-step window of the three QoS features and was trained for one hundred epochs on the training split. GPR was configured with default kernel settings. The QR model targeted the ninety-fifth percentile and used a regularization parameter α = 0.1. The MLP consisted of a single hidden layer with one hundred neurons and ReLU activation, optimized with the Adam optimizer for five hundred iterations. The RF used one hundred trees with no maximum depth constraint. The SVR employed an RBF kernel with default regularization parameters.
Model performance was evaluated with standard regression metrics, including MAE, RMSE, and R2. All scores were computed after applying the inverse standardization to restore the original units. In addition to point estimation, risk was assessed with Bayesian prediction intervals of the form  μ ^ ± k σ ^ , where  μ ^  and  σ ^  denote the predicted mean and standard deviation. The scaling factor k was tuned on the validation set to target approximately 95% coverage. For QR, risk evaluation used the 95th percentile prediction qτ with τ = 0.95, which directly provides the upper bound of the QoS indicators.
For models that support uncertainty estimation, including BR-LSTM, MC-Dropout LSTM, BNN, and GPR, additional metrics PICP, MPIW, and FAR were reported. These metrics were computed relative to the SLA thresholds of 100 ms for latency, 50 ms for jitter, and 10% for packet loss. While MAE, RMSE, and R2 assess point forecast accuracy, PICP captures the empirical coverage of the prediction intervals and MPIW measures their average width, reflecting sharpness. FAR quantifies the fraction of alarms that do not correspond to actual SLA violations. A reliable model should achieve high coverage with narrow intervals and a low FAR, thereby balancing calibration and informativeness.
In practice, the value of k is not fixed because the predictive uncertainty may deviate from a strict Gaussian distribution. Therefore, empirical calibration is performed by tuning k on the validation set. For a target of 95% coverage, we define the one-sided bounds as  L i = μ ^ i k σ ^ i  and  U i = μ ^ i + k σ ^ i , with k  [1, 3] chosen to minimize |PICP−0.95|. The calibrated interval is then evaluated on the test set. The risk-centric indicators are defined in (6)–(8), where N is the number of test samples,  I ·  is the indicator function, ϕ is the nominal miscoverage rate (for example, 0.05 for a 95% interval), and δ is the SLA upper bound for the corresponding QoS metric.
P I C P = 1 N i = 1 N I L i y i U i
M P I W = 1 N i = 1 N ( U i L i )
F A R = i = 1 N I { U i > δ , y i δ } m a x ( 1 , i = 1 N 1 { U i > δ } ) × 100
Calibration multiplies the interval half-width by the selected factor k, yielding calibrated interval width and alarm statistics (reported as MPIWcal and FARcal). Each model is evaluated separately for each QoS metric to enable a detailed assessment of performance and risk. All experiments were conducted in a controlled Python environment to ensure reproducibility.
Unless stated otherwise, baseline implementations used the default parameters of the scikit learn library (version 1.6.1). This applies to Bayesian Ridge, SVR, RF Regressor, and GPR. For the MLP regressor, the maximum number of iterations was set to 500, with all other parameters kept at their default values. The QR model used a quantile level of 0.95, a regularization parameter α = 0.1, and the “highs” solver.
The MC-Dropout LSTM baseline was implemented in PyTorch (version 2.6.0 + cu124) with a hidden size of 64 and a dropout rate of 0.3. It was trained for 30 epochs using the Adam optimizer with a learning rate of 0.001 and a batch size of 16. The implementation used T = 20 stochastic forward passes; in a pilot study,  σ ^  stabilized with less than 1% relative change beyond T = 20.
Although these configurations are common and reasonable, no systematic hyperparameter optimization was performed. Procedures such as cross-validated grid search or randomized search were not applied. Further tuning could potentially improve the performance of the baseline models.
MC-Dropout LSTM is reported as a single-pass measurement; if S stochastic passes are used to form a predictive average, the wall time on CPU scales approximately linearly with S. Bayesian neural networks similarly incur an S-dependent cost; when only a qualitative comparison is required, their inference is categorized as High under the stated thresholds.

4.3. Evaluation Results

Model performance was evaluated on three key QoS metrics, namely, latency, jitter, and packet loss rate. The evaluation employed standard regression metrics, including MAE, RMSE, and the coefficient of determination R2. Beyond point prediction accuracy, the reliability of calibrated risk intervals in maintaining compliance with SLA thresholds was also examined. Out-of-sample R2 on the test set was computed as  R 2 = 1 ( y y ^ ) 2 / ( y y ¯ test ) 2 , where  y ¯ test  denotes the test-set mean. Negative values arise when the squared error of the model exceeds the variance of the test set around its mean. This can occur on nonstationary QoS traces with abrupt regime changes and indicates that the model underperforms a mean predictor on that segment.
The results are interpreted with an emphasis on directional consistency and relative gaps rather than formal p-values. Specifically, Table A1 (in the Appendix A) summarizes point errors (MAE, RMSE, R2), Table A2 (in the Appendix A) reports calibrated interval quality (PICP, MPIW, FAR), and the SLA-oriented visuals in Figure 2, Figure 3 and Figure 4 contextualize upper-bound interactions with policy thresholds. Read together, these materials support model ranking and trade-off interpretation across latency, jitter, and loss.
In addition to MAE and RMSE, this study reports R2 as a scale-free complement. On highly nonstationary episodes with abrupt shifts, several models, including the proposed BR-LSTM, exhibit negative test set R2. Such negative values are expected in nonstationary and regime-shifting network traces of the type used in this study. Nevertheless, MAE and RMSE show that these models consistently outperform naive or persistence baselines, indicating practical forecasting value, even under challenging conditions. This behavior follows directly from the out-of-sample definition: for those segments, the squared error exceeds the variance around the test set mean. It does not contradict the absolute error improvements observed in MAE and RMSE.
As summarized in Table 4, the overall ranking consolidates point accuracy (Table A1) and risk-aware interval behavior (Table A2). BR–LSTM yields a balanced trade-off between accuracy and calibrated risk bounds; RF/MLP lead in terms of pure point accuracy but lack uncertainty; QR provides conservative upper bounds; and MC-Dropout/BNN undercovers despite the narrow intervals.
On latency, the local level state–space model attains the lowest error (MAE 24.99 ms), followed by GPR (MAE 25.58 ms). Among neural baselines, BNN reaches 28.70 ms, whereas BR-LSTM and MC-Dropout LSTM achieve 32.49 ms and 33.77 ms, respectively. Explained variance is positive for state–space (R2 = 0.20) and for GPR (R2 = 0.17), while most other methods yield smaller values; QR remains an outlier with markedly larger error (MAE 100.68 ms, R2 = −2.64). For jitter, RF provides the best point accuracy (MAE 3.05 ms, R2 = 0.74), followed by state–space (MAE 3.24 ms, R2 = 0.72) and GPR (MAE 3.33 ms, R2 = 0.71). Neural models are competitive but not leading in this regime (for example, MLP MAE 3.92 ms, R2 = 0.56; BNN MAE 4.61 ms, R2 = 0.43; BR-LSTM MAE 6.62 ms, R2 = 0.03). For packet loss, explained variance is generally limited; in MAE terms, state–space and ARIMA are strongest (7.15% and 7.23%), followed by standard LSTM (7.69%), MC-Dropout LSTM (7.68%), SVR (7.73%), and BR-LSTM (7.75%), with QR again worst (17.35%). Overall, deterministic accuracy favors state–space and GPR for latency and RF, state–space, and GPR for jitter, whereas state–space and ARIMA lead in packet loss MAE; the proposed BR-LSTM is competitive but not the top performer on MAE for this dataset.
Prediction interval quality was assessed using empirical coverage (PICP), mean interval width (MPIW), and FAR for both raw μ ± 1.96σ and calibrated μ ± bands, with k tuned to target 95% coverage. For latency, calibrated intervals reach near nominal PICP across the uncertainty-aware baselines: state–space 93.94% (MPIW 213.44 ms), GPR 93.43% (211.16 ms), ARIMA 93.43% (272.77 ms), and BR-LSTM 93.43% (198.17 ms). BR-LSTM, therefore, attains competitive coverage with the narrowest calibrated bands among these models, whereas ARIMA requires substantially wider intervals to achieve similar coverage.
For jitter, calibration yields tight, near nominal bands for GPR (PICP 95.45%, MPIW 15.57 ms) and state–space (94.95%, 16.27 ms). ARIMA also attains high coverage (97.47%) but with wider intervals (21.68 ms). BR-LSTM improves with calibration but remains under-covered on jitter (82.83%) and relatively wide (37.51 ms).
For packet loss, calibrated BR-LSTM reaches PICP 93.94%, with the tightest intervals among the four uncertainty-aware baselines (MPIW: 24.88% points), while GPR (93.43%, 26.70) and state–space (92.93%, 27.01) are comparable. ARIMA maintains high coverage (94.44%) at the cost of very wide bands (42.13). In contrast, MC-Dropout and BNN produce extremely narrow intervals but severe under-coverage on latency (for example, PICP 6.06% and 11.62%), making them unsuitable for risk-aware operations without additional uncertainty mechanisms. These outcomes illustrate the classical coverage versus width trade-off: nominal coverage can be achieved with broader bands, whereas overly tight bands degrade reliability.
In comparing calibrated upper risk bounds with the service level thresholds of 50 ms for jitter, 100 ms for latency, and 10% for loss yields consistent patterns, for jitter, the calibrated upper bounds of BR-LSTM, GPR, state–space, and ARIMA all lie below the 50 ms threshold (approximately 31.63 to 41.65 ms); the QR qτ is likewise below the threshold at 34.55 ms. For latency, BR-LSTM, at 84.36 ms, and MC-Dropout LSTM, at 75.68 ms, lie below 100 ms, whereas GPR, at 114.31 ms, state–space, at 118.87 ms, and ARIMA, at 209.71 ms, exceed the threshold, reflecting their wider calibrated bands. For packet loss, state–space at 10.22% and GPR at 10.65% sit near the 10% threshold, while BR-LSTM at 12.29%, MC-Dropout LSTM at 11.37%, ARIMA at 22.10%, and QR at 27.06% exceed it. Operationally, latency emerges as the primary risk driver. Models that maintain narrow yet calibrated envelopes during surge periods, such as BR-LSTM, provide clearer early warning signals, whereas very wide envelopes, such as ARIMA, are conservative but less actionable.
Taken together, Table 7 and Table A2 indicate that state–space and GPR lead in latency point accuracy; RF, state–space, and GPR lead in jitter accuracy; and state–space and ARIMA lead in packet loss MAE. After calibration, BR-LSTM provides competitive coverage with comparatively narrow bands for latency and loss, while GPR and state–space yield tight, near-nominal bands for jitter. ARIMA supplies a conservative envelope with high coverage but large widths. MC-Dropout and BNN exhibit severe under-coverage and would require stronger uncertainty modeling. These findings reconcile point accuracy with risk-aware behavior and can guide the choice of a forecaster depending on whether the priority is nominal tracking or SLA-oriented risk control.

5. Discussion

Across metrics, point accuracy and operational risk are only loosely coupled. Figure 4 shows that brief bursts can trigger SLA interactions even when global coverage is near nominal; conversely, very narrow bands can under-cover excursions despite favorable point errors. Together with Table A1 and Table A2, the plots show the trade-off between coverage and width. SLA monitoring should track the upper risk bound with its empirical coverage and low FAR; point errors alone are insufficient. Together with Table 4, these findings translate into an operational ranking that balances accuracy and actionable risk bounds for SLA-oriented decisions.
Given temporal dependence and regime shifts in the test traces, per-sample error tests that presume independence can misstate uncertainty. For this reason, this study refrains from reporting p-values and instead foregrounds effect directions and stable relative gaps across Table 7, Table A1 and Table A2 and Figure 2, Figure 3 and Figure 4, complemented by calibrated interval behavior against SLA thresholds. Because L mainly sets the temporal receptive field and H sets model capacity and computational cost, moderate adjustments primarily shift point errors without altering the directional conclusions emphasized in Table 7, Table A1 and Table A2 and Figure 2, Figure 3 and Figure 4. To preserve fairness across baselines, this study reports a fixed  ( L = 10 , H = 64 )  configuration and reserves exhaustive ablations for future work. Future work will incorporate paired testing (e.g., paired t-tests or Wilcoxon signed-rank) with serial-dependence-aware resampling (e.g., moving-block bootstrap) to produce confidence statements for model-to-model differences under realistic dependence.
Viewed through Figure 2 and Figure 3, BR-LSTM occupies a favorable operating point for latency assurance. After calibration, coverage is near the nominal target without uniformly wide intervals. For packet loss surveillance, among the well-calibrated baselines, it yields the tightest calibrated bands, which improves the precision of alarms during intermittent surges. For jitter, BR-LSTM is not leading; models with explicit stochastic smoothing, such as GPR and the state–space model, produce tighter yet reliable envelopes for small and frequent excursions. These patterns are consistent with the coverage and width frontier summarized in the evaluation.
Jitter is characterized by small and frequent excursions with a short correlation time. Kernel and state–space models encode local stochastic smoothing, which yields tighter and reliable jitter envelopes. In contrast, the BR-LSTM architecture increases variance where volatility emerges, such as burst onsets and regime transitions, which favors latency assurance and loss surveillance because these metrics exhibit trend-and-recovery dynamics with abrupt surges. This behavior explains why BR-LSTM is not the leading choice for jitter reliability yet attains competitive coverage with comparatively narrow calibrated bands for latency and loss. While tighter jitter control is required, context-conditioned or rolling calibration can further concentrate width on high-variance segments while preserving overall coverage.
Mechanistically, BR-LSTM couples a per-timestep BR, which supplies mean and variance cues, with a sequential LSTM that propagates these uncertainty signals through time. Calibration selects a single k on a validation split and then holds it fixed on the test split. The BR stage yields both epistemic and aleatoric components through the  σ n 2 + x Σ w x  decomposition; the sequence model propagates these signals to produce  σ ^ t + 1 , and the study calibrates the combined predictive band using the global k chosen on validation. The architecture tends to inflate variance where volatility emerges, such as burst onsets and regime transitions, rather than uniformly, which yields locally sharp yet adequately covered bands.
For jitter, GP and local level state–space models typically produce tight, well-calibrated envelopes, an advantage when small and frequent excursions dominate (see Figure 3 and Figure 4). Their memory footprint, however, is noticeably larger than compact neural baselines (Table 8). In contrast, BR-LSTM is burst-aware: it maintains calibration while avoiding excessive width during latency surges, which is preferable for SLA-centric alerting.
Classical ARIMA often attains coverage by broadening intervals. Figure 4 shows visibly wider bands, consistent with the larger MPIW required by classical baselines to meet coverage targets.
Dropout-based and weak Bayesian surrogates are overconfident. They produce narrow bands but under-cover across metrics (Figure 3), and Figure 4 shows that this under-coverage appears as missed excursions. Even when their apparent upper bounds are below the SLA in Figure 2, the intervals lack statistical reliability. BR-LSTM avoids this issue by combining calibrated coverage with operationally useful width.
Table 8 shows that BR-LSTM remains compact in memory (comparable to other lightweight neural baselines) while incurring moderate training time due to its two-stage procedure. GP and RF models are materially heavier in memory. Classical time-series models such as ARIMA and the local level state–space Kalman model exhibit minimal training time and moderate memory footprints, which makes them suitable as efficient statistical baselines. These costs are compatible with real-time dashboards and edge monitoring. Consequently, when the objective is to stay close to the latency SLA with calibrated guarantees, BR-LSTM is a sensible default, whereas for the tightest reliable jitter envelopes, GP and state–space components remain complementary.
In addition to training time and model size, Table 8 reports per-sample CPU inference latency (mean and p95). Under the stated thresholds, lightweight predictors fall into Low or Medium tiers and are thus compatible with tight edge budgets on CPU, whereas sampling- or kernel-based predictors generally occupy higher-latency tiers unless vectorized or parallelized. These measurements make the deployability implications explicit and clarify when accuracy gains warrant additional milliseconds on resource-constrained edge nodes.
A pragmatic selection rule is two-gated. The first gate enforces admissibility by calibration and accepts only models whose empirical coverage lies within a tight tolerance of the nominal target. The second gate prioritizes risk and, among the admissible models, selects the one that minimizes the upper risk bound relative to the metric-specific SLA.
Under this policy, BR-LSTM is typically preferred for latency and packet loss monitoring, whereas GP and state–space models are favored for jitter-specific refinement.
Beyond a global factor, dynamic or context-conditioned calibration can tighten intervals in stable regimes while preserving coverage during bursts. A time-varying scheme can recalibrate k on a rolling validation window to track gradual drift. A feature-conditioned scheme can assign  k ( x )  by context (e.g., mobility state or load bins) to approximate conditional coverage. Distribution-free conformal prediction provides another route by calibrating residuals on recent data, optionally with sliding windows or stratification by context. These options can reduce average interval width and align upper bounds with SLA thresholds more closely, subject to safeguards against small-sample instability and policies that prevent adaptation from using test outcomes. Empirical evaluation is left for future work. Second, exogenous drivers and regime-aware components, for example, state–space residuals or GP-style kernels layered on BR-LSTM, may reduce width at equal coverage, particularly for heavy-tailed loss bursts. These proposals are orthogonal to the reported results and are motivated directly by the coverage and sharpness trade-offs observed in Figure 2, Figure 3 and Figure 4.
Considering Figure 2, Figure 3 and Figure 4 and Table 8 jointly, BR-LSTM provides a balanced trade-off among calibrated uncertainty, burst-resilient risk control, and deployability. Classical ARIMA and Kalman state–space models are notable for computational efficiency and analytic uncertainty intervals; while they are generally less flexible for nonstationary or nonlinear dynamics than neural baselines, their low training cost and moderate memory usage make them practical for rapid prototyping and resource-constrained deployments.
After calibration, BR-LSTM achieves near-nominal coverage with the narrowest calibrated bands among uncertainty-aware baselines for latency and loss (Table A2), while GPR/state–space provides tighter and more reliable jitter envelopes (Table 4, Table A1 and Table A2). Extreme volatility or class imbalance in certain test traces can degrade interval coverage, especially for packet loss, which motivates future research on adaptive calibration and rare event detection.
Negative coefficients of determination arise on segments with pronounced nonstationarity or regime shifts where the squared error exceeds the variance around the test-set mean. In such cases, a constant baseline can outperform the fitted model locally. For decision making, point accuracy is therefore complemented with calibrated coverage (PICP), interval width (MPIW), false-alarm control (FAR), and SLA-oriented upper bounds  μ + k σ , which indicate whether alerts remain well-controlled under bursty conditions.
The Mininet dataset configures link-level QoS using tc netem without generating application-layer traffic, so feedback effects from real gaming workloads are not captured. This design isolates modeling effects under controlled, diverse conditions for fair method comparison, but external validity may be limited. Future work will incorporate traffic replay and semi-real testbeds and will evaluate the model on operational traces.

6. Conclusions

This study introduced a two-stage BR-LSTM framework for proactive QoS forecasting in dynamic esports networks. The model was evaluated against uncertainty-aware baselines (BNN, GPR, MC-Dropout LSTM, and QR) and classical time-series baselines (ARIMA and a local level state–space model). On point forecast accuracy (Table A1), BR-LSTM attains competitive errors across latency, jitter, and packet loss, for example, latency MAE 32.49 ms, jitter MAE 6.62 ms, and loss MAE 7.75%, while several baselines lead on specific metrics, such as state–space and GPR for latency and RF and state–space for jitter. These results position BR-LSTM as a strong neural alternative within the broader accuracy landscape.
With respect to risk-aware operation (Table A2), calibrated intervals of the form  μ ^ ± k σ ^  for BR-LSTM achieves near nominal coverage for latency and loss (PICP 93.43% and 93.94%, respectively) but sub-target coverage for jitter (82.83%). The corresponding upper risk bounds show that the latency envelope remains below the 100 ms SLA (84.36 ms) and the jitter envelope is comfortably below the 50 ms SLA (31.63 ms), whereas the packet loss bound remains above the 10% SLA (12.29%), reflecting the intrinsic difficulty of tail events in loss dynamics. These results highlight the persistent challenge of predicting packet loss under bursty and rare event conditions and motivate further work on tail risk modeling and adaptive calibration strategies. Taken together, the findings indicate a favorable coverage and sharpness trade-off for latency and a pragmatic envelope for loss, with jitter reliability being improvable through tighter local calibration.
From a deployment perspective, BR-LSTM maintains a compact footprint and a straightforward calibration workflow, which makes it suitable for real-time dashboards and edge scenarios. Evidence across Table 7 and Table A2 supports the following. For latency, BR-LSTM provides calibrated envelopes that remain under the SLA while preserving coverage. For jitter, GP and state–space models remain advantageous for the tightest reliable bands, with BR-LSTM competitive. For packet loss, all models encounter intermittent SLA interactions, which motivates loss-tail-aware enhancements. Classical ARIMA and Kalman models maintain extremely low training time and moderate memory, providing fast and interpretable benchmarks for practical monitoring systems, although with reduced flexibility for complex dynamics. The added per-sample latency in Table 8 indicates that models in the Low or Medium tiers meet sub-millisecond to few-millisecond CPU budgets on edge nodes without specialized accelerators, whereas higher-latency models remain appropriate when accuracy gains dominate and resources permit.
Calibrated upper bounds  μ + k σ  align naturally with SLA thresholds. Dashboards can surface early warnings when the upper bound approaches policy limits, while operators prioritize tickets or throttle nonessential traffic when risk persists across consecutive windows. For control loops, admissibility is first established by calibration; only models whose empirical PICP lies within a narrow tolerance of the nominal target are considered. Actions are then triggered when the upper bound exceeds metric-specific thresholds for latency, jitter, or loss. This workflow converts calibrated uncertainty into actionable signals for preemptive routing, pacing, or bitrate adaptation without relying on post hoc mitigation.
For network control, expose  μ + k σ  as a per-link risk score and incorporate it into a composite cost with latency, jitter, and packet loss. The controller selects routes that minimize this risk-weighted cost and triggers preemptive switching or pacing when the cost exceeds policy thresholds over consecutive windows.

Limitations and Future Work

While the proposed model demonstrates promising results, there are important limitations that remain, particularly in data realism and temporal calibration.
Data realism is a key challenge, as the current study relies on synthetic or controlled datasets. Real-world network congestion, unpredictable traffic, and heterogeneous access technologies are not fully captured. Incorporating live data and real-world traces will be a focus in improving robustness and generalization. Future work will evaluate rolling re-calibration, feature-conditioned (Mondrian) schemes, and distribution-free conformal prediction. The goal is to tighten intervals in stable regimes while preserving coverage during bursts, with safeguards against small-sample instability and leakage.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data available in a publicly accessible repository.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ARIMAAutoregressive Integrated Moving Average
BNNBayesian Neural Network
BRBayesian Regression
BR-LSTMBayesian Regression-augmented Long Short-Term Memory
FARFalse Alarm Rate
GNNGraph Neural Network
GPGaussian Process
GPRGaussian Process Regression
LSTMLong Short-Term Memory
MAEMean Absolute Error
MC-DropoutMonte Carlo Dropout
MLPMultilayer Perceptron
MPIWMean Prediction Interval Width
MSEMean Squared Error
PICPPrediction Interval Coverage Probability
QoEQuality of Experience
QoSQuality of Service
QRQuantile Regression
R2Coefficient of Determination (R-squared)
RBFRadial Basis Function
ReLURectified Linear Unit
RFRandom Forest
RMSERoot Mean Squared Error
RTPReal-time Transport Protocol
RTTRound-Trip Time
RUDPReliable User Datagram Protocol
SLAService Level Agreement
SVRSupport Vector Regression
TAQ-GRNNTopology-Aware QoS-GRNN
UDPUser Datagram Protocol
UDTUDP-based Data Transfer protocol
UQUncertainty Quantification

Appendix A

Table A1. Evaluation and risk analysis results.
Table A1. Evaluation and risk analysis results.
Jitter
MetricBNNBR-LSTMGPRMC-Dropout LSTMMLPQRRFSVRStandard LSTMARIMAState–Space
MAE4.616.623.336.803.928.123.054.186.713.853.24
RMSE5.937.814.248.175.198.894.015.388.094.914.24
R20.430.030.71−0.090.56−0.280.740.53−0.060.620.72
Latency
MetricBNNBR-LSTMGPRMC-Dropout LSTMMLPQRRFSVRStandard LSTMARIMAState–Space
MAE28.7032.4925.5833.7735.94100.6830.6328.4133.2131.6724.99
RMSE53.3246.5442.4658.6056.69111.3953.3254.8858.1748.1244.95
R20.160.010.17−0.010.06−2.640.160.120.010.080.20
Loss
MetricBNNBR-LSTMGPRMC-Dropout LSTMMLPQRRFSVRStandard LSTMARIMAState–Space
MAE8.037.757.777.689.4917.357.787.737.697.237.15
RMSE9.909.869.879.4312.1619.339.539.479.418.928.93
R2−0.110.000.00−0.01−0.68−3.24−0.03−0.020.00−0.01−0.01
Table A2. Risk assessment using  μ ^ ± 1.96 σ ^  and Q95 bounds. NA: not applicable (interval-based calibration metrics are not defined for the one-sided Q95 bound of QR).
Table A2. Risk assessment using  μ ^ ± 1.96 σ ^  and Q95 bounds. NA: not applicable (interval-based calibration metrics are not defined for the one-sided Q95 bound of QR).
MetricSLA ThresholdModelRisk BoundkMaxMeanFAR_Raw(%)FAR_Calibrated(%)PICP_Raw(%)PICP_Calibrated(%)MPIW_RawMPIW_Calibrated
Jitter50 msBNN   μ ^ ± 1.96 σ ^ 2.6423.7321.6004.557.071.241.67
BR-LSTM   μ ^ ± 1.96 σ ^ 3.1731.6331.25047.9855.5682.8323.1937.51
GPR   μ ^ ± 1.96 σ ^ 1.8133.4521.580096.4695.4516.8615.57
MC-Dropout LSTM   μ ^ ± 1.96 σ ^ 2.8422.8321.72002.535.560.771.11
QRQ95 34.5534.550NANANANANA
ARIMA   μ ^ ± 1.96 σ ^ 2.141.6520.890.510.5195.9697.4720.2421.68
State–Space   μ ^ ± 1.96 σ ^ 1.9134.9821.440095.4594.9516.716.27
Latency100 msBNN   μ ^ ± 1.96 σ ^ 2.8580.9169.91007.0711.627.4310.8
BR-LSTM   μ ^ ± 1.96 σ ^ 2.2484.368485.3585.3592.9393.43173.4198.17
GPR   μ ^ ± 1.96 σ ^ 2.41114.3171.9185.3585.3592.9393.43171.73211.16
MC-Dropout LSTM   μ ^ ± 1.96 σ ^ 2.8775.6870.53003.546.064.015.88
QRQ95 191.77191.7785.86NANANANANA
ARIMA   μ ^ ± 1.96 σ ^ 2.48209.7155.285.3585.3588.8993.43215.57272.77
State–Space   μ ^ ± 1.96 σ ^ 2.48118.8770.6485.3585.3592.9393.94168.69213.44
Loss Rate10%BNN   μ ^ ± 1.96 σ ^ 312.489.9235.3539.94.047.071.131.73
BR-LSTM   μ ^ ± 1.96 σ ^ 1.4512.2912.1354.5554.5597.4793.9433.6324.88
GPR   μ ^ ± 1.96 σ ^ 1.4510.6510.6554.5554.5597.4793.4336.0926.7
MC-Dropout LSTM   μ ^ ± 1.96 σ ^ 2.8911.379.8830.8138.891.013.030.630.93
QRQ95 27.0627.0653.54NANANANANA
ARIMA   μ ^ ± 1.96 σ ^ 1.8222.15.4154.5554.5594.9594.4445.3742.13
State–space   μ ^ ± 1.96 σ ^ 1.4810.229.9654.5554.5596.4692.9335.7727.01

References

  1. Tanenbaum, A.S.; Wetherall, D.J. Computer Networks, 5th ed.; Pearson Prentice Hall: Upper Saddle River, NJ, USA, 2011. [Google Scholar]
  2. Metzger, F.; Geißler, S.; Grigorjew, A.; Loh, F.; Moldovan, C.; Seufert, M.; Hoßfeld, T. An introduction to online video game QoS and QoE influencing factors. IEEE Commun. Surv. Tutor. 2022, 24, 1894–1925. [Google Scholar] [CrossRef]
  3. Seufert, M.; Egger, S.; Slanina, M.; Zinner, T.; Hoßfeld, T.; Tran-Gia, P. A survey on quality of experience of HTTP adaptive streaming. IEEE Commun. Surv. Tutor. 2015, 17, 469–492. [Google Scholar] [CrossRef]
  4. Quax, P.; Monsieurs, P.; Lamotte, W.; De Vleeschauwer, D.; Degrande, N. Objective and subjective evaluation of the influence of small amounts of delay and jitter on a recent first-person shooter game. In Proceedings of the 3rd ACM SIGCOMM Workshop on Network and System Support for Games, Portland, OR, USA, 30 August–3 September 2004; pp. 152–156. [Google Scholar] [CrossRef]
  5. Ries, M.; Svoboda, P.; Rupp, M. Empirical study of subjective quality for massive multiplayer games. In Proceedings of the 15th International Conference on Systems, Signals and Image Processing, Bratislava, Slovakia, 25–28 June 2008; pp. 181–184. [Google Scholar] [CrossRef]
  6. Arnab, A.A.; Shuvro, A.A.; Ma, K.; Leung, H. A deep learning approach for a QoS prediction system in cellular networks. In Proceedings of the 2023 IEEE 9th World Forum on Internet of Things (WF-IoT), Aveiro, Portugal, 12–27 October 2023; pp. 1–6. [Google Scholar] [CrossRef]
  7. Kääräinen, T.; Siekkinen, M.; Ylä-Jääski, A.; Zhang, W.; Hui, P. A measurement study on achieving imperceptible latency in mobile cloud gaming. In Proceedings of the 8th ACM on Multimedia Systems Conference, Taipei, Taiwan, 20–23 June 2017; pp. 88–99. [Google Scholar] [CrossRef]
  8. Sommers, J.; Barford, P. Cell vs WiFi: On the performance of metro area mobile connections. In Proceedings of the 2012 Internet Measurement Conference, Boston, MS, USA, 14–16 November 2012; pp. 301–314. [Google Scholar] [CrossRef]
  9. Sackl, A.; Schatz, R.; Hoßfeld, T.; Metzger, F.; Lister, D.; Irmer, R. QoE management made uneasy: The case of cloud gaming. In Proceedings of the 2016 IEEE International Conference on Communications Workshops (ICC), Kuala Lumpur, Malaysia, 23–27 May 2016; pp. 492–497. [Google Scholar] [CrossRef]
  10. Vasilev, V.; Leguay, J.; Paris, S.; Maggi, L.; Debbah, M. Predicting QoE factors with machine learning. In Proceedings of the 2018 IEEE International Conference on Communications (ICC), Kansas City, MO, USA, 20–24 May 2018; pp. 1–6. [Google Scholar] [CrossRef]
  11. Gu, Y.; Grossman, R.L. UDT: UDP-based data transfer for high-speed wide area networks. Comput. Netw. 2007, 51, 1777–1799. [Google Scholar] [CrossRef]
  12. Tran, D.T.; Choi, E. A reliable UDP for ubiquitous communication environments. In Proceedings of the 2007 Annual Conference on International Conference on Computer Engineering and Applications, Gold Coast Queensland, Australia, 17–19 January 2007; pp. 1–6. [Google Scholar]
  13. White, G.; Palade, A.; Clarke, S. Forecasting QoS attributes using LSTM networks. In Proceedings of the 2018 International Joint Conference on Neural Networks (IJCNN), Rio de Janeiro, Brazil, 8–13 July 2018; pp. 1–8. [Google Scholar] [CrossRef]
  14. Hameed, A.; Violos, J.; Leivadeas, A.; Santi, N.; Grünblatt, R.; Mitton, N. Toward QoS prediction based on temporal transformers for IoT applications. IEEE Trans. Netw. Serv. Manag. 2022, 19, 4010–4027. [Google Scholar] [CrossRef]
  15. Zou, G.; Lin, S.; Hu, S.; Duan, S.; Gan, Y.; Zhang, B.; Chen, Y. FHC-DQP: Federated hierarchical clustering for distributed QoS prediction. IEEE Trans. Serv. Comput. 2023, 16, 4073–4086. [Google Scholar] [CrossRef]
  16. Liu, M.; Xu, H.; Sheng, Q.Z.; Wang, Z. QoSGNN: Boosting QoS prediction performance with graph neural networks. IEEE Trans. Serv. Comput. 2023, 17, 645–658. [Google Scholar] [CrossRef]
  17. Wu, Z.; Ding, D.; Xiu, Y.; Zhao, Y.; Hong, J. Robust QoS prediction based on reputation integrated graph convolution network. IEEE Trans. Serv. Comput. 2023, 17, 1154–1167. [Google Scholar] [CrossRef]
  18. Williams, C.K.; Rasmussen, C.E. Gaussian Processes for Machine Learning; MIT Press: Cambridge, MA, USA, 2006. [Google Scholar]
  19. Gal, Y.; Ghahramani, Z. Dropout as a Bayesian approximation: Representing model uncertainty in deep learning. In Proceedings of the 33rd International Conference on International Conference on Machine Learning, New York, NY, USA, 19–24 June 2016; pp. 1050–1059. [Google Scholar]
  20. Kendall, A.; Gal, Y. What uncertainties do we need in Bayesian deep learning for computer vision? In Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 5580–5590. [Google Scholar]
  21. Eyobu, O.S.; Edwinah, E.K. A deep learning-based routing approach for wireless mesh backbone networks. IEEE Access 2023, 11, 49509–49518. [Google Scholar] [CrossRef]
  22. Moysen, J.; Giupponi, L.; Baldo, N.; Mangues-Bafalluy, J. Predicting QoS in LTE HetNets based on location-independent UE measurements. In Proceedings of the 2015 IEEE 20th International Workshop on Computer Aided Modelling and Design of Communication Links and Networks (CAMAD), Guildford, UK, 7–9 September 2015; pp. 124–128. [Google Scholar] [CrossRef]
  23. Wong, A.; Chiu, C.Y.; Hains, G.; Humphrey, J.; Fuhrmann, H.; Khmelevsky, Y.; Mazur, C. Gamers private network performance forecasting: From raw data to the data warehouse with machine learning and neural nets. arXiv 2021, arXiv:2107.00998. [Google Scholar] [CrossRef]
  24. Zhang, P.; Huang, W.; Chen, Y.; Zhou, M.; Al-Turki, Y. A novel deep-learning-based QoS prediction model for service recommendation utilizing multi-stage multi-scale feature fusion with individual evaluations. IEEE Trans. Autom. Sci. Eng. 2023, 21, 1740–1753. [Google Scholar] [CrossRef]
  25. Chen, Y.; Qi, Y.; Shen, J.; Yu, P.; Xiang, Z.; Zheng, Z. Dynamic QoS prediction for edge intelligence with multi-modal data fusion. In Proceedings of the 2024 IEEE Smart World Congress (SWC), Nadi, Fiji, 2–7 December 2024; pp. 1754–1761. [Google Scholar] [CrossRef]
  26. Drucker, H.; Burges, C.J.; Kaufman, L.; Smola, A.; Vapnik, V. Support vector regression machines. In Proceedings of the 10th International Conference on Neural Information Processing Systems, Denver, Colorado, 3–5 December 1996; pp. 155–161. [Google Scholar]
  27. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  28. Hornik, K.; Stinchcombe, M.; White, H. Multilayer feedforward networks are universal approximators. Neural Netw. 1989, 2, 359–366. [Google Scholar] [CrossRef]
  29. Zhang, G.; Patuwo, B.E.; Hu, M.Y. Forecasting with artificial neural networks: The state of the art. Int. J. Forecast. 1998, 14, 35–62. [Google Scholar] [CrossRef]
  30. Lakshminarayanan, B.; Pritzel, A.; Blundell, C. Simple and scalable predictive uncertainty estimation using deep ensembles. In Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 6405–6416. [Google Scholar]
  31. Wilson, A.G.; Hu, Z.; Salakhutdinov, R.; Xing, E.P. Deep kernel learning. In Proceedings of the 18th International Conference on Artificial Intelligence and Statistics, Cadiz, Spain, 9–11 May 2016; pp. 370–378. Available online: https://proceedings.mlr.press/v51/wilson16.html (accessed on 17 November 2025).
  32. Koenker, R. Quantile Regression; Cambridge University Press: Cambridge, UK, 2005. [Google Scholar] [CrossRef]
  33. Loevenich, J.F.; Lopes, R.R.F. DRL meets GNN to improve QoS in tactical MANETs. In Proceedings of the IEEE Network Operations and Management Symposium, Seoul, Republic of Korea, 6–10 May 2024; pp. 1–4. [Google Scholar] [CrossRef]
  34. Zhang, W.; Xu, L.; Yan, M.; Wang, Z.; Fu, C. A probability distribution and location-aware ResNet approach for QoS prediction. J. Web Eng. 2021, 20, 1189–1228. [Google Scholar] [CrossRef]
  35. Box, G.E.P.; Jenkins, G.M.; Reinsel, G.C.; Ljung, G.M. Time Series Analysis: Forecasting and Control, 5th ed.; John Wiley & Sons: Hoboken, NJ, USA, 2015. [Google Scholar]
  36. Durbin, J.; Koopman, S.J. Time Series Analysis by State Space Methods, 2nd ed.; Oxford University Press: Oxford, UK, 2012. [Google Scholar]
  37. Fritsch, T.; Ritter, H.; Schiller, J. The effect of latency and network limitations on MMORPGs: A field study of Everquest2. In Proceedings of the 4th ACM SIGCOMM Workshop on Network and System Support for Games, Hawthorne, NY, USA, 10–11 October 2005; pp. 1–9. [Google Scholar] [CrossRef]
  38. Yasui, T.; Ishibashi, Y.; Ikedo, T. Influences of network latency and packet loss on consistency in networked racing games. In Proceedings of the 4th ACM SIGCOMM Workshop on Network and System Support for Games, Hawthorne, NY, USA, 10–11 October 2005; pp. 1–8. [Google Scholar] [CrossRef]
  39. Lau, R.W.H.; Chan, A. Motion prediction for online gaming. In Proceedings of the 1st International Workshop on Motion in Games, Utrecht, The Netherlands, 14–17 June 2008; pp. 104–114. [Google Scholar] [CrossRef]
  40. Maskey, N.; Horsmanheimo, S.; Tuomimaki, L. Latency analysis of LTE network for M2M applications. In Proceedings of the 13th International Conference on Telecommunications (ConTEL), Graz, Austria, 13–15 July 2015; pp. 1–7. [Google Scholar] [CrossRef]
  41. Feng, Y.; Gao, M.; Zhang, Z. Web service QoS classification based on optimized convolutional neural network. In Proceedings of the IEEE 14th International Conference on Intelligent Systems and Knowledge Engineering (ISKE), Dalian, China, 14–16 November 2019; pp. 584–590. [Google Scholar] [CrossRef]
  42. Song, T.; Garza, P.; Meo, M.; Munafò, M.M. Modeling concurrent RTP flows for end-to-end predictions of QoS in real-time communications. In Proceedings of the 2024 International Symposium on Multimedia (ISM), Tokyo, Japan, 11–13 December 2024; pp. 63–70. [Google Scholar] [CrossRef]
  43. Cheng, Y.; Cao, W.; Fang, H.; Zang, S. A context-aware edge-cloud collaboration framework for QoS prediction. Tsinghua Sci. Technol. 2024, 30, 1201–1214. [Google Scholar] [CrossRef]
  44. Moreira, D.C.; Guerreiro, I.M.; Sun, W.; Cavalcante, C.C.; Sousa, D.A. QoS predictability in V2X communication with machine learning. In Proceedings of the IEEE 91st Vehicular Technology Conference (VTC2020-Spring), Antwerp, Belgium, 25–28 May 2020; pp. 1–5. [Google Scholar] [CrossRef]
  45. Wang, Y.; Jia, Z.; Zhang, X.; Shao, B.; Wang, H.; Xing, X. TAQ-GRNN: A topology-aware QoS prediction model based on gated recurrent neural networks. In Proceedings of the IEEE 13th Data Driven Control and Learning Systems Conference (DDCLS), Kaifeng, China, 17–19 May 2024; pp. 303–308. [Google Scholar] [CrossRef]
  46. Maroudis, A.C.; Theodoropoulos, T.; Violos, J.; Leivadeas, A.; Tserpes, K. Leveraging graph neural networks for SLA violation prediction in cloud computing. IEEE Trans. Netw. Serv. Manag. 2024, 21, 605–620. [Google Scholar] [CrossRef]
  47. Drainakis, G.; Pantazopoulos, P.; Katsaros, K.V.; Sourlas, V.; Amditis, A.; Kaklamani, D.I. Distributed predictive QoS in automotive environments under concept drift. In Proceedings of the IFIP Networking Conference, Thessaloniki, Greece, 3–6 June 2024; pp. 549–554. [Google Scholar] [CrossRef]
  48. Shen, L.; Pan, M.; Liu, L.; You, D.; Li, F.; Chen, Z. Contexts enhance accuracy: On modeling context aware deep factorization machine for web API QoS prediction. IEEE Access 2020, 8, 165551–165569. [Google Scholar] [CrossRef]
  49. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. In Proceedings of the 3rd International Conference on Learning Representations (ICLR 2015), San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
Figure 1. The pseudo code for model comparison.
Figure 1. The pseudo code for model comparison.
Information 16 01016 g001
Figure 2. Comparison of upper risk bounds with SLA thresholds for latency, jitter, and packet loss using  μ ^ + 1.96 σ ^  and Q95.
Figure 2. Comparison of upper risk bounds with SLA thresholds for latency, jitter, and packet loss using  μ ^ + 1.96 σ ^  and Q95.
Information 16 01016 g002
Figure 3. Calibration results of PICP and MPIW for six risk-aware models across latency, jitter, and packet loss.
Figure 3. Calibration results of PICP and MPIW for six risk-aware models across latency, jitter, and packet loss.
Information 16 01016 g003
Figure 4. Forecasted trajectories with calibrated 95% prediction intervals ( μ ^ ± k σ ^ ) and SLA thresholds for (a) jitter, (b) latency, and (c) loss. Each panel shows ground truth (thin black line), one model’s mean (grayscale line with markers), its 95% predictive band (light-gray region with diagonal hatching and a thin outline), and the SLA (thick black dash–dot line labeled “SLA”). The panels highlight that latency spikes are the primary drivers of SLA risk, while jitter remains mostly below its SLA and loss exhibits intermittent bursts.
Figure 4. Forecasted trajectories with calibrated 95% prediction intervals ( μ ^ ± k σ ^ ) and SLA thresholds for (a) jitter, (b) latency, and (c) loss. Each panel shows ground truth (thin black line), one model’s mean (grayscale line with markers), its 95% predictive band (light-gray region with diagonal hatching and a thin outline), and the SLA (thick black dash–dot line labeled “SLA”). The panels highlight that latency spikes are the primary drivers of SLA risk, while jitter remains mostly below its SLA and loss exhibits intermittent bursts.
Information 16 01016 g004aInformation 16 01016 g004b
Table 1. BR-LSTM parameters.
Table 1. BR-LSTM parameters.
SymbolDescription
LLook-back window length
  ν QoS metric index in {latency, jitter, loss}
  y t R 3 QoS vector at time t
  y t ( ν ) R Scalar QoS for metric ν at time t
  X t Past window
  x t R d BR input (e.g., time index)
  X R N × d BR design matrix
  y ( ν ) R N Targets for metric ν
  w ( ν ) R d Weights for metric ν
  λ   Prior   precision   of   w ( ν )
  σ n 2 Observation noise variance
  ε Observation noise
  μ w ( ν )   Posterior   mean   of   w ( ν )
  w ( ν )   Posterior   covariance   of   w ( ν )
  x * New input for prediction
  μ ^ * , ( ν ) BR predictive mean at x*
  ( σ ^ * , ( ν ) ) 2 BR predictive variance at x*
  μ ^ t * , ( ν ) Time-indexed predictive mean at xt
  σ ^ t * , ( ν ) Time-indexed predictive std. at xt
  z t R 6   [ μ t * , ( lat ) , σ t * , ( lat ) , μ t * , ( jit ) , σ t * , ( jit ) , μ t * , ( loss ) , σ t * , ( loss ) ]
  Z t Sequence input to LSTM
  LST M θ ( ) Sequence model
  μ ^ t + 1 R 3 Predicted mean QoS vector
  σ ^ t + 1 R 3 Predicted standard deviation vector
kInterval scaling factor
  λ c o v Weight for coverage term
HLSTM hidden size
NNumber of BR samples
Table 2. Implementation details of models.
Table 2. Implementation details of models.
CategoryImplementation Details
Standard LSTMOne stacked layer with hidden size 64; uses hyperbolic tangent and sigmoid activation gates; trained using the Adam optimizer [49] (learning rate = 1 × 10−3) for 30 epochs; outputs point forecasts only.
MC-Dropout LSTMSame architecture as Standard LSTM, with dropout rate 0.30 applied after the recurrent layer; 20 stochastic forward passes at inference to estimate predictive mean and standard deviation (the prediction interval uses  μ ^ ± 1.96 σ ^  (raw) or  μ ^ ± k σ ^  (calibrated, with k tuned on the validation set to target PICP = 95%), following Gal and Ghahramani [19]. For latency reporting, the single-pass measurement is used; per-sample wall time on CPU increases approximately linearly with the number of stochastic passes S.
BNNTwo fully connected layers (64 → 64 → 3) with dropout rate 0.20; trained for 100 epochs using Adam; 20 Monte Carlo samples used to derive predictive distributions.
QRQR with quantile level τ = 0.95 and regularization α = 0.1; solved using the HiGHS optimizer as implemented in SciPy 1.15.2 (Python 3.12.3) to directly provide Q95 upper bounds.
GPRUses Radial Basis Function (RBF) and white noise kernels; default initialization of length scale; posterior variance used to estimate predictive uncertainty.
SVRSVR with RBF kernel; parameters: C = 1.0 and epsilon = 0.1.
RFRF with 100 trees; no depth limitation; Mean Squared Error (MSE) criterion used for node splitting.
MLPOne hidden layer with 100 Rectified Linear Unit (ReLU) neurons; trained using the Adam optimizer with a maximum of 500 iterations.
ARIMA Seasonal-free ARIMA selected by auto_arima (AIC minimization); model orders re-estimated on the training split only; closed-form one-step forecast returns point prediction with 95% forecast interval from the analytic forecast error variance (no network epochs required).
Linear State–Space (Kalman)Local-level Gaussian state–space model fitted via Durbin–Koopman EM; system and measurement noise variances learned by maximum likelihood on training data; recursive Kalman filter yields  μ ^ t + 1  and  σ ^ t + 1 ; 95% interval reported as  μ ^ t + 1 ± 1.96 σ ^ t + 1  (optionally recalibrated to  μ ^ ± k σ ^ ).
Table 3. Availability of risk indicators.
Table 3. Availability of risk indicators.
CategoryAlgorithmsRisk Indicator
Risk-Aware ModelsBR-LSTM, BNN, MC-Dropout LSTM, GPR, Kalman (state–space), QR, ARIMA For   models   with   μ ^ ,   σ ^ :   μ ^ ± k σ ^  with k tuned to target about 95% PICP; for QR: qτ with τ = 0.95; for ARIMA or state–space: native 95% forecast interval, upper bound Unat
Deterministic ModelsSVR, RF, MLP, Standard LSTMPoint prediction only
Table 4. Ranking of models by accuracy and risk detection.
Table 4. Ranking of models by accuracy and risk detection.
Tier/CriterionModel(s)Representative ResultsRisk-Bound CharacteristicsRemarks
High Accuracy (non-probabilistic)RF, MLPJitter R2 = 0.74 (RF)
Jitter R2 = 0.56 (MLP)
Best point-forecast accuracy; no uncertainty estimates.
Balanced Accuracy + Calibrated UncertaintyBR-LSTMLatency MAE = 32.49 msCalibrated 95% prediction interval:
- jitter upper bound = 31.63 ms (<50 ms SLA);
- latency upper bound = 84.36 ms (<100 ms SLA);
- loss upper bound = 12.29% (>10% SLA).
Good trade-off between accuracy and reliable risk bounds.
Conservative/Wide Upper BoundQR (τ = 0.95) Q95 bounds:
- jitter 34.55 ms (<50 ms);
- latency 191.77 ms (>100 ms);
- loss 27.06% (>10%).
Percentile-based and wide; only jitter bound satisfies SLA. Suitable for worst-case provisioning; tends to over-reserve for latency and loss.
Conservative/Wide Upper BoundARIMA Calibrated bounds typically wide; latency upper bound > 100 ms; jitter upper bound often <50 ms; loss upper bound >10%.Conservative envelopes; coverage near nominal at the expense of width.
Well-calibrated and moderate widthState–space, GPR Calibrated coverage close to nominal; jitter upper bound <50 ms; latency upper bound is higher than 100 ms; loss upper bound is near 10%; moderate width.Balanced option with moderate width; provides tighter and more reliable jitter envelopes; useful as a complementary baseline.
Under-covered/Over-confidentMC-Dropout LSTM, BNNMean jitter PICP_cal = 5.56% (MC-Dropout LSTM); 7.07% (BNN)Prediction intervals too narrow; severe under-coverage.High point accuracy offset by unreliable risk estimates.
Moderate Accuracy (no uncertainty)SVRLatency R2 = 0.12; Jitter R2 = 0.53; Loss R2 = −0.02 Reasonable point forecasts but lacks uncertainty quantification.
Low AccuracyStandard LSTMMultiple metrics show negative R2 or high MAE/RMSENot applicable; deterministic model with no uncertainty estimates (no PICP, MPIW, FAR).Moderate point forecasts but lacks uncertainty; unsuitable for risk-aware operation; baseline only.
Table 5. Simulation components.
Table 5. Simulation components.
ComponentQuantityDescription
Player Hosts6Simulate individual esports players
Game Server1Central server for all players
Switch1Aggregates traffic between players and server
Table 6. Simulation results.
Table 6. Simulation results.
FeatureMeanStd. DevMinQ1MedianQ3Max
latency (ms)70.249.561.3835.146489.56299.46
jitter (ms)21.697.845.9615.0221.827.8647.25
loss rate (%)10.349.1201.849.4716.3748.57
Table 7. Model classification by nature.
Table 7. Model classification by nature.
CategoryAlgorithmsCharacteristics
Traditional MLSVR, RFNon-sequential; deterministic; no uncertainty estimates
Deterministic Neural NetworksMLPFeed forward; high point accuracy; no risk bound
Sequential DeterministicStandard LSTMCaptures temporal patterns; deterministic output
Bayesian SequentialBR-LSTM, MC-Dropout LSTMSequence modeling with explicit uncertainty
Probabilistic Feed-forwardBNN Fully   Bayesian   weights ;   μ ^ ± k σ ^  risk bounds
Kernel ProbabilisticGPRNon-parametric; predictive; mean μ ^ and σ ^
Quantile RegressionQR (qτ)Directly predicts upper quantile; τ = 0.95
Classical Statistical ARIMALinear autoregressive moving average with differencing; provides native 95% forecast interval
Sequential State–spaceLocal-level Kalman FilterLinear Gaussian state–space with Bayesian recursive update; provides native 95% forecast interval
Table 8. Computational cost comparison of models.
Table 8. Computational cost comparison of models.
ModelTraining Time (s)Model Size (MB)Latency (Mean) (ms)Latency (p95) (ms)Inference Cost
(Single-Sample, CPU)
BNN0.530.0149.2251.72High
BR-LSTM10.480.070.750.86Low
GPR0.2414.964.465.20Medium
MC-Dropout LSTM11.260.070.510.77Low
MLP13.540.332.463.02Medium
QR24.80.012.233.22Medium
RF29.7319.554.7178.24High
SVR0.560.574.636.21Medium
Standard LSTM23.110.070.541.16Low
ARIMA0.883.0315.7718.46High
Kalman0.721.636.987.98High
Notes: “Inference cost” labels follow the measured per-sample latency (Low ≤ 1 ms; 1–5 ms = Medium; >5 ms = High). MC-Dropout LSTM and BNN use S = 20 stochastic passes for predictive averaging in this study; Table 8 reports the single-pass latency, and executing S passes on CPU increases the per-sample wall time approximately S-fold.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yang, C.-F. Uncertainty-Aware QoS Forecasting with BR-LSTM for Esports Networks. Information 2025, 16, 1016. https://doi.org/10.3390/info16121016

AMA Style

Yang C-F. Uncertainty-Aware QoS Forecasting with BR-LSTM for Esports Networks. Information. 2025; 16(12):1016. https://doi.org/10.3390/info16121016

Chicago/Turabian Style

Yang, Ching-Fang. 2025. "Uncertainty-Aware QoS Forecasting with BR-LSTM for Esports Networks" Information 16, no. 12: 1016. https://doi.org/10.3390/info16121016

APA Style

Yang, C.-F. (2025). Uncertainty-Aware QoS Forecasting with BR-LSTM for Esports Networks. Information, 16(12), 1016. https://doi.org/10.3390/info16121016

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop