Next Article in Journal
Joint Effect of Signal Strength, Bitrate, and Topology on Video Playback Delays of 802.11ax Gigabit Wi-Fi
Previous Article in Journal
EEMD-TiDE-Based Passenger Flow Prediction for Urban Rail Transit
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Adaptive Bayesian System Identification for Long-Term Forecasting of Industrial Load and Renewables Generation

1
School of Internet of Things Engineering, Wuxi University, Wuxi 214105, China
2
Key Laboratory of Measurement and Control of CSE, Ministry of Education, Southeast University, Nanjing 210096, China
3
Jiangsu Provincial University Key Laboratory of Vehicle-Road Multimodal Perception and Control, Wuxi University, Wuxi 214105, China
*
Author to whom correspondence should be addressed.
Electronics 2026, 15(3), 530; https://doi.org/10.3390/electronics15030530
Submission received: 17 December 2025 / Revised: 20 January 2026 / Accepted: 20 January 2026 / Published: 26 January 2026
(This article belongs to the Section Systems & Control Engineering)

Abstract

The expansion of renewables in modern power systems and the coordinated development of upstream and downstream industrial chains are promoting a shift on the utility side from traditional settlement by energy toward operation driven by data and models. Industrial electricity consumption data exhibit pronounced multi-scale temporal structures and sectoral heterogeneity, which makes unified long-term load and generation forecasting while maintaining accuracy, interpretability, and scalability a challenge. From a modern system identification perspective, this paper proposes a System Identification in Adaptive Bayesian Framework (SIABF) for medium- and long-term industrial load forecasting based on daily freeze electricity time series. By combining daily aggregation of high-frequency data, frequency domain analysis, sparse identification, and long-term extrapolation, we first construct daily freeze series from 15 min measurements, and then we apply discrete Fourier transforms and a spectral complexity index to extract dominant periodic components and build an interpretable sinusoidal basis library. A sparse regression formulation with 1 regularization is employed to select a compact set of key basis functions, yielding concise representations of sector and enterprise load profiles and naturally supporting multivariate and joint multi-sector modeling. Building on this structure, we implement a state-space-implicit physics-informed Bayesian forecasting model and evaluate it on real data from three representative sectors, namely, steel, photovoltaics, and chemical, using one year of 15 min measurements. Under a one-month-ahead evaluation using one year of 15 min measurements, the proposed framework attains a Mean Absolute Percentage Error (MAPE) of 4.5% for a representative PV-related customer case and achieves low single-digit MAPE for high-inertia sectors, often outperforming classical statistical models, sparse learning baselines, and deep learning architectures. These results should be interpreted as indicative given the limited time span and sample size, and broader multi-year, population-level validation is warranted.

1. Introduction

The expansion of renewables in modern power systems and the ongoing industrial digital transformation are pushing power utilities to move from settling by energy to data- and model-driven operation for upstream and downstream industries [1]. Here, upstream industries refer to raw-material and basic-material sectors at the front of the value chain (e.g., mining, steel, chemicals, and cement), whereas downstream industries refer to mid- to end-manufacturing and application-oriented sectors, e.g., equipment manufacturing, building materials, consumer electronics, and renewables-related manufacturing. Industrial power demand data now form high-frequency, multi-year records across regions, sectors, and user levels and have become a key resource for describing the operating rhythm of industrial value chains and the structure of energy use [2]. Survey studies show that exploiting structural features and cross-scale temporal dependencies in user-side load data is essential for refined load forecasting and planning-oriented decision support [3,4]. However, the field still lacks a unified medium- and long-term modeling and forecasting framework for electricity demand across upstream and downstream industries that also balances forecasting accuracy, computational efficiency, and structural interpretability [5].
To further motivate the selection of representative sectors in this study, Table 1 summarizes the annual final energy consumption of major manufacturing sub-sectors in China (in 2019). It shows that the ferrous metals (steel) and chemical sub-sectors are among the largest energy consumers, and they are therefore adopted as representative high-intensity upstream sectors in our case studies. In contrast, lower-energy sub-sectors (e.g., textiles) are included to reflect downstream diversity and to stress-test the proposed framework under heterogeneous industrial rhythms.
In terms of load characteristics, electricity consumption in upstream and downstream industries simultaneously involves short-term intraday fluctuations at 15 min or even finer resolutions, as well as multi-scale periodic structures at weekly, monthly, quarterly, and seasonal horizons. At the same time, differences in production processes, operating conditions, and equipment configurations across industries lead to pronounced diversity and hierarchical structure in the resulting load series [6]. On the one hand, upstream sectors such as steel, chemicals, and cement typically exhibit high load levels, long operating cycles, and relatively stable shift patterns; their intraday fluctuations are relatively smooth, while weekly, monthly, and seasonal periodicity is more prominent [7]. On the other hand, downstream sectors such as equipment manufacturing, building materials, and consumer electronics are more strongly influenced by market demand, order fluctuations, and environmental factors, and they thus display more pronounced seasonality and uncertainty [8]. This multi-sector, multi-scale, multi-pattern nature imposes higher requirements on a unified modeling framework, which must be capable of capturing complex periodic structures while maintaining scalability and interpretability in multi-sector and multi-user scenarios.
Traditional time series methods, such as ARIMA and its seasonal extensions, have demonstrated certain advantages for nearly linear, weakly nonstationary load series [9,10]. By parameterizing trend and seasonal components, such models can achieve reasonably stable performance in short- and medium-term forecasting. However, they often struggle when confronted with strong nonlinearities, superposition of multiple periods, and structural breaks, and they have difficulty capturing complex patterns that span multiple time scales simultaneously. In recent years, deep learning methods based on LSTM, GRU, attention mechanisms, and Transformer architectures have achieved notable progress in short- and medium-term load forecasting [11]. These models can alleviate nonlinear and multi-scale issues by automatically extracting high-dimensional features from raw sequences through deep network structures. Nevertheless, they typically involve a large number of parameters, incur high training costs, and are sensitive to sample size and hyperparameter tuning. Moreover, their internal structures are difficult to interpret directly, along with the mapping from prediction results to physically and operationally meaningful concepts. This lack of transparency poses challenges in power system planning and learning algorithm analysis, where interpretability and communicability are particularly important [12,13,14].
To address these challenges, a promising line of research returns to the perspective of system identification and frequency-domain analysis, treating load time series as dynamic signals formed by the superposition of several dominant periodic components [15]. The imperative to transition from purely data-driven black-box models to interpretable, mechanistic representations of dynamic systems is driving a significant evolution in time series analysis [16]. This shift is particularly crucial in energy forecasting, where models must not merely predict but also offer profound mathematical insight into the underlying mechanisms and operational rhythms of complex industrial loads. Modern system identification (SI) is uniquely positioned to address this need by offering a principled approach to constructing parsimonious, real-world-compliant models that avoid the mere stacking of parameters.To address these challenges, a promising line of research returns to the perspective of system identification and frequency-domain analysis, treating load time series as dynamic signals formed by the superposition of several dominant periodic components. The foundation of modern SI was laid by seminal works that formalized the process of constructing mathematical models from observed data, particularly the maximum likelihood principle method [17] and prediction error method (PEM) [18,19]. These methods established rigorous statistical frameworks for estimating model parameters, moving beyond simple curve fitting. Contemporary system identification methods, such as Sparse Bayesian Learning (SBL) and Sparse Identification of Nonlinear Dynamics (SINDy), offer an alternative paradigm by focusing on discovering the underlying governing equations or sparse representations that describe the system’s evolution. Unlike purely data-driven black-box models, these techniques aim to distill complex dynamics into a compact, interpretable mathematical form, often involving ordinary or partial differential equations, or in the time series context, a minimal set of basis functions. The core concept of utilizing kernel methods and achieving sparsity through generalized priors, as seen in the Relevance Vector Machine (RVM) [20] and subsequent SBL extensions [21], provides a statistically robust way to select only the most relevant features. SBL, for instance, can be used for sparse feature selection in a regression context, effectively identifying the most relevant past observations or periodic components (e.g., Fourier basis) that contribute to the current load. Meanwhile, the development of SINDy [22,23] has revitalized the application of identification to complex, nonlinear time series, including those relevant to industrial processes and energy systems. These data-driven approaches leverage the power of 1 or sparsity-inducing norms to find the most parsimonious representation of the system dynamics. The success of these modern techniques relies heavily on the maturity of earlier identification works, such as the comprehensive methodology of Ljung’s system identification framework [24] and the fundamental principles of optimal control and estimation [25]. Similarly, the spirit of SINDy, which discovers parsimonious models by simultaneously identifying the relevant terms from a library of candidate functions and the coefficients for those terms, can be adapted to time series decomposition. By projecting the load data onto a carefully constructed dictionary (which may include trigonometric functions, wavelets, or other domain-specific patterns), these identification techniques can automatically select the minimal set of multi-scale periodic components and their corresponding amplitudes and phases. This approach inherently balances model complexity and interpretability: the resulting model is a sum of physically meaningful, identifiable periodic oscillations, offering a clear link between the forecast and the observed operational rhythms (e.g., daily shifts, weekly cycles, seasonal variations) of the upstream and downstream industries. This system-identification-based decomposition provides a structural understanding that is difficult to obtain from opaque deep learning models while explicitly handling the multi-scale periodicity that challenges traditional ARIMA models [26]. Furthermore, the application of subspace identification methods [27] has provided computationally efficient ways to model high-dimensional, multi-output systems, which is highly relevant for multi-sector load forecasting. The ongoing convergence of robust identification methods, sparse optimization, and control theory [28] promises to unlock truly interpretable and scalable solutions for new-type power systems, indicating that system identification currently holds immense potential for establishing sophisticated and highly reliable predictive models in time series forecasting across numerous critical domains.
Building on these insights, this paper adopts a system identification perspective and proposes a new load modeling approach for upstream and downstream industries that emphasizes both interpretability and scalability, centered on daily freeze electricity time series at the user and sector levels from SOTA learning methods [29,30,31]. Specifically, we propose the SIABF and follow a technical route of daily aggregation of high-frequency data, frequency-domain analysis, sparse identification, and long-term extrapolation. First, 15 min enterprise-side load measurements are aggregated to the daily scale to construct daily freeze series that reflect the electricity intensity of enterprises or sectors, which reduces modeling dimensionality while preserving long-term trends and periodic information and aligns with existing practice in day-level medium- and long-term load forecasting [32]. Second, discrete Fourier transforms are applied to the daily freeze series to extract dominant frequency components, from which a library of sinusoidal basis functions with clear frequency-domain interpretation is constructed, and a spectral complexity index is introduced to quantify the periodic complexity of different series. Third, a sparse regression framework with 1 regularization is employed to automatically select a small set of key basis functions and their coefficients from the candidate library, resulting in a compact and interpretable representation of the load series. Structurally, the SIABF model is explicitly represented as a linear combination of dominant periodic basis functions, which facilitates interpretation of industrial electricity consumption behaviors from the perspectives of daily, weekly, and seasonal cycles; from the identification standpoint, sparsity constraints enable automatic basis selection, thereby enhancing generalization and robustness in line with existing sparse system identification and compressive sensing methods [33,34]. In terms of computational complexity, both frequency-domain analysis and sparse regression scale well with data size, and the identified pairs of adaptive basis functions and coefficients can be used to extrapolate daily freeze series of upstream and downstream industries to medium- and long-term horizons, forming a stable day-level long-term forecasting mechanism suitable for large-scale deployment across multiple sectors and numerous users.
The remainder of this paper is organized as follows. Section 2 describes the structure of upstream–downstream industrial electricity data, the preprocessing workflow, and the forecasting task. Section 3 develops the adaptive basis sparse system identification framework, and Section 4 presents model-input design and extensions across upstream and downstream industries. Section 5 reports the experimental setup and results on real industrial datasets. Section 6 discusses empirical findings, limitations, and implications for deployment. Finally, Section 7 concludes this paper and outlines directions for large-scale validation and application-oriented extensions.

2. Electricity Data of Upstream and Downstream Industries and Problem Formulation

In the new-type power systems and coordinated development along industrial value chains, the large-scale user-side electricity data collected by power utilities not only reflect the internal production rhythms and equipment operation patterns of individual enterprises but also embody the supply–demand coupling among upstream and downstream industries. To construct a long-term forecasting model with good interpretability on this basis, it is necessary to systematically sort out the origin, structure, and quality of the electricity data for upstream and downstream industries and to clarify the modeling objectives and constraints within a unified mathematical framework. This section focuses on these issues, elaborating on the organization of upstream and downstream industry electricity data, the preprocessing pipeline, the construction of daily freeze values, and the formalization of the forecasting problem.

2.1. Industrial Chain Structure and User Hierarchy

Consider an industrial chain consisting of multiple sectors,
S = { steel , chemicals , cement , photovoltaics , renewables , } ,
where some sectors are located in the upstream part of the chain, such as steel, chemicals, and basic raw materials, while others are in the midstream or downstream part, such as equipment manufacturing, building materials, and end-product manufacturing. For each sector s S , the utility typically maintains electricity data for a number of users (enterprises); the corresponding user set is denoted by I s , and the global user set is then
I = s S I s .
At a finer granularity, each enterprise can be further decomposed into production line, workshop, or critical equipment. For example, an upstream steel enterprise may be divided into processes such as ironmaking, steelmaking, and rolling, whereas a downstream building materials enterprise may be divided into production lines, packaging lines, and warehousing and logistics subsystems. Correspondingly, electricity load data can be modeled at the whole-enterprise level or at the workshop/equipment level, forming a multi-level data structure spanning “sector–enterprise–equipment”.
From a modeling perspective, such a hierarchical structure has two important implications. First, load patterns differ significantly across sectors and enterprises. Upstream sectors are often characterized by high load levels, long cycles, and pronounced shift structures, while downstream sectors are more prone to demand volatility and seasonality. Second, even within the same sector, different enterprises may exhibit heterogeneity in baseline load levels, periodic structures, and process flows. This calls for a unified modeling framework with enterprise-specific parameterization. Therefore, subsequent modeling in this paper will mainly be described at the “sector–enterprise” two-level structure, and the coupling between upstream and downstream sectors will be understood as mutual influence and transmission among different sectors.

2.2. Measurement Data and Preprocessing Pipeline

For any user i I , the utility typically samples active power at fixed intervals (e.g., every 15 min), yielding 96 observations
p i , d , 1 , p i , d , 2 , , p i , d , 96 ,
for day d, where p i , d , k denotes the average active power during the k-th 15 min interval. These records, together with time stamps Time d , k , date labels Date d , user identifiers ID i , and other metadata, form the basis for aggregation at the daily, weekly, or sector level. To better characterize production and economic rhythms, auxiliary features can be derived from calendar (weekday/weekend, public holidays, month/quarter, and heating or high-temperature seasons) and environment (daily maximum, minimum and mean temperature, precipitation, humidity), and, where available, production-related information such as major maintenance windows, process changes, and order peaks. The sparse identification with adaptive basis functions proposed in this work focuses on extracting periodic structures from the load series themselves, but the framework is conceptually compatible with such exogenous variables and can be extended to include them as model inputs at the implementation stage.
In practice, meter data suffer from missing values, outliers, duplicated records and time misalignment, so a systematic preprocessing pipeline is required. Duplicates at the same user and time stamp are filtered using data-quality flags and acquisition priorities. Sporadic missing points are imputed by local time-neighborhood K-nearest-neighbor (KNN) averaging, while longer gaps are filled using typical-day load curves constructed from neighboring dates (e.g., separate weekday/weekend profiles); enterprises with pervasive data loss or long-term shutdowns can be removed or flagged according to business rules. Outliers—such as short spikes, prolonged abnormal loads, or patterns violating production logic—are detected using sliding-window statistics (e.g., z-scores based on mean and standard deviation) and robust measures (interquartile range or median absolute deviation), possibly in combination with comparisons to historical typical curves of the same day type. Erroneous points are then replaced by interpolated values or typical profiles, whereas genuine production events are retained and annotated. Finally, all multi-source series are aligned to a common time axis to correct hour shifts, time-zone differences, and resolution inconsistencies; for users with hourly data, 15 min series can be reconstructed by uniform distribution or refinement via typical curves. After missing-value handling, outlier correction, and time alignment, a high-quality, temporally consistent 15 min load series p ˜ i , d , k is obtained, which underpins the construction of daily freeze values and subsequent time series modeling.

2.3. Formalization of Forecasting Tasks and Modeling Objectives

Although 15 min data can finely depict an enterprise’s production status and equipment on/off behavior, long-term analysis of upstream and downstream industry electricity demand is more concerned with evolution at daily, weekly, monthly, and even longer time scales. It is therefore useful to aggregate high-frequency data to the daily scale and construct so-called “daily freeze values”.
For user i on day d, the daily freeze electricity consumption is defined as
y i , d = k = 1 96 p ˜ i , d , k ,
where p ˜ i , d , k is the preprocessed 15 min average active power. Equation (2) compresses the intraday fine-grained fluctuations into a single scalar that preserves the total consumption, thereby providing a concise and intuitive representation for medium- and long-term trend and dominant-period identification.
On this basis, multi-scale representations can be constructed as needed. At the weekly scale, the daily freeze values of a user in the same week can be aggregated to form a weekly freeze value y i , w ( week ) , defined as the sum of all daily freeze values within that week, thus characterizing the weekly load level. Similarly, monthly freeze values can be defined or sliding-window smoothing can be applied to the daily series to highlight seasonality and long-term trends while attenuating short-term random disturbances. In addition, the daily freeze series can be decomposed into a long-term baseline component and a short-term fluctuation component, thereby distinguishing structural changes from transient variations and supporting forecasting and scenario analysis at different time scales.
At the sector level, for sector s S , one can select M s representative users from I s and construct a sector-level load vector
y d ( s ) = y 1 , d ( s ) , , y M s , d ( s ) R M s , d = 1 , , D ,
thereby obtaining a multivariate time series { y d ( s ) } d = 1 D . In this way, the evolution of sector-level load and the heterogeneity among representative enterprises can be jointly captured.
Given the above data construction, the load forecasting problem for upstream and downstream industries can be formalized as a standard time series forecasting task. For a single user i, the observed daily freeze series over the study period is
y i = y i , 1 , y i , 2 , , y i , D .
During model training, the first T daily freeze values are taken as historical observations to build the forecasting model. The forecasting task is then to estimate the electricity consumption for the next H days given this historical window, that is,
y ^ i , T + 1 , y ^ i , T + 2 , , y ^ i , T + H ,
which yields the predicted trajectory of daily load for user i over the forecast horizon.
At the sector level, for the sector load vector series { y d ( s ) } d = 1 D , the forecasting task is to estimate the future H days of sector-level load vectors given a historical window of length T,
y ^ T + 1 ( s ) , y ^ T + 2 ( s ) , , y ^ T + H ( s ) ,
where each y ^ d ( s ) is an M s -dimensional vector representing the predicted electricity consumption of the representative enterprises in sector s.
Taking into account the upstream–downstream structure of the industrial chain, joint multi-sector forecasting tasks can also be defined. Let S up and S down denote the sets of upstream and downstream sectors, respectively. A joint time series
y d ( s ) : s S up S down , d = 1 , , D
can then be employed to build models that simultaneously capture the dynamics of multiple sectors and identify their shared periodic structures. On this basis, one can analyze how common dominant frequencies and long-term trends are reflected across different sectors and how these patterns can be exploited to improve forecasting performance in a joint modeling setting. Such modeling provides a quantitative basis for medium- and long-term planning and capacity allocation decisions of the power system.
Building upon this modeling foundation, the next section will develop a sparse system identification and forecasting framework with adaptive basis functions to extract frequency-domain structures and perform long-term extrapolation of industrial electricity load time series.

3. Sparse System Identification Framework with Adaptive Basis Functions

Based on the data construction and forecasting task definition in the previous section, this section develops a sparse system identification framework with adaptive basis functions for long-term electricity load modeling of upstream and downstream industries. The central idea is to first extract dominant periodic components from the daily freeze series through frequency-domain analysis, then construct a compact yet expressive basis function library, and finally perform sparse coefficient identification so that only a small number of basis components are retained. In this way, the model captures the essential periodic structure of industrial loads while remaining interpretable and computationally efficient.

3.1. Frequency-Domain Analysis and Adaptive Basis Construction

Consider a univariate daily freeze load series of length D for a given user or sector,
y 1 , y 2 , , y D .
For convenience, this series can be regarded as a realization of a discrete-time signal defined on the index set { 1 , , D } . To reveal its periodic structure, the series is first transformed into the frequency domain by the discrete Fourier transform (DFT). The DFT at frequency index k is given by
Y ( ω k ) = d = 1 D y d exp i 2 π k ( d 1 ) D , k = 0 , , D 1 ,
where ω k = 2 π k / D is the angular frequency and i is the imaginary unit. The amplitude spectrum is defined as
A k = | Y ( ω k ) | , k = 1 , , D 1 ,
and characterizes the contribution of each frequency component to the overall series.
In practice, only a finite number of frequencies carry substantial energy. To quantify the concentration of energy in the leading frequencies, a simple spectral complexity index I 10 is constructed. Let { A ( j ) } denote the amplitudes sorted in descending order, that is, A ( 1 ) A ( 2 ) A ( M ) , where M D is the number of distinct frequencies considered. The index I 10 is defined as
I 10 = 1 A ( 1 ) j = 1 10 max A ( j ) A ( j + 1 ) , 0 .
When I 10 is close to one, the spectrum is dominated by a few leading frequencies and the periodic structure is relatively simple. When I 10 is significantly smaller, energy is more evenly spread across frequencies and the periodic structure is more complex. This index provides a guideline for selecting the number of basis frequencies used in the subsequent modeling.
Based on the amplitude spectrum, a set of dominant frequencies is selected. Suppose that K frequency indices are chosen, denoted by k 1 , , k K , corresponding to angular frequencies ω k 1 , , ω k K . The normalized frequency of the j-th component can be defined as
f j = k j 1 D , j = 1 , , K .
For each normalized frequency f j , two real-valued basis functions are constructed,
ϕ 2 j 1 ( d ) = sin ( 2 π f j d ) , ϕ 2 j ( d ) = cos ( 2 π f j d ) , d = 1 , , D 1 .
This leads to a basis library containing 2 K sinusoidal functions, each corresponding to a dominant oscillatory mode of the series. These basis functions are interpretable because each pair captures a periodic pattern characterized by a specific period and phase shift.
Collecting all basis functions into a matrix yields
Φ = ϕ 1 ( 1 ) ϕ 2 ( 1 ) ϕ 2 K ( 1 ) ϕ 1 ( 2 ) ϕ 2 ( 2 ) ϕ 2 K ( 2 ) ϕ 1 ( D ) ϕ 2 ( D ) ϕ 2 K ( D ) R D × 2 K ,
which serves as the design matrix for the subsequent sparse identification step. The number of basis functions 2 K can be determined adaptively by starting from a moderate upper bound and adjusting it according to the spectral complexity index I 10 and cross-validation performance.

3.2. Sparse Coefficient Identification and Long-Term Forecasting Mechanism

Given the basis matrix Φ in (9) and the daily freeze vector
y = ( y 1 , , y D ) ,
the load series can be approximated as a linear combination of basis functions,
y d = 1 2 K θ ϕ ( d ) , d = 1 , , D ,
where θ is the coefficient associated with the -th basis function. In matrix form, this model can be written as
y = Φ θ + ε ,
where θ = ( θ 1 , , θ 2 K ) is the coefficient vector and ε is the residual term.
To avoid overfitting and enhance interpretability, the model encourages sparsity in θ by introducing an 1 -type regularization. The coefficient identification problem is formulated as
min θ R 2 K 1 2 y Φ θ 2 2 + λ θ 1 ,
where · 2 is the Euclidean norm, · 1 is the 1 norm, and  λ > 0 is a regularization parameter controlling the degree of sparsity. This optimization problem is a standard LASSO formulation. In practice, λ is selected by time series cross-validation on a validation segment (e.g., a blocked split or a rolling-origin split), and we choose λ * , which achieves low validation error while retaining a sparse set of dominant periodic bases (see Algorithm 1). We adopt the 1 penalty in this work due to its convexity and computational efficiency, which are important for large-scale deployment; other sparse penalties, such as elastic net to handle correlated bases and structured sparsity to promote shared components across enterprises, can be incorporated in the same framework, while exploring nonconvex penalties for reduced shrinkage bias is left for future work.
Algorithm 1 System Identification in Adaptive Bayesian Framework (SIABF)
Require: Daily freeze series { y d } d = 1 D for a user or sector; maximum number of dominant frequencies K max ; forecast horizon H; candidate regularization set Λ .
Ensure: Sparse coefficient vector (or matrix) θ ^ (or B ^ ( s ) ), adaptive basis functions { ϕ ( · ) } , and forecasts { y ^ d } d = D + 1 D + H (or { y ^ d ( s ) } d = D + 1 D + H ).
  1:
Daily series preparation: Construct the daily freeze series y 1 , , y D from preprocessed 15 min load data. Optionally split the series into a training segment and a validation segment.
  2:
Frequency-domain analysis: Compute the discrete Fourier transform Y ( ω k ) of the training segment using (6), obtain the amplitude spectrum A k = | Y ( ω k ) | , and sort { A k } in descending order. Compute the spectral complexity index I 10 via (7) and determine a candidate range for the number of dominant frequencies K K max .
  3:
Adaptive basis construction: Select K dominant frequencies with the largest amplitudes and compute their normalized frequencies f 1 , , f K . For each f j , construct the sinusoidal basis functions according to (8) and assemble the design matrix Φ R D × 2 K as in (9).
  4:
for each candidate λ Λ  do
  5:
   Sparse coefficient identification (univariate case): Solve the LASSO problem
min θ 1 2 y Φ θ 2 2 + λ θ 1
   to obtain a sparse coefficient vector θ ^ ( λ ) .
  6:
   Sparse coefficient identification (multivariate sector case, optional): For a sector
   matrix Y ( s ) R D × M s , solve
min B ( s ) 1 2 Y ( s ) Φ B ( s ) F 2 + λ B ( s ) 1
   to obtain the sparse coefficient matrix B ^ ( s ) ( λ ) .
  7:
   Validation (if a validation segment is available): Use the identified coefficients to
   generate forecasts on the validation segment, compute prediction errors (for example,
   mean absolute percentage error), and record the performance associated with ( K , λ ) .
  8:
end for
  9:
Hyperparameter selection: Select ( K * , λ * ) between forecasting accuracy and sparsity based on validation performance. Fix the corresponding basis matrix Φ * and coefficients θ ^ = θ ^ ( λ * ) .
10:
Long-term forecasting: For each future day d = D + 1 , , D + H , evaluate the basis vector
ϕ ( d ) = ϕ 1 ( d ) , , ϕ 2 K ( d )
and compute the forecast
y ^ d = ϕ ( d ) θ ^ .
11:
Long-term forecasting (sector case, optional): For each future day d = D + 1 , , D + H , compute
y ^ d ( s ) = ϕ ( d ) B ^ ( s ) R M s ,
yielding forecasts for all representative enterprises in sector s.
12:
Model evaluation and interpretation: Evaluate forecasting performance over the horizon H using appropriate error metrics.
This step also admits a Bayesian interpretation. Assume Gaussian observation noise ε N ( 0 , σ 2 I ) in (11) and independent Laplace priors on coefficients, p ( θ ) exp ( α θ 1 ) , where α is linked to λ and σ 2 . Then solving (12) yields the maximum a posteriori estimate of θ . After sparsity identifies an active set A = { : θ ^ 0 } , uncertainty can be quantified by refitting a Bayesian linear regression on the reduced design matrix Φ A . With a Gaussian prior θ A N ( 0 , τ 2 I ) , the posterior is closed-form with Σ A = ( Φ A Φ A / σ 2 + I / τ 2 ) 1 and μ A = Σ A Φ A y / σ 2 . For any future day d > D , the predictive distribution is Gaussian with mean y ^ d = ϕ A ( d ) μ A and variance s d 2 = ϕ A ( d ) Σ A ϕ A ( d ) + σ 2 , which directly yields prediction intervals.
From the viewpoint of system identification, the nonzero elements of θ reveal which periodic components are essential for reconstructing the observed load series. Their magnitudes determine the contribution of each component to the overall load profile, and the corresponding frequencies provide interpretable information about daily, weekly, or seasonal cycles.
The optimization problem in (12) can be solved using coordinate descent, proximal gradient methods, or other convex optimization algorithms. Since the dimension 2 K is usually much smaller than the number of samples D, the computational cost is moderate, which is advantageous for large-scale applications involving many users or sectors.
Once the sparse coefficient vector θ ^ is obtained by solving (12), the model can be used to produce long-term forecasts. For any future day index d > D , the basis functions in (8) can be evaluated at d by simple substitution. The forecast at day d is then given by
y ^ d = = 1 2 K θ ^ ϕ ( d ) = ϕ ( d ) θ ^ ,
where
ϕ ( d ) = ( ϕ 1 ( d ) , , ϕ 2 K ( d ) )
is the vector of basis evaluations at d. Because  ϕ ( d ) are sinusoidal functions with fixed frequencies, the forecast naturally extrapolates the periodic patterns identified from historical data into the future.
This forecasting mechanism has several desirable properties, where the forecast at any horizon depends only on the identified periodic structure and not on the accumulation of short-term prediction errors, which helps to maintain stability over long horizons. The periodic basis functions make the medium- and long-term behavior of the model transparent: the presence of a weekly component corresponds to a specific frequency, a seasonal component corresponds to another, and so on. Furthermore, the sparse coefficient vector highlights only a few dominant periodicities, making it easier to interpret forecasts in industrial applications such as planning and capacity allocation.

3.3. Bayesian State-Space Prediction with Implicit Physical Information

The sparse identification step provides an interpretable periodic representation. However, industrial load patterns may exhibit gradual drifts due to operational adjustments and seasonal transitions. To capture such implicit physical inertia and to quantify uncertainty, we formulate a Bayesian state-space model on the retained adaptive basis.
Let A = { : θ ^ 0 } denote the active set obtained from the sparse regression in (12). Define the reduced basis vector ϕ A ( d ) R | A | by keeping only the active components, and let θ d R | A | be the corresponding coefficient vector at day d.
We adopt the following linear Gaussian model:
y d = ϕ A ( d ) θ d + v d , v d N ( 0 , R ) ,
θ d = θ d 1 + w d , w d N ( 0 , Q ) ,
with a Gaussian prior θ 0 N ( μ 0 , P 0 ) . Here, Q controls the adaptivity of the coefficients (implicit physical inertia), and R represents measurement noise and unmodeled disturbances. In implementation, μ 0 can be initialized by the sparse estimate restricted to A , and  P 0 = κ I with κ > 0 .
Given posterior ( μ d 1 , P d 1 ) at day d 1 , the one-step prediction is
μ d = μ d 1 ,
P d = P d 1 + Q .
Let ϕ d ϕ A ( d ) . Define the innovation and its variance:
e d = y d ϕ d μ d ,
S d = ϕ d P d ϕ d + R .
The Kalman gain and posterior update are
K d = P d ϕ d S d ,
μ d = μ d + K d e d ,
P d = ( I K d ϕ d ) P d .
This yields adaptive parameter updating together with posterior uncertainty.
For any future day d > D , the predictive distribution is Gaussian:
p ( y d D D ) = N ( y ^ d , s d 2 ) ,
where
y ^ d = ϕ A ( d ) μ D ,
s d 2 = ϕ A ( d ) P D ϕ A ( d ) + R .
Therefore, a  ( 1 α ) prediction interval is given by
y ^ d ± z 1 α / 2 s d ,
where z 1 α / 2 is the standard normal quantile. We set Q = q I and tune ( q , R ) on the validation segment together with ( K , λ ) to balance adaptivity and long-horizon stability. Since | A | 2 K in typical cases, the Bayesian update is computationally efficient and scalable for multi-user and multi-sector deployment.

3.4. Multivariate and Multi-Sector Extension

The above development focuses on a single univariate series, such as the daily freeze load of one enterprise. For sector-level modeling, where multiple representative enterprises are considered simultaneously, the framework can be extended to a multivariate setting.
Suppose sector s has M s representative enterprises with daily freeze loads y m , d ( s ) , where m = 1 , , M s and d = 1 , , D . For each day d, define the sector-level load vector
y d ( s ) = ( y 1 , d ( s ) , , y M s , d ( s ) ) .
Stacking these vectors over time yields the matrix
Y ( s ) = y 1 , 1 ( s ) y M s , 1 ( s ) y 1 , D ( s ) y M s , D ( s ) R D × M s .
If the same basis matrix Φ in (9) is used for all enterprises within the sector, the multivariate model can be written as
Y ( s ) Φ B ( s ) ,
where B ( s ) R 2 K × M s is a coefficient matrix whose columns correspond to different enterprises.
To promote sparsity across the entire sector and to reveal shared periodic structure, a group-wise or element-wise regularization can be introduced. A simple choice is to penalize the element-wise 1 norm,
min B ( s ) 1 2 Y ( s ) Φ B ( s ) F 2 + λ B ( s ) 1 ,
where · F is the Frobenius norm and B ( s ) 1 denotes the sum of absolute values of all entries in B ( s ) . This formulation allows different enterprises to share a common set of basis functions, while their coefficients may differ in magnitude and sign. More structured regularization terms, such as row-wise group norms, can further encourage enterprises to share a subset of dominant periodic components.
Forecasting at the sector level is obtained in an analogous way. For a future day d > D , the basis vector ϕ ( d ) is evaluated and the sector-level load forecast is computed as
y ^ d ( s ) = ϕ ( d ) B ^ ( s ) R M s ,
where B ^ ( s ) is the estimated coefficient matrix. This provides forecasts for all representative enterprises in sector s and can be aggregated further if sector-level totals or averages are of interest.

3.5. Algorithmic Implementation

The SIABF procedure consists of daily aggregation, spectral analysis, adaptive basis construction, and sparse identification. To realize the adaptive Bayesian component, we further embed the retained basis expansion into a Bayesian inference step on the active coefficients. This step produces posterior uncertainty and predictive intervals in addition to point forecasts. The resulting workflow remains interpretable and scalable for multi-user and multi-sector deployment.
The complete sparse system identification procedure with adaptive basis functions can be summarized as follows. First, daily freeze series are obtained from preprocessed high-frequency data for the user or sector under consideration. Second, a DFT is computed and the amplitude spectrum is analyzed to identify dominant frequencies and to compute the spectral complexity index I 10 . Third, a set of K dominant frequencies is selected and the corresponding sinusoidal basis functions are constructed, yielding the basis matrix Φ . Fourth, the LASSO optimization problem in (12) or its multivariate counterpart in (28) is solved to obtain sparse coefficient estimates. Finally, the identified model is used to generate medium- and long-term forecasts by evaluating the basis functions at future time indices as in (13) and (29).
In practice, several implementation details deserve attention. The regularization parameter λ is selected by time series cross-validation by splitting the historical series into training and validation segments and choosing the value that balances validation forecasting accuracy and sparsity. The number of dominant frequencies K may be determined by combining the information from I 10 , visual inspection of the spectrum, and validation performance. When the daily series is short or exhibits nonstationarities, it may be helpful to restrict the maximum K or to update the basis functions in a rolling manner. These details do not change the core framework but are important for robust performance in real-world applications.
The SIABF algorithm takes as input a daily freeze load series
y 1 , y 2 , , y D
for a given user or sector, a maximum number of dominant frequencies K max , and a regularization parameter λ (or a search range for λ ). The primary outputs are the identified sparse coefficient vector (or matrix in the multivariate case), the set of dominant basis functions, and the corresponding long-term load forecasts over a prescribed horizon H. The detailed steps of SIABF are summarized in Algorithm 1.
The SIABF algorithm thus provides a complete and implementable procedure for sparse system identification and long-term load forecasting based on adaptive periodic basis functions. By explicitly separating the steps of spectral analysis, basis construction, and sparse identification, the algorithm offers both computational efficiency and structural interpretability and can be applied consistently at the enterprise level and at the sector level for upstream and downstream industrial electricity load modeling.

4. Model Input Design and Extensions to Upstream and Downstream Industries

The sparse identification framework with adaptive basis functions and the SIABF algorithm provide a generic mapping from historical daily freeze series to long-term load forecasts. To make this framework effective in real industrial settings, this section details how historical data are split into training and evaluation segments, how input windows and optional auxiliary features are organized, and how the SIABF model is extended from single-enterprise series to unified multi-sector upstream–downstream settings.

4.1. Data Partitioning and Temporal Organization

Let { y d } d = 1 D denote the daily freeze series of a given enterprise or the sector-level series defined in (3). In practice, the series is partitioned into three consecutive segments used for model training, validation, and testing. One common choice is to specify two indices
1 < T tr < T val < D ,
and to define the training segment as { y d } d = 1 T tr , the validation segment as { y d } d = T tr + 1 T val , and the test segment as { y d } d = T val + 1 D . The training segment is used to perform frequency-domain analysis, to construct the adaptive basis functions, and to identify sparse coefficients; the validation segment is used to select hyperparameters such as the number of dominant frequencies K and the regularization parameter λ ; and the test segment is reserved for final performance assessment.
Depending on the intended forecasting horizon H, the validation and test segments can be organized as either single-block horizons or rolling horizons. In a single-block setting, the model is trained once on the training segment and used to predict the entire validation or test segment in one shot, exploiting the closed-form extrapolation Formula (13). In a rolling setting, one may re-estimate the SIABF model in a sliding manner by gradually extending the training window and repeatedly generating forecasts; this allows the stability of the identified periodic structure to be inspected over time. The choice between these options is determined by operational requirements and the degree of nonstationarity in the industrial load series.

4.2. Input Window Length and Historical Information Usage

Although the SIABF can in principle use the entire training segment for frequency-domain analysis and coefficient identification, in some applications the effective historical horizon is restricted. For example, if process changes or structural shifts have occurred in the distant past, very early data may no longer be representative of current operation patterns. In such cases, a finite historical window of length L can be selected, and the training segment can be restricted to { y d } d = T tr L + 1 T tr for spectral analysis and sparse identification.
Formally, for a chosen L with 1 L T tr , define a truncated series
y ˜ 1 = y T tr L + 1 , , y ˜ L = y T tr ,
and apply the DFT (6) and the subsequent SIABF steps to { y ˜ d } d = 1 L instead of { y d } d = 1 T tr . In this way, the identified periodic structure reflects more recent load behavior. The window length L can itself be tuned based on validation performance, trading off between robustness (longer windows) and adaptability to recent changes (shorter windows).
In sector-level modeling where Y ( s ) in (27) is used, an analogous truncation can be applied along the temporal dimension, yielding a matrix of size L × M s as input to the multivariate SIABF procedure. This unified windowing scheme allows upstream and downstream sectors to be treated consistently while still respecting potential differences in the effective length of historical information.

4.3. Incorporation of Auxiliary Features

The core SIABF is built on the daily freeze series and the periodic basis functions derived from its spectrum. However, in practical applications, auxiliary features such as calendar variables (weekday, weekend, holidays), environmental variables (temperature, precipitation), and aggregated production indicators may also be available and informative. These features can be incorporated into the model in a simple and transparent way by augmenting the basis matrix Φ .
Suppose that for each day d an auxiliary feature vector z d R q is available, where q is the number of auxiliary variables. The sequence { z d } can be collected into a feature matrix
Z = z 1 z D R D × q .
An augmented design matrix can then be constructed as
Φ aug = [ Φ Z ] R D × ( 2 K + q ) .
The sparse regression problem (12) is replaced by
min θ aug 1 2 y Φ aug θ aug 2 2 + λ θ aug 1 ,
where θ aug includes both periodic basis coefficients and coefficients for auxiliary variables. The  1 penalty continues to promote sparsity so that only auxiliary variables that contribute meaningfully to the reconstruction and forecasting of the load series are retained.
For sector-level modeling, the same augmentation idea applies by replacing Φ with Φ aug in (28). This design enables upstream and downstream sectors to share a common periodic basis while capturing sector-specific sensitivities to calendar, environmental, or production-related features through additional sparse coefficients.

4.4. Joint Modeling of Upstream and Downstream Sectors

To extend the SIABF from single sectors to a joint modeling of multiple upstream and downstream sectors, the sector-level matrices introduced earlier can be organized into a block structure. Suppose that a set of sectors S joint S is selected for joint analysis, including both upstream and downstream sectors of interest. For each s S joint , assume that M s representative enterprises have been chosen and that the daily freeze series have been arranged into a matrix Y ( s ) R D × M s .
A straightforward joint representation can be obtained by concatenating the sector matrices along the column dimension,
Y ( joint ) = [ Y ( s 1 ) Y ( s 2 ) Y ( s r ) ] R D × M joint ,
where s 1 , , s r enumerate the sectors in S joint , and 
M joint = j = 1 r M s j
is the total number of representative enterprises across all sectors considered.
Using the same basis matrix Φ (or Φ aug if auxiliary features are included), the joint multivariate model can be written as
Y ( joint ) Φ B ( joint ) ,
where B ( joint ) R 2 K × M joint is the coefficient matrix to be identified. The sparse optimization problem is
min B ( joint ) 1 2 Y ( joint ) Φ B ( joint ) F 2 + λ B ( joint ) 1 ,
or the corresponding variant with Φ aug if auxiliary features are incorporated.
This joint modeling form allows upstream and downstream sectors to share a common periodic basis at the daily scale, reflecting global cycles such as work weeks and seasons while maintaining distinct coefficient profiles for individual enterprises and sectors. By inspecting the resulting coefficient matrix, one can identify periodic components that are common to many sectors as well as components that are specific to certain sectors or enterprises. In addition, forecasts for all sectors can be produced in a unified manner by applying the extrapolation Formula (29) to every column of B ( joint ) .
When the SIABF is applied to multiple sectors simultaneously, a consistent evaluation protocol is needed to assess performance at different aggregation levels. At the enterprise level, standard error metrics such as Mean Absolute Error or Mean Absolute Percentage Error can be computed for each time series individually. At the sector level, these errors can be aggregated across enterprises within the same sector to form sector-level performance indicators, such as average or median MAPE. For joint modeling, it is often useful to report both sector-specific errors and global errors across all enterprises in S joint . Let E ( s ) denote a sector-level error metric for sector s, and let E ( joint ) denote a global error metric computed on Y ( joint ) .

5. Experiments

5.1. Data Description and Preprocessing

To establish a power consumption forecasting model for electricity utility customers, power consumption information was collected from industrial customers of a regional power utility in China. The dataset comprises 15 min interval power consumption data for multiple customers within three distinct electricity-consuming sectors, namely, steel, photovoltaics (PV), and chemical. Customer IDs are anonymized for confidentiality. In particular, the steel-sector dataset contains multiple steel enterprises, and unless otherwise stated, the quantitative comparisons and the illustrative time series figures in Section 5.3 report results for a representative steel enterprise rather than an average across multiple steel companies. Data for each sector are stored in separate Excel files, structured as illustrated in Table 2.
Each row in the dataset records the power consumption of a customer from a specific industry across the 96 observation points for that day. Consumption records for most industry customers span a period of one year, typically from 1 July 2023 to 30 June 2024.
Prior to model construction, the missing values within the data must be addressed. Power consumption time series data for industrial users inevitably contain missing values due to sensor failures, transmission interruptions, human error, or external disturbances. If not handled properly, these missing values can compromise the stability and accuracy of the model. Simply deleting records containing missing values would not only result in the loss of valuable historical data but also damage the temporal continuity of the series, potentially leading to biased model predictions. Therefore, appropriate imputation of missing values is critical for enhancing both the predictive accuracy and the robustness of the model. Considering the characteristics of industrial data, this paper employs a hybrid imputation method. First, a KNN mean imputation approach, using eight neighboring points, is applied to fill the missing values. If a data point remains NaN after KNN imputation, a forward-filling method is used for secondary imputation. If the element is the first in the series and still NaN after the previous steps, the global mean is finally applied for imputation. The daily freezing value, which represents the total daily power consumption, is calculated by summing the 96 observation points from 00:00 to 24:00. In this study, the pandas 3.0.0 library in Python 3.10 is used to perform column-wise summation. The result is added as a new column named “Daily Freezing Value” to the dataset, thereby recording the total daily power consumption for each industrial customer.

5.2. Model Input and Evaluation Metric

To prevent data leakage, the dataset is partitioned into training and testing sets. The training set is used to construct and train the prediction model, while the testing set is used to evaluate the model’s performance. This division ensures that the model’s performance on unseen data can be effectively validated, guaranteeing the model’s generalization ability and predictive accuracy.
The daily freezing value is used as the one-dimensional feature for our time series, establishing a univariate time series forecasting model. The daily freezing value data from customers with a continuous one-year record (1 July 2023 to 30 June 2024) is selected for the task. The daily freezing value data from the first 11 months are used as the training set input, and the data from the last month are used as the testing set for final model performance evaluation.
The Mean Absolute Percentage Error (MAPE) is adopted as the primary evaluation metric for this experiment. The formula for MAPE is defined as
MAPE = 1 n t = 1 n A t F t A t × 100 % ,
where A t is the actual value in the testing period (all days in June 2024), F t is the value forecasted for days in June 2024 by the method trained with data of 1 July 2023 to 31 May 2024, and n is the number of observations. MAPE is particularly suitable for power consumption forecasting in industrial and utility scenarios compared to scale-dependent metrics such as Mean Absolute Error (MAE). The primary industrial value of MAPE lies in its scale independence and interpretability. Power consumption across different industrial sectors (e.g., cement vs. photovoltaics) and individual customers can vary by orders of magnitude. MAE reports error in the original units (e.g., kWh), making it impossible to directly compare the forecasting performance between a large-scale steel mill and a smaller renewables facility. Conversely, MAPE expresses the error as a percentage, providing a normalized measure of accuracy that allows for equitable comparison of model performance across heterogeneous customer groups, which is crucial for benchmarking in the utility sector. Furthermore, electric utilities and system operators typically manage risk and assess performance based on relative error. An absolute error of 1000 kWh might be negligible for a plant consuming 100 , 000 kWh per day ( 1 % error) but highly significant for a customer consuming only 5000 kWh per day ( 20 % error). MAPE’s direct translation into a percentage provides immediate industrial context and helps decision-makers, such as grid planners and energy traders, understand the relative impact of forecasting errors, thus aligning better with industrial Key Performance Indicators (KPIs).
In this work, the forecasting target is the daily freezing value, defined as the total daily electricity consumption obtained by summing the 96 measurements within each day. For each customer with a continuous one-year record (1 July 2023 to 30 June 2024), we adopt a one-month-ahead evaluation: the first eleven months (July 2023 to May 2024) are used for training, and the last month (June 2024) is used for testing. Note that the PV industry in our experiments refers to industrial electricity consumption of PV-related enterprises, e.g., manufacturing and supporting operations, rather than PV plant energy generation.

5.3. Forecasting Results and Comparative Analysis

This section presents the power consumption forecasting results obtained by the proposed SIABF and compares its performance against several established benchmark models. To comprehensively evaluate the robustness and predictive accuracy of the SIABF framework, the following models were selected as comparative baselines. These models represent the current state of the art and theoretically sound baseline methods widely adopted in the industrial power forecasting domain. The comparison set includes statistical models like ARIMA (Autoregressive Integrated Moving Average), which is a fundamental benchmark for linear time series analysis. It also incorporates machine learning and system identification models such as SBL (Sparse Bayesian Learning) and SINDy (Sparse Identification of Nonlinear Dynamics), which are valued for their potential in providing sparse and interpretable representations of the system dynamics. Finally, advanced deep learning models, specifically the LSTM (Long Short-Term Memory) network, the GRU (Gated Recurrent Unit) network, and the ResNet (Residual Network), are included to benchmark SIABF’s capability against methods known for capturing complex, long-term nonlinear temporal dependencies characteristic of modern energy consumption time series.
The SIABF model’s performance is systematically compared with the benchmark models across all three industrial sectors using the MAPE metric on the designated test set (the last month of data). The comparative results are summarized in Table 3, demonstrating the superior ability of the SIABF model to capture complex, potentially physics-informed, temporal dependencies in the energy consumption data.

5.3.1. Steel Industry

To provide deeper insight into the models’ behavior, we select representative forecasting results from specific customers that highlight the characteristic performance differences. These include high-load industrial customers (e.g., in the steel sector) to evaluate robustness under high variance and inertia, and volatile customers (e.g., in the PV or renewables sector) to assess the ability to handle irregular and rapidly changing consumption patterns. Visualizations depicting the curves of the daily freezing values are presented, comparing the forecasted daily freezing values against the actual values for the selected cases, in order to qualitatively demonstrate the superior accuracy and stability of the SIABF predictions compared to the leading benchmark models.
The initial comparison of SIABF with a sparse Bayesian baseline (SBL), shown in Figure 1, highlights the enhanced stability and closer fit of the SIABF predictions. While SBL exhibits pronounced deviations, particularly around consumption peaks, SIABF maintains a more reliable trajectory. Moreover, the ARIMA and SBL models fail to accurately learn the overall trend and struggle particularly with data volatility, exhibiting noticeable phase lag and amplitude errors.
As illustrated in Figure 2, the comparison between the system identification method SINDy and the representative deep learning model LSTM shows the challenges faced by both approaches. SINDy demonstrates high variance and substantial lag, failing to capture the time series structure effectively. In contrast, while LSTM shows better responsiveness, it still struggles with accurately predicting the sharp rises and falls in the industrial load profile.
The power consumption forecasting results for a representative steel industry customer are comprehensively illustrated in Figure 3. SIABF effectively tracks the genuine consumption curve (Ground Truth), even successfully predicting the sharp decline and rebound observed around June 22nd as observed. In contrast, traditional statistical and sparse learning methods struggle significantly. This result validates that by incorporating implicit structural information, the SIABF is highly effective at minimizing prediction errors and yields the most accurate and reliable forecast for the complex operational requirements of the steel industry.

5.3.2. Photovoltaic Power Generation Industry

The forecasting performance for a representative customer in the PV-related industry exhibits distinct characteristics due to high volatility and sensitivity to external factors, e.g., temperature and operational variability. We emphasize that this subsection studies electricity consumption (daily freezing value) of a PV-related enterprise rather than PV plant energy generation. The following figures report one-month-ahead forecasts for June 2024 using the previous eleven months as training.
Analyzing the first comparison in Figure 4, both the proposed SIABF model and the sparse baseline SBL struggle with the sharp, rapid fluctuations typical of the PV load profile. However, SIABF demonstrates a significant advantage, as its physically informed structure allows it to maintain a closer trajectory to the actual consumption, while SBL shows large, delayed deviations, failing to adapt quickly to sudden changes in consumption. ARIMA fails entirely, demonstrating a smooth, almost static line that misses all temporal dependencies, resulting in a large MAPE of 22.947%. SINDy’s performance is even worse, with MAPE 33.908%, yielding erratic and severely misaligned predictions.
The performance contrast between SINDy and the deep learning model LSTM is clearly depicted in Figure 5. SINDy, struggling with the high-frequency noise inherent in PV data, yields highly inaccurate and erratic predictions. LSTM, by leveraging its recurrent structure, manages to capture the overall trend much better but still suffers from oversmoothing the extreme peaks and valleys, a common limitation when dealing with nonstationary renewable energy data.
The forecasting performance in the PV industry is crucial due to the sector’s high sensitivity to external factors, leading to volatile load profiles. Figure 6 confirms the superior capability of the SIABF model in this challenging environment. Visually, the SIABF model consistently achieves an excellent fit to the true consumption curve, i.e., Ground Truth, successfully capturing both the overall trend and the sharp daily fluctuations characteristic of PV operations. In stark contrast, traditional and sparse models perform exceptionally poorly.
The PV-related customer exhibits the lowest MAPE among the tested sectors. This outcome is consistent with the mechanism of SIABF, which constructs an adaptive sinusoidal basis from dominant frequency components and then performs sparse identification. When the daily freezing series contains a small number of stable periodicities, such as weekly production schedules and smooth seasonal variations, its spectrum is more concentrated and can be represented by a compact set of basis functions. In such cases, the sparse periodic model tends to extrapolate reliably over a one-month-ahead horizon, which can lead to a low MAPE. Importantly, the PV case in this paper corresponds to industrial electricity consumption of a PV-related enterprise rather than PV power generation. For PV generation forecasting, the dominant drivers are meteorological conditions and physical irradiance mechanisms, and high daily accuracy typically requires exogenous weather forecasts and physics-informed features. Extending SIABF by integrating meteorological covariates and validating over multi-year and broader populations is therefore a necessary direction before drawing conclusions for PV generation forecasting.

5.3.3. Chemical Industry

The power consumption profile of the chemical industry is typically characterized by high inertia and strong seasonality, reflecting continuous process operations. The following figures present the comparative forecasting results for a representative chemical customer.
As illustrated in Figure 7, we observe the performance of the SIABF model against the SBL sparse learning baseline. The chemical industry’s consumption curve is generally smoother than the steel or PV sectors, yet SBL still exhibits noticeable deviations from the actual data, especially during transitions. In contrast, the SIABF model leverages its underlying structure to provide predictions that hug the actual values much more closely, demonstrating superior tracking capability. In addition, the numerical results from the table show that the ARIMA model yields the lowest MAPE at 8.420%, narrowly outperforming SIABF. Specifically, the two key sharp consumption drops (mutations) shown in the figure are entirely missed by the ARIMA prediction. This suggests that while ARIMA provides a slightly better average fit due to linear stability, it is highly inadequate for predicting critical, nonlinear operational events. Conversely, both SBL and SINDy models show a substantial divergence from the real data, confirming their limited applicability to the nonlinear dynamics of industrial processes.
Figure 8 compares SINDy with the representative deep learning model, LSTM. The SINDy framework faces substantial difficulties in modeling the complex, unobserved dependencies of the chemical process, resulting in highly amplified errors and poor temporal alignment. LSTM, benefiting from its architecture designed for sequential data, produces a much more reasonable forecast, effectively capturing the overall low-frequency trends in the consumption pattern.
The power consumption data for the chemical industry represents a relatively stable, high-inertia load with strong underlying periodicities. Analysis of Figure 9 reveals a nuanced performance landscape for this sector. Obviously, the SIABF model provides an excellent fit to the Ground-Truth curve, accurately tracing the consumption trajectory throughout the testing period.

6. Discussion

The empirical results on representative enterprises from the steel, PV-related, and chemical sectors indicate that the proposed sparse system identification framework, instantiated as the SIABF model, achieves competitive forecasting performance for industrial electricity consumption, i.e., daily freezing values, compared with classical statistical models and deep learning baselines. For the PV-related industrial customer, the load series exhibits higher volatility and sensitivity to external factors, and SIABF is able to exploit recurring temporal structures, e.g., weekly schedules and seasonal components, under our one-month-ahead evaluation. We emphasize that these PV case studies focus on the electricity consumption of PV-related enterprises rather than photovoltaic power generation. Therefore, this manuscript does not claim that SIABF alone can provide accurate PV generation forecasting, which typically relies on meteorological forecasts and physics-informed modeling; extending SIABF with exogenous weather drivers is left for future work.
At the same time, several limitations of the present study suggest directions for further research. The empirical evaluation is based on one year of data and a limited set of customers, and it mainly relies on daily freeze values without fully exploiting rich exogenous information such as detailed weather variables, production schedules, or market indicators. The current implementation also implicitly assumes local stationarity of the load series within the chosen time window; when structural breaks or process reconfigurations occur, the frequency content may shift, calling for online or rolling implementations with adaptive basis updates and time-varying regularization. Moreover, although the Bayesian formulation improves robustness, uncertainty quantification is not yet explicitly translated into probabilistic prediction intervals, which would be valuable for risk-aware decision-making. Future work will therefore focus on extending the framework to incorporate structured exogenous covariates, develop adaptive and probabilistic variants suitable for long-term deployment, and implement large-scale joint modeling of correlated sectors in real utility environments.

7. Conclusions

This paper has developed a sparse system identification framework with adaptive periodic basis functions for long-term load forecasting in upstream and downstream industrial chains. Centered on daily freeze electricity time series, the proposed SIABF approach combines frequency-domain analysis, a spectral complexity index for adaptive basis selection, and 1 -regularized sparse regression to construct interpretable models that explicitly decompose industrial loads into a small number of dominant daily, weekly, and seasonal components. The framework extends naturally from single-enterprise series to multivariate sector-level models and joint multi-sector formulations, thereby providing a unified and scalable route to modeling heterogeneous industrial users within new-type power systems.

Author Contributions

Conceptualization, L.S. and Z.W.; methodology, L.S. and L.Z.; software, L.S. and X.W.; validation, L.S. and Z.W.; investigation, X.W.; resources, L.Z.; supervision, L.Z. and X.W.; writing—original draft preparation, L.S. and Z.W.; writing—review and editing, X.W., L.Z. and Z.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported in part by the National Natural Science Foundation of China under Grant 42305158, and by the Startup Foundation for Introducing Talent of Wuxi University under Grant 2024r001.

Data Availability Statement

The data that support the findings of this study are available from the corresponding author upon reasonable request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Zhao, Z.; Li, Q.; Bo, B.; Yang, P.; Li, X.; Wu, Z.; Wang, G.; Ren, H. A Domain-Adversarial Mechanism and Invariant Spatiotemporal Feature Extraction Based Distributed PV Forecasting Method for EV Cluster Baseline Load Estimation. Electronics 2025, 14, 4709. [Google Scholar] [CrossRef]
  2. Sun, G.; Ma, W.; Wei, S.; Cai, D.; Wang, W.; Xu, C.; Zhang, K.; Wang, Y. A Fault Location Method for Medium Voltage Distribution Network Based on Ground Fault Transfer Device. Electronics 2023, 12, 4790. [Google Scholar] [CrossRef]
  3. Hammad, M.A.; Jereb, B.; Rosi, B.; Dragan, D. Methods and models for electric load forecasting: A comprehensive review. Logist. Supply Chain Sustain. Glob. Chall. 2020, 11, 51–76. [Google Scholar] [CrossRef]
  4. Liang, X.; Chen, L.; Zhou, S.; Xia, M.; Weng, L.; Liu, J. Interpretable and Accurate Multi-energy Loads Forecasting via Continuous-Time Modeling. IEEE Trans. Smart Grid 2025, 17, 255–268. [Google Scholar] [CrossRef]
  5. Dong, F.; Li, W. Research on the coupling coordination degree of “upstream-midstream-downstream” of China’s wind power industry chain. J. Clean. Prod. 2021, 283, 124633. [Google Scholar] [CrossRef]
  6. May, G.; Barletta, I.; Stahl, B.; Taisch, M. Energy management in production: A novel method to develop key performance indicators for improving energy efficiency. Appl. Energy 2015, 149, 46–61. [Google Scholar] [CrossRef]
  7. Rzeszowska, A.; Błażej, R. Effect of Sensor Head Orientation on the Accuracy of Magnetic Defect Detection in Steel-Cord Conveyor Belts. Sensors 2025, 25, 7364. [Google Scholar] [CrossRef] [PubMed]
  8. Belhadi, A.; Kamble, S.; Subramanian, N.; Singh, R.K.; Venkatesh, M. Digital capabilities to manage agri-food supply chain uncertainties and build supply chain resilience during compounding geopolitical disruptions. Int. J. Oper. Prod. Manag. 2024, 44, 1914–1950. [Google Scholar] [CrossRef]
  9. Shumway, R.H.; Stoffer, D.S. ARIMA models. In Time Series Analysis and Its Applications: With R Examples; Springer: Cham, Switzerland, 2017; pp. 75–163. [Google Scholar]
  10. Contreras, J.; Espinola, R.; Nogales, F.J.; Conejo, A.J. ARIMA models to predict next-day electricity prices. IEEE Trans. Power Syst. 2003, 18, 1014–1020. [Google Scholar] [CrossRef]
  11. Nilakanta Singh, K.; Robindro Singh, K. A review on deep learning models for short-term load forecasting. In Applications of Artificial Intelligence and Machine Learning: Select Proceedings of ICAAAIML 2020; Springer: Singapore, 2021; pp. 705–721. [Google Scholar]
  12. Henao, F.; Edgell, R.; Sharma, A.; Olney, J. AI in power systems: A systematic review of key matters of concern. Energy Inform. 2025, 8, 76. [Google Scholar] [CrossRef]
  13. Wei, M.; Yu, W.; Chen, D. AccDFL: Accelerated Decentralized Federated Learning for Healthcare IoT Networks. IEEE Internet Things J. 2025, 12, 5329–5345. [Google Scholar] [CrossRef]
  14. Jiang, L.; Ming, X.; Zhang, X. DT-DOFL: Digital-Twin-Empowered Decentralized Online Federated Learning for User-Centered Smart Healthcare Service Systems. IEEE Trans. Comput. Soc. Syst. 2025, 12, 4441–4455. [Google Scholar] [CrossRef]
  15. Tu, B.; Yang, X.; He, W.; Li, J.; Plaza, A. Hyperspectral Anomaly Detection Using Reconstruction Fusion of Quaternion Frequency Domain Analysis. IEEE Trans. Neural Netw. Learn. Syst. 2024, 35, 8358–8372. [Google Scholar] [CrossRef]
  16. Hassija, V.; Chamola, V.; Mahapatra, A.; Singal, A.; Goel, D.; Huang, K.; Scardapane, S.; Spinelli, I.; Mahmud, M.; Hussain, A. Interpreting black-box models: A review on explainable artificial intelligence. Cogn. Comput. 2024, 16, 45–74. [Google Scholar] [CrossRef]
  17. Aoki, M. Optimization of Stochastic Systems: Topics in Discrete-Time Systems; Elsevier: Amsterdam, The Netherlands, 2016. [Google Scholar]
  18. Lee, D.; Yannakakis, M. Testing finite-state machines: State identification and verification. IEEE Trans. Comput. 2002, 43, 306–320. [Google Scholar] [CrossRef]
  19. Wang, S.; Takyi-Aninakwa, P.; Jin, S.; Yu, C.; Fernandez, C.; Stroe, D.I. An improved feedforward-long short-term memory modeling method for the whole-life-cycle state of charge prediction of lithium-ion batteries considering current-voltage-temperature variation. Energy 2022, 254, 124224. [Google Scholar] [CrossRef]
  20. Tipping, M. The relevance vector machine. Adv. Neural Inf. Process. Syst. 1999, 12, 652–658. [Google Scholar]
  21. Bishop, C.M.; Nasrabadi, N.M. Pattern Recognition and Machine Learning; Springer: New York, NY, USA, 2006; Volume 4. [Google Scholar]
  22. Brunton, S. Discovering governing equations from data by sparse identification of nonlinear dynamics. In Proceedings of the APS March Meeting Abstracts, New Orleans, LA, USA, 13–17 March 2017; Volume 2017, p. X49-004. [Google Scholar]
  23. Kutz, J.N.; Brunton, S.L.; Brunton, B.W.; Proctor, J.L. Dynamic Mode Decomposition: Data-Driven Modeling of Complex Systems; SIAM: Philadelphia, PA, USA, 2016. [Google Scholar]
  24. Simpkins, A. System identification: Theory for the user, (ljung, l.; 1999) [on the shelf]. IEEE Robot. Autom. Mag. 2012, 19, 95–96. [Google Scholar] [CrossRef]
  25. Stengel, R.F. Optimal Control and Estimation; Courier Corporation: North Chelmsford, MA, USA, 1994. [Google Scholar]
  26. Box, G.E.; Jenkins, G.M.; Reinsel, G.C.; Ljung, G.M. Time Series Analysis: Forecasting and Control; John Wiley & Sons: Hoboken, NJ, USA, 2015. [Google Scholar]
  27. Van Overschee, P.; De Moor, B. Subspace Identification for Linear Systems: Theory—Implementation—Applications; Springer Science & Business Media: New York, NY, USA, 2012. [Google Scholar]
  28. Le-Yi, W.; Wen-Xiao, Z. System identification: New paradigms, challenges, and opportunities. Acta Autom. Sin. 2013, 39, 933–942. [Google Scholar]
  29. Wei, M.; Yang, J.; Zhao, Z.; Zhang, X.; Li, J.; Deng, Z. DeFedHDP: Fully Decentralized Online Federated Learning for Heart Disease Prediction in Computational Health Systems. IEEE Trans. Comput. Soc. Syst. 2024, 11, 6854–6867. [Google Scholar] [CrossRef]
  30. Zhang, H.; Liang, S.; Ou, M.; Wei, M. An asynchronous distributed gradient algorithm for economic dispatch over stochastic networks. Int. J. Electr. Power Energy Syst. 2021, 124, 106240. [Google Scholar] [CrossRef]
  31. Wei, M.; Yu, W.; Chen, D.; Kang, M.; Cheng, G. Privacy Distributed Constrained Optimization Over Time-Varying Unbalanced Networks and Its Application in Federated Learning. IEEE/CAA J. Autom. Sin. 2025, 12, 335–346. [Google Scholar] [CrossRef]
  32. La Bella, A.; Farina, M.; Sandroni, C.; Scattolini, R. Design of aggregators for the day-ahead management of microgrids providing active and reactive power services. IEEE Trans. Control Syst. Technol. 2019, 28, 2616–2624. [Google Scholar] [CrossRef]
  33. Messinger, S.; Fehl, M.; Miller, S.W.; Zugger, M.; Yukish, M. Applications of System Identification with Sparse Bayesian Regression Discovery of Unmodeled Dynamics of an Airship. In 2022 American Control Conference (ACC); IEEE: Piscataway, NJ, USA, 2022; pp. 1414–1419. [Google Scholar]
  34. Zhang, L.; Schaeffer, H. On the convergence of the SINDy algorithm. Multiscale Model. Simul. 2019, 17, 948–972. [Google Scholar] [CrossRef]
Figure 1. Time series forecasting comparison of ARIMA and SBL in the steel industry.
Figure 1. Time series forecasting comparison of ARIMA and SBL in the steel industry.
Electronics 15 00530 g001
Figure 2. Time series forecasting comparison of SINDy and LSTM in the steel industry.
Figure 2. Time series forecasting comparison of SINDy and LSTM in the steel industry.
Electronics 15 00530 g002
Figure 3. Time series forecasting comparison of GRU and ResNet in the steel industry.
Figure 3. Time series forecasting comparison of GRU and ResNet in the steel industry.
Electronics 15 00530 g003
Figure 4. Time series forecasting comparison of ARIMA and SBL in the photovoltaic-related manufacturing industry.
Figure 4. Time series forecasting comparison of ARIMA and SBL in the photovoltaic-related manufacturing industry.
Electronics 15 00530 g004
Figure 5. Time series forecasting comparison of SINDy and LSTM in the photovoltaic-related manufacturing industry.
Figure 5. Time series forecasting comparison of SINDy and LSTM in the photovoltaic-related manufacturing industry.
Electronics 15 00530 g005
Figure 6. Time series forecasting comparison of GRU and ResNet in the photovoltaic-related manufacturing industry.
Figure 6. Time series forecasting comparison of GRU and ResNet in the photovoltaic-related manufacturing industry.
Electronics 15 00530 g006
Figure 7. Time series forecasting comparison of ARIMA and SBL in the chemical industry.
Figure 7. Time series forecasting comparison of ARIMA and SBL in the chemical industry.
Electronics 15 00530 g007
Figure 8. Time series forecasting comparison of SINDy and LSTM in the chemical industry.
Figure 8. Time series forecasting comparison of SINDy and LSTM in the chemical industry.
Electronics 15 00530 g008
Figure 9. Time series forecasting comparison of GRU and ResNet in the chemical industry.
Figure 9. Time series forecasting comparison of GRU and ResNet in the chemical industry.
Electronics 15 00530 g009
Table 1. Annual final energy consumption of major manufacturing sub-sectors (China, 2019).
Table 1. Annual final energy consumption of major manufacturing sub-sectors (China, 2019).
Sub-SectorValue-Chain RoleFinal Energy (PJ)
Ferrous metals (steel)Upstream (high-intensity)20,715.56
ChemicalsUpstream (high-intensity)14,165.77
Non-metallic mineralsUpstream8163.61
Non-ferrous metalsUpstream3715.99
TextilesDownstream (representative)1229.89
Table 2. Data storage format example.
Table 2. Data storage format example.
No.Column NameMeaning
1–96Total Active Power P 1 , P 2 , , P 96 Power consumption in 96 observations with a 15 min interval over a single day.
97TimeRepresents the date of observation, e.g., 1 July 2023.
98Customer IDIdentifies the observed customer for the current date.
Table 3. MAPE comparison for a representative customer in the three industries.
Table 3. MAPE comparison for a representative customer in the three industries.
ModelSteelPVChemical
SIABF7.029%4.502%9.022%
ARIMA9.709%22.947%8.420%
SBL11.483%15.303%14.143%
SINDy55.953%33.908%64.315%
LSTM9.734%18.988%16.742%
GRU9.919%15.132%18.570%
ResNet10.640%27.771%19.064%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Sheng, L.; Wang, Z.; Wang, X.; Zhu, L. Adaptive Bayesian System Identification for Long-Term Forecasting of Industrial Load and Renewables Generation. Electronics 2026, 15, 530. https://doi.org/10.3390/electronics15030530

AMA Style

Sheng L, Wang Z, Wang X, Zhu L. Adaptive Bayesian System Identification for Long-Term Forecasting of Industrial Load and Renewables Generation. Electronics. 2026; 15(3):530. https://doi.org/10.3390/electronics15030530

Chicago/Turabian Style

Sheng, Lina, Zhixian Wang, Xiaowen Wang, and Linglong Zhu. 2026. "Adaptive Bayesian System Identification for Long-Term Forecasting of Industrial Load and Renewables Generation" Electronics 15, no. 3: 530. https://doi.org/10.3390/electronics15030530

APA Style

Sheng, L., Wang, Z., Wang, X., & Zhu, L. (2026). Adaptive Bayesian System Identification for Long-Term Forecasting of Industrial Load and Renewables Generation. Electronics, 15(3), 530. https://doi.org/10.3390/electronics15030530

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop