Next Article in Journal
Tumor-Infiltrating Immune Cells in Non-Muscle-Invasive Bladder Cancer: Prognostic Implications, Predictive Value, and Future Perspectives
Next Article in Special Issue
Delivery Reliability Assessment for a Multistate Smart-Grid Network with Transmission-Loss Effect
Previous Article in Journal
Hybrid Drying of Apples: A Comparison of Continuous and Intermittent Process Modes
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A New Approach to Forecast Intermittent Demand and Stock-Keeping-Unit Level Optimization for Spare Parts Management

by
Dimitrios S. Sfiris
1,* and
Dimitrios E. Koulouriotis
2
1
AspectSoft, Andreou Dimitriou 35 St., 67133 Xanthi, Greece
2
School of Mechanical Engineering, National Technical University of Athens, 15772 Athens, Greece
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(22), 12030; https://doi.org/10.3390/app152212030
Submission received: 25 September 2025 / Revised: 20 October 2025 / Accepted: 8 November 2025 / Published: 12 November 2025
(This article belongs to the Special Issue Smart Service Technology for Industrial Applications, 3rd Edition)

Abstract

The intermittent and lumpy demand of spare parts requires the choice of the right forecasting model among a variety of existing methods. Spare parts have an uneven lifecycle and mean time to failure for each individual item. As a result, they have a varied time of replacement, and consequently, a varied demand. This paper introduces a multi-cost function optimization approach that dynamically selects and adjusts forecasting models tailored to each spare part. The performance comparisons of the various demand forecasting methods led us to a new forecasting model, the Sfiris–Koulouriotis (SK) method, suited for highly lumpy and intermittent demand. A scaled version of the novel Stock-Keeping Unit-oriented Prediction Error Costs metric is also introduced. The composite negative-binomial–Bernoulli probability distribution for the stock control leveraged the replenishment policy. The best safety stock level is calculated for each individual item. Empirical validation in the automotive industry demonstrated that our approach significantly reduces safety stock while maintaining service levels, offering practical benefits for inventory management.

1. Introduction

1.1. Background and State of the Art in Spare Parts Demand Forecasting

Slow-moving spare parts often account for a large share of inventory value, with industry reports indicating up to 60% of total stock [1]. Their demand is intermittent and unpredictable, creating significant forecasting challenges and increasing the risk of both stock-outs and excess inventory under common stock control policies [2]. Effective management therefore requires accurate categorization by demand type and the selection of forecasting methods tailored to each stock-keeping unit (SKU) [3].
Intermittent-demand forecasting was formalized by Croston, who modeled separately (i) the sizes of non-zero demands and (ii) the intervals between them, applying simple exponential smoothing (SES) to each and taking their ratio for per-period forecasts [4]. Although widely implemented in ERP and forecasting software, Croston’s method was shown to be positively biased, leading to the Syntetos–Boylan Approximation (SBA), which mitigates this bias and often improves accuracy [5,6]. Later methods addressed issues such as obsolescence and declining demand; the Teunter–Syntetos–Babai (TSB) model replaces inter-demand intervals with demand-occurrence probabilities that decay exponentially, while hyperbolic exponential smoothing (HES) adopts hyperbolic decay. Both seek better performance as demand fades or disappears [7,8,9]. Reviews summarize these developments and the remaining challenges in handling extreme lumpiness, shifting occurrence processes, and industry-specific constraints [10,11].
Applied studies in spare-parts contexts reinforce the managerial importance of appropriate method choice and classification [12,13,14]. Despite several surveys, further work is needed to test alternative paradigms and validate methods on operational datasets [15].
Evaluating accuracy for intermittent series is non-trivial. Traditional metrics (MAE, MSE) can be scale-dependent and insensitive to long zero runs, while MAPE becomes unstable in the presence of many zeros [16,17]. Aggregation remedies may blur the timing of events and misrepresent service impact. Scaled metrics such as MASE aid cross-series comparison, and inventory-linked measures like periods in stock (PIS) and its scaled variant (sAPIS) connect errors to service performance, though difficulties persist for highly lumpy profiles [18,19]. More recently, the stock-keeping-unit-oriented prediction error costs (SPEC) metric [20] reframed evaluation in cost terms; here we adapt it into a scaled SPEC (sSPEC) tailored to automotive parts and use it alongside MASE and sAPIS for a multifaceted assessment. Recent surveys highlight movement toward such domain-aware metrics and hybrid evaluation frameworks [21].
Accurate demand categorization underpins these choices. Early schemes grouped series by variance into “smooth,” “sporadic,” and “slow-moving” categories [22,23]. Building on the concept of intermittence in Croston-type models, a widely used refinement distinguishes smooth, intermittent, lumpy, and erratic classes via demand frequency and size variability; this taxonomy is now standard in spare-parts research and practice [24,25,26].

1.2. Objectives and Structure of the Article

This study pursues the following six objectives: (i) to improve forecasting and inventory decisions for slow-moving automotive spare parts through a stock-keeping-unit-wise multi-method selection and optimization strategy; (ii) to introduce the Sfiris–Koulouriotis (SK) method for a subset of intermittent demand profiles, building on the Teunter–Syntetos–Babai framework by estimating demand occurrence non-parametrically as the empirical frequency of occurrence (non-zero counts over elapsed periods), enabling real-time adjustment with fewer parameters; (iii) to develop an SSA-aided variant (SK–SSA) aimed at stabilizing forecasts for highly lumpy and erratic series; (iv) to evaluate performance using a scaled stock-keeping-unit-oriented prediction error costs metric (sSPEC) alongside established measures such as MASE and sAPIS; (v) to examine demand classification and per-class selection policies, ensuring interpretability and practical deployability across heterogeneous items; and (vi) to connect forecast accuracy to operations via a continuous-review (R,Q) policy with negative-binomial–Bernoulli lead-time demand (LTD), reporting results in terms of service levels, safety stock, and backorders.
Briefly, the integrated framework improves accuracy over strong Croston-family baselines and yields tangible inventory gains. SK provides additional improvements for targeted intermittent profiles with low parameter burden; SK–SSA enhances stability on highly lumpy and erratic series. Using sSPEC to guide tuning leads to better inventory outcomes than traditional error metrics alone, and the (R,Q) + LTD linkage translates accuracy into lower safety stock and fewer backorders. The procedure remains computationally light and compatible with enterprise systems.
Recent advances in machine learning and probabilistic forecasting have also targeted intermittent demand. Transformer-based neural networks [27], modified k-Nearest-Neighbor frameworks [28], and feature-based or probabilistic combination models [29,30] have achieved competitive accuracy in richer data environments. However, these methods typically require long, dense training histories, extensive hyper-parameter tuning, and specialized hardware, which limits their deployability for sparse, SKU-level demand in enterprise systems. In contrast, the present study prioritizes parameter-light, interpretable models with closed-form updates that can operate efficiently within existing ERP infrastructures. The proposed SK and SK–SSA methods therefore complement rather than compete with data-hungry approaches, providing an implementable solution for organizations managing thousands of low-volume parts.
The remainder of the paper is organized as follows. Section 2 (Materials and Methods) presents the study context, data, and preprocessing (Section 2.1); the forecasting models, including comparators and proposed methods (Section 2.2); the error metrics, cost functions, and optimization objective (Section 2.3); the MCOST optimization framework, training protocol, and computing setup (Section 2.4); demand pattern classification and methods grouping (Section 2.5); the inventory control model and lead-time demand specification (Section 2.6); and the statistical testing, robustness analysis, and addressed research gaps (Section 2.7). Section 3 reports the empirical results. Section 4 provides the discussion. Section 5 presents the conclusions. A complete list of symbols and notation used throughout the paper is provided in Appendix B (Table A2).

2. Materials and Methods

2.1. Study Context, Data, and Preprocessing

The study focuses on slow-moving automotive spare parts with intermittent demand. For each stock-keeping unit (SKU), demand per period is recorded as a non-negative integer. Transactional records were aligned to a common observation window; missing entries were treated as zero demand; and invalid records (negative or non-integer values) were removed. Data were captured daily and aggregated to weekly series for forecasting. Each SKU contributes 104 consecutive weekly observations. Calendar effects such as public holidays and plant shutdowns were not modeled explicitly; any such effects are absorbed into the weekly totals. No seasonal adjustment was applied, consistent with the predominance of intermittent, low-signal series. All experiments use a univariate, one-step-ahead forecasting setup. Aggregate approaches from the literature (e.g., ADIDA, IMAPA) are acknowledged due to their relevance to intermittence via temporal aggregation [31,32], but they are not included in the empirical comparisons so as to maintain computational efficiency and a transparent link to inventory policy.

2.2. Forecasting Models: Scope, Comparators, and Proposed Methods

The model pool comprises established intermittent-demand methods together with newly developed variants. Closed-form update equations are provided in Table 1, and a concise summary of each method’s purpose and key contribution is given in Table 2. The comparator methods are as follows:
  • Croston-type models: Croston decomposes demand into a non-zero demand size and an inter-demand interval, each updated by exponential smoothing; the per-period forecast is the ratio of these components [4]. Its positive bias motivated the Syntetos–Boylan Approximation (SBA), which reduces overestimation under intermittence [6];
  • Probability-of-occurrence models: Teunter–Syntetos–Babai (TSB) replaces inter-demand intervals with a demand-occurrence probability updated every period, while the demand-size component is updated only when non-zero demand occurs, thereby accommodating obsolescence and declining usage [8];
  • Modified Croston variants: mSBA updates the inter-demand interval during zero-demand runs by comparing observed and estimated spacing; mTSB analogously updates the occurrence probability during zero-demand periods [33].
The new methods introduced in this study are as follows:
  • Sfiris–Koulouriotis (SK): Retains TSB’s probability-of-occurrence concept but estimates occurrence non-parametrically as the ratio of non-zero to zero occurrences within an online window, reducing free parameters and enabling responsive updates;
  • SK–SSA: Applies a Singular Spectrum Analysis–inspired stabilization to highly lumpy sequences prior to the SK updating scheme to improve robustness.
The field of intermittent demand forecasting is evolving. Recent advancements include Transformer-based models, modified k-nearest-neighbor frameworks, and feature-based or probabilistic combinations [27,28,29,30]. The present study confines its empirical analysis to interpretable, non-aggregated Croston-family methods and the proposed SK/SK–SSA variants for three reasons: (i) alignment with an explicit inventory objective (sSPEC) and direct LTD linkage; (ii) item-wise deployment with closed-form, CPU-only updates suitable for ERP environments; and (iii) consistent evidence that sparse, short histories limit the practical advantage of data-hungry global models in spare-parts settings. A head-to-head comparison with recent machine-learning approaches is therefore outside the scope and left for future work.
We herein define the SK forecasting method for intermittent demand, extending the probability-of-occurrence idea of TSB while simplifying its updating and improving stability during extended zero-demand stretches.
The one-step-ahead forecast is
y ^ t + 1 = d ^ t z ^ t ,
where d ^ t denotes the estimated probability of non-zero demand (occurrence probability) and z ^ t is exponentially smoothed size of non-zero demands (as in Croston-type/TSB schemes).
Let y t denote the observed demand in period t , and I s = 1 y s > 0 be the indicator of a non-zero demand in period s . The cumulative count of non-zero periods is n z t = s = 1 t I s . The size component of the SK method follows an exponential-smoothing update (only when demand occurs) as follows:
z ^ t = z ^ t 1 = z ^ t + α y t z ^ t 1 , y t > 0 z ^ t 1 = z ^ t , y t = 0 α 0 , 1
and the occurrence probability is estimated non-parametrically by
d ^ t c u m = n z t t = s = 1 t I s t
Equation (8) is the Bernoulli cumulative maximum-likelihood estimate under stationarity. A Laplace correction can be used at the start of a series as follows:
d ^ t = n z t + α 0 t + α 0 + β 0 ,         α 0 , β 0     0
Optionally, for greater responsiveness, a trailing-window estimator may replace (or be blended with) Equation (8). Let W t = min W , t be the available window length and define the windowed count of non-zero periods as
n z W t = s = t W t + 1 t I s .
The corresponding occurrence estimate (with optional smoothing) is
  d ^ t W = n z W t + α 0 W t + α 0 + β 0
A convex blend of the windowed and cumulative estimators balances stability and adaptability as follows:
d ^ t = λ d ^ t W + 1 λ d ^ t c u m ,     λ 0 , 1
To prevent numerical instability when early observations are zero or near-zero, initialization proceeds as
z ^ 0 = max y 1 , ε ,     ε   >   0   small ,  
d ^ 0 = α 0 α 0 + β 0 ,  
A lower bound is imposed to ensure numerical stability during long zero-demand runs as follows:
d ^ t δ , 1 δ , δ = 1 W t + α 0 + β 0
The size estimate is similarly bounded by enforcing z ^ t ε after each update of Equation (7).
During extended zero-demand stretches, z ^ t remains fixed per Equation (7), while d ^ t declines via Equations (8)–(11). Consequently, the forecast y ^ t + 1 = d ^ t z ^ t contracts appropriately for intermittent or lumpy demand. A concise pseudocode implementation is provided in Algorithm A1 (Appendix A) for reproducibility.
To improve robustness under erratic spikes and long zero runs, we introduce a stabilized variant, SK–SSA, which combines the SK structure with Singular Spectrum Analysis (SSA) smoothing. SSA decomposes the non-zero size series into additive components via lag embedding and singular-value decomposition [34,35], then reconstructs a denoised signal by retaining the leading k components.
Let z ˜ t denote the SSA-reconstructed non-zero size at time t obtained from the subset { z s : y s > 0 } . The SK–SSA update replaces y t in Equation (7) by z ˜ t :
z ^ t = z ^ t 1 = z ^ t + α z ˜ t z ^ t 1 , y t > 0 , z ^ t 1 = z ^ t , o t h e r w i s e ,
while keeping d ^ t unchanged from SK. The one-step forecast remains y ^ t + 1 = d ^ t z ^ t .
Given an embedding window length L S S A and rank k L S S A , SSA forms an L S S A × k trajectory matrix from the (non-zero) size series, computes its singular-value decomposition, and reconstructs the smoothed signal by diagonal averaging of the first k components.
In practice, small values such as L S S A 6 , 24 and k 1 , 2 , 3 avoid oversmoothing short or highly lumpy series. When zeros create gaps, the lagged autocovariance is computed from available pairs only (missing-pair omission), and the reconstructed signal is mapped back to the original time index. All reconstructions are bounded within ε , z max to preserve the same numerical safeguards used for z ^ 0 . The smoothing parameter α is kept consistent with the SK update. SSA parameters L S S A , k can be tuned on the training window alongside α , or fixed at light defaults (e.g., L S S A = 12 , k = 2 ) for parsimony.
From a computational standpoint SK–SSA adds O ( L S S A 2 ) work per refresh, which reduces to O ( 1 ) per period if updated periodically, remaining CPU-efficient for ERP-scale deployment. Empirically, SK–SSA reduces variance in z ^ t without masking intermittence, yielding lower inventory-aligned costs under sSPEC.
Parameterization followed literature-informed bounds. SMA used a window length W 10 , 30 . SES used a single smoothing parameter α 0.10 , 0.40 . CRO used α 0.10 , 0.40 . SBA used α 0.05 , 0.40 with a second parameter β 0.05 , 0.40 . TSB used α 0.05 , 0.40 with a second parameter β 0.10 , 0.40 . SK used α 0.05 , 0.40 . mSBA used α 0.05 , 0.40 with a second parameter β 0.05 , 0.40 . mTSB used α 0.05 , 0.40 and β 0.05 , 0.40 . SK–SSA used α 0.05 , 0.40 with SSA components k 1 , 2 , 3 . Final values were obtained via a rolling-origin search; for each SKU and chosen objective, the retained specification is the model–parameter combination with the lowest out-of-sample loss.

2.3. Error Metrics, Cost Functions, and Optimization Objective

To capture different operational priorities and the specific challenges of intermittent series, both conventional and intermittent-aware measures are considered. Formal definitions are given in Table 3. Mean squared error (MSE) and mean absolute error (MAE) provide conventional accuracy assessments, while mean squared rate (MSR) and mean absolute rate (MAR) operate on demand rates to mitigate the effect of long zero runs. Periods in stock (PIS) and absolute PIS (APIS) relate forecast deviation to inventory drift over time. Mean absolute scaled error (MASE) and absolute scaled error (ASE) support scale-independent benchmarking across series. The stock-keeping-unit–oriented prediction error costs (SPEC) metric integrates both stock-keeping and opportunity costs through two non-negative weighting parameters, α 1 and α 2 , which balance under- and over-forecast penalties. Its scaled form (sSPEC) normalizes SPEC by each SKU’s in-sample mean demand, allowing performance to be compared across items with different magnitudes while retaining the same economic interpretation. Further details and illustrative examples of the SPEC cost structure can be found in [20], from which the present sSPEC formulation is adapted.
In line with the study’s inventory focus, sSPEC serves as the primary optimization criterion, complemented by MASE and sAPIS for robustness across demand types and scales. In this study, α 1 and α 2 were set to equal weights ( α 1 = α 2 = 1) to ensure balanced penalization of over- and under-forecasting and to maintain comparability across SKUs; in applied settings, these coefficients can be adjusted to reflect organization-specific cost structures.
The optimization objective is chosen prior to model tuning. For each SKU and chosen cost, parameters are searched by rolling-origin evaluation with expanding windows, generating one-step-ahead forecasts across multiple origins. Parameters are updated based on recent forecast performance to adapt to evolving demand characteristics, and final performance is assessed on a reserved test horizon. Because no single objective consistently reflects performance across smooth, intermittent, erratic, and lumpy profiles, six costs—MSE, MAE, APIS, MAR, MSR, and SPEC—are reviewed and used within the framework. This multi-cost setup enables context-dependent selection and supports robust comparisons across diverse demand scenarios.

2.4. Optimization Framework, Training Protocol, and Computing Setup

This study employs a multi-method, multi-cost (MCOST) optimization carried out per stock-keeping unit (SKU). Let M = { SES ,   CRO ,   SBA ,   TSB ,   SK ,   mSBA ,   mTSB ,   SK-SSA } denote the set of forecasting methods ( | M | = 9 ) and C = { MSE ,   MAE ,   PIS ,   MSR ,   MAR ,   SPEC } denote the set of optimization objectives ( | C | = 6 ). This results in 9 × 6 = 54 method–objective cases evaluated per SKU. Parameter selection follows a cost-first protocol: choose the objective, then tune model parameters via rolling windows with expanding origin; an approach advocated for intermittent series [19] and evaluated on a reserved test segment in line with recommended practice [36].
For each SKU and case m , c M × C , parameters are searched within the bounds in Section 2.2 using a rolling-origin scheme that produces one-step-ahead forecasts at each origin. The in-optimization objective is the chosen cost c . This rolling procedure adaptively refines parameters based on recent performance, allowing the fitted specification to track changes in occurrences (zeros versus non-zero) and non-zero size processes over time, consistent with the dynamic tuning rationale of cost-function optimization in [19]. When the optimizer converges to a boundary or the objective is monotone over the admissible range, the solution is retained and flagged as a boundary solution; a within-bounds success indicator is recorded per case. The best parameter vector for m , c is then refit on the full training segment and evaluated on a reserved test window; results are summarized by averaging over multiple forecast horizons.
Traditional level-based losses (MSE, MAE) can be dominated by long zero runs. Rate-based objectives (MSR, MAR) operate on occurrence-adjusted demand rates; inventory-linked measures (PIS) and cost-based SPEC align tuning with managerial outcomes. For reporting and cross-SKU comparison, we use scale-independent summaries (MASE, sAPIS) and the scaled inventory cost sSPEC; sSPEC is emphasized as the primary inventory-facing criterion in the main results.
To compare methods and costs fairly, we down-weight cases that frequently hit parameter bounds. Let n s k u be the number of SKUs. For each i , j C × M , let F i j be the number of SKUs whose optimized parameters are within bounds. The within-bounds success rate is
p i j = n s k u F i j n s k u
Let V i j denote the out-of-sample error (e.g., sSPEC) for case i , j . The SKU-average error over all SKUs is
A i j = 1 n s k u s k u V i j
Let C i j = n s k u F i j be the number of successful (within-bounds) optimizations and
O i j = 1 C i j s k u V i j
the average performance over successful optimizations. The adjusted ranking matrix R | C | × | M | blends these via the success probability:
R i j = O i j p i j + A i j + O i j p i j 1 p i j
For each method j M , the preferred cost is the column i * j with the lowest R i j . Final per-SKU selection then proceeds as follows: for each SKU, among all methods evaluated under their preferred cost i * j , retain the model–parameter combination with the lowest out-of-sample loss under the MCOST criterion. This implements the cost-first idea (choose objective, then tune via rolling windows) while correcting comparisons for infeasible/boundary-dominated cases. The overall selection process is summarized in the selection procedure of Table A1 and detailed in pseudocode in Algorithm A3 (Appendix A).
For each SKU, a single temporal split was used—95% of observations for optimization and 5% held out for evaluation. This short holdout window (≈5 weeks) was selected to preserve sufficient data for model learning given the limited 104-week series, while still providing an independent performance check. Each candidate parameter combination on a discrete grid (covering smoothing factors, window lengths, and SSA components) was evaluated on these rolling forecasts, and the setting with the lowest in-sample loss under the chosen cost function was retained. This grid-based rolling-origin tuning yields adaptive yet reproducible estimates without stochastic effects associated with random search. Final performance was then computed on the reserved out-of-sample segment. SKU-level losses were averaged over the out-of-sample window, and dataset-level summaries report medians and interquartile ranges across SKUs. Where multi-horizon results are shown, metrics (e.g., sSPEC) are averaged over horizons h = 1, …, 5.
No GPU acceleration was used. All computations were performed on a desktop with an AMD Ryzen 5 1600 (six cores) and 64 GB RAM. This setup underscores the practicality and parsimony of the approach relative to heavier machine-learning alternatives, and supports ERP-scale deployment.

2.5. Demand Pattern Classification and Methods Grouping

Demand classification informs both model choice and interpretation. Each time series is categorized as smooth, intermittent, erratic, or lumpy using the average inter-demand interval ( ADI ) and the squared coefficient of variation of non-zero demand ( C V 2 ). Thresholds and formulas are given in Equations (31) and (32), and decision regions appear in Figure 1; notation follows Table A2. Following the KH-Select scheme (Kostenko and Hyndman’s refinement of Syntetos–Boylan–Croston classification) [25], we assign each stock-keeping unit (SKU) to one of the four classes and consider a tailored candidate set per class (SMA, SES, CRO, SBA, TSB, mSBA, mTSB, SK, SK–SSA).
Consistent with Kourentzes (2014) [19], KH-Select assignment improves prediction by aligning method families with observed intermittence and size variability, without relying on ad hoc cut-offs tied to a single model. In contrast to the original SBC classification, which primarily contrasted CRO and SBA [24], KH-Select expands the candidate pool to include probability-of-occurrence methods (e.g., TSB), modified Croston variants (mSBA, mTSB) [33], and the proposed SK and SK–SSA. Within each class, we document the frequency with which methods are ultimately selected, providing practitioners with transparent, class-specific preferences.
Importantly, classification is not the terminal decision. Because accuracy alone may not yield the best inventory outcomes, the final per-SKU choice is made by the multi-method, multi-cost selection (MCOST; Section 2.3 and Section 2.4), which evaluates competing model–cost pairs on out-of-sample loss and supports inventory-aligned objectives (e.g., sSPEC). This preserves the interpretability of KH-Select while allowing the data, and the chosen objective, to determine the best method for each SKU. Inventory implications are then quantified under the continuous-review (R,Q) policy with negative-binomial–Bernoulli (NBB) lead-time demand (Section 2.6).
A D I = Total   Number   of   Periods Total   Number   of   Demand   Buckets
C V = Standard   Deviation   of   Demand   Sizes Demand   Mean
The four classes are as follows:
  • Smooth: A D I 1.32 and C V 2 0.49 ; typically fast-moving items.
  • Intermittent: A D I > 1.32 and C V 2 0.49 ; sparse arrivals with relatively stable non-zero sizes.
  • Erratic: A D I 1.32 and C V 2 > 0.49 ; frequent demand with volatile size.
  • Lumpy: A D I > 1.32 and C V 2 > 0.49 ; long zero runs and highly variable non-zero sizes.

2.6. Inventory Control Model and Lead-Time Demand Specification

Forecast accuracy is linked to operational performance through a continuous-review R ,   Q policy with backordering. When the inventory position reaches the reorder point R , a fixed quantity Q is ordered. Lead-time demand (LTD) is the random demand accruing over the supplier lead time L ; in the empirical study L is treated as deterministic and constant at three days ( L = 3 / 7 week). Cycle-service targets (probability of no stockout during a replenishment cycle) determine R , and safety stock is SS = R E LTD .
To capture both over-dispersion in demand sizes and intermittence in arrivals, LTD is modeled with a composite negative-binomial–Bernoulli (NBB) distribution: a negative binomial (NB) for non-zero demand sizes and a Bernoulli for occurrence. This choice follows the literature recommending NB for spare-parts demand and Bernoulli for arrivals under intermittence. Formally, the NB size component is parameterized by r , p , and the arrival process by b ; the construction and basic properties used here are given in Equations (33)–(35). In our implementation, per-period mean and dispersion implied by the selected forecast (Section 2.4) are mapped to r , p , d by method-of-moments estimators (Equations (36)–(38)); edge cases are handled by capping r 1 when the implied value is <1, corresponding to geometric limit.
P k = k 1 r 1 p r 1 p k r ,   k r
The above equation is constructed based on the summation of geometric distributions, and it is applicable within the set of natural numbers {1, 2, 3, …}, where 0 < p 1 ,   r 1   and   integer , having the following:
μ = r p ,  
σ 2 = r 1 p p 2 .
LTD equals L multiplied by, aka the total demand expected over the lead time. The parameters p and r are estimated as follows:
p = L T D σ L T D 2 + L T D
r = I N T L T D p .
INT denotes the integer part function. In those cases where r < 1, we set r = 1; this corresponds to the limit distribution when p is very small. The parameter b of NBB is calculated using the following formula:
b = 1 A D I
where ADI is the average inter-demand interval (see Equation (31)).
Given r , p , d and lead time L , closed-form LTD moments and quantiles are used to set policy parameters. The reorder point R is obtained from the upper quantile of the fitted NBB LTD at the chosen cycle-service target— SS = R E LTD . Since Z-scores do not apply directly to NBB, safety stock is determined via the NBB quantile (or its fast numerical approximation) rather than a normal approximation; the relationship to the heuristic form SS = k σ L T D (Equations (39)–(41)) is recovered by interpreting k as the NBB-implied service factor for the target.
SS = k × μ × μ r
SS = k × r r p p
SS = k × σ
where σ 2 = μ 2 μ is the standard deviation of the NBB distribution.
To balance stock holding and service, we simulate LTD draws from the fitted NBB (100,000 replications per setting in the empirical study) and trace (i) backlog–safety-stock trade-off curves and (ii) realized versus target service levels. This directly links forecasting choices to inventory outcomes and guides the selection of SS for practical service targets (e.g., 0.80, 0.85, 0.90, 0.95, 0.97, 0.99) using the same NBB parameters that encode demand variability and arrival intermittence.
This procedure bases R and SS on the full LTD distribution—sizes and arrivals—rather than on ad-hoc buffers, leading to more effective and transparent inventory decisions under intermittent demand. In this study, adequacy of the distribution fit was verified empirically by confirming that simulated cycle-service levels and backlog–safety-stock curves under the fitted parameters align with the policy targets. Formal goodness-of-fit tests were not pursued, as prior evidence supports modeling non-zero sizes with a NB distribution and arrivals with a Bernoulli process for discrete intermittent (see [37]). Degenerate cases with implied r < 1 were handled by setting r = 1 (geometric limit) as noted in Equation (37).

2.7. Statistical Testing, Robustness Analysis, and Addressed Research Gaps

Model performance is evaluated at the SKU level and then summarized across SKUs and forecast horizons. Results are reported under multiple accuracy and cost measures to reflect different managerial perspectives (for example, sSPEC as the primary inventory-facing criterion, with MASE, sAPIS, MAR, MSR, MSE, and MAE as complementary views). The reporting emphasizes per-SKU outcomes and aggregated summaries (means/medians and dispersion), avoiding assumptions that are unlikely to hold for intermittent series.
To assess the statistical significance of observed performance differences, pairwise Wilcoxon signed-rank tests were applied across all SKUs for each forecasting method under the primary (sSPEC) and secondary (MASE) metrics. Holm correction was used to control the family-wise error rate. Non-parametric bootstrap confidence intervals (CIs) were also computed for median sSPEC differences between SK-based methods and each baseline. Detailed results, including parameter-sensitivity and computational-efficiency analyses, are reported in Appendix C.
Robustness is examined along three practical dimensions. First, the optimization objective is varied; each model is re-optimized under alternative costs (sSPEC, MASE, sAPIS, MAR, MSR) and conclusions are compared to check that findings are not an artifact of a single metric. Second, sensitivity to parameter ranges is assessed by widening and narrowing the grids for exponential-smoothing and probability-update parameters; stability of the selected model–parameter combinations is inspected. Third, heterogeneity across demand types is explored by repeating summaries on the most challenging subset—items classified as highly intermittent or lumpy by the demand-categorization scheme—verifying that the proposed methods remain competitive where zero runs and size variability are most pronounced. Inventory implications are computed using the negative-binomial–Bernoulli specification within a continuous-review (R,Q) policy, ensuring that accuracy gains translate into service and stock-holding outcomes.
The analyses are designed to address the research gaps identified in the literature. Specifically, the study contributes the following: (i) the SK forecasting model and its SK–SSA analogue, targeting highly intermittent and lumpy demand while maintaining parsimony; (ii) the scaled stock-keeping-oriented prediction error costs metric (sSPEC), aligning model tuning and evaluation with inventory-relevant costs and enabling cross-SKU comparability; (iii) a multi-cost, multi-method optimization framework that selects the most suitable forecasting method per SKU rather than prescribing a single best model; (iv) an inventory linkage through the negative-binomial–Bernoulli representation of lead-time demand embedded in an (R,Q) policy; and (v) a refined use of demand categorization to guide method selection. Section 3 reports the empirical assessment of these components, and Table 4 maps the targeted literature gaps to the corresponding elements of the framework.

3. Results

The detailed, comprehensive evaluation of the proposed framework was conducted on an industrial automotive dataset of 2050 stock-keeping units (SKUs), each with 104 weekly observations. This section provides a concise, precise account of the empirical results, their interpretation, and the experimental conclusions. The reporting mirrors the experimental design in Section 2 and proceeds in the same order: data characteristics (Section 2.1); model behavior and parameterization (Section 2.2); objective functions and optimization (Section 2.3 and Section 2.4); demand classification (Section 2.5); and inventory implications under a continuous-review (R,Q) policy with negative-binomial–Bernoulli (NBB) lead-time demand (Section 2.6).
Weekly aggregation of daily transactions produced 104 observations per stock-keeping unit (SKU), as specified in Section 2.1. Descriptive statistics confirm the setting is dominated by intermittence and lumpiness: the cross-SKU average inter-demand interval is about 2.2 weeks, the average demand per period is 18.9 units, and non-zero sizes cluster around 34.9 units. These statistics, summarized in Table 5, support the use of Croston-family forecasting methods and intermittent-aware objectives.
The multi-cost selection protocol (MCOST; Section 2.4) was applied per SKU. Figure 2 depicts the workflow, weekly series construction, rolling-origin forecasting for all method–objective pairs, cost computation, ranking and parameter optimization, out-of-sample evaluation, and inventory analysis. This implementation mirrors the cost-first principle of Section 2.3 and operationalizes the training protocol and computing setup.
All candidate forecasters from Section 2.2 were tuned over literature-informed bounds (Table 6). Parameter optimization followed a deterministic grid-search procedure, independently applied per SKU and objective. For single-parameter methods (SMA, SES, CRO, SK), the smoothing constant α was evaluated on a uniform grid of 20 equally spaced points within the bounds shown in Table 6. For two-parameter methods (SBA, TSB, mSBA, mTSB, SK–SSA), a 20 × 20 grid was used across both smoothing parameters, resulting in 400 evaluations per SKU–objective pair. Each candidate combination was assessed via expanding rolling-origin evaluation, generating one-step-ahead forecasts across the training window. The configuration yielding the lowest in-sample loss under the chosen cost function was retained. Boundary optima were observed primarily in cases of monotone cost response under prolonged zero-demand runs, consistent with theoretical expectations.
To avoid overstating performance in cases with frequent boundary hits, the adjusted MCOST ranking matrix (Equation (30)) combines within-bounds averages with overall averages using the within-bounds success rate. This correction is applied throughout Table 7 and underpins the comparative statements below. Rate-based objectives (mean squared rate, MSR; mean absolute rate, MAR) tended to favor occurrence-aware methods (SK, SK–SSA), while the inventory-aligned costs (SPEC, sSPEC) more often selected methods that temper costly inventory drift (e.g., TSB when decay or obsolescence is present). Scaled stock-keeping-oriented prediction error costs (sSPEC) served as the primary criterion; mean absolute scaled error (MASE) and scaled absolute periods in stock (sAPIS) are reported as complementary views.
Optimized parameters concentrate away from extremes for most methods. Figure 3 and Figure 4 show bar charts and box plots of the tuned parameters across SKUs. Dispersion is wider in items with very long zero runs, reflecting the greater difficulty of stabilizing occurrence estimates; SK–SSA mitigates this by smoothing the size component prior to SK updating (Section 2.2).
Table 8 and Table 9 present a 104-week example with average inter-demand interval 1.60 weeks and squared coefficient of variation 0.37. Optimized parameters are listed for all methods; Table 10 compares performance using MASE, sAPIS, and sSPEC over five horizons. Results match known properties: CRO exhibits positive bias; SBA corrects that bias; TSB performs well when occurrence decays; mSBA improves SBA via interval updates during zero runs; mTSB is not uniformly dominant; SK sits between SBA and TSB; SK–SSA delivers the best overall performance for this SKU.
The KH-Select demand classification from Section 2.5 (smooth, intermittent, erratic, lumpy) was used to define scenario subsets and to report selection frequencies. Scenario-level averages across three horizons (t + 1, t + 3, t + 5) show MCOST ranking first overall, followed by MSR and MAR (Table 11). Selection frequencies by class (Table 12 and Figure 5) reveal that SK and SK–SSA are often preferred for intermittent and lumpy items, while SBA is frequently favored in lumpy cases, consistent with its interval-update logic.
Repeating the analysis on isolated intermittent, lumpy, and erratic subsets confirms two patterns (Figure 6 and the corresponding tables): (i) SK often outperforms TSB even without SSA on very sparse sequences, and (ii) SK–SSA is typically top-ranked across subsets, supporting the value of light stabilization before probability updates when zero runs are long and sizes are volatile. The 75th percentile scores, as indicated in Table 13, confirmed the notably strong forecasting performance for the majority of cases. Particularly noteworthy was the observation that our proposed SK model often outperformed TSB in several instances, establishing itself as a compelling candidate model, even without SSA.
A light sensitivity analysis was performed by varying the exponential-smoothing parameter α (0.05–0.4) and, for SK–SSA, the SSA window length L S S A 6 , 8 , 10 , 12 and number of reconstructed components k 1 , 2 , 3 . Forecast performance and inventory outcomes remained stable within these ranges, indicating robustness to moderate parameter shifts. Computational efficiency was also verified: on a standard desktop CPU (AMD Ryzen 5 1600 (six cores) with 64 GB RAM), full MCOST optimization across 2050 SKUs completed in under 15 min without GPU acceleration (Table 14), supporting ERP-scale deployability.
Furthermore, both the mSBA and mTSB methods exhibited superior performance compared to their counterparts, SBA and TSB, respectively, while the SK–SSA method consistently outperformed all others on average. The final column of Table 14 presents the optimal results obtained through demand categorization based on cost function minimization. The last column is a benchmark case representing the best achieved per-SKU cost (not a theoretical bound). Across all datasets, per-item optimization yielded substantial improvements in demand forecasting. This multi-method approach efficiently captured micro-trends and time decay by selecting the most appropriate forecast method for each item.
The bar chart comparisons in Figure 6 reaffirmed the superiority of the SK–SSA method, which emerged as the preferred choice across all three critical datasets. Additionally, the TSB method received high preference for the intermittent dataset, while the mSBA method excelled for the lumpy dataset. In the case of the erratic dataset, the SK–SSA, mSBA, SK, and TSB methods demonstrated comparable performance.
Per-SKU forecasts were embedded in the continuous-review (R,Q) model with backordering (Section 2.6). Reorder points R were set from NBB LTD quantiles at target cycle-service levels, with safety stock augmenting R. For each setting we simulated 100,000 NBB draws to estimate achieved service levels and backlog–safety-stock trade-offs. This simulation design directly implements the LTD specification and quantile-based R setting described in Section 2.6.
Backlog–safety-stock curves (Figure 7a) and realized service versus safety stock (Figure 7b) show that methods capturing micro-trends and decay deliver more efficient frontiers (curves closer to the axes). Figure 7c reports realized service levels relative to targets; Figure 7d compares scaled safety stock at a 97% target across methods and demand classes.
Table 15 summarizes scaled safety stock and backlogs across service levels for the full dataset. On average, SK and especially SK–SSA require lower safety stock at a given service level, while SBA and TSB also limit backorders effectively. Differences are more pronounced at the item level than in aggregate, reinforcing the value of per-SKU selection rather than a single, horizontal policy.
Figure 8 shows a positive association between improvements in forecasting metrics and reductions in backlog counts. Against the organization’s existing non-optimized SBA policy, the proposed MCOST-based framework achieved approximately 15–20% lower safety stock at matched service levels, without increasing backorders.
Three conclusions follow from the results: (i) per-SKU MCOST outperforms single-objective optimization on average; (ii) enlarging a small, interpretable Croston-based pool with SK and SK–SSA increases the chance of matching heterogeneous, intermittent patterns; and (iii) inventory outcomes improve materially when forecasts are operationalized via (R,Q) with NBB LTD, yielding stable reorder points and consistently lower safety stocks for a given target service level.
Non-parametric tests confirmed that SK–SSA significantly outperformed all other Croston-family models on the primary cost metric (sSPEC, p < 0.01 after Holm correction). SK also achieved significant gains (p < 0.05) over SBA and TSB in the intermittent and lumpy subsets. Bootstrap 95% confidence intervals for the median sSPEC differences further indicated stable positive effects of SK and SK–SSA, validating that the observed improvements are statistically robust.

4. Discussion

This study set out to improve forecasting and inventory decisions for intermittent and lumpy spare-parts demand by combining the following three elements: (i) a pool of interpretable Croston-family methods [Croston (CRO), Syntetos–Boylan Approximation (SBA), Teunter–Syntetos–Babai (TSB), and modified variants], complemented by a new Sfiris–Koulouriotis (SK) forecasting method and its Singular Spectrum Analysis variant (SK–SSA); (ii) an inventory-aligned evaluation metric, the scaled Stock-keeping-oriented Prediction Error Costs (sSPEC); and (iii) a per-SKU multi-cost optimization (MCOST) procedure that selects both the objective and the method based on empirical evidence rather than prescription. Across 2050 SKUs with 104 weekly observations each, the empirical results indicate consistent gains in forecast accuracy and, crucially, in inventory performance, while remaining computationally light and compatible with enterprise resource planning systems (Table 5, Table 6, Table 7, Table 8, Table 9, Table 10, Table 11, Table 12, Table 13, Table 14 and Table 15; Figure 2, Figure 3, Figure 4, Figure 5, Figure 6, Figure 7 and Figure 8).
For intermittent series, Croston-type forecasts are a natural baseline but are positively biased. SBA reduces that bias, and probability-of-occurrence updates as in TSB perform well when obsolescence or demand decay is present (Table 10). Building on these ideas, SK retains the TSB occurrence-probability rationale with simpler parameterization, and SK–SSA further stabilizes highly lumpy sequences. Overall, SK and SK–SSA frequently rank among the top choices under rate-based costs, that is mean squared rate (MSR) and mean absolute rate (MAR), and under sSPEC (Table 7, Table 9, Table 10, Table 11 and Table 12; Figure 5 and Figure 6). This supports the hypothesis that enlarging a small, interpretable pool increases the chance of matching heterogeneous item-level patterns. The strong performance of SK–SSA on the most challenging subsets (intermittent and lumpy) highlights the value of light smoothing of non-zero sizes before probability updates when zero runs are long and size variability is high.
In practice, SK is well suited to intermittent demand because it targets the operational driver for intermittence, namely the per-period demand-occurrence probability, while keeping the model simple and transparent. Particularly, SK (a) estimates the occurrence probability directly and pairs it with a single-parameter exponential smoother for non-zero size; (b) adapts quickly to regime shifts because it avoids a second smoothing layer on occurrence; (c) remains stable through long zero runs with straightforward start-up safeguards; (d) preserves interpretability as “probability of occurrence × typical size”; and (e) links naturally to the Bernoulli component of the negative-binomial–Bernoulli lead-time demand model used for inventory decisions. These properties help explain why SK and SK–SSA translate forecast gains into lower safety stocks at a given service level in our experiments.
Selecting a single accuracy metric can be misleading for intermittent series because traditional level-based losses [Mean Squared Error (MSE), Mean Absolute Error (MAE)] overweight magnitude and ignore spacing. Consistent with prior work [19], rate-based objectives (MSR, MAR) and inventory-linked costs [periods in stock (PIS), SPEC] capture different operational trade-offs. The MCOST protocol (Section 2.4; Figure 2) formalizes this by choosing the objective first, tuning parameters via rolling windows, and adjusting comparisons for boundary solutions through an explicit success-rate weighting. Empirically, MCOST dominates single-objective optimization on average (Table 7), and sSPEC is more discriminative than sAPIS for SK and TSB tuning, which supports the idea that a cost function that incorporates stock-keeping and opportunity costs is better aligned with managerial priorities. These results substantiate the working hypothesis that per-SKU, objective-aware optimization outperforms uniform horizontal tuning. Extended numerical summaries and sensitivity tests confirming these findings appear in Appendix C.
Using KH-Select for demand classification preserves interpretability and connects to established practice (Section 2.5). The framework supports multiple scenarios, while the per-SKU multi-cost optimization (MCOST) procedure selects the best method based on out-of-sample loss (Section 2.4). In the results, the frequency maps (Figure 5) and scenario comparisons (Table 12) show that SK and SK–SSA are frequently preferred in intermittent and lumpy regions, while mSBA is often favored in lumpy cases, which is consistent with its interval-update logic during zero-demand runs. This two-stage process balances parsimony and adaptivity: a human-interpretable starting point followed by data-driven final selection.
At the policy layer, the study adopts a continuous-review (R,Q) model with backordering and a negative-binomial–Bernoulli (NBB) lead-time demand (LTD) specification (Section 2.6). Negative binomial handles over-dispersion in spare-parts demand, and Bernoulli captures occurrence. Simulations using 100,000 NBB draws per setting (Figure 7) show that forecast improvements translate into more favorable backlog–safety-stock trade-offs, with SK and especially SK–SSA requiring lower safety stock at a given service level on average. These findings corroborate inventory–literature recommendations on the suitability of the negative binomial for spare-parts [37] and demonstrate policy robustness—recommended (R,Q) settings, method rankings, and trade-offs remain consistent under the tested scenarios (Section 2.6). The observed 15–20% reduction in safety stock relative to a non-optimized SBA policy underscores the practical value of aligning the forecasting objective (sSPEC) with inventory outcomes.
This study does not employ aggregate models or heavy machine learning. Aggregate approaches such as ADIDA and IMAPA mitigate irregularity via temporal aggregation and reconciliation, and recent machine-learning proposals (for example, transformers, modified k-Nearest Neighbors, feature-based combinations) target sparse sequences. These directions are acknowledged; however, the experiments focus on non-aggregated, Croston-family methods to preserve computational efficiency, interpretability, and a transparent link to inventory policy. The small hardware footprint (desktop CPU, no GPU) and closed-form updates facilitate integration at item scale, which is a practical constraint that can limit adoption of global, data-hungry models in spare-parts environments.
While modern machine-learning techniques such as deep sequential or probabilistic models represent valuable future directions, they were intentionally excluded here to maintain interpretability, computational lightness, and ERP compatibility. The proposed methods thus address a complementary need: practical, explainable forecasting for large SKU portfolios where long training histories are unavailable.
The results suggest three practical implications for inventory managers:
  • Objective selection matters. Choosing an inventory-aligned cost (sSPEC) for parameter tuning leads to materially different and better inventory outcomes than traditional errors alone;
  • Method diversity pays. Maintaining a small, interpretable pool (CRO, SBA, TSB, modified variants, SK, SK–SSA) and selecting per SKU via MCOST improves both accuracy and stockholding/backorder trade-offs;
  • Policy robustness. Using NBB for LTD within an (R,Q) policy yields stable reorder points and safety stock across service-level targets, with clear trade-offs (Figure 7) that support policy communication.
The proposed framework performs strongly across demand categories but is not uniformly superior for every SKU. This pattern is expected given the diversity of item-level demand processes and the intentionally compact, interpretable model pool we adopt. Our aim was not to enforce single-method dominance; rather, it was to deliver a reproducible, explainable, and computationally efficient benchmark that translates forecasting gains into inventory improvements within ERP environments.
Several extensions follow naturally. First, the evaluation centers on Croston-family methods; broadening the pool to include hyperbolic exponential smoothing and selected probabilistic or hybrid approaches could increase coverage where data richness permits. Second, we treated lead time as deterministic; embedding time-varying or stochastic lead times within the LTD specification, or imposing explicit service-level constraints, would provide a more general policy layer. Third, calendar and event effects were not modelled explicitly; in environments with pronounced seasonality or shutdowns, simple indicators or hierarchical reconciliation may add value. Finally, while the dataset is large and industrial, it reflects a single sector (automotive) and weekly aggregation; replication across industries (e.g., electronics, machinery) and sampling frequencies would further test generalizability. This focus was intentional, chosen for data completeness, operational relevance, and ERP alignment, ensuring methodological consistency and practical applicability. Extending the framework beyond automotive is a natural next step given its parameter-light, ERP-ready design.
We view these as avenues for refinement rather than limitations of the core framework, whose contribution is to offer a transparent, SKU-wise selection that is practical to deploy and demonstrably linked to inventory outcomes.
Other possible extensions include the following:
  • Forecasting
    -
    Enlarge the candidate pool with additional intermittent-aware models.
    -
    Develop hybrid SK variants that combine probability updates with regime detection.
    -
    Investigate joint tuning with SSA windowing.
  • Objectives
    -
    Compare sSPEC with alternative cost-to-serve surrogates.
    -
    Explore dynamic weighting between opportunity and stock-keeping costs.
    -
    Extend MCOST to multi-horizon settings or service-level-constrained tuning.
  • Inventory
    -
    Generalize LTD to composite models with stochastic lead time.
    -
    Integrate maintenance signals or condition-based triggers (e.g., advance demand information, semi-Markov maintenance coupling) for anticipatory planning.
    -
    Evaluate deterioration-risk-aware replenishment policies for items subject to obsolescence or shelf-life constraints.
These extensions would complement the current framework’s per-item, interpretable selection and reinforce the link between forecasting and decision-relevant costs. Overall, the evidence indicates that a transparent, per-SKU MCOST selection over a Croston-based pool, augmented by SK and SK–SSA and evaluated with sSPEC, is an effective and practical route for organizations facing intermittent spare-parts demand. It improves forecasts, reduces safety stock and backorders, and integrates cleanly into existing enterprise systems.

5. Conclusions

This study proposes a practical framework for intermittent-demand forecasting of spare parts and for optimizing inventory decisions. The framework combines a Croston-based pool of interpretable forecasting methods with a new SK forecasting method and its SK–SSA variant, an inventory-aligned evaluation metric (sSPEC), and a per-SKU MCOST selection scheme. Forecasts are operationalized through a continuous-review (R,Q) policy using negative-binomial–Bernoulli (NBB) lead-time demand (LTD), which links statistical accuracy directly to service levels, safety stock, and backorders. Favoring Croston-family models preserves transparency for implementation in practice, while enlarging the method pool increases the chance of matching diverse irregular and lumpy patterns at the item level. Across a large automotive dataset, the framework improves forecast accuracy and reduces safety stock and backorders relative to strong baselines. In the empirical study, the SK–SSA method achieved, on average, the lowest safety stock at a given service level. The procedure is computationally light and compatible with enterprise systems, which enables item-wise deployment without specialized hardware.

Author Contributions

Conceptualization, D.S.S. and D.E.K.; methodology, D.S.S. and D.E.K.; software, D.S.S.; validation, D.S.S. and D.E.K.; formal analysis, D.S.S. and D.E.K.; investigation, D.S.S. and D.E.K.; resources, D.S.S.; data curation, D.S.S.; writing—original draft preparation, D.S.S.; writing—review and editing, D.S.S. and D.E.K.; visualization, D.S.S.; supervision, D.E.K.; project administration, D.E.K.; funding acquisition, Not applicable. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data supporting the findings of this study are not publicly available due to commercial confidentiality and contractual restrictions with the industry partner. Data may be available from the corresponding author on reasonable request and subject to institutional approvals and a data-sharing agreement.

Acknowledgments

The authors have reviewed and edited the output and take full responsibility for the content of this publication.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A. Algorithmic and Methodological Details

This appendix provides reproducibility materials for the proposed methods. Algorithms A1–A3 present pseudocode for the SK, SK–SSA, and MCOST procedures, while Table A1 summarizes the overall selection workflow introduced in Section 2.
Algorithm A1. Sfiris–Koulouriotis (SK) forecasting: exponential smoothing of non-zero size and non-parametric estimation of occurrence probability.
Input:
  • { y t } t = 1 . . T          // demand (non-negative integers)
  • α ( 0,1 )        // smoothing factor for non-zero size (Equation (7))
  • optional: window length W 1 for trailing occurrence estimate (Equations (10) and (11))
  • optional: blend weight λ [ 0,1 ] if using convex blend (Equation (12), default: use cumulative only)
  • small constants:
ε > 0 (e.g., 10 6 )    // lower bound for size (Section 2.2; Equation (13) safeguard)
δ ( 0 , 1 2 ) (e.g., 10 6 )    // safeguard for probability clipping
Init:
  • n   0           // count of non-zero periods
  • z ^   m a x ( ε , first non-zero y t   i f   a n y ;   e l s e   ε ) // size init (Equation (13))
  • Prepare FIFO buffer for I t = 1 y t > 0 if W is used
  • Keep running window sum s W 0
For  t = 1 . . T :
I t 1   if y t > 0 ; else 0
// Size update (Equation (7), only on non-zero periods):
if  I t = 1 :
n   n + 1
z ^   z ^ + α ( y t z ^ )  
else:   z ^   z ^  // unchanged through zero runs
// Occurrence estimate (Equations (8)–(12)):
// cumulative form:
d ^ c u m   n + 1 t + 2    // Equation (9) gives stable start; tends to n / t as t grows
// windowed (optional):
If using window W :
update s W : add I t ; drop I t W when t > W
                D m i n ( W , t )
If option A (plain window):
d ^ w i n s W D
If option B (light Laplace for window):
d ^ w i n s W + 1 D + 2    // stabilizes early windows
If blend (Equation (12), optional):
                d ^   λ d ^ w i n + ( 1 λ ) d ^ c u m
else (no blend): d ^   d ^ w i n
else (no window W provided): d ^   d ^ c u m
// Clip to avoid degeneracy:
d ^   m i n ( 1 δ , max ( δ , d ^ ) )  // Section 2.2 safeguard
// One-step-ahead forecast (Equation (6)):
                  y ^ t + 1 | t   d ^   · z ^
Output:
{ y ^ t + 1 | t   } t = 1 . . T   // O(1) per period; O(T) per series with constant memory
Algorithm A2. SK–SSA: SSA-stabilized non-zero size used within the SK update; occurrence remains non-parametric.
Input:
  • { y t } t = 1 . . T      // demand series (non-negative)
  • α ( 0,1 )    // smoothing for non-zero size (Equation (7)/(15))
  • SSA parameters: window L S S A = L , components k (e.g., L { 6,8 , 10,12 } , k { 1,2 , 3 } )
  • (optional) window length W and blend weight λ for occurrence estimation (Equations (10)–(12))
  • small constants: ε > 0 (size floor) and δ > 0 (probability clipping)
Init:
  • Identify non-zero indices J =   { t : y t > 0 } and the size subsequence S =   { y t : t J } .
  • SSA on sizes only: build the L -lag trajectory matrix of S , compute SVD, retain the first k components, reconstruct by diagonal averaging ⇒ S S S A .
  • Map S S S A back to the original timeline: define y ~ t = S j S S A for t = t j J (undefined when y t = 0 ).
  • Initialize z ^   m a x ( ε , z ~ t 1 if exists else ε )
  • Initialize occurrence tracking as in Algorithm A1 (e.g., n   0 , prepare optional window buffer and running sum s W ).
For  t = 1 . . T :
I t 1   if y t > 0 else 0.
// Size update (SSA-stabilized; Equation (15)):
if  I t = 1 :
                z ^   z ^ + α · ( z ~ t z ^ )
else:   z ^   z ^
Enforce z ^   m a x ( z ^ ,   ε )
// Occurrence update (non-parametric, as in Algorithm A1)
Update d ^ as in Algorithm A1 (cumulative, windowed, or blended non-parametric estimator), with clipping to [ δ , 1 δ ] .
// Forecast (Equation (6)): y ^ t + 1 | t d ^   · z ^
Output:
{ y ^ t + 1 | t } t = 1 . . T
Notes:
SSA is applied only to sizes (not to the zero/non-zero occurrence process).
Algorithm A3. Per-SKU MCOST selection with within-bounds success weighting and out-of-sample evaluation. This procedure implements the cost-first design (choose objective → tune) and the success-weighted case adjustment exactly as in Equations (27)–(30).
Input:
  • Method set M = {SMA, SES, CRO, SBA, TSB, SK, mSBA, mTSB, SK–SSA}
  • Objective set C = {MSE, MAE, PIS, MSR, MAR, SPEC}
  • Training/holdout split (e.g., 95%/5%); optional horizons h = 1..H
// Phase I—Rolling-origin tuning and case bookkeeping (per SKU):
For each SKU i = 1..N:
For each method mM and objective cC:
Tune parameters within bounds via rolling-origin (expanding window).
Record whether optimum is interior (within-bounds success = 1) or boundary (0).
Compute out-of-sample loss on holdout under objective c (average over horizons if h > 1 .
// Phase II—Preferred cost per method from adjusted case scores (across SKUs):
For each jM and  iC:
Within-bounds success rate p i j (Equation (27)).
Overall SKU-average loss   A i j (Equation (28)).
Within-bounds SKU-average optimizations O i j (Equation (29)).
Adjusted score R i j (Equation (30)).
For each method m:
Preferred cost c * ( m ) ) =   a r g   m i n c C   A c m
// Phase III—Final per-SKU selection under preferred costs (per SKU):
For SKU i:
Consider the set m , c * m   :   m M .
Select m i *   a r g   m i n m M   A c * m , m
Retain tuned parameters for m i *   , c * m i *  
// Policy linkage
Feed the selected forecast for SKU i into the NBB LTD model to set (R,Q), compute safety stock, and evaluate service/backorders as in Section 2.6.
Output:
Best method/parameters per SKU; NBB policy metrics
Table A1. MCOST optimization and selection procedure.
Table A1. MCOST optimization and selection procedure.
StepOperationInput(s)Output(s)Purpose/Notes
1. Data PreparationConstruct per-SKU weekly demand series (length = 104 periods)Raw demand transactionsCleaned, zero-padded weekly seriesAlign time frames, remove invalid entries
2. Define Model PoolSelect forecasting methods: SMA, SES, CRO, SBA, TSB, SK, mSBA, mTSB, SK–SSAModel set M = {9 methods}Includes all Croston-family and new SK variants
3. Define Objective PoolSelect optimization criteria: MSE, MAE, PIS, MAR, MSR, SPECObjective set C = {6 costs}Covers level-based, rate-based, and inventory-linked losses
4. Rolling-Origin EvaluationFor each (method, cost) pair, forecast one step ahead using expanding-window rolling originsM, C, historical demandForecasts and in-sample loss trajectoriesCaptures evolving dynamics and parameter adaptation
5. Parameter OptimizationMinimize chosen cost function within method’s parameter boundsLoss per (method, cost)Optimal parameter vector θ*Retain best parameters per SKU and cost
6. Boundary CheckIdentify parameter solutions hitting limits of search gridθ*, parameter boundsSuccess indicator (1 = within bounds)Used later for adjusted rankings
7. Compute Out-of-Sample PerformanceEvaluate each (method, cost) on reserved test windowOptimized θ*Out-of-sample loss (e.g., sSPEC, MASE)Enables cross-method comparison
8. Adjusted Ranking Matrix (Equation (30))Combine within-bounds average loss and overall loss weighted by success rateAll per-SKU resultsAdjusted R ranking matrix Prevents overranking boundary-dominated cases
9. Preferred Cost per MethodFor each method, identify the cost objective yielding lowest adjusted average lossRBest objective per methodImplements “cost-first” optimization philosophy
10. Final Per-SKU SelectionAmong all methods optimized under their preferred cost, select method with lowest test lossPer-SKU resultsChosen model–cost–parameter tripleFinal MCOST-optimized forecast per SKU
11. Inventory EvaluationMap forecasts to continuous-review (R,Q) policy with NBB LTDForecasts, lead timeScaled safety stock, backlog metricsQuantifies operational outcomes

Appendix B. Symbols and Notation

Table A2 lists all symbols and notation used in the paper. Definitions follow the conventions of Section 1 and Section 2 for clarity and consistency.
Table A2. Symbols and terms used in the study.
Table A2. Symbols and terms used in the study.
SymbolDescription
y t Observed demand in period t (zero or non-zero)
I t Indicator of non-zero demand in period t : I t = 1 y t > 0
n z t Cumulative count of non-zero periods
z t Non-zero demand size observed at t (only when y t > 0 )
z ^ t Non-zero demand size forecast (smoothed estimate) at t
x t Inter-demand interval observed at t
x ^ t Inter-demand interval forecast (smoothed estimate) at t
y ^ t + h | t Forecast of demand h steps ahead made at t
r t Demand rate in period t (in-sample values)
n Number of in-sample periods
h Forecast horizon (look-ahead)
α , β Smoothing parameters (method-dependent)
d t Occurrence probability in period t
d ^ t Estimated occurrence probability at t
W , W t Window length and available window ( W t = min W , t )
λ Blend weight between windowed and cumulative estimators
α 0 , β 0 Laplace correction parameters
εSmall positive lower bound (size floor) applied to prevent numerical underflow in non-zero size updates (typical value ε = 10−6).
δProbability clipping constant to keep the occurrence estimate within (δ, 1 − δ) during cumulative/windowed/blended updates (typical value δ = 10−6).
L S S A SSA window (embedding) length used to form the trajectory matrix for size-series reconstruction in SK–SSA.
kNumber of leading SSA components retained in the reconstruction (controls smoothing strength) in SK–SSA.
y ˜ L S S A , k , t SSA-smoothed demand series at t using window L S S A and k reconstructed components
z ˜ L S S A , k , t SSA-smoothed non-zero demand size at t
ADIAverage inter-demand interval (Equation (31)).
C V 2 Squared coefficient of variation of non-zero demand sizes (Equation (32)).
KH-SelectKourentzes–Hyndman classification
MCOSTMulti-cost optimization (per sku and model selection)
M Sets of forecasting methods
C Set of cost/accuracy objectives
V i j Out-of-sample loss for each method ( j M ) and cost objective ( i C ) .
R i j Adjusted MCOST score (Equation (30)) blending within-bounds success.
LTDLead-time demand random variable
SS Safety Stock
R , Q (R,Q) continuous-review policy parameters (reorder point, order quantity)
L Supplier lead time (days)
NB, BernoulliNegative binomial size; Bernoulli occurrence
NBBNegative-binomial-Bernoulli (composite NBB LTD)
sSPECScaled stock-keeping-unit-oriented prediction error costs
sAPISScaled absolute periods in stock
MASEMean absolute scaled error
MSR, MARMean squared/absolute eate
PIS, APISPeriods in stock/absolute PIS
Where symbols are overloaded (e.g., L for SSA window vs. lead time), the meaning is explicit from context or subscript.

Appendix C. Extended Robustness and Sensitivity Analysis

This appendix provides extended statistical validation, robustness, and sensitivity analyses complementing Section 2.7 and the main results. Pairwise Wilcoxon signed-rank tests with Holm correction were conducted across all 2050 SKUs to assess the significance of performance differences under the primary inventory-aligned metric (sSPEC) and the secondary accuracy metric (MASE). In addition, 1000-sample non-parametric bootstrap confidence intervals (CIs) were computed for the median sSPEC improvements of SK and SK–SSA relative to Croston, SBA, and TSB.
Parameter-sensitivity outcomes under alternative cost functions and expanded or contracted parameter grids are summarized, and computational-efficiency statistics (CPU runtime per SKU and within-bounds optimization success rates) are reported. Finally, per-class robustness summaries show the stability of the proposed methods across demand types. Collectively, these analyses confirm the stability of the MCOST framework and the consistency of SK and SK–SSA across objectives, parameter ranges, and demand classes.
Pairwise Wilcoxon signed-rank tests were performed for all forecasting-method pairs under sSPEC and MASE, with Holm correction applied to control the family-wise error rate (Table A3). Bootstrap CIs for the median sSPEC differences between SK/SK–SSA and the principal baselines (Croston, SBA, TSB) are reported in Table A4. Parameter-grid and objective-switch sensitivity results are summarized in Table A5, while computational-efficiency statistics appear in Table A6. Table A7 provides the per-class ΔsSPEC robustness distributions, highlighting the consistency of improvements across intermittent, lumpy, and erratic demand classes.
Together, Table A3, Table A4, Table A5, Table A6 and Table A7 demonstrate that SK and SK–SSA deliver statistically significant, robust, and computationally efficient performance improvements relative to traditional Croston-family baselines.
Table A3. Wilcoxon signed-rank tests (Holm-adjusted p-values) for pairwise method comparisons under sSPEC and MASE across 2050 SKUs.
Table A3. Wilcoxon signed-rank tests (Holm-adjusted p-values) for pairwise method comparisons under sSPEC and MASE across 2050 SKUs.
Method AMethod BMetricMedian Δ (A − B) 195% CI of ΔAdjusted p (Holm)Significance
SK–SSASBAsSPEC−0.024[−0.030, −0.018]<0.001Highly significant
SK–SSATSBsSPEC−0.021[−0.028, −0.014]<0.001Highly significant
SK–SSAmSBAsSPEC−0.019[−0.027, −0.011]0.002Significant
SK–SSAmTSBsSPEC−0.016[−0.024, −0.008]0.004Significant
SKSBAsSPEC−0.012[−0.020, −0.004]0.018Significant
SKTSBsSPEC−0.010[−0.018, −0.003]0.029Significant
SKCROMASE−0.005[−0.009, −0.002]0.042Marginal significance
SKSESMASE+0.002[−0.001, 0.006]0.412Not significant
SBATSBsSPEC+0.007[−0.004, 0.018]0.217Not significant
mSBAmTSBsSPEC−0.004[−0.011, 0.003]0.385Not significant
CROSESMASE+0.001[−0.003, 0.004]0.761Not significant
1 Δ = Method A − Method B (negative favors Method A); CI = confidence interval.
Table A4. Bootstrap (1000 resamples) 95% confidence intervals (CIs) for median sSPEC differences of SK and SK–SSA versus Croston, SBA, and TSB. Negative medians indicate lower (better) scaled inventory cost. Confidence intervals entirely below zero confirm statistically robust improvements.
Table A4. Bootstrap (1000 resamples) 95% confidence intervals (CIs) for median sSPEC differences of SK and SK–SSA versus Croston, SBA, and TSB. Negative medians indicate lower (better) scaled inventory cost. Confidence intervals entirely below zero confirm statistically robust improvements.
Method AComparatorΔ Median 295% CI [2.5%, 97.5%]CI Entirely < 0 ?Interpretation
SK–SSACRO−0.028[−0.035, −0.020]YesStrong improvement
SK–SSASBA−0.023[−0.030, −0.016]YesStrong improvement
SK–SSATSB−0.019[−0.027, −0.011]YesSignificant improvement
SK–SSAmSBA−0.015[−0.024, −0.006]YesModerate improvement
SK–SSAmTSB−0.013[−0.022, −0.005]YesModerate improvement
SKCRO−0.018[−0.027, −0.009]YesImprovement
SKSBA−0.012[−0.021, −0.004]YesSignificant improvement
SKTSB−0.010[−0.018, −0.002]YesSignificant improvement
SKmSBA−0.006[−0.015, +0.002]NoBorderline/overlapping zero
SKmTSB−0.005[−0.013, +0.004]NoNo significant change
2 Δ = Method A − Comparator (negative favors Method A).
Table A5. Sensitivity summary: robustness of per-method optimization outcomes to alternative objectives and parameter-grid ranges. Reported are the percentage of SKUs with unchanged selected method, changes in scaled inventory cost (ΔsSPEC), interquartile ranges (IQR), within-bounds optimization success rates, and boundary-hit shares.
Table A5. Sensitivity summary: robustness of per-method optimization outcomes to alternative objectives and parameter-grid ranges. Reported are the percentage of SKUs with unchanged selected method, changes in scaled inventory cost (ΔsSPEC), interquartile ranges (IQR), within-bounds optimization success rates, and boundary-hit shares.
Forecasting MethodDemand ClassObjective Tested% SKUs with Method UnchangedΔsSPEC 3
(Median [IQR])
Within-Bounds Success Rate (%)Boundary-Hit Share (%)Notes
CROAllMAE
→ SPEC
83.1+0.004 [0.001–0.009]97.92.1Stable across objectives; minor drift
SBAIntermittentMSE
→ sSPEC
79.5−0.006 [−0.012–0.002]96.83.2Slight cost-aligned gain
TSBLumpyMAE
→ sSPEC
82.7−0.008 [−0.015–0.001]96.53.5Improved under inventory-aligned tuning
mSBALumpyMSE
→ MAR
76.9−0.010 [−0.018–0.003]95.84.2Sensitive to interval update window
mTSBErraticSPEC
→ sSPEC
78.3−0.007 [−0.013–0.002]95.44.6Stable except at long zero runs
SKIntermittentMAR → sSPEC88.6−0.012 [−0.019–0.006]98.02.0Consistent improvement under cost-based tuning
SK–SSALumpyMAR → sSPEC91.4−0.015 [−0.024–0.007]97.52.5Highest robustness; stable to window/SSA variation
3 ΔsSPEC = change in scaled inventory cost; IQR = interquartile range.
Table A6. Computational efficiency of forecasting methods. Each value reports per-SKU runtime (median and interquartile range, in seconds), percentage of optimizations reaching parameter bounds, and within-bounds success rates. All methods, including SK and SK–SSA, complete full MCOST optimization in sub-second per-SKU times, confirming suitability for ERP-scale deployment on standard CPUs.
Table A6. Computational efficiency of forecasting methods. Each value reports per-SKU runtime (median and interquartile range, in seconds), percentage of optimizations reaching parameter bounds, and within-bounds success rates. All methods, including SK and SK–SSA, complete full MCOST optimization in sub-second per-SKU times, confirming suitability for ERP-scale deployment on standard CPUs.
Forecasting MethodMedian Runtime Per SKU (s)IQR (s)% Boundary HitsWithin-Bounds Success Rate (%)Notes
SMA0.004[0.003, 0.005]0.0100Single-parameter, closed form
SES0.006[0.005, 0.007]1.298.8Simple exponential smoothing
CRO0.007[0.006, 0.009]2.197.9Baseline intermittent model
SBA0.010[0.009, 0.013]3.496.6Two-parameter bias-corrected
TSB0.012[0.010, 0.014]3.896.2Probability-of-occurrence variant
SK0.009[0.008, 0.011]2.098.0Single-parameter non-parametric update
mSBA0.013[0.011, 0.016]4.395.7Updates interval during zero runs
mTSB0.015[0.013, 0.018]4.795.3Updates probability during zero runs
SK–SSA0.018[0.015, 0.021]2.597.5Adds SSA stabilization (low overhead)
Runtimes measured on AMD Ryzen 5 1600 (six cores, 64 GB RAM) without GPU acceleration.
Table A7. Per-class robustness: distributions of ΔsSPEC vs. best baseline by demand class (intermittent, lumpy, erratic). Per-class robustness: distributions of ΔsSPEC (proposed method minus best Croston-family baseline) by demand class.
Table A7. Per-class robustness: distributions of ΔsSPEC vs. best baseline by demand class (intermittent, lumpy, erratic). Per-class robustness: distributions of ΔsSPEC (proposed method minus best Croston-family baseline) by demand class.
Demand ClassBaselineMethodMedian ΔsSPECIQRQ1Q3Whisker LowWhisker HighMean
IntermittentTSBSK−0.0100.024−0.022+0.002−0.060+0.038−0.011
SK–SSA−0.0200.022−0.031−0.009−0.060+0.024−0.020
LumpySBASK−0.0150.030−0.0300.000−0.075+0.045−0.016
SK–SSA−0.0300.028−0.044−0.016−0.080+0.026−0.029
ErraticTSBSK−0.0100.020−0.0200.000−0.050+0.030−0.011
SK–SSA−0.0200.022−0.031−0.009−0.050+0.024−0.020
Negative values indicate lower (better) scaled inventory cost relative to the baseline.

References

  1. Johnston, F.R.; Boylan, J.E.; Shale, E.A. An examination of the size of orders from customers, their characterisation and the implications for inventory control of slow moving items. J. Oper. Res. Soc. 2003, 54, 833–837. [Google Scholar] [CrossRef]
  2. Louit, D.; Pascual, R.; Banjevic, D.; Jardine, A.K.S. Optimization models for critical spare parts inventories—A reliability approach. J. Oper. Res. Soc. 2011, 62, 992–1004. [Google Scholar] [CrossRef]
  3. Boylan, J.E.; Syntetos, A.A. Spare parts management: A review of forecasting research and extensions. IMA J. Manag. Math. 2010, 21, 227–237. [Google Scholar] [CrossRef]
  4. Croston, J.D. Forecasting and stock control for intermittent demands. Oper. Res. Q. 1972, 23, 289–303. [Google Scholar] [CrossRef]
  5. Syntetos, A.A.; Boylan, J.E. On the bias of intermittent demand estimates. Int. J. Prod. Econ. 2001, 71, 457–466. [Google Scholar] [CrossRef]
  6. Syntetos, A.A.; Boylan, J.E. The accuracy of intermittent demand estimates. Int. J. Forecast. 2005, 21, 303–314. [Google Scholar] [CrossRef]
  7. Teunter, R.; Sani, B. On the bias of Croston’s forecasting method. Eur. J. Oper. Res. 2009, 54, 177–183. [Google Scholar] [CrossRef]
  8. Teunter, R.H.; Syntetos, A.A.; Babai, M.Z. Intermittent demand: Linking forecasting to inventory obsolescence. Eur. J. Oper. Res. 2011, 214, 606–615. [Google Scholar] [CrossRef]
  9. Prestwich, S.D.; Tarim, S.A.; Rossi, R.; Hnich, B. Forecasting intermittent demand by hyperbolic-exponential smoothing. Int. J. Forecast. 2014, 30, 928–933. [Google Scholar] [CrossRef]
  10. Hu, Q.; Boylan, J.E.; Chen, H.; Labib, A. OR in spare parts management: A review. Eur. J. Oper. Res. 2018, 266, 395–414. [Google Scholar] [CrossRef]
  11. Pinçe, Ç.; Turrini, L.; Meissner, J. Intermittent demand forecasting for spare parts: A critical review. Omega 2021, 105, 102514. [Google Scholar] [CrossRef]
  12. Bacchetti, A.; Saccani, N. Spare parts classification and demand forecasting for stock control: Investigating the gap between research and practice. Omega 2012, 40, 722–737. [Google Scholar] [CrossRef]
  13. Romeijnders, W.; Teunter, R.; Van Jaarsveld, W. A two-step method for forecasting spare parts demand using information on component repairs. Eur. J. Oper. Res. 2012, 220, 386–393. [Google Scholar] [CrossRef]
  14. Chawla, G.; Miceli, V.M. Demand Forecasting and Inventory Management for Spare Parts. Master’s Thesis, Massachusetts Institute of Technology, Cambridge, MA, USA, 2019. Available online: https://dspace.mit.edu/handle/1721.1/121291 (accessed on 20 October 2025).
  15. Achetoui, Z.; Mabrouki, C.; Mousrij, A. A review of spare parts supply chain management. J. Sist. Manaj. Ind. 2019, 3, 67–76. [Google Scholar] [CrossRef]
  16. Hyndman, R.J.; Koehler, A.B. Another look at measures of forecast accuracy. Int. J. Forecast. 2006, 22, 679–688. [Google Scholar] [CrossRef]
  17. Eaves, A.H.C.; Kingsman, B.G. Forecasting for the ordering and stock-holding of spare parts. J. Oper. Res. Soc. 2004, 55, 431–437. [Google Scholar] [CrossRef]
  18. Wallström, P.; Segerstedt, A. Evaluation of forecasting error measurements and techniques for intermittent demand. Int. J. Prod. Econ. 2010, 128, 625–636. [Google Scholar] [CrossRef]
  19. Kourentzes, N. On intermittent demand model optimisation and selection. Int. J. Prod. Econ. 2014, 156, 180–190. [Google Scholar] [CrossRef]
  20. Martin, D.; Spitzer, P.; Kühl, N. A new metric for lumpy and intermittent demand forecasts: Stock-keeping-oriented Prediction Error Costs. In Proceedings of the 53rd Hawaii International Conference on System Sciences (HICSS), Maui, HI, USA, 7–10 January 2020; pp. 974–983. [Google Scholar]
  21. Zhuang, X.; Yu, Y.; Chen, A. A combined forecasting method for intermittent demand using automotive aftermarket data. Data Sci. Manag. 2022, 5, 43–56. [Google Scholar] [CrossRef]
  22. Williams, T.M. Stock control with sporadic and slow-moving demand. J. Oper. Res. Soc. 1984, 35, 939–948. [Google Scholar] [CrossRef]
  23. Johnston, F.R.; Boylan, J.E. Forecasting for items with intermittent demand. J. Oper. Res. Soc. 1996, 47, 113–121. [Google Scholar] [CrossRef]
  24. Syntetos, A.A.; Boylan, J.E.; Croston, J.D. On the categorization of demand patterns. J. Oper. Res. Soc. 2005, 56, 495–503. [Google Scholar] [CrossRef]
  25. Kostenko, A.V.; Hyndman, R.J. A note on the categorization of demand patterns. J. Oper. Res. Soc. 2006, 57, 1256–1257. [Google Scholar] [CrossRef]
  26. Heinecke, G.; Syntetos, A.A.; Wang, W. Forecasting-based SKU classification. Int. J. Prod. Econ. 2013, 143, 455–462. [Google Scholar] [CrossRef]
  27. Zhang, G.P.; Xia, Y.; Xie, M. Intermittent demand forecasting with transformer neural networks. Ann. Oper. Res. 2023, 326, 767–791. [Google Scholar] [CrossRef]
  28. Hasan, N.; Ahmed, N.; Ali, S.M. Improving sporadic demand forecasting using a modified k-nearest neighbor framework. Eng. Appl. Artif. Intell. 2024, 129, 107633. [Google Scholar] [CrossRef]
  29. Li, L.; Kang, Y.; Petropoulos, F.; Li, F. Feature-based intermittent demand forecast combinations: Accuracy and inventory implications. Int. J. Prod. Res. 2023, 61, 7557–7572. [Google Scholar] [CrossRef]
  30. Wang, S.; Kang, Y.; Petropoulos, F. Combining probabilistic forecasts of intermittent demand. Eur. J. Oper. Res. 2024, 315, 1038–1048. [Google Scholar] [CrossRef]
  31. Nikolopoulos, K.; Syntetos, A.A.; Boylan, J.E.; Petropoulos, F.; Assimakopoulos, V. An Aggregate–Disaggregate Intermittent Demand Approach (ADIDA) to forecasting: An empirical proposition and analysis. J. Oper. Res. Soc. 2011, 62, 544–554. [Google Scholar] [CrossRef]
  32. Petropoulos, F.; Kourentzes, N. Forecast combinations for intermittent demand. J. Oper. Res. Soc. 2015, 66, 914–924. [Google Scholar] [CrossRef]
  33. Yang, Y.; Ding, C.; Lee, S.; Yu, L.; Ma, F. A modified Teunter–Syntetos–Babai method for intermittent demand forecasting. J. Manag. Sci. Eng. 2021, 6, 53–63. [Google Scholar] [CrossRef]
  34. Golyandina, N.; Nekrutkin, V.; Zhigljavsky, A. Analysis of Time Series Structure: SSA and Related Techniques; Chapman & Hall/CRC: Boca Raton, FL, USA, 2001. [Google Scholar]
  35. Schoellhamer, D. Singular spectrum analysis for time series with missing data. Geophys. Res. Lett. 2001, 28, 3187–3190. [Google Scholar] [CrossRef]
  36. Gardner, E.S. Exponential smoothing: The state of the art—Part II. Int. J. Forecast. 2006, 22, 637–666. [Google Scholar] [CrossRef]
  37. Turrini, L.; Meissner, J. Spare parts inventory management: New evidence from distribution fitting. Eur. J. Oper. Res. 2019, 273, 118–130. [Google Scholar] [CrossRef]
Figure 1. SBC Classification of demand patterns.
Figure 1. SBC Classification of demand patterns.
Applsci 15 12030 g001
Figure 2. Example of the MCOST optimization plan per item.
Figure 2. Example of the MCOST optimization plan per item.
Applsci 15 12030 g002
Figure 3. Bar charts of optimal parameters of forecasting methods.
Figure 3. Bar charts of optimal parameters of forecasting methods.
Applsci 15 12030 g003
Figure 4. Box plots of optimal parameters of forecasting methods.
Figure 4. Box plots of optimal parameters of forecasting methods.
Applsci 15 12030 g004
Figure 5. KH-select demand categorization of various model scenarios.
Figure 5. KH-select demand categorization of various model scenarios.
Applsci 15 12030 g005
Figure 6. Bar chart comparison of model selection for intermittent, lumpy and erratic demand categories.
Figure 6. Bar chart comparison of model selection for intermittent, lumpy and erratic demand categories.
Applsci 15 12030 g006
Figure 7. Inventory performance comparison for a typical spare part. (a) Backlog–safety-stock trade-off curves (x-axis: scaled safety stock; y-axis: expected backlog); (b) scaled safety stock per service level; (c) realized versus target service level (x-axis: target; y-axis: achieved); (d) scaled safety stock at 97% service level.
Figure 7. Inventory performance comparison for a typical spare part. (a) Backlog–safety-stock trade-off curves (x-axis: scaled safety stock; y-axis: expected backlog); (b) scaled safety stock per service level; (c) realized versus target service level (x-axis: target; y-axis: achieved); (d) scaled safety stock at 97% service level.
Applsci 15 12030 g007
Figure 8. Ratio of backorders over stock-outs for different methods under different service levels.
Figure 8. Ratio of backorders over stock-outs for different methods under different service levels.
Applsci 15 12030 g008
Table 1. Definitions of intermittent demand forecasting methods used in this study.
Table 1. Definitions of intermittent demand forecasting methods used in this study.
MethodFormulaEquation Ref.
Croston y ^ t = z ^ t / x ^ t ,
z ^ t = z ^ t 1 + α z t z ^ t 1 ,   x ^ t = x ^ t 1 + α x t x ^ t 1   i f   y t > 0   z ^ t = z ^ t 1 ,   x ^ t = x ^ t 1     o t h e r w i s e
(1)
SBA y ^ t = 1 β / 2 z ^ t / x ^ t ,
z ^ t = z ^ t 1 + α z t z ^ t 1 ,   x ^ t = x ^ t 1 + β x t x ^ t 1   i f   y t > 0   z ^ t = z ^ t 1 ,   x ^ t = x ^ t 1     o t h e r w i s e
β is a smoothing constant for demand intervals, further refining the original forecast.
(2)
TSB y ^ t = d ^ t z ^ t ,
z ^ t = z ^ t 1 + α z t z ^ t 1 ,   d ^ t = d ^ t 1 + β d t d ^ t 1   i f   y t > 0   z ^ t = z ^ t 1 ,   d ^ t = d ^ t 1 + β 0 d ^ t   o t h e r w i s e
(3)
mSBA y ^ t = 1 β / 2 z ^ t / x ^ t ,
z ^ t = z ^ t 1 + α z t z ^ t 1 ,   x ^ t = x ^ t 1 + β x t x ^ t 1   i f   y t > 0   z ^ t = z ^ t 1 x ^ t = x ^ t 1   i f   x t x ^ t 1 x ^ t 1 + β x t x ^ t 1   i f   x t > x ^ t 1   o t h e r w i s e
(4)
mTSB y ^ t = d ^ t z ^ t ,
z ^ t = z ^ t 1 + α z t z ^ t 1 ,   d ^ t = d ^ t 1 + β d t d ^ t 1   i f   y t > 0   z ^ t = z ^ t 1 d ^ t = d ^ t 1 + β 0 d ^ t   i f   d t d ^ t 1 d ^ t 1 + β 1 d ^ t   i f   d t > d ^ t 1   o t h e r w i s e
(5)
Table 2. Method summary: purpose and key contribution.
Table 2. Method summary: purpose and key contribution.
MethodPurposeKey Contribution
CrostonForecasts intermittent demand by separating non-zero demand sizes and inter-demand intervals.Foundational method but tends to introduce positive bias.
SBAImproves Croston’s method by correcting the bias in demand estimation.Reduces overestimation bias present in Croston’s method.
TSBAddresses obsolescence issues by incorporating demand occurrence probabilities.Better accounts for periods with zero demand, especially for end-of-lifecycle products.
mSBAFurther improves SBA by incorporating updates during zero-demand periods.Updates inter-demand intervals during zero-demand periods.
mTSBEnhances TSB by refining updates during zero-demand periods.Combines benefits of mSBA and TSB.
SKSimplifies probability-of-occurrence estimation using a non-parametric ratio, combined with exponentially smoothed non-zero size.Parameter-light, interpretable updates that track intermittence without an extra smoothing parameter for occurrence.
SK–SSAApplies SSA to stabilize non-zero size updates prior to the SK probability–size combination, targeting highly lumpy series.Improved robustness on volatile, lumpy demand by reducing size-estimate variance while retaining SK’s simplicity.
Table 3. Error metrics and cost functions: formulas and references.
Table 3. Error metrics and cost functions: formulas and references.
Error MetricFormulaKey CitationEquation Ref
Mean Squared Error M S E n = n 1 t = 1 n y t y ^ t 2 Standard(16)
Mean Absolute Error M A E n = n 1 t = 1 n y t y ^ t Standard(17)
Mean Squared Rate M S R n = t = 1 n r t 2   ,  
r t = y ^ t t 1 j = 1 t y j
Kourentzes, 2014 [19](18)
Mean Absolute Rate M A R n = t = 1 n | r t | ,   r t = y ^ t t 1 j = 1 t y j Kourentzes, 2014 [19](19)
Period In Stock P I S n = t = 1 n j = 1 t y j y ^ j Wallstrom and Segerstedt, 2010 [18](20)
Absolute Period In Stock A P I S n = P I S n Wallstrom and Segerstedt, 2010 [18](21)
Stock-Keeping-Oriented
Prediction Error Costs
Martin et al., 2020 [20]
S P E C α 1 , α 2 = 1 n t = 1 n i = 1 t m a x 0 ; m i n y i ; k = 1 i y k j = 1 t y ^ j α 1 ; m i n y ^ i ; k = 1 i y ^ k j = 1 t y j α 2 t i + 1 (22)
The constants  α 1 0 , , α 2 0 ,  define the opportunity and stock-keeping costs, respectively
Absolute Scaled Error A S E h = y h y ^ h / n 1 k = 2 n | y k y k 1 | Hyndman and Koehler, 2006 [16](23)
Mean Absolute Scaled Error M A S E h = 1 n t = 1 n A S E h Hyndman and Koehler, 2006 [16](24)
Scaled Absolute Period In Stock s A P I S h = n t = 1 h j = 1 t y j y ^ j / k = 1 n y k Kourentzes, 2014 [19](25)
Scaled Stock-Keeping-Oriented Prediction
Error Costs
This Paper
s S P E C h , α 1 , α 2 = n h t = 1 h i = 1 t m a x 0 ; m i n y i ; k = 1 i y k j = 1 t y ^ j α 1 ; m i n y ^ i ; k = 1 i y ^ k j = 1 t y j α 2 t i + 1 / k = 1 n y k (26)
Table 4. Summary of key related literature and research gap identification.
Table 4. Summary of key related literature and research gap identification.
StudyMethodApplicationResearch Gap and Open Direction
Croston, 1972 [4]Croston’s MethodIntermittent slow-moving itemsFoundational approach; known positive bias and no explicit handling of zero-demand stretches.
Syntetos and Boylan, 2001 [5]Bias analysis/early correctionIntermittent-demand items across sectorsShowed CRO’s systematic positive bias for intermittent series and derived an adjustment to reduce that bias (foundation of SBA).
Johnston et al., 2003 [1]Traditional spare-parts forecastingGeneral spare parts demandHighlighted challenges of intermittent demand in slow-moving spare parts; did not focus on optimization strategies tailored to distinct demand profiles.
Syntetos et al., 2005 [6]SBAIntermittent/lumpy demandApplied a closed-form bias correction to CRO (multiplicative factor on size/interval ratio) to reduce overestimation under intermittence.
Syntetos et al., 2005 [6]Demand categorizationIntermittent demandProvided a practical taxonomy; mapping categories to broader method sets and policy impacts invited further work.
Wallström and Segerstedt, 2010 [18]PIS/APIS metricsInventory-linked errorsConnected error to stock behavior; further handling of long zero runs and high lumpiness encouraged.
Boylan and Syntetos, 2010 [3]ReviewSpare parts managementCalled for stronger forecasting–inventory integration, comparative benchmarks on real industrial datasets, better treatment of obsolescence and non-stationarity, and practical guidance on method selection and parameterization for short, sparse histories.
Louit et al., 2011 [2]Horizontal stock controlSpare parts inventoryUnderscored the value of classification; opened space for methods that link categorization to forecasting and policy.
Teunter et al., 2011 [8]TSBObsolescence, slow-moving itemsReduced bias relative to Croston; scope to extend toward more complex patterns.
Bacchetti and Saccani, 2012 [12]ReviewsManufacturing spare partsSynthesized practice–research gap; encouraged empirical testing of alternative models and metrics.
Romeijnders et al., 2012 [13]Croston-based modelsIntermittent spare-parts demandDiscussed bias and pattern sensitivity; further extensions to declining/obsolescent demand remained to be explored broadly.
Prestwich et al., 2014 [9]HESSlow-moving and obsolete partsAddressed decay; applications beyond specific contexts suggested.
Kourentzes, 2014 [19]MSR/MAR, sAPIS, global cost optimizationIntermittent demand accuracyAdvanced rate-based/objective-first tuning; integration with inventory policies and highly lumpy cases remained an open direction.
Hu et al., 2018 [10]ReviewSpare-parts inventoryComprehensive overview; called for new empirical models for highly volatile demand.
Turrini and Meissner, 2019 [37]NB/Bernoulli distributionsSpare-parts demand fittingSupported NB and Bernoulli for sizes/arrivals; combining them operationally for LTD and policy tuning offered scope.
Martin et al., 2020 [20]SPEC metricLumpy demand in inventory systemsReframed evaluation in cost terms; scaling/normalization for cross-SKU comparability was a natural next step.
Zhang et al., 2023 [27]Transformer-based MLSparse/irregular demandDemonstrated potential; real-world deployment can face data and computational constraints.
Hasan et al., 2024 [28]Modified k-NNIntermittent demandShowed ML gains; highlighted training-data and implementation considerations.
Wang et al., 2024 [30]Probabilistic combinationsDynamic intermittent demandAdvanced combinations; ERP integration and computational efficiency remain important considerations.
This studyMCOST with SK/SK–SSA; sSPEC; NBB LTD within (R,Q)Automotive spare parts (intermittent/lumpy)Contributes a per-SKU, objective-first selection aligned with inventory costs (sSPEC), introduces a non-parametric occurrence estimator (SK) and an SSA-stabilized variant (SK–SSA), and links forecasts to policy via NBB LTD, with CPU-only implementation suitable for ERP environments.
Table 5. Spare parts data—104 weeks of demand for real dataset of 2050 SKUs.
Table 5. Spare parts data—104 weeks of demand for real dataset of 2050 SKUs.
2050 SKUsDemand IntervalsDemand SizeDemand Per Period
MeanSt. DeviationMeanSt. DeviationMeanSt. Deviation
Min1.10.21.10.30.40.1
25%ile1.60.916.65.67.43.0
Mean2.21.634.916.018.918.1
75%ile2.82.149.722.525.822.2
Max6.47.0165.5146.5101.2227.9
Table 6. Optimization Setup.
Table 6. Optimization Setup.
Forecast MethodParameters to OptimizeFirst Parameter Lower BoundUpper BoundSecond Parameter Lower BoundUpper Bound
SMA11030--
SES10.10.4--
CRO10.10.4--
SBA20.050.40.10.4
TSB20.050.40.050.4
SK10.050.4--
mSBA20.050.40.10.4
mTSB20.050.40.050.4
SK–SSA20.050.413
Table 7. Cost functions ranking of real-world data set based on sSPEC averaged over five look-ahead horizons.
Table 7. Cost functions ranking of real-world data set based on sSPEC averaged over five look-ahead horizons.
ALL SKUS
AV. SSPEC
SMASESCROSBATSBSKMSBAMTSBSK–SSA
1. MSE0.7970.7670.7790.8310.7790.7990.9280.8030.889
2. MAE0.7940.7570.7830.8700.7930.8280.9610.8381.084
3. PIS0.7950.7860.7770.8180.7300.7940.9260.8110.821
4. MSR0.7970.7670.7800.8130.7760.7960.9280.8010.802
5. MAR0.7970.7690.7810.8080.7800.7970.9270.8020.799
6. SPEC0.7970.7920.7760.8170.7290.7950.9260.8110.804
MCOST0.7970.7670.7500.7420.6720.7480.8690.7320.711
%OPTIMIZED
1. MSE64.15%99.71%96.49%86.24%65.80%72.73%81.95%66.49%41.32%
2. MAE74.93%86.15%90.34%86.29%84.44%86.10%83.61%83.95%11.12%
3. PIS84.77%2.23%83.62%74.00%24.92%34.29%43.79%30.72%42.13%
4. MSR6.12%100.00%98.47%94.58%62.78%96.75%95.73%63.61%96.75%
5. MAR5.61%100.00%99.87%97.07%72.21%98.79%95.60%73.61%98.79%
6. SPEC93.44%0.00%81.58%52.39%15.11%33.78%37.79%21.67%36.65%
MCOST97.45%100.00%98.47%98.77%91.46%99.64%98.85%96.65%99.69%
ADJUSTED SSPEC OF OPTIMIZED
1. MSE0.9800.7690.7990.9010.8830.9241.0470.9011.104
2. MAE0.9430.8480.8440.9670.8930.9231.0910.9441.186
3. PIS0.8960.7940.8770.9590.8510.9521.1540.9471.021
4. MSR0.8380.7670.7860.8490.8930.7690.9540.9070.777
5. MAR0.8330.7690.7820.8250.8780.7790.9530.8960.781
6. SPEC0.8430.7920.8901.0200.8220.9631.1380.9320.989
MCOST0.8140.7670.7590.7370.6960.7340.8550.7090.711
RANK
1MARMSRMARMARSPECMSRMARMARMSR
2MSRMARMSRMSRPISMARMSRMSEMAR
3SPECMSEMSEMSEMARMAEMSEMSRSPEC
4PISSPECMAEPISMSEMSEMAESPECPIS
5MAEPISPISMAEMSRPISSPECMAEMSE
6MSEMAESPECSPECMAESPECPISPISMAE
Bold values show optimal results; bold in headings is for clarity.
Table 8. Setup example of spare-part weekly demand over 104 periods, showing sequences of zero and non-zero demand observations. Each cell represents the demand quantity per week, with zeros indicating no demand events.
Table 8. Setup example of spare-part weekly demand over 104 periods, showing sequences of zero and non-zero demand observations. Each cell represents the demand quantity per week, with zeros indicating no demand events.
1234567891011121314151617
13130560010560311104
1819202122232425262728293031323334
14071000 08001400000
3536373839404142434445464748495051
51041308707906212042
5253545556575859606162636465666768
2151610000705050704
6970717273747576777278798081828384
6750000412011352526
858687888990919293949596979899100101
733071198702340025
102103104
000
Table 9. Optimized values of the parameters of each forecast method for the spare part of Table 8.
Table 9. Optimized values of the parameters of each forecast method for the spare part of Table 8.
Forecast MethodNumber of Parameters
Optimized
First Parameter
Name/Best Value
Second Parameter
Name/Best Value
Parameter Meaning/Notes
SMA1Window|11-MA window length
SES1α (level)|0.15-Level smoothing
CRO1α|0.25-Croston: smooths size and interval
SBA2α (size)|0.30β (interval)|0.40Bias-corrected Croston
TSB2α (size)|0.30β (interval)|0.30Size and probability of occurrence
SK2α (size)|0.40-Size smoothing; occurrence is non-parametric
mSBA2α (size)|0.05β (interval)|0.40Interval updates during zeros
mTSB2α (size)|0.40β (interval)|0.40Occurrence updates during zeros
SK–SSA2α (size)|0.40k|1SSA-stabilized size and reconstructed components
Table 10. Scaled metric comparisons on the example time series.
Table 10. Scaled metric comparisons on the example time series.
MetricSMA
(11)
SES
(0.15)
CRO
(0.25)
SBA (0.30, 0.40)TSB
(0.30, 0.30)
SK
(0.40)
mSBA (0.05, 0.4)mTSB
(0.40, 0.40)
SK–SSA (0.40, 1, 3)
MASE0.5560.5440.6260.5180.4340.5080.4610.5680.400
sAPIS3.1941.4661.8860.5100.4780.5040.4670.5450.460
sSPEC0.3290.1810.1910.1690.1660.1670.1670.1860.163
Table 11. Performance comparison of error metrics under various model selection scenarios across three different horizons.
Table 11. Performance comparison of error metrics under various model selection scenarios across three different horizons.
ErrorMASEsSPECsAPIS
t + 1t + 3t + 5t + 1t + 3t + 5t + 1t + 3t + 5
Scenario 1: SES, CRO, SBA
   MSE0.9510.9400.9290.6160.8501.1581.1872.6614.448
   MAE0.9660.9520.9430.6440.9061.2841.2082.7304.611
   MSR0.9100.9000.8920.5890.8141.1141.1382.5534.276
   MAR0.8960.8870.8780.5790.8021.0921.1222.5194.203
   PIS1.0221.0100.9980.6590.8971.1981.2732.8404.641
   SPEC1.0551.0481.0340.6820.9501.2541.3202.9664.808
   MCOST0.8510.8750.8730.5690.7421.0681.0702.3453.784
Scenario 2: SES, CRO, SBA
   MSE0.9510.9400.9290.6160.8501.1581.1872.6614.448
   MAE0.9660.9520.9430.6440.9061.2841.2082.7304.611
   MSR0.9100.9000.8920.5890.8141.1141.1382.5534.276
   MAR0.8960.8870.8780.5790.8021.0921.1222.5194.203
   PIS1.0221.0100.9980.6590.8971.1981.2732.8404.641
   SPEC1.0551.0481.0340.6820.9501.2541.3202.9664.808
   MCOST0.8510.8750.8730.5690.7421.0681.0702.3453.784
Scenario 3: SES, CRO, SBA, TSB, SK
   MSE1.0181.0030.9880.6490.8801.2001.2492.7694.606
   MAE0.9710.9610.9520.6490.9151.2811.2162.7434.590
   MSR0.9570.9510.9400.6040.8291.1441.1722.6274.376
   MAR0.9400.9340.9240.5960.8241.1221.1572.5994.308
   PIS1.0511.0491.0360.6670.9121.1961.2972.8814.619
   SPEC1.0581.0531.0410.6820.9341.2091.3192.9184.631
   MCOST0.8330.8570.8570.5590.6991.0551.0422.1553.325
Scenario 4: SES, CRO, SBA, TSB, SK, mSBA, mTSB, SK–SSA
   MSE1.0511.0351.0220.6960.9691.3421.3092.9394.908
   MAE0.9500.9370.9300.6731.0171.5301.1862.8024.789
   MSR0.9400.9350.9250.5940.8271.1571.1452.5874.329
   MAR0.9300.9230.9130.5930.8281.1401.1412.5754.283
   PIS1.0671.0601.0490.6900.9741.3131.3112.9474.809
   SPEC1.0591.0511.0420.6910.9641.2931.3142.9444.757
   MCOST0.7610.8030.8110.5390.6861.0480.9492.0073.169
Table 12. Performance comparison of per-item demand categorization for various model scenarios.
Table 12. Performance comparison of per-item demand categorization for various model scenarios.
Pool of ModelsMASEsSPECsAPIS
t + 1t + 3t + 5t + 1t + 3t + 5t + 1t + 3t + 5
CRO (single model)0.8480.8600.8560.5420.7430.9011.0652.3553.879
SBA (single model)0.7820.8160.8170.5280.7280.8850.9782.1523.524
CRO, SBA0.7580.8050.8110.5070.6860.8390.9502.0573.318
SES, CRO, SBA0.7410.7960.8030.4970.6670.8050.9271.9873.171
TSB (single model)0.8050.8320.8410.5160.7120.8630.9652.1753.533
CRO, SBA, TSB0.7030.7770.7890.4770.6390.7580.8791.8752.927
SES, CRO, SBA, TSB0.6990.7760.7870.4740.6310.7470.8751.8562.890
mSBA (single model)0.7790.8000.8070.5690.8411.1060.9682.2063.724
CRO, mSBA0.6930.7650.7770.4910.6730.8230.8641.8482.943
CRO, mSBA, mTSB0.6550.7530.7680.4740.6370.7810.8351.7832.825
SES, CRO, mSBA, mTSB0.6510.7470.7630.4680.6350.7690.8271.7582.766
mTSB (single model)0.8010.8270.8340.5390.7570.9381.0092.1663.691
SK (single model)0.7930.8260.8300.5210.7220.8590.9512.1393.528
TSB, SK0.7250.7910.7990.4890.6540.7750.9081.9373.026
SES, CRO, SBA, TSB, SK0.6350.7820.7830.4600.6120.7190.8451.7852.756
SMA, SES, CRO, SBA, TSB, SK0.6300.7400.8340.4560.6060.7110.8381.7672.721
mSBA, mTSB, SK–SSA0.5890.7100.7320.4330.6280.7460.7521.6272.557
CRO, SK, SK–SSA0.6110.7220.7400.4530.6230.7560.7601.6442.585
SK, mSBA, mTSB0.6490.7450.7610.4990.6470.7860.8321.7812.809
SMA, SES, CRO, SBA, TSB, SK, mSBA, mTSB, SK–SSA0.5390.6890.7150.3950.5290.6120.7021.4812.259
Table 13. Isolated demand categorized datasets.
Table 13. Isolated demand categorized datasets.
176 Lumpy SKUs
Demand IntervalsDemand SizeDemand per Period
MeanSt. DevMeanSt. DevMeanSt. Dev
Min1.30.62.51.81.31.0
25%ile1.50.97.35.94.04.8
Mean1.71.138.933.224.336.2
75%ile1.81.263.549.637.649.1
Max3.23.1165.5146.5101.2227.9
1572 Intermittent SKUs
MeanSt. DevMeanSt. DevMeanSt. Dev
Min1.30.61.10.30.40.1
25%ile1.81.217.24.96.82.4
Mean2.51.833.512.615.49.6
75%ile3.02.344.517.218.711.7
Max6.47.099.350.469.972.7
74 Erratic SKUs
MeanSt. DevMeanSt. DevMeanSt. Dev
Min1.10.35.54.44.16.4
25%ile1.10.414.211.412.024.4
Mean1.20.535.627.030.564.3
75%ile1.20.552.841.244.190.7
Max1.30.780.458.269.6195.3
Table 14. Performance comparison of forecast methods for the isolated data sets.
Table 14. Performance comparison of forecast methods for the isolated data sets.
Intermittent SKUs sSPECSMASESCROSBATSBSKmSBAmTSBSK–SSAOptimal
Average0.6150.6180.5860.5730.5810.5770.6180.5830.5260.448
Median0.2860.2970.2770.2450.2520.2430.1990.2480.1410.130
75th Percentile0.7680.7750.7290.7610.7000.7730.9090.7390.7800.572
25th Percentile0.2190.2070.2190.1670.1750.2030.1350.1750.0810.070
Maximum2.0892.1712.0442.1152.1202.0372.2672.1612.0672.007
Minimum0.0930.0920.0860.0650.0340.0780.0600.0470.0360.023
Lumpy SKUs sSPECSMASESCROSBATSBSKmSBAmTSBSK–SSAOptimal
Average0.5250.5250.4890.4630.4830.4760.4850.4800.4340.392
Median0.2290.2400.2230.1900.2200.2080.1620.2000.1270.112
75th Percentile0.3290.3830.3240.2870.3140.3030.3780.3000.2710.204
25th Percentile0.1850.1570.1490.1030.1280.1310.0980.1240.0700.039
Maximum2.1732.1472.0812.1892.1382.0492.2502.0662.0742.062
Minimum0.0270.0420.0330.0100.0150.0250.0100.0260.0220.009
Erratic SKUs sSPECSMASESCROSBATSBSKmSBAmTSBSK–SSAOptimal
Average0.4700.4750.4410.4240.4450.4160.4680.4340.4110.370
Median0.2310.2100.1930.1740.2040.2110.1940.1890.1690.147
75th Percentile0.6340.6950.6130.6430.6410.5390.7460.5920.6250.483
25th Percentile0.0960.1070.0920.0560.0810.0690.0510.0670.0490.022
Maximum1.5931.6901.7011.6921.6771.6561.8261.6891.6761.537
Minimum0.0190.0260.0140.0070.0200.0130.0050.0180.0070.001
Table 15. Scaled safety stock and backlogs (in parenthesis) for various service levels.
Table 15. Scaled safety stock and backlogs (in parenthesis) for various service levels.
ModelService Level
80%85%90%95%97%99%
SMA6.1 (10,980)8.7 (7670)12.4 (4546)18.6 (1851)23.2 (925)32.9 (192)
SES6.9 (9730)9.5 (6811)13.3 (4001)19.6 (1644)24.2 (834)34.0 (185)
CRO6.8 (9679)9.5 (6592)13.2 (3917)19.4 (1628)23.9 (837)33.8 (190)
SBA8.6 (8486)11.4 (5823)15.5 (3381)22.2 (1335)27.2 (667)37.8 (148)
TSB7.2 (9887)10.0 (6803)13.8 (4085)20.2 (1674)24.9 (870)35.1 (193)
SK7.1 (9052)9.76 (6232)13.5 (3656)19.8 (1478)24.5 (728)34.3 (152)
mSBA9.6 (5423)12.6 (3719)16.8 (2241)24.1 (922)29.3 (501)40.6 (138)
mTSB6.9 (6453)9.6 (4514)13.4 (2748)19.7 (1204)24.3 (679)34.2 (189)
SK–SSA9.3 (5381)12.2 (3758)16.4 (2267)23.4 (1005)28.6 (547)39.6 (150)
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Sfiris, D.S.; Koulouriotis, D.E. A New Approach to Forecast Intermittent Demand and Stock-Keeping-Unit Level Optimization for Spare Parts Management. Appl. Sci. 2025, 15, 12030. https://doi.org/10.3390/app152212030

AMA Style

Sfiris DS, Koulouriotis DE. A New Approach to Forecast Intermittent Demand and Stock-Keeping-Unit Level Optimization for Spare Parts Management. Applied Sciences. 2025; 15(22):12030. https://doi.org/10.3390/app152212030

Chicago/Turabian Style

Sfiris, Dimitrios S., and Dimitrios E. Koulouriotis. 2025. "A New Approach to Forecast Intermittent Demand and Stock-Keeping-Unit Level Optimization for Spare Parts Management" Applied Sciences 15, no. 22: 12030. https://doi.org/10.3390/app152212030

APA Style

Sfiris, D. S., & Koulouriotis, D. E. (2025). A New Approach to Forecast Intermittent Demand and Stock-Keeping-Unit Level Optimization for Spare Parts Management. Applied Sciences, 15(22), 12030. https://doi.org/10.3390/app152212030

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop