1. Introduction
Karst spring discharge time series are known to exhibit highly irregular, heterogeneous, and multifractal dynamics [
1,
2]. These complex patterns reflect the underlying geologic heterogeneity, including the interaction between the porous matrix, fracture networks, and conduit systems [
3]. Traditional linear hydrological models, which often rely on simplified assumptions of stationarity and Gaussian-like fluctuations, fail to capture the intricacies inherent in karst hydrology, such as long-range dependence, sudden jumps in flow rates, and the influence of extreme events [
4].
The recession curves of karst spring hydrographs, frequently decomposed into distinct flow components (e.g., conduit-driven quick flow and matrix-driven base flow) serve as important indicators of aquifer properties, infiltration processes, and hydraulic connectivity [
1]. Although conceptual models such as those proposed by Mangin [
5] and Fiorillo [
1] provide useful insights by distinguishing various storage and drainage compartments, they often lack the ability to represent the non-linear and fractal or multifractal nature of observed discharge records [
2,
5]. Empirical studies have shown that karstic systems display scale-invariant behavior: the complexity of flow paths and conduit geometries induces fractal patterns in spring discharge time series [
6]. The fractal dimension (
D) characterizes the degree of roughness and complexity on multiple scales [
7,
8] and, thus, serves as a crucial quantitative measure to understand karst processes [
9,
10].
Moment scaling for several springs, including the ones that are the subject of the present study, was presented in [
10] indicating that most hydrographs are monofractals rather than genuine multifractals (cf. [
11], section 6.3). Hence, a single number for the fractal dimension is fully informative. Theoretically, box-count and information dimensions coincide for monofractals; therefore, they similarly characterize the springs. As a result, the information dimension does not provide valuable new information but strengthens that of the box-count (capacity) dimension, which we consider throughout the present study.
Motivated by these findings, our primary objective here is to develop a stochastic modeling framework capable of reproducing both the fractal scaling and the jump-like features observed in karst spring discharges. In this study, we employ a fractional Ornstein–Uhlenbeck process (fOUp) enriched with jump components to model and simulate karst spring discharge dynamics. The fOUp is a continuous-time Gaussian process with long-range dependence characterized by a Hurst exponent (
H, with
), allowing it to capture persistent (
) and slowly decaying autocorrelations or anti-persistent (
), rough fluctuations inherent in karst systems [
12].
The usual, mean-reverting Ornstein–Uhlenbeck process (OUp;
) is defined by the stochastic differential equation (SDE):
where
is the mean reversion parameter,
is the long-term mean,
is the volatility, and
is a classic Brownian motion. Extending the OUp to a fractional context involves replacing
with a fractional Brownian motion (
) of the Hurst index (
H). This introduces long-memory characteristics into the dynamics when
and rough fluctuation when
.
The ultimate source for renewing the karstic hydrogeological system is infiltration from precipitation or from melting snow. To address sudden discharge spikes originating from infiltration, we incorporate a jump component into the driving force of the fractional Ornstein–Uhlenbeck (fOU) equation, resulting in a hybrid model that can be schematically written as follows:
where
represents the jump component. In general modeling practice, the jump component (
) often follows a compound Poisson process. However, a shortcoming is that the jump times and jump sizes are independent, and the process includes the Markov property, enforcing exponential interarrival times (i.e., times between jumps, also called sojourn times in renewal theory). This property is counterintuitive in our application, since a long delay between jumps—induced mainly by long delays in precipitation events—leaves the aquifer drained. Hence, the water from the next precipitation event—or a quick snowmelt for its turn—is kept back in the aquifer, filling up the porous matrix and the fracture networks, allowing for only lesser peaks in spring water discharge. Inspired by renewal theory and supported by statistical evidence (see
Section 6), we suggest using a semi-Markov jump process for
with Weibull-distributed interarrival times and Weibull-distributed peaks [
13]. In hydrology, the two-parameter Weibull (EV Type III) distribution is well known to fit low-flow and spring-flow extremes. For example, Sugiyama et al. (2007) [
14] noted that ”the Weibull distribution is very suitable for the probability plot of low stream flows”. Likewise, in our context, Leone et al. (2021) [
15] found that fitting spring discharge with a Weibull distribution ”fits the extreme values of both tails well”. Furthermore, a physical rationale is that the interarrival times between flow pulses can be viewed as “time-to-threshold” or “time-to-failure” processes (analogous to reliability theory), for which Weibull laws often arise. Contrary to Markov models, semi-Markov models allow for interdependence of the process state and the interarrival time and essentially place no restriction on the interarrival time distribution (see, e.g., ref. [
13]). The proper correlation of the interarrival times and the jump sizes is secured by the probability integral transform and its inverse, along with the Gaussian copula. To the best of our knowledge, this model specification is novel in the literature. The jump component describes the abrupt short term, whereas the fractional component fine tunes the long-term behavior.
We apply this fractional jump Ornstein–Uhlenbeck framework to daily discharge data from karst springs in northeast Hungary. The associated analysis illustrates the methodology’s capacity to replicate both the statistical and scaling features of the observed time series. By comparing simulated and observed discharge data, we demonstrate that this modeling strategy can reproduce essential empirical characteristics, such as fractal scaling, non-Gaussian tails, and sudden discharge extremes. Thus, the jump–fOUp approach provides new insights into the structure and complexity of karst aquifers and represents a promising avenue for understanding and predicting the non-linear, non-stationary, and fractal-like behavior of karstic spring discharges.
The remainder of this article is organized as follows.
Section 2 introduces the modeling framework—a fractional Ornstein–Uhlenbeck process augmented by a renewal–reward jump term—together with its mathematical properties and a short proof of its well posedness, which can be found in
Appendix A.
Section 3 reviews existing estimation techniques for the pure fOU process and adapts them into a two-tier procedure that later underpins our jump–fOUp inference.
Section 4 describes the geological setting, the 19-year daily discharge dataset from four Hungarian karst springs, and the exploratory analyses that motivate our model choices.
Section 5 presents the complete estimation–simulation workflow: detection of surges, copula-based generation of correlated jump pairs, fGn simulation for the fractional kernel, and the iterative scheme that couples both components.
Section 6 compares observed and simulated hydrographs, fractal dimensions, and cumulative discharges, demonstrating that jump–fOUp reproduces key statistical and visual features markedly better than short-memory alternatives.
Section 7 discusses hydrological implications, the necessity of long-range dependence, and possible extensions, such as time-varying parameters or precipitation-driven jump intensities. Finally,
Section 8 concludes with the main findings and outlines directions for future research.
3. Parameter Estimation Methods
Much effort has been devoted to the parameter estimation problem for fOU processes. Assuming that a continuous record of observations is available in the non-mean-reverting case (
), Kleptsina and Le Breton [
16] were the first to obtain the formula for the estimation of the drift parameter’s maximum likelihood (ML) and proved almost sure convergence. Bercu, Courtin, and Savy [
17] proved a central-limit theorem for the MLE in the case of
. They claimed, without proof, that the above convergence is also valid for
. Finally, Tanaka et al. [
18] completed the case by studying ML estimates when
and
.
Turning to mean reversion with
and keeping the assumption of continuous observation, Lohvinenko and Ralchenko [
19] obtained ML estimates of
and
when
and
.
Estimating the parameter vector (
) of the fOUp in the more realistic case of discrete, time-equidistant observations (
) poses significant challenges due to the fractional nature of the driving noise. In the non-mean-reverting case, specialized statistical methods were developed in [
20,
21], using the relationship between the Skorokhod (or divergence) and Stratonovich integrals. These works also provide limit theorems for the estimation, as both the time horizon and the mesh size grow to infinity simultaneously and the horizon increases faster than the mesh size. However, this condition is difficult to verify in a real application with a finite sample size.
3.1. A Two-Tier Estimation Framework for the fOU Parameters
This section consolidates, in a single presentation, the continuous-record methodology of [
20,
21] and its exact, discrete-sample analog introduced in [
22]. Both procedures target the parameter vector (
) of a fOUp.
3.1.1. Tier I—Continuous Record (Ref. [21])
Assuming full observation of the path (
), the mean reversion speed (
) can be consistently recovered by either least squares, i.e.,
or the energy-type statistic, i.e.,
where
denotes Euler’s gamma function. Both estimators are strongly consistent and asymptotically normal as
.
3.1.2. Tier II—Exact Discrete Record (Ref. [22])
Let
be an equally spaced sample of size
N with a grid step of
. We define second-order differences and their quadratic variations.
Closed-form, strongly consistent estimators are then obtained as
6. Application: Jump Distribution Fitting and fOUp Parameter Estimation for Karstic Spring Discharges
6.1. Separating the Jumps
The surges in spring discharges are not merely isolated spikes; rather, they exert a dampening effect on the recession curve. Consequently, it is not sufficient to model these as simple jumps in the process itself—they must be incorporated into the driving force of the dynamics, specifically in the differential component of the stochastic differential equation (SDE). Therefore, to locate the jump positions and sizes, the differenced series has to be considered. The quantiles and maxima of the differenced spring discharge series are presented in
Table 2.
The comparison of quantiles, maxima, and their relative proportions in
Table 2 reveals the heavy-tailed nature of the underlying distribution. This observation rules out the adequacy of a pure fOUp for modeling. Replacing fractional Brownian motion (fBm) with a Lévy flight in the SDE could address heavy-tail behavior, but it would fail to capture the correlation between extreme values and their timing. This limitation motivated the development of our proposed approach.
To corroborate the jump threshold used in this study,
Table 2 lists the empirical
,
, and
quantiles. Across all springs, the largest observed surge is roughly 20 to 450 times the
quantile and 10 to 200 times the
quantile, while the
quantile is only 2 to 3 times the
quantile. So, choosing the
quantile level as the cut-off effectively separates the extreme outliers, interpreted as hydrogeological ’jumps’, from the long-memory background. A total of 155 values are observed above the
quantile level, and it is still suitable to fit a distribution to them with sufficient accuracy. Hence, we choose this threshold to determine the jumps.
6.2. Jump Distributions and Waiting-Time Dependence
We analyze karst spring discharges through two complementary layers. First, we model the jump component by semi-Markov dynamics, whose interarrival times and magnitudes follow Weibull laws. Jump parameters (interarrival time and jump size distributions) were fitted by maximum likelihood and validated—for jump size, against a gamma distribution, too—with Kolmogorov–Smirnov tests reported in
Table 3. In addition, we compared Weibull fits to alternatives (e.g., gamma and log-normal) and found Weibull giving the lowest AIC and the most balanced fit to small and large flows. We report these statistics values in
Table 4.
Another possible alternative, the generalized gamma distribution, is known for its mathematical complexity and its requirement of large sample sizes for parameter estimation convergence and for clear distinction from the gamma and Weibull distributions. Indeed, the ML fit did not converge in the majority of our cases; therefore, we omitted it from our analysis.
After excising these jumps, the residual series is represented by a fractional Ornstein–Uhlenbeck baseline whose parameters are inferred from second-order differences as set out in
Section 3. All statistical decisions (choice of distribution, selection of the jump threshold, and verification of long-memory parameters) are supported by formal goodness-of-fit tests and by the scaling diagnostics summarized below.
6.3. fOU Background: Continuous Versus Truncated Records
After removing every surge that exceeds the 97.5th empirical quantile, a fractional Ornstein–Uhlenbeck (fOU) model is fitted to the resulting jump-filtered series.
Table 5 lists the estimated volatility (
), damping coefficient (
), memory scale (
), and Hurst exponent (
).
Across all springs, is small, so the background discharge reverts slowly to its mean; the associated memory () indicates pronounced persistence, while the conditional variance is far lower than in the raw data. The Hurst exponents fall between and , confirming long-range dependence in the smooth baseline component.
To quantify the improvement brought about by the jump–fOUp specification, we calculated both the mean absolute error (MAE) and the root-mean-square error (RMSE) and contrasted them with those of an AIC-selected short-memory AR(p) benchmark. Because the MAE figures mirror the RMSE pattern, we discuss them in the text rather than reproduce every value in
Table 6, which would add clutter without changing the conclusion.
The choice of the 97.5th-percentile jump threshold in the jump–fOUp minimizes the RMSE for every spring—e.g., from 756.28 to 661.73 at Csörgő and from 7590.31 to 2902.02 at Jósva—and similarly reduces the MAE in all cases, except for a slight increase at Csörgő (218.88 → 226.68). That isolated rise likely resulted from occasional extreme surges that influence squared errors more strongly than absolute errors. Overall, incorporating jumps and long memory delivers a clear accuracy gain over the AR(p) baseline, confirming the evidence discussed in
Section 7 for cumulative discharge.
The RMSE gains in
Table 6 show that adding jumps and long memory improves predictive skill. However, a rigorous assessment of how robust these gains are to parameter uncertainty is beyond the scope of this paper. Future work could combine Morris elementary-effects screening [
34] with Sobol indices for the most influential parameters while also testing a fractal-dimension objective (
D; hence,
H) as suggested by Bai et al. [
35]. Such analyses would clarify which coefficients control forecast skill and further validate the jump–fOUp framework under varied hydrological conditions.
6.4. Assessing Model Adequacy
Figure 2 displays the simulated and observed hydrographs of the four springs and demonstrates that simulations based on the fitted model exhibit strong visual agreement with the observed time series. This directly addresses a common criticism from hydrological and hydrogeological experts, who often contend that mathematical models—despite achieving good statistical fits in certain aspects—fail to produce realistic-looking hydrographs.
Although the ’eye test’ alone can be deceptive and misleading, plotting a classical ensemble mean or a ±SD ribbon is uninformative for a process driven by random time jumps; any goodness-of-fit measure based on point-by-point differences is inadequate. To illustrate, consider that in one year, the observed series might feature a flash flood on June 2, while the simulated series could exhibit a similar event on August 14. A difference-based metric would likely report a large mismatch, simply because one series contains an extreme event at a different time than the other. Since flash floods represent very large excursions from the base flow, they inevitably dominate such metrics—even though highlighting the exact time and extent of these excursions is not our objective and does not truly indicate overall fit. Among others reason, that is how the Nash–Sutcliffe Efficiency (NSE), a widely accepted performance metric in hydrology, suffers from exactly the same sensitivity to the timing and magnitude of individual peaks and is, therefore, unsuitable here.
Instead, we need a criterion that is insensitive to the timing of individual peaks. Therefore, we compare the quantiles of the simulated discharge process with those of the observed process. In this respect,
Table 7 displays the mean ratio of simulated to observed discharge quantiles. The slight overestimation of these quantiles by the simulation, represented in the table by values greater than one, is consistent with a conservative stance: future flash floods may well exceed anything observed to date.
We do not attempt to match the entire distribution because
- 1.
Stationarity of the process is unknown and not under investigation; and
- 2.
Even if the process were stationary, the strong temporal dependence of daily discharges would bias distribution-wide estimates.
High-end values, however, are nearly independent under broad conditions, so high-quantile estimates are far more reliable than full-distribution estimates, providing the rationale of our method.
6.5. Scaling Check via Fractal Dimension
Because a distribution theory for fractal-dimension estimators is not yet available, constructing a formal hypothesis test is not feasible. To verify that the combined jump + fOUP model captures the multiscale roughness of the observed discharge time series, we perform a Monte-Carlo re-estimation exercise: 500 synthetic datasets are generated from the fitted model, and the fractal dimension is re-estimated on each run. The empirical spread (reported in
Table 6) indicates that the point estimates are stable. While this is not a formal confidence interval, it provides a transparent indication of parameter uncertainty without invoking unverified asymptotics. We call it the empirical 90% confidence interval (CI), and the observed
D values are evaluated against these intervals.
Table 8 presents the observed
D alongside the mean and dispersion of the simulated distribution, while
Figure 3 visualizes the sampling distributions.
For every spring, the observed
D falls within the simulated 90% CI, consistently near its upper bound, indicating a slight positive bias, as also observable in
Figure 3, but no statistically significant discrepancy. For example, at Csörgő Spring, the observed
contrasts with a simulated mean of 1.291 (90% CI: 1.266–1.316).
Comparable agreement is obtained for Jósva, Kecskekút, and Komlós. This analysis confirms that the combined model faithfully reproduces the fractal roughness and the associated Hurst exponent () across scales.
The
values in
Table 8 describe the full jump + fOU process, so they are systematically lower (0.54–0.71) than the background
in
Table 5 (0.68–0.91); this confirms the well-known “spurious roughness” effect of jumps [
36]: high-frequency jump spikes shorten the effective memory, whereas the jump-filtered baseline retains the longer-range dependence quantified in the
Section 6.1. Simulation studies in the literature show that the madogram remains unbiased, but its variance increases when large jumps contaminate the signal [
37]. In contrast, once jumps are removed, the fOUp-specific estimator regains full efficiency.
6.6. Validation on Accumulated Discharge
From a water resource management perspective, cumulative discharge is of paramount importance, as it directly reflects the volume of water available for various uses over a given period. Therefore, its accurate estimation is a key objective in hydrological modeling.
Figure 4 demonstrates the performance of our model in this regard.
To highlight the critical role of long-range dependence, we replace the background flow component with an autoregressive (AR) process while retaining the same jump sequence used in the jump–fOUp simulation. The outcome is striking: the 100 cumulative discharge trajectories generated by the AR model (blue lines) significantly overshoot the observed discharge (red line)—in some cases, more than doubling it by the end of the period. In contrast, simulations based on the fOUp model closely track the empirical data, underscoring the model’s accuracy. Moreover, the variability among the jump–AR simulations is substantially higher than that of the jump–fOUp simulations, reflecting the latter’s smoother trajectories. These results illustrate that the jump–fOUp model yields more reliable cumulative discharge estimates, enhancing its utility for water resource planning and management.