Next Article in Journal
Multivariate Universal Local Linear Kernel Estimators in Nonparametric Regression: Uniform Consistency
Previous Article in Journal
Two Block Splitting Iteration Methods for Solving Complex Symmetric Linear Systems from Complex Helmholtz Equation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Change-Point Detection in Functional First-Order Auto-Regressive Models

by
Algimantas Birbilas
*,† and
Alfredas Račkauskas
*,†
Institute of Applied Mathematics, Vilnius University, Naugarduko g. 24, LT-03225 Vilnius, Lithuania
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Mathematics 2024, 12(12), 1889; https://doi.org/10.3390/math12121889
Submission received: 30 May 2024 / Revised: 11 June 2024 / Accepted: 14 June 2024 / Published: 18 June 2024

Abstract

:
A sample of continuous random functions with auto-regressive structures and possible change-point of the means are considered. We present test statistics for the change-point based on a functional of partial sums. To study their asymptotic behavior, we prove functional limit theorems for polygonal line processes in the space of continuous functions. For some situations, we use a block bootstrap procedure to construct the critical region and provide applications. We also study the finite sample behavior via simulations. Eventually, we apply the statistics to a telecommunications data sample.

1. Introduction

A change-point in a sample is present when at least one statistical parameter changes, which may be caused by a variation in the mean, trend, or other parameters. Change-point detection has a broad range of applications, including financial time series analyses [1,2], econometrics [3], and climate change [4], to name a few. Given a sample of functional observations x 1 ( t ) , , x n ( t ) , t [ 0 , 1 ] , our goal is to detect the existence of a change-point such that before the change-point, samples are generated by a stationary functional auto-regressive process of the first-order, while after the change-point, stationarity is destroyed by the variation in the mean.
Statistical methods for functional observations were pioneered by Ramsey [5] in 1982. As the amount of high-resolution data has increased dramatically over the years, interest in functional data analysis (FDA) has emerged. In recent decades, multiple authors have contributed to the improvement of the theoretical FDA background—Kokoszka and Reimherr [6] and Ramsey et al. [7], to name few—and the practical implementation background of FDA in various fields—Ramsey et al. [8], Aneiros et al. [9], Alaya et al. [10], and Koerner et al. [11].
The first contribution to structural breaks in functional data was made by Berkes et al. [12], where a cumulative sum (CUSUM) test was proposed for independent functional data by using projections of the sample onto some principal components of covariance. Aue et al. [13] analyzed the limit distribution of a change-point estimator in the same setting. Horváth et al. [14] constructed a test for the stability of the autoregressive operator in a Hilbert-space-valued autoregressive process. Tests that do not rely on dimension reduction were derived by Horváth et al. [14] for general testing of stationarity and by Aue et al. [13] for functional break detection of m-dependent innovations. Structural break detection in the context of functional linear models was studied by Horváth and Reeder [15] and Aue et al. [13]. Smooth deviations from stationarity of functional time series in the frequency domain were investigated by Aue and van Delft [16]. Danielius and Račkauskas [17] suggested tests based on p-variation of functional CUSUM processes.
Still, a limited number of papers use functional approaches for telecommunications data. Aspirot et al. [18] studied a non-parametric regression model with the explanatory variable as a non-stationary dependent functional and the response variable as a scalar. Yu and Lambert [19] analyzed records for completed international calls. Birbilas and Račkauskas [20] modeled call data records and predicted mobile product consumption using different approaches in a functional framework. Change-point detection plays a crucial role in telecommunications as it can be monitored and implemented at different levels depending on the business model, e.g., the operational level (system failures in cell towers), the technical level (signal changes and wireless connection stability), and the marketing level (a promotion’s effect on usage). Eventually, it has a significant effect at the business level, as accurate and timely detection of change-points is imperative for network operators to proactively address issues, optimize resource allocation, and ensure seamless service delivery. In other words, a change-point might identify improvement or, worse, system failure. System failures usually cost more than improvements. As telecommunications systems continue to evolve in complexity and scale, the application of sophisticated change-point detection techniques becomes increasingly indispensable for maintaining the integrity and functionality of this critical communication infrastructure.
Change-points are a relevant topic in both the scientific and telecommunications fields. In Shields et al. [21], the authors analyze collective behavior within smart cities using time-series-type call data record (CDR) data. The authors identify multiple change-points using three methods: binary segmentation, segmentation neighborhoods, and pruned extract linear time (PELT). They suggest that deviation from normal collective mobility patterns can be explained by known significant events in the city. Loreh John [22] provides a study using change-point analysis to detect changes in the mean for the wireless telecommunications field. The author uses a previously known likelihood method and suggests a new maximum distance of the running means method to identify locations of the change-points within the 2G network. One of the latest works in this area is by Aleksiejunas and Garuolis [23] and is devoted to traffic change-points; it utilizes machine learning methods such as long short-term memory (LSTM) and recurrent neural networks (RNNs). The authors apply change-point identification algorithms for synthetic data and reuse algorithms for the spatial traffic distributions of LTE (Long Term Evolution) mobile networks.
Recall that for a separable Banach space B with a norm x B , x B , C ( [ 0 , 1 ] , B ) is a separable Banach space of continuous functions f : [ 0 , 1 ] B endowed with the norm f B , = max 0 t 1 x ( t ) B , f C ( [ 0 , 1 ] , B ) . We abbreviate C ( [ 0 , 1 ] , R ) as C [ 0 , 1 ] and x R , as x . For a given random sample X 1 ( t ) , , X n ( t ) , t [ 0 , 1 ] , with values in the space C [ 0 , 1 ] , we consider the following model:
X k μ = β ( X k 1 μ ) + g ( k / n ) + Y k ,
where β : C [ 0 , 1 ] C [ 0 , 1 ] is a linear bounded operator, μ C [ 0 , 1 ] , the function g : [ 0 , 1 ] C [ 0 , 1 ] is deterministic, and Y i = ( Y i ( t ) , t [ 0 , 1 ] ) , i Z , is a sequence of independent and identically distributed (iid) random variables with values in the space C [ 0 , 1 ] and defined on a complete probability space ( Ω , F , P ) .
We analyze this model while keeping in mind the telecommunications data coming from call data records (CDRs). This data set contains the usage of three mobile products: voice calls, SMS, and mobile data. For the case study, voice consumption in minutes and mobile data consumption in MB are used. This type of date was analyzed in [20] using first-order auto-regressive models, and it was confirmed that models of type (1) are appropriate.
Throughout, Y = ( Y ( t ) , t [ 0 , 1 ] ) denotes a generic element of the sequence ( Y i , i Z ) for which the following assumption shall be used.
Assumption 1 (Assumption (A)).
Y is a mean-zero C [ 0 , 1 ] -random variable with covariance function q, q ( t , s ) = E Y ( t ) Y ( s ) , t , s [ 0 , 1 ] , and satisfies the central limit theorem (denoted Y C L T ( C [ 0 , 1 ] ) .
Recall that Y C L T ( C [ 0 , 1 ] ) means that there exists a mean-zero Gaussian random variable (denoted N q ) in C [ 0 , 1 ] with covariance q such that the sequence ( n 1 / 2 ( Y 1 + + Y n ) , n 1 ) converges in a distribution to N q . It is well known that if there exists a random variable M such that E M 2 < and
| Y ( t ) Y ( s ) | M | t s | α , s , t [ 0 , 1 ] ,
with some α ( 0 , 1 ] , then Y C L T ( C [ 0 , 1 ] ) .
Throughout, we assume
β : = sup x < 1 β ( x ) < 1 .
This assumption guaranties that the inverse operator ( I β ) 1 exists, where I is the identity operator, and has the expansion
( I β ) 1 = k = 0 β k .
Our main aim is to test the hypothesis
H 0 : g = 0 versus H 1 : g 0 ,
with emphasis on the case of change-point detection, which corresponds to a piecewise constant function g with respect to the first argument.
The model under consideration covers both abrupt as well continuous variations of a sample mean. Test statistics and their asymptotic distributions under the null hypothesis as well under various alternatives are established and presented in Section 2. Section 3 contains a description of the testing procedure. The main challenge here is to find critical levels. The solution for this problem is proposed via the Monte Carlo method. Section 4 contains a simulation study. Finally, Section 5 deals with a concrete data set related to telecommunication problems.

2. Auxiliary Results

For each n 1 , consider the random sample Z 0 , Z 1 , , Z n in C [ 0 , 1 ] satisfying
Z 0 = 0 , Z i = β Z i 1 + f ( i / n ) r n + Y i , i = 1 , , n ,
where r n C [ 0 , 1 ] , and f : [ 0 , 1 ] R . Consider a polygonal line process ζ n = ( ζ n ( t ) , t [ 0 , 1 ] ) defined by
ζ n ( t ) = k = 1 n t Z k + ( n t n t ) Z n t + 1 = k = 1 n Z k e n k ( t ) ,
where
e n k ( t ) = 0 if t < ( k 1 ) / n , t n ( k 1 ) if ( k 1 ) / n t k / n , 1 if t > k / n .
Let W q = ( W q ( t ) , t [ 0 , 1 ] ) , be a C [ 0 , 1 ] -valued Wiener process corresponding to covariance q in such a way that W q is a C [ 0 , 1 ] -valued mean-zero Gaussian process with independent increments such that W q ( s ) W q ( t ) has the same distribution as N q | t s | 1 / 2 .
Recall, that for a function f : [ 0 , 1 ] R ,
v 1 ( f , [ a , b ] ) : = sup k = 1 m | f ( τ k ) f ( τ k 1 ) | : a = τ 0 < τ 1 < < τ m 1 < τ m = b , m 1 .
The function f has finite variation if v 1 ( f ) : = v 1 ( f : [ 0 , 1 ] ) < .
Lemma 1.
Assume that Y satisfies Assumption 1, and
(a) 
There is a continuous function r : [ 0 , 1 ] R such that
lim n n r n r = 0 ;
(b) 
The function f : [ 0 , 1 ] R is bounded and has finite variation.
Then
n 1 / 2 ( I β ) ζ n D W q + h r , f in the space C ( [ 0 , 1 ] , C [ 0 , 1 ] ) ,
where
h r , f ( t ) = 0 t f ( s ) d s r , t [ 0 , 1 ] .
Proof of Lemma 1.
Consider a linear process ( Z ˜ k , k 1 ) with values in C [ 0 , 1 ] defined by
Z ˜ k = j = 0 β j Y k j = j = k β k j Y j , k = 0 , 1 , 2 , .
Note that the series above converge a.s. in C [ 0 , 1 ] so that Z ˜ k is correctly defined and is a C [ 0 , 1 ] -valued random variable for each k 0 . Indeed,
E j = m n β j Y k j j = m n β j E Y 0 as m , n
yields convergence in the probability of the series j β j Y k j . As ( Y i ) are independent, the Ito–Nisio theorem ensures a.s.-convergence as well. Next, we consider a C [ 0 , 1 ] -valued polygonal line process ζ ˜ n defined by
ζ ˜ n ( t ) = k = 1 n t Z ˜ k + ( n t n t ) Z ˜ n t + 1 = k = 1 n Z ˜ k e n k ( t ) , t [ 0 , 1 ] .
We have by [24] (see Theorem 5 therein)
n 1 / 2 ( I β ) ζ ˜ n D W q in the space C ( [ 0 , 1 ] , C [ 0 , 1 ] ) ,
Since Z k can be expressed by
Z k = j = 1 k β k j Y j + j = 1 k β k j f ( j / n ) r n = j = 1 k β k j Y j + j = 0 k 1 β j f ( ( k j ) / n ) r n ,
we have
Z k Z ˜ k = j = 0 k 1 β j f ( ( k j ) / n ) r n j = 0 β k j Y j .
This yields
n 1 / 2 ζ n ( t ) = n 1 / 2 ζ ^ n ( t ) + Δ n ( t ) , t [ 0 , 1 ] ,
where Δ n ( t ) = n 1 / 2 [ ζ n ( t ) ζ ˜ n ( t ) ] . We claim that
max 0 t 1 Δ n ( t ) ( I β ) 1 r 0 t f ( s ) d s C [ 0 , 1 ] n P 0 .
To prove this claim, first we decompose
Δ n ( t ) = Δ n ( t ) + Δ n ( t ) , t [ 0 , 1 ] ,
where
Δ n ( t ) = n 1 / 2 k = 1 n j = 0 k 1 β j f ( ( k j ) / n ) r n e n k ( t ) , Δ n ( t ) = n 1 / 2 k = 1 n j = 0 β k j Y j e n k ( t ) .
Since Δ ( t ) = n 1 / 2 Z ˜ 0 k = 1 n β k e n k ( t ) , we see that max 0 t 1 Δ n C [ 0 , 1 ] n P 0 , and (9) reduces to
max 0 t 1 Δ n ( t ) ( I β ) 1 r 0 t f ( s ) d s C [ 0 , 1 ] n P 0 .
Write n 1 / 2 Δ n ( t ) = Δ n 1 ( t ) + Δ n 2 ( t ) , where
Δ n 1 ( t ) = n 1 k = 1 n j = 0 k 1 β j f ( ( k j ) / n ) r e n k ( t )
and
Δ n 2 ( t ) = n 1 k = 1 n j = 0 k 1 β j f ( ( k j ) / n ) ( n r n r ) e n k ( t ) .
Next, observe that
Δ n 2 ( t ) C [ 0 , 1 ] f ( 1 β ) 1 n r n r C [ 0 , 1 ] ,
since k = 1 n e n k ( t ) = n t . By condition (5), lim n max 0 t 1 Δ n 2 ( t ) C [ 0 , 1 ] = 0 and (10) reduces to
max 0 t 1 Δ n 1 ( t ) ( I β ) 1 r 0 t f ( s ) d s C [ 0 , 1 ] 0 .
Consider the functions
I n ( h , t ) = n 1 k = 1 n f ( k / n h ) e n k ( t ) , I ( h , t ) = 0 t f ( s h ) d s ,
We have
| I n ( h , t ) I ( h , t ) | = | k = 1 n ( k 1 ) / n k / n [ f ( k / n h ) f ( s h ) ] 1 [ 0 , t ] ( s ) d s | k = 1 n ( k 1 ) / n k / n | f ( k / n h ) f ( s h ) | d s n 1 k = 1 n v ( f ( · h ) : [ ( k 1 ) / n , k / n ] ) n 1 v ( f : [ h , 1 h ] ) n 1 v ( f : [ 0 , 1 ] )
Next, we observe that
| I ( h , t ) I ( 0 , t ) | 0 t | f ( s h ) f ( s ) | d s 0 as h 0 .
Fix J 1 . Then for each ε > 0 , we find N ε 1 such that
| I ( j / n , t ) I ( 0 , t ) | 0 t | f ( s h ) f ( s ) | d s < ε if n N ε , j J .
Since f ( t ) = 0 for t 0 ,
Δ n 1 ( t ) = n 1 k = 1 n j = 0 k β j r f ( ( k j ) / n ) e n k ( t ) = k = 1 n j = 0 n β j r f ( ( k j ) / n ) e n k ( t ) = j = 0 n β j r k = 1 n f ( ( k j ) / n ) e n k ( t ) = j = 0 β j r I n ( j / n , t ) .
This yields
max 0 t 1 Δ n 1 ( t ) ( I β ) 1 r 0 t f ( s ) d s C [ 0 , 1 ] =   j = 0 β j r I n ( j / n , t ) j = 0 β j r I ( 0 , t ) C [ 0 , 1 ] j = 0 J β j r ε + 2 f j = J β j r
Letting J gives
max 0 t 1 Δ n 1 ( t ) ( I β ) 1 r 0 t f ( s ) d s C [ 0 , 1 ] r 1 β ε .
Since ε > 0 is arbitrary, this yields (11) and completes the proof.    □
Consider an operator L : C ( [ 0 , 1 ] , C [ 0 , 1 ] ) C ( [ 0 , 1 ] , C [ 0 , 1 ] ) defined by L x ( t ) = x ( t ) t x ( 1 ) , t [ 0 , 1 ] , x C ( [ 0 , 1 ] , C [ 0 , 1 ] ) . Define a C [ 0 , 1 ] -valued Brownian bridge B q = ( B q ( t ) , t [ 0 , 1 ] ) , by
B q ( t ) = W q ( t ) t W q ( 1 ) = L W q ( t ) .
Lemma 2.
Under the conditions of Lemma 1,
n 1 / 2 ( I β ) L ζ n D B q + L h r , f in the space C ( [ 0 , 1 ] , C [ 0 , 1 ] ) ,
where
L h r , f ( t ) = h r , f ( t ) t h f ( 1 ) = 0 t f ( s ) d s t 0 1 f ( s ) d s r , t [ 0 , 1 ] ) .
Proof of Lemma 2.
Since L is a continuous operator, the result follows from Lemma 1 and the continuous mapping theorem (Billingsley [25] Theorem 2.7).    □

3. Test Statistics and Their Asymptoticity

Define
M n , β : = max 1 k n max 0 t 1 | ( I β ) j = 1 k X j ( t ) | ,
and
T n , β = max 1 k n max 0 t 1 | ( I β ) j = 1 k ( X j ( t ) X ¯ n ( t ) ) | ,
where X ¯ n = n 1 ( X 1 + + X n ) . Statistic M n , β is adopted for testing (3) if μ = 0 , whereas T n , β can be used for testing (3) in the case where we have μ 0 .
Theorem 1.
For the model (1), assume that Y satisfies Assumption 1 and μ = 0 , g = 0 . Then
n 1 / 2 M n , β D M q : = max 0 t 1 W q ( t ) C [ 0 , 1 ] .
Proof of Theorem 1.
By Lemma 1 and the continuous mapping theorem,
n 1 / 2 ( I β ) ζ n C [ 0 , 1 ] , D W q C [ 0 , 1 ] , .
Observing that M n , β = ( I β ) ζ n C [ 0 , 1 ] , and M q = W q C [ 0 , 1 ] , completes the proof.    □
Theorem 2.
For the model (1), assume that Y satisfies Assumption 1 and g = 0 . Then
n 1 / 2 T n , β D T q : = max 0 t 1 B q ( t ) C [ 0 , 1 ] .
Proof of Theorem 2.
Since substituting X k with X k μ does not change the statistic T n , β , we assume μ = 0 . The function L : C ( [ 0 , 1 ] , C [ 0 , 1 ] ) is continuous; hence, by the continuous mapping theorem, n 1 / 2 L ζ n D L W q . Applying the continuous mapping theorem once more yields
n 1 / 2 max 0 t n ( I β ) L ζ n ( t ) C [ 0 , 1 ] D max 0 t 1 L W q ( t ) C [ 0 , 1 ] .
Next, we notice that max 0 t n ( I β ) L ζ n ( t ) C [ 0 , 1 ] = T n , β , whereas L W q = B q . This completes the proof.    □
Theorem 3.
Assume for the model (1) that Y satisfies Assumption 1, g ( s , t ) = g n ( s , t ) = f ( s ) r n ( t ) satisfies n r n r in C [ 0 , 1 ] , and the function f : [ 0 , 1 ] R is bounded and has finite variation. Then
n 1 / 2 T n , β D T q , r : = max 0 t 1 B q ( t ) + 0 t f ( s ) d s t 0 1 f ( s ) d s r C [ 0 , 1 ] ,
Particularly, if lim n n max 0 t 1 | r n ( t ) | = , then
n 1 / 2 T n , β n P .
Proof of Theorem 3.
The proof follows from Lemma 2 and the continuous mapping theorem.    □
Theorem 4.
Assume for model (1) that Y satisfies Assumption 1, g = 0 , and μ = μ n satisfies n μ n r in C [ 0 , 1 ] . Then
n 1 / 2 M n , β D M q , r : = max 0 t 1 W q ( t ) + r C [ 0 , 1 ] .
Particularly, if lim n n max 0 t 1 | μ n ( t ) | = , then
n 1 / 2 M n , β n P .
Proof of Theorem 4.
The proof follows from Lemma 1 and the continuous mapping theorem.    □
Remark 1.
In practice, the auto-regression operator β is unknown and has to be substituted with its consistent estimator β ^ . This substitution can be controlled through the following convergence:
| n 1 / 2 M n , β n 1 / 2 M n , β ^ |   β β ^ · ( I β ) 1 n 1 / 2 M n , β n P 0
provided β β ^   n P 0 .
The simulations below (Figures 16 and 17) confirm this remark.

4. Testing Procedure

Assume that the functional sample X 1 , , X n follows model (1). Consider the null hypothesis H 0 : g = 0 . For a fixed 0 α < 1 , let C α and C α be solutions of equations
P ( M q > C α ) = α and P ( T q > C α ) = α
respectively. According to Theorems 1 and 2, the tests
n 1 / 2 M n , β > C α and n 1 / 2 T n , β > C α
have an asymptotic significance level of α .
If q is known, then by the Kuelbs [26] invariance principle, as N ,
F q , N ( x ) : = P N 1 / 2 max 1 k N max 0 s 1 | j = 1 k ξ q j ( s ) | x P ( M q x ) , x 0 ,
where ξ g j , j = 1 , , N are iid Gaussian random variables in C [ 0 , 1 ] with mean zero and covariance q. Hence, an approximate solution of the first equation in (21) can be found from
1 F q , N ( C α ) = α
by the Monte Carlo method. Similarly, as N ,
F ˜ q , N ( x ) : = P N 1 / 2 max 1 k N max 0 s 1 | j = 1 k ξ q j ( s ) M 1 j = 1 n ξ q j ( s ) | x P ( T q x ) , x 0 ,
and an approximate solution of the second equation in (21) can be found by the Monte Carlo method from
1 F ˜ q , N ( C α ) = α .
Particularly, if ( Y k , k 1 ) are iid standard Wiener processes, then q ( s , t ) = min { s , t } , s , t [ 0 , 1 ] , and again, denoting
F M , N ( x ) = P N 1 / 2 M 1 / 2 max 1 k N max 1 m M | j = 1 k i = 1 m γ i j | x ,
where ( γ i j , i = 1 , , N , j = 1 , , M ) are iid standard normal random variables, and using the Donsker invariance principle, the solution of (23) can be approximated by the solution of the following equation:
1 F M , N ( C α ) = α .
Similarly, (24) can be approximated by
1 F ˜ M , N ( C α ) = α ,
where
F ˜ M , N ( x ) = P N 1 / 2 M 1 / 2 max 1 k N max 1 m M | j = 1 k i = 1 m γ i j M 1 i = 1 M γ i j | x , x > 0 .
Density functions of the distribution functions F N , M and F ˜ M , N (with M = N = 1000 ) were generated using kernel (Epanechnikov) density estimates and M C { 100 , 400 , 500 , 800 , 1000 } replications of the corresponding random variables (see Figure 1). We see that the effect of the number of MC replications is minimal.
Equations (25) and (26) can be solved by the Monte Carlo method. The obtained quantile values of C α and C α are presented in Table 1 and Table 2.
The true critical value C n ( α ) for the test statistic M n , β corresponding to the pre-designated significance level α ( 0 , 1 ) is defined as a solution of the following equation:
P ( M n , β > C n ( α ) | H 0 ) = α
the exact solution of which can be obtained only in some very special cases. Therefore, critical values obtained from asymptotic distribution theory are widely used in applications. However, this usually requires a very large sample size since first-order asymptotic theory often gives poor approximations of the distributions of test statistics.
Another way to solve Equation (27) is the block bootstrap procedure, as it can assist us with the approximation of critical values. In the works of Härdle et al. [27], Kunsch [28], and Liu and Singh [29], we find multiple algorithms on how to construct the block bootstrap for time series data. There exist numerous variants of the block bootstrap method that can mainly be distinguished by whether the blocks are overlapping or not and whether the block length is fixed or increasing. The overall block bootstrap idea, however, involves partitioning of the sample into sub-blocks and then independently drawing these sub-blocks until the sample size is attained. Adapted to our case, a block bootstrap method with fixed length l and overlapping blocks is performed based on the work of Nyarige [30]. We take the algorithm from Nyarige [30] and adapt it to the overlapping case: see Algorithm 1. The application of the procedure will be presented in the simulation study.
Algorithm 1 Block bootstrap procedure
  • Given the sample { X 1 , ..., X n }, using a suitable rule, choose the fixed block length, l, and partition the sample into B overlapping blocks, each with the given length l. As blocks can overlap, B = n l + 1 .
  • Draw randomly for a replacement from the resulting blocks and order the drawn blocks from end-to-end to form a new bootstrap sample { X 1 * , ..., X n * }.
  • Compute the statistic of interest in the same way as for the original sample.
We discuss asymptotic critical value adjustments afterwards. In the case where ( Y k ) are iid standard Brownian motions and the operator β is multiplied by a function, we generate 500 Monte Carlo simulations M n , β under H 0 to have a test statistics sample { M n , β 1 , …, M n , β 500 }. Epanechnikov kernel density plots of the sample { M n , β 1 , …, M n , β 500 } and of M q are shown in Figure 2. The Monte Carlo experiment provides evidence that the true and nominal probabilities that the test makes a Type I error can be notably dissimilar when an asymptotic critical value is used. Critical value adjustments can be made with the help of a Bonferroni correction: see Haynes [31] for more details and Sedgwick [32] for an application example. One can see that the Monte Carlo experiment density (blue density) is pushed to the left in comparison to the asymptotic critical density (yellow density). The graph provides a suggestion that quantiles can be adjusted to have the Type I (false-positive) error rejection rate improved. Adjustment of the asymptotic critical value means finding a C α * that satisfies
P ( M q > C α * ) = α * < α
such that P ( M n , β > n 1 / 2 C α * | H 0 ) is as close as possible to α .
This can be done by solving the following equation:
P ( M q > C α * ) P ( M n , β > n 1 / 2 C α * | H 0 ) .
The actual adjusted asymptotic critical values and applications will be presented in the simulation study.
In the case where the operator β is unknown, the critical values can be adjusted also. Considering the case where the operator β in (1) is a multiplication by a function β ( t ) , t [ 0 , 1 ] , we define the following Algorithm 2.
Algorithm 2 Adjusted quantiles procedure
  • Calculate β ^ (t) from the functional auto-regressive model defined in Equation (1)
  • Calculate Y ^ k (t) by equation Y ^ k (t) = X k + 1 (t) − β ^ (t) X k (t), k = 1 , , n . Afterwards, center the residuals Y ˜ k (t) = Y ^ k (t) − 1 n   n = 1 k   Y ^ k (t), k = 1 , , n .
  • Draw randomly (with repetition) and with probability P ( Y ˜ n = Y k * ) = 1 n , k = 1 , , n , from { Y ˜ 1 , Y ˜ 2 , , Y ˜ n } to generate { Y 1 * , Y 2 * , , Y n * } .
  • Generate new sample data { X 1 * , X 2 * , , X n * } with the first moment by equation X 1 * ( t ) = X 1 ( t ) = Y 1 * (t) and the following moments by equation X k * (t) = β ^ (t) X k 1 * + Y k * , k = 1 , , n .
  • Use block bootstrap procedure, see Algorithm 1, with the length of the block l to draw from { X 1 * , X 2 * , , X n * } new values { X 1 , X 2 , ..., X n }. Then, calculate M n , β ^ from { X 1 , X 2 , ..., X k }, k = 1 , , n , using Equation (12).
  • Repeat previous step as Monte Carlo simulations for MC times to construct { M n , β ^ 1 , M n , β ^ 2 , ..., M n , β ^ M C }.
  • Adjust M q quantile values with { M n , β ^ 1 , M n , β ^ 2 , ..., M n , β ^ M C } to have Type I error minimized.
Algorithm 2 is built to solve the following equation:
P ( M n , β > C α ) P ( M n , β > C α | X 1 , X 2 , , X n ) P ( M q > C α ) .
A graphical interpretation of Algorithm 2 is presented in Figure 3. One can see that the initial M n , β ^ density generated using the evaluated β ^ (t) is positioned with notable dispersion and a heavy tail to the right. Density improves with the following steps generating M n , β ^ * . Eventually, we have M n , β ^ density, which is close to M q . An application of this algorithm will be presented in the case study.

5. Simulation Study

In this section, we consider the random functional sample X i , i = 1 , , n , generated by the model
X k ( t ) μ ( t ) = β ( t ) ( X k 1 ( t ) μ ( t ) ) + g ( k / n , t ) + Y k ( t ) , k = 1 , , n ,
where the functions Y k , k = 1 , , n are iid standard Wiener processes, and β = ( β ( t ) , t [ 0 , 1 ] ) is a continuous function. We analyze test (22) in a Monte Carlo simulation study, distinguishing two cases:
  • Case 1: g = 0 , with the aim to test
    H 0 : μ = 0 versus H 1 : μ 0 .
  • Case 2: μ = 0 , with the aim to test
    H 0 : g = 0 versus H 1 : g 0 .
In both cases, the following β functions are used:
β 1 ( t ) = 0.7 t , β 2 ( t ) = 0.7 t , β 3 ( t ) = 0.7 1 t , β 4 ( t ) = 0.7 1 t , t [ 0 , 1 ] .
First we analyze the test statistic M n , β for the model (31) under assumptions g = 0 and μ = 0 . Figure 4 shows that coefficient b in β ( t ) = b t , t [ 0 , 1 ] has a minor effect on distribution of M n , β if b < 0.7 . If b exceeds 0.7 and approaches one, the M n , β density is flattened and forced to the right. Moreover, Figure 5 indicates minor differences between the densities if different β i (t), i = 1, …, 4, from Equation (32) are used. Therefore, in the following analysis, model (31) with β 1 ( t ) = 0.7 t t [ 0 , 1 ] will be used. Simulations are performed with various n values: 100, 200, 500, 1000. These simulations are done for the null hypothesis with functions μ (t) = 0 and g ( k / n , t ) = 0; β 1 (t) is used in the following graphical analysis, and summary values are provided in a table for β 1 (t), β 2 (t), β 3 (t), and β 4 (t).

5.1. False-Positive Rate

In this subsection, we are analyzing the false-positive or Type I error rate. We start with a comparison of the kernel density plots for M n , β under H 0 versus M q . Figure 6a shows that a higher n indicates that the density of the null hypothesis is moving towards the M q density while n < 1000. This identifies a false-positive issue that can be seen from the graph—the test will conduct a significant number of false-positive errors if n ≤ 1000.
Figure 7a implies that the Type I error rate decreases dramatically once n ≥ 1000. We adjust asymptotic critical values by solving Equation (29) in order to minimize the Type I error rate. The adjusted asymptotic critical values corresponding to β 1 ( t ) , β 2 ( t ) , β 3 ( t ) , β 4 ( t ) are given in Table 3. Indeed, the adjusted asymptotic critical values provide better results for false-positive cases: see Figure 7. This defines our upcoming steps: for further testing in simulations, adjusted one-sided asymptotic critical values will be used for n ≥ 1000.
If n < 1000 , then we perform a block bootstrap by Algorithm 1. The great question is which block length is optimal. The question is answered by using generated bootstrap data with different block lengths l = {2, 3, …, 10, 20, 50, 100, 200}. The decision is reached by two criteria for errors (mean square error and mean absolute error): the most notable change in error by the elbow rule and the lowest error value. Eventually, this indicated that the best block length is equal to 4. After ascertaining all the parameters for Algorithm 1, we perform it MC = 500 times to calculate the sample of the test statistics Z = { M n , β 1 , M n , β 2 , …, M n , β 500 }. We calculate quantile values from Z and decide by the Z 1 α value: we reject the null hypothesis with a significance level α if M n , β is higher than Z 1 α .
The density plots in Figure 6b provide evidence that the block bootstrap is a good choice for the n = 100 situation. The density plot for M n , β generated with n = 100 is observed in both Figure 6a,b in green color. If one compares this density with the M q density graph in Figure 6a, it becomes evident that the densities corresponding to n = 100 exhibit a sub-optimal degree of overlap with their associated M q density values (dotted blue color). In Figure 6b, comparing the density in the relation between the n = 100 density and the block bootstrap values (dotted black color), a noteworthy enhancement in the degree of overlap is observed. We ask how the block bootstrap procedure can help to avoid false-positive errors. Figure 8 implies that the procedure decreases the false positive rate by a notable amount. Figure 8a adds that the Type I error rate tested with block bootstrap critical values is stable when n ≤ 100, and the false-positive rate stays under 0.4. Improved outcomes are evident for lower values of n if we apply the block bootstrap procedure.
To conclude the M n , β under H 0 graphical analysis, for cases with n ≥ 1000, it is advised to use M q , and for n < 100, use block bootstrap critical values of M n , β . For n (100;1000), we suggest using the block bootstrap procedure to generate quantile values or using adjusted asymptotic critical values.

5.2. Power Analysis

Y k (t) are iid standard Wiener processes. In this section, we have H 1 : g 0 or H 1 : μ 0 depending on the case:
  • Case 1
    g = 0 a n d μ ( a ; t ) = a t .
  • Case 2
    μ = 0 a n d g a ( k / n ; t ) = a t 1 [ k * , n ] ( k ) .
Simulation example graphs are provided in Figure 9 for deeper understanding. One can see that we might encounter issues as the grey curves (simulations with jumps) and the black curves (simulations with no jumps) overlap.
By testing different a values from the interval, we observe the test power, which is defined by the following probability:
P ( M n , β > C α | H 1 ) = β .
Power calculations in the following paragraph will be provided for two situations: n = 100 and n = 1000. Moreover, calculations are done using block bootstrap critical values, see Algorithm 1, for n = 100 and adjusted asymptotic critical values, see Algorithm 2, for n = 1000.
Simulations commenced with the Case 1 scenario for different n values. Prior to delving into a comprehensive power analysis, we conduct an initial examination of the density plots. The examination is started with the simulated density plots in Figure 10, where the M n , β density is simulated using jump parameter a = 0.5. Although the density graph in Figure 10a identifies that M q quantiles can be used to test the change-point, the results are dissatisfying with n = 100. If the M n , β block bootstrap is used, see Figure 10b, the results are better and the density plots have a higher degree of overlap than in the M q case. The problem is solved when the jump parameter is higher, a = 1, and both M q and the M n , β block bootstrap can be used to identify the jump. Density plots as proof are provided in Figure 11a,b.
Simulations were extended to the Case 2 scenario. The jump, k * , is monitored after 25% of the sample size, n, is simulated. In other words, k * = 0.25n. For example, if the simulation contains X 1 ( t ) , X 2 ( t ) , …, X 100 ( t ) , then g a ( k / n , t ) ≠ 0 in X 25 ( t ) , …, X 100 ( t ) , and g a ( k / n , t ) = 0 elsewhere. We present density plots for the test statistics: see Figure 12 and Figure 13. As before, two different jump parameter values are used for simulations: a = 0.5 and a = 1. The density plots are compared between asymptotic critical values M q and block bootstrap from M n , β critical values. One can instantly identify that the density plots for the test statistics with both jump parameters (a = 0.5 and a = 1) in the Case 2 scenario tend to have increased dispersion in comparison to the Case 1 scenario. Wider dispersion and flat density plots identify a challenge to recognize the change-point for n ≤ 500 simulations using the critical values M q : see Figure 12a. However, the density plots suggest that change-point identification can be improved if the block bootstrap values from M n , β are used: see Figure 12b. If the jump value is increased to 1, the situation improves as the density plot tweaks to the right side: see Figure 13.
A power analysis is started for the Case 1 scenario. Figure 14 provides power changes for different a values. As before, n = 100 and n = 1000 cases are provided and tested with the block bootstrap critical values from M n , β and the adjusted asymptotic critical values from M q , respectively. As one can see from Figure 14a, parameter a needs to reach 1.2 in order to gain power almost equal to 1 in the n = 100 situation. As for n = 1000, the test can have a lower a value, 0.5, to gain power equal to 1. We continue with the power analysis for Case 2: see Figure 15. The results are similar to those for Case 1 for both n = 100 and n = 1000. The main difference can be seen in Figure 15b, where the desired power value, 0.95, is reached faster when the jump parameter value a is equal to 0.3.
As we mentioned in the previous section, β ( t ) is usually unknown for M n , β . The solution for such a problem requires β ( t ) to be replaced by β ^ ( t ) , and eventually M n , β ^ needs to be used for the testing procedure. First, let us examine if we have significant fluctuations between the theoretical and evaluated β ( t ) . In Figure 16, one can find the theoretical β ( t ) provided in green color. Moreover, β ^ M C ( t ) is provided for every Monte Carlo simulation in grey color, and the averaged value from all Monte Carlo simulations is provided in black color. The plot indicates that, on average, the difference between theoretical and evaluated coefficients is minor. However, some β ^ M C ( t ) variations from the average can be noted. To conclude, in situations where β ( t ) is unknown, one needs to use β ^ ( t ) and calculate M n , β ^ and compare it against M q when n = 1000.
We might encounter a problem when data are n = 100 as we need to calculate block bootstrap values from test statistics under the null hypothesis. In other words, for hypothesis testing the M n , β ^ value needs to be compared against quantile values to compute M n , β ^ , which we define as M n , β ^ under H 0 : see Algorithm 2. For the M n , β ^ calculation, we use the block bootstrap procedure defined in Algorithm 1 with l = 4. Eventually, M n , β ^ is compared with quantile values from M n , β ^ to either reject or accept H 0 in any case scenario for n = 100.
Afterwards, we analyze how the test power is affected if we take β ^ ( t ) instead of β ( t ) . Figure 17 can answer this question. When the jump parameter value was elevated from 0 to 2, power, calculated with M n , β , achieved a value equal to 1 with a = 1.2. For the power generated with M n , β ^ , desired power level is gained with a = 1.7. To conclude, the test power faces minor drawbacks while moving from M n , β to M n , β ^ . However, M n , β ^ manages to provide stable and desired power results.
Size-adjusted power graphs are provided for every scenario: see Figure 18 and Figure 19. Size-adjusted power is the rate at which the null hypothesis is correctly rejected when the used critical value is stable with various n values. What we want to ensure here is that the test power is not affected by size distortions. Again, different a values, a = 0.1, a = 0.5 and a = 1, are used to understand how fast the desired power is reached using the predefined size of simulations when the jump is known and located proportionally to the sample size. For example, 0.2n is 20% of the simulated sample data for the sample size n, and the jump is located at position k * = 0.2n× 0.25. The power is calculated empirically by using Monte Carlo simulations. Moreover, a significance level α = 0.01 is used for power calculation as it manages to provide the desired results.
Figure 18a identifies the Case 1, see Equation (34), scenario for n = 100. The graph implies that a = 0.5 faces power instability until 20% of the sample and increases afterwards. The power reached a value of 0.4, with fluctuations during the increase. The results are better with a = 1: see Figure 18b. In the plot, fluctuations can be seen until 15% of the sample. Later increases are stable, without high fluctuations, and the plot eventually reaches power equal to 0.8 with 100% of the data. The size-adjusted power is calculated for the n = 1000 situation: see the results in Figure 19. Figure 19a provides the size-adjusted power with a jump parameter equal to 0.1. The empirical power reaches a 0.76 power value after 60 % of the sample size. The figure also indicates higher fluctuations than for the n = 100 case. The results improve with a = 0.5, where the empirical power increases to 0.99 after reaching 20% of the sample size. Moreover, the rise of the curve faces only minor fluctuations, as can be seen in Figure 19b.
We continue with the Case 2 scenario and the jump defined by Equation (34). Empirical power results are presented in Figure 20 and Figure 21 for n = 100 and n = 1000, respectively. As before, for n = 100, block bootstrap critical values from M n , β are used for power calculation. The experimental findings indicate that the sample characterized by the jump parameter a = 0.5 yields outcomes with statistical power achieving a value of 0.5 when the entire sample size is employed: see Figure 20b. In contrast, the sample featuring the jump parameter equal to 1 overcomes this power threshold, achieving a power value equal to 0.97 at the end of the sample: see Figure 20a. We resume with the n = 1000 situation, where power was tested using adjusted asymptotic critical values from M q . Figure 21a, simulated using a = 0.1, implies that the empirical power gain value of 0.6 using 40% of the sample size. However, the variation is notable in the graph. For a = 0.5, the results are significantly better: an empirical power of 0.98 is achieved with 25% of the sample size with minor variation.
We perform several additional analyses for Case 2. To begin with, a border analysis is conducted in order to understand overall test capabilities for the presented simulations at the borders of the sample with full period information. Simulations are run with jumps conducted at starting points (0.02n, 0.04n, …, 0.1n) and endpoints (0.9n, 0.92n, …, 0.98n). In these simulations, we use n = 100 for the block bootstrap critical values and n = 1000 for the adjusted asymptotic critical values. The jump parameter, a, is equal to 1 for the n = 100 situation and 0.5 for the n = 1000 situation. For better visual understanding, simulated FAR(1) curves with defined n and a values are presented in Figure 22. A jump is conducted on the k * = 0.02n observation. We can see that although the jump moves the curves upward, most of the grey curves (simulations with jump) overlap with the black curves (simulations with no jump). This means we might encounter issues if conducting an eye test. The jump test is conducted, and the results are presented in Figure 23. What needs to be emphasized is that the test is conducted with full information n. As one can notice, both situations suggest that the test is not sensitive to the jump location. Figure 23a indicates that the desired power is reached from 0.02n with no issues at the end of the interval. Figure 23b implies that power close to 1 is reached from 0.02n to 0.98n with slightly lower power if the jump is monitored at the end.
We continued additional analyses with the offline method. The method retrospectively detects changes when all samples are received. In other words, in this situation, testing is conducted on the partial information k n . For both situations, n = 100 and n = 1000, simulations are fixed with the jump, Equation (34), at X 25 ( t ) and X 250 ( t ) , respectively. Graphical results are presented in Figure 24. For n = 100, one can notice that the first M n , β values reach the red line, the C α value, at n = 30, while after n > 50, a significant number of curves reaches the same red line: see Figure 24a. For n = 1000, the results indicate a similar trend: the M n , β values reach the C α value moments after the jump, at n > 250, and the number of curves over the red line gradually increases while n increases. This implies that for both n = 100 and n = 1000 in the Case 2 scenario, the test provides H 0 rejections for some number of simulations moments after the jump. The number of curves over C α increases faster in relation to k with n = 1000 than with n = 100.
To sum up, the test performed for situations where FAR(1), Equation (31), is generated using Y k ( t ) was a standard normal Wiener process. Two types of scenarios, Case 1 and Case 2, helped to evaluate the test power. Multiple additional analyses were performed for Case 2. Moreover, two types of situations, n = 100 and n = 1000, were used for calculating different test statistics. For the jumps defined in Equation (33) and Equation (34) for various a values, we indicated where the test performed well and where the test encountered issues. Overall, the test provided a positive outcome.

6. Case Study

In order to perform the test on real data, we employed a telecommunications data set. The data set was used by Birbilas and Račkauskas [20], who showed that the FAR(1) model is appropriate for prediction purposes. In this work, we continue the analysis of this data set by proposing a test for the detection of structural changes. The data are taken from a telecommunications company that is operating in southern Europe. The data set contains voice and mobile data daily consumption for approximately 13k subscribers. The total monitored period was 31 months, starting December 2019. Consumption is aggregated by a summation procedure on a monthly basis. In order to move data to a functional space, consumption is converted to functions using b-spline functions with a basis equal to 10. Moreover, consumption is encapsulated to the interval t = [0,1] in order to have the same length during all months.
In order to have better understanding of how smoothing is performed, Figure 25a,b are provided as example cases for 20 randomly selected subscribers. As an example for better understanding of the data, Figure 26a has a whole year’s voice calls in minutes for 20 randomly selected subscribers. The same graph with mobile data consumption is provided in Figure 26b.
Indeed, the graphs identify that subscribers have various consumption patterns. One might notice that a number of jumps can be seen with an eye test despite which metric, voice consumption or mobile data consumption, is used. For further analysis, we will use voice consumption in minutes as a metric. Moreover, we do not find value in testing each of the 13k subscriber for jump patterns, and for further analyses, subscribers had to be clustered. Clustering is done by following these steps:
  • A functional first-order auto-regression, FAR(1), is created for every subscriber. The total number of subscribers is z = 13,862.
  • β ^ s z ( t ) are taken from every FAR(1).
  • β ^ s z ( t ) are clustered into four groups using the Fisher-EM algorithm with k-means initialization. The best cluster is chosen by AIC criteria.
Additional filtering β ^ s z ( t ) ∈ (−1,1) is done in order to discard β ^ s z ( t ) from spurious auto-regressions. After filtering, we created four clusters using the Fisher-EM method. The results with the averaged β ^ s ( t ) for clusters are provided in Figure 27.
Analyses conducted on the clusters provided insights. In the first cluster, the majority of subscribers are old subscribers with stable consumption patterns. Most of them do not change their bundle plans, nor do they have their telecommunication services suspended. Figure 28a provides a better understanding of the average consumption within the first cluster. The second cluster contains subscribers who are frequent add-on users. We conclude that these subscribers tend to exhaust their bundle and purchase add-ons afterwards. This increases usage and creates average fluctuations: see Figure 28b. The third cluster contains more-or-less stable users. The only difference between this group and the first one is that subscribers in the third group tended to change their plans for the better. Telecommunications operators provided attractive offers during some months. This can be identified by increased average consumption in Figure 28c. In the last, fourth, cluster, we find that most of the subscribers are unstable users. For subscribers in this cluster, a stable activity period for several months is accompanied by no activity afterwards. Moreover, a significant number of subscribers within this cluster face service suspension due to insufficient funds during bundle renewal. Figure 28d provides averaged consumption for this cluster.
The jump test is conducted on the created clusters. As one can see, only the first cluster (after excluding the last three months as outliers) is a candidate for no jump, while the second, third, and fourth clusters may even contain several jumps. A point-wise Box–Ljung test suggests (p-value > 0.05) that all residuals of the model are not auto-correlated, as we cannot reject the null hypothesis of independence for model residuals. Moreover, the KPSS test suggests that residual increments are stationary, as we cannot reject the null hypothesis with confidence level α = 0.01. Finally, visual interpretation and the Shapiro–Wilk test suggest that residual increments follow a normal distribution, as we cannot reject the null hypothesis of normality with the same confidence level. By completing the previous steps, we can state that our residual increments follow a Wiener (Brownian motion) process.
The jump test is performed for all four clusters. To have the Case 2 scenario have Equation (34) with μ = 0 and g ( k / n ; a ; t ) 0 , we centered our case study data. As n = 31, we used the blocked bootstrap critical values calculated using Algorithm 1 with the presented data. Although the strategy to identify jumps is the same between clusters, β t ^ ( s ) differed between the clusters: see Figure 27. This led to different block bootstrap quantile values from M n , β ^ statistics and different H 0 rejection critical values. Moreover, for the real-data situation, we used an estimated FAR(1) parameter: β t ^ ( s ) . Eventually, the test implied no jump in the first cluster, see Figure 29a, and jumps in the other clusters, see Figure 29b–d, with the significance levels α = 0.01. The results did match the eye test—the first cluster contains few outliers at the end of the interval and no jump, while the other clusters contain one or several jumps.
To conclude, eye-test assumptions were confirmed by the jump test. The first cluster had no proof to reject H 0 with the defined significance level. All the other clusters had H 0 rejected. The jump test performed well for the case study data.

Author Contributions

Conceptualization, A.R. and A.B.; methodology, A.R. and A.B.; software, A.R. and A.B.; validation, A.R. and A.B.; formal analysis, A.R. and A.B.; investigation, A.R. and A.B.; data curation, A.R. and A.B.; writing—original draft preparation, A.R. and A.B.; visualization, A.R. and A.B.; supervision, A.R. and A.B. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Aue, A.; Horváth, L. Structural breaks in time series. J. Time Ser. Anal. 2013, 34, 1–16. [Google Scholar] [CrossRef]
  2. Zhu, X.; Li, Y.; Liang, C.; Chen, J.; Wu, D. Copula based change-point Detection for Financial Contagion in Chinese Banking. Procedia Comput. Sci. 2013, 17, 619–626. [Google Scholar] [CrossRef]
  3. Garcia, R.; Perron, P. An Analysis of the Real Interest Rate Under Regime Shifts. Rev. Econ. Stat. 1996, 78, 111–125. [Google Scholar] [CrossRef]
  4. Beaulieu, C.; Chen, J.; Sarmiento, J.L. Change-point analysis as a tool to detect abrupt climate variations. Philos. Trans. R. Soc. 2012, 370, 1228–1249. [Google Scholar] [CrossRef]
  5. Ramsay, J.O. When the data are functions. Psychometrika 1982, 47, 379–396. [Google Scholar] [CrossRef]
  6. Kokoszka, P.; Reimherr, M. Introduction to Functional Data Analysis; Chapman & Hall/CRC Texts in Statistical Science; Chapman and Hall/CRC: Boca Raton, FL, USA, 2018. [Google Scholar]
  7. Ramsay, J.O.; Hooker, G.; Graves, S. Functional Data Analysis; Springer Series in Statistics; Springer: Berlin/Heidelberg, Germany, 2009. [Google Scholar]
  8. Ramsay, J.O.; Hooker, G.; Graves, S. Functional Data Analysis with R and MATLAB; Springer: Berlin/Heidelberg, Germany, 2009. [Google Scholar]
  9. Aneiros, G.; Cao, R.; Vilar-Fernandez, J.M.; Munoz-San-Roque, A. Functional Prediction for the Residual Demand in Electricity Spot Markets. Recent Adv. Funct. Data Anal. Relat. Top. 2011, 1, 4201–4208. [Google Scholar] [CrossRef]
  10. Alaya, M.; Ternynck, C.; Dabo-Niang, S.; Chebana, F.; Ouardas, T. Change-point detection of flood events using a functional data framework. Adv. Water Resour. 2020, 137. [Google Scholar] [CrossRef]
  11. Koerner, F.; Anderson, J.; Fincham, J.; Kassa, R. Change-point detection of cognitive states across multiple trials in functional neuroimaging. Stat. Med. 2016, 36, 618–642. [Google Scholar] [CrossRef]
  12. Berkes, I.; Gabrys, R.; Horváth, L.; Kokoszka, P. Detecting changes in the mean of functional observations. J. R. Stat. Soc. Ser. (Stat. Methodol.) 2009, 71, 927–946. [Google Scholar] [CrossRef]
  13. Aue, A.; Gabrys, R.; Horváth, L.; Kokoszka, P. Estimation of a change-point in the mean function of functional data. J. Multivar. Anal. 2009, 100, 2254–2269. [Google Scholar] [CrossRef]
  14. Horváth, L.; Kokoszka, P.; Rice, G. Testing stationarity of functional data. J. Econom. 2014, 179, 66–82. [Google Scholar] [CrossRef]
  15. Horváth, L.; Reeder, R. Detecting changes in functional linear models. J. Multivar. Anal. 2012, 111, 310–334. [Google Scholar] [CrossRef]
  16. Aue, A.; van Delft, A. Testing for stationarity of functional time series in the frequency domain. Ann. Stat. 2020, 48, 2505–2547. [Google Scholar] [CrossRef]
  17. Danielius, T.; Račkauskas, A. p-Variation of CUSUM process and testing change in the mean. Commun. Stat. Simul. Comput. 2020, 52, 1–13. [Google Scholar] [CrossRef]
  18. Aspirot, L.; Bertin, K.; Perera, G. Asymptotic normality of the Nadaraya–Watson estimator for nonstationary functional data and applications to telecommunications. J. Nonparametric Stat. 2009, 21, 535–551. [Google Scholar] [CrossRef]
  19. Yu, Y.; Lambert, D. Fitting Trees to Functional Data, with an Application to Time-of-Day Patterns. J. Comput. Graph. Stat. 1999, 8, 749–762. [Google Scholar] [CrossRef]
  20. Birbilas, A.; Račkauskas, A. Functional modelling of telecommunications data. Math. Model. Anal. 2022, 27, 117–133. [Google Scholar] [CrossRef]
  21. Shields, A.; Doody, P.; Scully, T. Application of multiple change-point detection methods to large urban telecommunication networks. In Proceedings of the 28th Irish Signals and Systems Conference (ISSC), Killarney, Ireland, 20–21 June 2017; pp. 1–6. [Google Scholar] [CrossRef]
  22. Loreh, J. Changepoint Analysis in the Wireless Telecommunications Industry; Colorado School of Mines, ProQuest Dissertations Publishing: Ann Arbor, MI, USA, 2013; Available online: https://hdl.handle.net/11124/80121 (accessed on 1 February 2023).
  23. Aleksiejunas, R.; Garuolis, D. Usage of Published Network Traffic Datasets for Anomaly and Change Point Detection. Wireless Pers. Commun. 2023, 133, 1281–1303. [Google Scholar] [CrossRef]
  24. Račkauskas, A.; Suquet, C. On limit theorems for Banach-space-valued linear processes. Lith. Math. J. 2010, 50, 71–87. [Google Scholar] [CrossRef]
  25. Billingsley, P. Convergence of Probability Measures; John Willey & Sons: Hoboken, NJ, USA, 1999. [Google Scholar]
  26. Kuelbs, J. The invariance principle for Banach space valued random variables. J. Multivar. Anal. 1973, 3, 161–172. [Google Scholar] [CrossRef]
  27. Härdle, W.; Horowitz, J.; Kreiss, J.P. Bootstrap Methods for Time Series. Int. Stat. Rev. 2003, 71, 435–459. [Google Scholar] [CrossRef]
  28. Kunsch, H.R. The jackknife and the bootstrap for general stationary observations. Ann. Stat. 1993, 17, 1217–1261. [Google Scholar] [CrossRef]
  29. Liu, R.Y.; Singh, K. Moving blocks jackknife and bootstrap capture weak dependence. Explor. Limits Bootstrap 1992, 1, 225–248. [Google Scholar]
  30. Nyarige, E.U. The Bootstrap for the Functional Autoregressive Model FAR(1). Ph.D. Thesis, Kaiserslautern Technical University, Kaiserslautern, Germany, 2016. Available online: https://d-nb.info/1106250273/34 (accessed on 1 April 2023).
  31. Haynes, W. Bonferroni Correction. Encycl. Syst. Biol. 2013, 1, 154. [Google Scholar] [CrossRef]
  32. Sedgwick, P. Multiple significance tests: The Bonferroni correction. BMJ 2012, 344, 509. [Google Scholar] [CrossRef]
Figure 1. Kernel (Epanechnikov) density functions of F M , N obtained with different numbers of Monte Carlo simulations: MC = (400, 600, 1000), and N = 1000, M = 1000.
Figure 1. Kernel (Epanechnikov) density functions of F M , N obtained with different numbers of Monte Carlo simulations: MC = (400, 600, 1000), and N = 1000, M = 1000.
Mathematics 12 01889 g001
Figure 2. M n , β under H 0 kernel (Epanechnikov) density versus M q kernel (Epanechnikov) density. Grey dotted line indicates α = 0.01 one-sided quantile, while grey dotted line provides adjusted quantile for α = 0.01.
Figure 2. M n , β under H 0 kernel (Epanechnikov) density versus M q kernel (Epanechnikov) density. Grey dotted line indicates α = 0.01 one-sided quantile, while grey dotted line provides adjusted quantile for α = 0.01.
Mathematics 12 01889 g002
Figure 3. Algorithm 2 kernel (Epanechnikov) densities M n , β ^ , M n , β ^ * , and M n , β ^ versus M q kernel (Epanechnikov) density.
Figure 3. Algorithm 2 kernel (Epanechnikov) densities M n , β ^ , M n , β ^ * , and M n , β ^ versus M q kernel (Epanechnikov) density.
Mathematics 12 01889 g003
Figure 4. M n , β kernel (Epanechnikov) densities under null hypothesis with β 1 ( t ) = b ( t ) , where b = 0.1, b = 0.3, b = 0.5, b = 0.7, b = 0.9.
Figure 4. M n , β kernel (Epanechnikov) densities under null hypothesis with β 1 ( t ) = b ( t ) , where b = 0.1, b = 0.3, b = 0.5, b = 0.7, b = 0.9.
Mathematics 12 01889 g004
Figure 5. M n , β kernel (Epanechnikov) densities under null hypothesis with β 1 (t), β 1 (t), β 3 (t), β 4 (t) and n = 1000.
Figure 5. M n , β kernel (Epanechnikov) densities under null hypothesis with β 1 (t), β 1 (t), β 3 (t), β 4 (t) and n = 1000.
Mathematics 12 01889 g005
Figure 6. M n , β under H 0 densities with n = 100 , n = 200 , n = 100 , n = 1000 versus M q (a) or block bootstrap M n , β (b) densities. All densities estimated using Epanechnikov kernels.
Figure 6. M n , β under H 0 densities with n = 100 , n = 200 , n = 100 , n = 1000 versus M q (a) or block bootstrap M n , β (b) densities. All densities estimated using Epanechnikov kernels.
Mathematics 12 01889 g006
Figure 7. False-positive rate calculations conducted using different values for n , n 200 for adjusted (b) and non-adjusted (a) critical values. Colors stands for significance level α ; hypothesis rejected using M q .
Figure 7. False-positive rate calculations conducted using different values for n , n 200 for adjusted (b) and non-adjusted (a) critical values. Colors stands for significance level α ; hypothesis rejected using M q .
Mathematics 12 01889 g007
Figure 8. False-positive rate using M q critical values (a) and M n , β (b) block bootstrap for different significance levels α .
Figure 8. False-positive rate using M q critical values (a) and M n , β (b) block bootstrap for different significance levels α .
Mathematics 12 01889 g008
Figure 9. Simulations generated using Equation (31) with n = 200 for both Case 1 and Case 2 jumps: (a) Case 1: μ ( a ; t ) = a t ; (b) Case 2: g a ( k / n ; t ) = a t 1 [ k * , n ] ( k ) . Jump parameters for both cases are a = 0 and a = 0.5.
Figure 9. Simulations generated using Equation (31) with n = 200 for both Case 1 and Case 2 jumps: (a) Case 1: μ ( a ; t ) = a t ; (b) Case 2: g a ( k / n ; t ) = a t 1 [ k * , n ] ( k ) . Jump parameters for both cases are a = 0 and a = 0.5.
Mathematics 12 01889 g009
Figure 10. Case 1: H 1 :a = 0.5 densities with n = 100, n = 200, n = 500, n = 1000 versus M q (a) or block bootstrap M n , β (b) densities. All densities estimated using Epanechnikov kernels.
Figure 10. Case 1: H 1 :a = 0.5 densities with n = 100, n = 200, n = 500, n = 1000 versus M q (a) or block bootstrap M n , β (b) densities. All densities estimated using Epanechnikov kernels.
Mathematics 12 01889 g010
Figure 11. Case 1: H 1 :a = 1 densities with n = 100, n = 200, n = 500, n = 1000 versus M q (a) or block bootstrap M n , β (b) densities. All densities estimated using Epanechnikov kernels.
Figure 11. Case 1: H 1 :a = 1 densities with n = 100, n = 200, n = 500, n = 1000 versus M q (a) or block bootstrap M n , β (b) densities. All densities estimated using Epanechnikov kernels.
Mathematics 12 01889 g011
Figure 12. Case 2: H 1 :a = 0.5 densities with n = 100, n = 200, n = 500, n = 1000 versus M q (a) or block bootstrap M n , β (b) densities. All densities estimated using Epanechnikov kernels.
Figure 12. Case 2: H 1 :a = 0.5 densities with n = 100, n = 200, n = 500, n = 1000 versus M q (a) or block bootstrap M n , β (b) densities. All densities estimated using Epanechnikov kernels.
Mathematics 12 01889 g012
Figure 13. Case 2: H 1 :a = 1 densities with n = 100, n = 200, n = 500, n = 1000 versus M q (a) or block bootstrap M n , β (b) densities. All densities estimated using Epanechnikov kernels.
Figure 13. Case 2: H 1 :a = 1 densities with n = 100, n = 200, n = 500, n = 1000 versus M q (a) or block bootstrap M n , β (b) densities. All densities estimated using Epanechnikov kernels.
Mathematics 12 01889 g013
Figure 14. Case 1: Power change with H 1 simulated using different jump parameters a. Power calculated using M q (b) or block bootstrap M n , β (a).
Figure 14. Case 1: Power change with H 1 simulated using different jump parameters a. Power calculated using M q (b) or block bootstrap M n , β (a).
Mathematics 12 01889 g014
Figure 15. Case 2: Power change with H 1 simulated using different jump parameters a. Power calculated using M q (b) or block bootstrap M n , β (a).
Figure 15. Case 2: Power change with H 1 simulated using different jump parameters a. Power calculated using M q (b) or block bootstrap M n , β (a).
Mathematics 12 01889 g015
Figure 16. β ^ M C ( t ) in grey color provides evaluated β ( t ) function from every simulation, while the average of all evaluated β ^ M C ( t ) is in black color. Theoretical β ( t ) in black.
Figure 16. β ^ M C ( t ) in grey color provides evaluated β ( t ) function from every simulation, while the average of all evaluated β ^ M C ( t ) is in black color. Theoretical β ( t ) in black.
Mathematics 12 01889 g016
Figure 17. Simulations performed to understand how jump parameter affects power with α = 0.01. Green color stands for power calculated with theoretical FAR(1) coefficient, M n , β ; black color stands for power evaluated with FAR(1) coefficient M n , β ^ . Jump parameter, a, advanced from 0 to 2.
Figure 17. Simulations performed to understand how jump parameter affects power with α = 0.01. Green color stands for power calculated with theoretical FAR(1) coefficient, M n , β ; black color stands for power evaluated with FAR(1) coefficient M n , β ^ . Jump parameter, a, advanced from 0 to 2.
Mathematics 12 01889 g017
Figure 18. Case 1: Size-adjusted power calculated using using block bootstrap from M n , β with α = 0.01, n = 100. Power with a = 0.5 (a), and power with a = 1 (b). Calculations are done for MC = 500.
Figure 18. Case 1: Size-adjusted power calculated using using block bootstrap from M n , β with α = 0.01, n = 100. Power with a = 0.5 (a), and power with a = 1 (b). Calculations are done for MC = 500.
Mathematics 12 01889 g018
Figure 19. Case 1: Size-adjusted power calculated using using critical values from M q with α = 0.01, n = 1000. Power with a = 0.1 (a), and power with a = 0.5 (b). Calculations are done for MC = 200.
Figure 19. Case 1: Size-adjusted power calculated using using critical values from M q with α = 0.01, n = 1000. Power with a = 0.1 (a), and power with a = 0.5 (b). Calculations are done for MC = 200.
Mathematics 12 01889 g019
Figure 20. Case 2: Size-adjusted power calculated using block bootstrap from M n , β with α = 0.01, n = 100. Power with a = 0.5 (a), and power with a = 1 (b). Calculations are done for MC = 500.
Figure 20. Case 2: Size-adjusted power calculated using block bootstrap from M n , β with α = 0.01, n = 100. Power with a = 0.5 (a), and power with a = 1 (b). Calculations are done for MC = 500.
Mathematics 12 01889 g020
Figure 21. Case 2: Size-adjusted power calculated using using critical values from M q with α = 0.01, n = 1000. Power with a = 0.1 (a), and power with a = 0.5 (b). Calculations are done for MC = 200.
Figure 21. Case 2: Size-adjusted power calculated using using critical values from M q with α = 0.01, n = 1000. Power with a = 0.1 (a), and power with a = 0.5 (b). Calculations are done for MC = 200.
Mathematics 12 01889 g021
Figure 22. Simulation curves generated using Equation (31) with defined n and a values and jump on 0.02n. (a) n = 100 and a = 1; (b) n = 1000 and a = 0.5.
Figure 22. Simulation curves generated using Equation (31) with defined n and a values and jump on 0.02n. (a) n = 100 and a = 1; (b) n = 1000 and a = 0.5.
Mathematics 12 01889 g022
Figure 23. Simulations performed with jumps at the start and end of the interval in order to understand the location effect on power using significance level α = 0.01. (a) n = 100 and a = 1; (b) n = 1000 and a = 0.5.
Figure 23. Simulations performed with jumps at the start and end of the interval in order to understand the location effect on power using significance level α = 0.01. (a) n = 100 and a = 1; (b) n = 1000 and a = 0.5.
Mathematics 12 01889 g023
Figure 24. Total of 500 Monte Carlo simulations performed with jumps 0.25n with significance level α = 0.01. (a) n = 100, a = 1, and (b) n = 1000, a = 0.5.
Figure 24. Total of 500 Monte Carlo simulations performed with jumps 0.25n with significance level α = 0.01. (a) n = 100, a = 1, and (b) n = 1000, a = 0.5.
Mathematics 12 01889 g024
Figure 25. Raw (a) versus smoothed (b) data for 20 randomly selected subscribers. Colors represent different subscribers.
Figure 25. Raw (a) versus smoothed (b) data for 20 randomly selected subscribers. Colors represent different subscribers.
Mathematics 12 01889 g025
Figure 26. Voice call consumption in minutes (a) and mobile data consumption in MB (b) for 20 randomly selected subscribers. Colors represent different subscribers.
Figure 26. Voice call consumption in minutes (a) and mobile data consumption in MB (b) for 20 randomly selected subscribers. Colors represent different subscribers.
Mathematics 12 01889 g026
Figure 27. Clustering results with cluster centers. Legend provides cluster number with total subscribers within the cluster.
Figure 27. Clustering results with cluster centers. Legend provides cluster number with total subscribers within the cluster.
Mathematics 12 01889 g027
Figure 28. Averaged monthly consumption by cluster. (a) First cluster; (b) Second cluster; (c) Third cluster; (d) Fourth cluster.
Figure 28. Averaged monthly consumption by cluster. (a) First cluster; (b) Second cluster; (c) Third cluster; (d) Fourth cluster.
Mathematics 12 01889 g028
Figure 29. Test statistics versus calculated statistics by cluster. (a) First cluster; (b) Second cluster; (c) Third cluster; (d) Fourth cluster.
Figure 29. Test statistics versus calculated statistics by cluster. (a) First cluster; (b) Second cluster; (c) Third cluster; (d) Fourth cluster.
Mathematics 12 01889 g029
Table 1. α quantiles of F M , N , M = N = 1000 .
Table 1. α quantiles of F M , N , M = N = 1000 .
α 0.0050.010.0250.050.250.50.750.950.9750.990.995
MC = 4000.80820.82660.86060.89821.16221.41631.76892.54142.69212.90712.9927
MC = 6000.80620.85620.89650.94121.19571.43581.80702.51882.72423.23693.3202
MC = 8000.76360.81200.87790.94131.18531.43171.75762.45682.68812.94273.1688
MC = 10000.78970.81760.88570.93221.16901.42401.79192.44782.70432.95653.1000
Table 2. α quantiles of F ˜ M , N , M = N = 1000 .
Table 2. α quantiles of F ˜ M , N , M = N = 1000 .
α 0.0050.010.0250.050.250.50.750.950.9750.990.995
MC = 4000.74890.76820.79880.83131.04011.22761.47001.93172.03332.13892.2285
MC = 6000.75090.78090.82140.86491.04251.23011.48601.92902.12132.31262.6485
MC = 8000.75140.77450.82660.87511.04011.22571.46061.91882.06612.20022.4172
MC = 10000.73130.75200.81110.86061.03971.22301.43361.85102.07772.29002.3669
Table 3. Adjusted C α significance level α values for β 1 ( t ) , β 2 ( t ) , β 3 ( t ) , β 4 ( t ) .
Table 3. Adjusted C α significance level α values for β 1 ( t ) , β 2 ( t ) , β 3 ( t ) , β 4 ( t ) .
C α for β 1 C α for β 2 C α for β 3 C α for β 4
α  = 0.013.513.513.543.51
α  = 0.023.413.413.423.41
α  = 0.053.103.153.123.1
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Birbilas, A.; Račkauskas, A. Change-Point Detection in Functional First-Order Auto-Regressive Models. Mathematics 2024, 12, 1889. https://doi.org/10.3390/math12121889

AMA Style

Birbilas A, Račkauskas A. Change-Point Detection in Functional First-Order Auto-Regressive Models. Mathematics. 2024; 12(12):1889. https://doi.org/10.3390/math12121889

Chicago/Turabian Style

Birbilas, Algimantas, and Alfredas Račkauskas. 2024. "Change-Point Detection in Functional First-Order Auto-Regressive Models" Mathematics 12, no. 12: 1889. https://doi.org/10.3390/math12121889

APA Style

Birbilas, A., & Račkauskas, A. (2024). Change-Point Detection in Functional First-Order Auto-Regressive Models. Mathematics, 12(12), 1889. https://doi.org/10.3390/math12121889

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop