Next Article in Journal
Dual-Leverage Effects of Embeddedness and Emission Costs on ESCO Financing: Engineering-Driven Design and Dynamic Decision-Making in Low-Carbon Supply Chains
Previous Article in Journal
Nonlinear Smooth Sliding Mode Control Framework for a Tumor-Immune Dynamical System Under Combined Radio-Chemotherapy
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Change Point Monitoring in Wireless Sensor Networks Under Heavy-Tailed Sequence Environments

1
School of Artificial Intelligence & Computer Science, Xi’an University of Science and Technology, Xi’an 710054, China
2
School of Sciences, Xi’an University of Science and Technology, Xi’an 710054, China
*
Author to whom correspondence should be addressed.
Mathematics 2026, 14(3), 523; https://doi.org/10.3390/math14030523
Submission received: 8 December 2025 / Revised: 29 January 2026 / Accepted: 29 January 2026 / Published: 1 February 2026
(This article belongs to the Section D1: Probability and Statistics)

Abstract

In the special case of a heavy-tailed sequence environment, change point monitoring in wireless sensor networks faces many serious challenges, such as high communication overhead, particularly sensitivity to sparse changes, and dependence on strict parameter assumptions. In order to solve these limitations, a distributed robust M-estimator-based change point monitoring (DRM-CPM) method is proposed. This method combines ratio statistics with sliding window technology so that in online detection, there is no need to know the distribution before and after changes in advance. A threshold-triggered communication strategy is introduced, where sensors exchange local statistics only when exceeding predefined thresholds, significantly reducing energy consumption. By means of theoretical analysis, the asymptotic characteristics of the statistics are confirmed, and the robustness of the algorithm to heavy-tail noise and unknown parameters is also proved. Simulation results show that the algorithm is better than the existing methods in terms of empirical size control, empirical power, and communication efficiency, particularly in the face of sparse variation or heavy-tailed data. This framework provides a scalable solution for real-time anomaly monitoring with non-Gaussian data characteristics in industrial and environmental applications.

1. Introduction

In an era of ubiquitous data, a particularly critical challenge is identifying the dynamic changes in those large datasets [1]. Change point monitoring (CPM), has become a key technology to solve this problem. Acting as a “data radar”, CPM effectively identifies shifts in the data distribution, which can reveal possible changes in the system state, and provide the necessary support for risk assessment and informed decision-making. The theoretical roots of the change point can be traced back to Page [2]’s pioneering work; the utility of the change point has since extended to a myriad of fields, encompassing environmental monitoring [3,4], industrial process control [5], financial risk mitigation [6], and biomedical analysis [7]. It is of great significance to conduct a comprehensive study on CPM methods. Finally, the proposed monitoring procedures are evaluated through a practical application.
For sensor networks, the CPM algorithm has gradually evolved into a variety of deployment strategies to adapt to the needs of different applications. In the early days, related research mainly focused on the centralized sensor network architecture because it has the characteristics of direct global information aggregation, which has attracted relatively large attention. In such a configuration, each sensor node will collect data at each time instance, and then transmit these observed data to the fusion center for analysis so that global change point detection and decision-making can be achieved; as shown in Figure 1a, the fusion center will receive and dispose data streams from all sensors. The point of change refers to the point at which the data distribution of the affected subset of sensors deviates from the original statistical pattern. The fundamentals of centralized CPM algorithm design include constructing valid statistics that aggregate information from multiple sensors and set appropriate thresholds for anomaly identification. Although the centralized method is relatively mature in terms of theoretical analysis and algorithm development, its inherent centralized data transmission mode will bring a lot of communication bottlenecks and energy consumption problems. This problem is very serious in large-scale, high-density sensor networks.
To mitigate the limitations of centralized architectures, researchers have shifted their research focus to decentralized distributed sensor network architectures as shown in Figure 1b, while also studying the corresponding distributed change point detection algorithms (distributed CPM). Distributed CPM eliminates the need for a fusion center by equipping each sensor node with the ability to make autonomous decisions. In a distributed network, individual sensor nodes are tasked with identifying change points using only their own collected data. Simultaneously, they participate in a restricted exchange and collaboration of information with nearby nodes.

1.1. Related Works

Over the past few decades, research on centralized CPM algorithms has built a theoretical foundation that emphasizes the importance of optimizing communication and computation at the same time. Traditional methods typically rely on quantitative techniques [8,9], involving local decision transmission strategies [10,11] to reduce communication requirements. Quantization compresses data by regulating precision in transmitted values, while the local decision method reduces the communication rate through limited information exchange. Kurt [12] introduced the threshold triggering mechanism, changing the communication mode from continuous communication to event-triggered transmission, and the asymptotic optimality can be maintained. This idea was further explored in the work of Yang [13], which presented a detection framework combining moving average and moving sum (MOSUM) methods, achieving performance close to an ideal centralized system without needing prior knowledge of the pre- and post-change mean. In addition to communication efficiency, energy considerations have also begun to receive attention. An energy-harvesting wireless sensor network has been studied [14], which presents new challenges to data collection and processing and affects the performance of CPM. Sun and Zou [15] developed lightweight algorithms for cases of unknown posterior distributions and anonymous networks, using parameter estimation to reduce computational complexity. However, these algorithms are primarily designed for centralized architectures, and when used in large, growing wireless sensor networks, they fall short in terms of robustness, scalability, and energy efficiency.
Early investigations on distributed change point detection generally focus on finding a balance between effective test statistics and ensuring effective communication. Braca and Paolo [16] proposed two distributed CPM frameworks: the Fused Consensus CUSUM (FC-CUSUM) algorithm, which approximates cumulative sum (CUSUM) statistics via a distributed consensus mechanism, and a non-cooperative algorithm that relies on independent local CUSUM decisions to trigger global alerts. From theoretical analysis, it can be seen that when changes have an impact on multiple sensors, the effect of the FC-CUSUM algorithm is relatively good, but when the changes are sparse, its effectiveness will be reduced, and the performance of non-cooperative algorithms is contrary to that of the FC-CUSUM algorithm. In order to solve this problem, Liu et al. [17] introduced the C-CUSUM algorithm, which uses dynamic average local CUSUM statistics to achieve reliable detection; while this approach provides some stability in terms of the number of sensors affected by the change, it requires constant communication at a higher frequency for the consensus process to work.
The increasing prevalence of Internet of Things (IoT) applications has placed a significant emphasis on the need to lower energy consumption associated with communication. This urgency stems from the fundamental constraints of wireless sensor networks, where frequent communication, as highlighted by [18], leads to substantial energy consumption. This ultimately shortens deployment lifespans due to the inherent difficulties in battery replacement, as noted by [19]. To address this problem, Hsu et al. [20] proposed a quantified transmission scheme that reduces the communication load by reducing the accuracy of the data, but this method requires an understanding of the entire network structure, which makes it unsuitable for networks that change over time. Gu et al. [21] integrated a threshold-trigger mechanism to develop the ECO-CUSUM algorithm (a request-response and censoring scheme). While this approach significantly improves energy efficiency when changes affect a large number of sensors, it remains sensitive to sparse variations. Fellouris and Sokolov [22] extended the research scope to include any unknown sensor subset, and Jiao et al. [23] combined the C-CUSUM algorithm to propose the TR-CUSUM algorithm, achieving strong robustness with respect to the number of affected nodes.
A key point to consider is that current algorithms generally rely on strict parameter assumptions. The log-likelihood ratio statistic given in Equation (1), which is the basis of these algorithms, requires precise provisions for density parameters before and after changes. Although historical data can be used to estimate before changes, the inherent unobservability of post-change parameters introduces a risk of detection failure. In order to solve this problem, Liu  et al. [24] proposed a recursive method to estimate the average value. However, this method has to pay a price in terms of statistical efficiency. Although Yang et al. [13] relied on the MOSUM method to eliminate the dependence on the changed parameters, it can only be used in centralized systems:
L t , i = log f 1 ( y t , i ) f 2 ( y t , i ) ,
In reality, the data in the real world is becoming more and more complex, and time series will show very prominent spikes and heavy tails in many practical application scenarios. Such phenomena are common in many fields such as seismology [25], network traffic monitoring [26], material fault detection [27], and mining operations [28]. In detecting change points in autocorrelation coefficients, the ratio statistic method developed by Zhang et al. [29] is noteworthy, especially for its transformation technique toward the mean, which can be adopted in related contexts. In these fields, data modeling based on heavy-tail distribution will have a better effect. However, most prior studies have focused on Gaussian or sub-Gaussian series; extending the analytical framework of time series to heavy-tailed series holds key theoretical significance and can provide more reliable support for risk management and decision-making in related fields.
In order to solve the above challenges, this paper draws inspiration from the framework proposed in [30,31], and then proposes a method for online detection of change points based on ratio statistics. By integrating the sliding window method, the proposed method can carry out the effective real-time monitoring of data flow. No prior information about the distribution before and after the change is required. To adapt to complex real-world scenarios, we incorporate the M-estimation technique, which extends the algorithm’s applicability beyond standard Gaussian or sub-Gaussian sequences to include heavy-tailed sensor data streams. This improvement makes the algorithm significantly improve its robustness when processing scenarios involving sharp tail and heavy-tail data.

1.2. Contributions

In this paper, we propose a distributed change point detection (CPM) algorithm, termed the distributed robust M-estimator-based change point detection (DRM-CPM). The main features of the proposed method are summarized as follows:
  • Synergy of Low Energy and Robustness: In wireless sensor networks, achieving robustness usually comes at the cost of high communication overhead. DRM-CPM breaks this trade-off through a distributed event-triggered architecture. Unlike centralized robust methods that require transmitting raw data, each sensor node i exchanges only local statistical information V t , i with its immediate neighbors j, and strictly only when necessary (when V t , i > η ). This significantly reduces energy consumption by avoiding the quadratic path loss of long-range transmission while maintaining the statistical rigor required for heavy-tailed detection.
  • Robustness to Change Sparsity: A key achievement is the algorithm’s consistent performance across various levels of change sparsity. As demonstrated in the subsequent analysis, the detection remains satisfactory regardless of whether the change is local or global, addressing the challenge of sparse change capture.
  • Unknown Change Parameters: A DRM-CPM online detection algorithm based on ratio-type statistics is developed. The algorithm does not need to know the pre- and post-change distributions. Addressing the problem of change point detection under unknown probability density functions is one of the advantages of the proposed algorithm.
  • Robustness to Heavy-tailed Data: The primary theoretical contribution is breaking the reliance on Gaussian assumptions inherent in traditional distributed detection methods. By generalizing distributed CPM to heavy-tailed sequences via bounded M-estimation, the algorithm resolves the detection failure issue in non-Gaussian environments—a capability that standard energy-efficient protocols typically lack.
  • Detection Accuracy and Robustness: The DRM-CPM algorithm obtains high detection accuracy and shows robust, rapid and precise change point monitoring performance for different kinds of data sequences.

1.3. Organization

The structure of the paper is as follows. Section 2 models the distributed CPM problem. Section 3 elaborates on the proposed algorithm. Section 4 establishes the theoretical properties of the utilized statistics. Section 5 presents the simulation studies. A real-world application using wind speed data is detailed in Section 6. Finally, Section 7 concludes the paper and discusses potential future work.

2. Materials and Methods

Consider a network comprising N sensors, where each sensor observes data y t , i at each time instant t = 1 , 2 , , n , representing dynamically evolving data with specific temporal characteristics. To establish a rigorous foundation for the proposed test, we explicitly define the following three fundamental assumptions: first, all sensor nodes within the network are assumed to maintain time synchronization; second, the anomaly is modeled as a location shift occurring at an unknown time instant [ n τ ] , which affects an unknown and time-varying subset of sensors S t { 1 , , N } while the noise distribution remains stationary; and third, the innovation process ε t , i follows a heavy-tailed mixed sequence characterized by a tail index κ . Consequently, under these explicit conditions, the data model can be represented as follows:
y t , i = μ i + δ t , i I { t > [ n τ ] } + ε t , i , t = 1 , 2 , , n , 1 i N ,
where μ i is the unknown pre-change location parameter, δ t , i is the location shift, and  I { · } is the indicator function. The innovation process ε t , i follows a heavy-tailed mixed sequence with tail index κ . After the time instant [ n τ ] , the location parameter of the S sensor’s observations shifts from μ S to μ S + δ S . Propose the following hypothesis-testing questions:
H 0 : δ t , i = 0 , t > 0 , i [ N ] , H 1 : δ t , i = 0 , t = 1 , 2 , , [ n τ ] , δ t , i 0 , t = [ n τ ] + 1 , , n , i S ,
where the null hypothesis H 0 posits the absence of any change point, while the alternative hypothesis H 1 posits the existence of a single location change point occurring at time [ n τ ] , and the set of affected sensors S is fixed but unknown. When S = N , this means all sensors experience a change, referred to as a global change.
To ensure the asymptotic validity of the proposed test, we impose the following assumptions on the innovation process ε t , i , the M-estimation function ψ M ( · ) , and the sensor network topology.
Assumption 1. 
The distribution F ε of { ε t } lies within the stable domain of attraction, with a tail index κ ( 0 , 2 ] . When κ > 1 , we have E ( ε t ) = 0 . When κ 1 , { ε t } follows a symmetric distribution.
Assumption 2. 
For each sensor i { 1 , , N } , let { ε t , i } t 1 denote the local noise sequence. We assume that { ε t , i } is a strictly stationary α-mixing sequence uniformly over i. Specifically, for some finite χ > 0 and χ > 0 , there exists a constant C ( χ , χ ) > 0 , independent of i, such that
e = 0 ( e + 1 ) 1 / 2 α i ( e ) χ / ( 2 + χ + χ ) C ( χ , χ ) ,
where α i ( e ) ( e = 1 , 2 , ) are the α-mixing coefficients of the i-th sensor. This condition ensures that the dependence between the sequence elements decays sufficiently fast as the time lag increases, and this decay rate is consistent across the entire network.
Assumption 3. 
The M-estimation function ψ M ( · ) satisfies the following conditions:
(1) 
E [ ψ M ( ε t ) ] = 0 ;
(2) 
E [ | ψ M ( ε t ) | 2 + p ] < , for some p > 0 ;
(3) 
0 < E [ ψ M ( ε t ) ] < and for some β > 1 , E [ | ψ M ( ε t ) | β ] < ;
(4) 
0 < σ 2 ( ψ M ) = E [ ψ M 2 ( ε t ) ] + 2 j = 1 E [ ψ M ( ε t ) ψ M ( ε t + j ) ] < .
Assumption 4. 
The influence function ψ M ( · ) is a bounded, non-decreasing, odd function with a unique root at zero, and  ψ M ( · ) is Lipschitz continuous.
Remark 1. 
The boundedness of ψ M (Assumption 4) ensures that the moment conditions in Assumption 3 are satisfied for ψ M ( ε t ) , even if the raw innovations have infinite variance.
Assumption 5. 
In order to improve the detection performance, the sensors communicate with their neighbors via a static undirected graph to exchange information, and then integrate the information with weighted consensus matrix W . If there is a communication link between sensor i and sensor j, generally W satisfies:
(1) 
w i j > 0 (the ( i , j ) -th entry of W is positive);
(2) 
w i j = w j i 0 , indicating that the weight of information flow from sensor j to sensor i is the same as that from sensor i to sensor j;
(3) 
j = 1 N w i j = 1 , for all i, representing a row stochastic normalization.
where denoted by N i is the set of neighbors of sensor i, i.e.,  N i = { j [ N ] : w i j > 0 } .
The goal of the paper is to detect the change at time [ n τ ] with high accuracy, low false alarm rate, small detection delay and low communication cost, where [ n τ ^ ] is the time at which the algorithm issues an alarm. According to the tradition in hypothesis testing, we use two performance metrics to evaluate the algorithm: empirical size (Size) and empirical power (Power). Size is the probability of incorrectly rejecting the null hypothesis H 0 , i.e., the probability of Type I error. Power is the probability of correctly rejecting the null hypothesis H 0 under the alternative hypothesis H 1 so as to avoid Type II error. The estimation is accurate if the alarm time [ n τ ^ ] is in a neighborhood of true change point [ n τ ] .
Size : = P ( Reject H 0 H 0 ) , Power : = P ( Reject H 0 H 1 ) .
To evaluate the detection performance, we use the expected detection delay (EDD), which quantifies the expected time elapsed between the occurrence of the change and the issuance of an alarm:
EDD : = E ( [ n τ ^ ] [ n τ ] ) ,
The communication cost is measured by ζ , defined as the long-term average communication rate per sensor:
ζ = lim n 1 n t = 1 T M t i [ N ] # N i , n > τ ,
where M t denotes the number of communications at time t, and  # N i represents the cardinality of set N.

3. Distributed Change Point Detection Method

To address the aforementioned challenges and reduce the communication overhead, this study proposes a M-estimator-based ratio-type statistic, building upon the statistic introduced in [30,32] and incorporating a sliding window technique; this approach is then extended for the online monitoring of sensor data. This score-based formulation ensures a Brownian motion limit rather than a Stable one, effectively circumventing the bottleneck of unknown tail indices. Following the construction above, the local statistic for each sensor is computed as specified in Equation (3):
V t , i ( s ) = max [ n s ] [ n h ] + 1 [ n s ] [ n s ] + [ n h ] t = [ n s ] [ n h ] + 1 [ n s ] ψ M y t , i μ ^ i ( ψ M ) max [ n s ] [ n h ] + 1 [ n v ] [ n s ] t = [ n s ] [ n h ] + 1 [ n v ] ψ M y t , i μ ^ 1 , i ( ψ M ) + max [ n s ] + 1 [ n v ] [ n s ] + [ n h ] t = [ n v ] + 1 [ n s ] + [ n h ] ψ M y t , i μ ^ 2 , i ( ψ M ) .
where [ n h ] = m / 2 represents half of the sliding window size m, and [ n v ] is a dummy index representing the candidate split points within the current sliding window. The maximization over [ n v ] is performed to capture the local fluctuation of the score function. As new streaming data arrives, the window slides forward by one unit to achieve online monitoring, where μ ^ i ( ψ M ) is the M-estimator of μ , obtained from the observations y [ n s ] [ n h ] + 1 , i , , y [ n s ] + [ n h ] , i :
μ ^ i = argmin β R t = [ n s ] [ n h ] + 1 [ n s ] + [ n h ] ρ ( y t , i β ) ,
where ρ is a loss function such that ρ = ψ M . Equivalently, the M-estimator in the above equation can also be defined as the solution to the following equation:
t = [ n s ] [ n h ] + 1 [ n s ] + [ n h ] ψ M ( y t , i β ) = 0 .
Similarly, μ ^ 1 , i ( ψ M ) and μ ^ 2 , i ( ψ M ) are the M-estimators of the pre-change and post-change location parameter, respectively, computed from the observations y [ n s ] [ n h ] + 1 , i , , y [ n s ] , i and y [ n s ] + 1 , i , , y [ n s ] + [ n h ] , i , respectively, obtained as the solutions to
t = [ n s ] [ n h ] + 1 [ n s ] ψ M ( y t , i β ) = 0 , t = [ n s ] + 1 [ n s ] + [ n h ] ψ M ( y t , i β ) = 0 .
The function ψ M can take various forms. When ψ M ( x ) = x , x R , it reduces to the classical least squares (LS) procedure. If  ψ M ( x ) = x I { | x | K } + K sgn ( x ) I { | x | > K } , it is equivalent to the standard Huber loss function. And  ψ M ( x ) = x I { | x | K } + 0 I { | x | > K } is a more aggressive truncation loss function, where the parameter K is a fixed tuning parameter that determines the robustness and efficiency of the M-estimator.
The preceding content focuses on change point detection implemented at a single sensor. The following section will discuss a collaborative algorithm for distributed wireless sensor networks to achieve energy-efficient and robust online change point detection. To illustrate the communication architecture, a schematic diagram is provided in Figure 2.
Each sensor locally performs a ratio-type test and computes its local ratio statistic V t , i . This paper employs a threshold-based communication strategy to regulate the communication frequency. Specifically, an elevated statistic (i.e., V t , i > η ) signals a potential anomaly, identifying the sensor as a candidate node; collectively, these nodes form the set S t , which estimates the subset of affected nodes at time t:
S t = { i [ N ] : V t , i > η } .
Sensors within S t proactively initiate data exchange requests with all their neighbors to fuse the neighbors’ statistical information. Upon receiving the data, neighboring nodes fuse their own statistical information with the received information to update their monitoring data. Conversely, sensors with local statistic values below the threshold assist their neighbors in completing data exchange tasks. After exchanging data, each sensor updates its decision statistic V ˜ t , i based on the following consensus rule:
V ˜ t , i = j N i w i j V t , j + w i i V t , i , i S t j N i S t w i j V t , j + j N i S t , c w i j V t , i , i S t , c
where S t , c is the complement of S t , and  N i represents the neighborhood of sensor i. The algorithm raises an alarm for sensor i when V ˜ t , i > b . According to Olfati-Saber et al. [33], a distributed consensus is achieved asymptotically on a connected graph, with the convergence rate determined by the algebraic connectivity. Therefore, under Assumptions 3 and 5, the proposed distributed monitoring algorithm is theoretically guaranteed to converge.
In essence, the DRM-CPM algorithm adopts a “local-to-global” detection strategy. First, each sensor performs preliminary local screening to flag potential anomalies. Candidate sensors then share diagnostic information with neighbors to collaboratively validate changes, suppressing random noise through consensus-based fusion. This approach ensures alarms are triggered only upon network-wide agreement, improving both sensitivity and robustness. See Algorithm 1 for details.
Algorithm 1 Robust ratio-type distributed change point detection (DRM-CPM).
Input: 
Sensor data streams { y t , i } , consensus matrix W , window size m, threshold η , and critical value b.
Output: 
Alarm time [ n τ ^ ] and location of change.
  1:
 Online detection:
  2:
 for  t = 1 , 2 , , n   do
  3:
       // Stage 1: Local screening via sliding window
  4:
       for  i [ N ]  do
  5:
             Compute local ratio statistic V t , i based on m and Equation (3).
  6:
       end for
  7:
       Identify candidate set S t = { i [ N ] : V t , i > η } .
  8:
       // Stage 2: Distributed information exchange
  9:
       for each sensor i S t  do
10: 
             Send V t , i to all neighboring nodes j N i .
11: 
       end for
12: 
       for  i [ N ]  do
13: 
             if node i receives V t , j from neighboring nodes then
14: 
                   Retrieve and cache its own local statistic V t , i .
15: 
             end if
16: 
       end for
17: 
       // Stage 3: Global fusion and consensus decision
18: 
       for  i [ N ]  do
19: 
             Update decision statistic V ˜ t , i based on W and Equation (4).
20: 
             if  V ˜ t , i > b  then
21: 
                   Issue an alarm;
22: 
                   return Alarm time [ n τ ^ ] = t .
23: 
             end if
24: 
       end for
25: 
 end for

Complexity Analysis

The computational efficiency of the DRM-CPM algorithm is evaluated per time step t. Specifically, the first stage involves calculating the local ratio statistics { V t , i } for all N sensors based on Equation (3). Since statistics requires performing M-estimation within a sliding window of size m, this stage incurs a time complexity of O ( N · m ) . The subsequent stages, including candidate identification, neighborhood information exchange, and distributed decision updating, have a complexity of O ( | S t | · d m a x + N · d m a x ) , where d m a x denotes the maximum degree of the network graph. Consequently, the total time complexity per time step is expressed as O ( N ( m + d m a x ) ) . Regarding the memory requirements, each sensor only needs to store its local sliding window data and the statistics received from its neighbors, resulting in a per-node storage complexity of O ( m + d m a x ) . This analysis demonstrates that the algorithm is computationally efficient and spatially scalable for large-scale sensor networks.

4. Theoretical Properties

Lemma 1. 
If Assumptions 1–4 hold, and under the null hypothesis H 0 , then
[ n h ] 1 / 2 ( μ ^ ( ψ M ) μ ) d σ ( ψ M ) · [ B ( s + h ) B ( s h ) ] 2 E ( ψ M ( ε t ) ) , [ n h ] 1 / 2 ( μ ^ 1 ( ψ M ) μ ) d σ ( ψ M ) [ B ( s ) B ( s h ) ] h E ( ψ M ( ε t ) ) , [ n h ] 1 / 2 ( μ ^ 2 ( ψ M ) μ ) d σ ( ψ M ) [ B ( s + h ) B ( s ) ] h E ( ψ M ( ε t ) ) .
where B ( · ) denotes a standard Brownian motion.
Theorem 1. 
Under the null hypothesis, assuming y t , i are derived from the model (2) and Assumptions 1–4 hold, as 2 [ n h ] , we have
V n d sup s h s s + h B ( s ) B ( s h ) B ( s + h ) B ( s h ) 2 C + D , C = sup s h + 1 < v s | B ( v ) B ( s h ) v s + h h ( B ( s ) B ( s h ) ) | , D = sup s < ν s + h | B ( s + h ) B ( ν ) s + h ν h ( B ( s + h ) B ( s ) ) | .
Remark 2. 
Theorem 1 establishes that under the null hypothesis, the limiting distribution of V n involves the supremum of Brownian motion functionals. Crucially, due to the use of bounded M-estimation, this limiting distribution is Gaussian based, distinct from the Lévy stable nature of the raw heavy-tailed noise. This implies that the asymptotic behavior is invariant to the specific tail index of the noise. As a consequence, at any fixed significance level, there exists a unique critical value, making the test robust to nuisance parameters and heavy-tailed outliers. This is an appealing feature of the ratio-type test for real data analysis where the tail index of the noise is typically unknown.
The detailed proofs of Lemma 1 and Theorem 1 are provided in Appendix A.

5. Simulations

To evaluate the performance of the proposed algorithm, we conducted experiments on synthetic datasets and provided a comparative analysis against benchmark approaches, specifically MOSUM and TR-CUSUM, in centralized and distributed frameworks, respectively. Additionally, by varying the parameters, including the magnitude of change δ , tail index κ , and window size m, the influence of different parameter settings on the detection performance is investigated. All experimental results, namely, the empirical size (Size), empirical power (Power), expected detection delay (EDD), etc., are based on 2000 Monte Carlo repetitions.
In this experiment, a randomly generated geometric graph is used to generate the distributed sensor network topology. That is, 50 sensor nodes are randomly scattered in a 1 × 1 unit square region, and the communication radius is set to be 0.3 to construct the neighborhood. Following the consensus weight design in [34], the largest connected component is adopted, with all edge weights being uniformly distributed. The affected sensor is randomly selected as the center, and all sensors located within a circular region of radius r = 0.3 centered at this sensor are considered affected. The number of sensors within this affected region ranges approximately from 14 to 17. The following data generation model is used to simulate the data sequences for each sensor node:
y t i = ε t , i , 1 t [ n τ ] , δ i + ε t , i , [ n τ ] + 1 t n ,
where ε t , i are heavy-tailed random variables. Unless otherwise specified, the parameters are set as follows: the threshold parameter K = 1.345 for the Huber function was determined through numerical simulations based on previous studies [35,36] and corresponds to 95% asymptotic efficiency under the standard normal distribution; sample size n = 1000 ; tail index κ = 0.6 , 0.8 , 1.2 , 1.6 , 2 ; magnitude of change δ = 0.5 , 1 , 1.5 ; window size m = 2 [ n h ] = 30 , 40 , 50 , 60 , 70 , 80 , 100 ; and affected region radius r = 0.1 , 0.2 , 0.3 , 0.5 , 0.7 .

5.1. Asymptotic Critical Values

Experiment I determines the threshold and critical value through Monte Carlo simulations. In the TR-CUSUM algorithm, the critical value must be estimated by fitting no-change sample data during the monitoring process, which significantly increases the algorithm’s computational complexity; moreover, the TR-CUSUM algorithm does not provide explicit guidance for selecting the threshold η . In view of this, based on the theoretical proof that the test statistic converges to a functional of Brownian motion under the null hypothesis (Theorem 1), the threshold η and critical value b are set according to the significance level α = 0.1 and α = 0.05 , the corresponding quantile of the limiting distribution, respectively. These significance levels are chosen to rigorously control the false alarm rate while ensuring the detection results maintain standard statistical confidence. In practice, the algorithm is robust to the choice of η within a reasonable range of significance levels, as the global aggregation mechanism effectively compensates for the information filtered by local thresholds. This ensures a stable trade-off between communication efficiency and detection sensitivity without requiring sensitive manual tuning. This approach is consistent with the critical value selection method used in the MOSUM algorithm; however, the threshold η setting in MOSUM remains somewhat subjective. Additionally, since the limiting distribution under the null hypothesis is a function of window size, we investigate the thresholds and critical values under m = 40 , 50 , 60 , 70 . A larger window size m enhances estimation accuracy and reduces false alarms by increasing local samples, while a smaller m minimizes detection delay; detailed experimental analysis of this trade-off is provided in later sections. The specific results are summarized in Table 1.
According to the experimental results presented in Table 1, it can be observed that the proposed change point detection algorithm maintains relatively stable threshold and critical values across different tail index distributions. This observation is highly consistent with the recent findings [37,38], further demonstrating the robustness of the proposed method in handling complex and diverse data environments.

5.2. Performance Evaluation

In this section, the empirical size and power of the proposed distributed robust M-estimator-based change point detection (DRM-CPM) method are compared with those of the benchmark algorithms, MOSUM and TR-CUSUM, to demonstrate the effectiveness of the proposed approach.
Experiment II: Under the null hypothesis H 0 , the empirical size is evaluated by setting the sample size n = 1000 and the window size m = 2 [ n h ] = 40 , 50 , 60 , 70 , and then analyzing the empirical rejection rates of different change point detection methods under heavy-tailed noise. According to the tail index κ , the sequences exhibit different distributional characteristics: when κ = 2 , the sequence exhibits the characteristics of a Gaussian distribution; when κ < 2 , the tail of the sequence becomes heavier as κ decreases, leading to more frequent extreme events. The specific results are presented in Table 2.
The results of the empirical size comparison show that when κ = 2 , TR-CUSUM can maintain a relatively stable empirical size close to the significance level, but when κ < 2 , the empirical size of this method is obviously different from the significance level, which indicates that the method is completely ineffective in controlling the false alarm rate under the heavy-tail distribution. This is mainly because TR-CUSUM statistics rely on the log-likelihood ratio, which requires accurate knowledge of the probability density function and its parameters before and after the change. Although it may be possible to accurately estimate these parameters for normal distribution (e.g., κ = 2 ), for a heavy-tail distribution with infinite variance, accurately estimating these parameters becomes challenging, severely limiting the TR-CUSUM method’s usefulness in heavy-tail scenarios.
When κ = 2 , the MOSUM can still keep a reasonable error control level as shown in the significance level because of the heavy fluctuation of heavy-tail process. While, when κ < 2 , its empirical size is much larger than the significance level, and it is very sensitive to tail index. Since the heavy-tailed data will lead to large fluctuations, it will thus easily cause a false alarm to appear in methods sensitive to extreme values.
In contrast, as shown in Table 2, the empirical size of our proposed DRM-CPM method remains consistently stable around the significance level of 0.05, regardless of the tail index κ . This demonstrates the excellent robustness and error control capability of our method under heavy-tailed noise conditions. Both the experimental results and the theoretical analysis (Theorem 1) support the effectiveness and practical applicability of the DRM-CPM method for online change point detection in heavy-tailed data.
Experiment III is conducted under the alternative hypothesis H 1 , with a sample size of n = 1000 . The primary goal is to investigate and discuss the Power performance of different change point methods for heavy-tailed distributions with varying parameters. As shown in Figure 3a–i, the empirical performance of DRM-CPM, D-LAD, TR-CUSUM and MOSUM under different tail indices κ when δ = 1 , 1.5 , 2 .
From the numerical results, we can see that TR-CUSUM and MOSUM suffer significant failure when κ < 2 , with their empirical power approaching 0. The underlying reasons for this phenomenon have been previously discussed. To further evaluate the performance against robust alternatives, we introduced the Distributed Least Absolute Deviations (D-LAD) algorithm (which minimizes the L 1 norm) as a benchmark. The results indicate that while D-LAD achieves comparable power to DRM-CPM at larger window sizes (e.g., m = 60 ), its performance degrades notably as the window size decreases (e.g., m = 40 ). This decline is attributed to the lower statistical efficiency of the median-based approach in small samples, whereas DRM-CPM maintains higher power by effectively utilizing bounded influence functions. Consequently, DRM-CPM proves to be more suitable for heavy-tailed distributions.
However, when the δ is small, (e.g., as shown in Figure 3a), the detection power of DRM-CPM becomes bad. It may be caused by the fact that its signal strength when small is relatively weak when heavy tailed, which makes it more difficult to be detected. When increased, (Figure 3b,c), the empirical power of DRM-CPM greatly improves. It shows that DRM-CPM has a strong detection ability in structural changes with a larger magnitude of change.
Based on the results of Experiment III, in Experiment IV, we would like to further explore how different methods will affect the EDD and average communication cost. Table 3 shows the EDD and ζ values of different methods under different δ and κ .
From the EDD values in Table 3, when κ = 2 , the EDD values of MOSUM and DRM-CPM are relatively close. However, EDD of TR-CUSUM is much lower than the other two. This is mainly because, compared with MOSUM and DRM-CPM, TR-CUSUM is more sensitive to the structural change of sequence, which leads to the change point appearing faster when the change happens with known pre- and post-change distributions. Regarding the average communication cost ζ , the values for DRM-CPM and TR-CUSUM are essentially consistent, while MOSUM exhibits a notably lower ζ . However, when κ < 2 , the TR-CUSUM method completely fails, resulting in the absence of EDD and ζ values and rendering them without practical significance. Similarly, the MOSUM method nearly fails within this range, and its EDD and ζ values lack practical meaning.
DRM-CPM demonstrates relatively stable EDD and ζ across different magnitudes of change. Nevertheless, the EDD value increases with decreasing δ , indicating that smaller δ increases the difficulty of change point identification, consequently leading to a longer EDD.

5.3. The Numerical Dependency on Parameters

Experiment V aims to compare the performance of different detection methods under the fixed conditions of κ = 2 , δ = 2 , and m = 50 , while varying the number of affected sensors S. Specifically, the proportion of affected sensors, denoted as r, is set to 0.15 , 0.24 , 0.33 , 0.50 , 0.70 , corresponding to 3 , 8 , 14 , 36 , 50 affected sensors, respectively.
As depicted in Figure 4, the solid lines represent the empirical power (Power) of each method, and the dashed lines represent the EDD. The observation results indicate that when the number of affected sensors is very small, the Power of the DRM-CPM method experiences a certain degree of decrease but remains above 0.9 . Additionally, the TR-CUSUM method also exhibits an increase in EDD when the number of affected sensors is small; however, its performance remains relatively stable under other conditions. Contrarily, the EDD of the MOSUM method rises when the number of affected sensors becomes smaller, and the detection power of this method decreases. Remarkably, the EDD of the DRM-CPM method is remarkably stable against the number of affected sensors. Note that in Figure 4, the Y-axis starts from 0.9 to highlight the differences in high-power scenarios, which visually accentuates minor Monte Carlo fluctuations. This demonstrates that DRM-CPM effectively resists variations in spatial scale, maintaining robust detection timeliness in both local failure and global change scenarios.
Since the sliding window size m is a crucial parameter for the DRM-CPM method, Experiment VI examines its impact on detection performance. In this experiment, with δ = 2 , we set m = 30 , 40 , 50 , 60 , 70 , 80 , 100 and examine how the method behaves for different window sizes. The experimental results are shown in Figure 5. The solid lines show the Power for different change point indices κ , and the dashed lines show the corresponding EDD. The convergence of Power values and the linear predictability of EDD collectively confirm the high operational stability of the proposed method under parameter perturbations.
Observations from the figure reveal that when m = 30 , the Power value significantly decreases. This is mainly because an excessively small m fails to extract sufficient effective change information from the data, thereby increasing the false alarm rate. When m 50 , the Power tends to stabilize. Furthermore, the experiment demonstrates that the EDD exhibits a near-linear increasing trend with the increase in m. Practically, selecting m requires balancing statistical stability against detection latency. While m must be sufficient (typically 30 ) to satisfy asymptotic normality, it should be kept minimal to reduce the Average Detection Delay (ADD). We recommend performing an offline calibration using stable historical data to identify the minimum m that maintains the target false alarm rate.

6. Empirical Applications

In this section, the effectiveness of the proposed method is evaluated using daily average wind speed data from Atlantic City International Airport in New Jersey, USA, along with five nearby monitoring stations. Covering a full year from 1 November 2022, to 31 October 2023, the data comprises 365 observations in total (source: NCEI NOAA https://www.ncei.noaa.gov/ (accessed on 20 January 2026). Using the R package stable (version 3.4.2) provided by Nolan [39], we estimate the tail index as κ = 1.705 , which confirms that the wind speed data exhibits notable heavy-tailed behavior.
Three change point detection methods, DRM-CPM, TR-CUSUM, and MOSUM, were applied to monitor the dataset. The results are shown in Figure 6 below. The dashed line represents the original standardized data, highlighting its heavy-tailed nature. The solid line shows the 7-day moving average, which helps visualize shifts in the location over time. The three methods gave different results: DRM-CPM detected a change point at k DRM - CPM = 230 (18 June 2023), TR-CUSUM flagged one much earlier at k TR - CUSUM = 108 (16 February 2023), and MOSUM identified a later change point at k MOSUM = 263 (21 July 2023).
To validate the empirical findings, we cross-referenced the detected change points with the National Climate Report from the National Centers for Environmental Information (NCEI/NOAA). The results are highly consistent with recorded extreme weather events in the region. Specifically, NOAA reports confirm that on 8 June, two significant tornadoes (EF-1 and EF-0) struck the southern part of the study area. These were followed on 16 and 17 June by intense straight-line winds with speeds reaching up to 100 mph in neighboring Pennsylvania. Such localized, high-intensity atmospheric disturbances represent abrupt structural shocks to the wind speed distribution. The ability of DRM-CPM to identify these specific time points—amidst the heavy-tailed noise inherent in storm data—demonstrates its superior sensitivity to “regime shifts” compared to TR-CUSUM and MOSUM, which failed to accurately localize these brief but violent transitions.

7. Conclusions

This paper addresses the change point detection problem in distributed sensor networks and proposes a distributed robust M-estimator-based change point detection (DRM-CPM). The core achievement of this study lies in resolving the detection failure issue in heavy-tailed data scenarios by overcoming the traditional dependence on Gaussian assumptions. The algorithm significantly reduces the communication overhead in wireless sensor networks by introducing an event-triggered communication strategy and Ratio-type test statistics, and achieves robust detection of change points with unknown pre- and post-change distribution parameters and heavy-tailed noise. Theoretical analysis and experimental results demonstrate that DRM-CPM maintains a high detection power under different tail indices κ ( 0 , 2 ] while keeping a low false alarm rate, and its EDD and ζ are superior to existing methods. Compared with TR-CUSUM and MOSUM algorithms, DRM-CPM fulfills the research objectives set out in the introduction by providing a solution for sparse changes in non-Gaussian environments.
However, the current study is not without its limitations. First, when the magnitude of change δ is relatively small, the detection delay tends to increase as data fluctuations become more pronounced, which may affect real-time performance in high-noise scenarios. Second, the selection of the sliding window size m currently relies on a manual trade-off between detection sensitivity and timeliness, lacking a fully automated optimization mechanism.
Building upon these findings, future research will focus on several key directions to enrich the proposed framework. We intend to explore adaptive window adjustment strategies that can dynamically respond to varying network conditions. Furthermore, extending the algorithm to multivariate changes and non-homogeneous network environments will be a priority to enhance its applicability and scientific depth in increasingly complex sensing scenarios. The research findings of this paper provide theoretical support and technical references for online anomaly detection in fields such as industrial monitoring and environmental sensing.

Author Contributions

Conceptualization, H.J.; methodology, H.H.; validation, H.J.; formal analysis, L.W.; investigation, L.W. and H.H.; data curation, H.H.; writing—original draft preparation, H.H.; writing—review and editing, L.W. and H.J.; supervision, H.J.; funding acquisition, H.J. All authors have read and agreed to the published version of the manuscript.

Funding

The authors also thank the financial support from NNSF (Nos.71473194), and SNSF (Nos.2020JM-513).

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

Proof of Lemma 1. 
Assuming that [ n h ] 1 / 2 ( μ μ ^ ( ψ M ) ) = O p ( 1 ) , for [ n h ] , it holds that [ n h ] H 2 h k 2 . In conjunction with
[ n h ] 1 / 2 t = 1 [ n k ] ψ M ( ε t ) h 1 / 2 σ ( ψ M ) · B ( k ) ,
and (2), we obtain
[ n h ] 1 / 2 t = 1 [ n h ] { ψ M ( y t μ ^ ( ψ M ) ) ψ M ( y t μ ) } = O p ( 1 ) .
Since ψ M , the derivative of the convex function ρ , is monotonically increasing, an application of the law of large numbers yields
sup ϕ > H [ n h ] 1 / 2 t = 1 [ n h ] [ ψ M ( y t μ ϕ [ n h ] 1 / 2 ) ψ M ( y t μ ) ] = H E ( ψ M ( ε ) ) + o p ( 1 ) ,
where ε t ( ε t H [ n h ] 1 / 2 , ϵ t ) . This demonstration indicates that by selecting a sufficiently large H, the probability
lim sup n P [ n h ] 1 / 2 ( μ μ ^ ( ψ M ) ) > H ,
can be made arbitrarily small. An analogous argument is applicable to
lim sup n P [ n h ] 1 / 2 ( μ μ ^ ( ψ M ) ) < H ,
This establishes that ( μ μ ^ ( ψ M ) ) = O c ( [ n h ] 1 / 2 ) . Given that [ n h ] 1 / 2 ( μ μ ^ ( ψ M ) ) = O p ( 1 ) , it follows that for any arbitrarily high probability, [ n h ] 1 / 2 ( μ μ ^ ( ψ M ) ) H . Define the process Z n ( u ) as:
Z n ( u ) = t = 1 n { ρ ( ε t + u [ n h ] 1 / 2 ) ρ ( ε t ) } ,
Let u = [ n h ] 1 / 2 ( μ μ ^ ( ψ M ) ) , and assume 2 [ n h ] . Applying a Taylor expansion to the terms within the summation of Z n ( u ) around u = 0 yields
Z n ( u ) = u [ n h ] 1 / 2 t = [ n s ] [ n h ] + 1 [ n s ] + [ n h ] ψ M ( ε t ) + 1 2 u 2 [ n h ] 1 t = [ n s ] [ n h ] + 1 [ n s ] + [ n h ] ψ M ( ε t ) + R n ( u ) ,
where ε t ( ε t , ε t ± | u [ n h ] 1 / 2 | ) , R n ( u ) = 1 2 u 2 [ n h ] 1 { ψ M ( ε t ) ψ M ( ε t ) } . Applying the Lipschitz condition | ψ M ( ε t ) ψ M ( ε t ) | R | u [ n h ] 1 / 2 | , as n : | R n ( u ) | 1 2 u 2 · R | u [ n h ] 1 / 2 | · [ n h ] 1 · 2 [ n h ] = R | u | 3 [ n h ] 1 / 2 p 0 . This confirms that terms of order higher than two vanish in probability.
[ n h ] 1 t = [ n s ] [ n h ] + 1 [ n s ] + [ n h ] | ψ M ( ε t ) ψ M ( ε t ) | R | u [ n h ] 1 / 2 | 0 .
By the law of large numbers (LLN), it follows that
sup | u | M 1 [ n h ] t = [ n s ] [ n h ] + 1 [ n s ] + [ n h ] ψ M ( ε t ) 2 E ( ψ M ( ε t ) ) p 0 .
Combining (A1) and (A2) leads to
Z n ( u ) = u [ n h ] 1 / 2 t = [ n s ] [ n h ] + 1 [ n s ] + [ n h ] ψ M ( ε t ) + 1 2 u 2 [ n h ] 1 t = [ n s ] [ n h ] + 1 [ n s ] + [ n h ] ψ M ( ε t ) = u [ n h ] 1 / 2 t = [ n s ] [ n h ] + 1 [ n s ] + [ n h ] ψ M ( ε t ) + 1 2 u 2 [ n h ] 1 t = [ n s ] [ n h ] + 1 [ n s ] + [ n h ] ( ψ M ( ε t ) ) + o p ( 1 ) ,
Minimizing the objective function defined in (A3) with respect to u, we set the derivative Z n ( u ) to zero:
Z n ( u ) = [ n h ] 1 / 2 t = [ n s ] [ n h ] + 1 [ n s ] + [ n h ] ψ M ( ε t ) + u [ n h ] 1 t = [ n s ] [ n h ] + 1 [ n s ] + [ n h ] E ( ψ M ( ε t ) ) = 0 .
Now, consider the linear term. According to Assumption 3, although the original innovations { ε t } exhibit heavy-tailed characteristics (potentially infinite variance), the transformation via the bounded and antisymmetric score function ψ M ensures that the variable ψ M ( ε t ) possesses finite variance and higher-order moments. Given the mixing conditions specified in Assumption 2, the invariance principle holds. By the Functional Central Limit Theorem (FCLT) for dependent sequences,
[ n h ] 1 / 2 t = [ n s ] [ n h ] + 1 [ n s ] + [ n h ] ψ M ( ε t ) d σ ( ψ M ) · ( B ( s + h ) B ( s h ) ) .
For the second-order term, we need to address the mean reversion of ψ M ( ε t ) . Since u is random, it is necessary to prove that this term satisfies a uniform law of large numbers (ULLN) over a compact set of u. Combining the Lipschitz continuity of ψ M and the condition of finite expectations in Assumption 3, it can be shown that
sup | u | M [ n h ] 1 t = [ n s ] [ n h ] + 1 [ n s ] + [ n h ] ψ M ( ε t ) 2 E ( ψ M ( ε t ) ) P 0 .
According to convex optimization theory, since ρ ( · ) is a convex function, Z n ( u ) is also a convex process with respect to u. Its global minimum can be obtained by solving the first-order derivative equation Z n ( u ) = 0 :
Z n ( u ) = [ n h ] 1 / 2 ψ M ( ε t ) + u [ n h ] 1 ψ M ( ε t ) + o p ( 1 ) = 0 ,
which leads to the asymptotic expression for u:
u = [ n h ] 1 / 2 t = [ n s ] [ n h ] + 1 [ n s ] + [ n h ] ψ M ( ε t ) + o p ( 1 ) [ n h ] 1 t = [ n s ] [ n h ] + 1 [ n s ] + [ n h ] E ( ψ M ( ε t ) ) .
Substituting the convergence results of the random term and the deterministic term, and considering u = [ n h ] 1 / 2 ( μ μ ^ ( ψ M ) ) , we finally obtain
[ n h ] 1 / 2 ( μ ^ ( ψ M ) μ ) d σ ( ψ M ) · ( B ( s + h ) B ( s h ) ) 2 E ( ψ M ( ε t ) ) .
Similarly, the other two terms can be proven:
[ n h ] 1 / 2 ( μ ^ 1 ( ψ M ) μ ) d σ ( ψ M ) ( B ( s ) B ( s h ) ) h E ( ψ M ( ε t ) ) ,
[ n h ] 1 / 2 ( μ ^ 2 ( ψ M ) μ ) d σ ( ψ M ) [ B ( s + h ) B ( s ) ] h E ( ψ M ( ε t ) ) .
The proof of Lemma 1 is now complete. □
Proof of Theorem 1. 
The structure of this proof follows that of Theorem 1.1 in [40]. Applying the mean value theorem yields
ψ M ( y t μ ^ ( ψ M ) ) = ψ M ( ε t + μ μ ^ ( ψ M ) ) = ψ M ( ε t ) + ψ M ( ε t ) ( μ μ ^ ( ψ M ) ) ,
where ε t lies strictly between ε t and ε t + ( μ μ ^ ( ψ M ) ) . To evaluate the partial sum of these localized observations, we perform a first-order expansion of the sum over the left sub-window:
[ n h ] 1 / 2 t = [ n s ] [ n h ] + 1 [ n s ] ψ M ( y t μ ^ ( ψ M ) ) = [ n h ] 1 / 2 t = [ n s ] [ n h ] + 1 [ n s ] ψ M ( ε t ) + [ n h ] 1 / 2 ( μ μ ^ ( ψ M ) ) · 1 n h t = [ n s ] [ n h ] + 1 [ n s ] ψ M ( ε t ) + o p ( 1 ) .
To establish the asymptotic behavior, we invoke the invariance principle for dependent sequences. Under Assumptions 2 (mixing conditions) and 3 (finite moments of bounded ψ M ), the Functional Central Limit Theorem (FCLT) holds (see Herrndorf [41]). Thus, the partial sum of innovations converges weakly to a Brownian motion B ( · ) :
[ n h ] 1 / 2 t = [ n s ] [ n h ] + 1 [ n s ] ψ M ( y t μ ^ ( ψ M ) ) = [ n h ] 1 / 2 t = [ n s ] [ n h ] + 1 [ n s ] ψ M ( ε t ) + [ n h ] 1 / 2 ( μ μ ^ ( ψ M ) ) t = [ n s ] [ n h ] + 1 [ n s ] ψ M ( ε t ) + o p ( 1 ) d h 1 / 2 σ ( ψ M ) [ B ( s ) B ( s h ) ] h 1 / 2 σ ( ψ M ) · B ( h ) E ( ψ M ( ε t ) ) E ( ψ M ( ε t ) ) = h 1 / 2 σ ( ψ M ) · B ( s ) B ( s h ) B ( s + h ) B ( s h ) 2 ,
Next, we control the second term. Using the Lipschitz continuity of ψ M (Assumption 4), we have | ψ M ( ε t ) ψ M ( ε t ) | R | μ μ ^ ( ψ M ) | . The remainder term satisfies
1 n h t ( ψ M ( ε t ) ψ M ( ε t ) ) 1 n h t R | μ μ ^ ( ψ M ) | = R | μ μ ^ ( ψ M ) | = O p ( ( n h ) 1 / 2 ) p 0
Combined with the Weak Law of Large Numbers (WLLN) for mixing processes, which ensures 1 n h ψ M ( ε t ) p E [ ψ M ( ε ) ] , we obtain
1 n h t = [ n s ] [ n h ] + 1 [ n s ] ψ M ( ε t ) p E [ ψ M ( ε ) ]
The derivations for the first and second terms in the denominator (of a preceding expression) are largely analogous; we focus here on the analysis of the second term:
[ n h ] 1 / 2 i = [ n v ] + 1 [ n s ] + [ n h ] ψ M ( y i μ ^ 2 ( ψ M ) ) = [ n h ] 1 / 2 i = [ n v ] + 1 [ n s ] + [ n h ] ψ M ( ε i ) + [ n h ] 1 / 2 ( μ μ ^ 2 ( ψ M ) ) i = [ n v ] + 1 [ n s ] + [ n h ] ψ M ( ε i ) = [ n h ] 1 / 2 i = [ n v ] + 1 [ n s ] + [ n h ] ψ M ( ε i ) + [ n h ] 1 / 2 ( μ μ ^ 2 ( ψ M ) ) 1 n h i = [ n v ] + 1 [ n s ] + [ n h ] ψ M ( ε i ) + o p ( 1 ) = [ n h ] 1 / 2 i = 1 [ n s ] + [ n h ] ψ M ( ε i ) [ n h ] 1 / 2 i = 1 [ n v ] ψ M ( ε i ) + [ n h ] 1 / 2 [ ( μ μ ^ 2 ( ψ M ) ) ] ( [ n s ] + [ n h ] [ n v ] ) E ( ψ M ( ε i ) ) p σ ( ψ M ) B ( s + h ) B ( v ) B ( s + h ) B ( s ) h E ( ψ M ( ε i ) ) ( s + h v ) E ( ψ M ( ε i ) ) ,
where ε i ( ε i , ± [ n h ] 1 / 2 | μ μ ^ 2 ( ψ M ) | ) . Similarly, we can derive
[ n h ] 1 / 2 i = [ n s ] [ n h ] + 1 [ n v ] ψ M ( y i μ ^ 1 ( ψ M ) ) d h 1 / 2 σ ( ψ M ) · B ( v ) B ( s h ) v s + h h ( B ( s ) B ( s h ) ) .
Therefore, combining (A4), (A5) and (A6) yields the limiting distribution of the statistic as
V n d sup s h s s + h B ( s ) B ( s h ) B ( s + h ) B ( s h ) 2 C + D , C = sup s h + 1 < v s | B ( v ) B ( s h ) v s + h h ( B ( s ) B ( s h ) ) | , D = sup s < ν s + h | B ( s + h ) B ( ν ) s + h ν h ( B ( s + h ) B ( s ) ) | .
The proof of Theorem 1 is now complete. □

References

  1. Yu, Z.; Liang, H.; Ye, O.; Zhang, Y. ESOD-YOLOv8: Small object detection enhanced with auto-disturbance rejection convolution. Expert Syst. Appl. 2025, 296, 129046. [Google Scholar] [CrossRef]
  2. Page, E.S. Continuous inspection schemes. Biometrika 1954, 41, 100–115. [Google Scholar] [CrossRef]
  3. Hemraj, D.A.; Carstensen, J. Towards ecosystem-based techniques for tipping point detection. Biol. Rev. 2025, 100, 892–919. [Google Scholar] [CrossRef]
  4. Moradi, M.; Montesino-SanMartin, M.; Ugarte, M.D.; Militino, A.F. Locally adaptive change-point detection (LACPD) with applications to environmental changes. Stoch. Environ. Res. Risk Assess. 2022, 36, 251–269. [Google Scholar] [CrossRef]
  5. Wu, Z.; Li, Y.; Hu, L. A synchronous multiple change-point detecting method for manufacturing process. Comput. Ind. Eng. 2022, 169, 108114. [Google Scholar] [CrossRef]
  6. Li, M.; Gao, L.; Lv, G.; Tong, X. Random change point model with an application to the China Household Finance Survey. Sci. China Math. 2024, 67, 2373–2386. [Google Scholar] [CrossRef]
  7. Jiang, F.; Zhao, Z.; Shao, X. Time series analysis of COVID-19 infection curve: A change-point perspective. J. Econom. 2023, 232, 1–17. [Google Scholar] [CrossRef]
  8. Mei, Y. Information bounds and quickest change detection in decentralized decision systems. IEEE Trans. Inf. Theory 2005, 51, 2669–2681. [Google Scholar] [CrossRef]
  9. Tabella, G.; Ciuonzo, D.; Yilmaz, Y.; Wang, X.; Rossi, P.S. Time-aware distributed sequential detection of gas dispersion via wireless sensor networks. IEEE Trans. Signal Inf. Process. Netw. 2023, 9, 721–735. [Google Scholar] [CrossRef]
  10. Quarisa, M.W.; Blostein, S.D. Change Detection for Large Distributed Sensor Networks with Multitriggered Local Sensors. IEEE Access 2022, 10, 44789–44803. [Google Scholar] [CrossRef]
  11. Ciuonzo, D.; Salvo Rossi, P. DECHADE: Detecting slight changes with hard decisions in wireless sensor networks. Int. J. Gen. Syst. 2018, 47, 535–548. [Google Scholar] [CrossRef]
  12. Kurt, M.N.; Yılmaz, Y.; Wang, X. Distributed quickest detection of cyber-attacks in smart grid. IEEE Trans. Inf. Forensics Secur. 2018, 13, 2015–2030. [Google Scholar] [CrossRef]
  13. Yang, Z.; Eckley, I.A.; Fearnhead, P. A communication-efficient, online changepoint detection method for monitoring distributed sensor networks. Stat. Comput. 2024, 34, 115. [Google Scholar] [CrossRef]
  14. Biswas, S.; Dey, S. Asymptotic performance analysis of distributed non-Bayesian quickest change detection with energy harvesting sensors. IEEE Trans. Aerosp. Electron. Syst. 2022, 58, 3697–3707. [Google Scholar] [CrossRef]
  15. Sun, Z.; Zou, S.; Zhang, R.; Li, Q. Quickest change detection in anonymous heterogeneous sensor networks. IEEE Trans. Signal Process. 2022, 70, 1041–1055. [Google Scholar] [CrossRef]
  16. Braca, P.; Marano, S.; Matta, V.; Willett, P. Consensus-based Page’s test in sensor networks. Signal Process. 2011, 91, 919–930. [Google Scholar] [CrossRef]
  17. Liu, Q.; Zhang, R.; Xie, Y. Distributed change detection via average consensus over networks. In Proceedings of the Stochastic Models, Statistics and Their Applications, Dresden, Germany, 6–8 March 2019; Springer: Cham, Switzerland, 2019; pp. 177–192. [Google Scholar] [CrossRef]
  18. Chen, F.; Guo, L.; Chen, C. A survey on energy management in the wireless sensor networks. IERI Procedia 2012, 3, 60–66. [Google Scholar] [CrossRef]
  19. Meena, S.B.; Patil, P.R.; Kandharkar, S.; Hemalatha, N.; Khade, A.; Dixit, K.K.; Chinthamu, N. The Evolution of Smart Grid Technologies: Integrating Renewable Energy Sources, Energy Storage, and Demand Response Systems for Efficient Energy Distribution. Nanotechnol. Perceptions 2024, 20, 1098–1109. [Google Scholar] [CrossRef]
  20. Hsu, M.X.; Huang, Y.C.; Li, W.H.; Chu, C.F.; Wu, P.J. An efficient algorithm and quantization for fully distributed sequential change detection. IEEE Trans. Commun. 2023, 71, 6088–6099. [Google Scholar] [CrossRef]
  21. Gu, Y.; Jiao, Y.; Xu, X.; Yu, Q. Request–response and censoring-based energy-efficient decentralized change-point detection with IoT applications. IEEE Internet Things J. 2020, 8, 6771–6788. [Google Scholar] [CrossRef]
  22. Fellouris, G.; Sokolov, G. Second-order asymptotic optimality in multisensor sequential change detection. IEEE Trans. Inf. Theory 2016, 62, 3662–3675. [Google Scholar] [CrossRef]
  23. Jiao, Y.; Meng, L.; Li, Y.; Yu, Q.; Gu, Y. Efficient Change-Point Detection Over Fully Decentralized Wireless Networks with Low Communication Rate. IEEE Trans. Veh. Technol. 2024, 74, 1626–1642. [Google Scholar] [CrossRef]
  24. Liu, K.; Zhang, R.; Mei, Y. Scalable sum-shrinkage schemes for distributed monitoring large-scale data streams. Stat. Sin. 2019, 29, 1–22. [Google Scholar] [CrossRef]
  25. Dutfoy, A. Estimation of tail distribution of the annual maximum earthquake magnitude using extreme value theory. Pure Appl. Geophys. 2019, 176, 527–540. [Google Scholar] [CrossRef]
  26. Kim, G. Holistic In-Network Acceleration for Heavy-Tailed Storage Workloads. IEEE Access 2023, 11, 77416–77428. [Google Scholar] [CrossRef]
  27. Arslan, S.S.; Zeydan, E. On the distribution modeling of heavy-tailed disk failure lifetime in big data centers. IEEE Trans. Reliab. 2020, 70, 507–524. [Google Scholar] [CrossRef]
  28. Reyes, J.; Rojas, M.A.; Venegas, O.; Gómez, H.W. Nakagami distribution with heavy tails and applications to mining engineering data. J. Stat. Theory Pract. 2020, 14, 55. [Google Scholar] [CrossRef]
  29. Zhang, X.; Jin, H.; Yang, Y. Quasi-autocorrelation coefficient change test of heavy-tailed sequences based on M-estimation. AIMS Math. 2024, 9, 19569–19596. [Google Scholar] [CrossRef]
  30. Peštová, B.; Pešta, M. Abrupt change in mean using block bootstrap and avoiding variance estimation. Comput. Stat. 2018, 33, 413–441. [Google Scholar] [CrossRef]
  31. Jin, H.; Tian, S.; Hu, J.; Zhu, L.; Zhang, S. Robust ratio-typed test for location change under strong mixing heavy-tailed time series model. Commun. Stat.-Theory Methods 2025, 54, 5760–5783. [Google Scholar] [CrossRef]
  32. Jin, H.; Hu, J.; Zhu, L.; Tian, S.; Zhang, S. M-procedures robust to structural changes detection under strong mixing heavy-tailed time series models. J. Stat. Plan. Inference 2025, 239, 106295. [Google Scholar] [CrossRef]
  33. Olfati-Saber, R.; Fax, J.A.; Murray, R.M. Consensus and cooperation in networked multi-agent systems. Proc. IEEE 2007, 95, 215–233. [Google Scholar] [CrossRef]
  34. Boyd, S.; Diaconis, P.; Xiao, L. Fastest mixing Markov chain on a graph. SIAM Rev. 2004, 46, 667–689. [Google Scholar] [CrossRef]
  35. Qin, R.; Liu, W. A robust test for mean change in dependent observations. J. Inequalities Appl. 2015, 2015, 48. [Google Scholar] [CrossRef]
  36. Hušková, M.; Picek, J. Bootstrap in detection of changes in linear regression. Sankhyā Indian J. Stat. 2005, 67, 200–226. [Google Scholar] [CrossRef]
  37. Zhang, S.; Jin, H.; Su, M. Modified block bootstrap testing for persistence change in infinite variance observations. Mathematics 2024, 12, 258. [Google Scholar] [CrossRef]
  38. Jin, H.; Wang, A.; Zhang, S.; Liu, J. Subsampling ratio tests for structural changes in time series with heavy-tailed AR (p) errors. Commun. Stat.-Simul. Comput. 2024, 53, 3721–3747. [Google Scholar] [CrossRef]
  39. Nolan, J.P. Numerical calculation of stable densities and distribution functions. Commun. Statistics. Stoch. Model. 1997, 13, 759–774. [Google Scholar] [CrossRef]
  40. Horváth, L.; Horváth, Z.; Hušková, M. Ratio tests for change point detection. In Beyond Parametrics in Interdisciplinary Research: Festschrift in Honor of Professor Pranab K. Sen; Institute of Mathematical Statistics: Beachwood, OH, USA, 2008; Volume 1, pp. 293–305. [Google Scholar] [CrossRef]
  41. Herrndorf, N. A Functional Central Limit Theorem for Weakly Dependent Sequences of Random Variables. Ann. Probab. 1984, 12, 141–153. [Google Scholar] [CrossRef]
Figure 1. Illustration of the CPM problem for under different network topologies.
Figure 1. Illustration of the CPM problem for under different network topologies.
Mathematics 14 00523 g001
Figure 2. Schematic diagram of DRM-CPM communication architecture.
Figure 2. Schematic diagram of DRM-CPM communication architecture.
Mathematics 14 00523 g002
Figure 3. Power under varying magnitude of change δ and tail index κ .
Figure 3. Power under varying magnitude of change δ and tail index κ .
Mathematics 14 00523 g003
Figure 4. Empirical power and EDD versus number of affected nodes S.
Figure 4. Empirical power and EDD versus number of affected nodes S.
Mathematics 14 00523 g004
Figure 5. The influence of window sizes.
Figure 5. The influence of window sizes.
Mathematics 14 00523 g005
Figure 6. Atlantic City International Airportś daily average standardized wind speed from 1 November 2022 to 31 October 2023.
Figure 6. Atlantic City International Airportś daily average standardized wind speed from 1 November 2022 to 31 October 2023.
Mathematics 14 00523 g006
Table 1. Thresholds and critical values.
Table 1. Thresholds and critical values.
κ m = 40 m = 50 m = 60 m = 70
η b η b η b η b
2.01.91352.06241.83361.95971.77641.89171.75531.8837
1.61.90102.06851.81771.94111.76541.88011.75871.8776
1.21.90812.04381.82921.94401.77281.89621.74961.8794
0.81.88492.00681.80471.94931.76241.87301.75321.8719
0.41.92092.08221.79371.94641.74841.87981.74311.8848
Table 2. Empirical sizes under H 0 .
Table 2. Empirical sizes under H 0 .
κ m = 40 m = 50
DRM-CPMTR-CUSUMMOSUMDRM-textbfTR-CUSUMMOSUM
2.00.08180.04340.02180.05070.04340.0365
1.60.06830.09470.06610.2471
1.20.05380.10760.03690.1360
0.80.06250.38640.02070.4625
0.60.09820.31630.02540.2125
κ m = 60 m = 70
DRM-CPMTR-CUSUMMOSUMDRM-CPMTR-CUSUMMOSUM
2.00.04560.04340.01310.02050.04340.0157
1.60.03180.10870.03170.1952
1.20.02410.23920.01930.3164
0.80.03370.10970.01460.3631
0.60.01840.27380.00930.2065
Table 3. EDD and ζ under H 1 ( m = 50 ).
Table 3. EDD and ζ under H 1 ( m = 50 ).
δ kEDD ζ
DRM-CPM R-CUSUM MOSUM DRM-CPM TR-CUSUM MOSUM
2.024.0610.6723.860.27380.22510.0758
1.623.8136.720.28090.1010
11.224.2219.090.29020.1792
0.824.5120.640.30210.2117
0.625.3117.110.32770.2153
2.019.965.2122.570.25330.22810.0531
1.619.8814.950.25770.0932
1.51.220.1925.270.25880.1076
0.821.0716.590.26340.1645
0.621.097.210.28640.1032
2.018.973.4125.970.25480.22890.0493
1.618.6637.820.25010.1082
21.218.5935.080.25110.0625
0.818.6420.630.25230.0994
0.618.839.580.25530.1133
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, L.; Hu, H.; Jin, H. Change Point Monitoring in Wireless Sensor Networks Under Heavy-Tailed Sequence Environments. Mathematics 2026, 14, 523. https://doi.org/10.3390/math14030523

AMA Style

Wang L, Hu H, Jin H. Change Point Monitoring in Wireless Sensor Networks Under Heavy-Tailed Sequence Environments. Mathematics. 2026; 14(3):523. https://doi.org/10.3390/math14030523

Chicago/Turabian Style

Wang, Liwen, Hongbo Hu, and Hao Jin. 2026. "Change Point Monitoring in Wireless Sensor Networks Under Heavy-Tailed Sequence Environments" Mathematics 14, no. 3: 523. https://doi.org/10.3390/math14030523

APA Style

Wang, L., Hu, H., & Jin, H. (2026). Change Point Monitoring in Wireless Sensor Networks Under Heavy-Tailed Sequence Environments. Mathematics, 14(3), 523. https://doi.org/10.3390/math14030523

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop