Next Article in Journal
Utilizing the Honeybees Mating-Inspired Firefly Algorithm to Extract Parameters of the Wind Speed Weibull Model
Previous Article in Journal
Forecasting Convective Storms Trajectory and Intensity by Neural Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Forecasting and Anomaly Detection in BEWS: Comparative Study of Theta, Croston, and Prophet Algorithms

by
Aleksandr N. Grekov
,
Elena V. Vyshkvarkova
* and
Aleksandr S. Mavrin
Institute of Natural and Technical Systems, 299011 Sevastopol, Russia
*
Author to whom correspondence should be addressed.
Forecasting 2024, 6(2), 343-356; https://doi.org/10.3390/forecast6020019
Submission received: 10 April 2024 / Revised: 15 May 2024 / Accepted: 20 May 2024 / Published: 21 May 2024

Abstract

:
Evaluation of water quality and accurate prediction of water pollution indicators are key components in water resource management and water pollution control. The use of biological early warning systems (BEWS), in which living organisms are used as biosensors, allows for a comprehensive assessment of the aquatic environment state and a timely response in the event of an emergency. In this paper, we examine three machine learning algorithms (Theta, Croston and Prophet) to forecast bivalves’ activity data obtained from the BEWS developed by the authors. An algorithm for anomalies detection in bivalves’ activity data was developed. Our results showed that for one of the anomalies, Prophet was the best method, and for the other two, the anomaly detection time did not differ between the methods. A comparison of methods in terms of computational speed showed the advantage of the Croston method. This anomaly detection algorithm can be effectively incorporated into the software of biological early warning systems, facilitating rapid responses to changes in the aquatic environment.

1. Introduction

Water, a vital resource for humanity, faces increasing demand due to population growth and economic development. However, the problem is not just the increasing demand for water; poor water quality also makes it unsuitable for many purposes [1]. Therefore, it is important to ensure not only the quantity but also the quality of water to prevent shortages. Thus, the sixth sustainable development goal of the United Nations [2] aims to ensure the availability of water and sanitation, as well as the sustainable management of these resources. Given the presence of more than 350,000 chemicals and their combinations in the world [3], water pollution stands out as one of the key environmental problems we face [4]. This directly affects human health, especially in the context of providing safe drinking water. Given the complexity and diversity of chemicals, analyzing each of them with high frequency is expensive and practically impossible. An alternative to such analysis is the use of living organisms as bioindicators, which provides a comprehensive assessment of the state of the aquatic environment. Biological early warning systems, known as BEWS [5], have become widespread in recent decades in assessing the quality of water resources and activating alarms in emergency situations [6]. Freshwater mussels are a widespread and relatively long-lived group of aquatic organisms [7]. Mussels are considered effective indicators of environmental change, responding to both long-term and acute changes caused by stressors [8]. Behavioral characteristics of mussels, such as valves opening, reflect their natural circadian rhythms, feeding and breathing activity, and can also serve as an indicator of external stress factors [9]. Assessing anomalies from the effects of stress factors using experts is possible, but this approach does not provide a quick response to emergency situations. The key challenge is still to effectively detect abnormal situations and to trigger alarms in a timely manner.
Time series anomaly detection models use a calculated numerical metric, known as an anomaly score, to determine the status of each data point as an anomaly or normal value [10]. The anomaly score is determined based on the difference between the predicted and actual values. If the score value exceeds a set threshold, the data point is classified as an anomaly [11]. In some situations, a static threshold may be applicable, but often, this approach is unsuitable, due to the variability of the variance and the mathematical expectation of the time series.
Machine learning algorithms allow the building of models to predict data with seasonal variability. Algorithms are widely used in making business forecasts, for example, forecasting purchasing power [12], product prices [13,14], and others. In the last decade, algorithms began to be used to predict climatic, environmental, and biological processes. Solving time series forecasting problems is possible using different types of machine learning algorithms, for example, forecasting algorithms, algorithms based on Bayesian statistics, and others. The Prophet forecasting model was used to predict both short-term and long-term air pollution in Seoul, Korea and China [15,16]. Using sensor data from a wastewater treatment plant system, Kramar and Alchakov [17] applied various machine learning methods (seasonal autoregressive integrated moving average SARIMA, Holt–Winters exponential smoothing, ETS, Facebook Prophet, extreme gradient boosting XGBoost, and long short-term memory) for time series forecasting. The XGBoost method turned out to be the best method for their purposes. An important direction for future research is to explore how these approaches can better support forecasting in practice [18]. A Bayesian approach is used to predict the possibility of large earthquakes using thermal anomalies from satellite observations [19].
In a previous work, we investigated the ability to detect anomalies in bivalves’ activity data using four unsupervised machine learning algorithms (elliptic envelope, isolation forest (iForest), one-class support vector machine (SVM), and local outlier level (LOF) [20], the autoregressive integrated moving average (ARIMA) forecast model with a seasonal component in the same data [21], obtained by biological early warning system [22]). In this paper, we further analyzed machine learning forecasting algorithms for identifying anomalies and generating alarms in bivalves’ activity data. The main parameter for the success of forecasting the developed models will be the speed of identifying an anomaly, since data from the BEWS is sent to the server in real time, and the algorithm is planned to be implemented in the software of BEWS. In addition, the developed algorithm for detecting an anomaly and subsequently generating an alarm signal should not require significant computing resources. The novelty of the work lies in the application of machine learning algorithms for activity data of bivalve mollusks used as biosensors in the biomonitoring system of water bodies. Previously, Valletta et al. [23] and Bertolini et al. [24] demonstrated the feasibility of using machine learning algorithms in animal behavior studies, particularly to identify consistent behavioral patterns in the activity of the bivalves Mytilus galloprovincialis and Mytilus edulis. Meyer et al. [25] combined statistical-analysis techniques with machine learning approaches to study the prediction of the movement patterns of ruminants. However, to the best of our knowledge, no study has applied machine learning algorithms to forecast bivalves’ activity data.
Section 2 describes data, three machine learning algorithms, and data preprocessing. The block diagram of the anomaly detection algorithm is presented. Section 3 provides the main results. The graphical results of choosing a fixed RMSE threshold are given. In Section 4, we discuss results and compare the anomaly detection time obtained by different methods. Finally, Section 4 summarizes the findings and outlines their scope.

2. Materials and Methods

2.1. Data

Measured activity data from the freshwater bivalve Unio pictorum (Linnaeus, 1758) collected between March and April 2017 were used to create forecasting models. The data were obtained using the BEWS developed by the authors, based on recording the behavioral reactions of bivalve mollusks [22]. The BEWS consisted of surface and underwater parts (Figure 1). The system automatically, in real time, provided digital information about the mollusks’ valve movements using Hall sensors and a magnet attached to them (Figure 1b). Individual reactions of 16 animals were recorded simultaneously. The sensitivity of the device was 0.1 mm. Continuous retrieval of information, with the registration of signals, was carried out every 10 s. The synchronous reaction of a group of mollusks (at least 70% of the total number) was an alarm signal. Computer programs allowed the real-time transmission of information in on-line mode. The latent period of the system response to pollution, depending on the nature and concentration, ranged from several seconds to several minutes. The synchronized closing of mollusk valves indicated the occurrence of a stressful situation, in particular, pollution of the aquatic environment [26,27]. In addition, natural factors, such as changes in environmental characteristics, may be the reasons for the reactions of mollusks. In this regard, in addition to the opening amplitude of the valves, the BEWS also measured water temperature and light level.
The biomonitoring complex was deployed at hydroelectric complex No. 14 of the Chernaya River in the city of Sevastopol.

2.2. Machine Learning Algorithms

The following machine learning algorithms were the ones we applied in the investigation presented in this article.

2.2.1. Theta Method

The Theta method, described in [28], is a variant of simple exponential smoothing (SES) that takes into account drift (as shown in [29]). The Theta method is applied to non-seasonal or deseasonalized time series. Multiplicative classical decomposition is often used for deseasonalization. The essence of the method is to decompose the original time series into two new lines using the so-called theta coefficients. The advantage of this approach is the ability to use information that usually cannot be fully taken into account and modeled by extrapolating the original time series [30]. Theta lines can be thought of as separate time series that are extrapolated independently using an appropriate forecasting method.
The Theta method is highly flexible: the number of theta lines, theta coefficients, and extrapolation methods may vary, and they can be combined to obtain reliable forecasts. However, Assimakopoulos and Nikolopoulos [28] proposed a simplified version of the method that uses only two theta lines with predefined coefficients θ. The first theta line with the coefficient θ1 = 0 is extrapolated over time using a linear regression model, while for the second theta line with the coefficient θ2 = 2, the simple exponential smoothing method is applied. The final predictions are generated by combining the predictions of both theta lines with equal weights.

2.2.2. Croston Method

Croston’s method [31] is a modification of simple exponential smoothing (SES), specifically designed for working with discontinuous time series and described by the following equation:
y t = o t z t
where yt is the actual observation, ot is a Bernoulli distributed binary variable that takes a value of one when demand occurs and is zero otherwise, zt is the potential quantity of demand with a conditional distribution (becomes real only in those moments when ot =1), and t is the observation time.
Croston proposes dividing the model (1) into two parts, working with each of them separately. This allows us to estimate the probability of demand occurring using intervals between demands [32]. Thus, instead of one series t = 1, …, T, we have two series: demand intervals q j t and demand sizes z j t , where jt = 1, …, N reflects the successive numbers of intervals and demand sizes, and N is the number of non-zero demands. If q j t represents the time since the last non-zero observation and reflects the demand interval, then it serves as an indicator for the next non-zero observation. Croston [31] assumes that the probability of occurrence is constant between non-zero demands, while average demand sizes are assumed to be the same during zero demands. In this method, both demand sizes z j t and demand intervals q j t are forecast using SES [33], resulting in the following system of equations:
y ^ t = y ^ j t = 1 q ^ j t z ^ j t z ^ j t = α z z j t 1 + ( 1 α z ) z ^ j t 1 q ^ j t = α q q j t 1 + ( 1 + α q ) q ^ j t 1 j t = j t 1 + o t
where y ^ j t is the projected average demand, z ^ j t is the projected demand size, q ^ j t is the forecast demand interval, and αq and αz are the smoothing options for intervals and sizes, respectively. The system of Equation (2) shows how each observation at a certain time t is transformed into the corresponding jt element of the size and demand interval. Croston’s original formulation assumed that αq = αz, but later, Schultz [34] proposed using separate smoothing parameters, which was supported by other researchers [35,36].
However, the Croston method has a significant drawback: the forecast is updated only after periods with positive demand. This means that the method becomes less effective after successive periods of zero demand. Therefore, it cannot be used to assess the risk of product obsolescence or to manage excess inventory [37].

2.2.3. Prophet Method

Prophet is a method for forecasting time series data based on an additive model in which non-linear trends correspond to annual, weekly, and daily seasonality, as well as anomaly effects. It works best with time series that have strong seasonal effects and multiple seasons of historical data. The method is robust to missing data and shifts in trends and generally handles outliers well.
Prophet is an open-source time series regression algorithm originally designed to predict various Facebook data, such as platform usage and growth, and can be used in both Python and R [38].
The proposed approach to time series forecasting uses a decomposable model based on the work of Harvey and Peters [39]. The model consists of three main components: trend, seasonality, and holidays [38]:
y t = g t + s t + h t + e ( t )
where y(t) is a value of a time series at a point in time t, g(t) is a trend component taking into account non-periodic changes, s(t) is a seasonal component reflecting periodic changes, h(t) represents the consequences of the holidays, and e(t) is an error term t, which represents any idiosyncratic changes that are not taken into account by the model.
The Prophet algorithm was initially developed for business time series data with annual, seasonal, monthly, daily, or hourly trends, but various recent studies have explored the potential of the Prophet algorithm with environmental time series data.
Unlike ARIMA or SARIMAX, which use autocorrelation and partial autocorrelation to capture temporal dependencies, Prophet decomposes the input time series into additive components [40].

2.3. Data Preprocessing

The activity data of bivalve mollusks in our case were the opening values of the valves (in mm) of 14 mollusks, which were simultaneously recorded and transmitted to the server with a sampling step of 10 s. The average value of the valve openings of all 14 mollusks was used as input data for our models. The data were first checked for omissions and duplicates. Data gaps were filled by interpolation, and duplicates were excluded from the data set. Data gaps were caused by hardware failures.
The ultimate goal of the algorithm is to determine the time of occurrence of the anomaly. The time difference between the detected anomaly point and the actual anomaly point represents a measure of performance. However, the exact time when the anomaly occurred is unknown. The days on which anomalies occurred were identified using the expert assessment method. A biologist and two hydrologists took part in the anomaly analysis. The biologist assessed the monotony of the behavior of mollusks in the daily cycle and determined deviations. Hydrologists assessed the condition of the water (analysis of water turbidity and temperature).
The entire data set (except for anomalies) was divided into intervals from one to five days, with a shift from one minute to an hour. Within each interval, with the exception of the last interval (from one minute to an hour), the models were trained with different data averagings. The final set of variable parameters common to all models was as follows: averaging time (from 10 s to 30 min), number of prediction points (or interval) (from 1 to 6 or from 1 minute to 1 hour), and training sample size (from 1 to 5 days). Multiplying the averaging time and the number of prediction points gave the forecasting horizon (in minutes). The data decompositions studied earlier in the work [21] were used to select the seasonality period parameter required for the Theta and Prophet models. The published article provides a description of the methodology and a detailed analysis of the results. Note that based on the results of the decomposition analysis of the valve opening value (Figure 2), the daily seasonal component of the model was identified (Figure 2b). For the Theta model, the additive method of combining theta lines was used. For the Croston method, the classic version was chosen.
The models were compared with each other based on the time of anomaly detection. The best solution is to detect the anomaly as quickly as possible with a minimum number of false alarms. Thus, the optimal solution is based on a trade-off between detection speed or detection latency and the false alarm rate.
As a measure for detecting anomalies in the data, the root mean squared error (RMSE) was used [41]. The RMSE is the square root of the variance, evaluated by the equation:
R M S E = i = 1 n ( y ^ i y i ) n
where yi and y i ^ are actual and predicted values, and n is the number of samples.
The calculation of the RMSE was carried out for each prediction interval for all data averaging times. The fact that the threshold value of the RMSE was exceeded was a signal of an anomaly.
The block diagram of the algorithm described above is presented in Figure 3.
Data analysis and implementation of the developed algorithm were carried out in the Python programming language (V 3.9.12) using the machine learning library scikit-learn (V 1.2.2) [42], statistical analysis library statsmodels (V 0.14.0) [43], and library for forecasting and anomaly detection in time series Darts (V 0.27.1) [44].

3. Results

The fixed threshold of the RMSE, at which the largest error occurred in the absence of an anomaly, plus a ten percent safety margin required to work in a real system for each method, was selected to exclude false positives (Table 1, Table 2 and Table 3). It turned out to be impossible to detect anomalies using the Theta method with a 10 s averaging, since false positives turned out to be higher than real anomalies in the data. With one-minute data averaging, false positives of the models were obtained at the same level with real anomalies by all three methods. For example, Figure 4 shows RMSE graphs for a false alarm above the real anomaly (a) (Theta method, 10 s averaging), for a false alarm at the same level with real anomalies (b) (Prophet method, 1 min data averaging), as well as for the normal operation of the model, where an anomaly was clearly visible in the RMSE values (c) (Prophet, 20 min data averaging).
Using the calculated values of RMSE for certain model settings, the detection times of three anomalies were determined (Table 4). The best time (17:20) for detecting the first anomaly (19 March 2017) was obtained by the Prophet method, with a 20 min averaging and with the number of forecast points equal to 3 (one-hour forecast). The best time to detect the second anomaly (14 April 2017) turned out to be the same for the three methods, with the same model settings (5 min of data averaging and 6 prediction points, with a one-hour forecast). The third anomaly (24 April 2017), similar to the second, was detected using three methods at the same time. The number of days in the training set did not affect the time it took to detect an anomaly, although it did affect the threshold level.
Figure 5 shows the RMSE values and their fixed thresholds for highlighting the anomaly on 19 March 2017.

4. Discussion and Conclusions

The algorithms used separately or together with other models (hybrid approaches) were actively applied to different types of data. For example, algorithms were applied to predict the amount of waste per capita in the states of the European Union using the Prophet model [45]. De Oliviera et al. [46] proposed a new multi-functional methodology based on auto-encoder and traditional time series models, including ARIMA and Prophet, for COVID-19 time series forecasting. Time series forecasting models (SARIMA and Prophet) and technologies based on artificial neural networks were compared to model the dynamics and conditions of soils and the efficiency of agroecosystems [47]. A hybrid model combining wavelet transform, support vector regression (SVR), and Prophet was used for a precipitation forecast [48]. Using several methods, including Croston, an inventory planning system was developed for a tire retailer [49].
In the paper, we compared three machine learning prediction methods (Theta, Croston, and Prophet) for detecting anomalies in bivalves’ activity data. The results showed that for one of the anomalies, Prophet was the best method, and for the other two, the anomaly detection time did not differ between the methods. A comparison of the results obtained in the work with the results of predicting anomalies in the activity data of bivalve mollusks by the SARIMA model [21] showed the preferable use of the selected algorithms. The Prophet method was able to detect the first anomaly 1 hour and 20 min faster than the SARIMA model. The detection time of the second anomaly by the three analyzed algorithms improved by 25 min, and the third anomaly improved by 40 min, compared to the SARIMA model.
The results of the methods in terms of computation time were compared. Calculations were performed on Windows 10 Pro System Version 22H2 Build 19045.4170, processor Intel(R) Core(TM) i5-10500 CPU @ 3.10GHz, RAM 16 GB, NVIDIA GeForce GTX 1650 video card. Croston’s method turned out to be the fastest for all model settings (averaging time and number of prediction points) (Table 5). The computation time for the best model parameters obtained for the second anomaly (5 min, 6 points) showed that the fastest was the Croston method, which was 5.4 times faster compared to the Theta method (40 s, 3 min, and 36 s, respectively) and 18 times faster compared to the Prophet method (40 s and 12 min, respectively). The shortest computation time for the best model parameters obtained for the third anomaly (10 min, 6 points, one-hour forecast) was also obtained for the Croston method (27 s), while the Theta method gave the result in 46 s, and the Prophet in 7 min (Table 5). The SARIMA model turned out to be the slowest computationally. Computation time for the SARIMA model, with settings of averaging time 5 min and 6 forecasting points (the best detection time for the second anomaly by Theta, Croston, and Prophet methods), was 1 h and 40 min.
Comparison with estimates presented in [20], which describes the possibility of detecting anomalies in the same bivalves’ data using four unsupervised machine learning algorithms: elliptic envelope, isolation forest (iForest), one-class support vector machine (SVM), and local outlier level (LOF), showed using the example of the first anomaly (anomaly 3 in [20]) that the use of machine learning algorithms lost by almost an hour in the speed of anomaly detection, compared to the Prophet method (Table 6).
Thus, the algorithms discussed in the article turned out to be better in terms of anomaly detection time and faster in computational complexity, compared to the SARIMA model.
The advantage of the algorithms used in the article compared to the ARIMA model has been shown in many works using different data. For example, the advantage (less errors) of the Prophet model compared to SARIMA was obtained when predicting solar and wind resources in the Doomadgee area of Far North Queensland, Australia [50]. Despite being extremely useful, empirical evaluations, such as the M4 competition [51], have shown that no automatic selection algorithm (e.g., ETS, or automatic ARIMA) has been able to significantly outperform established benchmarks such as Theta [28], which became popular due to its superiority over other competitors in the M3 competition [52]. Prophet outperformed two other algorithms, ARIMA and ThymeBoost, in predicting monthly rainfall data in India [53]. Xiao et al. [54] found that Prophet improved runoff modeling in the Zhou River Basin. Bolick et al. [55] used the Prophet method to predict hourly water level changes in an urban stream (South Carolina, USA) and obtained very accurate estimates, with coefficient of determination values greater than 0.9.
The work carried out is important for improving the efficiency of forecasting anomalies in the activity data of bivalve mollusks used in the environmental monitoring of the state of the aquatic environment, which will help reduce costs associated with operational, tactical, and strategic planning.

Author Contributions

Conceptualization, A.N.G. and E.V.V.; methodology, A.N.G.; software, A.S.M.; validation, A.N.G. and A.S.M.; formal analysis, A.N.G. and A.S.M.; writing—original draft preparation, A.N.G.; writing—review and editing, A.N.G. and E.V.V.; visualization, E.V.V.; supervision, E.V.V.; project administration, E.V.V.; funding acquisition, E.V.V. All authors have read and agreed to the published version of the manuscript.

Funding

The study was supported by a grant from the Russian Science Foundation NO. 23-29-00558, https://rscf.ru/en/project/23-29-00558/.

Data Availability Statement

Data are available upon request.

Acknowledgments

The authors are grateful to the editor and anonymous reviewers for the remarks and comments which led to the improvement of the paper.

Conflicts of Interest

The authors declare no conflicts of interest.

Acronyms

ARIMA (SARIMA)(seasonal) autoregressive integrated moving average
ARVarray of RMSE values
BEWSbiological early warning systems
DCCdata cleaning counter
iForestisolation forest
LOFlocal outlier level
RMSEroot mean squared error
SESsimple exponential smoothing
SVMsupport vector machine
TAtraining array
VVOvalve opening value

References

  1. Van Vliet, M.T.H.; Jones, E.R.; Flörke, M.; Franssen, W.H.P.; Hanasaki, N.; Wada, Y.; Yearsley, J.R. Global water scarcity including surface water quality and expansions of clean water technologies. Environ. Res. Lett. 2021, 16, 024020. [Google Scholar] [CrossRef]
  2. The Sustainable Development Goals Report 2022. Available online: https://unstats.un.org/sdgs/report/2023/The-Sustainable-Development-Goals-Report-2023.pdf (accessed on 20 March 2024).
  3. Wang, Z.; Walker, G.W.; Muir, D.C.; Nagatani-Yoshida, K. Toward a global understanding of chemical pollution: A first comprehensive analysis of national and regional chemical inventories. Environ. Sci. Technol. 2020, 54, 2575–2584. [Google Scholar] [CrossRef] [PubMed]
  4. Lemm, J.U.; Venohr, M.; Globevnik, L.; Stefanidis, K.; Panagopoulos, Y.; van Gils, J.; Posthuma, L.; Kristensen, P.; Feld, C.K.; Mahnkopf, J.; et al. Multiple stressors determine river ecological status at the European scale: Towards an integrated understanding of river status deterioration. Glob. Change Biol. 2021, 27, 1962–1975. [Google Scholar] [CrossRef] [PubMed]
  5. Kramer, K.J.M.; Botterweg, J. Aquatic biological early warning systems: An overview. In Bioindicators and Environmental Management; Jeffrey, D.W., Madden, B., Eds.; Academic Press Inc.: London, UK, 1991; pp. 95–126. [Google Scholar]
  6. Bae, M.J.; Park, Y.S. Biological early warning system based on the responses of aquatic organisms to disturbances: A review. Sci Total Environ. 2014, 466, 635–649. [Google Scholar] [CrossRef] [PubMed]
  7. Haag, W.R.; Rypel, A.L. Growth and longevity in freshwater mussels: Evolutionary and conservation implications. Biol. Rev. 2011, 86, 225–247. [Google Scholar] [CrossRef] [PubMed]
  8. Hartmann, J.T.; Beggel, S.; Auerswald, K.; Stoeckle, B.C.; Geist, J. Establishing mussel behavior as a biomarker in ecotoxicology. Aquat. Toxicol. 2016, 170, 279–288. [Google Scholar] [CrossRef]
  9. Tran, D.; Ciret, P.; Ciutat, A.; Durrieu, G.; Massabuau, J.-C. Estimation of potential and limits of bivalve closure response to detect contaminants: Application to cadmium. Environ. Toxicol. Chem. 2003, 22, 914–920. [Google Scholar] [CrossRef] [PubMed]
  10. Aggarwal, C.C. Data Mining: The Textbook; Springer: Berlin/Heidelberg, Germany, 2015. [Google Scholar]
  11. Ahmad, S.; Lavin, A.; Purdy, S.; Agha, Z. Unsupervised real-time anomaly detection for streaming data. Neurocomputing 2017, 262, 134–147. [Google Scholar] [CrossRef]
  12. Chuwang, D.D.; Chen, W. Forecasting Daily and weekly passenger demand for urban rail transit stations based on a time series model approach. Forecasting 2022, 4, 904–924. [Google Scholar] [CrossRef]
  13. Menculini, L.; Marini, A.; Proietti, M.; Garinei, A.; Bozza, A.; Moretti, C.; Marconi, M. Comparing Prophet and deep learning to ARIMA in forecasting wholesale food prices. Forecasting 2021, 3, 644–662. [Google Scholar] [CrossRef]
  14. Stefenon, S.F.; Seman, L.O.; Mariani, V.C.; Coelho, L.d.S. Aggregating Prophet and seasonal trend decomposition for time series forecasting of Italian electricity spot prices. Energies 2023, 16, 1371. [Google Scholar] [CrossRef]
  15. Shen, J.; Valagolam, D.; McCalla, S. Prophet forecasting model: A machine learning approach to predict the concentration of air pollutants (PM2.5, PM10, O3, NO2, SO2, CO) in Seoul, South Korea. PeerJ 2020, 8, e9961. [Google Scholar] [CrossRef] [PubMed]
  16. Hasnain, A.; Sheng, Y.; Hashmi, M.Z. Time series analysis and forecasting of air pollutants based on Prophet forecasting model in Jiangsu Province, China. Front. Environ. Sci. 2022, 10, 1044. [Google Scholar] [CrossRef]
  17. Kramar, V.; Alchakov, V. Time-series forecasting of seasonal data using machine learning methods. Algorithms 2023, 16, 248. [Google Scholar] [CrossRef]
  18. Petropoulos, F.; Spiliotis, E. The wisdom of the data: Getting the most out of univariate time series forecasting. Forecasting 2021, 3, 478–497. [Google Scholar] [CrossRef]
  19. Jiao, Z.; Shan, X. A Bayesian Approach for Forecasting the Probability of Large Earthquakes Using Thermal Anomalies from Satellite Observations. Remote Sens. 2024, 16, 1542. [Google Scholar] [CrossRef]
  20. Grekov, A.N.; Kabanov, A.A.; Vyshkvarkova, E.V.; Trusevich, V.V. Anomaly detection in biological early warning systems using unsupervised machine learning. Sensors 2023, 23, 2687. [Google Scholar] [CrossRef] [PubMed]
  21. Grekov, A.N.; Vyshkvarkova, E.V.; Mavrin, A.S. Anomaly detection algorithm using the SARIMA model for the software of an automated complex for the aquatic environment biomonitoring. Artif. Intell. Decis. Mak. 2024, 1, 52–67. (In Russian) [Google Scholar] [CrossRef]
  22. Grekov, A.N.; Kuzmin, K.A.; Mishurov, V.Z. Automated early warning system for water environment based on behavioral reactions of bivalves. In Proceedings of the 2019 International Russian Automation Conference (RusAutoCon), Sochi, Russia, 8–14 September 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 1–5. [Google Scholar]
  23. Valletta, J.J.; Torney, C.; Kings, M.; Thornton, A.; Madden, J. Applications of machine learning in animal behavior studies. Anim. Behav. 2017, 124, 203–220. [Google Scholar] [CrossRef]
  24. Bertolini, C.; Capelle, J.; Royer, E.; Milan, M.; Witbaard, R.; Bouma, T.J.; Pastres, R. Using a clustering algorithm to identify patterns of valve-gaping behavior in mussels reared under different environmental conditions. Ecol. Inform. 2022, 69, e101659. [Google Scholar] [CrossRef]
  25. Meyer, P.G.; Cherstvy, A.G.; Seckler, H.; Hering, R.; Blaum, N.; Jeltsch, F.; Metzler, R. Directedeness, correlations, and daily cycles in springbok motion: From data via stochastic models to movement prediction. Phys. Rev. Res. 2023, 5, 043129. [Google Scholar] [CrossRef]
  26. Gnyubkin, V.F. An early warning system for aquatic environment state monitoring based on an analysis of mussel valve movement. Russ. J. Mar. Biol. 2009, 35, 431–436. [Google Scholar] [CrossRef]
  27. Borcherding, J. Ten years of practical experience with the Dreissena-Monitor, a biological early warning system for continuous water quality monitoring. Hydrobiologia 2006, 556, 417–426. [Google Scholar] [CrossRef]
  28. Assimakopoulos, V.; Nikolopoulos, K. The Theta model: A decomposition approach to forecasting. Int. J. Forecast. 2000, 16, 521–530. [Google Scholar] [CrossRef]
  29. Hyndman, R.J.; Billah, B. Unmasking the Theta method. Int. J. Forecast. 2003, 19, 287–290. [Google Scholar] [CrossRef]
  30. Fiorucci, J.A.; Pellegrini, T.R.; Louzada, F.; Petropoulos, F.; Koehler, A.B. Models for optimising the theta method and their relationship to state space models. Int. J. Forecast. 2016, 32, 1151–1161. [Google Scholar] [CrossRef]
  31. Croston, J.D. Forecasting and Stock Control for Intermittent Demands. Oper. Res. Q. 1972, 23, 289–303. [Google Scholar] [CrossRef]
  32. Svetunkov, I.; Boylan, J.E. iETS: State space model for intermittent demand forecasting. Int. J. Prod. Econ. 2023, 265, 109013. [Google Scholar] [CrossRef]
  33. Prestwich, S.D.; Tarim, S.A.; Rossi, R. Intermittency and obsolescence: A Croston method with linear decay. Int. J. Forecast. 2021, 37, 708–715. [Google Scholar] [CrossRef]
  34. Schultz, C.R. Forecasting and inventory control for sporadic demand under periodic review. J. Oper. Res. Soc. 1987, 38, 453–458. [Google Scholar] [CrossRef]
  35. Snyder, R.D. Forecasting sales of slow and fast moving inventories. Eur. J. Oper. Res. 2002, 140, 684–699. [Google Scholar] [CrossRef]
  36. Kourentzes, N. On intermittent demand model optimisation and selection. Int. J. Prod. Econ. 2014, 156, 180–190. [Google Scholar] [CrossRef]
  37. Teunter, R.; Syntetos, A.A.; Babai, M.Z. Intermittent demand: Linking forecasting to inventory obsolescence. Eur. J. Oper. Res. 2011, 214, 606–615. [Google Scholar] [CrossRef]
  38. Taylor, S.J.; Letham, B. Forecasting at scale. Am. Stat. 2018, 72, 37–45. [Google Scholar] [CrossRef]
  39. Harvey, A.C.; Peters, S. Estimation procedures for structural time series models. J. Forecast. 1990, 9, 89–108. [Google Scholar] [CrossRef]
  40. Fronzi, D.; Narang, G.; Galdelli, A.; Pepi, A.; Mancini, A.; Tazioli, A. Towards groundwater-level prediction using Prophet forecasting method by exploiting a high-resolution hydrogeological monitoring system. Water 2024, 16, 152. [Google Scholar] [CrossRef]
  41. Aminikhanghahi, S.; Cook, D.J. A survey of methods for time series change point detection. Knowl. Inf. Syst. 2017, 51, 339–367. [Google Scholar] [CrossRef] [PubMed]
  42. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Müller, A.; Nothman, J.; Louppe, G.; et al. Scikit-learn: Machine learning in python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  43. Seabold, S.; Perktold, J. Statsmodels: Econometric and statistical modeling with python. Proc. 9th Python Sci. Conf. 2010, 57, 10–25080. [Google Scholar]
  44. Herzen, J.; Lässig, F.; Piazzetta, S.G.; Neuer, T.; Tafti, L.; Raille, G.; Van Pottelbergh, T.; Pasieka, M.; Skrodzki, A.; Huguenin, N.; et al. Darts: User-friendly modern machine learning for time series. J. Mach. Learn. Res. 2022, 23, 1–6. [Google Scholar]
  45. Scriosteanu, A.; Criveanu, M.M. Reverse Logistics of Packaging Waste under the Conditions of a Sustainable Circular Economy at the Level of the European Union States. Sustainability 2023, 15, 14727. [Google Scholar] [CrossRef]
  46. De Oliveira, E.V.; Aragão, D.P.; Gonçalves, L.M.G. A New Auto-Regressive Multi-Variable Modified Auto-Encoder for Multivariate Time-Series Prediction: A Case Study with Application to COVID-19 Pandemics. Int. J. Environ. Res. Public Health 2024, 21, 497. [Google Scholar] [CrossRef] [PubMed]
  47. Mirpulatov, I.; Gasanov, M.; Matveev, S. Soil Dynamics and Crop Yield Modeling Using the MONICA Crop Simulation Model and Time Series Forecasting Methods. Agronomy 2023, 13, 2185. [Google Scholar] [CrossRef]
  48. Li, D.; Ma, J.; Rao, K.; Wang, X.; Li, R.; Yang, Y.; Zheng, H. Prediction of Rainfall Time Series Using the Hybrid DWT-SVR-Prophet Model. Water 2023, 15, 1935. [Google Scholar] [CrossRef]
  49. Neves, D.; Monteiro, M.; Felício, M.J. Inventory Improvement in Tyre Retail through Demand Forecasting. Eng. Proc. 2023, 39, 1. [Google Scholar] [CrossRef]
  50. Islam, M.K.; Hassan, N.M.S.; Rasul, M.G.; Emami, K.; Chowdhury, A.A. Forecasting of solar and wind resources for power generation. Energies 2023, 16, 6247. [Google Scholar] [CrossRef]
  51. Makridakis, S.; Spiliotis, E.; Assimakopoulos, V. The M4 Competition: 100,000 time series and 61 forecasting methods. Int. J. Forecast. 2020, 36, 54–74. [Google Scholar] [CrossRef]
  52. Spiliotis, E.; Assimakopoulos, V.; Makridakis, S. Generalizing the Theta method for automatic forecasting. Eur. J. Oper. Res. 2020, 284, 26. [Google Scholar] [CrossRef]
  53. Chowdari, K.K.; Barma, S.D.; Bhat, N.; Girisha, R.; Gouda, K.C. Evaluation of ARIMA, Facebook Prophet, and a boosting algorithm framework for monthly precipitation prediction of a semi-arid district of north Karnataka, India. In Proceedings of the Fourth International Conference on Emerging Research in Electronics, Computer Science, and Technology (ICERECT), Mandya, India, 26–27 December 2022; pp. 1–5. [Google Scholar]
  54. Xiao, Q.; Zhou, L.; Xiang, X.; Liu, L.; Liu, X.; Li, X.; Ao, T. Integration of hydrological model and time series model for improving the runoff simulation: A case study on BTOP modeling in Zhou River Basin, China. Appl. Sci. 2022, 12, 6883. [Google Scholar] [CrossRef]
  55. Bolick, M.M.; Post, C.J.; Naser, M.Z.; Forghanparast, F.; Mikhailova, E.A. Evaluating urban stream flooding with machine learning, LiDAR, and 3D Modeling. Water 2023, 15, 2581. [Google Scholar] [CrossRef]
Figure 1. Biological early warning system diagram (a) and the diagram of attaching mussels to the block (b).
Figure 1. Biological early warning system diagram (a) and the diagram of attaching mussels to the block (b).
Forecasting 06 00019 g001
Figure 2. Seasonal component on a two-month interval (a) after the decomposition of the averaged bivalves’ activity data and for the period 10–21 April 2017 (b).
Figure 2. Seasonal component on a two-month interval (a) after the decomposition of the averaged bivalves’ activity data and for the period 10–21 April 2017 (b).
Forecasting 06 00019 g002
Figure 3. The block diagram of the algorithm.
Figure 3. The block diagram of the algorithm.
Forecasting 06 00019 g003
Figure 4. Results of choosing a fixed RMSE threshold: (a) for false positives using the Theta method; (b) in case of a false positive using the Prophet method, 1 min; and (c) during normal operation of the Prophet model, 20 min.
Figure 4. Results of choosing a fixed RMSE threshold: (a) for false positives using the Theta method; (b) in case of a false positive using the Prophet method, 1 min; and (c) during normal operation of the Prophet model, 20 min.
Forecasting 06 00019 g004
Figure 5. Detection time of anomaly 1 (19 March 2017) by three methods, with 20 min averaging and three prediction points.
Figure 5. Detection time of anomaly 1 (19 March 2017) by three methods, with 20 min averaging and three prediction points.
Forecasting 06 00019 g005
Table 1. The RMSE values plus 10% for different averaging times and the number of prediction points using the Theta method.
Table 1. The RMSE values plus 10% for different averaging times and the number of prediction points using the Theta method.
Averaging TimeNumber of Prediction PointsForecasting Horizon (min)Training Sample Size, Days
12345
10 s61*****
1 min66**********
5 min6300.3130.3140.3140.3140.286
10 min110**********
10 min2200.3160.3140.3150.3160.275
10 min3300.2950.2920.2560.2530.264
10 min6600.3110.3090.2130.2060.209
20 min3600.2940.2890.2920.2750.286
30 min2600.2430.2330.2310.2310.264
* Cannot be distinguished—false positives are higher than real anomalies. ** Cannot be distinguished—false positives at the same level as real anomalies.
Table 2. The RMSE values plus 10% for different averaging times and the number of prediction points using the Croston method.
Table 2. The RMSE values plus 10% for different averaging times and the number of prediction points using the Croston method.
Averaging TimeNumber of Prediction PointsForecasting Horizon (min)Training Sample Size, Days
12345
10 s61**********
1 min66**********
5 min6300.3220.3220.3220.3220.314
10 min1100.3630.3610.3610.3610.361
10 min2200.410.410.410.410.41
10 min3300.3130.3130.2920.2920.292
10 min6600.3810.3810.4810.4810.481
20 min3600.3710.5710.3690.3710.373
30 min2600.520.5140.510.520.508
** Cannot be distinguished—false positives at the same level as real anomalies.
Table 3. The RMSE values plus 10% for different averaging times and the number of prediction points using the Prophet method.
Table 3. The RMSE values plus 10% for different averaging times and the number of prediction points using the Prophet method.
Averaging TimeNumber of Prediction PointsForecasting Horizon (min)Training Sample Size, Days
12345
10 s61**********
1 min66**********
5 min6300.2810.3610.4250.4910.578
10 min3300.2910.3650.4910.5410.629
10 min6600.1920.3230.3810.4220.496
20 min3600.1910.4050.4150.4860.534
30 min2600.2640.2730.2900.2920.284
** Cannot be distinguished—false positives at the same level with real anomalies.
Table 4. Anomaly detection time by three algorithms at different data averaging times and number of prediction points.
Table 4. Anomaly detection time by three algorithms at different data averaging times and number of prediction points.
Averaging Time, Number of Prediction PointsMethod
ThetaCrostonProphet
Anomaly 1 (19 March 2017)
5 min, 6 points18:3518:3518:35
10 min, 1 point 18:40
10 min, 2 points18:3018:30
10 min, 3 points18:4018:4018:40
10 min, 6 points19:1018:1018:10
20 min, 3 points19:2018:2017:20
30 min, 3 points18:3019:3018:30
Anomaly 2 (14 April 2017)
5 min, 6 points12:0512:0512:05
10 min, 1 point 12:30
10 min, 2 points12:3012:30
10 min, 3 points12:4012:4012:40
10 min, 6 points12:1012:1012:10
20 min, 3 points12:2012:2012:20
30 min, 3 points12:3012:3012:30
Anomaly 3 (24 April 2017)
5 min, 6 points17:3517:3517:35
10 min, 1 point 17:50
10 min, 2 points17:5017:50
10 min, 3 points17:4017:4017:40
10 min, 6 points17:1017:1017:10
20 min, 3 points17:2017:2017:20
30 min, 2 points17:3017:3017:30
The best anomaly detection time is highlighted in red.
Table 5. Computation time by all methods.
Table 5. Computation time by all methods.
Averaging TimeMethod
ThetaCrostonProphetSARIMA
10 s, 6 points4 h40 min1 h 10 min2 h 22 min
1 min, 6 points13 min1 min 25 s15 min1 h 43 min
5 min, 6 points3 min 36 s40 s12 min1 h 40 min
10 min, 1 point3 min 33 s55 s12 min1 h 27 min
10 min, 2 points1 min 54 s37 s10 min1 h 10 min
10 min, 3 points1 min 20 s33 s9 min1 h 2 min
10 min, 6 points46 s27 s7 min52 min
20 min, 3 points35 s23 s6 min50 min
30 min, 2 points29 s21 s5 min44 min
Table 6. Anomaly detection time for Anomaly 1 (19 March 2017) by Prophet and unsupervised machine learning algorithms.
Table 6. Anomaly detection time for Anomaly 1 (19 March 2017) by Prophet and unsupervised machine learning algorithms.
MethodAnomaly Detection TimeModel Setting (Averaging Time)
Prophet17:2020 min
Elliptic envelope18:3015 min
iForest18:1515 min
LOF18:355 min
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Grekov, A.N.; Vyshkvarkova, E.V.; Mavrin, A.S. Forecasting and Anomaly Detection in BEWS: Comparative Study of Theta, Croston, and Prophet Algorithms. Forecasting 2024, 6, 343-356. https://doi.org/10.3390/forecast6020019

AMA Style

Grekov AN, Vyshkvarkova EV, Mavrin AS. Forecasting and Anomaly Detection in BEWS: Comparative Study of Theta, Croston, and Prophet Algorithms. Forecasting. 2024; 6(2):343-356. https://doi.org/10.3390/forecast6020019

Chicago/Turabian Style

Grekov, Aleksandr N., Elena V. Vyshkvarkova, and Aleksandr S. Mavrin. 2024. "Forecasting and Anomaly Detection in BEWS: Comparative Study of Theta, Croston, and Prophet Algorithms" Forecasting 6, no. 2: 343-356. https://doi.org/10.3390/forecast6020019

APA Style

Grekov, A. N., Vyshkvarkova, E. V., & Mavrin, A. S. (2024). Forecasting and Anomaly Detection in BEWS: Comparative Study of Theta, Croston, and Prophet Algorithms. Forecasting, 6(2), 343-356. https://doi.org/10.3390/forecast6020019

Article Metrics

Back to TopTop