Next Article in Journal
A New Accelerated Forward–Backward Splitting Algorithm for Monotone Inclusions with Application to Data Classification
Previous Article in Journal
Dark Energy from Time Crystals
Previous Article in Special Issue
Intuitive Innovation: Unconventional Modeling and Systems Neurology
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep Learning for Bifurcation Detection: Extending Early Warning Signals to Dynamical Systems with Coloured Noise

by
Yazdan Babazadeh Maghsoodlo
1,2,*,
Daniel Dylewsky
3,
Madhur Anand
2 and
Chris T. Bauch
1
1
Department of Applied Mathematics, University of Waterloo, Waterloo, ON N2L 3G1, Canada
2
School of Environmental Sciences, University of Guelph, Guelph, ON N1G 2W1, Canada
3
Department of Biology, Carleton University, Ottawa, ON K1S 5B6, Canada
*
Author to whom correspondence should be addressed.
Mathematics 2025, 13(17), 2782; https://doi.org/10.3390/math13172782
Submission received: 15 May 2025 / Revised: 28 July 2025 / Accepted: 19 August 2025 / Published: 29 August 2025
(This article belongs to the Special Issue Innovative Approaches to Modeling Complex Systems)

Abstract

Deep learning models have demonstrated remarkable success in recognising tipping points and providing early warning signals. However, there has been limited exploration of their application to dynamical systems governed by coloured noise, which characterizes many real-world systems. In this study, we show that it is possible to leverage the normal forms of three primary types of bifurcations (fold, transcritical, and Hopf) to construct a training set that enables deep learning architectures to perform effectively. Furthermore, we showed that this approach could accommodate coloured noise by replacing white noise with red noise during the training process. To evaluate the classifier trained on red noise compared to one trained on white noise, we tested their performance on mathematical models using Receiver Operating Characteristic (ROC) curves and Area Under the Curve (AUC) scores. Our findings reveal that the deep learning architecture can be effectively trained on coloured noise inputs, as evidenced by high validation accuracy and minimal sensitivity to redness (ranging from 0.83 to 0.85). However, classifiers trained on white noise also demonstrate impressive performance in identifying tipping points in coloured time series. This is further supported by high AUC scores (ranging from 0.9 to 1) for both classifiers across different coloured stochastic time series.

1. Introduction

Dynamical systems are employed to describe various phenomena across disciplines, from studying the climate system [1,2,3] to epidemiological studies [4,5,6], rumour propagation mechanisms [7], and coupled behaviour–disease dynamics [8]. While dynamical systems exhibit numerous intriguing aspects, one particularly captivating focus in recent years has been the possibility of bifurcation and tipping points [8,9,10,11]. In natural systems, a local bifurcation can occur as external conditions gradually change, bringing the system near an equilibrium state and potentially triggering a qualitative shift in behaviour when a tipping point is reached [12,13].
Bifurcation theory is a rich branch of mathematics that delves into the theoretical properties of dynamical systems experiencing different types of bifurcations [14]. One of its most significant theories, the Centre Manifold Theorem, along with the Normal Form Theorem, simplifies nonlinear dynamical systems by transforming them into their simplest equivalent form, termed the “normal form”. Through coordinate transformations, it eliminates unnecessary nonlinear terms while preserving the system’s essential behaviour, making it easier to examine features like bifurcations [15,16,17].
While these theorems are generally valid, this paper specifically focuses on three types of local bifurcations, as they represent distinct behaviours in dynamical systems: the abrupt jump from one equilibrium state to another (fold bifurcation) [18], smoothly shifting stability (transcritical bifurcation) [19], and the transition to oscillatory behaviour (Hopf bifurcation) [20]. For these bifurcation types, the dominant eigenvalue of the derivative matrix crosses the imaginary axis at the bifurcation point. Leading up to the bifurcation, the system can exhibit critical slowing down, corresponding to prolonged recovery from perturbations and reduced resilience to disturbances, as measured by increased auto-correlation and variance [21]. Numerous generic indicators of local bifurcations have been proposed based on these consequences [22,23,24,25]. For instance, the variance and lag-1 auto-correlation of a time series prior to bifurcation shows trends that can serve as early warning signals (EWSs) [26]. While such indicators have been instrumental in identifying and predicting tipping points and abrupt shifts, they also possess limitations [27], such as an inability to specify the type of impending bifurcation. With recent advances in artificial intelligence (AI) and the development of deep learning (DL) models, it was anticipated that these models could eventually provide EWSs with greater accuracy and fewer weaknesses than generic indicators.
The first study applying AI to the detection of tipping points in dynamical systems is by Bury et al. [28], who showed that convolutional and recurrent neural networks could identify EWS in time series drawn from a random library of differential equations. Their approach demonstrated that DL architectures could outperform traditional statistical indicators. Deb et al. [29] introduced EWSNet, a hybrid Convolutional Neural Network–Long Short-Term Memory (CNN-LSTM) network trained on simulated time series from ecological, biological, and climate models. EWSNet showed high accuracy in classifying catastrophic and non-catastrophic transitions and proved robust to noise, short time series, and irregular sampling. These previous works were extended in a follow-up study by Bury et al. [30] to discrete-time systems, further validating the generalizability of DL models to diverse bifurcation scenarios. Dylewsky et al. [31] demonstrated that universal EWSs could be extracted from DL models trained on climate system simulations, supporting the broader applicability of these methods. Dylewsky et al. [25] advanced this work by embedding bifurcations in high-dimensional systems, showing that DL models remain effective even when the underlying signals are distributed across many variables. Most recently, Huang et al. [32] addressed the challenge of rate-induced tipping (R-tipping), where classical indicators like critical slowing down fail. Their interpretable DL framework predicted transition probabilities in noisy, time-varying systems and extracted higher-order fingerprints of instability. In parallel, Dylewsky et al. [33] tackled spatially patterned phase transitions using neural architectures tailored for spatiotemporal systems. Together, these studies illustrate a growing body of work that supports the use of deep learning, with CNN–LSTM architectures being one of the main and most effective models for robust and generalizable detection of critical transitions in noisy, nonlinear dynamical systems.
In contrast to previous approaches that generate extensive libraries of stochastic differential equations and discard those that do not exhibit bifurcations [30], we propose a theoretically grounded method by constructing the training dataset directly from the normal forms of canonical bifurcations. This significantly reduces computational overhead while ensuring that the training data captures the essential dynamical features associated with each bifurcation type. The first objective of this study is to demonstrate that such a targeted construction is not only feasible but also more efficient than relying on random dynamical systems. Furthermore, we extend the DL-based early warning framework of [28] to accommodate coloured noise, a feature often overlooked despite its prevalence in empirical systems. Many ecological and climate systems that undergo critical transitions are influenced by auto-correlated, coloured noise [34,35], prompting recent interest in its ecological impacts [36,37,38,39,40,41,42,43,44,45]. By generating training data with varying levels of redness, we evaluate the robustness of our CNN-LSTM classifier under both white and coloured noise, enhancing the applicability of EWS frameworks to real-world, noisy, nonlinear systems.

2. Methods

2.1. Model Selection Rationale

To construct the deep learning model, we employed a hybrid architecture combining a CNN with an LSTM network [46,47,48]. Theoretically, the CNN layers serve to extract local patterns and short-term features from the time series by applying convolutional filters, while the LSTM layers are designed to capture long-range temporal dependencies through their recurrent structure. This layered combination is particularly well-suited for our task, as it enables both local pattern detection and memory retention—key components in identifying EWSs indicative of critical transitions.
The CNN-LSTM architecture has been widely adopted in the broader literature on time series forecasting and classification. For instance, ref. [49] demonstrated that CNN–LSTM achieved the highest forecasting accuracy across multiple benchmarks, outperforming standalone models such as Multilayer Perceptron (MLP), CNN, recurrent neural network (RNN), and LSTM. Similarly, ref. [50] evaluated CNN, LSTM, and CNN–LSTM on ice-jam event classification and reported superior F1-scores for the CNN–LSTM model. In the context of missing data imputation, ref. [51] showed that CNN–LSTM achieved better predictive accuracy than either CNN or LSTM alone when forecasting electricity consumption. More specifically, architectures similar to ours have been increasingly applied to problems involving bifurcation detection and phase transitions in dynamical systems. Recent studies such as [25,28,30,31,33] have demonstrated that CNN–LSTM models are effective at detecting EWSs in stochastic systems and outperforming simpler baselines.
To justify our choice of DL model, we conducted an ablation study comparing the performance of LSTM-only and CNN-LSTM architectures. Both models were trained and tested on time series generated with white noise, using identical hyperparameter settings to ensure a fair comparison. To ensure a fair evaluation, we tested both the LSTM-only and CNN-LSTM models on time series generated using the same process as the training dataset. The results, presented in Table 1, show that while the LSTM-only model achieves reasonably strong performance, the CNN-LSTM consistently outperforms it in terms of accuracy, precision, recall, and F1-score. This supports our choice of architecture and aligns with findings from previous work, such as [28], which also highlighted the superiority of CNN-LSTM over LSTM-only models in similar dynamical settings. Both the CNN–LSTM and LSTM-only models were trained on the same dataset from normal forms with white noise, using identical settings and pre-processing. The LSTM-only model shares the recurrent layers of the CNN–LSTM but omits the convolutional block. Hyperparameters were guided by [28] and further tuned through experiments to optimize performance. Full details are provided in the Section 5.

2.2. Constructing Training Data Set

The training dataset was created by stochastically simulating the normal forms of different types of bifurcations (fold, transcritical, Hopf). In each case, we began at the system’s equilibrium point, far from the bifurcation point, and then gradually and linearly moved the system closer to the bifurcation point. These time series were classified into three categories, with an additional “null” category created by simulating white/coloured noise alone. Using this dataset, we trained the CNN-LSTM model to classify time series into the four defined categories.
This method can be generalised to coloured noise by replacing white noise with red noise in the additive stochastic process of simulations. We generated red noise using the Auto-Regressive model of order 1 (AR(1)) method.
X t + 1 = ( redness · X t ) + 1 redness 2 · ϵ t
With X 0 N ( 0 ,   1 ) . This recursive equation is controlled by a redness parameter between 0 and 1. If the redness equals zero, the process generates white noise, and the larger the redness, the more coloured and auto-correlated the process becomes. Additionally, ϵ t is a stochastic variable drawn from a Gaussian distribution N ( 0 ,   1 ) .
We simulated 20,000 time series in total, using 90% of them for the training process. The remaining 10% was used to test the performance of the DL model. Table 2 summarises the model performance metrics across redness levels.
The redness values range from 0.0 to 0.5, where higher values indicate stronger auto-correlation in the input noise. The validation accuracy and other performance metrics remain relatively stable across this range, with a peak of 85% observed at redness values of 0.0, 0.1, 0.2, and 0.3. A slight drop to 83% is seen at the extreme values (0.4 and 0.5), suggesting minimal sensitivity of the model’s performance to changes in noise persistence. This demonstrates that the DL architecture can be trained effectively on coloured noise inputs as well. The details of the confusion matrix and classification metrics are provided in the Section 5.
In addition, we include further discussion on the robustness of the training process, the minimum data requirements, and full technical details on simulating the training dataset.

3. Results

We have demonstrated that the DL model can provide EWS for bifurcations in mathematical models. For this purpose, we utilised six well-known mathematical models: neural activation and May’s harvesting model for fold bifurcation, the Rosenzweig–MacArthur and SIS models for transcritical bifurcation, and the Rosenzweig–MacArthur and Van der Pol models for Hopf bifurcation. We simulated these mathematical models with an additive stochastic term, starting from the equilibrium point and linearly varying the bifurcation parameter toward the bifurcation point throughout the simulation. Since the tipping point can occur earlier than the theoretical prediction, we employed the binary segmentation algorithm to locate the tipping point. The DL model assigns four probabilities (fold, transcritical, Hopf, and null) to each point in the simulated time series. The null category in our study represents stochastic time series that either do not include a tipping point or are relatively distant from it. Assigning a relatively high probability to the correct type of bifurcation before the actual location of the tipping point is considered a signal in this study. Figure 1 and Figure 2 depict the results of the DL analysis of the introduced mathematical models. The estimated location of the tipping point is indicated on the plots with a dashed line. As shown, the DL model is not only capable of providing EWS but also accurately identifies the correct type of bifurcation. Furthermore, at the beginning of the simulated time series, where the tipping point is relatively distant, the DL model assigns the highest probability to null. As the series approaches the tipping point, the model transitions to selecting the correct type of bifurcation and then returns to assigning null after the tipping point has passed. This behaviour is consistent with the observations from the analysis of all six mathematical models. The mathematical descriptions, full details of the simulations, and null analyses are provided in the Section 5.
In addition, we have demonstrated that the DL model exhibits robust performance in providing EWS on empirical data containing tipping points. We analysed two sets of empirical data: one representing the transition to thermoacoustic instability in a horizontal Rijke tube, a prototypical thermoacoustic system, and the other capturing episodes of anoxia in the Eastern Mediterranean from sedimentary archives. To ensure proper resolution and quality for the analysis, we expanded the size of the time series for these datasets using the Linear Interpolation method. Figure 3 illustrates the results of the DL analysis on four examples of these time series (two from each dataset). The estimated location of the tipping point is indicated with a dashed line on each subplot. As shown, the DL model successfully provides EWS on these time series before the actual tipping point, transitioning from assigning null to identifying one of the types of bifurcation. Notably, for the anoxia data, the DL classifier predicts a Hopf transition, whereas the empirical data suggest a transition more akin to a fold-type bifurcation. Additional analysis on empirical data is provided in the Section 5.
Moreover, the DL model demonstrates consistent success with stochastic time series that include coloured noise. To illustrate this, we repeated the same training process using the same DL architecture (CNN-LSTM), but this time incorporated red noise (with specific redness values) as an additive stochastic term while simulating the deterministic normal forms of different types of bifurcation. As a result, we generated five additional DL classifiers, each trained on normal forms combined with red noise (redness values of 0.1, 0.2, 0.3, 0.4, and 0.5). A larger redness value in red noise signifies greater dominance of low-frequency components, stronger persistence, and smoother, long-term variability in the time series. Note that the baseline DL model used in previous sections is based on white noise and can be considered a special case of this process with a redness value of 0. To quantify and compare the performance of the red noise classifier against the white noise classifier, we simulated stochastic time series for the six mathematical models using different redness values (ranging from 0 to 0.5). Each mathematical model was simulated 50 times, and the ROC curves for all classifiers were plotted. By calculating the AUC score for each classifier across mathematical models and redness values, we could compare their performance. The objective was to demonstrate that the DL model trained on white noise can maintain strong performance in detecting bifurcations on time series that include red noise. This expectation is confirmed by the results shown in Figure 4. Each of the six subplots in this figure corresponds to one of the mathematical models. The y-axis represents the AUC score, while the x-axis represents redness values. The performance of the white noise classifier is shown in blue, while the red noise classifier is depicted in orange. The figure shows that, for each redness value, the classifier trained on white noise exhibited comparable performance to the classifier trained on red noise with the corresponding redness value. Further explanations are provided in the Section 5.
The results presented in this section demonstrate that our approach to developing a DL model for providing the EWSs of tipping points can be generalised to coloured noise and applied to highly coloured simulated or empirical time series. Moreover, these findings indicate that a classifier trained on white noise can still maintain a high level of performance on coloured time series.

4. Discussion

4.1. Conclusions

In this study, we developed a DL framework for detecting the EWSs of tipping points in complex dynamical systems. Our primary objective was to construct a training methodology grounded in theory by generating time series exclusively from the normal forms of canonical bifurcations. This approach eliminates the need for large, randomly sampled libraries of stochastic differential equations and ensures that the training data directly reflects the core dynamical patterns associated with critical transitions.
We first demonstrated that a CNN-LSTM model trained on these theoretically constructed time series is capable of accurately identifying EWS in out-of-sample data. This included both synthetic time series from unseen mathematical models and real-world empirical datasets, highlighting the model’s ability to generalise beyond the training distribution.
A second objective was to assess the robustness of this framework under realistic noise conditions. To this end, we replaced white noise in the training simulations with coloured (red) noise, introducing temporal auto-correlation using AR(1) processes. We found that the performance of the DL model remains stable across a wide range of redness values.
Finally, we performed a systematic evaluation using ROC curves and their corresponding AUC scores. These experiments confirmed that models trained on coloured noise perform exceptionally well when tested on coloured stochastic time series with bifurcations. Importantly, we also showed that models trained on white noise generalise well to coloured noise settings, underscoring the flexibility of the training approach. Taken together, our results demonstrate that deep learning models trained on theoretically grounded simulations can serve as reliable tools for bifurcation detection, even in the presence of realistic stochasticity and noise correlations.

4.2. Limitation and Future Work

Nevertheless, the model and assumptions used in this study limited its scope. For instance, we focused exclusively on three specific types of local bifurcations. Global bifurcations could also be studied by expanding the training set to include these, as DL models can only recognise patterns they have been trained on. Furthermore, the study of coloured noise in our work was limited. Coloured noise refers to noise with a power spectral density (PSD) proportional to 1 / f β , where f is frequency and β is the spectral exponent. Each value of β characterises a different type of noise. While our method creates a continuum between white and red noise ( β = 0 and β = 2 ), it does not address other types of noise, such as blue noise ( β = 1 ) and violet noise ( β = 2 ). Moreover, the development of the training set relied entirely on normal forms, whereas incorporating higher-order terms could increase the accuracy of the DL model.
In future work, we aim to address these limitations. Specifically, we intend to expand the scope of this study to include other types of local and global bifurcations. Although climate empirical data predominantly rely on red noise, we plan to incorporate other types of coloured noise to create a more comprehensive classifier. Additionally, adding higher-order terms to the training set is expected to enhance the accuracy and efficiency of the DL classifier, which will be a focus of our future work.

5. Supporting Information

5.1. Model Architecture and Data Preparation

5.1.1. Generation of Training Data for the DL Classifier

We used the normal forms of three types of bifurcation with an additive stochastic term to construct the training dataset. For fold bifurcation, we used
d x d t = μ + x 2 + σ · ϵ t
The bifurcation occurs at μ = 0 . For transcritical bifurcation, we used
d x d t = r x + x 2 + σ · ϵ t
The bifurcation occurs at r = 0 . For Hopf bifurcation, we used
d z d t = ( μ + i ω ) z | z | 2 z + σ · ϵ t
The bifurcation occurs at μ = 0 . For the null case, we used
d x d t = σ · ϵ t
Here, σ is a constant, and ϵ t is a stochastic term, which can be either white noise drawn from a Gaussian distribution with a mean of zero and standard deviation of one, or red noise derived using Equation (1).
For each type of bifurcation, we simulated 5000 time series, starting from an equilibrium point far from bifurcation (e.g., μ = a random number between 3 and 2 , and x 0 = μ for fold bifurcation). The system was gradually and linearly brought closer to the bifurcation point over 1000 steps. The simulation was halted before the bifurcation point to ensure that the deep learning (DL) model was exposed only to the pre-bifurcation phase.
Once the training set was constructed using white noise (equivalent to redness = 0 ), we repeated the process for redness values of 0.1 , 0.2 , 0.3 , 0.4 , and 0.5 . The Euler–Maruyama method was used to simulate these time series with Δ t = 0.01 .
Before feeding these time series into the DL model, we pre-processed them using a Gaussian kernel. Additionally, we tested various pre-processing methods such as moving average, differencing, and Z-score normalisation, but achieved the best performance with a Gaussian kernel, where σ was set to 0.1 times the length of the time series (=100).

5.1.2. Data Requirements and Sequence Length Sensitivity

In the left panel of Figure 5, we evaluate how the amount of training data affects model performance. Starting with a dataset of 20,000 time series, we incrementally increase the fraction used for training while keeping the test set fixed, and assess the model’s performance using the F1-score. We observe a significant improvement in performance as the training fraction increases from 0.1 to 0.7, with the F1-score rising sharply in the early stages. Beyond a training fraction of 0.7, the performance plateaus, indicating that the model reaches a saturation point in learning. This behaviour suggests that the training process is robust and that a subset of the full dataset (around 70 percent) is sufficient to achieve near-optimal performance, highlighting the data efficiency of our DL framework.
In the right panel of Figure 5, we examine the impact of time series length on model precision. Using the full set of 20,000 time series, we simulate the same experimental setting across various sequence lengths to test the model’s sensitivity to temporal resolution. The results show that for relatively short time series, model precision is notably lower—likely due to insufficient temporal information to capture the dynamics preceding a bifurcation. However, as the sequence length increases, performance improves steadily, eventually plateauing beyond 800 time steps. This indicates a minimum required length for capturing the relevant precursors to tipping points. While longer sequences beyond this threshold do not significantly improve performance, they may be advantageous when training classifiers for use with higher-resolution simulated or empirical datasets.

5.1.3. Robustness of Training Process

To evaluate the stability and reproducibility of our model, we trained the same CNN-LSTM architecture 10 times using the same training dataset containing white noise and identical hyperparameter settings. Table 3 reports the mean and standard deviation of the key performance metrics—accuracy, F1-score, precision, and recall—across these repeated runs. The results show strong and consistent performance, with mean values centered around 0.82 and standard deviations ranging from 0.009 to 0.010. This low variability across identical training conditions demonstrates that the model’s performance is robust to random initialization and other sources of stochasticity inherent in the training process, reinforcing the reliability of our deep learning framework.

5.2. Deep Learning Architecture and Training Process

5.2.1. Deep Learning Model Architecture

The deep learning model employed in this study is implemented as a sequential model using the Keras API (TensorFlow 2.12.1, Keras 2.12.0), tailored for multi-class sequence classification. The architecture begins with two one-dimensional convolutional (Conv1D) layers. The first layer comprises 50 filters, followed by a second layer with 100 filters—both using a kernel size of 10, ReLU activation, and ‘same’ padding to preserve the temporal dimension of the input. These convolutional layers are followed by a dropout layer with a rate of 0.05 to mitigate overfitting. A MaxPooling1D layer with a pool size of 2 and stride of 2 is then applied to downsample the feature maps.
The convolutional block is followed by two standard, stacked unidirectional Long Short-Term Memory (LSTM) layers. The first LSTM layer includes 50 memory cells and returns sequences to preserve temporal structure for the next LSTM layer, which contains 10 units. Each LSTM layer is followed by an additional dropout layer with a rate of 0.05. The network concludes with a fully connected dense output layer with four units, corresponding to the number of classes, and uses a softmax activation function to output class probabilities. All layers utilize the LeCun kernel initializer for weight initialization.
For training, the model uses the Adam optimizer with a learning rate of 0.01. Training is conducted for up to 200 epochs with a batch size of 128. An early stopping callback is employed to halt training if the validation accuracy does not improve for 10 consecutive epochs, and a Model Checkpoint is configured to save the best-performing model based on validation accuracy. The loss function used is ‘sparse categorical crossentropy’, suitable for multi-class classification tasks with integer labels. Model performance is monitored using both ‘accuracy’ and ‘sparse categorical accuracy’ metrics during training.
After experimenting with CNN layers, LSTM layers, and Inception layers, we found that the combination of CNN and LSTM layers produced the best results. We also implemented a simplified LSTM-only model to perform an ablation study. The LSTM-only architecture consisted of two stacked LSTM layers (50 and 10 units) with the same dropout, dense output, and training procedure as the CNN–LSTM model. All models were trained using the same dataset and evaluated under consistent conditions. Our choice of hyperparameters was informed by prior work by [28], and we further refined these settings via internal experiments to ensure optimal performance. We used the pre-processed time series for each redness value, allocating 90% of the time series for the training process and 10% for validation. Figure 6 provides a detailed overview of the deep learning (DL) architecture used in this study. The learning rate was set to 0.0005, training was performed in 200 epochs, and early stopping was employed to prevent overfitting. The code is written in python 3.12.11 using TensorFlow 2.12.1. The validation accuracy at the end of the training process is provided in Table 2 for different redness values.

5.2.2. Training Process Information

Here we provide information on training process, most specifically the confusion and classification matrix for different redness values.
  • Confusion Matrices Across Redness Levels
Redness = 0 Fold Transcritical Hopf Null Fold 249 0 0 3 Transcritical 0 205 26 18 Hopf 0 28 191 14 Null 2 49 34 181
Redness = 0 . 1 Fold Transcritical Hopf Null Fold 252 1 0 1 Transcritical 0 207 12 18 Hopf 0 34 184 28 Null 4 39 32 188
Redness = 0 . 2 Fold Transcritical Hopf Null Fold 238 1 0 1 Transcritical 0 213 15 13 Hopf 0 33 183 29 Null 6 40 29 199
Redness = 0 . 3 Fold Transcritical Hopf Null Fold 228 1 0 1 Transcritical 0 245 11 15 Hopf 0 47 176 24 Null 5 46 16 185
Redness = 0 . 4 Fold Transcritical Hopf Null Fold 246 1 1 1 Transcritical 0 234 12 6 Hopf 0 39 192 23 Null 7 41 28 169
Redness = 0 . 5 Fold Transcritical Hopf Null Fold 246 0 0 1 Transcritical 0 206 15 31 Hopf 0 18 197 32 Null 5 10 24 215
  • Classification matrix
Redness = 0 P r e c i s i o n R e c a l l F 1 - Score F o l d 0.992 0.988 0.990 T r a n s c r i t i c a l 0.727 0.823 0.772 H o p f 0.761 0.820 0.789 N u l l 0.838 0.680 0.751 Redness = 0 . 1 P r e c i s i o n R e c a l l F 1 - Score F o l d 0.984 0.992 0.988 T r a n s c r i t i c a l 0.737 0.873 0.799 H o p f 0.807 0.748 0.776 N u l l 0.80 0.715 0.755
Redness = 0 . 2 P r e c i s i o n R e c a l l F 1 - Score F o l d 0.975 0.992 0.983 T r a n s c r i t i c a l 0.742 0.884 0.807 H o p f 0.806 0.747 0.775 N u l l 0.822 0.726 0.771 Redness = 0 . 3 P r e c i s i o n R e c a l l F 1 - Score F o l d 0.979 0.991 0.985 T r a n s c r i t i c a l 0.723 0.904 0.803 H o p f 0.867 0.713 0.782 N u l l 0.822 0.734 0.776
Redness = 0 . 4 P r e c i s i o n R e c a l l F 1 - Score F o l d 0.972 0.988 0.98 T r a n s c r i t i c a l 0.742 0.929 0.825 H o p f 0.824 0.756 0.789 N u l l 0.849 0.69 0.761 Redness = 0 . 5 P r e c i s i o n R e c a l l F 1 - Score F o l d 0.980 0.996 0.988 T r a n s c r i t i c a l 0.880 0.817 0.848 H o p f 0.835 0.798 0.816 N u l l 0.771 0.846 0.807

5.3. Theoretical Models Used for Testing

We have used some well-known mathematical models to test the performance of our trained DL model for providing early warning signals (EWSs) of bifurcations:
May’s Harvesting Model:
d x d t = r x ( 1 x ) h x 2 x 2 + s 2
where x represents the population size, r is the intrinsic growth rate, and h is the harvesting rate. The term x 2 x 2 + s 2 introduces a saturating effect in the harvesting process, with s controlling the steepness of the saturation. A fold bifurcation occurs as h is varied.
Neural Activation Model:
d x d t = x + S 1 + e β ( x θ )
where x is the activation level, S represents the stimulus strength, β is the steepness of the sigmoid function, and θ is the threshold. A fold bifurcation occurs as S is varied.
Rosenzweig–MacArthur Consumer–Resource Model: Transcritical Bifurcation
d x d t = r x 1 x K a x y b + x , d y d t = c a x y b + x d y
Here, x is the resource population, y is the consumer population, r is the intrinsic growth rate of the resource, K is the carrying capacity, a is the maximum predation rate, b is the half-saturation constant, c is the conversion efficiency, and d is the death rate of the consumer. A transcritical bifurcation occurs as d is varied.
Rosenzweig–MacArthur Consumer–Resource Model: Hopf Bifurcation
d x d t = r x 1 x K a x y b + x , d y d t = c a x y b + x d y
The same equations govern this case, but a Hopf bifurcation occurs when d crosses a critical threshold, leading to oscillatory dynamics.
Van der Pol Oscillator: Hopf Bifurcation
d x d t = y , d y d t = μ ( 1 x 2 ) y x
Here, x and y are the state variables, and μ is the bifurcation parameter. A Hopf bifurcation occurs at μ = 0 , where the system transitions from a stable equilibrium to sustained oscillations.
SIS Model: Transcritical Bifurcation
d S d t = β S I + γ I , d I d t = β S I γ I
In this model, S is the susceptible population, I is the infected population, β is the infection rate, and γ is the recovery rate. A transcritical bifurcation occurs at the critical threshold R 0 = β γ = 1 , where R 0 is the basic reproduction number.
All these models were simulated with an additive stochastic term. After simulation, the DL model analysed each time series using a sliding window of 100 steps, assigning probabilities to each segment. To achieve a smoother result, we applied a moving average with a window size of 100 as post-processing on the DL analysis output.

Validation on Bifurcating and Null Dynamical Systems

To evaluate the model’s ability to avoid false positives and correctly assign low confidence to bifurcation classes when no critical transition is present, we simulate time series from six nonlinear dynamical systems. In each case, control parameters are held fixed such that the system remains outside any bifurcation regime throughout the simulation. These represent purely null dynamics, meaning no tipping point occurs. This setup allows us to assess whether the classifier can reliably withhold predictions of bifurcations when presented with stable, non-transitioning systems—an essential capability for reducing false alarms in real-world applications.
  • Bifurcating Systems:
  • Rosenzweig–MacArthur Predator–Prey Model (Hopf or Transcritical):
    d x d t = r x 1 x k a x y 1 + a h x , d y d t = e a x y 1 + a h x m y
    with the fixed parameters r = 4 , k = 1.7 , e = 0.5 , h = 0.15 , and m = 2 . For the transcritical bifurcation, the predation rate a is held constant at 6, and for Hopf it is held constant at 12.
  • Fold Bifurcation Model (Harvesting):
    d x d t = x ( 1 x ) h · x 2 x 2 + 0.01
    where the harvesting rate h is held constant at 0.1 over the course of the simulation.
  • Van der Pol Oscillator (Hopf):
    d x d t = y , d y d t = μ ( 1 x 2 ) y x
    with μ = 7.0 .
  • SIS Epidemic Model (Transcritical):
    d I d t = β I ( 1 I ) γ I
    where β = 4 , γ = 5 .
  • Neural Activation Model (Fold):
    d u d t = u + r 1 + exp ( 5 ( u 1 ) )
    with r = 6 .
Null Dynamics: The null simulations are derived by fixing the control parameters in each system, preventing any bifurcation. For example, in the Rosenzweig–MacArthur model, a = 6 is held constant to avoid crossing a bifurcation threshold. Similarly, in the fold and transcritical systems, the control parameters are set below the bifurcation point and remain unchanged throughout the simulation. The results are presented in Figure 7 and Figure 8.

5.4. Empirical Systems Used for Testing

In this paper, we have used two sources of empirical data to test the DL classifier.
  • Sedimentary Archives (Mediterranean Sea)
This dataset provides high-resolution reconstructions of oxygen dynamics from sediment cores in the eastern Mediterranean Sea [52]. It captures transitions between oxic and anoxic states, with evidence of EWS preceding these transitions [53]. Data consist of molybdenum (Mo) and uranium (U) concentrations, proxies for anoxic and suboxic conditions, respectively, spanning eight anoxic events across three cores. A total of 26 time series are analyzed with a resolution of 10–50 years, depending on the core.
  • Thermoacoustic Instability (Rijke Tube Experiments)
This dataset investigates transitions to self-sustained oscillations caused by thermoacoustic instability. Experiments were conducted in a horizontal Rijke tube with controlled voltage across a heated wire mesh to trigger transitions via subcritical Hopf bifurcations [54]. Data include 19 forced trajectories with varying voltage ramp rates and 10 steady-state trajectories at fixed voltages. Downsampled data (4 kHz or 10 kHz to 2 kHz) capture 1500 points prior to transitions.
Both datasets offer insights into detecting early signs of critical transitions in distinct physical systems.

5.5. Evaluation via ROC Curve Analysis

To assess the discriminative performance of our deep learning classifier in detecting early warning signals of bifurcations, we compute Receiver Operating Characteristic (ROC) curves across a range of simulated dynamical systems. Each system is configured to either gradually approach a bifurcation point (signal present) or remain in a stable regime with fixed parameters (signal absent). This binary labeling enables quantitative evaluation of the classifier’s ability to distinguish between pre-bifurcation and null dynamics.
Each time series is then classified using two CNN-LSTM models: one trained on white noise ( r = 0 ) and one trained on coloured (red) noise generated using an AR(1) process with varying auto-correlation levels ( r [ 0 ,   0.5 ] ). The model outputs softmax probabilities for four classes: transcritical, Hopf, fold, and null.
For ROC construction, we treat the mean probability of the correct bifurcation class as the classifier score. By sweeping over 200 thresholds τ [ 0 ,   1 ] , we compute binary predictions at each level by classifying a window as “signal present” if the score exceeds τ . The true positive rate (TPR) and false positive rate (FPR) are then defined as:
TPR ( τ ) = Correctly classified signal windows Total signal windows , FPR ( τ ) = Incorrectly classified null windows Total null windows .
Varying the threshold τ traces out the ROC curve, revealing the sensitivity–specificity trade-off. To ensure robustness against stochasticity in simulation and training, this procedure is repeated over 50 Monte Carlo runs, and the resulting TPR and FPR values are averaged. The area under the ROC curve (AUC) is used to summarize performance: values close to 1 indicate strong discriminative ability, while values near 0.5 suggest random guessing. This evaluation is conducted separately for classifiers trained on white and red noise to compare their generalization under different noise structures.

5.6. Extended Evaluation of the DL Model Using Empirical Datasets

We evaluated the deep learning model on both experimental and geological datasets to demonstrate its ability to generalize from synthetic to empirical time series. Specifically, the analysis includes thermoacoustic transition signals from combustor experiments at ROF 3 and ROF 4, as well as geological anoxia indicators from sediment cores at site 64PE. Figure 9 shows the corresponding time series and the model’s bifurcation predictions.

Author Contributions

Conceptualization, M.A. and C.T.B.; Methodology, Y.B.M., D.D., M.A. and C.T.B.; Software, Y.B.M.; Validation, Y.B.M.; Formal analysis, Y.B.M., M.A. and C.T.B.; Investigation, Y.B.M.; Resources, M.A. and C.T.B.; Data curation, Y.B.M.; Writing—original draft, Y.B.M.; Writing—review & editing, Y.B.M., M.A. and C.T.B.; Visualization, Y.B.M., M.A. and C.T.B.; Supervision, M.A. and C.T.B. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Natural Sciences and Engineering Research Council of Canada (NSERC) Discovery Grants to Chris T. Bauch and Madhur Anand.

Data Availability Statement

All the codes and empirical data used in this research can be found in the following repository: https://github.com/Yazdan-Babazadeh/DL-model-EWS-coloured-noise (accessed on 18 August 2025).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Bury, T.M.; Bauch, C.T.; Anand, M. Charting pathways to climate change mitigation in a coupled socio-climate model. PLoS Comput. Biol. 2019, 15, e1007000. [Google Scholar] [CrossRef]
  2. Lenton, T.M. Land and ocean carbon cycle feedback effects on global warming in a simple Earth system model. Tellus B Chem. Phys. Meteorol. 2000, 52, 1159–1188. [Google Scholar] [CrossRef]
  3. Babazadeh, Y.; Anand, M.; Bauch, C.T. Social dynamics can delay or prevent climate tipping points by speeding the adoption of climate change mitigation. Proc. R. Soc. A 2025, in press. [Google Scholar] [CrossRef]
  4. Wang, Z.; Bauch, C.T.; Bhattacharyya, S.; d’Onofrio, A.; Manfredi, P.; Perc, M.; Perra, N.; Salathé, M.; Zhao, D. Statistical physics of vaccination. Phys. Rep. 2016, 664, 1–113. [Google Scholar] [CrossRef]
  5. Kwuimy, C.; Nazari, F.; Jiao, X.; Rohani, P.; Nataraj, C. Nonlinear dynamic analysis of an epidemiological model for COVID-19 including public behavior and government action. Nonlinear Dyn. 2020, 101, 1545–1559. [Google Scholar] [CrossRef]
  6. Babazadeh Maghsoodlo, Y.; Safaeesirat, A.; Ghanbarnejad, F. The Big Bang of an epidemic: A metapopulation approach to identify the spatiotemporal origin of contagious diseases and their universal spreading pattern. Sci. Rep. 2025, 15, 5809. [Google Scholar] [CrossRef] [PubMed]
  7. Satheesh Kumar, A.; Bauch, C.T.; Anand, M. Climate-denying rumor propagation in a coupled socio-climate model: Impact on average global temperature. PLoS ONE 2025, 20, e0317338. [Google Scholar] [CrossRef]
  8. He, Z.; Bauch, C.T. Effect of homophily on coupled behavior-disease dynamics near a tipping point. Math. Biosci. 2024, 376, 109264. [Google Scholar] [CrossRef] [PubMed]
  9. Russill, C.; Nyssa, Z. The tipping point trend in climate change communication. Glob. Environ. Chang. 2009, 19, 336–344. [Google Scholar] [CrossRef]
  10. Evangelou, N.; Cui, T.; Bello-Rivas, J.M.; Makeev, A.; Kevrekidis, I.G. Tipping points of evolving epidemiological networks: Machine learning-assisted, data-driven effective modeling. Chaos Interdiscip. J. Nonlinear Sci. 2024, 34, 063128. [Google Scholar] [CrossRef]
  11. Farahbakhsh, I.; Bauch, C.T.; Anand, M. Tipping points in coupled human–environment system models: A review. Earth Syst. Dyn. 2024, 15, 947–967. [Google Scholar] [CrossRef]
  12. Luo, D. Bifurcation Theory and Methods of Dynamical Systems; World Scientific: Singapore, 1997; Volume 15. [Google Scholar]
  13. O’Keeffe, P.E.; Wieczorek, S. Tipping phenomena and points of no return in ecosystems: Beyond classical bifurcations. SIAM J. Appl. Dyn. Syst. 2020, 19, 2371–2402. [Google Scholar] [CrossRef]
  14. Crawford, J.D. Introduction to bifurcation theory. Rev. Mod. Phys. 1991, 63, 991. [Google Scholar] [CrossRef]
  15. Jezequel, L.; Lamarque, C.H. Analysis of non-linear dynamical systems by the normal form theory. J. Sound Vib. 1991, 149, 429–459. [Google Scholar] [CrossRef]
  16. Touzé, C. Normal form theory and nonlinear normal modes: Theoretical settings and applications. In Modal Analysis of Nonlinear Mechanical Systems; Springer: Vienna, Austria, 2014; pp. 75–160. [Google Scholar]
  17. Bressan, A. Tutorial on the center manifold theorem. Hyperbolic Syst. Balance Laws 2003, 1911, 327–344. [Google Scholar]
  18. Kuznetsov, Y.A. Saddle-node bifurcation. Scholarpedia 2006, 1, 1859. [Google Scholar] [CrossRef]
  19. Müller, M.A.; Waldherr, S.; Allgöwer, F. The transcritical bifurcation in absolutely stable feedback systems. In Proceedings of the 2009 European Control Conference (ECC), Budapest, Hungary, 23–26 August 2009; IEEE: Piscataway, NJ, USA, 2009; pp. 2146–2151. [Google Scholar]
  20. Marsden, J.E.; McCracken, M. The Hopf Bifurcation and Its Applications; Springer Science & Business Media: New York, NY, USA, 2012; Volume 19. [Google Scholar]
  21. Tredicce, J.R.; Lippi, G.L.; Mandel, P.; Charasse, B.; Chevalier, A.; Picqué, B. Critical slowing down at a bifurcation. Am. J. Phys. 2004, 72, 799–809. [Google Scholar] [CrossRef]
  22. Bury, T.M.; Bauch, C.T.; Anand, M. Detecting and distinguishing tipping points using spectral early warning signals. J. R. Soc. Interface 2020, 17, 20200482. [Google Scholar] [CrossRef] [PubMed]
  23. Dakos, V.; Carpenter, S.R.; Brock, W.A.; Ellison, A.M.; Guttal, V.; Ives, A.R.; Kéfi, S.; Livina, V.; Seekell, D.A.; van Nes, E.H.; et al. Methods for detecting early warnings of critical transitions in time series illustrated using simulated ecological data. PLoS ONE 2012, 7, e41010. [Google Scholar] [CrossRef]
  24. Dakos, V.; van Nes, E.H.; Scheffer, M. Flickering as an early warning signal. Theor. Ecol. 2013, 6, 309–317. [Google Scholar] [CrossRef]
  25. Dylewsky, D.; Anand, M.; Bauch, C.T. Early warning signals for bifurcations embedded in high dimensions. Sci. Rep. 2024, 14, 18277. [Google Scholar] [CrossRef]
  26. Dakos, V.; Boulton, C.A.; Buxton, J.E.; Abrams, J.F.; Armstrong McKay, D.I.; Bathiany, S.; Blaschke, L.; Boers, N.; Dylewsky, D.; López-Martínez, C.; et al. Tipping point detection and early-warnings in climate, ecological, and human systems. EGUsphere 2023, 2023, 1–35. [Google Scholar] [CrossRef]
  27. Dakos, V.; Van Nes, E.H.; d’Odorico, P.; Scheffer, M. Robustness of variance and autocorrelation as indicators of critical slowing down. Ecology 2012, 93, 264–271. [Google Scholar] [CrossRef]
  28. Bury, T.M.; Sujith, R.; Pavithran, I.; Scheffer, M.; Lenton, T.M.; Anand, M.; Bauch, C.T. Deep learning for early warning signals of tipping points. Proc. Natl. Acad. Sci. USA 2021, 118, e2106140118. [Google Scholar] [CrossRef] [PubMed]
  29. Deb, S.; Sidheekh, S.; Clements, C.F.; Krishnan, N.C.; Dutta, P.S. Machine learning methods trained on simple models can predict critical transitions in complex natural systems. R. Soc. Open Sci. 2022, 9, 211475. [Google Scholar] [CrossRef]
  30. Bury, T.M.; Dylewsky, D.; Bauch, C.T.; Anand, M.; Glass, L.; Shrier, A.; Bub, G. Predicting discrete-time bifurcations with deep learning. Nat. Commun. 2023, 14, 6331. [Google Scholar] [CrossRef]
  31. Dylewsky, D.; Lenton, T.M.; Scheffer, M.; Bury, T.M.; Fletcher, C.G.; Anand, M.; Bauch, C.T. Universal early warning signals of phase transitions in climate systems. J. R. Soc. Interface 2023, 20, 20220562. [Google Scholar] [CrossRef]
  32. Huang, Y.; Bathiany, S.; Ashwin, P.; Boers, N. Deep learning for predicting rate-induced tipping. Nature Machine Intelligence 2024, 6, 1556–1565. [Google Scholar] [CrossRef]
  33. Dylewsky, D.; Kéfi, S.; Anand, M.; Bauch, C.T. Neural models for prediction of spatially patterned phase transitions: Methods and challenges. Theor. Ecol. 2025, 18, 1–13. [Google Scholar] [CrossRef]
  34. Pal, K.; Deb, S.; Dutta, P.S. Tipping points in spatial ecosystems driven by short-range correlated noise. Phys. Rev. E 2022, 106, 054412. [Google Scholar] [CrossRef] [PubMed]
  35. Liu, Z.; Gao, M.; Li, Z.; Liu, H. Dispersal, colored environmental noise, and spatial synchrony in population dynamics: Analyzing a discrete host–parasitoid population model. Ecol. Res. 2009, 24, 383–392. [Google Scholar] [CrossRef]
  36. Grove, M.; Timbrell, L.; Jolley, B.; Polack, F.; Borg, J.M. The importance of noise colour in simulations of evolutionary systems. Artif. Life 2021, 27, 164–182. [Google Scholar] [CrossRef] [PubMed]
  37. Gilljam, D.; Knape, J.; Lindén, A.; Mugabo, M.; Sait, S.M.; Fowler, M.S. The colour of environmental fluctuations associated with terrestrial animal population dynamics. Glob. Ecol. Biogeogr. 2019, 28, 118–130. [Google Scholar] [CrossRef]
  38. Spanio, T.; Hidalgo, J.; Muñoz, M.A. Impact of environmental colored noise in single-species population dynamics. Phys. Rev. E 2017, 96, 042301. [Google Scholar] [CrossRef]
  39. Fung, T.; O’Dwyer, J.P.; Chisholm, R.A. Species-abundance distributions under colored environmental noise. J. Math. Biol. 2017, 74, 289–311. [Google Scholar] [CrossRef] [PubMed]
  40. Lögdberg, F.; Wennergren, U. Spectral color, synchrony, and extinction risk. Theor. Ecol. 2012, 5, 545–554. [Google Scholar] [CrossRef]
  41. van de Pol, M.; Vindenes, Y.; Sæther, B.E.; Engen, S.; Ens, B.J.; Oosterbeek, K.; Tinbergen, J.M. Poor environmental tracking can make extinction risk insensitive to the colour of environmental noise. Proc. R. Soc. B Biol. Sci. 2011, 278, 3713–3722. [Google Scholar] [CrossRef]
  42. Ruokolainen, L.; Lindén, A.; Kaitala, V.; Fowler, M.S. Ecological and evolutionary dynamics under coloured environmental variation. Trends Ecol. Evol. 2009, 24, 555–563. [Google Scholar] [CrossRef]
  43. Petchey, O.L.; Gonzalez, A.; Wilson, H.B. Effects on population persistence: The interaction between environmental noise colour, intraspecific competition and space. Proc. R. Soc. London. Ser. B Biol. Sci. 1997, 264, 1841–1847. [Google Scholar] [CrossRef]
  44. Ripa, J.; Lundberg, P. Noise colour and the risk of population extinctions. Proc. R. Soc. London. Ser. B Biol. Sci. 1996, 263, 1751–1753. [Google Scholar]
  45. Vasseur, D.A. Populations embedded in trophic communities respond differently to coloured environmental noise. Theor. Popul. Biol. 2007, 72, 186–196. [Google Scholar] [CrossRef]
  46. Sainath, T.N.; Vinyals, O.; Senior, A.; Sak, H. Convolutional, long short-term memory, fully connected deep neural networks. In Proceedings of the 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), South Brisbane, QLD, Australia, 19–24 April 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 4580–4584. [Google Scholar]
  47. Alzubaidi, L.; Zhang, J.; Humaidi, A.J.; Al-Dujaili, A.; Duan, Y.; Al-Shamma, O.; Santamaría, J.; Fadhel, M.A.; Al-Amidie, M.; Farhan, L. Review of deep learning: Concepts, CNN architectures, challenges, applications, future directions. J. Big Data 2021, 8, 53. [Google Scholar] [CrossRef]
  48. Hochreiter, S. Long Short-Term Memory; Neural Computation MIT-Press: Cambridge, MA, USA, 1997. [Google Scholar]
  49. Lu, W.; Li, J.; Li, Y.; Sun, A.; Wang, J. A CNN-LSTM-based model to forecast stock prices. Complexity 2020, 2020, 6622927. [Google Scholar] [CrossRef]
  50. Madaeni, F.; Chokmani, K.; Lhissou, R.; Homayouni​​​​​​​, S.; Gauthier, Y.; Tolszczuk-Leclerc, S. Convolutional neural network and long short-term memory models for ice-jam predictions. Cryosphere 2022, 16, 1447–1468. [Google Scholar] [CrossRef]
  51. Hussain, S.N.; Aziz, A.A.; Hossen, M.J.; Aziz, N.A.A.; Murthy, G.R.; Mustakim, F.B. A novel framework based on cnn-lstm neural network for prediction of missing values in electricity consumption time-series datasets. J. Inf. Process. Syst. 2022, 18, 115–129. [Google Scholar]
  52. Hennekam, R.; Van der Bolt, B.; van Nes, E.; de Lange, G.J.; Scheffer, M.; Reichart, G.J. Calibrated XRF-Scanning Data (mm Resolution) and Calibration Data (ICP-OES and ICP-MS) for Elements Al, Ba, Mo, Ti, and U in Mediterranean Cores MS21, MS66, and 64PE406E1; Netherlands Institute for Sea Research (NIOZ): Texel, The Netherlands, 2020. [Google Scholar]
  53. Hennekam, R.; van der Bolt, B.; van Nes, E.H.; de Lange, G.J.; Scheffer, M.; Reichart, G.J. Early-warning signals for marine anoxic events. Geophys. Res. Lett. 2020, 47, e2020GL089183. [Google Scholar] [CrossRef]
  54. Pavithran, I.; Sujith, R. Effect of rate of change of parameter on early warning signals for critical transitions. Chaos Interdiscip. J. Nonlinear Sci. 2021, 31, 013116. [Google Scholar] [CrossRef] [PubMed]
Figure 1. DL model providing EWS prior to bifurcation points for Rosenzweig–MacArthur and May’s harvesting model. (AC) Synthetic time series from three canonical bifurcation models with additive coloured noise: (A) Rosenzweig–MacArthur predator–prey system near a transcritical bifurcation; (B) Rosenzweig–MacArthur with Hopf bifurcation; (C) May’s harvesting model near a fold bifurcation. (DF) DL model predictions for the systems in (AC), showing temporal evolution of bifurcation class probabilities. Probabilities are represented with distinguishable markers and line styles for print clarity: triangle (fold), square (transcritical), circle (Hopf), and x (null). The vertical dashed line indicates the approximate location of the tipping point.
Figure 1. DL model providing EWS prior to bifurcation points for Rosenzweig–MacArthur and May’s harvesting model. (AC) Synthetic time series from three canonical bifurcation models with additive coloured noise: (A) Rosenzweig–MacArthur predator–prey system near a transcritical bifurcation; (B) Rosenzweig–MacArthur with Hopf bifurcation; (C) May’s harvesting model near a fold bifurcation. (DF) DL model predictions for the systems in (AC), showing temporal evolution of bifurcation class probabilities. Probabilities are represented with distinguishable markers and line styles for print clarity: triangle (fold), square (transcritical), circle (Hopf), and x (null). The vertical dashed line indicates the approximate location of the tipping point.
Mathematics 13 02782 g001
Figure 2. DL model providing EWS prior to bifurcation points for Van der Pol, SIS, and neural activation model. (AC) Synthetic time series from three canonical bifurcation models with additive white noise: (A) Van der Pol oscillator approaching a Hopf bifurcation; (B) SIS model undergoing a transcritical bifurcation; (C) neural activation model with a fold bifurcation. The vertical dashed lines indicate the location of the bifurcation. (DF) Output of the DL classifier showing predicted bifurcation probabilities over time for the corresponding systems in (AC). Distinct line styles and markers denote bifurcation types: triangle (fold), square (transcritical), circle (Hopf), and x (null).
Figure 2. DL model providing EWS prior to bifurcation points for Van der Pol, SIS, and neural activation model. (AC) Synthetic time series from three canonical bifurcation models with additive white noise: (A) Van der Pol oscillator approaching a Hopf bifurcation; (B) SIS model undergoing a transcritical bifurcation; (C) neural activation model with a fold bifurcation. The vertical dashed lines indicate the location of the bifurcation. (DF) Output of the DL classifier showing predicted bifurcation probabilities over time for the corresponding systems in (AC). Distinct line styles and markers denote bifurcation types: triangle (fold), square (transcritical), circle (Hopf), and x (null).
Mathematics 13 02782 g002
Figure 3. DL model providing EWS before bifurcation points. (AD) Real-world time series examples: (A,B) thermoacoustic instability signals from combustor experiments at ROF 5 and ROF 10; (C,D) geological anoxia indicators from sediment cores at two locations. (EH) Corresponding DL-based bifurcation predictions for (AD), with line styles and markers indicating the inferred probability of each class. This illustrates how the model generalizes from synthetic to empirical data. Distinct line styles and markers denote bifurcation types: triangle (fold), square (transcritical), circle (Hopf), and x (null). The vertical dashed lines are the approximate location of the tipping point.
Figure 3. DL model providing EWS before bifurcation points. (AD) Real-world time series examples: (A,B) thermoacoustic instability signals from combustor experiments at ROF 5 and ROF 10; (C,D) geological anoxia indicators from sediment cores at two locations. (EH) Corresponding DL-based bifurcation predictions for (AD), with line styles and markers indicating the inferred probability of each class. This illustrates how the model generalizes from synthetic to empirical data. Distinct line styles and markers denote bifurcation types: triangle (fold), square (transcritical), circle (Hopf), and x (null). The vertical dashed lines are the approximate location of the tipping point.
Mathematics 13 02782 g003
Figure 4. AUC scores of white noise and red noise classifiers across six mathematical models Classifier performance across six dynamical systems under varying levels of noise correlation (redness). Each subplot displays the AUC score as a function of redness for white noise conditions (WNCs, blue circles) and red noise conditions (RNCs, orange squares). Top row: Rosenzweig–MacArthur model with transcritical and Hopf bifurcations, and May’s harvesting model with a fold bifurcation. Bottom row: SIS model (transcritical), Van der Pol oscillator (Hopf), and neural activation model (fold). The results highlight that while most systems retain high classification accuracy across redness levels, performance in certain fold bifurcations—especially under WNCs in the neural activation model and under RNCs in the May model—shows degradation as temporal correlation increases.
Figure 4. AUC scores of white noise and red noise classifiers across six mathematical models Classifier performance across six dynamical systems under varying levels of noise correlation (redness). Each subplot displays the AUC score as a function of redness for white noise conditions (WNCs, blue circles) and red noise conditions (RNCs, orange squares). Top row: Rosenzweig–MacArthur model with transcritical and Hopf bifurcations, and May’s harvesting model with a fold bifurcation. Bottom row: SIS model (transcritical), Van der Pol oscillator (Hopf), and neural activation model (fold). The results highlight that while most systems retain high classification accuracy across redness levels, performance in certain fold bifurcations—especially under WNCs in the neural activation model and under RNCs in the May model—shows degradation as temporal correlation increases.
Mathematics 13 02782 g004
Figure 5. (Left) Model performance as a function of training data fraction. (Right) Precision as a function of time series length.
Figure 5. (Left) Model performance as a function of training data fraction. (Right) Precision as a function of time series length.
Mathematics 13 02782 g005
Figure 6. Deep learning architecture.
Figure 6. Deep learning architecture.
Mathematics 13 02782 g006
Figure 7. (AC) Synthetic null time series from three canonical bifurcation models with additive white noise all far from the bifurcation point: (A) Rosenzweig–MacArthur predator–prey system, (B) Rosenzweig–MacArthur, and (C) May’s harvesting model. (DF) Deep learning model predictions for the systems in (AC), showing temporal evolution of bifurcation class probabilities. Probabilities are represented with distinguishable markers and line styles for print clarity: triangle (fold), square (transcritical), circle (Hopf), and x (null). In all subplots, the null class consistently exhibits the highest predicted probability compared to the other bifurcation types.
Figure 7. (AC) Synthetic null time series from three canonical bifurcation models with additive white noise all far from the bifurcation point: (A) Rosenzweig–MacArthur predator–prey system, (B) Rosenzweig–MacArthur, and (C) May’s harvesting model. (DF) Deep learning model predictions for the systems in (AC), showing temporal evolution of bifurcation class probabilities. Probabilities are represented with distinguishable markers and line styles for print clarity: triangle (fold), square (transcritical), circle (Hopf), and x (null). In all subplots, the null class consistently exhibits the highest predicted probability compared to the other bifurcation types.
Mathematics 13 02782 g007
Figure 8. (AC) Synthetic null time series from three canonical bifurcation models with additive white noise all far from the bifurcation point: (A) Van der Pol system, (B) SIS model, and (C) neural activation model. (DF) Deep learning model predictions for the systems in (AC), showing temporal evolution of bifurcation class probabilities. Probabilities are represented with distinguishable markers and line styles for print clarity: triangle (fold), square (transcritical), circle (Hopf), and x (null). The null class consistently exhibits the highest predicted probability in all subplots compared to the other bifurcation types.
Figure 8. (AC) Synthetic null time series from three canonical bifurcation models with additive white noise all far from the bifurcation point: (A) Van der Pol system, (B) SIS model, and (C) neural activation model. (DF) Deep learning model predictions for the systems in (AC), showing temporal evolution of bifurcation class probabilities. Probabilities are represented with distinguishable markers and line styles for print clarity: triangle (fold), square (transcritical), circle (Hopf), and x (null). The null class consistently exhibits the highest predicted probability in all subplots compared to the other bifurcation types.
Mathematics 13 02782 g008
Figure 9. Deep learning model providing early warning signals (EWSs) prior to bifurcation points. (AD) Real-world time series examples: (A,B) thermoacoustic instability signals from combustor experiments at ROF 3 and ROF 4; (C,D) geological anoxia indicators from sediment cores at two locations. (EH) Corresponding deep learning-based bifurcation predictions for (AD), with line styles and markers indicating the inferred probability of each class. This illustrates how the model generalizes from synthetic to empirical data. The vertical dashed lines are the approximate location of the tipping point.
Figure 9. Deep learning model providing early warning signals (EWSs) prior to bifurcation points. (AD) Real-world time series examples: (A,B) thermoacoustic instability signals from combustor experiments at ROF 3 and ROF 4; (C,D) geological anoxia indicators from sediment cores at two locations. (EH) Corresponding deep learning-based bifurcation predictions for (AD), with line styles and markers indicating the inferred probability of each class. This illustrates how the model generalizes from synthetic to empirical data. The vertical dashed lines are the approximate location of the tipping point.
Mathematics 13 02782 g009
Table 1. Comparison of overall performance metrics between CNN-LSTM and LSTM-only models.
Table 1. Comparison of overall performance metrics between CNN-LSTM and LSTM-only models.
MetricCNN-LSTMLSTM-Only
Accuracy Score0.8230.773
F1-Score0.8220.769
Precision0.8340.778
Recall0.8230.773
Table 2. Model performance metrics across redness levels.
Table 2. Model performance metrics across redness levels.
RednessValidation AccuracyAccuracy ScoreF1-ScorePrecisionRecall
0.00.8520.8260.8250.8290.827
0.10.8570.8310.8290.8320.832
0.20.8570.8330.8340.8360.837
0.30.8530.8340.8360.84760.8355
0.40.8310.8410.8380.8470.84
0.50.8320.8640.8640.8660.864
Table 3. Model performance metrics with mean and standard deviation over multiple runs.
Table 3. Model performance metrics with mean and standard deviation over multiple runs.
MetricMean ScoreStandard Deviation
Accuracy0.819±0.009
F1-Score0.819±0.010
Precision0.826±0.010
Recall0.821±0.009
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Babazadeh Maghsoodlo, Y.; Dylewsky, D.; Anand, M.; Bauch, C.T. Deep Learning for Bifurcation Detection: Extending Early Warning Signals to Dynamical Systems with Coloured Noise. Mathematics 2025, 13, 2782. https://doi.org/10.3390/math13172782

AMA Style

Babazadeh Maghsoodlo Y, Dylewsky D, Anand M, Bauch CT. Deep Learning for Bifurcation Detection: Extending Early Warning Signals to Dynamical Systems with Coloured Noise. Mathematics. 2025; 13(17):2782. https://doi.org/10.3390/math13172782

Chicago/Turabian Style

Babazadeh Maghsoodlo, Yazdan, Daniel Dylewsky, Madhur Anand, and Chris T. Bauch. 2025. "Deep Learning for Bifurcation Detection: Extending Early Warning Signals to Dynamical Systems with Coloured Noise" Mathematics 13, no. 17: 2782. https://doi.org/10.3390/math13172782

APA Style

Babazadeh Maghsoodlo, Y., Dylewsky, D., Anand, M., & Bauch, C. T. (2025). Deep Learning for Bifurcation Detection: Extending Early Warning Signals to Dynamical Systems with Coloured Noise. Mathematics, 13(17), 2782. https://doi.org/10.3390/math13172782

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop