2.1. Energy Dissipation Mutation Criterion
During the progressive failure of slopes, the energy field within the slope system and its surrounding environment remains in a state of dynamic equilibrium [
15]. Studying slope failure from the perspective of energy evolution allows us to bypass complex stress–strain analyses and avoid tracking detailed damage states of geomaterials at each intermediate step. Instead, the focus shifts to the pathways of energy input, storage, transformation, and dissipation. In slope systems, part of the externally applied energy is stored as elastic strain energy within the material, whilst the remaining energy is gradually dissipated as damage-induced energy through deformation and damage processes [
16,
17]. According to the first law of thermodynamics, this process can be expressed as follows:
Under static equilibrium, the energy input is primarily driven by the self-weight of the slope. Thus, Equation (1) can be rewritten as follows:
where
represents the total energy input;
denotes the dissipated energy;
is the elastic strain energy; and
stands for the gravitational potential energy.
Both the storage of elastic strain energy and the dissipation of energy through deformation in a slope have inherent capacity limits. When the total energy input from external sources exceeds the slope’s energy-bearing threshold, the surplus energy that cannot be stored or dissipated rapidly transforms into kinetic energy. At this point, the slope reaches a state of limit equilibrium, triggering sliding and leading to failure. By simplifying slope stability analysis to the judgment of energy evolution patterns, the critical conditions for slope instability can be identified more efficiently and accurately. The energy evolution process of the slope can then be expressed by the following equation:
where
is the kinetic energy.
Taking any volumetric element within the slope system as the research object, the gravitational potential energy
, dissipated energy
elastic strain energy,
and kinetic energy
can be calculated using the following formulas:
where
represents the density;
is the gravitational acceleration;
is the height of the element’s centroid;
is the volume of the element;
is the stress tensor; “:” is the inner product operation;
is the plastic strain tensor;
,
, and
are the principal stresses in the x, y, and z directions, respectively;
,
, and
are the principal strains in the x, y, and z directions, respectively;
is the elastic modulus of the material;
is the Poisson’s ratio of the material; and
is the velocity at the element’s centroid.
The energy dissipation process of slopes exhibits significant differences between unstable and stable states. During the stable phase, the dissipated energy typically follows a relatively smooth trend, with fluctuations within a certain range when influenced by external factors. However, as the slope approaches the critical state of instability, the internal structure of the rock and soil undergoes drastic changes, such as rapid crack propagation and imbalances in particle interactions, causing the energy dissipation to deviate from its original smooth pattern and exhibit abrupt changes [
18]. Although the theoretical understanding of the distinct energy dissipation behaviors in different slope states is clear, accurately identifying the mutation point is crucial for the precision of subsequent safety factor calculations. Previous studies have attempted to locate the mutation point mathematically by constructing regression equations between reduction coefficients and dissipated energy [
19]. However, these methods struggle to achieve precise identification in complex scenarios involving dual reduction coefficients due to the inherent limitations of the equations and the complexity of multi-variable interactions.
Unsupervised clustering algorithms [
20], as important techniques in the field of data mining, can achieve automatic classification based on similarity metrics among data points under the condition of unlabeled datasets. This approach does not require pre-setting target variables and mines the inherent data structure with high adaptability. Common unsupervised clustering algorithms have distinct characteristics. The K-means algorithm realizes data partitioning by iteratively updating cluster centers, featuring high computational efficiency and rapid convergence, which is suitable for datasets with uniform sample distribution and approximately spherical cluster shapes; hierarchical clustering algorithms intuitively present the hierarchical relationships of data through a dendrogram structure, allowing flexible adjustment of the number of clusters, especially suitable for exploratory data analysis and scenarios requiring the revelation of clustering hierarchical relationships; the Bayesian Gaussian Mixture Model [
21] (BGMM), based on probability theory, completes clustering by estimating the probability that data belongs to each Gaussian distribution. It can not only output the probability values of samples belonging to different categories but also effectively addresses data uncertainty and fuzziness, which has been widely applied in fields such as image recognition and speech processing.
In slope stability analysis, compared with the method of constructing regression equations between reduction coefficients and dissipated energy to find mathematical mutation points, unsupervised clustering methods in the field of machine learning show unique potential in identifying dissipated energy mutations under dual-parameter reduction. The stability (or instability) of slopes is essentially a probabilistic problem—with minor changes in reduction coefficients, their mechanical responses (such as dissipated energy) exhibit continuous gradual changes. Near the critical region, dissipated energy data points do not abruptly jump from “stable” to “unstable” but exist in a fuzzy region with overlapping states, reflecting the uncertainty of state evolution in slope systems. The core advantage of the Bayesian Gaussian Mixture Model lies in modeling data as a weighted mixture of several probability distributions, which can accurately describe the overlapping data distributions generated by different potential states. In this study, “stable” and “unstable” are regarded as two potential states. By assigning each data point the probability (membership degree) of belonging to the “stable class” and “unstable class”, BGMM perfectly matches the physical essence of state uncertainty in the critical region. Therefore, this study uses this model to perform clustering analysis on slope dissipated energy data, achieving a scientific classification of slope states into two categories: “stable” and “unstable”. To objectively demonstrate the applicability of BGMM,
Table 1 systematically compares the core characteristics of each algorithm.
2.2. Gaussian Mixture Model (GMM)
The Gaussian Mixture Model (GMM) is a probabilistic model and an unsupervised clustering machine learning algorithm [
22]. It assumes that all data points
are independently sampled from the same probability distribution. Its probability density function can be expressed as a linear combination of multiple Gaussian functions, which are as follows:
where
represents the number of Gaussian distributions, i.e., the number of mixture components;
is the prior probability, i.e., the mixing coefficient;
is the probability density function of the k-th Gaussian distribution, where
is the mean vector of the k-th Gaussian distribution; and
is the covariance matrix of the k-th Gaussian distribution.
The specific form of the probability density function
for the Gaussian distribution is as follows:
where
is the dimensionality of the data point x;
is the determinant of the covariance matrix
; and
is the inverse matrix of the covariance matrix
.
The core idea of the Gaussian Mixture Model lies in the assumption that the data is generated from a mixture of multiple Gaussian distributions, with each Gaussian distribution corresponding to a cluster. The model fits the probability distribution of the data by estimating the mean, covariance matrix, and mixing coefficient for each cluster. Its operational mechanism involves first identifying the Gaussian function center corresponding to each sample point, then calculating the posterior probability of the sample point belonging to each cluster, and finally assigning the sample point to the appropriate cluster based on these probabilities [
23]. In practical implementation, within the Gaussian Mixture Model the mean, covariance matrix, and mixing coefficient for each Gaussian distribution can be estimated using either the maximum likelihood estimation method or the Expectation Maximization (EM) algorithm. Once these parameters have been estimated, the sample points can be assigned to the cluster corresponding to the Gaussian function with the highest probability density by comparing the probability density of each sample point across the Gaussian distributions.
The likelihood estimation method is a commonly used statistical inference approach. Its fundamental idea is that, given a probability model and a set of observed data, the goal of likelihood estimation is to find the model parameter values that are most likely to have generated the observed data. The likelihood function is the core tool of likelihood estimation, representing the probability of observing the data given the parameter values. It is typically denoted as
, where
are the unknown parameters of the model, and the observed data is
, with the probability density function
. The expression for the likelihood function is as follows:
By replacing the probability density function
with the probability density function of the Gaussian Mixture Model, the likelihood function of the Gaussian Mixture Model can be derived as follows:
2.3. Bayesian Estimation Algorithm
The likelihood equation of the Gaussian Mixture Model typically lacks an analytical solution. Combining the Bayesian algorithm for iterative computation is one method for solving the Gaussian Mixture Model. Bayesian estimation [
24], as a statistical inference method based on Bayes’ theorem, is fundamentally different from the traditional frequentist approach. In the Bayesian estimation framework, the unknown parameters of the model are treated as random variables, unlike the traditional frequentist approach, which considers them as fixed but unknown constants. The key point lies in accurately calculating the posterior distribution of the parameters based on prior knowledge and observed data. Here, the prior distribution carries the accumulated knowledge and assumptions about the parameters before observing the data. Subsequently, using Bayes’ formula the prior distribution and the likelihood function of the observed data are integrated to derive the posterior distribution. Finally, parameter estimation and inference are conducted based on the obtained posterior distribution. The specific steps for estimating the parameters of the Gaussian Mixture Model using the Bayesian algorithm are as follows.
In the Gaussian Mixture Model, the parameters to be estimated are the mixing coefficient
, the mean vector
, and the covariance matrix
. The probability density functions of their prior distributions can be expressed as follows:
where the mixing coefficient follows a Dirichlet distribution,
, where
is the parameter vector of the Dirichlet distribution; the mean vector follows a normal distribution,
, where
is the prior mean and
is the prior covariance; the covariance matrix follows an inverse Wishart distribution,
, where
and
are the parameters of the inverse Wishart distribution; and
is the data dimensionality.
Assuming a set of observed vectors as
, and letting the unknown parameter be
, the posterior probability of
can be expressed as follows:
where
is the posterior probability of
;
is the probability density function of
;
is the prior estimate of
; and
is the observed probability of
.
Under the given observed event
,
is a measurable constant, and
can be considered the likelihood function of event
. Thus, the posterior probability
is proportional to
, which is shown as follows:
From Equation (17), the posterior distribution of the mixing coefficient is , where , and is the number of samples belonging to the k-th Gaussian component. The posterior distribution of the mean vector is where and . The posterior distribution of the covariance matrix is , where and .
By calculating the expected value of the posterior distribution, the parameter estimates can be obtained. The specific principle of the Bayesian estimation algorithm for computing the parameter values of the Gaussian Mixture Model is illustrated in
Figure 1: