Next Article in Journal
Microplastics in Inland Saline Lakes of the Central Ebro Basin, NE Spain
Previous Article in Journal
Influence of Precipitation on the Estimation of Karstic Water Storage Variation
Previous Article in Special Issue
Reliability Analysis of the Bearing Performance of Corroded Piles Subjected to Scour Action
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Application of a Multi-Teacher Distillation Regression Model Based on Clustering Integration and Adaptive Weighting in Dam Deformation Prediction

1
POWERCHINA Guiyang Engineering Corporation Limited, Guiyang 550000, China
2
College of Big Data and Information Engineering, Guizhou University, Guiyang 550025, China
*
Author to whom correspondence should be addressed.
Water 2025, 17(7), 988; https://doi.org/10.3390/w17070988
Submission received: 25 February 2025 / Revised: 18 March 2025 / Accepted: 25 March 2025 / Published: 27 March 2025

Abstract

:
Deformation is a key physical quantity that reflects the safety status of dams. Dam deformation is influenced by multiple factors and has seasonal and periodic patterns. Due to the challenges in accurately predicting dam deformation with traditional linear models, deep learning methods have been increasingly applied in recent years. In response to the problems such as an excessively long training time, too-high model complexity, and the limited generalization ability of a large number of complex hybrid models in the current research field, we propose an improved multi-teacher distillation network for regression tasks to improve the performance of the model. The multi-teacher network is constructed using a Transformer that considers global dependencies, while the student network is constructed using Temporal Convolutional Network (TCN). To improve distillation efficiency, we draw on the concept of clustering integration to reduce the number of teacher networks and propose a loss function for regression tasks. We incorporate an adaptive weight module into the loss function and assign more weight to the teachers with more accurate prediction results. Finally, knowledge information is formed based on the differences between the teacher networks and the student network. The model is applied to a concrete-faced rockfill dam located in Guizhou province, China, and the results demonstrate that, compared to other knowledge distillation methods, this approach exhibits higher accuracy and practicality.

1. Introduction

As the core structure of hydraulic engineering, dams play a critical role not only in flood control, power generation, and water supply but also in effectively regulating the water environment and aquatic ecosystems. They have a significant impact on the safety of life and property, economic development, and the ecological environment of downstream populations [1]. Over time, the structural performance of dams inevitably degrades, and, in the event of a failure, the consequences are often catastrophic, leading to substantial loss of life and property for downstream residents. In fact, most dam failures are not sudden events but rather a gradual evolutionary process [2]. Therefore, it is essential to adopt reasonable and effective methods to monitor and evaluate the operational conditions of dams, with deformation prediction being a crucial component of the monitoring process [3,4,5]. The comparison between predicted values and actual measurements can help identify abnormal conditions in a timely manner, enabling prompt response measures and ensuring the safe operation of dams [6,7].
Traditional dam deformation prediction methods are predominantly based on regression-based statistical models. These statistical approaches primarily establish functional relationships between effect quantities (such as deformation, seepage, and stress) and independent variables (such as water level, temperature, and time). Due to their simplicity and ease of implementation, they have been widely used in the field of dam safety monitoring [8,9,10]. However, during the actual operation of a dam, the relationship between the independent variables and the effect quantities is often extremely complex, exhibiting strong nonlinear characteristics. Consequently, traditional linear models are unable to accurately describe such nonlinear relationships, which further restricts the prediction accuracy. With the rapid development of machine learning and deep learning, techniques including but not limited to artificial neural networks, support vector machines, random forests, and genetic algorithms have been applied to dam deformation analysis and prediction. These algorithms exhibit significant advantages in addressing nonlinear problems and have demonstrated promising results [11,12,13]. Yang et al. employed long short-term memory (LSTM) networks to address the long-term dependencies present in dam deformation [14]. Building on this foundation, Yang et al. introduced Convolutional Neural Networks (CNNs) to identify and learn the spatial relationships within-time series data, developing a reliable dam prediction model [15]. Jiedeerbieke et al. proposed a dam deformation prediction model based on clustering partitioning and bidirectional long short-term memory (BiLSTM) networks. By considering the intrinsic correlations among monitoring points and integrating multiple feature information, the model achieves more comprehensive deformation monitoring [16]. Su et al. proposed a data processing framework that uses a long short-term memory (LSTM) model coupled with an attention mechanism to predict the deformation response of a dam structure in order to obtain more accurate nonlinear prediction results [17]. Zhou et al. established a dam displacement prediction model based on a multi-expert network. The model employs long short-term memory (LSTM) to extract temporal features and utilizes a multi-head attention mechanism to capture adjacency values, constructing a spatial graph to describe spatial relationships within the multi-expert network. Additionally, graph convolutional networks (GCNs) are applied to extract temporal relationships, enabling the prediction of multi-point dam displacements [18]. Wang et al. proposed a hybrid algorithm that combines backpropagation with a genetic algorithm (GA-BP) and a multiple population genetic algorithm (MPGA) to optimize the BP neural network. This hybrid algorithm has improved convergence speed and prediction accuracy [19]. The current research shows that hybrid neural networks often achieve better performance than single-structure models. However, this improvement in performance comes at a cost. As the model structure becomes more complex, the number of parameters in the neural network also increases significantly. This means that more computational resources are required, and longer training times are needed. The increase in computational overhead directly affects the application efficiency of the model, especially in real-world application scenarios with extremely high real-time requirements. Therefore, it is necessary to design a method that can significantly reduce the number of model parameters while improving the model’s accuracy, thus shortening the training and inference times.
The principle of knowledge distillation (KD) [20] is effectively optimizin the original model by developing a lightweight, efficient, and small model. A lightweight student network is used to mimic the behavior and output of a powerful teacher network. However, currently, knowledge distillation has mostly been applied to classification problems [21,22,23,24], where knowledge has been transferred through the soft labels generated by the teacher model. Nevertheless, this rule is not applicable in regression problems.
In response to the above-mentioned problems, this paper proposes a knowledge distillation method for regression problems, which is used for dam deformation prediction. This method consists of multiple modules, such as the pre-training of teacher networks, clustering integration based on information theory, the knowledge distillation process, and the prediction module. The main content is as follows: First, considering the robustness of teacher networks, we pre-trained teacher networks based on the original data to expand the teacher team. Subsequently, we adopted the clustering ensemble method to effectively reduce the number of redundant and low-precision teacher networks, thus avoiding the confusion in guidance caused by an excessive number of teacher networks with varying qualities during the subsequent distillation process. Then, we used the proposed weighted loss function for knowledge transfer to train a lightweight student network. Finally, the high-precision prediction of the student network was achieved. We used an actual concrete-faced rockfill dam as the experimental object, and the experimental results showed that the proposed model had higher accuracy and practicality in specific engineering cases.
Based on the above description, the contributions of this study are as follows:
( 1 )
A multi-teacher distillation framework for regression problems is proposed. Traditional knowledge distillation methods have certain limitations when dealing with regression problems. We provide a new paradigm for this field. With the help of our framework, it is possible to provide the student network with richer and more comprehensive knowledge information.
( 2 )
We adopted a clustering integration method based on information theory to reduce the number of redundant and low-precision teacher networks. This avoids knowledge redundancy and conflicts caused by too many teacher networks, further enhancing the performance and stability of the model.
( 3 )
We propose of a novel weighted loss function for regression to guide the student network. Different weights are assigned according to the performance of different teacher networks, enabling the student network to pay more attention to the knowledge content of high-quality teachers during the learning process.
( 4 )
Our method was applied to a concrete-faced rockfill dam in Guizhou province, China. The results show that, through our distillation algorithm, the prediction accuracy (MSE) of the student network can be effectively reduced by 40% to 73%. Compared with other knowledge distillation methods, our method has higher accuracy and practicality.
Next, an outline of the remaining parts of this paper is presented. Section 2 introduces the theoretical basis and traditional knowledge distillation (KD) methods. In Section 3, we present the proposed method. Section 4 reports the experimental results. In Section 5, the conclusions and future work are presented.

2. Related Works

2.1. HST Statistical Model

First, we briefly introduce the classic HST statistical model [25], According to the current knowledge system and relevant experience, the displacement of a dam is mainly related to three major influencing factors, namely, the upstream water level H, the temperature T, and the aging θ . We further express this as the water pressure component y H , the temperature y T , the aging component y θ , and the constant term c as follows:
y = y H + y T + y θ + c
The water pressure component y H of the horizontal displacement y of a gravity dam can be expressed by the upstream water levels H as follows:
y H = i = 1 3 a i H i
where a i ( i = 1 ,   2 ,   3 ) are the regression coefficients.
The actual situation of a dam is often complex. Since it is difficult to install a sufficient number of thermometers on the bedrock and the dam body, when the internal temperature field of the dam reaches a stable state, we usually use multi-cycle harmonic variations as the factors influencing the temperature component y T as follows:
y H = i = 1 3 a i H i y T = i = 1 2 b 1 i sin 2 π i t 365 + b 2 i cos 2 π i t 365
where b 1 i and b 2 i represent regression coefficients, and t represents the number of days from the detection date to the initial monitoring date.
The aging component y θ of dam deformation is irreversible. For most dams, the displacement caused by aging develops rapidly during the initial operation period but tends to stabilize in the later stages. Therefore, y θ is usually generalized as a combination of a logarithmic function and a linear function as follows:
y θ = c 1 θ + c 2 ln θ
where θ = t/100; t is calculated as the cumulative number of days from the monitoring day to the start of the monitoring period. Combining Equations (1)–(4), the complete mathematical expression of the HST model is as follows:
y = i = 1 3 a i H i + i = 1 2 ( b 1 i s i n 2 π i t 365 + b 2 i c o s 2 π i t 365 ) + c 1 θ + c 2 ln θ + c
where the other symbols have the same meanings as in the previous equation.

2.2. Transformer Model

This section provides a comprehensive introduction to the Transformer. The Transformer [26] model was initially designed to address the challenges in natural language processing (NLP). Traditional models perform poorly in handling long-range dependencies. In contrast, the Transformer, with its unique architecture and innovative mechanisms, demonstrates outstanding performance and has extended its influence to various fields, including time-series prediction [27,28,29,30,31]. In time-series data, accurately capturing long-term dependencies and complex patterns is crucial for improving prediction accuracy. The Transformer can precisely meet this requirement. By modeling the relationships between different time steps in the time series, it can make more accurate predictions. The Transformer mainly consists of two core components: the encoder and the decoder. The encoder module is used to capture the complex relationships and contextual information among the elements in the input sequence, laying the foundation for subsequent processing. The decoder module focuses on feature reconstruction. Figure 1 shows the detailed framework of the Transformer.
The encoder is composed of multiple identical layers, and its primary function is to capture the dependencies within the input sequence. Each encoding layer takes the output of the previous layer as its input, and adjacent layers are connected through residual connections. This structure can effectively transmit information and mitigate the vanishing gradient problem during the training process. A normalization operation is performed after each residual module. Each layer of the encoder contains two core sub-layers: multi-head attention and a fully-connected feed-forward network. Different from a single self-attention mechanism, multi-head attention compares each query with a set of key vectors from multiple representation sub-spaces. Similar to the encoder, the decoder is also composed of multiple identical layers, with each layer taking the output of the previous layer as its input. Each decoding layer consists of three core sub-layers: masked multi-head attention, a fully-connected feed-forward network, and multi-head attention. The outputs of these sub-layers are passed through residual links and combined with layer normalization to ensure the stability of the input for each layer. The computation processes of the fully connected feed-forward network and multi-head attention are the same as those in the encoder. Masked multi-head attention is the same as multi-head attention, except for the self-attention computation process.
Compared with Recurrent Neural Networks (RNNs) that process data sequentially, the Transformer can process all data simultaneously to improve efficiency and performance. However, when applied to complex tasks such as dam deformation prediction, the Transformer faces a series of challenges, which can be mainly summarized as high computational resource requirements, large memory usage, and limited processing speed.

2.3. Multi-Teacher Knowledge Distillation

Knowledge distillation (KD), as an important method for model compression, mainly aims to effectively transfer the “dark knowledge” from complex and large-scale models to concise and small-scale models, enabling the smaller models to achieve performance similar to that of large models. In 2014, reference [20] proposed a method that allows the student network to learn the output of the teacher network, defining the results of the teacher model as “soft labels”. Knowledge distillation works well in classification tasks because of its advantage in “dark knowledge”, which refers to the softened hidden layer representations (logits) output of the teacher model. The specific process of knowledge distillation is shown in Figure 2a,b.
The objective function of knowledge distillation for classification is as follows:
L KD = α · L CE ( O s T , O t T ) + ( 1 α ) L CE O s 1 , y l a b e l
where α is the balance coefficient of the loss function, which is used to adjust the balance between the soft loss and the hard loss. Its value ranges from [0, 1]; L CE is the cross-entropy loss function, T is the temperature coefficient, O s T is the probability distribution of the output of the student network after passing through the Softmax layer with the temperature coefficient T, O t T is the probability distribution of the output of the teacher network after passing through the Softmax layer with the temperature coefficient T; O s 1 is the probability distribution of the output of the student network after passing through the Softmax layer; y l a b e l is the label of the input data, which is defined as the hard label.
The objective function of knowledge distillation commonly used for regression is as follows:
L K D = α L M S E O s , O t + ( 1 α ) L M S E ( O s , y l a b e l )
where O s represents the direct output information of the student network; O t represents the soft-label information directly output by the teacher network; L M S E is the mean squared error loss function.
Multi-teacher distillation [33,34,35,36], as a variant of knowledge distillation, employs multiple teacher models to jointly guide the training of the student model. Compared with the traditional single-teacher knowledge distillation method, the multi-teacher distillation strategy provides more comprehensive and stable knowledge for the student model by aggregating the expertise from different teachers. Under the multi-teacher distillation framework, the learning process typically includes the following key steps:
( 1 )
Teacher model training: Multiple teacher models are trained independently. Each teacher model may be based on a different model architecture or a different training subset, ensuring that they possess unique knowledge and capabilities.
( 2 )
Teacher model integration: in multi-teacher distillation, weighted averaging of the prediction results is usually adopted to ensure more accurate knowledge transfer.
( 3 )
Student model training: The student model is trained to mimic the outputs of multiple teacher models as closely as possible. In classification tasks, the cross-entropy loss function is commonly used to measure the consistency between the outputs of the student model and the teacher models so as to achieve efficient knowledge transfer.
Compared with the traditional knowledge distillation that relies on a single teacher model, multi-teacher distillation can provide more diverse and robust knowledge, thereby significantly improving the performance and generalization ability of the student model [37,38,39]. When dealing with complex tasks and large-scale model training, multi-teacher distillation shows excellent results. It not only optimizes the efficiency of knowledge transfer but also enhances the adaptability and robustness of the model in the face of different data and tasks, demonstrating its wide application potential and important value in the field of deep learning.

2.4. The Basic Concept of Information Entropy

Here, we present some basic concepts of information entropy. Let v be a random variable with k possible values. D ( v ) = { v 1 ,   v 2 ,   ,   v k } is the set containing all possible values of the variable. If v is a discrete variable, its information entropy is defined as
H ( ν ) = i = 1 k P ( ν i ) log P ( ν i )
where P ( ν i ) is the probability of event v = v i occurring. If v is a continuous variable, then the information entropy is defined as
H ( ν ) = f ( ν ) log f ( ν ) d ν
where f ( v ) is the probability density function of v; the smaller the value of H ( ν ) , the more certain the value of this variable.
Conditional entropy quantifies the uncertainty of a random variable, assuming the value of another random variable is known. Let u be a random variable with k possible values and  D ( u ) = { u 1 ,   u 2 ,   ,   u k } be the set containing all possible values of this variable. At this time, the conditional entropy of v given u can be defined as
H ( ν | u ) = j = 1 k P ( u j ) H ( ν | u j )
If v is discrete variable, it is defined as
H ν | u j = i = 1 k P ν i | u j log P ν i | u j
If v is continuous, it is defined as
H v | u j = f v | u j log f v | u j d v
Therefore, information entropy can be used to evaluate the consistency of an object set over a given feature set. The more objects have the same or similar feature values, the more certain these feature values of the objects.

3. Regression Method Based on Multi-Teacher Distillation

3.1. Method Overview

Traditional single-teacher networks may encounter bottlenecks in generalization ability and expressiveness when guiding student networks [40]. However, when adopting the multi-teacher distillation strategy, we may face challenges such as knowledge conflicts and high resource consumption [41]. To overcome these limitations, we propose an innovative method: using the idea of clustering ensemble to accurately cluster and efficiently integrate multi-teacher networks, ensuring that teacher networks with similar structures can be clustered into the same category. This strategy helps reduce model redundancy and improve the efficiency of knowledge transfer to form a more powerful guiding model. To further optimize the application of the multi-teacher distillation method in regression problems, we introduce a new loss function, aiming to solve the knowledge transfer problems faced by multi-teacher network distillation in regression tasks and improve the accuracy and stability of the student model when dealing with complex regression problems.
In the entire knowledge distillation process, first, we pre-train N teacher networks. Then, we perform clustering ensemble on the pre-trained N teacher networks. The clustering method is explained in detail in Section 3.2. Through our method, the N teacher networks are clustered into k clusters. In each cluster, we select the top b teacher networks for weighted averaging, thereby integrating multiple teacher networks into a small number of high-quality teacher networks, which helps improve the subsequent distillation efficiency. Next, the predicted outputs of the aggregated k teacher networks are used to guide the learning process of the student network.
The detailed process is shown in Figure 3. The learning process of the student network consists of two parts. One part comes from the predicted outputs of the teacher networks, denoted as Limit, sand the other part comes from the real labels, Label, denoted as L M S E . We use the balance factor α to adjust the weights of the student network for different knowledge sources. In addition, we provide Table 1 to summarize the main symbols used.

3.2. Multi-Teacher Network Clustering Ensemble Based on Information Theory

Our goal is to integrate the wisdom of multiple teacher networks to form a small number of elite teacher networks, thereby enhancing the learning efficiency of the student network. To make the clustering results more reliable, we draw on the idea of clustering ensemble, aggregating multiple clustering results to generate more robust and stable clustering outcomes.
In our approach, we draw inspiration from the information–theoretic clustering ensemble methodology [42], as illustrated in Figure 4. We consider each teacher network as embedded within two feature spaces: the original feature space and the basic clustering feature space.
First, we generate a pool of teacher networks based on the original data. Then, through similarity calculation, we form the original data feature set T ( A ) . Subsequently, using multiple clustering methods on the original data feature set, we construct the base clustering feature set T ( Δ ) . After that, we obtain the final clustering result through hierarchical clustering. Our requirement for the final clustering result is to have high consistency with the base clustering feature set. We use information entropy to evaluate the uncertainty between them to obtain high-quality clustering results. Finally, we obtain the final clustering result T ( λ * ) for the subsequent weighted integration of teacher networks. The detailed definition and process of this method are presented in Figure 4.
T ( A ) is an N × N matrix, where N is the number of teacher networks. We define it as the original data feature set. First, we present the construction processes of the original feature set and the basic clustering-feature set. For teacher networks t i and t j , we flatten the weights of the linear layers in the last layer of the teacher networks into one-dimensional vectors v i and v j . We use cosine similarity to evaluate the similarity between the two teacher networks:
cos i j = v i · v j v i   v j
In addition, we define D i s i j as representing the distance between the two teacher networks T i and T j :
D i s i j = 1 cos i j
We construct the original data feature set of teacher networks based on the distances between different teacher networks. The purpose of this is to fully consider the structural similarity among the different teacher networks when generating the basic clustering feature set. Subsequently, we select M clustering algorithms and generate M clustering results according to the original data feature set T ( A ) . These results form the base clustering feature set T ( Δ ) . T ( Δ ) is an N × M matrix, where M represents the number of clustering methods. It denotes the base clustering feature set composed of all initial clustering results.
Next, we introduce how to define information entropy to measure the consistency between the clustering results and the base clustering feature set Δ . We define each λ h as a random variable, which represents the feature value formed by the h-th base clustering result. Let D ( λ h ) = { λ h l } l = 1 k h , where λ h l is the label of the clustering result of C h l , and  P ( λ h l ) is the probability that λ h l , where λ h is the feature value of the h-th base clustering. Suppose S ⊆ T is a cluster that contains multiple objects and S is a subset of the set T. We calculate P( λ h l |S) by means of the relative frequency of the label value λ h l occurring in the cluster S:
P ( λ h l S ) = | { x i x i ( λ h ) = λ h l , x i S } | | S |
The information entropy of λ h given S is
H ( λ h S ) = i = 1 k j P ( λ h l S ) log P ( λ h l S )
In addition, we assume that each λ h is independent of the others. Then, Δ can be regarded as a set composed of M random variables. Given cluster S, the information entropy of Δ can be expressed as:
H ( Δ S ) = h = 1 M H ( λ h S )
Suppose S = { S l l = 1 p } is the set of S, including p clusters, where S l is the l-th cluster, 1 l p . Therefore, according to the definition of conditional entropy, given S, the information entropy of Δ is
H ( Δ S ) = l = 1 p P ( S l S ) H ( Δ S l )
P ( S l | S ) is the probability of S l in S , and the calculation process is as follows:
P ( S l S ) = S l l = 1 p S l
H ( Δ | S ) can be regarded as the sum of the consistencies of all clusters S , which we define as the consensus sum. If the consistency of each cluster S l with respect to Δ is very high, then the value of H ( Δ | S ) is lower. Based on Formulas (17) and (18), we can define the consensus sum of S on Δ as follows:
I ( Δ | S ) = H ( Δ | S ) H ( Δ | S )
We believe that if set S of clusters is a good set for cluster S, it can enhance the consensus of cluster S on Δ . The larger the I ( Δ | S ) , the higher the clustering consistency on Δ .
Given T, H ( Δ | T ) is determined. Then, the more objects in each cluster of the set S share the same labels in the basic clustering feature set Δ , the lower the value of H ( Δ | S ). Therefore, clustering results with high consistency on Δ are considered to have good performance. In addition, if a set of clusters S = { S i , S j } contains two clusters S i and S j , then, according to Formula (20), we have
I Δ | { S i , S j } = H Δ | S i S j H Δ | { S i , S j }
Given S i and S j , H ( Δ | S i S j ) is determined. If most of the objects in S i S j have the same labels on Δ , then I ( Δ | { S i , S j } ) decreases significantly. At this time, we consider that these two clusters are similar on Δ .
The clustering-ensemble task can be regarded as a clustering operation performed on the basic clustering feature set. We use information entropy as the objective function to screen out the optimal final clustering results. Assuming S = C is the set of clusters formed by T, where C is defined as a partition, we ultimately use H ( Δ | C * ) as the objective function of the clustering ensemble method and design the corresponding optimization model to perform the clustering ensemble task on the basic clustering feature set.
The formal description of the clustering ensemble optimization model is
min C * H ( Δ | C * )
According to the above-mentioned content, we can conclude that, given T,
max C * I ( Δ C * ) = H ( Δ T ) min C * H ( Δ C * )
According to Formula (23), we know that minimizing the objective function H ( Δ | C * ) is equivalent to maximizing the consensus sum between the final clustering results and the basic clustering feature set.
We adopt a hierarchical clustering strategy to generate the final clustering results. Define C ( z ) as the partition generated by merging cluster C i ( z 1 ) and cluster C j ( z 1 ) at the z-th time. First, we construct an initial partition C 0 , where each teacher network in T = { t 1 ,   t 2 ,   ,   t N } serves as a separate cluster; that is, C ( 0 ) = { C 1 ,   C 2 ,   ,   C N } . According to Formula (18), at this time, H ( Δ | C 0 ) = 0 . Next, we continuously combine two partitions to form a new partition C ( z + 1 ) . Based on the properties of conditional entropy, we have
H ( Δ | C ( z + l ) ) H ( Δ | C ( z ) )
As the number of clusters decreases, the consensus sum of the clustering results also gradually decreases. Therefore, each time we merge partitions, we hope that the consensus sum decreases as little as possible. The optimization problem can be transformed into
min C ( z + 1 ) H ( Δ C ( z ) ) H ( Δ C ( z + 1 ) ) = min C i ( z ) , C j ( z ) C ( z ) I ( Δ { C i ( z ) , C j ( z ) } )
Therefore, each time we need to merge the two clusters C i ( z ) and C j ( z ) with the highest consistency in C ( z ) , the merging condition is
min C i ( z ) , C j ( z ) C ( z ) I ( Δ | { C i ( z ) , C j ( z ) } )
When the number of clusters in C ( z ) is k, the merging process terminates, and the final clustering result is C * = C ( z ) .
The aforementioned steps effectively group multiple teacher networks, which are followed by our screening process to enhance the overall model’s performance and efficiency. We obtain the final clustering result T ( λ * ) from C * , which ensures that the clustering result maximally reflects the similarities and differences among the teacher networks. Here, λ * denotes the final clustering result, where 1 λ * k . Our objective is to further refine the team of teacher networks. To achieve this, we rank the teacher networks within each cluster based on their mean squared error (MSE), allowing us to clearly understand the performance of each teacher network within its cluster. Subsequently, we select the top b teacher networks from each cluster (where b is a predefined parameter, typically adjusted according to specific requirements and model characteristics). This selection aims to retain the best-performing teacher networks in each cluster while eliminating those with lower prediction accuracy, thereby ensuring the quality of the refined teacher network team.
After completing the screening of teacher networks, we determine the weighted average of the predictions from the selected teacher networks. This weighted average takes into account the performance differences among the teacher networks, assigning higher weights to those with superior performance. Based on the final clustering result, we rank the teacher networks in each cluster according to the MSE evaluation metric, aiming to further streamline the teacher network team. We sequentially select the top b teacher networks from each cluster and compute a weighted average of their predictions, ultimately generating k strong teacher networks. These strong teacher networks not only retain the advantages of multiple teacher networks but also eliminate redundant and low-performance networks. Consequently, this step provides the student network with more accurate and effective knowledge guidance, enhancing the learning efficiency and predictive performance of the student network. Our proposed method is formally described in Algorithm 1.
Algorithm 1 Multi-teacher clustering ensemble based on information theory.
  • Input: T ( A ) , T ( Δ ) , k
  • Output: D ( O t )
  • Set z = 0 and produce an initial partition C ( z ) = { C i ( z ) } i = 1 N where C i ( z ) = { t i } , 1 i N ;
  • while C ( z ) |   > k do
  •    Select set two clusters C i ( z ) and C j ( z ) which satisfy min   C i ( z ) , C j ( z ) C ( z ) I w ( Δ | { C i ( z ) , C j ( z ) } ) ;
  •    Execute clustering merging C i ( z ) = C i ( z ) C j ( z ) ;
  •    Update partition results C ( z + 1 ) = C ( z ) C j ( z ) ;
  •    The number of merges increases automatically z = z + 1;
  • end while
  • Get T ( λ * ) corresponding to C ( z ) ;
  • Cluster the teacher networks into k clusters according to T ( λ * ) ;
  • while i <   k do
  •    Select the top b teacher networks in the i-th cluster;
  •    Calculate the weighted average result O t i ;
  • end while
  • Return D ( O t ) = { O t 1 ,   O t 2 ,   ,   O t k } ;

3.3. Knowledge Distillation

Under the framework of muti-teacher distillation, the design of the loss function is of vital importance. It not only determines the learning efficiency of the student network of the knowledge of the teacher networks but also directly affects the performance of the final network. Next, the carefully designed multi-teacher distillation loss function for the regression task is elaborated in detail.
Inspired by [43], we chose to use Limit to replace the mean squared error (MSE) used in traditional regression tasks. When the performance of the student network is better than that of the teacher network, we did not add the loss of the teacher network at this time as an additional loss of the student network. In this way, we could effectively guide the model learning.
At the same time, the loss function we propose fully considers the performance differences among multiple teacher models. We set parameter λ to balance their impacts on the student model. This weighting strategy ensures that teacher models with excellent performance can play a greater role in the training process, thereby accelerating the learning efficiency of the student model and improving its generalization ability. The objective function we propose is as follows:
L K D = 1 n i n j m α O s y l a b e l i 2 + ( 1 α ) λ j L m t ( O s , O t j ) i
L i m i t O s , O t = O s O t 2 , if O s y l a b e l > O t y l a b e l 0 , otherwise
λ j = 1 ( m s e j ) s u m ( M )
M = { 1 n i n ( O i t y l a b e l ) j 2 : j = 1 ,   ,   k }
where λ j is the set weight coefficient, α is the balance coefficient, n is the number of samples, k is the number of clustered teacher networks, and M = { m s e j } j = 1 k is the set of mean squared errors of the clustered teacher networks.
The following presents its mathematical derivative-derivation process:
= L K D W = 1 n i N α ( W x + b y l a b e l ) 2 + j k ( 1 α ) λ j ( w x + b O t j ) 2 W = 2 n α x ( W x + b y l a b e l ) + 2 ( 1 α ) x j k λ j ( w x + b O t j )
W = W 0 η · L K D ( W 0 )
where W 0 is the parameter of the independent variable, η is the learning factor, and W is the updated W 0 .

4. Results

4.1. Project Overview

This study focused on a concrete-faced rockfill dam located in Guizhou province, China. The dam had a crest elevation of 154 m, a crest width of 10 m, a crest length of 428.93 m, a maximum dam width of 454.085 m, an installed capacity of 90,000 kilowatts, and a total reservoir capacity of 1.325 billion cubic meters. To monitor the dam’s operational status, various sensors were installed internally to collect data and assess its performance. Horizontal displacement and settlement monitoring sensors were placed at the cross-section 0+005.000, with four monitoring lines. For a visual representation, refer to Figure 5. Monitoring data from four monitoring points (SG19, SG20, SG25, and SG26) from September 2021 to May 2024 were selected for modeling, with a data interval of 5 days. The dataset was divided into 80% for training and 20% for testing. During model training, the training set was further split into 0.8 for training and 0.2 for validation. The deformation data of the monitoring points are shown in Figure 6.
The dam deformation prediction model developed in this study used influencing factors such as water level, temperature, and time effects as inputs, with dam deformation as the output. A total of 10 influencing factors were selected, including water pressure factors H 1 , H 2 , H 3 , and H 4 ; temperature factors s i n ( 2 π t / 365 ) , c o s ( 2 π t / 365 ) , s i n ( 2 · 2 π t / 365 ) , and c o s ( 2 · 2 π t / 365 ) ; as well as time-effect factors θ and l n ( θ ) .

4.2. Evaluation Metrics

To comprehensively evaluate the prediction accuracy of the model, various indicators are usually used. We adopted RMSE (root mean squared error), MAPE (mean absolute percentage error), MAE (mean absolute Eerror), and the coefficient of determination ( R 2 ) to evaluate the model. The specific formulas are given in Equations (31)–(34):
R M S E = 1 N i = 1 N ( y i y ^ i ) 2
M A P E = 1 n i = 1 N y i y ^ i y i × 100 %
M A E = 1 N i = 1 N y i y ^ i
R 2 = 1 i = 1 N ( y i y ^ i ) 2 i = 1 N ( y i y ¯ i ) 2
where N is the total number of samples, y i is the predicted value of the i-th sample, y ^ i is the observed value of the i-th sample, and y ¯ i is the average value of the observed values of the i-th sample.
RMSE evaluates the prediction accuracy, MAPE evaluates the model’s accuracy, and MAE reflects the actual situation of the prediction error. The smaller the values of these three parameters, the higher the model’s accuracy. R 2 represents the fitting effect of the model, wehre the higher its value, the better the prediction performance.

4.3. Experimental Settings

We adopted the Transformer model as the teacher network, which included a position-encoding module, encoder layers, and a decoder composed of a simple linear layer. It was specifically designed for processing and predicting complex sequential data. We used a Temporal Convolutional Network (TCN) as the student network model. We selected the Temporal Convolutional Network (TCN) as the student network because it adopts a convolutional structure. Compared with traditional recurrent neural networks, convolutional operations can process data in parallel, thereby improving computational efficiency. Moreover, convolutional kernels can automatically extract local features from the data. Most importantly, the TCN has a relatively simple structure with a small number of parameters and thus has a relatively low demand for computing resources. Due to space limitations, we omit the description of the TCN and recommend readers refer to [44] for detailed information.In addition, we provide the algorithmic complexities of the Transformer and TCN in Table 2. On the one hand, it served as a reference for selecting the teacher network and the student network. On the other hand, it was used to verify the degree of reduction in model complexity after distillation. We set the initial number N of teacher networks to 50, the number k of clustering categories of teacher networks to 3, and the basic clustering set M composed of basic clustering results to 15. We used three methods, namely, the classic k-means, spectral clustering, and Gaussian mixture clustering, to cluster the 50 teacher networks according to the original feature set, with the number of clusters being 5. Each method selected 5 different random seeds to obtain 5 different clustering results, and b was set to 6 during the clustering selection process. Both the teacher network and the student network used the Adam optimizer. We initialized the learning rate at 0.0001, set the batch size to 8, and set α to 0.8 during the knowledge distillation process.

4.4. Parameter Sensitivity Experiment

In the experiments presented in this paper, parameter α was incorporated into the objective function. Its purpose is to regulate the contribution of the knowledge extracted from the teacher network. The value of α has a significant and intuitive impact on the knowledge distillation process.Theoretically, a larger value of α means that the teacher network has more influence on the distillation process. Conversely, when α is smaller, the student network places more emphasis on its own learning ability during the distillation. In the experiments in this study, to comprehensively and thoroughly investigate the impact of parameter α on the model’s performance, we selected a set of representative values from the set {0.05, 0.1, 0.2, 0.4, 0.6, 0.8, 0.9, 0.99}. To more intuitively demonstrate the relationship between the model’s performance and parameter α , we plotted a trend graph showing the model’s R 2 (coefficient of determination; the closer the R 2 value to ond, the better the model fits the data) against α , as shown in Figure 7.
We found that when α was set to 0.8, the model achieved the optimal accuracy. This result indicates that when α = 0.8 , an ideal balance is reached between the teacher network and the student network. On the one hand, it can fully utilize the rich knowledge provided by the teacher network; on the other hand, it gives the student network enough room to exert its own learning ability, enabling the model to achieve the best performance during the knowledge distillation process. Based on this important experimental result, we made a reasonable and scientific decision. In the subsequent experiments, we fixed the value of α at 0.8.

4.5. Distillation Performance Evaluation

Table 3 shows the prediction results of the teacher networks after clustering, the student network, and CLMTKD on the test set. TeacherCL1,TeacherCL2 and TeacherCL3 are advanced teacher networks generated by weighted averaging the top six teacher networks after clustering. From SG19, SG20, SG25, and SG26, it can be seen that the proposed CLMTKD provides a significant improvement compared with the student network and is close to or even better than the clustering results of the advanced teacher networks at some measurement points.
Figure 8 presents a comparison between the student network and the student network after knowledge distillation. It is evident from the figure that the fitting performance of the student network significantly improved after knowledge distillation. Specifically, the distilled student network demonstrates superior fitting capability in the regression fitting curves, better capturing the distribution characteristics of the data. This indicates that the knowledge distillation process effectively enhances the performance of the student network, enabling it to learn and predict more accurately while maintaining lower complexity. This result further validates the effectiveness of the proposed CLMTKD algorithm in improving the learning efficiency and prediction accuracy of the student network.

4.6. Comparative Experiment

To compare the effectiveness of our method with traditional knowledge distillation methods for regression, we selected a total of 5 comparison methods, including single-teacher network distillation and multi-teacher network distillation. Methods such as AIL [45], TBR [46], and TOR [47] use a single teacher network for knowledge distillation. Since the research on multi-teacher distillation models for regression is still in its infancy, we conducted two comparative experiments. The first method was forming a strong teacher network by weighted averaging all teacher networks and then selecting the regression loss of the AIL method to guide the learning of the student network (MTKD). The second method involved using all teacher networks simultaneously and selecting the regression loss in Formula (28) to guide the learning of the student network (ALLMTKD).
According to the comparison of the experimental results in Table 4, it can be found that the proposed CLMTKD achieves ideal results on the four datasets. Figure 9 is the radar chart of all algorithms, where the larger the area, the better the performance. It is not difficult to find that our method outperforms the other methods on all indicators.
Further analyzing the experimental results, compared with traditional single-teacher distillation networks (AIL, TBR, TOR), we find that relying solely on a single teacher network to transmit knowledge to students has a single source and limited features that it can cover. In contrast, our method aggregates the advantages of multiple teacher networks through clustering and integrated selection, ensuring performance while maintaining model diversity. This enables the student network to learn more comprehensive and rich knowledge, effectively demonstrating the necessity of introducing multi-teacher networks in knowledge distillation. At the same time, our method dynamically sets weights according to the performance of each teacher network, meaning that more capable teacher networks have more say in the knowledge transfer process, allowing the student network to learn more from the “high-quality information” of its outputs. However, single-teacher networks do not have such a mechanism for dynamically adjusting weights according to performance.
Compared with traditional multi-teacher networks, our method also achieves performance improvements. MTKD uses multiple teacher networks to form a single teacher network through weighted averaging. However, it is difficult to effectively distinguish and utilize the truly valuable parts in multi-teacher networks, and, instead, low-quality or repetitive information may interfere with the learning of the student network. Compared with ALLMTKD, which directly allows all teacher networks to guide students, due to the large number of teacher networks, the guidance weights are overly dispersed, meaning that the influence of each teacher network on the student network is very weak, and it is difficult to effectively transmit their own knowledge to the student network. The student network is likely to be lost in the large amount of weak and messy guiding knowledge, resulting in poor learning effects, effectively demonstrating the importance of directly introducing clustering and integrated selection.
The results of the time and performance comparison of different models are shown in Figure 10. We can see that the training time of the single-teacher distillation method is shorter than that of multi-teacher networks. The main reason is that the multi-teacher distillation process needs to integrate knowledge from multiple teachers. Therefore, it usually requires a longer training time than the single-teacher distillation method. ALLMKD has the longest training time because multiple teacher networks participate in the guidance simultaneously, resulting in chaos in the knowledge transfer process, greatly increasing the training cost, and achieving the worst performance. This also demonstrates that the number of teacher networks in the distillation process is not the more, the better. The training time of the proposed CLMTKD is longer than that of MTKD. The reason is that we introduce a clustering and integration method, which consumes a part of the training time, but we achieve the best distillation results.

5. Conclusions

This paper proposed a multi-teacher distillation method based on clustering integration for dam displacement prediction. We discussed the advantages of multi-teacher networks over single-teacher networks in the knowledge distillation process. On this basis, a clustering selection and integration algorithm for multi-teacher networks was proposed to address issues such as the confusion and inefficiency caused by multi-teacher networks during the distillation process. Knowledge is transferred from the teacher network to the student network by minimizing the loss function. The experiments on all datasets showed that our model could effectively compress the model while improving the accuracy of the student model. Further analysis showed that we could significantly improve students’ performance, and the prediction results were more consistent with the true values. In conclusion, the method we proposed can effectively improve the prediction accuracy, capture the subtle changes in dam deformation in advance, provide strong data support for dam safety management, reduce the risk of safety accidents, and ensure the safety of the lives and property of people downstream as well as the normal operation of water conservancy facilities. At present, this method was only applied to concrete face rockfill dams. In the future, we will consider exploring this framework for use with more general infrastructure such as concrete gravity dams and arch dams.

Author Contributions

Conceptualization, F.G. and J.Y.; methodology, J.Y.; software, F.G.; validation, F.G., J.Y. and D.L.; formal analysis, X.Q.; investigation, F.G.; resources, X.Q.; data curation, J.Y.; writing—original draft preparation, F.G.; writing—review and editing, J.Y.; visualization, J.Y.; supervision, J.Y.; project administration, D.L.; funding acquisition, X.Q. All authors have read and agreed to the published version of this manuscript.

Funding

This research was supported by the Guizhou Science and Technology Projects (QiankeheSupport [2023] General 251), the Guizhou Provincial Major Scientific and Technological Program (grant number QiankeheMajor[2024]008), and the Guizhou Provincial Basic Research Program (Natural Science) Youth Guidance Project (Qiankehe Basic-[2024] Youth 095).

Data Availability Statement

Data are contained within this article.

Conflicts of Interest

Author Fawang Guo was employed by the company POWERCHINA Guiyang Engineering Corporation Limited, Guiyang, China. The remaining authors declare that this research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflicts of interest.

References

  1. Aydemir, A. Modified risk assessment tool for embankment dams: Case study of three dams in Turkey. Civ. Eng. Environ. Syst. 2017, 34, 53–67. [Google Scholar] [CrossRef]
  2. Sivasuriyan, A. Health assessment of dams under various environmental conditions using structural health monitoring techniques: A state-of-art review. In Environmental Science and Pollution Research; Springer Nature: Berlin/Heidelberg, Germany, 2022; pp. 1–12. [Google Scholar]
  3. Yang, J. An intelligent prediction model of settlement deformation for the earth-rock dam. In Proceedings of the 2021 International Conference on Artificial Intelligence, Big Data and Algorithms (CAIBDA), Xi’an, China, 28–30 May 2021; pp. 198–201. [Google Scholar]
  4. Wang, Q. Research on dam deformation prediction based on variable weight combination model. In Proceedings of the 2024 4th International Conference on Neural Networks, Information and Communication (NNICE), Guangzhou, China, 19–21 January 2024; pp. 900–903. [Google Scholar]
  5. Luo, C. Research on Dam Deformation Prediction Model Based on Wavelet Neural Network. In Proceedings of the 2021 3rd International Conference on Applied Machine Learning (ICAML), Changsha, China, 23–25 July 2021; pp. 362–365. [Google Scholar]
  6. Zhu, Y. Automatic damage detection and diagnosis for hydraulic structures using drones and artificial intelligence techniques. Remote Sens. 2023, 15, 615. [Google Scholar] [CrossRef]
  7. Zhu, M. Optimized multi-output LSSVR displacement monitoring model for super high arch dams based on dimensionality reduction of measured dam temperature field. Eng. Struct. 2022, 268, 114686. [Google Scholar] [CrossRef]
  8. Mata, J. Constructing statistical models for arch dam deformation. Struct. Control Health Monit. 2014, 21, 423–437. [Google Scholar] [CrossRef]
  9. Chen, R. Construction and selection of deformation monitoring model for high arch dam using separate modeling technique and composite decision criterion. Struct. Health Monit. 2024, 23, 2509–2530. [Google Scholar] [CrossRef]
  10. Wu, B. A Multiple-Point Deformation Monitoring Model for Ultrahigh Arch Dams Using Temperature Lag and Optimized Gaussian Process Regression. Struct. Control Health Monit. 2024, 2024, 2308876. [Google Scholar] [CrossRef]
  11. Xing, Y. Research on dam deformation prediction model based on optimized SVM. Processes 2022, 10, 1842. [Google Scholar] [CrossRef]
  12. Li, X. An approach using random forest intelligent algorithm to construct a monitoring model for dam safety. Eng. Comput. 2021, 37, 39–56. [Google Scholar] [CrossRef]
  13. Hipni, A. Daily forecasting of dam water levels: Comparing a support vector machine (SVM) model with adaptive neuro fuzzy inference system (ANFIS). Water Resour. Manag. 2013, 27, 3803–3823. [Google Scholar] [CrossRef]
  14. Yang, D. A concrete dam deformation prediction method based on LSTM with attention mechanism. IEEE Access 2020, 8, 185177–185186. [Google Scholar] [CrossRef]
  15. Yang, J. A CNN-LSTM model for tailings dam risk prediction. IEEE Access 2020, 8, 206491–206502. [Google Scholar]
  16. Jiedeerbieke, M. Gravity Dam Deformation Prediction Model Based on I-KShape and ZOA-BiLSTM. IEEE Access 2024, 12, 50710–50722. [Google Scholar] [CrossRef]
  17. Su, Y. Dam deformation interpretation and prediction based on a long short-term memory model coupled with an attention mechanism. Appl. Sci. 2021, 11, 6625. [Google Scholar] [CrossRef]
  18. Zhou, Y. Multi-expert attention network for long-term dam displacement prediction. Adv. Eng. Informatics 2023, 57, 102060. [Google Scholar]
  19. Wang, X. Study on MPGA-BP of gravity dam deformation prediction. Math. Probl. Eng. 2017, 2017, 2586107. [Google Scholar]
  20. Hinton, G. Distilling the knowledge in a neural network. arXiv 2015, arXiv:1503.02531. [Google Scholar]
  21. Sun, T. Explainability-based knowledge distillation. Pattern Recognit. 2025, 159, 111095. [Google Scholar]
  22. Chu, Y. Bi-directional contrastive distillation for multi-behavior recommendation. In Joint European Conference on Machine Learning and Knowledge Discovery in Databases; Springer International Publishing: Cham, Switzerland, 2022; pp. 491–507. [Google Scholar]
  23. Lopez-Paz, D. Unifying distillation and privileged information. arXiv 2015, arXiv:1511.03643. [Google Scholar]
  24. Yuan, Y. The Impact of Knowledge Distillation on the Energy Consumption and Runtime Efficiency of NLP Models. In Proceedings of the IEEE/ACM 3rd International Conference on AI Engineering-Software Engineering for AI, Lisbon, Portugal, 14–15 April 2024; pp. 129–133. [Google Scholar]
  25. Léger, P. Hydrostatic, temperature, time-displacement model for concrete dams. J. Eng. Mech. 2007, 133, 267–277. [Google Scholar]
  26. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30, 5998–6008. [Google Scholar]
  27. Xu, C. A financial time-series prediction model based on multiplex attention and linear transformer structure. Appl. Sci. 2023, 13, 5175. [Google Scholar] [CrossRef]
  28. Ahmed, S. Transformers in time-series analysis: A tutorial. Circuits Syst. Signal Process. 2023, 42, 7433–7466. [Google Scholar] [CrossRef]
  29. Sasal, L. W-transformers: A wavelet-based transformer framework for univariate time series forecasting. In Proceedings of the 2022 21st IEEE International Conference on Machine Learning and Applications (ICMLA), Nassau, Bahamas, 12–14 December 2022; pp. 671–676. [Google Scholar]
  30. Wen, Q. Transformers in time series: A survey. arXiv 2022, arXiv:2202.07125. [Google Scholar]
  31. Hittawe, M.M. Stacked Transformer Models for Enhanced Wind Speed Prediction in the Red Sea. In Proceedings of the 2024 IEEE 22nd International Conference on Industrial Informatics (INDIN), Beijing, China, 18–20 August 2024; pp. 1–7. [Google Scholar]
  32. Jiang, W. Applicability analysis of transformer to wind speed forecasting by a novel deep learning framework with multiple atmospheric variables. Appl. Energy 2024, 353, 122155. [Google Scholar] [CrossRef]
  33. Lin, Y. ATMKD: Adaptive temperature guided multi-teacher knowledge distillation. Multimed. Syst. 2024, 30, 292. [Google Scholar] [CrossRef]
  34. Zhang, H. Adaptive multi-teacher knowledge distillation with meta-learning. In Proceedings of the 2023 IEEE International Conference on Multimedia and Expo (ICME), Brisbane, Australia, 10–14 July 2023; pp. 1943–1948. [Google Scholar]
  35. Sarıyıldız, M.B. UNIC: Universal Classification Models via Multi-teacher Distillation. In European Conference on Computer Vision; Springer Nature: Cham, Switzerland, 2024; pp. 353–371. [Google Scholar]
  36. Li, R. Cross-View Gait Recognition Method Based on Multi-Teacher Joint Knowledge Distillation. Sensors 2023, 23, 9289. [Google Scholar] [CrossRef]
  37. Yang, Y. MKDAND: A network flow anomaly detection method based on multi-teacher knowledge distillation. In Proceedings of the 2022 16th IEEE International Conference on Signal Processing (ICSP), Beijing, China, 21–24 October 2022; Volume 1, pp. 314–319. [Google Scholar]
  38. Pham, C. Collaborative multi-teacher knowledge distillation for learning low bit-width deep neural networks. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–7 January 2023; pp. 6435–6443. [Google Scholar]
  39. Ye, X. Knowledge distillation via multi-teacher feature ensemble. IEEE Signal Process. Lett. 2024, 31, 566–570. [Google Scholar] [CrossRef]
  40. Wu, C. One teacher is enough? Pre-trained language model distillation from multiple teachers. arXiv 2021, arXiv:2106.01023. [Google Scholar]
  41. Zhang, J. Badcleaner: Defending backdoor attacks in federated learning via attention-based multi-teacher distillation. IEEE Trans. Dependable Secur. Comput. 2024, 21, 4559–4573. [Google Scholar] [CrossRef]
  42. Bai, L. An information-theoretical framework for cluster ensemble. IEEE Trans. Knowl. Data Eng. 2018, 31, 1464–1477. [Google Scholar] [CrossRef]
  43. Lee, K. Fast and accurate facial expression image classification and regression method based on knowledge distillation. Appl. Sci. 2023, 13, 6409. [Google Scholar] [CrossRef]
  44. Bai, S. An empirical evaluation of generic convolutional and recurrent networks for sequence modeling. arXiv 2018, arXiv:1803.01271. [Google Scholar]
  45. Saputra, M.R.U. Distilling knowledge from a deep pose regressor network. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 263–272. [Google Scholar]
  46. Chen, G.; Choi, W.; Yu, X.; Han, T.; Chandraker, M. Learning efficient object detection models with knowledge distillation. Adv. Neural Inf. Process. Syst. 2017, 30, 742–751. [Google Scholar]
  47. Takamoto, M. An efficient method of training small models for regression problems with knowledge distillation. In Proceedings of the 2020 IEEE Conference on Multimedia Information Processing and Retrieval (MIPR), Shenzhen, China, 6–8 August 2020; pp. 67–72. [Google Scholar]
Figure 1. Architecture diagram of the Transformer [32].
Figure 1. Architecture diagram of the Transformer [32].
Water 17 00988 g001
Figure 2. Knowledge distillation processes for classification tasks and regression tasks. (a) Knowledge distillation for classification. (b) Knowledge distillation for regression.
Figure 2. Knowledge distillation processes for classification tasks and regression tasks. (a) Knowledge distillation for classification. (b) Knowledge distillation for regression.
Water 17 00988 g002
Figure 3. Multi-teacher knowledge distillation architecture for regression tasks.
Figure 3. Multi-teacher knowledge distillation architecture for regression tasks.
Water 17 00988 g003
Figure 4. The basic process of the clustering ensemble method based on information theory.
Figure 4. The basic process of the clustering ensemble method based on information theory.
Water 17 00988 g004
Figure 5. Monitoring profile diagram of the dam body on the left side of the cross-section 0+005.000.
Figure 5. Monitoring profile diagram of the dam body on the left side of the cross-section 0+005.000.
Water 17 00988 g005
Figure 6. Deformation monitoring data of the monitoring points.
Figure 6. Deformation monitoring data of the monitoring points.
Water 17 00988 g006
Figure 7. The influence of α on the model’s R 2 : (a) monitoring point SG19; (b) monitoring point SG20; (c) monitoring point SG25; (d) monitoring point SG26.
Figure 7. The influence of α on the model’s R 2 : (a) monitoring point SG19; (b) monitoring point SG20; (c) monitoring point SG25; (d) monitoring point SG26.
Water 17 00988 g007
Figure 8. Prediction fitting images of the student network and the student network after distillation: (a) monitoring point SG19; (b) monitoring point SG20; (c) monitoring point SG25; (d) monitoring point SG26.
Figure 8. Prediction fitting images of the student network and the student network after distillation: (a) monitoring point SG19; (b) monitoring point SG20; (c) monitoring point SG25; (d) monitoring point SG26.
Water 17 00988 g008
Figure 9. Radar chart showing comparison with other baseline models: (a) monitoring point SG19; (b) monitoring point SG20; (c) monitoring point SG25; (d) monitoring point SG26.
Figure 9. Radar chart showing comparison with other baseline models: (a) monitoring point SG19; (b) monitoring point SG20; (c) monitoring point SG25; (d) monitoring point SG26.
Water 17 00988 g009
Figure 10. Deformation monitoring data of the monitoring points.
Figure 10. Deformation monitoring data of the monitoring points.
Water 17 00988 g010
Table 1. Explanations of some symbols used in this paper.
Table 1. Explanations of some symbols used in this paper.
SymbolDescription
λ h The feature value of the h-th base clustering
D(v)The domain of variable (feature) v
λ h l The clustering label of C h l
C h The h-th base clustering set
C h l The l-th cluster in C h
S ⊆ TA cluster composed of multiple objects in T
S = { S l } l = 1 p A set composed of clusters S where S l is the l-th cluster
C S = C The set composed of the objects in T, which is defined as a partition
C * The partition result formed by the final clusters
kThe number of clusters in partition C *
λ * The final clustering feature vector
Table 2. Complexity of teacher networks and student network.
Table 2. Complexity of teacher networks and student network.
Network#ParamsFLOPs
Transformer87,0078735
TCN20.735 M0.9928 M
Table 3. Performance comparison among teacher networks after clustering, student network, and student network after distillation.
Table 3. Performance comparison among teacher networks after clustering, student network, and student network after distillation.
DatasetModelMSE (mm)RMSE (mm)MAE (mm)MAPE R 2
TeacherCL10.060490.245950.184832.362030.97679
TeacherCL20.099140.314870.252753.352160.96196
SG19TeacherCL30.108560.329490.259643.572590.95834
Student0.139550.373570.293993.888490.94644
CLMTKD0.075780.275270.213252.772320.97092
TeacherCL10.002670.05130.040720.594780.96749
TeacherCL20.003170.056340.044320.641940.96088
SG20TeacherCL30.005070.071170.053860.784960.93758
Student0.005200.072100.059230.857570.93594
CLMTKD0.003080.055540.042650.622910.96199
TeacherCL10.014460.120270.090451.141680.97692
TeacherCL20.035640.188790.152371.938990.94313
SG25TeacherCL30.037830.194490.158851.986120.93964
Student0.044630.211270.173932.180140.92878
CLMTKD0.011710.108190.078370.987880.98132
TeacherCL10.010180.100900.081281.010310.96411
TeacherCL20.014460.120260.097291.218370.94901
SG26TeacherCL30.014820.121740.095541.176030.94775
Student0.021760.147530.120291.504590.92327
CLMTKD0.007660.087540.067810.844330.97298
Table 4. Comparison with other benchmark models.
Table 4. Comparison with other benchmark models.
DatasetModelMSE (mm)RMSE (mm)MAE (mm)MAPE R 2
SG19AIL0.117980.343480.265973.250200.954724
TBR0.090400.300660.236793.072380.965308
TOR0.111350.336380.277383.537470.95658
MTKD0.087140.295200.225492.942860.96656
ALLMTKD0.295880.543940.473936.325270.88645
CLMTKD0.075780.275270.213252.772320.97092
SG20AIL0.005340.073070.057800.839950.93421
TBR0.004230.065040.054660.794100.94788
TOR0.004320.058480.045330.661930.95786
MTKD0.003980.063050.049520.718760.95102
ALLMTKD0.012450.111570.100121.443190.84661
CLMTKD0.003080.055540.042650.622910.96199
SG25AIL0.040950.202370.161391.969940.93465
TBR0.026050.161390.129531.610690.95844
TOR0.020770.144130.113861.414030.96685
MTKD0.013360.115610.085601.073100.97867
ALLMTKD0.101770.319000.278083.440800.83762
CLMTKD0.011710.108190.078370.987880.98132
SG26AIL0.020300.142460.119361.470080.92845
TBR0.010380.101860.078540.971470.96342
TOR0.012120.110100.083041.024460.95727
MTKD0.009320.096480.077260.966720.96718
ALLMTKD0.052610.229370.192742.327230.81451
CLMTKD0.007660.087540.067810.844330.97298
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Guo, F.; Yuan, J.; Li, D.; Qin, X. Application of a Multi-Teacher Distillation Regression Model Based on Clustering Integration and Adaptive Weighting in Dam Deformation Prediction. Water 2025, 17, 988. https://doi.org/10.3390/w17070988

AMA Style

Guo F, Yuan J, Li D, Qin X. Application of a Multi-Teacher Distillation Regression Model Based on Clustering Integration and Adaptive Weighting in Dam Deformation Prediction. Water. 2025; 17(7):988. https://doi.org/10.3390/w17070988

Chicago/Turabian Style

Guo, Fawang, Jiafan Yuan, Danyang Li, and Xue Qin. 2025. "Application of a Multi-Teacher Distillation Regression Model Based on Clustering Integration and Adaptive Weighting in Dam Deformation Prediction" Water 17, no. 7: 988. https://doi.org/10.3390/w17070988

APA Style

Guo, F., Yuan, J., Li, D., & Qin, X. (2025). Application of a Multi-Teacher Distillation Regression Model Based on Clustering Integration and Adaptive Weighting in Dam Deformation Prediction. Water, 17(7), 988. https://doi.org/10.3390/w17070988

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop