Next Article in Journal
Piezoelectric Effect of k-Carrageenan as a Tool for Force Sensor
Previous Article in Journal
Early Bearing Fault Diagnosis in PMSMs Based on HO-VMD and Weighted Evidence Fusion of Current–Vibration Signals
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Timestamp-Guided Knowledge Distillation for Robust Sensor-Based Time-Series Forecasting

1
School of Computer Science and Technology, Beijing Jiaotong University, Beijing 100044, China
2
Inner Mongolia High Tech Holdings Co., Ltd., Ordos 017000, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(15), 4590; https://doi.org/10.3390/s25154590
Submission received: 25 June 2025 / Revised: 18 July 2025 / Accepted: 23 July 2025 / Published: 24 July 2025
(This article belongs to the Section Intelligent Sensors)

Abstract

Accurate time-series forecasting plays a vital role in sensor-driven applications such as energy monitoring, traffic flow prediction, and environmental sensing. While most existing approaches focus on extracting local patterns from historical observations, they often overlook the global temporal information embedded in timestamps. However, this information represents a valuable yet underutilized aspect of sensor-based data that can significantly enhance forecasting performance. In this paper, we propose a novel timestamp-guided knowledge distillation framework (TKDF), which integrates both historical and timestamp information through mutual learning between heterogeneous prediction branches to improve forecasting robustness. The framework comprises two complementary branches: a Backbone Model that captures local dependencies from historical sequences, and a Timestamp Mapper that learns global temporal patterns encoded in timestamp features. To enhance information transfer and reduce representational redundancy, a self-distillation mechanism is introduced within the Timestamp Mapper. Extensive experiments on multiple real-world sensor datasets—covering electricity consumption, traffic flow, and meteorological measurements—demonstrate that the TKDF consistently improves the performance of mainstream forecasting models.

1. Introduction

Time-series forecasting plays a vital role in various sensor-based application, including transportation [1,2,3], energy [4], and weather [5]. Since time series consist of sequential data collected at fixed time intervals, capturing both local and global dependencies is crucial for accurate prediction. Recently, neural networks have realized promising achievements owing to their strong data fitting and automatic learning capabilities. However, most of the existing studies primarily focus on local observations within historical windows, underestimating abundant global information contained in the timestamp.
Deep learning approaches, such as Recurrent Neural Networks (RNNs), Convolutional Neural Networks (CNNs), and Transformers, have been extensively validated for their capability to automatically capture temporal dependencies [6,7,8]. Nevertheless, most of existing methods adopt a sequence-to-sequence prediction paradigm, where future sequences are predicted based on continuous historical sequences. These works typically focus on historical observations within local predefined windows, while overlooking the significant role of timestamps in forecasting. In fact, the timestamp contains substantial global and seasonal information, especially in sensor-generated time series where periodicity and external environmental patterns are often prominent.
For each timestamp, temporal features such as the month, day, weekday, hour, minute, and second can be extracted, all of which are highly relevant to accurate prediction. For example, traffic volume collected from roadside sensors often differs between weekdays and weekends, as well as between peak and off-peak hours, reflecting distinct temporal patterns. Owing to their rich informational content, timestamps have the potential to provide robust global guidance for forecasting models and can even independently generate prediction outputs.
Through a systematic review of state-of-the-art methods, existing works can be categorized into four groups based on how timestamps are utilized. As shown in Figure 1, the first category primarily concentrates on historical sequential observations, completely overlooking timestamps, such as Decomposition-Linear (DLinear) [9]. A significant portion of the existing works fall into this category. Instead of relying on historical sequences, the second category uses only the multiple temporal features contained in the timestamp, such as the Multi-Feature Spatial–Temporal fusion Network (MFSTN) [10]. For example, traffic volumes can be approximately estimated based on factors such as weather, holidays, day of the week, and peak hours. The third category, such as Informer [11], begins to consider timestamps by summing temporal embeddings with data embeddings. These intricate correlations prompt networks to extract information from more intuitive observations. The fourth category handles the historical sequence and timestamp independently through different models and combines two outputs using fusion weights, such as in the Global–Local Adaptive Fusion Framework (GLAFF) [12].
Despite the remarkable progress of existing models in time-series forecasting, the rich semantic information embedded in timestamps has not been fully exploited [13]. First, many approaches rely heavily on historical observations within fixed-size windows, which limits their ability to capture long-term seasonal or global patterns. Second, although timestamp features contain rich temporal semantics, they are often processed in overly simplistic ways (e.g., through summation with data embeddings), failing to unleash their potential in guiding predictions. For instance, Informer attempts to incorporate timestamp information by summing timestamp embeddings with data embeddings; the GLAFF processes the historical sequence and timestamp independently through separate models and then combines the outputs via fusion weights. However, these approaches fail to enable mutual learning between different models, which restricts their ability to fully capture the complementary knowledge within temporal data.
To overcome these limitations, we leverage knowledge distillation to facilitate the collaboration between heterogeneous models. This approach allows models to extract complementary temporal knowledge—local patterns from historical sequences and global patterns from timestamps. Through mutual guidance, these models can distill and integrate complementary information, leading to improved forecasting accuracy and robustness.
In this paper, we propose a timestamp-guided knowledge distillation framework (TKDF) that facilitates effective knowledge transfer among heterogeneous prediction models, aiming to enhance both robustness and predictive accuracy. This framework consists of a Backbone Model, which can be any sequence-based prediction model designed to capture local dependencies within a specific historical window. Additionally, a Timestamp Mapper is introduced, specifically designed to capture global dependencies by individually mapping timestamps to observation values. Unlike traditional methods, our approach integrates both local and global dependencies through mutual learning, which encourages the above two prediction branches to learn from one another, fostering a more comprehensive understanding. Notably, we incorporate self-distillation in the Timestamp Mapper to facilitate the knowledge transfer between deeper and shallower layers, thus enhancing the model’s capability. The contributions of this paper are summarized as follows:
(1)
We propose a novel temporal knowledge distillation framework consisting of two prediction branches that learn from each other. The framework is highly flexibility: the Backbone Model can be any sequence-based prediction model, and the Timestamp Mapper serves as a plug-and-play component that seamlessly collaborates with the Backbone Model.
(2)
We design a unique Timestamp Mapper with self-distillation. This self-distillation facilitates knowledge transfer from deeper to shallower layers within the same model, which is beneficial to enhancing its capacity to learn from broader contextual information in the timestamps.
(3)
Experiments are conducted using several real-world datasets collected from sensor-based systems to demonstrate the effectiveness of the proposed model.
The subsequent sections of this paper are organized as follows: Section 2 reviews the related work on knowledge distillation and time-series prediction. Section 3 describes the overall design and details of the proposed framework. Section 4 presents the experimental results, demonstrating the effectiveness of our approach in comparison to existing methods. Section 5 concludes the paper.

2. Related Work

This section reviews the relevant literature about knowledge distillation, and time-series prediction, with a focus on the underlying technical principles and their associated applications.

2.1. Knowledge Distillation

The original concept of knowledge distillation, proposed by Hinton et al., involves transferring knowledge from a complex teacher model with strong learning capacity to a simpler student model, thereby forming a teacher–student network [14]. The student model, typically lightweight, learns from both the ground truth (as the hard label) and the teacher’s output (as the soft label). By mimicking the behavior of the teacher through soft label supervision, the student model can achieve comparable performance while significantly reducing computational and memory overhead. Due to its superior performance, knowledge distillation has been successfully applied in the field of time-series prediction. For instance, Ma et al. proposed a multi-stage catalytic distillation framework for time-series forecasting which leverages knowledge distillation to facilitate efficient knowledge transfer in edge computing environments [15]. However, it may be challenging for a simpler student model to absorb all the knowledge from a complex teacher model. To address this issue, researchers have introduced several variations of the original knowledge distillation framework, such as mutual learning and self-distillation, as illustrated in Figure 2.
Mutual learning is a collaborative learning paradigm that enables multiple models to exchange knowledge, facilitating bidirectional interaction during the training process. This approach disrupts the traditional teacher–student relationship, allowing models to serve as both teachers and students to one another. Such interaction promotes knowledge sharing, ultimately improving prediction accuracy and enhancing the model’s generalization capability. For example, Wu et al. proposed a mutual learning module comprising multiple network branches, where interleaved multi-task supervision was employed to significantly boost performance [16]. Li et al. applied mutual learning to better capture complex correlations in traffic flow data and improve forecasting accuracy [17].
Self-distillation is a specialized form of knowledge distillation where a single network simultaneously functions as both teacher and student. Typically, the deeper layers of the network are used to supervise the shallower ones, as their outputs encode more complex and informative representations. This enables the shallower layers to acquire higher-level knowledge. Self-distillation facilitates self-optimization without the need for an external teacher model, thereby significantly reducing computational overhead and inference cost. For example, Kim et al. introduced a method called progressive self-knowledge distillation, applicable to any supervised learning task, which enhances prediction performance and generalization by distilling a model’s own knowledge [18]. Similarly, Ji et al. proposed a self-distillation mechanism during the pre-training phase, which can be regarded as an implicit regularization technique that transfers knowledge within the model itself [19].

2.2. Time-Series Prediction

As a significant practical challenge, time-series forecasting has attracted wide attention [20]. Time-series forecasting approaches have evolved into four categories: traditional statistics-based methods, machine learning-based methods, deep learning-based methods, and other emerging methods. The earliest methods based on traditional statistics, such as the auto-regressive integrated moving average (ARIMA) model [21], rely on temporal smoothness assumptions and are difficult to apply in practice [22,23]. Subsequently, machine learning methods, such as Support Vector Regression (SVR), became widely adopted but require large amounts of data for manual feature extraction, limiting their capacity [24]. Ensemble learning paradigms, such as bagging and boosting, have been widely adopted in time-series forecasting tasks to improve predictive performance by aggregating multiple weak learners [25]. With the development of deep learning, neural networks, including RNN-based, CNN-based, Transformer-based, and other architectures, have demonstrated satisfactory prediction performance. Among them, RNN-based methods capture temporal dependencies in time-series data dynamically through their recurrent structure [26,27,28], with examples including Long Short-Term Memory (LSTM) and Gated Recurrent Unit (GRU). Unfortunately, this architecture often suffers from the vanishing gradient problem when handling long sequences. Meanwhile, CNN-based methods rely on convolutional kernels to perform local operations [29], capturing detailed information within their limited receptive fields. This makes it difficult to capture long-term dependencies and multiple repetitive patterns using traditional convolution. To alleviate this problem, Wu et al. proposed TimesNet [30], which transforms 1D inputs into 2D space and employs 2D convolution to capture multiple periodicities. Unlike RNNs and CNNs, Transformer-based methods utilize self-attention mechanisms [31], allowing them to effectively capture long-range dependencies. However, the self-attention mechanism has quadratic complexity with respect to sequence length, making it computationally expensive for extremely long sequences. To address this, Zhou et al. proposed Informer [11], which introduces a ProbSparse self-attention mechanism that reduces both computational complexity and memory usage. As a recent variant, the iTransformer model [32] repurposes the traditional Transformer architecture by embedding time series from different channels into variate tokens which are then processed by the attention mechanism. Recently, other emerging methods have also demonstrated promising performance. Zeng et al. proposed a lightweight model, named DLinear [9], which employs only simple linear operations to accelerate both training and inference speed. With the rapid advancement of artificial intelligence, large language models (LLMs) have recently demonstrated remarkable performance in time-series forecasting [33]. Despite strong capabilities, LLMs incur high deployment and inference costs, making them less suitable for resource-constrained or low-latency prediction scenarios.
In terms of timestamps, most existing methods focus on local observations within historical windows, treating timestamps merely as auxiliary information whose global significance remains largely underestimated. For example, DLinear completely overlooks timestamps. Informer and TimesNet incorporate timestamps by adding their embeddings to position and data embeddings. iTransformer embeds timestamp features separately into variate tokens, which may compromise the physical significance of timestamps. To fully exploit timestamps, we construct heterogeneous prediction branches that independently process historical sequences and timestamps, and then integrate both local and global temporal knowledge through mutual learning to generate the final predictions.

3. Methods

We propose a timestamp-guided knowledge distillation framework that can efficiently integrate the local and global temporal dependencies from historical sequences and timestamp information. In time-series forecasting, given the historical sequence of c channels within h time steps X = { x 1 , x h } h × c , our goal is to predict the future sequence in subsequent p time steps Y = { y 1 , y p } p × c . In addition to sequential observations, we extract several timestamp features to provide global information, such as month, day, weekday, hour, minute, and second. For example, for the timestamp “2018-06-02 12:00:00” at the sampling moment t , its features can be expressed as s t = [ 06 , 02 , 05 , 12 , 00 , 00 ] 1 × 6 . Unlike observations, both the historical timestamps within the predefined window S = { s 1 , s h } h × 6 and the future timestamps within the prediction window T = { t 1 , t p } p × 6 are known. Therefore, we can incorporate the local observations X and global timestamps S , T to provide robustness prediction for the values of Y .

3.1. Overview

The proposed timestamp-guided knowledge distillation framework is illustrated in Figure 3, comprising two primary components: Timestamp Mapper and Backbone Model. Our framework offers highly flexibility: the Backbone Model can be any sequence-based model that convers historical sequences into future sequences, while the Timestamp Mapper serves as a plug-and-play component that maps timestamp features to observation values.
During the pre-training stage, the capacity of the Timestamp Mapper is enhanced through self-distillation, enabling it to learn the relationship between timestamps and observations. Since both historical observations and timestamps are available, the Timestamp Mapper can transform historical timestamps S h × 6 into the final historical mapping X ^ h × c , aligning them with corresponding observations X h × c . Self-distillation feeds deep feature representations back to the shallow layers, allowing the model to acquire global temporal contextual knowledge more effectively.
During the multi-branch prediction stage, the well-trained Timestamp Mapper further maps future timestamps T p × 6 to the final future mapping Y ^ p × c , while the Backbone Model offers the future prediction Y ¯ p × c based on the historical sequence X h × c . Both prediction results are aligned with the ground truth Y p × c . Mutual learning facilitates two heterogeneous prediction branches to interact with each other, enabling them to exchange complementary temporal knowledge. In this way, the proposed framework integrates the local observations and global timestamps to provide accurate and robust predictions.

3.2. Timestamp Mapper

As depicted in Figure 3, the Timestamp Mapper utilizes the Transformer encoder structure, which consists of an embedding layer, attention blocks, and a projection layer. Initially, each timestamp is embedded to describe its properties, then passed through attention blocks for correlation calculation, and processed by a projection layer to obtain the initial mappings. The initial historical and future mappings are denoted as X ˜ h × c and Y ˜ p × c , respectively, conforming to standard distribution. Subsequently, the Denormalizer Module separately inverse normalizes the initial mappings X ˜ and Y ˜ to produce the final mappings X ^ and Y ^ , thereby alleviating the impact of data drift.
The Timestamp Mapper can sufficiently extract the global temporal information embedded in timestamps. Specifically, the procedure of transforming historical timestamps S into an initial mapping X ˜ is succinctly outlined as follows:
H 0 = Embedding ( S )
H i + 1 = Attention ( H i ) , i = 0 , , l 1
X ˜ = Projection ( H l )
where H i h × d denotes the intermediate representation output by the i-th network layer, and d represents the dimension of the intermediate output. The attention blocks are stacked in l layers to capture the high-level temporal information embedded within the timestamps. The primary computation process for the i-th attention block is delineated as follows:
H i = LayerNorm ( H i + MSA ( H i , H i , H i ) )
H i + 1 = LayerNorm ( H i + FeedForward ( H i ) )
where LayerNorm ( ) represents the common normalization layer and FeedForward ( ) denotes the feedforward network. MSA ( Q , K , V ) indicates the multi-head self-attention mechanism, where Q , K , V represent the query, key, and value matrix, respectively. To simplify implementation, both the embedding and projection layers consist of a single linear layer. Additionally, a dropout mechanism is adopted to alleviate overfitting problems and enhance the network’s generalization. Similarly, the process of transforming future timestamps T into an initial mapping Y ˜ mirrors the aforementioned procedure, achieved by simply substituting S and X ˜ in Equations (1)–(3) with T and Y ˜ , respectively.
To alleviate the impact of data drift, the conventional inverse normalization approach considers mean and variance to measure the distribution deviations. However, this approach is sensitive to extreme values and lacks robustness when the observations contain anomalies. Instead of relying on the mean and standard deviation, the Denormalizer Module uses median and quantile ranges to enhance the robustness of the Timestamp Mapper against anomalies. This procedure can be succinctly expressed as follows:
X ^ = X ˜ μ ˜ σ ˜ × σ + μ
Y ^ = Y ˜ μ ˜ σ ˜ × σ + μ
where μ ˜ 1 × c and μ 1 × c denote the median of the initial mapping X ˜ and the actual observation X , respectively. Similarly, σ ˜ 1 × c and σ 1 × c represent the quantile range (the distance between the q quantile and the 1 q quantile) of the initial mapping X ˜ and the actual observation X . When q = 0.75 , the quantile range corresponds to the interquartile range (IQR3). This process not only helps alleviate the problem of data drift, but also makes the distribution of the final mappings more consistent with the actual observations.

3.3. Self-Distillation Phase

As shown in Figure 3, self-distillation is incorporated into the Timestamp Mapper during the pre-training stage, enabling it to acquire higher-level contextual knowledge more effectively. According to the relevant definitions, the knowledge distillation loss function consists of hard label loss and soft label loss. This self-distillation process can be represented as follows:
L s = α L s ( H ) + ( 1 α ) L s ( S )
where α is the weight hyperparameter, L s ( H ) and L s ( S ) , respectively, represent the hard label loss and soft label loss of self-distillation. During the pre-training stage, the Timestamp Mapper maps historical timestamps S to initial historical mappings X ˜ , which are then transformed into final historical mappings X ^ through inverse normalization, followed by alignment with the corresponding observations X . Therefore, hard label loss can be expressed as follows:
L s ( H ) = d i s t ( X ^ , X ) = X ^ , X 2
where d i s t ( ) is the distance function, which is used to measure the distance between the final historical mappings X ^ and actual observations X . Additionally, the soft label loss can be expressed as follows:
L s ( S ) = d i s t ( H 0 , X ˜ ) = H 0 , X ˜ 2
where H 0 and X ˜ represent the first and last layer of the Timestamp Mapper, respectively. In this way, the deep feature representations are fed back to the shallow layers, allowing the attention blocks to focus on more crucial information. The total self-distillation loss function is as follows:
L s = X ^ , X 2 + H 0 , X ˜ 2

3.4. Mutual Learning Phase

Mutual learning involves two student networks learning from each other, as illustrated in the multi-branch prediction stage in Figure 3. Two neural networks, Net - Θ 1 and Net - Θ 2 , have both hard label loss and soft label loss. As shown in Figure 3, the Timestamp Mapper is denoted as Net - Θ 1 , whose mutual learning loss function can be expressed as follows:
L θ 1 = β L θ 1 ( H ) + ( 1 β ) L θ 1 ( S )
where β is the weight hyperparameter. L θ 1 ( H ) and L θ 1 ( S ) , respectively, represent the hard label loss and soft label loss of Net - Θ 1 , whose calculation process is as follows:
L θ 1 ( H ) = d i s t ( Y ^ , Y ) = Y ^ , Y 2
L θ 1 ( S ) = d i s t ( Y ^ , Y ¯ ) = Y ^ , Y ¯ 2
where L θ 1 ( H ) is used to measure the distance between the final future mapping Y ^ and the ground truth Y , and L θ 1 ( S ) is used to measure the distance between the final future mapping Y ^ and the sequence-based prediction Y ¯ . Similarly, the Backbone Model is denoted as Net - Θ 2 , whose mutual learning loss function is as follows:
L θ 2 = β L θ 2 ( H ) + ( 1 β ) L θ 2 ( S )
where L θ 2 ( H ) and L θ 2 ( S ) , respectively, represent the hard label loss and soft label loss of Net - Θ 2 , whose calculation process is as follows:
L θ 2 ( H ) = d i s t ( Y ¯ , Y ) = Y ¯ , Y 2
L θ 2 ( S ) = d i s t ( Y ¯ , Y ^ ) = Y ¯ , Y ^ 2
where L θ 2 ( H ) is used to measure the sequence-based prediction Y ¯ and the ground truth Y , and L θ 2 ( S ) is used to measure the distance between the sequence-based prediction Y ¯ and the final future mapping Y ^ . Mutual guidance between heterogeneous branches allows them to distill and integrate complementary knowledge, thereby improving prediction accuracy and robustness.
Overall, we employ the Timestamp Mapper and Backbone Model to extract the local and global temporal knowledge and enhance the accuracy and robustness of prediction. Algorithm 1 presents the overall process of the prediction algorithm in the TKDF.
Algorithm 1: Prediction algorithm process in TKDF—TKDF for Time-Series Prediction
Input: Historical timestamp S , future timestamp T , historical sequence X
Output: Future sequence Y
Hyperparameter: Learning rate r , weight α , β
1. Initialization: Network θ 1 and Network θ 2
2. [The pre-training stage]: utilize S to predict X
3. for i t e r < i t e r s do:
4. Calculate self-distillation loss: L s = α L s ( H ) + ( 1 α ) L s ( S )
5. Update parameters of the timestamp mapper: θ s = θ s + r L s θ s
6. end for
7. [The Multi-branch Prediction Stage]: utilize T and X to predict Y
8. for i t e r < i t e r s do:
9. Calculate mutual learning loss of network θ 1 : L θ 1 = β L θ 1 ( H ) + ( 1 β ) L θ 1 ( S )
10. Calculate mutual learning loss of network θ 2 : L θ 2 = β L θ 2 ( H ) + ( 1 β ) L θ 2 ( S )
11. Update parameters of network θ 1 : θ 1 = θ 1 + r L θ 1 θ 1
12. Update parameters of network θ 2 : θ 2 = θ 2 + r L θ 2 θ 2
13. Update the prediction result
14. End for

4. Experiments

To verify the effectiveness of the proposed framework, three types of experiments are conducted to address the following research questions:
RQ1: Compared with the existing models for time-series prediction, does the TKDF achieve improved predictive performance?
RQ2: What are the appropriate values for the key hyperparameters?
RQ3: How do the learning strategies and individual components adopted in the TKDF contribute to predictive performance?

4.1. Datasets

The performance of the proposed method is evaluated on seven real-world benchmark datasets collected from sensor-based systems, including Electricity, Traffic, Weather, and four ETT datasets. These datasets are publicly available and have been widely used for performance evaluation [12]. Table 1 summarizes the statistics of these datasets. The basic information of these datasets is as follows.
  • Electricity comprises hourly electricity consumption data from 321 customers, collected between 2012 and 2014.
  • Traffic contains hourly road occupancy rates collected by 862 sensors installed on freeways in the San Francisco Bay Area, covering the period from 2015 to 2016.
  • Weather contains 21 meteorological variables recorded every 10 min at stations across Germany in 2020.
  • ETT records the oil temperature and load characteristics of two power transformers from 2016 to 2018, with measurements collected at two different resolutions (15 min and 1 h) for each transformer, resulting in four datasets: ETTh1, ETTh2, ETTm1, and ETTm2.
We follow the standard segmentation protocol by chronologically dividing each dataset into training, validation, and testing sets to prevent information leakage. The segmentation ratio is set to 6:2:2. The Adam optimizer is employed to train the model. The batch size and learning rate are set to 32 and 0.001, respectively. The number of training epochs is fixed at 50. Specifically, the length of the history window is set to 96 for the Electricity, Traffic, Weather, and all four ETT datasets, while the prediction length is varied among {96, 192, 336, 720}. The number l of attention blocks in the Timestamp Mapper is set to 2, and the dropout rate p is set to 0.1. The quantile q used in the Denormalizer Module is configured to 0.75.

4.2. Backbone Model

To demonstrate the effectiveness of our framework, we select several existing mainstream networks to serve as the Backbone Model, including the Transformer-based Informer (2021) [11] and iTransformer (2024) [32], the linear-based DLinear (2023) [9], and the convolution-based TimesNet (2023) [30]. All aforementioned models represent previous state-of-the-art methods for time-series forecasting tasks. (1) Informer leverages the ProbSparse attention mechanism to efficiently handle exceedingly long input sequences. It also incorporates a generative decoder to alleviate error accumulation inherent in autoregressive forecasting methods. (2) DLinear employs decomposition-enhanced linear networks to achieve competitive forecasting performance. (3) TimesNet captures two-dimensional dependencies by transforming one-dimensional time series into a collection of two-dimensional tensors. It leverages multiple periodicities to embed intra-periodic and inter-periodic variations along the columns and rows of each tensor, respectively. (4) iTransformer embeds each individual channel into a token processed by the attention mechanism, facilitating the capture of inter-channel multivariate correlations. A feedforward network is then applied to each token to learn nonlinear series representations.
As described in Section 2, these models encompass three different strategies to handle timestamps in forecasting techniques, namely summation (Informer, TimesNet), concatenation (iTransformer), and omission (DLinear).

4.3. Evaluation Metrics

The Mean Square Error (MSE) and Mean Absolute Error (MAE) are used to evaluate the prediction performance of the proposed method. The smaller the value of the metrics, the better the predictive performance of the model.

4.4. Experiment 1: Prediction Performance of the Proposed Framework

Aiming to answer RQ1, the TKDF is compared with various mainstream baselines across different architectures. The experimental results are shown in Table 2.
In Table 2, lower values indicate better predictive performance. The best results for each setting are highlighted in bold. Across most datasets and all prediction horizons, the TKDF significantly enhances the performance of nearly all baseline models, demonstrating strong generalization capability. This demonstrates that the TKDF effectively integrates global information from timestamps and local information from historical windows, thereby enhancing the robustness and predictive capability of mainstream forecasting models. This makes the TKDF well suited for long-range and ultra-long-range time-series forecasting.
On the Traffic dataset, the TKDF is significantly superior to TimesNet, DLinear, and Informer, with particularly obvious improvements at steps 336 and 720 (e.g., MAE decreases from 0.437 to 0.341). It is shown that the TKDF can better capture the temporal patterns in complex and dynamic traffic data. On the ETT datasets, the TKDF consistently achieves the best performance in both MSE and MAE for long-horizon predictions (e.g., horizon 720), indicating its superior ability to model long-term dependencies. On the Electricity and Weather datasets, although the performance gains are relatively modest, the TKDF still maintains a leading position. This indicates that the TKDF can provide stable advantages even on datasets characterized by strong periodicity and relatively stable patterns.
Figure 4a–d illustrate the prediction performance of the four backbone networks of the TKDF. In Figure 4, historical data are marked in gray. The Traffic dataset records traffic volume in hourly granularity. We deliberately selected data with clear periodic patterns and used working-day data (marked in gray) as historical input to predict weekend traffic peaks. The results demonstrate that our method can effectively capture periodicity and provide accurate forecasting. Because the TKDF fully leverages the global information embedded in timestamps, it effectively captures temporal correlations and periodic patterns, enabling accurate prediction of both peak and off-peak values. It can be observed that iTransformer achieves the best prediction performance, demonstrating that the proposed TKDF is also effective when applied to state-of-the-art models.

4.5. Experiment 2: Parameter Determination

To address RQ2, we conduct a comprehensive hyperparameter analysis. It is worth noting that, in theory, the TKDF introduces the two most important core hyperparameters, the self-distillation weight α and the mutual learning weight β . To comprehensively compare the different parameter configurations, we also explore the quantile q of the Denormalizer Module, the number l of attention blocks, and the dropout proportion p in the Timestamp Mapper. Since iTransformer represents the previous state of the art in time-series forecasting, we implement our method using it as the backbone. The history window length is fixed at 96, while the prediction window length is set to 192. The experimental results are depicted below.
As shown in Figure 5, the self-distillation weight α is selected from {0.1, 0.3, 0.5, 0.7, 0.9}. We analyze the influence of different self-distillation weights. When α approaches 1, no soft label supervision is involved, meaning that the model relies solely on hard labels. In this case, it cannot benefit from deep semantic representations generated by the model itself and ultimately degenerates into a standard supervised learning paradigm. When α approaches 0, the model relies solely on soft labels, resulting in reduced robustness due to potential overfitting to internal representations. Therefore, the results demonstrate that moderate values (e.g., α = 0.5 ) yield the best performance, suggesting that both soft and hard labels contribute to improved predictive performance.
As shown in Figure 6, the mutual learning weight β is selected from {0.1, 0.3, 0.5, 0.7, 0.9}. We analyze the influence of different mutual learning weights. The model exhibits the worst performance when β = 0.1 or β = 0.9 . A high value of β indicates that the weight assigned to the mutual learning loss is insufficient, weakening the knowledge exchange between networks and consequently degrading the prediction performance. On the contrary, a low value of β indicates that the mutual learning loss is overemphasized, while the supervision from the ground truth (hard labels) is underweighted, which significantly compromises the model’s predictive performance. When β = 0.5 , both networks contribute complementary information, and the model achieves its best predictive performance.
As shown in Figure 7, the number l of attention blocks is selected from {1, 2, 3, 4, 5}. The impact of the number l of attention blocks on performance can be observed. Overall, the model exhibits low sensitivity to changes in layer depth, with two or three layers generally enough to achieve strong performance. However, increasing the network depth also leads to higher computational costs and greater difficulty in convergence. Considering training cost and model performance, we choose 2 as the number l of attention blocks.
As shown in Figure 8, the dropout proportion p is selected from {0.0, 0.1, 0.2, 0.3, 0.4}. We can observe the change trend of dropout proportion p . Most datasets perform better at a lower dropout rate (such as 0.1 or 0.2), indicating that an excessively high dropout rate may introduce excessive information loss. To simplify the parameter selection challenge, we adopt a dropout proportion p of 0.1.
As shown Figure 9, the quantile q is selected from {0.55, 0.65, 0.75, 0.85, 0.95}. The quantile parameter in the Denormalizer Module enables the TKDF to better adapt to data drift, particularly in the presence of anomalies within the sliding window. However, higher quantile values may cause the TKDF to become overly sensitive to magnitude variations, reducing its robustness to outliers. In contrast, lower quantiles may limit the TKDF’s ability to capture distributional changes, thereby impairing its adaptability to data drift. We can observe that the performance around q = 0.75 is better on almost all datasets, supporting the rationality of setting q = 0.75 .

4.6. Experiment 3: Ablation Experiments

To validate the effectiveness of the different learning strategies and core components, a series of ablation experiments are conducted below. We implement our proposed framework and its several variants using the iTransformer as the Backbone Model. The results of the ablation studies conducted on three representative benchmark datasets are summarized in Table 3.
Without self-distillation, the self-distillation mechanism within the Timestamp Mapper is removed, disabling the internal knowledge transfer between deep and shallow layers. As a result, the model exhibits a moderate performance drop across datasets, indicating that self-distillation plays a supportive role in enhancing the representation capacity of the Timestamp Mapper. This mechanism helps lower layers to incorporate richer temporal context from deeper layers, thereby improving the model’s generalization ability, especially when the timestamp information is complex or weakly periodic.
Without mutual learning, the mutual learning strategy between the Timestamp Mapper and the Backbone Model is disabled. Each component makes predictions independently without exchanging soft label guidance, equivalent to without the Backbone Model and without the Timestamp Mapper. This modification results in a noticeable degradation in performance, highlighting the importance of cross-branch collaboration. Compared to self-distillation, mutual learning leads to a more significant enhancement in predictive accuracy, as it facilitates bidirectional knowledge exchange between components, thereby capturing both local and global temporal dependencies more effectively.
Without the backbone, the Backbone Model is entirely removed from the TKDF, and predictions are made solely based on timestamps. Surprisingly, the TKDF still achieves competitive performance using only the Timestamp Mapper, highlighting its ability to effectively capture global temporal patterns embedded in the timestamps. This observation aligns with human cognitive processes when predicting traffic states, where we often do not rely on large amounts of historical data but instead make reasonably accurate predictions based on temporal features such as whether it is a weekend or peak hour. This fully demonstrates that the rich semantic information in timestamps can support independent prediction.
Without the Timestamp Mapper, the stacked attention blocks are replaced with MLP networks of the equivalent size. After this substitution, the TKDF exhibits a noticeable decline in performance, indicating that the model struggles to capture temporal dependencies among timestamps without the self-attention mechanism.
Without the quantile, the Denormalizer Module is replaced with conventional inverse normalization. The experimental results demonstrate that our proposed design is more effective. When the historical window contains anomalies, conventional normalization methods often produce inaccurate distribution estimates.

5. Discussion

The proposed timestamp-guided knowledge distillation framework (TKDF) has demonstrated significant improvements in forecasting robustness and accuracy across various sensor-based time-series datasets. In this section, we discuss the strengths and potential limitations of the framework, highlight its differences from existing models, and analyze its computational complexity.

5.1. Strengths and Innovations

The TKDF introduces a dual-branch architecture with a Backbone Model for capturing local dependencies and a Timestamp Mapper for modeling global temporal semantics. The mutual learning mechanism between these two branches enables effective integration of complementary information. Furthermore, the incorporation of self-distillation in the Timestamp Mapper allows for enhanced extraction of timestamp semantics, improving the generalization capability across different datasets.

5.2. Comparison with Existing Models

Unlike the GLAFF, which treats historical sequences and timestamp features independently and fuses their outputs with fixed weights, the TKDF allows for collaborative knowledge exchange between the two branches during training. This design facilitates deeper integration of local and global temporal patterns. Compared to the MFSTN, which directly maps timestamp features to predictions and neglects historical sequences, the TKDF leverages both components synergistically, ensuring richer temporal context awareness.

5.3. Computational Complexity Analysis

The TKDF is designed to balance accuracy and efficiency. By limiting the number of attention blocks to two and adopting lightweight components for both the Backbone Model and Timestamp Mapper, the framework maintains a manageable parameter size and training time. Preliminary observations indicate that increasing attention depth beyond two provides no noticeable improvement in accuracy while introducing additional computational overhead. This trade-off was a key consideration in finalizing the architecture.

5.4. Limitations and Future Work

One limitation of the TKDF is its reliance on pre-training for the Timestamp Mapper, which could increase the overall training time in large-scale applications. Additionally, while the TKDF has shown robustness across multiple datasets, its performance on highly irregular or sparse time-series data requires further investigation. Future work will explore alternative fusion strategies, adaptive attention mechanisms, and more efficient distillation techniques to address these challenges.

6. Conclusions

In this paper, we propose a novel timestamp-guided knowledge distillation framework (TKDF) that integrates both local and global temporal dependencies through a multi-branch mutual learning architecture. Specifically, the TKDF consists of two heterogeneous prediction branches: a Backbone Model that captures local dependencies from historical sequential data, and a Timestamp Mapper designed to model global dependencies inherent in timestamps. The two branches interact through mutual learning, allowing for bidirectional knowledge transfer via soft label supervision. Furthermore, self-distillation is incorporated into the Timestamp Mapper to enhance its capacity by transferring deep representations to shallower layers. Extensive experiments on multiple sensor-based datasets demonstrate that the TKDF significantly boosts the performance of mainstream forecasting models, particularly in capturing periodic patterns, improving robustness, and handling long-range dependencies. On the Traffic dataset, for example, the TKDF is significantly superior to TimesNet, DLinear, and Informer, with particularly obvious improvements at steps 336 and 720 (e.g., MAE decreases from 0.437 to 0.341).

Author Contributions

Conceptualization, J.Y.; methodology, J.Y.; formal analysis, H.L. (Honghui Li); data curation, Y.B. (Yanhui Bai) and J.L.; validation and provision of technical standards compliance review, Y.B. (Yang Bai); writing—original draft, J.Y.; project administration, H.L. (Hairui Lv); writing—review and editing, H.L. (Honghui Li) and Y.B. (Yanhui Bai); supervision, H.L. (Honghui Li); funding acquisition, H.L. (Honghui Li). All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Major Scientific Project “Unveiling and Leading” of Ordos City, grant number JBGS-2023-002.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data can be provided upon request.

Conflicts of Interest

Authors Hairui Lv and Yang Bai are employed by the company Inner Mongolia High Tech Holdings Co., Ltd. Author Honghui Li is partially supported by a collaborative project between Beijing Jiaotong University and the company Inner Mongolia High Tech Holdings Co., Ltd. The authors declare no additional commercial or financial conflicts of interest beyond this acknowledged project funding.

References

  1. Zhang, J.; Zheng, Y.; Qi, D. Deep Spatio-Temporal Residual Networks for Citywide Crowd Flows Prediction. In Proceedings of the 31st AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017. [Google Scholar]
  2. Liu, Y.; Yu, J.J.Q.; Kang, J.; Niyato, D.; Zhang, S. Privacy-Preserving Traffic Flow Prediction: A Federated Learning Approach. IEEE Internet Things J. 2020, 7, 7751–7763. [Google Scholar] [CrossRef]
  3. Zhang, L.; Zhu, J.; Jin, B.; Wei, X. Multiview Spatial-Temporal Meta-Learning for Multivariate Time Series Forecasting. Sensors 2024, 24, 4473. [Google Scholar] [CrossRef] [PubMed]
  4. Pinto, T.; Praça, I.; Vale, Z.; Silva, J. Ensemble Learning for Electricity Consumption Forecasting in Office Buildings. Neurocomputing 2021, 423, 747–755. [Google Scholar] [CrossRef]
  5. Du, S.; Li, T.; Yang, Y.; Horng, S.J. Deep Air Quality Forecasting Using Hybrid Deep Learning Framework. IEEE T. Knowl. Data En. 2019, 33, 2412–2424. [Google Scholar] [CrossRef]
  6. He, Q.Q.; Siu, S.W.I.; Si, Y.W. Instance-based Deep Transfer Learning with Attention for Stock Movement Prediction. Appl. Intell. 2023, 53, 6887–6908. [Google Scholar] [CrossRef]
  7. Liang, P.P.; Zadeh, A.; Morency, L.P. Foundations & Trends in Multimodal Machine Learning: Principles, Challenges, and Open Questions. ACM Comput. Surv. 2024, 56, 1–42. [Google Scholar] [CrossRef]
  8. Yin, X.; Wu, G.; Wei, J.; Shen, Y.; Qi, H.; Yin, B. Deep Learning on Traffic Prediction: Methods, Analysis, and Future Directions. IEEE Trans. Intell. Transp. Syst. 2022, 23, 3054840. [Google Scholar] [CrossRef]
  9. Zeng, A.; Chen, M.; Zhang, L.; Xu, Q. Are Transformers Effective for Time Series Forecasting? In Proceedings of the AAAI Conference on Artificial Intelligence, Washington, DC, USA, 7–14 February 2023.
  10. Yan, J.; Li, H.; Zhang, D.; Bai, Y.; Xu, Y.; Han, C. A Multi-feature Spatial–temporal Fusion Network for Traffic Flow Prediction. Sci. Rep. 2024, 14, 14264. [Google Scholar] [CrossRef] [PubMed]
  11. Zhou, H.; Zhang, S.; Peng, J.; Zhang, S.; Li, J.; Xiong, H.; Zhang, W. Informer: Beyond Efficient Transformer for Long Sequence Time-Series Forecasting. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 2–9 February 2021. [Google Scholar]
  12. Wang, C.; Qi, Q.; Wang, J.; Sun, H.; Zhuang, Z.; Wu, J.; Liao, J. Rethinking the Power of Timestamps for Robust Time Series Forecasting: A Global-Local Fusion Perspective. In Proceedings of the Annual Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 10–15 December 2024. [Google Scholar]
  13. Pereira, L.M.; Salazar, A.; Vergara, L.A. Comparative Analysis of Early and Late Fusion for the Multimodal Two-Class Problem. IEEE Access 2023, 11, 84283–84300. [Google Scholar] [CrossRef]
  14. Hinton, G.; Vinyals, O.; Dean, J. Distilling the knowledge in a neural network. arXiv 2015, arXiv:1503.02531. [Google Scholar] [PubMed]
  15. Ma, R.; Zhang, C.; Kang, Y.; Wang, X.; Qiu, C. MCD: Multi-Stage Catalytic Distillation for Time Series Forecasting. In Proceedings of the International Conference on Database Systems for Advanced Applications, Gifu, Japan, 2–5 July 2024. [Google Scholar]
  16. Wu, R.; Feng, M.; Guan, W.; Wang, D.; Lu, H.; Ding, E. A Mutual Learning Method for Salient Object Detection with Intertwined Multi-Supervision. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, California, CA, USA, 16–20 June 2019. [Google Scholar]
  17. Li, Y.; Li, P.; Yan, D.; Liu, Z. Deep Knowledge Distillation: A Self-Mutual Learning Framework for Traffic Prediction. Expert Syst. Appl. 2024, 252, 124138. [Google Scholar] [CrossRef]
  18. Kim, K.; Ji, B.M.; Yoon, D.; Hwang, S. Self-knowledge Distillation with Progressive Refinement of Targets. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montréal, QC, Canada, 11–17 October 2021. [Google Scholar]
  19. Ji, J.; Yu, F.; Lei, M. Self-supervised Spatiotemporal Graph Neural Networks with Self-Distillation for Traffic Prediction. IEEE T. Intell. Transp. 2022, 24, 1580–1593. [Google Scholar] [CrossRef]
  20. Li, W.; Law, K.L.E. Deep learning models for time series forecasting: A review. IEEE Access 2024, 12, 92306–92327. [Google Scholar] [CrossRef]
  21. Ariyo, A.A.; Adewumi, A.O.; Ayo, C.K. Stock Price Prediction Using the ARIMA Model. In Proceedings of the International Conference on Computer Modelling and Simulation, Cambridge, UK, 26–28 March 2014. [Google Scholar]
  22. Ghaderpour, E.; Dadkhah, H.; Dabiri, H.; Bozzano, F.; Scarascia Mugnozza, G.; Mazzanti, P. Precipitation Time Series Analysis and Forecasting for Italian Regions. Eng. Proc. 2023, 39, 23. [Google Scholar]
  23. Puri, C.; Kooijman, G.; Vanrumste, B.; Luca, S. Forecasting Time Series in Healthcare with Gaussian Processes and Dynamic Time Warping Based Subset Selection. IEEE J. Biomed. Health 2022, 26, 6126–6137. [Google Scholar] [CrossRef] [PubMed]
  24. Agarap, A.F.M. A Neural Network Architecture Combining Gated Recurrent Unit (GRU) and Support Vector Machine (SVM) for Intrusion Detection in Network Traffic Data. In Proceedings of the ACM International Conference Proceeding Series, Macau, China, 26–28 February 2018. [Google Scholar]
  25. Bouhali, A.; Zeroual, A.; Harrou, F. Enhancing Traffic Flow Prediction with Machine Learning Models: A Comparative Study. In Proceedings of the 2024 International Conference of the African Federation of Operational Research Societies, Tlemcen, Algeria, 3–5 November 2024. [Google Scholar]
  26. Hochreiter, S.; Schmidhuber, J. Long Short-Term Memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  27. Li, Y.; Yu, R.; Shahabi, C.; Liu, Y. Diffusion Convolutional Recurrent Neural Network: Data-Driven Traffic Forecasting. arXiv 2018, arXiv:1707.01926. [Google Scholar] [CrossRef]
  28. Jia, Y.; Lin, Y.; Hao, X.; Lin, Y.; Guo, S.; Wan, H. WITRAN: Water-Wave Information Transmission and Recurrent Acceleration Network for Long-Range Time Series Forecasting. In Proceedings of the Annual Conference on Neural Information Processing Systems, New Orleans, LA, USA, 10–16 December 2023. [Google Scholar]
  29. Zhao, L.; Song, Y.; Zhang, C.; Liu, Y.; Wang, P.; Lin, T.; Deng, M.; Li, H. T-GCN: A Temporal Graph Convolutional Network for Traffic Prediction. IEEE Trans. Intell. Transp. Syst. 2020, 21, 3848–3858. [Google Scholar] [CrossRef]
  30. Wu, H.; Hu, T.; Liu, Y.; Zhou, H.; Wang, J.; Long, M. TimesNet: Temporal 2D-Variation Modeling for General Time Series Analysis. In Proceedings of the International Conference on Learning Representations, Kigali, Rwanda, 1–5 May 2023. [Google Scholar]
  31. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is all you need. In Proceedings of the Annual Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  32. Liu, Y.; Hu, T.; Zhang, H.; Wu, H.; Wang, S.; Ma, L.; Long, M. iTransformer: Inverted Transformers Are Effective for Time Series Forecasting. In Proceedings of the International Conference on Learning Representations, Vienna, Austria, 7–11 May 2024. [Google Scholar]
  33. Kaushik, S.; Choudhury, A.; Sheron, P.K.; Dasgupta, N.; Natarajan, S.; Pickett, L.A.; Dutt, V. AI in Healthcare: Time-Series Forecasting Using Statistical, Neural, and Ensemble Architectures. Front. Big Data 2020, 3, 4. [Google Scholar] [CrossRef] [PubMed]
Figure 1. A comparison between our method and existing approaches.
Figure 1. A comparison between our method and existing approaches.
Sensors 25 04590 g001
Figure 2. The original knowledge distillation framework and its main variations.
Figure 2. The original knowledge distillation framework and its main variations.
Sensors 25 04590 g002
Figure 3. An overview of the proposed timestamp-guided knowledge distillation framework.
Figure 3. An overview of the proposed timestamp-guided knowledge distillation framework.
Sensors 25 04590 g003
Figure 4. Prediction results of mainstream baseline models enhanced by TKDF.
Figure 4. Prediction results of mainstream baseline models enhanced by TKDF.
Sensors 25 04590 g004aSensors 25 04590 g004b
Figure 5. Hyperparameter analysis of self-distillation weight.
Figure 5. Hyperparameter analysis of self-distillation weight.
Sensors 25 04590 g005
Figure 6. Hyperparameter analysis of mutual learning weight.
Figure 6. Hyperparameter analysis of mutual learning weight.
Sensors 25 04590 g006
Figure 7. Hyperparameter analysis of attention block number.
Figure 7. Hyperparameter analysis of attention block number.
Sensors 25 04590 g007
Figure 8. Hyperparameter analysis of dropout rate.
Figure 8. Hyperparameter analysis of dropout rate.
Sensors 25 04590 g008
Figure 9. Hyperparameter analysis of quantile.
Figure 9. Hyperparameter analysis of quantile.
Sensors 25 04590 g009
Table 1. The statistics of each dataset. Channel indicates the number of variates in each dataset. Length refers to the total number of time points. Frequency denotes the sampling interval.
Table 1. The statistics of each dataset. Channel indicates the number of variates in each dataset. Length refers to the total number of time points. Frequency denotes the sampling interval.
DatasetChannelLengthFrequencyInformation
Electricity32126,3041 hEnergy
Traffic86217,5441 hTransportation
Weather2152,69610 minClimate
ETTh1 and ETTh2717,4201 hEnergy
ETTm1 and ETTm2769,68015 minEnergy
Table 2. The forecasting errors on multivariate datasets for the TKDF and mainstream baselines.
Table 2. The forecasting errors on multivariate datasets for the TKDF and mainstream baselines.
ModelsInformer+OursTimesNet+OursDLinear+OursiTransformer+Ours
MetricMSEMAEMSEMAEMSEMAEMSEMAEMSEMAEMAEMAEMSEMAEMSEMAE
Traffic960.6410.3470.5980.3320.5980.3210.5650.3060.6020.3900.5610.3680.3970.2730.3850.265
1920.7080.3980.6770.3750.6230.3390.6010.3250.6370.3980.5100.3200.4380.2810.4290.274
3360.8400.4390.7020.3910.6450.3430.6270.3310.6530.4370.5140.3410.4560.3150.4370.289
7200.8640.4520.7980.4060.6620.3580.6420.3440.6970.4690.5350.3840.4690.3320.4410.297
Electricity960.3340.4200.2960.3830.1860.2790.1640.2450.1980.2810.1520.2420.1500.2410.1420.231
1920.3650.4340.3430.4060.2030.2980.1800.2670.2040.2900.1750.2710.1690.2570.1680.250
3360.3870.4410.3690.4170.2240.3190.1920.2930.2110.3090.1940.2860.1780.2670.1700.237
7200.4060.4490.3940.4280.2400.3270.2220.3190.2530.3380.2300.3140.2290.3240.2170.304
ETTh1960.9380.7420.6320.6390.4560.4870.4370.4620.4190.4420.4030.4120.4000.4270.3980.401
1921.2970.8360.8720.6980.5430.5310.5230.5130.4850.5070.4700.4830.4620.5410.4410.528
3361.3680.8780.9030.7040.6210.5720.5810.5560.5260.5230.4980.5040.5030.4580.4740.512
7201.3640.8920.9730.7370.8010.6910.7040.6530.7120.6120.6970.6000.7210.6310.6980.603
ETTh2960.7890.6710.7680.6540.3470.3740.2960.3290.3420.3740.3240.3520.3060.3490.2980.347
1921.1970.6991.1320.6370.4320.4250.3670.3960.4800.4730.4530.4510.3830.4130.3740.409
3361.5411.0171.5071.0010.4670.4580.3920.4310.5920.5170.5670.4980.4300.4380.4230.421
7201.7971.2081.7631.1070.4830.4690.4160.4020.6430.5960.6090.5070.4330.4540.4240.443
ETTm1960.5980.5610.5100.4830.3690.4170.3490.3810.3320.3870.3120.3500.3960.4070.3430.372
1920.6210.5820.5390.5320.4420.4380.4110.3940.3920.3920.3670.3830.4180.4330.3920.403
3360.8920.7810.7240.6390.5030.4990.4820.4630.4490.4360.4380.4240.4720.4760.4390.441
7201.0830.7940.9280.7320.5510.5270.5130.5130.5010.4830.4930.4870.5230.5160.5030.498
ETTm2960.2370.3120.2180.2700.1830.2670.1610.2320.1970.2980.1900.2930.1890.2680.1790.263
1920.3140.3470.2970.3230.2400.3130.2270.2680.2930.3710.2870.3690.2640.3180.2480.309
3360.4780.4820.4630.3840.3170.3350.3040.2970.3750.4350.3630.4270.3200.3640.3070.342
7200.8780.6730.7320.5910.4140.3980.4030.3720.5680.5380.5500.5280.4230.4270.4010.413
Weather961.2480.8670.6490.5670.1780.2310.1740.2240.1980.2570.1720.2470.1800.2250.1610.220
1921.2720.8820.8720.6780.2230.2790.2210.2790.2420.2310.2240.2290.2280.2630.2130.261
3361.3061.0120.9980.8610.2890.3130.2860.2970.3910.3390.3690.3170.2850.3130.2750.308
7201.3631.1311.3200.8920.3680.3690.3520.3620.3500.3860.3340.3650.3610.3500.3500.346
Table 3. Ablation study on TKDF and its variants.
Table 3. Ablation study on TKDF and its variants.
MethodiTransformer+Oursw/o Mutual Learningw/o Quantilew/o Self-Distillation
w/o Backbonew/o Timestamp Mapper
MetricMSEMAEMSEMAEMSEMAEMSEMAEMSEMAEMSEMAE
Traffic960.3970.2730.3850.2650.4320.2970.3950.2710.3890.2710.3890.270
1920.4380.2810.4290.2740.4630.3030.4430.2850.4330.2800.4310.279
3360.4560.3150.4370.2890.4780.3360.4570.3130.4460.2970.4440.293
7200.4690.3320.4410.2970.5100.3510.4700.3360.4590.3220.4510.310
Electricity960.1500.2410.1420.2310.1890.2680.1530.2440.1460.2350.1440.233
1920.1690.2570.1680.2500.2100.2790.1720.2590.1690.2540.1680.256
3360.1780.2670.1700.2370.2350.2940.1800.2710.1740.2480.1720.241
7200.2290.3240.2170.3040.2670.3480.2310.3300.2230.3170.2200.310
Weather960.1800.2250.1610.2200.2260.2490.1820.2270.1710.2230.1690.224
1920.2280.2630.2130.2610.2490.2890.2290.2660.2200.2620.2170.261
3360.2850.3130.2750.3080.3170.3430.2870.3150.2790.3110.2790.308
7200.3610.3500.3500.3460.3840.3760.3650.3510.3580.3490.3570.347
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yan, J.; Li, H.; Bai, Y.; Liu, J.; Lv, H.; Bai, Y. Timestamp-Guided Knowledge Distillation for Robust Sensor-Based Time-Series Forecasting. Sensors 2025, 25, 4590. https://doi.org/10.3390/s25154590

AMA Style

Yan J, Li H, Bai Y, Liu J, Lv H, Bai Y. Timestamp-Guided Knowledge Distillation for Robust Sensor-Based Time-Series Forecasting. Sensors. 2025; 25(15):4590. https://doi.org/10.3390/s25154590

Chicago/Turabian Style

Yan, Jiahe, Honghui Li, Yanhui Bai, Jie Liu, Hairui Lv, and Yang Bai. 2025. "Timestamp-Guided Knowledge Distillation for Robust Sensor-Based Time-Series Forecasting" Sensors 25, no. 15: 4590. https://doi.org/10.3390/s25154590

APA Style

Yan, J., Li, H., Bai, Y., Liu, J., Lv, H., & Bai, Y. (2025). Timestamp-Guided Knowledge Distillation for Robust Sensor-Based Time-Series Forecasting. Sensors, 25(15), 4590. https://doi.org/10.3390/s25154590

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop