Next Article in Journal
Fog-Enabled Machine Learning Approaches for Weather Prediction in IoT Systems: A Case Study
Previous Article in Journal
Fault Diagnosis Method for Shearer Arm Gear Based on Improved S-Transform and Depthwise Separable Convolution
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Unlabeled-Data-Enhanced Tool Remaining Useful Life Prediction Based on Graph Neural Network

School of Mechanical Engineering, Jiangsu University of Science and Technology, Zhenjiang 212000, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(13), 4068; https://doi.org/10.3390/s25134068
Submission received: 28 April 2025 / Revised: 5 June 2025 / Accepted: 25 June 2025 / Published: 30 June 2025
(This article belongs to the Section Industrial Sensors)

Abstract

Remaining useful life (RUL) prediction of cutting tools plays an important role in modern manufacturing because it provides the criterion used in decisions to replace worn cutting tools just in time so that machining deficiency and unnecessary costs will be restrained. However, the performance of existing deep learning algorithms is limited due to the smaller quantity and low quality of labeled training datasets, because it is costly and time-consuming to build such datasets. A large amount of unlabeled data in practical machining processes is underutilized. To solve this issue, an unlabeled-data-enhanced tool RUL prediction method is proposed to make full use of the abundant accessible unlabeled data. This paper proposes a novel and effective method for utilizing unlabeled data. This paper defines a custom criterion and loss function to train on unlabeled data, thereby utilizing the valuable information contained in these unlabeled data for tool RUL prediction. The physical rule that tool wear increases with the increasing number of cuts is employed to learn knowledge crucial for tool RUL prediction from unlabeled data. Model parameters trained on unlabeled data contain this knowledge. This paper then transfers the parameters through transfer learning to another model based on labeled data for tool RUL prediction, thus completing unlabeled data enhancement. Since multiple sensors are frequently used to simultaneously collect cutting data, this paper uses a graph neural network (GNN) for multi-sensor data fusion, extracting more useful information from the data to improve unlabeled data enhancement. Through multiple sets of comparative experiments and validation, the proposed method effectively enhances the accuracy and generalization capability of the RUL prediction model for cutting tools by utilizing unlabeled data.

1. Introduction

Statistics show that downtime due to tool wear and damage accounts for approximately 20% of total production time [1]. Furthermore, relevant research shows [2] that only 50% to 80% of tool useful life is utilized efficiently. It is not only necessary to replace tools before they become blunt but also to fully utilize the RUL of tools to avoid resource waste and additional cutting costs. As a result, tool RUL prediction during the process of mechanical cutting is a necessary method. Tool RUL prediction usually depends on tool wear measured during machining to evaluate a tool’s degradation and estimate its residual lifespan.
Tool RUL prediction mainly includes physics-driven and data-driven methods. Li et al. [3] introduced a physics-informed meta-learning framework that employs physical modeling strategies to predict tool wear across varying wear rates. Karandikar et al. [4] developed a physics-guided logistic classification method using logarithmic transformation of inputs to model the nonlinear relationship between cutting speed and tool life for the purposes of tool life modeling and cutting parameter optimization. Physics-driven methods require deep domain expertise and complex mathematical formulations, which can limit their adaptability to different tools and operating conditions, resulting in poor generalization capability.
Data-driven methods mainly include machine learning, especially deep learning methods. Sensors collect signals during the cutting process. Feature vectors are then extracted from these signals and input into machine learning or deep learning algorithms to predict tool RUL. Traditional machine learning methods include principal component analysis (PCA), signal processing, and classifiers [5,6,7,8,9,10]. Deep learning methods include convolutional neural networks (CNNs) [11,12], recurrent neural networks (RNNs) [13,14], autoencoders (AEs) [15], attention mechanisms [16], and others. Deep learning methods offer advantages like automatic feature extraction, the handling of complex nonlinear relationships, the processing of large datasets, robust generalization capability, and end-to-end learning, leading to superior performance in tool RUL prediction.
The research studies mentioned above on predicting tool wear and RUL with deep learning methods were mostly conducted on labeled data. However, in practical cutting processes, labeled data are scarce because accurately measuring tool wear and labeling the corresponding cutting data is a costly and time-consuming task, resulting in limited labeled data and abundant unlabeled data. This unlabeled data also contains valuable information about tool condition and wear, making the effective utilization of unlabeled data crucial for predicting tool RUL. Some researchers have utilized unlabeled data to enhance the performance of their tool wear and RUL prediction models. Shu et al. [17] introduced a self-supervised contrastive learning framework that uses multi-view contrastive learning to predict tool wear effectively with only a few labeled samples. Wang et al. [18] developed a physics-guided neural network that integrates physical knowledge with data-driven techniques, reducing reliance on abundant labeled data. Sun et al. [19] proposed a semi-supervised method using unreliable pseudo-labels, enhancing model generalization capability when there is a lack of reliable annotations. Warke et al. [20] and Zhu et al. [21] designed a hybrid physics–data-driven model fusion framework and an unsupervised dual-regression domain adversarial network, respectively, to adapt to tool wear prediction across varying working conditions, thereby improving robustness and generalization capability on unseen data. Li et al. [22] employed a semi-supervised multi-source meta-domain generalization method with instance-based domain adaptation to address tool wear state prediction under changing cutting conditions. Liu et al. [23] proposed an unsupervised chatter detection method combining autoencoders with a Gaussian mixture model (GMM) and K-means clustering, enabling effective detection without requiring abundant labeled data. In summary, these research studies show that the effective use of unlabeled data can enhance the accuracy and generalization capability of models for tool wear and RUL prediction.
Although these research studies use unlabeled data to varied degrees, limitations remain. Some rely on predefined physical models or domain-specific knowledge, limiting their flexibility and broad applicability across different contexts. Others, while using self-supervised or semi-supervised methods, may require substantial computational resources and time to process large-scale unlabeled data, resulting in inefficiencies. Moreover, several methods may be sensitive to initial parameters or fail to maintain stable performance in complex and variable working conditions. Since tool wear increases with increasing cuts, this paper captures the latent feature in cutting signals that evolves with increasing cuts. This paper defines a criterion to describe the trend of this evolving feature and then trains models on unlabeled datasets to match outputs with the defined criterion. To train for tool RUL prediction using labeled data, this paper first transfers trained parameters from an unlabeled data model via transfer learning, thereby transferring the knowledge learned from unlabeled data. This method is simpler, more efficient, and more generalizable, aiming to significantly enhance the accuracy and robustness of tool RUL prediction by fully utilizing abundant unlabeled data.
Multiple sensors are often used to collect signals during cutting tool operations, and the fusion of multi-sensor data is important to improve unlabeled data enhancement (better suited for the purpose to fully utilize unlabeled data) for tool RUL prediction. CNNs and their variants [24,25,26,27,28,29] have been widely used in the fusion of multi-sensor data for efficient tool RUL prediction as well as certain other models [30,31,32]. However, CNNs are primarily suited for regular grid-like data structures, such as 2D images, and have limited capability in processing non-Euclidean data like complex sensor networks. In multi-sensor data fusion, different types of sensor data often have complex topological relationships and dependencies, which CNNs may overlook by focusing on local spatial information extraction. To extract more valuable information from multi-sensor data for unlabeled-data-enhanced tool RUL prediction, this paper proposes the use of graph neural networks (GNNs) to aggregate multi-sensor signal features. GNNs have been successfully applied in various industrial research areas [33,34,35,36] with excellent performance. GNNs can process non-Euclidean data directly and effectively capture complex dependencies between sensors, making them suitable for the fusion of heterogeneous information from multi-sensor data.
In summary, this paper proposes an unlabeled-data-enhanced tool RUL prediction method based on GNNs. The main contributions of this paper are as follows:
(1)
To better achieve unlabeled data enhancement, this paper innovatively introduces a GNN to multi-sensor data fusion for tool RUL prediction. The GNN can directly process non-Euclidean data and capture complex dependencies between sensors, thus extracting more information from the intricate spatial and dynamic correlations in multi-sensor data.
(2)
By training on unlabeled data, this paper defines a criterion and loss function to learn knowledge crucial for tool RUL prediction. Model parameters trained on unlabeled data contain the knowledge and are then transferred to another model based on labeled data for tool RUL prediction, completing the unlabeled data enhancement.
The remainder of this paper is organized as follows: Section 2 details the proposed method; Section 3 outlines the experimental design; Section 4 provides results and discussion, validating the method; and Section 5 concludes this paper.

2. Methodology

To complete unlabeled-data-enhanced tool RUL prediction, this method primarily consists of the following aspects.
Firstly, feature extraction is applied on multi-sensor signals collected during each cut, converting each sensor signal into a fixed-dimensional feature vector. Secondly, given that graph neural networks (GNNs) require graph-structured data as input, the data from a cut is constructed into a graph. These graphs can be classified as either unlabeled or labeled data. This paper constructs 2 models: the ULDTM (unlabeled-data-trained model) and LDTM (labeled-data-trained model). Thirdly, the unlabeled data is input into the ULDTM for training based on the defined criterion and a custom loss function. After training on unlabeled data, parameters of the ULDTM learn certain knowledge crucial for enhancing tool RUL prediction. The labeled data is input into the LDTM for tool RUL prediction. Fourthly, the parameters from the GNN part of the ULDTM are transferred via transfer learning to initialize the GNN parameters in the LDTM, thereby transferring the knowledge learned from the unlabeled data. Finally, the LDTM outputs the predicted tool RUL and uses this output for training and validation. The overall methodology is illustrated in Figure 1.

2.1. Feature Extraction

Tool cutting data typically consist of multi-sensor time series (signals). We use feature extraction to convert the time series into fixed-dimensional feature vectors. This paper uses 30 time–frequency-domain characteristics to extract features from time series collected during the tool cutting process.
This paper uses 16 time–frequency-domain characteristics from reference [37]. In addition to the 16 time–frequency-domain characteristics mentioned in the reference, this paper supplemented 14 more time–frequency-domain characteristics. A total of 30 time–frequency-domain characteristics were extracted from the sensor signals, comprising 17 time-domain characteristics and 13 frequency-domain characteristics. In addition to the 16 characteristics mentioned in the citation, the remaining characteristics are as follows.
For a single sample, the collected signal is denoted as x = [x(1), …, x(i), …, x(N)], where N represents the number of sampling points in that sample. The time-domain characteristics are calculated using the formulas in Table 1.
To calculate frequency-domain characteristics from multi-sensor time series, this paper uses the Fast Fourier Transform (FFT). Let s(k) be the spectrum of the signal x, where k = 1, 2, …, K and K is the number of spectral lines. The frequency value of the k-th spectral line is denoted as f(k). The frequency-domain characteristics are calculated using the following formulas in Table 2.

2.2. Graph Neural Network

To improve unlabeled data enhancement, this paper uses a GNN for multi-sensor data fusion. The tool cutting process usually employs multiple sensors to collect signals (time series). Using the feature extraction methods described above, each sensor’s signal collected during a single cutting process is converted to a 30-dimensional feature vector. If n sensors are used, a single cutting process signal is converted into n 30-dimensional feature vectors.
In multi-sensor signals, different sensors capture different information, and their data is complexly interrelated. To effectively characterize these relationships and aggregate multi-sensor signal feature vectors for predicting tool RUL, this paper constructs graph-structured data from the multi-sensor network. This graph structure naturally captures the complex dependencies between sensors.
Graph Construction
  • Node: Each sensor is viewed as a node in the graph.
  • Node Embedding: The 30-dimensional feature vector extracted from each sensor’s signal is used to embed the corresponding node.
  • Edges: Edges between nodes represent the dependencies or correlations among sensors.
Thus, a single cutting process signal is converted into a graph, which consists of nodes with 30-dimensional feature vectors as node embeddings and edges between nodes. This conversion enables a more comprehensive and accurate representation of multi-sensor signals, enhancing effective tool RUL prediction. This method takes advantage of the inherent complexity of multi-sensor signals, providing a robust data framework for tool RUL prediction. The conversion of a single cutting process signal into graph data is illustrated in Figure 2.
Graph neural networks (GNNs) are deep learning models that process graph-structured data. GNNs can capture the complex interactions between sensor nodes using graph convolutional layers, combining signals from different sensors into a unified framework and accomplishing multi-sensor signal feature vector fusion. By more effectively aggregating information from multi-sensor data compared to other deep learning methods, the GNN improves unlabeled data enhancement as well as the accuracy of tool RUL prediction.
In the GNN, each node’s representation is updated by aggregating node embeddings from its neighboring nodes using graph convolutional layers. This aggregation process can be performed iteratively, allowing nodes to capture information from neighbors at longer distances. The GNN takes node-embedding vectors and edge indices of a graph as input and returns a single feature vector that represents the entire graph. Specifically, the structure of the GNN is illustrated in Figure 3.
Graph convolution layers: These layers work as follows. They input node embeddings and the edge index of a graph. They aggregate node embeddings through 2 layers of graph convolutions (GConv1 and GConv2), followed by batch normalization (BatchNorm). Selecting 2 layers of graph convolutions allows for effective information extraction during node-embedding aggregation and updating, enhancing model accuracy. The 2-layer graph convolution also ensures the model’s running time does not increase significantly, maintaining efficiency. If using GraphSAGE as a graph convolution layer, the process of aggregating node embeddings in GraphSAGE is as follows:
Input: Graph G = ( V , E ) , where V is the set of nodes and E is the set of edges. Input node embeddings for each node { x v v V } ; search depth K; weight matrices W k ,   k = 1 , , K ; activation function σ ; differentiable aggregation functions AGGREGATE k ,   k = 1 , , K ; neighbor selection function N ( v ) , which returns the set of neighboring nodes of node v.
The output is the updated node embedding z v for each node v.
The algorithm steps are as follows: Initialize the node embedding for each node:
h v 0 = x v ,   v V
GraphSAGE offers several different aggregation function options, which can capture the local features of neighboring nodes and send them to the target node. Commonly used aggregation functions include the following: mean aggregator, pooling aggregator, and LSTM aggregator. For each depth k = 1 , , K and each node v, aggregate the node embeddings from its neighbors:
h N ( v ) k = AGGREGATE k h u k 1 u N ( v )
Update the node embedding of node v by combining its own information and that of its neighbors:
h v k = σ W k CONCAT ( h v k 1 , h N ( v ) k )
h v k represents the representation of node v at layer k. In each layer, the new node embedding h v k + 1 of the target node v is generated by combining its own embedding h v k with the aggregated embeddings of its neighbors. The specific update rule is usually as follows:
h v k + 1 = σ W v k h v k + W N k Aggregate ( { h u k , u N ( v ) } )
Here, W v k and W N k are learnable weight matrices, and Aggregate is the aggregation function for the features of neighboring nodes. This process is iterated over multiple layers, with each layer further aggregating the local and neighbor information of the nodes.
Normalize the node embedding:
h v k = h v k h v k 2 , v V
Output the final node embedding for each node:
z v = h v K , v V
Activation function: Use the ReLU activation function to introduce nonlinearity.
Graph pooling layer: After each graph convolution, use a graph pooling layer (PoolLayer) to downsample the graph structure and reduce the number of nodes. After the graph pooling layer, use global mean pooling to convert all node embeddings into a single readout vector (output).
The proposed method using a GNN for multi-sensor data fusion and unlabeled-data-enhanced tool RUL prediction offers a robust and efficient solution for processing complex and interrelated data in practical cutting.
In this paper, the learning rate is set to 0.001 with an Adam optimizer, and a step learning rate scheduler is employed to adjust the learning rate during training. The batch size is fixed at 256 for the training process. The GNN architecture, specifically a GraphSage model, consists of two graph convolutional layers with 1024 hidden dimensions each.

2.3. Unlabeled-Data-Trained Model Training

To train a model for capturing latent features from unlabeled data, each unlabeled graph is input into a GNN, which processes the input and outputs an output vector. The GNN outputs B output vectors for a batch size of B unlabeled graphs. The L2 norms of the output vectors are then calculated. These output vectors subsequently pass through the L2 Norm Layer to obtain their corresponding L2 norms, denoted as l1, l2, …, lB. The L2 norm l of an output vector v = (v1, v2, …, vn) is calculated using the following formula:
l = v 2 = i = 1 n v i 2
This paper constructs a norm vector L = (l1, l2, …, lB) using L2 norms for each graph in a batch to calculate the monotonicity loss. Based on the fact that tool wear increases with the number of cuts increasing, we define a criterion that the L2 norm should increase with the cuts increasing. To meet this criterion, we define a monotonicity loss. This norm vector L is then input into a custom MonotonicityLoss layer, which calculates the monotonicity loss over the elements of L. In this paper, the monotonicity loss for L elements is defined as follows: for a norm vector, if the preceding element is greater than the succeeding one (i.e., it does not follow monotonically increasing order), the absolute difference between these two elements is calculated and added to the monotonicity loss. Finally, the monotonicity loss is the sum of all such losses. The formula for calculating the monotonicity loss is given by
MonotonicityLoss = i = 1 B 1 | l i l i + 1 | δ ( l i > l i + 1 )
To minimize the monotonicity loss, backpropagation is used to gradually reduce its value. A smaller monotonicity loss shows that elements in the norm vector L are increasing more monotonically, which suggests that the GNN is efficiently capturing latent features from unlabeled tool cutting data. The latent features evolve as the number of cuts increases. The forward and backward propagation processes for unlabeled data through the ULDTM are illustrated in Figure 4.

2.4. Tool RUL Prediction and Transfer Learning

Labeled data is used for tool RUL prediction. The labeled dataset is divided into training and validation sets. Specifically, the labeled data is input into the GNN and processed to output vectors (the same as in the ULDTM). These output vectors then pass through a fully connected layer, resulting in a single value that represents the predicted RUL of the tool.
The predicted RUL is compared to the actual labels to calculate the mean squared error loss (MSELoss), which is then backpropagated to update the network parameters. The LDTM is structured as shown in Figure 5.
During the unlabeled-data-training process, state dictionaries (model parameters) of the GNN part of the ULDTM are saved at each epoch. During the training process on unlabeled data, an epoch with a smaller monotonicity loss is chosen. The state dictionaries (model parameters) of the GNN part from this epoch are used as the initial parameters for the GNN part of the LDTM, completing transfer learning. Labeled data is then input into the LDTM to predict tool RUL, train, and validate.
The GNN parts of both the ULDTM and LDTM are identical. The part of the GNN used in the ULDTM is followed by a monotonicity loss layer to evaluate the model’s capability to capture latent features throughout the cutting process, while the part of the GNN used in the LDTM is followed by a fully connected layer for tool RUL prediction. The comparison between the two models and transfer learning can be described as shown in Figure 6.

3. Experimental Design

This paper uses the PHM2010 dataset to predict tool RUL of a three-flute ball nose cutter under identical cutting conditions. Additionally, the JUST_T05 dataset is used as a supplementary experiment.

3.1. Experimental Design of PHM2010

The structure of the PHM2010 experiment is shown in Figure 7.
The PHM2010 dataset includes cutting experiments for six tools, denoted as C1 to C6, with each experiment including 315 cuts. The experiments C2, C3, and C5 are unlabeled, while C1, C4, and C6 are labeled. Each dataset contains data collected by seven sensors: three for measuring cutting forces along the X, Y, and Z axes, three for vibration signals along the X, Y, and Z axes, and one for acoustic emission signals. The structure of the PHM2010 dataset is shown in Table 3.
To apply the proposed method to the PHM2010 dataset, signal feature extraction is performed on multi-sensor data, followed by the construction of a graph. Each cut from experiments C1 to C6 is converted into a single graph. A graph consists of seven nodes, each representing one of seven sensors: three for cutting forces, three for vibration signals, and one for acoustic emission signals. Feature vectors extracted from multi-sensor data are used to embed corresponding nodes. Nodes in each graph are fully connected (edges between nodes can be deleted, directed, or weighted according to further research to better reflect real-world conditions; in this study, we focus on undirected, fully connected graphs with a weight of 1 for all edges), forming a complete graph structure, as illustrated in Figure 8.
To begin with, the ULDTM is trained on unlabeled data to improve its capability to capture latent features from unlabeled data. Specifically, the cutting data from experiments C2 (cuts 1–289), C3 (cuts 1–238), and C5 (cuts 1–238) are used to construct graphs that are then input into the GNN. (The selection of the number of samples for the unlabeled datasets from C2, C3, and C5 is based on an estimation of the cutting samples before tool wear failure, and this number is flexible and can be adjusted according to research requirements.) The GNN output vectors’ L2 norms are calculated, then a monotonicity loss is calculated and backpropagated. During each training epoch, the parameters of the GNN of the ULDTM are saved. The parameters from the epoch with the smallest monotonicity loss are chosen as the initial parameters for the GNN of the LDTM.
In this paper, 0.16 mm of tool wear is used to define the tool failure threshold. Tool RUL labels are calculated as the total number of cuts from the start of the cutting process to the tool failure threshold minus the current cut number. For example, if the first 306 cuts from experiment C1 are chosen, the tool RUL label for the first cut is 305, the second cut is 304 and so on, with the 305th cut labeled 1 and the 306th cut labeled 0. Based on the tool failure threshold, 306 cuts are chosen from experiment C1 and 278 cuts from experiment C4. These chosen cuts form a training dataset of 584 samples (306 from C1 and 278 from C4). Additionally, the first 238 cuts from experiment C6 are chosen to form a validation dataset of 238 samples. The structure of the PHM2010 training/validation dataset is shown in Table 4.

3.2. Experimental Design of JUST_T05

Additionally, the JUST_T05 dataset is used as a supplementary experiment. JUST_T05 is a milling dataset built in a local laboratory. The cutting tool is a four-flute milling cutter. Similar to the PHM2010 dataset (Figure 7), the JUST_T05 dataset adopts a multi-sensor signal acquisition approach. In this paper, seven sensor signals are used to predict tool RUL. Three force sensors collect force signals in X, Y, and Z directions (N), three acceleration sensors collect vibration signals in X, Y, and Z directions (g), and one AE sensor collects the acoustic emission signal (dB). All cutting data are collected under the same working conditions.
The dataset is structured into three parts:
T_1: Unlabeled data used for training by the ULDTM.
T_2: Labeled data, containing 35 cuts, used for the LDTM as the training dataset.
T_3: Labeled data, containing 35 cuts, used for the LDTM as the validation dataset to predict tool RUL. Each cut in T_3 is structured as a graph, similar to the one shown in Figure 8.

4. Results and Discussion

4.1. PHM2010

To evaluate the capability of the ULDTM in capturing latent features from tool cutting data, this paper uses the unlabeled datasets from C2, C3, and C5. A smaller monotonicity loss indicates that the ULDTM is more able to capture latent features which evolve as the number of cuts increases. Specifically, the ULDTM separately receives 289 graphs from dataset C2, 238 graphs from C3, and 238 graphs from C5. The monotonicity loss output by each training epoch for datasets C2, C3, and C5 is recorded. The monotonicity loss output by the ULDTM for C2, C3, and C5 is shown in Figure 9.
The results indicate that as the number of epochs increases, the monotonicity loss gradually decreases and eventually converges. This trend shows that guided by the custom criterion, the model can learn to extract latent features from unlabeled data.
To improve the accuracy and generalization capability of tool RUL prediction LDTM, this paper uses the ULDTM trained on unlabeled datasets. The parameters of the GNN part are first trained on unlabeled data, and these trained parameters are then used as the initial parameters for the GNN part of the LDTM, accomplishing transfer learning. This method considerably reduces the RMSE (root mean square error) of the LDTM calculated between predicted tool RUL and actual labels.
We use the LDTM to train and validate tool RUL prediction, respectively, on the training dataset (C1 and C4) and validation dataset (C6) [38]. This paper compares the performance of the LDTM under four conditions: without transfer learning (non-transfer) and with transfer learning enhancement from unlabeled datasets C2 (C2-transfer, etc.), C3, and C5. The RMSE is recorded during both training and validation phases. To ensure the reliability and credibility of the experimental results, each result presented is the median of multiple trials. The RMSE values of training and validation phases without/with unlabeled dataset enhancement are shown in Figure 10 and Figure 11.
The RMSE is calculated by the formula given below:
RMSE = 1 n i = 1 n ( y ^ i y i ) 2
where y = { y 1 , y 2 , , y n } represents the real tool RUL and y ^ = { y ^ 1 , y ^ 2 , , y ^ n } represents the predicted tool RUL.
According to the results shown above, it is evident that the use of transfer learning leads to a more stable convergence of RMSE, as well as oscillation reduction and a smaller convergence value during both training and validation of tool RUL prediction. This indicates that the proposed method can effectively utilize unlabeled data to enhance the training and validation of labeled data tool RUL prediction.
A comparison chart is drawn to compare the real RUL (labels), the RUL prediction without/with transfer learning, and the absolute error between labels and tool RUL prediction on the validation dataset C6. The chart clearly illustrates that the RUL prediction using transfer learning is more accurate than that without transfer learning, with a smaller RMSE and more generalization capability. The comparison of tool RUL prediction without/with unlabeled data enhancement is shown in Figure 12.
In addition to the formula given above for calculating RMSE, the formulas for calculating MAE (mean absolute error) and MAPE (mean absolute percentage error) are as follows (because the real tool RUL of MAPE cannot appear as 0, the term with real tool RUL 0 is skipped):
MAE = 1 n i = 1 n | y ^ i y i |
MAPE = 100 % n i = 1 n y ^ i y i y i
This paper conducts additional experiments to study the broad applicability of the proposed method for using unlabeled data to improve tool RUL prediction on labeled data, as well as to demonstrate that a GNN can extract more useful information from multi-sensor data, thereby improving feature extraction quality. This paper replicates the proposed method in tool RUL prediction experiments using a convolutional neural network (CNN), multi-layer perceptron (MLP), transformer (Trans.), generative adversarial network (GAN), autoencoder (AE), and long short-term memory (LSTM). These models are also trained under two conditions: without transfer learning (no unlabeled data enhancement) and with transfer learning (unlabeled data enhancement). In these experiments, this paper continues to use C1 + C4 as the training dataset and C6 as the validation dataset. Model performance is evaluated by calculating the RMSE, MAE, and MAPE between the labels and the predicted tool RUL of C6.
The experimental results, as illustrated in the figure below, show that models achieve much smaller indicators (RMSE, MAE, and MAPE) in tool RUL prediction when using unlabeled data enhancement (the best result among C2-transfer, C3-transfer, and C5-transfer) than no unlabeled data enhancement (non-transfer). This shows that the proposed method improves the tool RUL prediction accuracy and generalization capability of the models. Moreover, the GNN outperforms other models in feature extraction from multi-sensor data, with smaller indicators under all conditions. In conclusion, the proposed method for unlabeled data enhancement not only improves the tool RUL prediction accuracy but also enhances the model’s generalization capability, confirming the effectiveness and broad applicability of the proposed method. The RMSE, MAE, and MAPE of tool RUL prediction for these models are shown in Figure 13.

4.2. JUST_T05

To verify the broad applicability of the proposed unlabeled data enhancement method for tool RUL prediction, this paper conducts additional experiments using the self-built dataset JUST_T05. In these experiments, T_1 is used to train the ULDTM. T_2 is used as the training dataset and T_3 as the validation dataset to predict tool RUL using the LDTM. The comparison of tool RUL prediction without/with unlabeled data enhancement on JUST_T05 is shown in Figure 14.
This paper focuses on enhancing tool RUL prediction accuracy and the generalization capability of models using the proposed unlabeled data enhancement method. Therefore, we applied the proposed method to other models to study its wide applicability. This paper replicates the proposed method on JUST_T05 using CNN, MLP, Trans., GAN, AE, LSTM. These models are also trained under two conditions: without transfer learning (no unlabeled data enhancement) and with transfer learning (unlabeled data enhancement). In these experiments, T_1 is used to train the ULDTM. T_2 is used as the training dataset and T_3 as the validation dataset. Model performance is evaluated by calculating the RMSE, MAE, and MAPE between the labels and the predicted tool RUL of T_3. As is shown in Figure 15, the proposed method also achieves smaller indicators on the self-built dataset JUST_T05. This demonstrates that the proposed method can effectively use unlabeled data, thereby improving the tool RUL prediction accuracy and generalization capability of the models.

5. Conclusions

This paper proposes a method for an unlabeled-data-enhanced tool remaining useful life prediction method based on a graph neural network. Compared with the tool RUL prediction model without unlabeled data enhancement, the model using unlabeled data enhancement achieves significant performance improvement. On the PHM2010 validation dataset, the RMSE of the tool RUL prediction model with unlabeled data enhancement is reduced by up to 50.6%, the MAE is reduced by up to 53.9%, and the MAPE is reduced by up to 23.66%. On the JUST_T05 validation dataset, the RMSE of the tool RUL prediction model with unlabeled data enhancement is reduced by up to 37.3%, the MAE is reduced by up to 29.1%, and the MAPE is reduced by up to 1.82%.
This method presents an innovative way for using unlabeled data to enhance tool RUL prediction on labeled data. Moreover, this paper uses a GNN to extract more useful information from multi-sensor data, thereby improving unlabeled data enhancement. The experimental results show that the proposed method improves the tool RUL prediction accuracy and generalization capability of the models. This paper offers a novel perspective for other researchers on utilizing unlabeled data and proves the applicability of GNNs in the field of unlabeled-data-enhanced tool RUL prediction.
In this paper, the graph-structured data is built as an undirected, fully connected graph with a weight of 1. In future research, some edges can be deleted, a directed graph constructed, and weights added to better reflect real-world conditions. Also, the custom monotonicity loss function in this paper can be replaced with other criteria (or does not need to be replaced), such as the coefficient of determination and trend strength, etc., to describe latent features that change when the number of cuts increases.

Author Contributions

Conceptualization, H.Z. and L.S.; methodology, D.G.; software, D.G.; validation, D.G., G.L. and L.S.; formal analysis, D.G.; investigation, D.G.; resources, H.Z.; data curation, H.Z.; writing—original draft preparation, D.G.; writing—review and editing, G.L.; visualization, D.G.; supervision, H.Z.; project administration, H.Z.; funding acquisition, L.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Natural Science Foundation of China grant number No. 62203193.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhu, K.; Guo, H.; Li, S.; Lin, X. Online tool wear monitoring by super-resolution based machine vision. Comput. Ind. 2023, 144, 103782. [Google Scholar] [CrossRef]
  2. Thangamuthu, M.; Tamilvanan, A. Decision support system for tool condition monitoring in milling process using artificial neural network. J. Eng. Res. 2022, 10, 142–155. [Google Scholar] [CrossRef]
  3. Li, Y.; Wang, J.; Huang, Z.; Gao, R.X. Physics-informed meta learning for machining tool wear prediction. J. Manuf. Syst. 2022, 62, 17–27. [Google Scholar] [CrossRef]
  4. Karandikar, J.; Schmitz, T.; Smith, S. Physics-guided logistic classification for tool life modeling and process parameter optimization in machining. J. Manuf. Syst. 2021, 59, 522–534. [Google Scholar] [CrossRef]
  5. Duan, J.; Hu, C.; Zhan, X.; Zhou, H.; Liao, G.; Shi, T. MS-SSPCANet: A powerful deep learning framework for tool wear prediction. Robot. Comput.-Integr. Manuf. 2022, 78, 102391. [Google Scholar] [CrossRef]
  6. Gharesi, N.; Arefi, M.M.; Razavi-Far, R.; Zarei, J.; Yin, S. A neuro-wavelet based approach for diagnosing bearing defects. Adv. Eng. Inform. 2020, 46, 101172. [Google Scholar] [CrossRef]
  7. Painuli, S.; Elangovan, M.; Sugumaran, V. Tool condition monitoring using K-star algorithm. Expert Syst. Appl. 2014, 41, 2638–2643. [Google Scholar] [CrossRef]
  8. Cai, W.; Zhang, W.; Hu, X.; Liu, Y. A hybrid information model based on long short-term memory network for tool condition monitoring. J. Intell. Manuf. 2020, 31, 1497–1510. [Google Scholar] [CrossRef]
  9. Ma, Z.; Zhao, M.; Dai, X.; Chen, Y. A hybrid-driven probabilistic state space model for tool wear monitoring. Mech. Syst. Signal Process. 2023, 200, 110599. [Google Scholar] [CrossRef]
  10. Wang, M.; Zhou, J.; Gao, J.; Li, Z.; Li, E. Milling tool wear prediction method based on deep learning under variable working conditions. IEEE Access 2020, 8, 140726–140735. [Google Scholar] [CrossRef]
  11. Gao, K.; Xu, X.; Jiao, S. Measurement and prediction of wear volume of the tool in nonlinear degradation process based on multi-sensor information fusion. Eng. Fail. Anal. 2022, 136, 106164. [Google Scholar] [CrossRef]
  12. Li, G.; Jiang, R.; Sun, L.; Zhou, H.; Liu, Y.; Shang, X. Prediction of bidirectional milling forces based on spindle current signals by using deep learning algorithms. Proc. Inst. Mech. Eng. Part E J. Process Mech. Eng. 2025. [Google Scholar] [CrossRef]
  13. Marani, M.; Zeinali, M.; Songmene, V.; Mechefske, C.K. Tool wear prediction in high-speed turning of a steel alloy using long short-term memory modelling. Measurement 2021, 177, 109329. [Google Scholar] [CrossRef]
  14. Wang, J.; Yan, J.; Li, C.; Gao, R.X.; Zhao, R. Deep heterogeneous GRU model for predictive analytics in smart manufacturing: Application to tool wear prediction. Comput. Ind. 2019, 111, 1–14. [Google Scholar] [CrossRef]
  15. Sun, H.; Zhang, J.; Mo, R.; Zhang, X. In-process tool condition forecasting based on a deep learning method. Robot. Comput.-Integr. Manuf. 2020, 64, 101924. [Google Scholar] [CrossRef]
  16. Li, W.; Fu, H.; Han, Z.; Zhang, X.; Jin, H. Intelligent tool wear prediction based on Informer encoder and stacked bidirectional gated recurrent unit. Robot. Comput.-Integr. Manuf. 2022, 77, 102368. [Google Scholar] [CrossRef]
  17. Shu, R.; Xu, Y.; He, J.; Yang, X.; Zhao, Z.; Huang, G.Q. Multi-view contrastive learning framework for tool wear detection with insufficient annotated data. Adv. Eng. Inform. 2024, 62, 102666. [Google Scholar] [CrossRef]
  18. Wang, J.; Li, Y.; Zhao, R.; Gao, R.X. Physics guided neural network for machining tool wear prediction. J. Manuf. Syst. 2020, 57, 298–310. [Google Scholar] [CrossRef]
  19. Sun, Y.; He, J.; Gao, H.; Song, H.; Guo, L. A New Semi-supervised Tool-wear Monitoring Method using Unreliable Pseudo-Labels. Measurement 2024, 226, 113991. [Google Scholar] [CrossRef]
  20. Warke, V.; Kumar, S.; Bongale, A.; Kotecha, K. Robust Tool Wear Prediction using Multi-Sensor Fusion and Time-Domain Features for the Milling Process using Instance-based Domain Adaptation. Knowl.-Based Syst. 2024, 288, 111454. [Google Scholar] [CrossRef]
  21. Zhu, Y.; Zi, Y.; Xu, J.; Li, J. An unsupervised dual-regression domain adversarial adaption network for tool wear prediction in multi-working conditions. Measurement 2022, 200, 111644. [Google Scholar] [CrossRef]
  22. Li, W.; Fu, H.; Zhuo, Y.; Liu, C.; Jin, H. Semi-supervised multi-source meta-domain generalization method for tool wear state prediction under varying cutting conditions. J. Manuf. Syst. 2023, 71, 323–341. [Google Scholar] [CrossRef]
  23. Liu, B.; Liu, C.; Zhou, Y.; Wang, D.; Dun, Y. An unsupervised chatter detection method based on AE and merging GMM and K-means. Mech. Syst. Signal Process. 2023, 186, 109861. [Google Scholar] [CrossRef]
  24. Yan, B.; Zhu, L.; Dun, Y. Tool wear monitoring of TC4 titanium alloy milling process based on multi-channel signal and time-dependent properties by using deep learning. J. Manuf. Syst. 2021, 61, 495–508. [Google Scholar] [CrossRef]
  25. Kou, R.; Lian, S.W.; Xie, N.; Lu, B.E.; Liu, X.M. Image-based tool condition monitoring based on convolution neural network in turning process. Int. J. Adv. Manuf. Technol. 2022, 119, 3279–3291. [Google Scholar] [CrossRef]
  26. Hou, W.; Guo, H.; Luo, L.; Jin, M. Tool wear prediction based on domain adversarial adaptation and channel attention multiscale convolutional long short-term memory network. J. Manuf. Process. 2022, 84, 1339–1361. [Google Scholar] [CrossRef]
  27. Li, G.; Shang, X.; Sun, L.; Fu, B.; Yang, L.; Zhou, H. Application of audible sound signals in tool wear monitoring: A review. J. Adv. Manuf. Sci, Technol. 2025, 5, 2025003. [Google Scholar] [CrossRef]
  28. Traini, E.; Bruno, G.; D’antonio, G.; Lombardi, F. Machine learning framework for predictive maintenance in milling. IFAC-PapersOnLine 2019, 52, 177–182. [Google Scholar] [CrossRef]
  29. Ma, J.; Luo, D.; Liao, X.; Zhang, Z.; Huang, Y.; Lu, J. Tool wear mechanism and prediction in milling TC18 titanium alloy using deep learning. Measurement 2021, 173, 108554. [Google Scholar] [CrossRef]
  30. Lin, M.; Wanqing, S.; Chen, D.; Zio, E. Evolving connectionist system and hidden semi-Markov model for learning-based tool wear monitoring and remaining useful life prediction. IEEE Access 2022, 10, 82469–82482. [Google Scholar] [CrossRef]
  31. Li, X.; Liu, X.; Yue, C.; Wang, L.; Liang, S.Y. Data-model linkage prediction of tool remaining useful life based on deep feature fusion and Wiener process. J. Manuf. Syst. 2024, 73, 19–38. [Google Scholar] [CrossRef]
  32. Wang, Z.; Zhou, G.; Zhang, C.; Liu, J.; Chang, F.; Zhou, Y.; Han, C.; Zhao, D. An adaptive RUL prediction approach for cutting tools incorporated with interpretability and uncertainty. Reliab. Eng. Syst. Saf. 2025, 256, 110705. [Google Scholar] [CrossRef]
  33. Li, C.; Mo, L.; Yan, R. Fault diagnosis of rolling bearing based on WHVG and GCN. IEEE Trans. Instrum. Meas. 2021, 70, 3519811. [Google Scholar] [CrossRef]
  34. Yu, X.; Tang, B.; Zhang, K. Fault diagnosis of wind turbine gearbox using a novel method of fast deep graph convolutional networks. IEEE Trans. Instrum. Meas. 2021, 70, 6502714. [Google Scholar] [CrossRef]
  35. Chen, K.; Hu, J.; Zhang, Y.; Yu, Z.; He, J. Fault location in power distribution systems via deep graph convolutional networks. IEEE J. Sel. Areas Commun. 2019, 38, 119–131. [Google Scholar] [CrossRef]
  36. Li, X.; Yan, X.; Gu, Q.; Zhou, H.; Wu, D.; Xu, J. Deepchemstable: Chemical stability prediction with an attention-based graph convolution network. J. Chem. Inf. Model. 2019, 59, 1044–1049. [Google Scholar] [CrossRef]
  37. Li, G.; Xu, S.; Jiang, R.; Liu, Y.; Zhang, L.; Zheng, H.; Sun, L.; Sun, Y. Physics-informed inhomogeneous wear identification of end mills by online monitoring data. J. Manuf. Process. 2024, 132, 759–771. [Google Scholar] [CrossRef]
  38. Li, T.; Zhou, Z.; Li, S.; Sun, C.; Yan, R.; Chen, X. The emerging graph neural networks for intelligent fault diagnostics and prognostics: A guideline and a benchmark study. Mech. Syst. Signal Process. 2022, 168, 108653. [Google Scholar] [CrossRef]
Figure 1. Overall methodology: ➀ Cut signal feature extraction. ➁ Graph-structured data construction. ➂ Use of unlabeled data for ULDTM training. ➃ Parameter transfer. ➄ LDTM outputs the predicted tool RUL of labeled data, followed by training and validating.
Figure 1. Overall methodology: ➀ Cut signal feature extraction. ➁ Graph-structured data construction. ➂ Use of unlabeled data for ULDTM training. ➃ Parameter transfer. ➄ LDTM outputs the predicted tool RUL of labeled data, followed by training and validating.
Sensors 25 04068 g001
Figure 2. Graph-structured data construction.
Figure 2. Graph-structured data construction.
Sensors 25 04068 g002
Figure 3. Structure and data processing of GNN.
Figure 3. Structure and data processing of GNN.
Sensors 25 04068 g003
Figure 4. Structure and training process of ULDTM.
Figure 4. Structure and training process of ULDTM.
Sensors 25 04068 g004
Figure 5. Structure and training and validation processes of LDTM.
Figure 5. Structure and training and validation processes of LDTM.
Sensors 25 04068 g005
Figure 6. Comparison between the two models and transfer learning.
Figure 6. Comparison between the two models and transfer learning.
Sensors 25 04068 g006
Figure 7. Schematic diagram of the PHM2010 experiment.
Figure 7. Schematic diagram of the PHM2010 experiment.
Sensors 25 04068 g007
Figure 8. Graph-structured data for each cut in PHM2010.
Figure 8. Graph-structured data for each cut in PHM2010.
Sensors 25 04068 g008
Figure 9. The monotonicity loss output by ULDTM for C2, C3, C5.
Figure 9. The monotonicity loss output by ULDTM for C2, C3, C5.
Sensors 25 04068 g009
Figure 10. The RMSE of training phase without/with unlabeled data enhancement (no unlabeled data enhancement and enhancement from unlabeled dataset C2, C3, or C5).
Figure 10. The RMSE of training phase without/with unlabeled data enhancement (no unlabeled data enhancement and enhancement from unlabeled dataset C2, C3, or C5).
Sensors 25 04068 g010
Figure 11. The RMSE of validation phase without/with unlabeled dataset enhancement (no unlabeled data enhancement and enhancement from unlabeled dataset C2, C3, or C5).
Figure 11. The RMSE of validation phase without/with unlabeled dataset enhancement (no unlabeled data enhancement and enhancement from unlabeled dataset C2, C3, or C5).
Sensors 25 04068 g011
Figure 12. Comparison of the real tool RUL (labels), the tool RUL prediction without/with transfer learning (non-transfer, C2-transfer, etc.), and the absolute error between labels and tool RUL prediction on the validation dataset C6.
Figure 12. Comparison of the real tool RUL (labels), the tool RUL prediction without/with transfer learning (non-transfer, C2-transfer, etc.), and the absolute error between labels and tool RUL prediction on the validation dataset C6.
Sensors 25 04068 g012
Figure 13. The RMSE, MAE, and MAPE of tool RUL prediction for models using C1 + C4 as the training dataset and C6 as the validation dataset. (The indicators are calculated based on the labels and predicted tool RUL of C6. As is shown in the figure, the indicators are smaller with unlabeled data enhancement across all these models).
Figure 13. The RMSE, MAE, and MAPE of tool RUL prediction for models using C1 + C4 as the training dataset and C6 as the validation dataset. (The indicators are calculated based on the labels and predicted tool RUL of C6. As is shown in the figure, the indicators are smaller with unlabeled data enhancement across all these models).
Sensors 25 04068 g013
Figure 14. Comparison of the real tool RUL (labels), the tool RUL prediction without/with transfer learning (non-transfer, transfer), and the absolute error between labels and tool RUL prediction on the validation dataset T_3.
Figure 14. Comparison of the real tool RUL (labels), the tool RUL prediction without/with transfer learning (non-transfer, transfer), and the absolute error between labels and tool RUL prediction on the validation dataset T_3.
Sensors 25 04068 g014
Figure 15. The RMSE, MAE, and MAPE of tool RUL prediction for models using T_2 as the training dataset and T_3 as the validation dataset. (The indicators are calculated based on the labels and predicted tool RUL of T_3. As is shown in the figure, the indicators are smaller with unlabeled data enhancement across all these models).
Figure 15. The RMSE, MAE, and MAPE of tool RUL prediction for models using T_2 as the training dataset and T_3 as the validation dataset. (The indicators are calculated based on the labels and predicted tool RUL of T_3. As is shown in the figure, the indicators are smaller with unlabeled data enhancement across all these models).
Sensors 25 04068 g015
Table 1. Time-domain characteristics.
Table 1. Time-domain characteristics.
Time-Domain CharacteristicsFormula
Peak value y 1 = max ( | x ( i ) | )
Form factor y 2 = 1 N i = 1 N x ( i ) 2 1 N i = 1 N | x ( i ) |
Clearance factor y 3 = y 1 1 N i = 1 N | x ( i ) | 2
Shape factor y 4 = max ( x ( i ) ) min ( x ( i ) ) 1 N i = 1 N x ( i ) 2
Table 2. Frequency-domain characteristics.
Table 2. Frequency-domain characteristics.
Frequency-Domain CharacteristicsFormula
Mean value F 1 = 1 K k = 1 K s ( k )
Centroid frequency F 2 = k = 1 K s ( k ) f k k = 1 K s ( k )
Variance of frequency F 3 = k = 1 K s ( k ) ( f k F 2 ) 2 K
Frequency spread F 4 = k = 1 K s ( k ) | f k F 2 | K F 3
Fourth moment of frequency F 5 = k = 1 K s ( k ) f k 4 k = 1 K s ( k ) f k 2
Form factor F 6 = k = 1 K s ( k ) f k 2 k = 1 K s ( k ) k = 1 K s ( k ) f k 4
Coefficient of skewness F 7 = F 3 F 2
Third moment of the mean F 8 = k = 1 K s ( k ) ( f k F 2 ) 3 K F 3 3
Fourth moment of the mean F 9 = k = 1 K s ( k ) ( f k F 2 ) 4 K F 3 4
Energy concentration factor F 10 = k = 1 K s ( k ) f k 2 2 k = 1 K s ( k ) f k 4
Table 3. Structure of PHM2010 dataset.
Table 3. Structure of PHM2010 dataset.
UnlabeledLabeled
Experiment numberC2C3C5C1C4C6
Number of cuts315315315315315315
Number of sensors777777
Number of signals777777
Table 4. Structure of PHM2010 training/validation dataset.
Table 4. Structure of PHM2010 training/validation dataset.
PHM2010 DatasetChosen SamplesDescription
C2289ULDTM training dataset
C3238ULDTM training dataset
C5238ULDTM training dataset
C1306LDTM training dataset_1
C4278LDTM training dataset_2
C6238LDTM validation dataset
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Guo, D.; Zhou, H.; Sun, L.; Li, G. Unlabeled-Data-Enhanced Tool Remaining Useful Life Prediction Based on Graph Neural Network. Sensors 2025, 25, 4068. https://doi.org/10.3390/s25134068

AMA Style

Guo D, Zhou H, Sun L, Li G. Unlabeled-Data-Enhanced Tool Remaining Useful Life Prediction Based on Graph Neural Network. Sensors. 2025; 25(13):4068. https://doi.org/10.3390/s25134068

Chicago/Turabian Style

Guo, Dingli, Honggen Zhou, Li Sun, and Guochao Li. 2025. "Unlabeled-Data-Enhanced Tool Remaining Useful Life Prediction Based on Graph Neural Network" Sensors 25, no. 13: 4068. https://doi.org/10.3390/s25134068

APA Style

Guo, D., Zhou, H., Sun, L., & Li, G. (2025). Unlabeled-Data-Enhanced Tool Remaining Useful Life Prediction Based on Graph Neural Network. Sensors, 25(13), 4068. https://doi.org/10.3390/s25134068

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop