1. Introduction
CT scanners are important medical imaging devices that provide three-dimensional images of internal body tissues through X-ray tomography, which play a key role in disease screening, localization, and diagnosis [
1]. However, these critical and expensive medical assets face significant operational challenges, with X-ray tube failures representing the most costly and disruptive maintenance issue. Unexpected tube failures can result in equipment downtime lasting days to weeks, directly impacting patient care delivery and hospital operations. Traditional reactive and preventive maintenance approaches are inadequate for managing such critical equipment, motivating the development of intelligent predictive maintenance solutions specifically designed for medical environments. Currently, healthcare institutions are increasingly focusing on the application of predictive maintenance (PdM) strategies for medical equipment. CT scanners frequently experience downtime due to sudden failures of components such as X-ray tubes, which not only affect the normal healthcare operations of hospitals but also delay timely diagnosis for patients [
2]. Traditional preventive maintenance (PvM) strategies have limited ability to handle random sudden failures [
3]. Therefore, accurately predicting the remaining useful life of CT tubes helps hospitals develop reasonable tube replacement plans and further reduce the risk of unexpected CT equipment downtime [
4].
Predictive maintenance (PdM) has been widely applied in modern industrial production [
5], especially in predicting equipment’s remaining useful life (RUL) in engineering, mechanical, and automation applications [
6,
7]. Equipment maintenance spans the entire lifecycle of the equipment, and a good maintenance strategy can significantly extend equipment lifespan and reduce failure rates [
8,
9]. PdM, based on historical data collected during equipment operation, utilizes advanced data analysis and machine learning technologies to achieve the precise prediction of equipment RUL, providing a scientific basis for equipment maintenance decisions [
10,
11].
Compared to traditional physics-based methods, data-driven degradation prediction methods can automatically learn equipment degradation patterns and rules from large-scale heterogeneous data without an in-depth understanding of complex failure mechanisms, thus possessing stronger universality and adaptability [
12,
13]. With the rapid development of sensor technology, IoT, and big data processing capabilities, the acquisition and storage of large amounts of operational data have become easier, providing a solid foundation for data-driven degradation prediction methods [
14,
15].
Currently, data-driven degradation prediction methods can be categorized into three main types: statistical learning methods, machine learning methods, and deep learning methods [
16,
17,
18]. Statistical learning methods model the degradation process through probability and statistics theory [
19], such as hidden Markov models, Weibull distribution models, Gamma processes [
20], Wiener processes [
21], and random filtering [
22,
23]. However, these methods rely on prior assumptions about the degradation process and have limitations in handling complex non-linear relationships, making them difficult to adapt to the multi-stage variation characteristics in CT tube life prediction. Machine learning methods such as support vector machines and random forests extract features from historical data and learn degradation patterns [
24]. Wei et al. [
25] proposed a method combining support vector regression (SVR) with particle filters to predict the RUL of lithium-ion batteries. This method could improve prediction accuracy to some extent but still struggles to capture complex temporal dependencies when processing long sequence data, making it difficult to learn features directly from raw data. Deep learning methods such as long short-term memory (LSTM) networks and convolutional neural networks can automatically learn complex temporal features from raw data, achieving end-to-end degradation prediction [
26]. For example, Ciani et al. [
27] used LSTM and a single exponential degradation model to accurately estimate the RUL of lithium batteries in wireless sensor networks based on a limited number of data records. Ma et al. [
28] combined the CNN with feature extraction capability and LSTM with recurrent memory capability to carry out bearing degradation trend prediction. Deep learning methods, such as LSTM networks and convolutional neural networks, can automatically learn complex temporal features from raw data, achieving end-to-end degradation prediction. However, single deep learning models have limitations in CT tube life prediction. For example, LSTM may encounter gradient vanishing or explosion problems with long sequences, while CNN lacks the ability to model temporal dependencies when processing time series data.
Despite significant progress in data-driven degradation prediction methods, the development of a prediction method suitable for X-ray tubes of CT equipment still faces enormous challenges [
29]. On the one hand, there are evident issues of data non-linearity and multi-phase characteristics for the collected CT tube data. The tube lifespan data exhibits complex non-linear degradation patterns and shows different degradation features at different usage stages. There are significant differences in data distribution between early stages, mid-use periods, and near-failure phases, making it difficult for a traditional model to capture the variation patterns throughout the entire lifecycle. Additionally, factors such as the uncertainty of the various working environments, parameter fluctuations, and load changes further increase the complexity of data modeling, leading to insufficient prediction accuracy of traditional or single-stage models in practical applications. On the other hand, from the perspective of prediction model construction, traditional static model structures cannot dynamically adjust prediction strategies based on different data characteristics, lacking adaptive mechanisms to handle different types of temporal patterns [
30]. The architecture of a single model is also unable to process different types of temporal patterns simultaneously. For example, a pure LSTM model encounters gradient vanishing or exploding problems with long sequences, making it difficult to capture long-term dependencies. A pure temporal convolutional network (TCN) model, although capable of handling long sequences, lacks the ability to process sequential and state information. In addition, simple loss functions result in the model being insensitive to state transitions at turning points in the data, and the use of standard MSE or MAE loss functions fails to distinguish the importance of different types of samples. Fixed-weight loss functions also cannot adapt to the requirements of different training stages [
31,
32].
To address the above issues, this paper proposes a CT tube life intelligent prediction architecture based on a dual-branch neural network (DDLNet). Firstly, an adaptive phase-aware framework is applied to identify the different phases of tube, enabling the model to adaptively adjust prediction strategies according to the current phase, which can address the problem of insufficient model adaptability. Secondly, a dual-branch collaborative architecture is applied to enhance the modeling capability of long-term dependencies from temporal data, effectively capturing non-linear changes in the degradation phase and overcoming the limitations of a single model. Finally, a dynamically weighted triplet mixed loss function is designed to adaptively adjust the weight ratios of different prediction tasks, solving the problems of sample imbalance and uneven prediction accuracy.
Moreover, the evaluation methodology presents another critical challenge in CT tube RUL prediction research. Due to the limited availability of complete lifecycle data from medical equipment and the substantial costs associated with CT tube replacement, most studies rely on small datasets. This necessitates robust evaluation strategies that can provide reliable performance assessment despite limited sample sizes. Traditional train–test splits may lead to optimistic performance estimates when dealing with such constrained datasets, making it essential to employ rigorous cross-validation methodologies that ensure model generalization across different equipment and operating conditions.
Despite these advances, several critical research gaps remain in CT tube RUL prediction. First, existing single-model approaches cannot effectively capture the multi-phase degradation characteristics of CT tubes, as they lack adaptive mechanisms to handle the distinct patterns between stable and degradation phases. Second, current methods fail to synergistically combine long-term dependency modeling with multi-scale temporal feature extraction, limiting their effectiveness in complex industrial environments. Third, traditional loss functions do not adequately address the sample imbalance issues inherent in CT tube datasets, where stable phase data significantly outnumber degradation phase samples.
To better clarify our contributions, the key innovations of this work include the following.
(1) A dual-branch neural network architecture that integrates RSA-BiLSTM with D-TCN for CT tube RUL prediction. The RSA-BiLSTM branch uses a CNN-QKV generator to transform input sequences into query, key, and value representations and employs a forward-looking attention mechanism where query vectors at time step t + 1 interact with key vectors at time step t to capture state transitions. The D-TCN branch uses dilated convolutions with dilation rates of 1, 2, 4, and 8 to extract multi-scale temporal features. This dual-branch design addresses gradient vanishing problems in single LSTM models and temporal dependency limitations of pure CNN models.
(2) An adaptive phase-aware framework with dynamic stage detection for binary CT tube degradation states. The framework identifies the stable phase (≥30 days) and the degradation phase (<30 days) to enable adaptive prediction strategies across different operational states.
(3) A dynamically weighted triplet mixed loss function combining base value loss, phase classification loss, and smoothness loss. The dynamic weighting mechanism adjusts each component’s contribution based on real-time metrics, including prediction error, change point density, and prediction smoothness. This automatically adapts the loss function emphasis during training to handle sample imbalance and improve prediction accuracy across different RUL ranges.
The remainder of this paper is organized as follows:
Section 2 introduces the proposed method and related knowledge,
Section 3 introduces the experimental dataset sources and data processing.
Section 4 demonstrates the superiority and effectiveness of the proposed method through comparison with existing methods and ablation experiments.
Section 5 concludes the paper.
2. Experiment Dataset and Preprocessing Step
2.1. Data Description
The Internet of Medical Things (IoMT) system achieves the collection, processing, and analysis of medical data through a multi-layer architecture, including sensor layer, transmission layer, and application layer [
33,
34]. The sensor layer consists of various sensors monitoring patient physiological data and medical device status. The transmission layer utilizes wireless communication protocols to transmit data to central servers. The application layer is responsible for data storage, analysis, and ensuring secure access.
All data used in this research comes from the IoMT system of West China Hospital. The research team selected 6 X-ray tubes with complete lifecycles from 6 CT devices for RUL prediction research.
Table 1 shows the specific information of 6 X-ray tubes, including CT number, tube number, start time, end time, and data volume.
The collected data includes a total of 45 parameters, categorized into device identification, time information, device control parameters, interruption information, and lifespan-related parameters. These parameters reflect various aspects of CT tube operation and degradation, including scan counts that indicate cumulative usage, voltage fluctuations that affect tube stress, temperature measurements that reflect thermal conditions, and arc events that signal electrical anomalies. Data sampling frequency is not fixed, with sampling intervals ranging from 20 to 60 s. Among these, the device control parameters include 5 variables: “focus”, “kind”, “mode”, “region”, and “dom_type”.
The experiment primarily uses the control parameters of 6 tubes. “Focus” refers to the position where the electron beam hits the target material, with the values “3” and “1” representing large focus and small focus, respectively. The data shows that small focus is used more frequently, indicating a greater demand for high-resolution imaging in clinical applications. Scan types (“kind”) mainly include the SEQ, SPI, TOP, and ROT modes. Exposure modes (“mode”) primarily include “A”, “AB”, and “B”, with this research mainly using the “A” mode. Scanning regions (“region”) are divided into “head” and “body”, with the data showing that body scans are predominant, which aligns with clinical practice as body scans encompass chest, abdomen, and pelvic examinations. “Dom_type” represents the tube current dynamic adjustment mode, including “AEC” (360° rotation tube current mode), “ZEC” (Z-direction automatic tube current adjustment mode), and “No” (standard tube current mode without adjustment). These detailed parameter data provide a solid foundation for studying the performance degradation characteristics of CT X-ray tubes and predicting their remaining useful life.
The CT tube lifespan data in the dataset is divided into two phases: the stable phase and the degradation phase. The stable phase is characterized by a lifespan value of 30 days or more, indicating that the equipment is operating normally. The degradation phase is marked by a lifespan value of less than 30 days, indicating that the performance of the equipment is declining and approaching failure. This binary phase division reflects the characteristic changes of CT tubes at different usage stages and provides a basis for the model to distinguish between different working states, thereby enhancing prediction accuracy and adaptability.
2.2. Preprocessing Step
The initial phase requires collecting CT operation log data through the medical IoMT system. Our dataset contains complete lifecycle data for 6 CT X-ray tubes, with each tube generating approximately 45 parameters. These parameters can be divided into three main categories: 8 parameters related to device identification and time information, 7 parameters related to device control and operational interruptions, and 30 parameters directly related to tube lifespan characteristics.
For effective analysis, we first apply min–max normalization to 30 lifespan-related parameters to address the issue of varying measurement units and scales between different parameters, which can not only improve the prediction performance of the model but also accelerate convergence during the training process. The
min–max normalization is calculated as follows:
After normalization, feature engineering is performed to derive new features from the normalized parameters. This process includes creating interaction terms between existing features through basic arithmetic operations (such as differences, sums, products, and ratios). These derived features capture complex relationships that might not be evident in the original parameters, effectively revealing hidden patterns in the data and enhancing the ability of the model to recognize non-linear relationships.
To determine the most relevant parameters for predicting RUL, this research employs Pearson correlation analysis. This technique evaluates the linear relationship between each feature and the target variable. The Pearson correlation coefficient ranges from −1 to 1, with values closer to either extreme indicating stronger relationships. The Pearson correlation coefficient is defined as
This research sets a correlation threshold of 0.75, selecting only those features with absolute correlation coefficients exceeding this value. The Pearson correlation method provides computational efficiency and direct interpretability, allowing the research to quickly identify the most important predictors without involving the complexity or potential instability of other techniques. The correlation analysis reveals several highly correlated features, including parameters related to scan counts, voltage fluctuations, temperature measurements, and arc events.
Finally, we apply sliding window processing to the screened features and data, but original window data often contains significant fluctuations that may mask potential degradation trends. To address this issue, two key processing techniques are applied.
Downsampling: By selecting points at regular intervals (sampling interval T = 20), data density is reduced, which helps smooth short-term fluctuations.
Exponential Moving Average (EMA): Applying EMA smoothing operations further reduces noise and highlights gradual degradation patterns.
4. Case Study
4.1. Evaluation Metrics
To evaluate the performance of the RUL model, this research selects three key evaluation metrics, i.e., Mean Squared Error (MSE), Mean Absolute Error (MAE), and Coefficient of Determination (R2). These metrics reflect the prediction accuracy of the model from different perspectives, as follows:
- 1.
Mean Squared Error (MSE): MSE is particularly suitable for RUL prediction as it heavily penalizes large prediction errors, which are critical in maintenance planning. In CT tube maintenance, a significant prediction error (e.g., predicting 20 days when actual RUL is 5 days) could lead to equipment failure and patient care disruption.
MSE is calculated as Equation (45):
where
is the actual value,
is the predicted value, and
is the sample size.
MSE evaluates the performance of prediction models through squared errors, which is sensitive to large errors.
- 2.
Mean Absolute Error (MAE): MAE provides an intuitive understanding of average prediction accuracy in the same units as the target variable (days). This is directly interpretable by maintenance personnel who need to understand the typical prediction uncertainty when planning maintenance schedules. Unlike MSE, MAE is less sensitive to outliers, providing a robust measure of typical model performance. MAE calculated as
- 3.
Coefficient of Determination (R
2): R
2 indicates how well the model explains the variance in RUL values, which is crucial for assessing model reliability across different CT tubes and operating conditions. In medical equipment maintenance, understanding the proportion of RUL variance that can be explained by the model helps in risk assessment and decision-making confidence.
R2 is calculated as
where
is the mean of the actual values. The
R2 value typically ranges between 0 and 1, with values closer to 1 indicating better model fit, meaning the model can explain more variance in the dependent variable. These three metrics provide complementary perspectives: MSE emphasizes accuracy in critical situations, MAE provides practical interpretability, and R
2 assesses overall model reliability.
To ensure robust evaluation and avoid overfitting, this study employs a LOOCV strategy. Given the limited number of CT tubes (six tubes), LOOCV provides the most comprehensive evaluation by using each tube as a test set while training on the remaining five tubes. This approach ensures that the model’s performance is evaluated on completely unseen data from different CT scanners, providing a more realistic assessment of the model’s generalization capability across different equipment and operating conditions. While LOOCV provides rigorous evaluation within our dataset, it cannot address the fundamental limitation of domain diversity. The claims of generalizability across different healthcare institutions and CT equipment types remain insufficiently substantiated.
4.2. Comparative Experiments
The experiments are conducted on a system with the following configuration: a Windows 10 (22H2) operating system, equipped with an Intel(R) i7-9750H processor and an Nvidia RTX2060 graphics card. The software environment included Python 3.12.1 and PyTorch 2.3.1 for implementing and training the neural network models. All experimental results reported in this study are obtained using LOOCV. For each iteration, one tube dataset is held out as the test set, while the remaining five tube datasets are used for training and validation. This process is repeated six times, ensuring each tube serves as the test set exactly once. The reported metrics (MSE, MAE, and R2) represent the average performance across all six cross-validation folds. To evaluate the performance of DDLNet in the CT tube RUL task, this research compared it with five network models, i.e., LSTM, Transformer, CNN-BiLSTM-Attention, CNN-LSTM, and CNN-LSTM-Attention.
Table 2 shows the MSE, MAE, and R
2 metric values of each model on six different CT tube datasets (Tube 1–6). From the experimental data, DDLNet obtains the best performance on the MSE, MAE, and R
2 metrics. Notably, on Tube 5, other models generally have low R
2 values (such as LSTM at 0.0018, CNN-BiLSTM-Attention at 0.0643), while DDLNet reaches a 0.8576 R
2 value. On all datasets, the R
2 of DDLNet remained above 0.62, while other models had multiple instances below 0.25. The inclusion of standard deviations reveals that DDLNet demonstrates not only superior mean performance but also the lowest variability across different CT tubes (std: 1.39 for MSE, 0.18 for MAE, 0.10 for R
2), indicating more consistent and reliable predictions compared to baseline methods. Through comparative analysis, DDLNet showed good performance on all three evaluation metrics, especially when handling complex datasets such as Tube 4 and Tube 5, maintaining relatively stable prediction performance with lower MSE and MAE and higher R
2.
To visually demonstrate the performance of the DDLNet model in the CT tube RUL prediction,
Figure 2 shows the prediction results, actual label values, absolute errors, and relative errors on six test datasets. The red line represents the predicted values, the orange line represents the actual label values, the green bars represent absolute errors, and the blue dashed line represents relative errors. From the visualization results, it can be observed that the prediction values can generally track the changes in the actual RUL values. At most time points, the differences between predicted and actual values are small, with absolute errors generally kept at low levels. The model shows high prediction accuracy in the mid-life stage of CT tubes (RUL values between 10–25). It is worth noting that in all six datasets, when the RUL value is low (i.e., when the tube is approaching the end of its life), the relative error increases. This is mainly because the RUL value itself is small, so even small absolute errors can lead to large relative error percentages. When the RUL approaches 5 or below, increased fluctuations in relative error can be observed, occasionally exceeding 200% relative error peaks. Additionally, in the initial phase of the time series (RUL around 30), the fit between the predicted values of the model and actual values is generally good, with consistently small absolute errors. As time progresses, the prediction curve can capture the downward trend of actual RUL values. Overall, the DDLNet model demonstrates good prediction capability across all six test datasets, effectively tracking the changing trends in CT tubes’ remaining life.
To further visually compare the differences in prediction capabilities between DDLNet and comparison methods,
Figure 3 shows the comparison of prediction results and actual label values for various models on six test datasets. Overall, the DDLNet model achieved significant advantages in MSE, MAE, and R
2 evaluation metrics, particularly showing higher prediction accuracy and stability when handling complex datasets. Tube 4 presents a special case where all comparative models show obvious prediction stability issues in the high RUL region (20–30), with predicted values exhibiting large fluctuations around the actual values. While the DDLNet model also shows some fluctuations, it can better reflect the downward trend of RUL; especially in the region where RUL is below 15, its prediction accuracy is significantly better than those of other models. This indicates that the DDLNet model can effectively address the challenges posed by the multi-phase variation characteristics of CT tubes, providing reliable CT tube remaining life predictions for medical institutions and helping to formulate scientific maintenance plans.
4.3. Ablation Study
To further verify the contribution of each component of DDLNet, this research designed a series of ablation experiments using the same LOOCV strategy.
Figure 4 shows the comparison of prediction performance between the complete DDLNet model (marked as “Ours”) and variant models with specific components removed or modified on six test datasets. This experiment mainly examined the following variants:
No pre-training model (No_Pretraining): Removed the pre-training phase, directly training with randomly initialized model parameters to verify the impact of the pre-training strategy on model performance.
Residual self-attention model (RSA): Only retained the RSA-BiLSTM branch of DDLNet to evaluate the effectiveness of the RSA-BiLSTM branch acting alone.
Simplified loss function model (Simplified_Loss): Used standard mean squared error (MSE) to replace the composite loss function proposed in this paper, removing the relative error term to verify the impact of loss function design on prediction accuracy.
Temporal convolutional network model (TCN): Only retained the TCN branch to evaluate the effectiveness of the TCN branch acting alone.
According to the data in
Table 3, there are significant performance differences between the variant models and the complete DDLNet. Regarding the pre-training strategy, the No_Pretraining variant has an average MSE of 4.5034, higher than DDLNet’s 2.8016, and its average R
2 is 0.6479, lower than DDLNet’s 0.7714. Specifically for each dataset, on Tube 3, the MSE of the No_Pretraining model is 3.8784, 71.9% higher than DDLNet (2.2568), and on Tube 4, the MSE of the No_Pretraining model is 3.6361, 195.1% higher than DDLNet (1.2323). However, on Tube 5, the No_Pretraining model’s R
2 reached 0.7980, close to DDLNet’s 0.8576, indicating that the impact of the pre-training strategy varies across different datasets.
Regarding the contribution of the RSA-BiLSTM branch, the variant model that only retained this branch showed inconsistent performance across different datasets. The RSA variant has an average MSE of 6.5116 and an average MAE of 0.7711, higher than the complete DDLNet. On Tube 2, the RSA variant’s R2 value reaches 0.6857, close to DDLNet, but on Tube 1, the R2 value is only 0.4081, 40.6% lower than DDLNet’s 0.6866. On Tube 5, the RSA variant’s MAE is 1.5266, 118.4% higher than DDLNet’s 0.6989, indicating that using the RSA-BiLSTM branch alone has limited adaptability to different feature patterns.
The loss function design also has a significant impact on prediction performance. The Simplified_Loss variant has an average MSE of 5.3374 and an average MAE of 0.6119, 90.5% and 53.8% higher than DDLNet, respectively. On Tube 1 and Tube 6, the Simplified_Loss model’s MSE is 8.0198 and 6.8265, 86.3% and 73.8% higher than DDLNet, respectively. Notably, on Tube 5, the Simplified_Loss model’s R2 reaches 0.7651, indicating that standard MSE loss can also achieve good results on certain datasets.
Regarding the contribution of the TCN branch, the variant model that only retained the TCN branch has an average MSE of 5.0932 and an average MAE of 0.5983. Its performance is also inconsistent across datasets. For instance, on Tube 5, the TCN variant’s R2 is 0.7813, close to DDLNet, while on Tube 1, R2 is only 0.4341, significantly lower than DDLNet’s 0.6866. The MSE of the TCN variant on Tube 1 is 7.7832, 80.6% higher than DDLNet, indicating that the capability of the TCN structure varies when processing different temporal features.
Figure 4 shows the comparison of prediction results between the complete DDLNet model and various variant models on the six test datasets. Through visual analysis, it can be seen that the pre-training strategy, RSA-BiLSTM branch, TCN branch, and composite loss function each play different roles in enhancing model performance. The pre-training strategy (as shown in tubes 1 and 4) provides better parameter initialization, significantly reducing prediction fluctuations, with an average MSE reduction of 60.7%. The RSA-BiLSTM branch (Tube 2) excels at capturing global dependencies in temporal data with its residual self-attention mechanism, achieving an R
2 of 0.6857 on Tube 2. The TCN branch (Tubes 5 and 6) provides multi-scale feature extraction capability through multi-layer dilated convolutions, performing well in handling smooth change regions, with an R
2 of 0.7813 on Tube 5. The composite loss function (tubes 3 and 5) improves the sensitivity of the model to low RUL value regions by combining absolute and relative errors, reducing MSE by 86.3% and 73.8% on Tube 1 and Tube 6, respectively.
To further validate the rationality of our loss function design, we conducted a comprehensive sensitivity analysis on the dynamic weighting mechanism of the triplet mixed loss function. Using leave-one-out cross-validation across all six CT tube datasets, we systematically evaluated the impact of weight variations on model performance. Our experimental design included the following.
(1) Individual parameter sensitivity testing by varying each weight parameter while keeping others constant;
(2) Parameter interaction analysis through simultaneous variations;
(3) Extreme boundary testing to assess model robustness limits.
As demonstrated in
Table 4, our empirically selected configuration (α_base = 0.8, β_base = 0.15, γ_base = 0.05) achieves optimal performance across all evaluation metrics. The individual parameter sensitivity analysis reveals distinct characteristics: α_base exhibits the highest sensitivity, with performance degrading significantly when deviating from 0.8 (MSE increases from 2.92 to 3.45 when α_base = 0.6, representing an 18.2% deterioration). β_base shows moderate sensitivity, with MSE increasing to 3.28 and 3.19 when changed to 0.10 and 0.20, respectively. γ_base demonstrates the least individual impact, though still meaningful, with MSE rising to 3.12 and 3.25 for values of 0.03 and 0.07. The parameter interaction analysis (rows 10–11) shows that simultaneous moderate changes can maintain reasonable performance, while extreme combinations (rows 12–13) result in substantial performance degradation, with MSE increasing by up to 33.2%. This systematic analysis validates that our weight selection represents a well-optimized configuration that effectively balances the contributions of base value loss, phase classification loss, and smoothness regularization.
The collaborative effect of these components enables DDLNet to demonstrate excellent prediction accuracy and stability across datasets with different characteristics, confirming the effectiveness and necessity of the multi-component collaborative design.
4.4. Discussion
While our experimental results demonstrate the effectiveness of DDLNet on real CT tube data, several important directions warrant further investigation for practical deployment:
- (1)
Domain Adaptation and Transferability
Future work should focus on enhancing the model’s adaptability across different CT scanner manufacturers, tube types, and operational environments. The development of domain adaptation techniques could enable our approach to generalize effectively to new hospital settings with minimal retraining, addressing the inherent variability in medical equipment configurations and usage patterns.
- (2)
Practical System Integration and Deployment
Translating our research findings into operational clinical systems requires comprehensive engineering efforts. This includes developing robust data pipelines for continuous parameter collection, implementing real-time prediction algorithms with appropriate failover mechanisms, establishing integration protocols with existing hospital information systems, and conducting extensive validation studies across multiple institutions to ensure reliability and safety standards.
- (3)
Model Interpretability and Clinical Trust
A critical aspect for medical equipment applications is enhancing model interpretability to build clinical trust and support regulatory approval. Future research should explore explainable AI techniques that can provide clear reasoning for RUL predictions, develop uncertainty quantification methods to communicate prediction confidence, and establish clinical decision support frameworks that integrate predictions with existing maintenance protocols while maintaining human oversight and accountability.