Skip to Content
  • Article
  • Open Access

5 October 2025

Med-Diffusion: Diffusion Model-Based Imputation of Multimodal Sensor Data for Surgical Patients

,
,
,
,
,
,
,
and
1
Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen 518055, China
2
University of Chinese Academy of Sciences, Beijing 100049, China
3
Department of Mathematics, University of Shanghai for Science and Technology, 334 Campus, Shanghai 200093, China
*
Authors to whom correspondence should be addressed.
This article belongs to the Section Biomedical Sensors

Abstract

The completeness and integrity of multimodal medical data are critical determinants of surgical success and postoperative recovery. However, because of issues such as poor sensor contact, small vibrations, and device discrepancies during signal acquisition, there are frequent missing values in patients’ medical data. This issue is especially prominent in rare or complex cases, where the inherent complexity and sparsity of multimodal data limit dataset diversity and degrade predictive model performance. As a result, clinicians’ understanding of patient conditions is restricted, and the development of robust algorithms to predict preoperative, intraoperative, and postoperative disease progression is hindered. To address these challenges, we propose Med-Diffusion, a diffusion-based generative framework designed to enhance sensor data by imputing missing multimodal clinical data, including both categorical and numerical variables. The framework integrates one-hot encoding, simulated bit encoding, and feature tokenization to improve adaptability to heterogeneous data types, utilizing conditional diffusion modeling for accurate data completion. Med-Diffusion effectively learns the underlying distributions of multimodal datasets, synthesizing plausible data for incomplete records, and it mitigates the data sparsity caused by poor sensor contact, vibrations, and device discrepancies. Extensive experiments demonstrate that Med-Diffusion accurately reconstructs missing multimodal clinical information and significantly enhances the performance of downstream predictive models.

1. Introduction

Multimodal data from surgical patients, encompassing preoperative, intraoperative, and postoperative information, play a crucial role in ensuring successful surgeries and improving patient prognosis [1,2]. These data offer invaluable insights into patients’ physiological status, potential risk factors, and responses to various surgical and anesthetic interventions and treatments. The effective utilization of cardiovascular data can inform clinical decision making, guide personalized treatment strategies, and optimize resource allocation in healthcare settings [3].
However, the acquisition of comprehensive cardiovascular data for patient risk assessment often encounters numerous challenges. Due to technical limitations, disease progression, and other factors, some data modalities may be missing or uncollected, which affects the accuracy and reliability of the assessment [4,5,6]. Complete and accurate modality data is vital for constructing effective predictive models, as missing data can lead to the loss of critical information or introduce bias during the risk evaluation process. Healthcare data imputation not only improves the utilization of existing data—maximizing available resources—but also enhances the efficiency of model training [7]. Risk assessment of patients during surgery usually requires the integration of multiple modalities to gain a comprehensive understanding of a patient’s health condition. Missing modality data can limit the model’s ability to fully analyze the patient’s status [8]. There are imputation techniques that generate missing modality data and align them with existing data ensure data integrity, thereby improving the reliability and accuracy of predictive models.
By exploring methods for generating and aligning missing data, we can effectively impute the missing data and integrate it with other modalities [9], facilitating the synergistic use of multimodal data and enhancing the comprehensiveness and accuracy of the evaluation.
Data augmentation is a powerful technique to address the issue of data scarcity. Researchers have explored the potential of medical data augmentation to compensate for the limited datasets of surgical patients [10]. Recent advancements in generative models have shown promising results in generating high-quality synthetic data, such as MissForest Random Forest Imputation [11] and GAINs (Generative Adversarial Imputation Networks) [12], which can significantly augment existing datasets [13]. Specifically, diffusion models have been proven to outperform Generative Adversarial Networks (GANs) in image generation tasks, offering advantages, such as broader distribution coverage, stable training objectives, and strong scalability [14,15]. The application of generative models in medical data augmentation has increasingly drawn the attention of researchers [16]. By leveraging these technologies, healthcare providers can overcome missing data issues and build more accurate and reliable predictive models [17], supporting clinical decision making and improving the care of surgical patients [18]. Integrating diffusion-based augmentation techniques into multimodal medical data workflows enhances data availability and diversity, thereby enabling informed clinical decisions, personalized treatments, and improved patient outcomes [19].
To address the issue of missing data in multimodal medical datasets, we propose Med-Diffusion, a diffusion model-based conditional imputation and data augmentation framework designed to tackle missing and sparse data challenges in multimodal medical data. This method is carefully designed to consider both the complexity of multimodal features and the practical challenges of data scarcity and frequent missing data in medical settings. First, the framework preprocesses tabular data and constructs training missing data based on a Missing Completely at Random (MCAR) mechanism to ensure that the model learns more robust imputation strategies in a controlled environment. At the representation layer, the framework systematically compares three different approaches for handling categorical variables: one-hot encoding, simulated bit encoding, and feature tokenization (FT, feature embedding). The FT approach, which demonstrated superior performance, was ultimately selected. This approach maps both numerical and categorical variables into the same-dimensional embedding space, allowing the entire multimodal data to be fully numerically represented and input into the diffusion model. Subsequently, Med-Diffusion performs a reversible “noising → denoising” process: during the training phase, the model learns a conditional denoising network on complete data using the observed sub-vectors as conditions and minimizing the mean squared error (MSE) of the noise prediction to approximate the true distribution. During the inference phase, the model only performs stepwise denoising on the missing dimensions of incomplete samples, while keeping the observed dimensions unchanged. The imputation result is then obtained through the decoding strategy corresponding to one-hot encoding, simulated bit encoding, or FT. This process allows the framework to handle both numerical variables (such as heart rate, systolic/diastolic blood pressure, temperature, and respiratory rate) and categorical variables (such as cancer stages I–IV and smoking status at light/moderate/severe), ensuring the imputation’s plausibility while enhancing data diversity and utilization. Compared to traditional methods, Med-Diffusion not only provides a more comprehensive and flexible solution, but also demonstrates clear advantages in terms of training stability, distribution coverage, and the quality of generated samples. Experimental results, assessed using RMSE and error rates, validate the framework’s effectiveness and its competitiveness relative to state-of-the-art methods.
Our primary contributions are as follows. The experimental results are shown in Figure 1.
Figure 1. To address the issue of missing data arising from the heterogeneity of sensors and medical devices, we propose Med-Diffusion. The red box indicates the comparison table of data before and after imputation, and the blue box represents the comparison line charts of data imputation across different examples. The imputation process and the corresponding results obtained by Med-Diffusion are illustrated in Figure 1. Med-Diffusion effectively learns the underlying distributions of multimodal datasets and synthesizes plausible data for incomplete records, thus improving overall data quality and utility.
  • We propose a novel diffusion model-based multimodal medical data imputation method, Med-Diffusion, and implement it for the tabular imputation of static medical sequences.
  • We propose a comprehensive pipeline for effectively training Med-Diffusion on categorical features, utilizing advanced encoding and embedding techniques.
  • Our experimental results show that Med-Diffusion reduces error rates and the RMSE (Root Mean Square Error) by 4.4% and 5.1%, respectively, compared to existing probabilistic methods for healthcare and environmental data.

3. Methodology

We propose a multimodal medical data augmentation method based on diffusion models. The core idea of this approach is to treat the process of data generation as a gradual transition from data distribution to a Gaussian distribution. This process can be conceptualized as the data progressively losing its structural characteristics and eventually becoming disordered, analogous to the diffusion phenomenon in physics. During the generation of new samples, the model must reverse this process, gradually restoring the ordered data structure from the disordered Gaussian distribution. Experimental results are shown in Table 1.
Table 1. Comparison of the RMSE and error rate performance on three mixed-variable datasets (Diabetes, COVID-19, and Heart Disease indicators). It should be noted that, for datasets without multi-category variables, one-hot encoding and simulated bit encoding are equivalent. The bold numbers represent the state-of-the-art results.

3.1. Overview

For multimodal medical data imputation, the proposed method includes four main stages, executed as follows.
  • Handling Categorical Variables: Categorical variables are processed using one-hot encoding, simulated bit encoding, and embedding encoding methods. Among these, the embedding method (feature tokenization/FT) is found to perform better. Thus, categorical variables are processed using FT in the experimental implementation, and this is reflected in Table 2.
    Table 2. Comparison of the RMSE and error rate performance on three mixed-variable datasets (Diabetes, COVID-19, and Heart Disease indicators). It should be noted that, for datasets without multi-category variables, one-hot encoding and simulated bit encoding are equivalent.
  • Forward Diffusion Process: Both the complete dataset and the incomplete dataset are gradually transformed into a Gaussian distribution by adding noise, which generates a series of progressively noisy states.
  • Reverse Generation Process: A neural network is trained on the complete dataset to model the reverse process. This process enables the model to gradually recover the original data from noisy states, essentially reversing the effects of the forward diffusion process.
  • Imputing Missing Data: For the incomplete dataset, the reverse process, which is trained on the complete dataset, is used to gradually recover the missing data points from the Gaussian noise state. This process combines the reverse diffusion model with the conditional observed values.
Before handling categorical variables, data preprocessing is required. Preprocessing mainly involves distinguishing between categorical and numerical variables, as well as randomly masking some feature entries to construct training and evaluation samples. Specifically, this study employed the Missing Completely at Random (MCAR) mechanism to mask the original data. The mechanism operates as follows: for each row of data, a random subset of the originally observable entries is uniformly selected and masked. Specifically, for each row of data, we randomly select N obs ( col ) × 0.2 entries from the non-missing samples to mask (rounded down). Therefore, unless the dataset is particularly small, the actual missing rate for all features is roughly the same. If a row of data has no missing entries initially, the final missing rate for that row is 20%. If some entries are already missing, then the final missing rate for that row is slightly above 20%.
To facilitate subsequent training and evaluation, we placed all continuous features (numerical variables) at the beginning of the data, while categorical features (categorical variables) were placed at the end. This arrangement does not affect the masking mechanism, as all columns are independently masked in a uniformly random manner. Additionally, the randomness of the entire process is controlled by a fixed random seed to ensure reproducibility of the results.
Moreover, the process treats each row of data the same way, applying the same missing rate to all features of each row, meaning that the missing rate for each feature is approximately the same. To make processing more convenient for subsequent training and evaluation, we place continuous variables (numeric features) in the front and categorical variables (categorical features) in the back. This does not affect the handling of missing values because all data columns independently follow the same MCAR missing pattern. The framework of the Med-diffusion and the training process are shown in Figure 2.
Figure 2. (a) In the rectangle on the middle left, the green and white areas represent observed and missing values, respectively. The observed values are divided into the red imputation target y 0 ta and the blue conditional observed values y 0 co , and they are used to train ϵ θ . The colored regions in each rectangle represent the presence of values. (b) During the training process, the training dataset randomly masks some data as the prediction target data, and multiple training epochs are used to compute and minimize this loss. During inference, the model with the minimum loss is used for predictions, with the actual missing data serving as the prediction target data.

3.2. Categorical Variable Processing

Before processing categorical variables, it is necessary to distinguish between categorical and numerical variables. During the preprocessing stage, we explicitly specify which columns are categorical variables for encoding, while the remaining data is treated as numerical variables, skipping the processing of categorical variables for missing data imputation.
To effectively handle categorical variables, this method explores three techniques (as shown in Figure 3): one-hot encoding, simulated bit encoding, and feature tokenization. In the original data, 45 and 0.7 are numerical variables (represented by blue and green blocks), while 3 is a categorical variable (represented by a yellow block), assuming there are three possible categories.
Figure 3. An example of handling categorical variables using one-hot encoding, simulated bit encoding, and feature tokenization. The two numerical variables in the example are labeled in red and green, while the categorical variable is labeled in blue. It is good practice to explain the significance of the figure in the caption.
a.
In one-hot encoding, 45 and 0.7 remain unchanged, while the categorical variable 3 is converted into three binary bits, [1, 0, 0], indicating the third category is selected (1) and the other two are unselected (0).
b.
In simulated bit encoding, 45 and 0.7 remain unchanged, with each category assigned a unique fixed-length binary code. The categorical variable 3 is encoded as two binary bits, [1, 1], where a more compact representation requires fewer bits.
c.
In processing categorical variables, embedding encoding is used to convert all variables (both numerical and categorical) into embeddings of the same length. E1, E2, and E3 represent the embedded representations of the three original variables. This approach maps all variables to a unified dimensional space. During output processing, the embedding vector is converted back to the original numerical or categorical values. For numerical variables, the final value is obtained by averaging the elements of the embedding vector divided by the output of the diffusion model. For categorical variables, the predicted category is determined by calculating the distance between the embedding vector and the diffusion model output, selecting the nearest embedding vector.

3.3. Diffusion Probabilistic Model for Numerical Imputation

The framework of this study is primarily based on diffusion models, which serve as the mathematical foundation for both categorical and numerical imputation tasks, enabling a seamless transition between noisy and clean data. The following diffusion probabilistic model serves as the mathematical foundation for numerical imputation in this paper. It defines how data progressively transitions from an ordered state to a disordered state, and it also describes the process for generating data in reverse. The key components of the diffusion probabilistic model are as follows (the detailed schematic illustration is provided in Figure 2).

3.3.1. Objective and Evaluation Metric

Let Y = ( R { } ) d represent the input space, where R denotes the real number space and represents missing values. In the missing value imputation task, we are given a d-dimensional training dataset Y t r = { y i } i = 1 n , where n is the number of data points. Without loss of generality, the j-th feature of data point y i (with j { 1 , , d } ) is defined as y i j Y , and this feature can be a missing value, numerical variable, or categorical variable. This study focused on inductive learning (a type of learning where general rules or models are derived from training data and applied to unseen data) with the goal of finding an imputation function f : Y R d , which will allow the conversion of the input space Y with missing values into a d-dimensional real-valued space, replacing missing values with reasonable values. The specific steps for finding the imputation function include the following.
(1)
Defining the Conditional Diffusion Model: Med-Diffusion introduces the Conditional Diffusion Model to impute the missing values conditioned on observed data. This includes the following. Conditioning Setup: For each sample y i , define the observed part y 0 c o (non-missing values) and the missing part y. The goal is to learn the conditional distribution p ( y | y 0 c o ) . In the diffusion model, the conditional information y 0 c o is incorporated into the reverse process. Forward Process: The forward process is the same as in the standard diffusion model, where noise is gradually added to the true data y 0 , generating Y 1 , , Y T . For numerical variables, Gaussian noise is directly added; for categorical variables, one-hot encoding is applied first, and then noise is added. Reverse Process: The reverse process is conditioned on the observed values y 0 c o and progressively denoised to generate imputed values. Med-Diffusion modifies the reverse process to focus on imputing the missing part while keeping the observed part unchanged.
(2)
Processing Numerical and Categorical Variables: Med-Diffusion employs distinct modeling strategies for numerical and categorical variables:
a.
Numerical Variables (e.g., heart rate and blood pressure). Modeling: Numerical variables are treated as continuous values, fitting the Gaussian noise assumption of the diffusion model. Imputation: During the reverse diffusion process, missing numerical values are progressively denoised to reasonable real values. For example, heart rate might recover from a noisy value to 80 beats per minute, and blood pressure may return to 120/80 mmHg. Loss Function: Mean squared error (MSE) is used to optimize the imputation results, ensuring the imputed values are close to the true distribution.
b.
Categorical Variables (e.g., alcohol consumption level and heart rhythm status). Modeling: Categorical variables are transformed into continuous vectors through one-hot encoding. However, the Gaussian noise of the diffusion model may disrupt the discrete nature. Therefore, Med-Diffusion introduces a Discrete Diffusion Model or a post-processing step. Imputation: During the reverse diffusion process, the one-hot (embedding) vector for categorical variables is denoised, generating a probability distribution (e.g., alcohol consumption level might produce [0.7, 0.2, 0.1]). Post-processing: The probability distribution is converted into discrete categories by selecting the category with the highest probability (e.g., [0.7, 0.2, 0.1] becomes “mild”). Loss Function: Cross-entropy loss is used to optimize the imputation results for categorical variables, ensuring the imputed category aligns with the true distribution.
(3)
Training the Imputation Function f: Neural Network Design: Med-Diffusion employs a neural network based on a diffusion architecture to parameterize the mean μ θ ( y t , t ) of the reverse diffusion process. Input: Current time step t, noisy data y t , and conditional information y 0 c o . Output: Denoised data y t 1 , which corresponds to the imputed data. Training Goal: The goal is to optimize the neural network parameters θ so that the predicted noise approaches the true noise, thus making the denoising process as realistic as possible. The objective function is
f sample ( θ ) = E t , y 0 , ϵ ϵ ϵ θ ( y t , y 0 c o ) 2 ,
where ϵ θ is the noise predicted by the neural network, and y 0 c o represents the conditional information. t Uniform { 1 , , T } : A time step t is randomly selected from the diffusion process. y 0 q ( y 0 ) : An original data point y 0 is sampled from the training set, which may contain missing or non-missing values. ϵ N ( 0 , I ) : Noise sampled from a standard normal distribution. y t = α t y 0 + 1 α t ϵ : Given an original sample y 0 and a time step t, the noisy version y t is obtained using this formula. α t is a time-dependent noise decay factor, which is predetermined in this experiment (usually provided by a pre-defined linear noise schedule). ϵ θ ( y t , y 0 c o ) : The model predicts the most likely original noise ϵ given the noisy data y t , the current time step t, and the conditional information y 0 c o (i.e., the observed non-missing part). ϵ ϵ θ ( y t , y 0 c o ) 2 : The standard mean squared error (MSE) between the model’s predicted noise and the true noise. Minimizing this error is equivalent to instructing the model to “reverse the noise I added.” Training Process. a. Sample y 0 from the training data Y t r . b. Randomly select a time step t and generate the noisy version y t . c. Use the observed values y 0 c o as conditioning information to predict the noise ϵ θ . d. Optimize the loss function and update the network parameters θ .
(4)
Inference Process (Generation Process):
a.
Input Sample. The input sample y new , which contains missing values, is represented as y new ( R { } ) d . This vector has several missing dimensions to be predicted (denoted by a mask M, where M j = 0 indicates the j-th dimension is missing and M j = 1 indicates the j-th dimension is observed). We divided the data into two parts—the conditional observed values y 0 c o as the observable part and the target missing values y as the part to be imputed. The observed and missing sections are represented by the blue and red areas in Figure 2.
b.
Initialization of Missing Value Region, Forming z T . To initiate the reverse diffusion process, we need an initial “fully noisy” vector. The observed region retains the original y 0 c o ; the missing region y is initialized with random noise (Gaussian distribution). We constructed an initial state z T
z T = α ¯ T y 0 + 1 α ¯ T ϵ ,
where ϵ N ( 0 , I ) , introducing noise in only the missing region, while the observed region retained the true values.
c.
Conditional Control Mechanism Input—Feeding into Multi-layer Denoising U-Net for T Steps. Next, we feed the current noisy data z t , the current time step t, and the conditional observed values y 0 c o into a neural network function ϵ θ (in this paper, U-Net). θ represents the network parameters, and the network predicts the noise term ϵ θ . In other words, the network asks the following: “How should I remove the noise to make the missing region more realistic?”
ϵ θ = ϵ θ ( z t , t , y 0 c o )
The “conditioning mechanism” (highlighted in red in the figure) is crucial as it ensures that the network knows which areas are the observed values and which are missing, allowing for the imputation of the missing values according to the completion process of the U-Net.
d.
Step-by-Step Reverse Diffusion, from z T to z 0 . Using the noise output ϵ θ from the forward process, we update according to the reverse formula
z t 1 = M y 0 + ( 1 M ) 1 α t z t 1 α t 1 α ¯ t ϵ θ ( z t , t , y 0 c o ) + σ t ϵ ϵ N ( 0 , I ) ,
Note that the observed region is “condition-locked” and does not participate in sampling or noise updating. The reverse diffusion mechanism generates the predicted missing values based on the conditional observed data. Starting from t = T , we update step-by-step to t = 0 , gradually denoising the data from a fully noisy state to the true data space, generating the imputed values for the missing regions.
e.
Obtaining the Imputed Result y ˜ . Finally, when t = 0 , the network outputs the predicted values for the features z 0 , which are decoded to obtain x ˜ . At this point, the missing regions have been filled through multiple steps of denoising, and the denoised output will be as close as possible to the true data distribution, with the observed regions remaining unchanged. This output can directly serve as the imputed result for further analysis.
Through these two steps—the training phase and the inference phase—the Med-Diffusion model effectively learns to recover the missing parts of samples with missing values, providing an accurate and efficient imputation function f.
f.
Post-processing. The processed input is then used to train the model. After obtaining the raw output, different processing schemes require different recovery processes.
For one-hot encoding, the index of the largest element is selected as the model’s inferred category. The model outputs a set of continuous values corresponding to the scores or activation values for each category. Categorical Variables: During prediction, the category corresponding to the largest index in the output vector is selected as the final predicted category. Numerical Variables: Since there is no encoding process for numerical variables, the model output is directly taken as the recovered numerical value.
For simulated bit encoding, if an output element is greater than 0, each output element is converted to 1; otherwise, it is converted to −1. The model output consists of continuous values corresponding to the simulated bit encoding. For subsequent categorical variables, each output element is thresholded and set to 1 if it is greater than 0, or −1 otherwise; then, the encoding vector is decoded back to the original category using a predefined mapping. Numerical Variables: Similar to one-hot encoding, the output is directly restored to a continuous value.
In the FT scheme (feature tokenization), we need to recover both numerical and categorical variables from the embeddings [27,28]. The model outputs an embedding vector of the same dimension.
Categorical Variables: We calculate the Euclidean distance between the Med-Diffusion output and each categorical embedding. The closest embedding (1-nearest neighbor) is selected as the final model output.
Numerical Variables: Since they have not undergone an encoding process, the model’s output is directly used as the recovered numerical value.
FT Decoding Example for Categorical Variables: To predict sleep stages (awake, light sleep, and deep sleep), each category has an “embedding vector” (a list of numbers). For example, see the following:
Category 1 embedding vector: e 1 = ( 0.2, 0.1) . Category 2 embedding vector: e 2 = ( 0.6, 0.2) . Category 3 embedding vector: e 3 = ( 0.4, 0.7) . After the CSDI model performs calculations, it outputs a predicted embedding vector e pred = ( 0.5, 0.3) . We calculate the Euclidean distance between e pred and each category’s embedding vector as follows:
d 1 = ( 0.5 0.2) 2 + ( 0.3 0.1) 2 = 0.5 d 2 = ( 0.5 0.6) 2 + [ 0.3 ( 0.2) ] 2 = 0.02 d 3 = ( 0.5+ ( 0.4) ) 2 + ( 0.3 0.7) 2 = 1.81
Since d 2 is the smallest (approximately 0.141), we selected Category 2 as the final prediction.

3.3.2. Evaluation Process

To evaluate the performance of f, we were given the test input data Y t e = { y i } i = 1 n and the true values X t e = { x i j R : y i j = } . We defined y ^ i j as the j-th feature imputed by f ( y i ) . Let M j = { i : y i j = } denote the index set of missing values for the j-th feature, and let N miss j = | M j | be the number of missing values for the j-th feature.
To compute the error of f, if j is a numerical variable, we use the Root Mean Square Error (RMSE); if j is a categorical variable, we use the error rate (Err):
RMSE ( j ) = i M j ( y ^ i j x i j ) 2 N miss j ,
Err ( j ) = 1 N miss j i M j y [ y ^ i j x i j ] ,
where y [ x ^ i j = y i j ] is an indicator function that returns 1 if the condition holds, otherwise 0. This result is used to optimize the prediction function f sample ( θ ) .

4. Experiments

4.1. Datasets

Five datasets were used to evaluate various health indicators and disease predictions. The PhysioNet Challenge 2012 dataset [39] comes from The PhysioNet/Computing in Cardiology Challenge 2012. The Wisconsin Breast Cancer dataset is from the UCI Machine Learning Repository on Kaggle [40]. The COVID-19, diabetes, and heart disease datasets are also from Kaggle. It is important to note that the diabetes and COVID-19 datasets contain only binary classification variables, and all numerical variables in the datasets are preprocessed using min-max normalization.
PhysioNet Challenge 2012 (PhysioNet): PhysioNet is a publicly available medical dataset that includes data from 12,000 ICU hospital admissions. All patients are adults admitted to cardiac, medical, surgical, or trauma ICUs for various reasons. Each record consists of approximately 48 h of multivariate time series data, which captures 37 metrics, such as respiratory rate and blood glucose, from multimodal sensor data during different time periods of the patient’s hospital stay. Each observation is associated with a timestamp in hours and minutes, indicating the observation time since the patient was admitted to the ICU. For example, a timestamp of 35:19 means the observation was made 35 h and 19 min after the patient’s ICU admission.
Indicators of Heart Disease [41]: This dataset integrates data from 401,958 adult respondents in the 2020 BRFSS survey (Kaggle, 2021), including heart disease diagnoses (yes/no) and 14 key predictors (e.g., smoking history, BMI, and exercise frequency). Of the samples, 8.1% were diagnosed with heart disease. The data has been weighted by the CDC to reflect the U.S. population distribution. The focus of this study was on the subgroup aged ≥ 50 years (n = 187,342) to match the target population. Researchers can predict heart disease risk and provide early interventions. This dataset is widely used in cardiovascular disease research.
Diabetes Health Indicators Dataset [42]: Sourced from the CDC’s BRFSS annual telephone survey (Kaggle, 2021), this dataset includes health indicators from 253,680 respondents. The data covers diabetes diagnosis (Type 1/2/gestational/no), BMI, blood pressure, cholesterol levels, and other clinical and behavioral features. A total of 13.4% of the samples were diabetic. This dataset is extensively used in diabetes prediction research, analyzing health conditions of diabetic patients through these indicators.
Breast Cancer Wisconsin Dataset [43]: Provided by the UCI Machine Learning Repository this dataset consists of 569 breast tumor biopsy samples with digital features. Each sample contains 30 cell morphology parameters (e.g., radius, texture, and number of concave points) obtained by pathologists through Fine Needle Aspiration (FNA) image analysis. The dataset is labeled as malignant (212 samples) or benign (357 samples), and it was split into a training and test set with a 7:3 ratio. This dataset is widely used in cancer diagnosis model research, particularly in the fields of machine learning and data mining.
COVID-19 patient pre-condition dataset: This dataset was collected by Tanmoy et al. and includes clinical records of COVID-19 patients from medical institutions across multiple countries (Kaggle, 2022). The data covers key indicators, such as underlying conditions (e.g., diabetes and cardiovascular disease), symptom severity, hospitalization methods, and final outcomes (recovery/death), with a sample size exceeding 100,000 cases. We extracted 20 features including age, underlying conditions, and blood oxygen saturation for this study. All data were anonymized by medical institutions and cross-center standardized by a professional team to ensure data authenticity and reliability. This dataset is used for predicting patient prognosis and has provided foundational data support for various studies on patient recovery.

4.2. Experimental Details

In our experiment, we used the Python 3.10.9 programming language in conjunction with the PyTorch 2.5.1+cu121 machine learning library to build an efficient diffusion probabilistic model aimed at optimizing the data encoding and generation process. The experimental environment was equipped with an NVIDIA GeForce RTX 2080 Ti graphics card, which significantly accelerates the training of deep learning models through its powerful floating-point computation capabilities.
For the training process of the tabular mixed data imputation task, the optimizer for the diffusion model employs a learning rate of 1× for updating the entire model parameters to ensure training stability. The batch size is set to 64 to balance computational efficiency and sample diversity, avoiding overfitting. The hyperparameter T for the diffusion process is set to 1000, enabling a more detailed simulation of the data diffusion process. A quadratic (quad) schedule is adopted for the diffusion noise range β [ 10 4 , 0.5] . The number of training epochs was set to 250 to ensure stable convergence and good generalization with the current data size and model capacity. In terms of model architecture, the diffusion network consists of four layers, each with 128 channels, and it features four heads of multi-head attention. The time embedding dimension was set to 128 and the feature embedding dimension was set to 8. Categorical features were tokenized and jointly modeled with numerical features for training and evaluation. The 500 training cycles ensure that the model sufficiently learns the data characteristics and achieves strong generalization, leading to more accurate predictions in real-world applications. During both the training and evaluation stages, the missing rate was set to 0.2 (test_missing_ratio = 0.2), and five-fold cross-validation was employed for model evaluation.
These experimental settings and parameter configurations ensure that the diffusion probabilistic model demonstrates efficient and accurate performance in missing data imputation tasks. Not only do these settings enhance training efficiency, but they also increase the model’s robustness and reliability in practical applications.

5. Results and Discussion

This section presents a performance comparison of Med-Diffusion with several existing methods, including mean/mode imputation, Multiple Imputation by Chained Equations (MICE, linear), Random Forest-based multiple imputation (MissForest), and Generative Adversarial Imputation Networks (GAINs). The performance comparison is based on two evaluation metrics: the Root Mean Square Error (RMSE) and error rate. The RMSE is used to assess the imputation accuracy of numerical variables, while the error rate is used to evaluate the imputation accuracy of categorical variables. Each main value in the table is followed by a bracket containing a decimal number. This number represents the standard error (SE), which serves as an indicator of the reliability of statistical estimates (such as mean and RMSE). A smaller SE indicates more stable estimates with higher confidence.
Table 1 presents a comparison of the RMSE and error rate performance on three mixed-variable datasets (diabetes, COVID-19, and cardiovascular datasets). Med-Diffusion achieved the lowest RMSE on the COVID-19 and cardiovascular datasets, while MissForest achieved the lowest error rate on the diabetes and cardiovascular datasets. For categorical variables, Med-Diffusion uses three different encoding techniques: one-hot encoding, simulated bit encoding, and feature tokenization (FT). The experimental results show that Med-Diffusion with FT achieved the lowest error rate on the COVID-19 dataset, indicating that FT is more effective than the other two techniques in handling categorical variables.
Table 2 shows a comparison of the RMSE performance on a pure numerical dataset (breast cancer dataset). Med-Diffusion demonstrated the best performance on this dataset, highlighting its effectiveness in imputing numerical variables. The comparison results are illustrated in Figure 4 and Figure 5.
Figure 4. Left: Method-wise performance across three datasets. Right: Standard error across five baseline and three ablation studies on all datasets.
Figure 5. Comparison of the RMSE and error rate performance of the different methods on different datasets.

Discussion

This study presents Med-Diffusion, a diffusion-based framework that combines one-hot encoding, simulated bit encoding, and feature tokenization for multimodal medical data imputation, addressing both categorical and numerical variables. To effectively handle categorical and numerical variables, the method explores the integration of these three techniques with the diffusion-based framework. Experimental results show that the feature tokenization (FT) technique performs better in handling categorical variables. This method leverages the latent structure and distribution information of the data, enabling effective imputation even in the presence of missing values. This is particularly valuable for complex, incomplete, and multimodal medical datasets that involve diverse variables and intricate dependencies. The diffusion probabilistic model better captures latent patterns, improving the quality and accuracy of the imputations.
Comparison Methods In this experiment, we compared Med-Diffusion with the following four classical missing data imputation methods.
(1)
Mean/Mode Baseline Method:
The mean/mode method is a simple imputation strategy. For numerical features, missing values are replaced with the global mean of that feature from the training set, while for categorical features, the most frequent category (mode) is used for imputation. This method does not require training complex models as it simply computes the mean or mode for each column, making it computationally inexpensive. Despite its simplicity, it provides a reasonable reference level for evaluating the performance of more advanced imputation algorithms. In this study, the mean/mode method showed significantly higher RMSE and error rates on mixed-variable datasets compared to deep and iterative models, highlighting the advantages of the other methods.
(2)
MICE (Multiple Imputation by Chained Equations) Linear Regression Version [44]:
MICE is a classic iterative multiple imputation method that treats each feature column as an imputation target. It uses all other columns as predictor variables to train a conditional distribution estimator, and it iteratively generates predicted imputed values for each missing value. In the “linear” version, linear regression is used to predict missing numerical feature values, while logistic regression is employed for categorical features. The process is as follows: (1) initialize all missing values (e.g., using mean/mode); (2) train the corresponding regression or classification model for each column and predict missing values; (3) use the newly imputed values to train the next column’s model; and (4) repeat the process until convergence or the preset iteration limit is reached. MICE captures the linear relationships between features and is suitable for datasets with strong correlations between features. However, its limitations lie in its reliance on strong parametric assumptions based on linear and generalized linear models. For the complex nonlinear relationships (e.g., age and disease risk not being a simple linear relationship), higher-order interaction effects (e.g., a drug’s effectiveness being influenced by both genotype and weight), and non-normal data distributions common in medical data, MICE’s modeling capacity is insufficient and may result in inaccurate imputations.
(3)
MissForest Random Forest Imputation [11]:
MissForest is another iterative imputation method, but it uses a non-parametric random forest model as a conditional distribution estimator, addressing the representational deficiencies caused by the linear model assumptions. Its core process is similar to MICE: starting with initial imputed values, a random forest is trained for the current feature to be imputed using other (imputed) features as inputs to predict the missing values and update the imputed values. This process is repeated until the error converges. Since random forests naturally handle mixed-type data (numerical and categorical) and have strong capabilities for modeling nonlinear and interaction relationships, MissForest shows excellent robustness for complex and diverse tabular data. However, its imputation process is autoregressive, meaning features are imputed sequentially, one by one, which may lead to error propagation and accumulation during iterations (errors in early features directly affect subsequent feature models’ training). Moreover, it approximates the joint distribution via the chain rule, rather than directly modeling the joint distribution of the data, which may limit its ability to generate globally consistent and diverse samples.
(4)
GAINs (Generative Adversarial Imputation Networks) [12]:
GAINs is a deep generative model based on GANs, where the missing data imputation problem is treated as an adversarial learning task. The framework consists of a generator and a discriminator: the generator receives input with masks (where missing values are randomly set as noise) and attempts to generate imputed values, while the discriminator learns to distinguish real observations from the “pseudo-imputed values” generated by the generator. During training, the generator and discriminator engage in a game where the generator tries to deceive the discriminator, ultimately learning an imputation strategy that matches the true data distribution. Compared to traditional iterative methods, GAINs captures more complex joint distributions and higher-order correlations, providing significant advantages for high-dimensional heterogeneous data. However, the game between the generator and discriminator requires fine hyperparameter tuning. Compared to diffusion models, GANs theoretically have weaker distribution coverage and may fail to fully learn the true distribution of complex multimodal medical data, potentially overlooking some “rare but plausible” imputed values. The performance may not meet expectations when dealing with non-time series data.
(5)
CSDI [45]:
CSDI is an innovative method for imputing multivariate time series data. It uses a score-based diffusion model conditioned on the observed data, making it particularly suitable for missing value imputation. Compared to previous score-based methods, CSDI enhances the imputation performance by explicitly training for the imputation task and leveraging the correlations between observations. In CSDI, the model starts from random noise and gradually transforms the noise into reasonable time series values through a reverse diffusion process. The reverse process is conditioned on the observed values, allowing it to better capture the temporal dependencies and feature relationships in the time series. Unlike earlier models that approximate conditional distributions, CSDI directly learns the conditional distribution, providing more accurate imputations. However, CSDI is primarily designed to handle continuous data and does not fully address the importance of categorical variables in data imputation, lacking direct handling capability for multimodal datasets (e.g., categorical variables). Additionally, the score-based diffusion process requires more samples and more complex computations, potentially leading to longer training times.

6. Conclusions

Our results demonstrate that Med-Diffusion effectively addresses missing data, significantly enhancing the availability and diversity of multimodal medical datasets—critical for accurate risk assessment and improved surgical outcomes. By incorporating techniques such as one-hot encoding, simulated bit encoding, and feature tokenization, Med-Diffusion enhances the model’s adaptability to both numerical and categorical data types. The framework generates high-quality synthetic data, effectively imputes missing values, and improves the overall data quality, making it a valuable tool for healthcare applications.

Author Contributions

Conceptualization, investigation, writing—original draft preparation, Z.C. (Zhenyu Cheng); methodology, B.Z.; formal analysis, Y.H. ; project administration, Y.D.; writing—review and editing, Z.Z., T.L. and Z.C. (Zhuoxu Cui); supervision, Z.Z. and T.L.; funding acquisition, C.L.; resources, S.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Key-Area Research and Development Program of Guangdong Province (No. 2025B1111020001), in part by the Natural Science Foundation of Guangdong Province (No. 2023A1515010673), in part by the Shenzhen Science and Technology Innovation Bureau key projects (No. JSGG20220831110400001, No. CJGJZD20230724093303007, No. KJZD20240903101259001), in part by the Shenzhen Medical Research Fund (No. D2404001), in part by the Shenzhen Engineering Laboratory for Diagnosis & Treatment Key Technologies of Interventional Surgical Robots (XMHT20220104009), and the Key Laboratory of Biomedical Imaging Science and System, CAS, for the research platform support.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding authors.

Acknowledgments

We gratefully acknowledge the generous support provided by the above funding agencies, whose contributions have made it possible to conduct the theoretical development, experimental validation, and analyses presented in this work. Their support has been instrumental in facilitating collaboration among team members, providing essential research resources, and enabling the progress of this study.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Moon, K.S.; Lee, S.Q. A wearable multimodal wireless sensing system for respiratory monitoring and analysis. Sensors 2023, 23, 6790. [Google Scholar] [CrossRef]
  2. Choi, S.H.; Yoon, H.; Baek, H.J.; Long, X. Biomedical Signal Processing and Health Monitoring Based on Sensors. Sensors 2025, 25, 641. [Google Scholar] [CrossRef] [PubMed]
  3. Yashudas, A.; Gupta, D.; Prashant, G.C.; Dua, A.; AlQahtani, D.; Reddy, A.S.K. Deep-cardio: Recommendation system for cardiovascular disease prediction using IoT network. IEEE Sens. J. 2024, 24, 14539–14547. [Google Scholar] [CrossRef]
  4. Shieh, A.; Tran, B.; He, G.; Kumar, M.; Freed, J.A.; Majety, P. Assessing ChatGPT 4.0’s test performance and clinical diagnostic accuracy on USMLE STEP 2 CK and clinical case reports. Sci. Rep. 2024, 14, 9330. [Google Scholar] [CrossRef]
  5. Abdulhussain, S.H.; Mahmmod, B.M.; Alwhelat, A.; Shehada, D.; Shihab, Z.I.; Mohammed, H.J.; Abdulameer, T.H.; Alsabah, M.; Fadel, M.H.; Ali, S.K.; et al. A Comprehensive Review of Sensor Technologies in IoT: Technical Aspects, Challenges, and Future Directions. Computers 2025, 14, 342. [Google Scholar] [CrossRef]
  6. Campello, V.M.; Gkontra, P.; Izquierdo, C.; Martín-Isla, C.; Sojoudi, A.; Full, P.M. Multi-centre, multi-vendor and multi-disease cardiac segmentation: The M&Ms challenge. IEEE Trans. Med. Imag. 2021, 40, 3543–3554. [Google Scholar]
  7. Vaduganathan, M.; Mensah, G.A.; Turco, J.V.; Fuster, V.; Roth, G.A. The global burden of cardiovascular diseases and risk: A compass for future health. J. Am. Coll. Cardiol. 2022, 80, 2361–2371. [Google Scholar] [CrossRef]
  8. Behrad, F.; Abadeh, M.S. An overview of deep learning methods for multimodal medical data mining. Expert Syst. Appl. 2022, 200, 117006. [Google Scholar] [CrossRef]
  9. Soenksen, L.R.; Ma, Y.; Zeng, C.; Boussioux, L.; Carballo, K.V.; Na, L.; Wiberg, H.M.; Li, M.L.; Fuentes, I.; Bertsimas, D. Integrated multimodal artificial intelligence framework for healthcare applications. NPJ Digit. Med. 2022, 5, 149. [Google Scholar] [CrossRef]
  10. Bardak, B.; Tan, M. Improving clinical outcome predictions using convolution over medical entities with multimodal learning. Artif. Intell. Med. 2021, 117, 102112. [Google Scholar] [CrossRef]
  11. Hong, S.; Lynn, H.S. Accuracy of random-forest-based imputation of missing data in the presence of non-normality, non-linearity, and interaction. BMC Med. Res. Methodol. 2020, 20, 199. [Google Scholar] [CrossRef] [PubMed]
  12. Yoon, J.; Jordon, J.; Schaar, M. Gain: Missing data imputation using generative adversarial nets. In Proceedings of the 35th International Conference on Machine Learning, Stockholm, Sweden, 10–15 July 2018; pp. 5689–5698. [Google Scholar]
  13. Li, Y.; Daho, M.E.H.; Conze, P.; Zeghlache, R.; Boité, H.L.; Tadayoni, R.; Cochener, B.; Lamard, M.; Quellec, G. A review of deep learning-based information fusion techniques for multimodal medical image classification. Comput. Biol. Med. 2024, 108635. [Google Scholar] [CrossRef] [PubMed]
  14. Li, Y.; Liao, I.Y.; Zhong, N.; Toshihiro, F.; Wang, Y.; Wang, S. Generative AI enables the detection of autism using EEG signals. In Proceedings of the Chinese Conference on Biometric Recognition, Xuzhou, China, 1–3 December 2023; pp. 375–384. [Google Scholar]
  15. Ho, J.; Jain, A.; Abbeel, P. Denoising diffusion probabilistic models. Adv. Neural Inf. Process. Syst. 2020, 33, 6840–6851. [Google Scholar]
  16. Zheng, S.; Charoenphakdee, N. Diffusion models for missing value imputation in tabular data. arXiv 2022, arXiv:2210.17128. [Google Scholar]
  17. Mahmmod, B.M.; Abdulhussain, S.H.; Naser, M.A.; Alsabah, M.; Hussain, A.; Al-Jumeily, D. 3D object recognition using fast overlapped block processing technique. Sensors 2022, 22, 9209. [Google Scholar] [CrossRef]
  18. Alsabah, M.; Naser, M.A.; Albahri, A.S.; Albahri, O.S.; Alamoodi, A.H.; Abdulhussain, S.H.; Alzubaidi, L. A comprehensive review on key technologies toward smart healthcare systems based IoT: Technical aspects, challenges and future directions. Artif. Intell. Rev. 2025, 58, 343. [Google Scholar] [CrossRef]
  19. Hameed, I.M.; Abdulhussain, S.H.; Mahmmod, B.M.; Hussain, A. Content based image retrieval based on feature fusion and support vector machine. In Proceedings of the 14th International Conference on Developments in eSystems Engineering (DeSE), Sharjah, United Arab Emirates, 7–10 December 2021; pp. 552–558. [Google Scholar]
  20. Chen, P.-F.; Chen, L.; Lin, Y.-K.; Li, G.-H.; Lai, F.; Lu, C.-W.; Yang, C.-Y.; Chen, K.-C.; Lin, T.-Y. Predicting postoperative mortality with deep neural networks and natural language processing: Model development and validation. JMIR Med. Inform. 2022, 10, e38241. [Google Scholar] [CrossRef]
  21. Hui, V.; Litton, E.; Edibam, C.; Geldenhuys, A.; Hahn, R.; Larbalestier, R.; Wright, B.; Pavey, W. Using machine learning to predict bleeding after cardiac surgery. Eur. J. Cardio-Thorac. Surg. 2023, 64, ezad297. [Google Scholar] [CrossRef]
  22. McLean, K.A.; Sgrò, A.; Brown, L.R.; Buijs, L.F.; Mountain, K.E.; Shaw, C.A.; Drake, T.M.; Pius, R.; Knight, S.R.; Fairfield, C.J.; et al. Multimodal machine learning to predict surgical site infection with healthcare workload impact assessment. NPJ Digit. Med. 2025, 8, 121. [Google Scholar] [CrossRef]
  23. Liu, T.; Zhang, Z.; Fan, G.; Li, N.; Xu, C.; Li, B.; Zhao, G.; Zhou, S. FocusMorph: A novel multi-scale fusion network for 3D brain MR image registration. Pattern Recognit. 2025, 167, 111761. [Google Scholar] [CrossRef]
  24. Rao, P.K.; Chatterjee, S.; Nagaraju, K.; Khan, S.B.; Almusharraf, A.; Alharbi, A.I. Fusion of graph and tabular deep learning models for predicting chronic kidney disease. Diagnostics 2023, 13, 1981. [Google Scholar] [CrossRef]
  25. Psychogyios, K.; Ilias, L.; Ntanos, C.; Askounis, D. Missing value imputation methods for electronic health records. IEEE Access 2023, 11, 21562–21574. [Google Scholar] [CrossRef]
  26. Fan, C.-C.; Peng, L.; Wang, T.; Yang, H.; Zhou, X.-H.; Ni, Z.-L.; Wang, G.; Chen, S.; Zhou, Y.-J.; Hou, Z.-G. TR-Gan: Multi-session future MRI prediction with temporal recurrent generative adversarial Network. IEEE Trans. Med. Imag. 2022, 41, 1925–1937. [Google Scholar] [CrossRef] [PubMed]
  27. Ahn, J.C.; Noh, Y.-K.; Hu, M.; Shen, X.; Shah, V.H.; Kamath, P.S. Use of Artificial Intelligence–Generated Synthetic Data to Augment and Enhance the Performance of Clinical Prediction Models in Patients with Alcohol-Associated Hepatitis and Acute Cholangitis. Gastro Hep Adv. 2025, 4, 100643. [Google Scholar] [CrossRef] [PubMed]
  28. Xie, F.; Yuan, H.; Ning, Y.; Ong, M.E.H.; Feng, M.; Hsu, W.; Chakraborty, B.; Liu, N. Deep learning for temporal data representation in electronic health records: A systematic review of challenges and methodologies. J. Biomed. Inform. 2022, 126, 103980. [Google Scholar] [CrossRef]
  29. Liaghat, B.; Ussing, A.; Petersen, B.H.; Andersen, H.K.; Barfod, K.W.; Jensen, M.B.; Hoegh, M.; Tarp, S.; Juul-Kristensen, B.; Brorson, S. Supervised training compared with no training or self-training in patients with subacromial pain syndrome: A systematic review and meta-analysis. Arch. Phys. Med. Rehabil. 2021, 102, 2428–2441. [Google Scholar] [CrossRef]
  30. Fortuin, V.; Baranchuk, D.; Rätsch, G.; Mandt, S. Gp-vae: Deep probabilistic time series imputation. In Proceedings of the International Conference on Artificial Intelligence and Statistics, Online, 26–28 August 2020; pp. 1651–1661. [Google Scholar]
  31. Waikel, R.L.; Othman, A.A.; Patel, T.; Hanchard, S.L.; Hu, P.; Tekendo-Ngongang, C.; Duong, D.; Solomon, B.D. Recognition of genetic conditions after learning with images created using generative artificial intelligence. JAMA Netw. Open 2024, 7, e242609. [Google Scholar] [CrossRef]
  32. Stekhoven, D.J.; Bühlmann, P. MissForest—Non-parametric missing value imputation for mixed-type data. Bioinformatics 2012, 28, 112–118. [Google Scholar] [CrossRef]
  33. Liu, T.; Zhang, Z.; Fan, G.; Li, B.; Zhou, S.; Xu, C. MambaVesselNet: A Novel Approach to Blood Vessel Segmentation Based on State-Space Models. IEEE J. Biomed. Health Inform. 2024, 29, 2034–2047. [Google Scholar] [CrossRef]
  34. Yang, D.; Tian, J.; Tan, X.; Huang, R.; Liu, S.; Guo, H.; Chang, X.; Shi, J.; Zhao, S.; Bian, J.; et al. Uniaudio: Towards universal audio generation with large language models. In Proceedings of the International Conference on Machine Learning, Vienna, Austria, 21–27 July 2024. [Google Scholar]
  35. Zhang, Z.; Zhang, J.; Zhang, X.; Mai, W. A comprehensive overview of Generative AI (GAI): Technologies, applications, and challenges. Neurocomputing 2025, 632, 129645. [Google Scholar] [CrossRef]
  36. Guo, X.; Zhao, L. A systematic survey on deep generative models for graph generation. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 5370–5390. [Google Scholar] [CrossRef]
  37. Kollovieh, M.; Ansari, A.F.; Bohlke-Schneider, M.; Zschiegner, J.; Wang, H.; Wang, Y. Predict, refine, synthesize: Self-guiding diffusion models for probabilistic time series forecasting. Adv. Neural Inf. Process. Syst. 2023, 36, 28341–28364. [Google Scholar]
  38. Alcaraz, J.M.L.; Strodthoff, N. Diffusion-based time series imputation and forecasting with structured state space models. arXiv 2022, arXiv:2208.09399. [Google Scholar]
  39. Silva, I.; Moody, G.; Mark, R.; Celi, L.A. Predicting Mortality of ICU Patients: The PhysioNet/Computing in Cardiology Challenge 2012. Available online: https://physionet.org/content/challenge-2012/1.0.0/ (accessed on 1 August 2025).
  40. Dua, D.; Graff, C. UCI Machine Learning Repository. Available online: https://archive.ics.uci.edu/dataset/17/breast+cancer+wisconsin+diagnostic (accessed on 1 August 2025).
  41. Pytlak, K. Personal Key Indicators of Heart Disease dataset. Available online: https://www.kaggle.com/datasets/kamilpytlak/personal-key-indicators-of-heart-disease (accessed on 1 August 2025).
  42. Teboul, A. Diabetes Health Indicators Dataset. Available online: https://www.kaggle.com/datasets/alexteboul/diabetes-health-indicators-dataset (accessed on 1 August 2025).
  43. Tanmoy, X. COVID-19 Patient Pre-Condition Dataset. Available online: https://www.kaggle.com/datasets/tanmoyx/covid19-patient-precondition-dataset (accessed on 1 August 2025).
  44. Royston, P.; White, I.R. Multiple imputation by chained equations (MICE): Implementation in Stata. J. Stat. Softw. 2011, 45, 1–20. [Google Scholar] [CrossRef]
  45. Tashiro, Y.; Song, J.; Song, Y.; Ermon, S. Csdi: Conditional score-based diffusion models for probabilistic time series imputation. Adv. Neural Inf. Process. Syst. 2021, 34, 24804–24816. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.