Next Article in Journal
A New Depth-Based Test for Multivariate Two-Sample Problems
Previous Article in Journal
On the Classification–Causal Tradeoff in Neural Network Propensity Score Estimation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multiple Imputation of a Continuous Outcome with Fully Observed Predictors Using TabPFN

Center for Health, Policy and Economics, Faculty of Health Sciences and Medicine, University of Lucerne, 6002 Luzern, Switzerland
Stats 2026, 9(2), 38; https://doi.org/10.3390/stats9020038
Submission received: 24 February 2026 / Revised: 24 March 2026 / Accepted: 29 March 2026 / Published: 1 April 2026
(This article belongs to the Special Issue Statistical Methods for Hypothesis Testing)

Abstract

Handling missing data is a central challenge in quantitative research, particularly when datasets exhibit complex dependency structures, such as nonlinear relationships and interactions. Multiple imputation (MI) via fully conditional specification (FCS), as implemented in the MICE R package, is widely used but relies on user-specified models that may fail to capture complex dependency structures, especially in high-dimensional settings, or on more sophisticated algorithms that are considered data-hungry. This paper investigates the performance of TabPFN, a transformer-based, pretrained foundation model developed for tabular prediction tasks, for MI. TabPFN is pretrained on millions of synthetic datasets and approximates posterior predictive distributions without dataset-specific retraining, offering a compelling solution for imputing complex missing data in small to moderately sized samples. We conduct a simulation study focusing on univariate missingness in a continuous outcome with complete predictors, comparing TabPFN with standard MI methods. Performance is evaluated using bias, standard error, and coverage of the marginal mean estimand across a range of data-generating and missingness mechanisms. Our results show that TabPFN yields competitive or superior performance relative to Classification and Regression Trees and Predictive Mean Matching. These findings highlight TabPFN as a promising tool for missing data imputation, with particular relevance to health research.

1. Introduction

Quantitative analyses in health research, including clinical trials and observational studies, commonly assume that all variables are fully observed. In practice, however, missing data in covariates or outcomes are ubiquitous and can arise even in carefully designed randomized controlled trials (RCTs), such as when patients withdraw from follow-up and outcome measurements cannot be obtained. If not appropriately handled, missing data can lead to biased estimates, loss of efficiency, and invalid statistical inference [1,2]. Regulatory guidance therefore emphasizes the need for principled approaches to missing data that support valid estimation and uncertainty quantification, particularly for pre-specified estimands in confirmatory analyses. More broadly, the literature on missing data methods is extensive, and identifying appropriate imputation strategies is of particular interest in the life sciences, where a large share of comparative studies has been conducted [3]. Illustrative examples include applications to electronic health records and clinical measurements [4], binary classification problems [5], and time-series health data [6]. Beyond classical approaches, a growing body of work investigates machine learning (ML) and deep learning-based imputation methods in healthcare data [7], as well as methods designed for high-dimensional settings [8]. These studies consistently emphasize that no single method performs uniformly best, with performance depending strongly on the data structure, dimensionality, and the analysis objective. At the same time, much of the existing literature focuses on predictive performance metrics, whereas relatively fewer studies assess the ability of imputation methods to recover the underlying data distribution or support valid statistical inference [9,10,11]. This distinction is particularly relevant in applications where valid inference, rather than predictive accuracy alone, is the primary objective.
Multiple imputation (MI) provides a widely accepted framework for addressing missing data under valid uncertainty quantification by generating multiple plausible completed datasets, analyzing each dataset separately, and combining results using Rubin’s rules to reflect both sampling variability and uncertainty due to missingness [12]. The performance of MI depends critically on the adequacy of the imputation models used to approximate the joint distribution of the data. When important complex dependency structures, such as nonlinear relationships or interactions, are omitted from the imputation models, MI can yield biased point estimates and misleading measures of uncertainty [13,14]. This challenge is particularly relevant when several variables are partially observed, a setting commonly addressed using fully conditional specification (FCS) algorithms, such as those implemented in the widely used MICE (Multivariate Imputation by Chained Equations) package [15]. FCS MI proceeds by iteratively imputing each incomplete variable using a univariate conditional model given the current values of all other variables, thereby implicitly defining a joint distribution. However, unless the conditional models correspond to a coherent joint distribution, incompatibilities may arise, particularly when nonlinear relationships or interactions are present [16]. In many applications, scientific interest focuses on parameters that are more remote from the full joint distribution, such as regression coefficients or prevalence estimates. In such settings, the joint distribution primarily acts as a working model for generating imputations, and empirical studies suggest that FCS performs well in many applied contexts ([16] Chapter 4.5). A key advantage of FCS compared with joint modeling approaches is its flexibility, as it allows the imputation model for each variable to be specified separately. Nevertheless, this flexibility requires careful specification of functional forms, transformations, and interactions in each conditional model to ensure congeniality with the substantive analysis model [17]. This task becomes increasingly challenging in high-dimensional settings or when relationships between variables exhibit complex dependency structures [18].
Motivated by these challenges, a growing body of research has explored the use of flexible predictive algorithms for missing data imputation. Tree-based methods such as Classification and Regression Trees (CART) [19], random forests [20], and gradient boosting approaches, including XGBoost [21], have been incorporated into imputation frameworks such as MICE and MIXGB. Simulation studies suggest that these approaches can better preserve complex dependency structures than traditional parametric imputation models, leading to improved finite-sample performance in some settings [22,23,24,25]. Related work includes ensemble-based MI using super learners, as implemented in the MISL package [26], and MI based on multivariate adaptive regression splines (MARS), which have demonstrated favorable confidence-interval coverage under complex data-generating mechanisms [27].
Despite these advances, highly flexible models typically require moderate-to-large sample sizes to reliably estimate complex dependency structures [28,29]. Hence, ML and neural-network-based imputation methods should not be viewed as universally preferable. Their performance depends strongly on the data structure and application context, and in some specialized settings more direct imputation approaches have been shown to outperform more complex ML-based methods [30]. While large observational datasets may support such approaches, clinical trials and other confirmatory studies are often limited in size, making it difficult to fit highly flexible imputation models reliably without imposing strong structural assumptions. Nevertheless, appropriately addressing missing data remains essential for maintaining valid statistical inference in these settings, particularly when regulatory decisions depend on accurate estimation and uncertainty quantification [31].
Recent developments in foundation models for tabular data offer a promising alternative to conventional task-specific modeling approaches, which often struggle to generalize under limited data settings [32]. The Tabular Prior-Data Fitted Network (TabPFN), introduced by Hollmann et al. [33], is a transformer-based model trained on millions of synthetic tabular datasets. By learning a broad prior over data-generating processes, TabPFN is designed to generalize across a wide range of prediction tasks, effectively encoding structural assumptions learned during pretraining rather than relying solely on estimation from the observed data. Even without task-specific fine-tuning, TabPFN has demonstrated strong predictive performance in small-to-moderately sized datasets (up to approximately 10,000 observations and 500 features) [34,35], making TabPFN particularly attractive for data-scarce domains [35].
Naturally, promising prediction algorithms are adapted for missing data imputation tasks often in the context of single imputation (SI), where missing values are replaced by a single predicted value, as in the TabPFN-based imputation approach proposed by Feitelberg et al. [36], which is specifically pretrained for imputing missing values. While such SI approaches may perform well in terms of predictive accuracy, strong predictive performance alone does not guarantee valid statistical inference after imputation. SI does not propagate uncertainty due to missingness and therefore generally yields overly optimistic standard errors and confidence intervals [1]. Because of this inadequate uncertainty quantification, MI is generally recommended over SI for inferential purposes [31,37]. To date, the use of pretrained foundation models such as TabPFN within a MI framework explicitly designed to preserve inferential validity remains largely unexplored.
The aim of this paper is to investigate the performance of TabPFN as an imputation engine within a MI framework suitable for valid statistical inference. Our contributions are threefold. First, we propose a principled integration of TabPFN into an MI procedure that explicitly accounts for uncertainty due to missing data. Second, we conduct an extensive simulation study to compare the proposed approach with established MI methods implemented in MICE, focusing on bias, standard error, and confidence-interval coverage under a range of data-generating and missingness mechanisms. Third, we illustrate the practical relevance of the method using a real-data example motivated by applied epidemiological research. Section 2 introduces the TabPFN model and its use for MI. Section 3 describes the simulation design, competing methods, and evaluation metrics. Section 4 presents the simulation results, Section 5 provides a real-data application, Section 6 discusses the numerical findings, and Section 7 concludes with broader implications and directions for future research.

2. Methods

2.1. Brief Introduction to TabPFN

TabPFN is a transformer-based neural-network architecture incorporating self-attention mechanisms [38], originally developed for sequence modeling in natural language processing and later extended to various other domains, including tabular data. TabPFN is designed for classification and regression tasks and aims to approximate predictive distributions induced by a broad prior over data-generating processes in a non-parametric and computationally efficient manner [39].
Unlike traditional models that must be retrained for each new dataset, TabPFN is pretrained on a wide range of synthetic datasets. These datasets are generated by sampling from a distribution over a variety of data-generating processes, including structural causal models and Bayesian neural networks with varying parameters and noise levels [40]. This distribution of data-generating processes—referred to by the authors as the prior —reflects a broad but structured set of assumptions about plausible relationships in tabular data. Thus, rather than learning a dataset-specific model from scratch, TabPFN performs prediction under a learned prior that can be especially useful in small-sample settings. Through this pretraining, TabPFN learns to approximate Bayesian-style predictive inference across diverse data scenarios without requiring dataset-specific retraining.
Intuitively, TabPFN does not reuse specific relationships from external datasets for the dataset at hand. Instead, through pretraining on many synthetic datasets drawn from a broad range of plausible data-generating processes, it learns a general prior over how variables in tabular data may relate to one another. When applied to a new dataset, this prior is combined with the observed complete cases in that dataset to generate predictive distributions for missing values. In this sense, TabPFN imposes an inductive bias, much like a conventional regression model imposes structural assumptions through its functional form. Such bias can be advantageous in small-sample settings, where highly flexible methods may be unstable despite being asymptotically less restrictive.
For a specific task, TabPFN is provided with complete cases, consisting of observed outcomes and covariates ( Y comp , X comp ) , and incomplete cases, with the observed covariates X incomp but missing the outcomes Y incomp * . A single forward pass through the pretrained model yields a predictive distribution for Y incomp * that reflects both the observed data and the prior assumptions [33].
In practice, TabPFN has been shown to support efficient predictive inference on small-to-moderately sized datasets (e.g., up to approximately 10,000 observations on GPU or 1000 observations on CPU with hundreds of variables), as reported in Hollmann et al. [33], making it particularly well suited to experimental settings where data collection is costly. TabPFN’s output is a full predictive distribution rather than a single point estimate, from which MIs can be generated by repeated sampling. Hence, MIs are obtained by independently sampling from this predictive distribution for each missing value, thereby introducing stochastic variation across imputations conditional on the observed data. Because the model is pretrained on a diverse range of generative processes, it can capture complex structures such as nonlinear associations, variable interactions, and heteroskedasticity. These features make TabPFN a promising method for handling missing data in biomedical and epidemiological research, where such complexities and limited sample sizes often challenge more traditional imputation methods.

2.2. Using TabPFN for Multiple Imputation

Focusing on univariate missingness, we defined Y R n × 1 as the partially missing outcome variable, with corresponding predictors X R n × p that were fully observed. The missingness indicator vector was denoted by M, where m i = 1 indicated that y i was missing and m i = 0 indicated that y i was observed. Thus, Y comp and X comp represent the observed subset of Y and X (i.e., where m i = 0 ), while Y incomp * and X incomp represent the subset with missing Y (i.e., where m i = 1 ).
Missing data mechanisms are commonly classified as missing completely at random (MCAR) when the probability of missingness is independent of both observed and unobserved data; missing at random (MAR) when it depends only on observed data; and missing not at random (MNAR) when it depends on unobserved information such as the partially missing variable Y itself. Here, we focus on imputation under MAR (or MCAR).
TabPFN performs classification. Therefore, continuous outcomes are, by default, discretized into 5000 classes prior to imputation, thereby reframing the regression problem as a multi-class classification task (see Müller et al. [40] for details on the discretization procedure). For each observation with m i = 1 , TabPFN outputs a vector of class probabilities, representing a predictive distribution for Y incomp * conditional on X incomp , X comp , and Y comp . This discretization constitutes an approximation to the underlying continuous outcome distribution and may affect variance estimation and tail behavior. Its impact on inferential performance is therefore evaluated empirically in the simulation study.
Thus, under the MAR (or MCAR) assumption, when all variables driving the missingness mechanism are observed and included in the imputation model, MIs can be generated by sampling from these discrete predictive distributions without specifying a parametric error distribution. After generating m imputed datasets using TabPFN-based predictive draws, the resulting parameter estimates and standard errors are combined using Rubin’s rules in the same manner as in conventional MI procedures. For additional details, including feature pre-processing, acceleration, and post-hoc ensembling, we refer to Hollmann et al. [33]. We do not claim formal theoretical validity of Rubin’s rules under TabPFN-based imputation. Instead, inferential adequacy of the proposed approach is evaluated empirically through bias, standard error behavior, and confidence-interval coverage in controlled simulation settings. Rubin’s rules are applied under the working assumption that TabPFN’s prediction draws approximate proper posterior predictive distributions for the missing values, such that between-imputation variability reflects uncertainty about the missing data. This assumption is pragmatic: in contrast to conventional MI approaches that rely on parametric or dataset-specific predictive models, TabPFN is trained offline to approximate Bayesian predictive inference across a broad class of tabular prediction tasks, but formal guarantees for a specific dataset are not available.

3. Simulation Study

The simulation setup employed in this study was closely aligned with a design previously used by Sepin [27], which itself builds upon the second simulation study by Little and An [41]. The design varied sample size, number of noise variables, outcome structure, and missingness mechanism, resulting in a fully factorial set of scenarios. All simulations were implemented in R using GPU acceleration.
In line with the recommendation by Oberman and Vink [42], we used the same simulated dataset across all strategies within each simulation run to eliminate unnecessary variation and ensure fair comparison.

3.1. Data and Missingness Generation

We considered a single partially observed outcome Y, with the estimand of interest being the marginal mean μ = E ( Y ) , a quantity that depends on correctly recovering the marginal distribution of Y under missing data ([16] Chapter 2.7). The covariates available for the missing data imputation consisted of two informative predictors along with two different amounts of irrelevant noise variables. For each simulation run, we independently generated n observations according to the following data-generating process. Two covariates relevant for the outcome and missingness mechanisms, U 1 and U 2 , were generated as independent draws from the uniform distribution on 1 , 1 . In addition, p irrelevant noise variables X 1 , , X p were independently sampled from a standard normal distribution, X 1 , , X p N p ( 0 , I ) .
The outcome Y was then generated as a function of U 1 and U 2 , according to one of three structural forms:
  • Linear:  Y N ( 10 ( 1 + U 1 + 3 U 2 ) , 2 2 ) ;
  • Additive:  Y N ( 118 + ( 3 U 1 3 ) 3 + ( 3 U 2 3 ) 3 , 2 2 ) ;
  • Non-additive:  Y N ( 10 ( 1 + U 1 + U 2 + 4 U 1 U 2 ) , 2 2 ) .
This means that each structure implies a marginal mean of E ( Y ) = 10 . Missingness in Y was imposed under both MCAR and MAR mechanisms with approximately 50% missingness. The probability of observing Y, conditional on U 1 and U 2 , was defined as follows:
  • MCAR:  P ( M = 0 U 1 , U 2 ) = 0.5 ;
  • Linear MAR:  P ( M = 0 U 1 , U 2 ) = expit ( U 1 + U 2 ) ;
  • Additive MAR:  P ( M = 0 U 1 , U 2 ) = expit ( U 1 3 + U 2 3 ) ;
  • Non-additive MAR:  P ( M = 0 U 1 , U 2 ) = expit ( U 1 + U 2 + 3 U 1 U 2 ) .
Here, expit ( x ) = exp ( x ) 1 + exp ( x ) . Investigating the behavior under MNAR mechanisms was outside the scope of this study. In total, the simulation design formed a full factorial combination of:
  • p { 10 , 100 } : number of noise variables;
  • n { 100 , 200 , 400 } : number of observations;
  • 3 mean structures: linear, additive, non-additive;
  • 4 missingness mechanisms: MCAR, linear MAR, additive MAR, non-additive MAR.
This results in 72 experimental conditions, each repeated 250 times.

3.2. Comparator Multiple Imputation Methods

Comparator MI strategies were implemented using the MICE package (version 3.19.0) [15] with default hyperparameters. Since missingness was limited to a single variable, the number of iterations (maxit) was set to 1, and the number of imputations to m = 20 , consistent with common MI practice ([16] Chapter 2.8). To avoid unnecessary filtering of collinear predictors, the eps parameter was fixed at 0.
The comparator MI approaches used here comprised Predictive Mean Matching (PMM) [43], a semi-parametric method that selects donor values from observed cases and is the default for continuous variables in MICE, and Classification and Regression Trees (CART) [19], a non-parametric recursive partitioning approach implemented in MICE and previously studied for MI in Shah et al. [23] and Doove et al. [24]. CART was selected as a representative tree-based imputation method because previous studies have shown that it can perform competitively in MI settings while preserving interaction structures in the data [24].
For context, two non-imputation baselines were also analyzed. The full-data estimate used the complete simulated dataset prior to introducing missingness, providing a benchmark for the best achievable performance. The complete-case (CC) analysis included only observations without missing values in the target variable.

3.3. Performance Evaluation

Let μ ^ s denote the pooled estimate of μ = E Y obtained via Rubin’s rules after MI in simulation repetition s = 1 , , N sim , where N sim = 250 . To assess the performance of each MI method, we computed the following summary statistics across simulation runs:
  • Mean and Standard Deviation of Bias:
    For each simulation repetition, the bias was defined as μ ^ s 10 . For each method and simulation condition, we reported the mean and standard deviation of the resulting biases.
  • Mean and Standard Deviation of the Standard Error (SE):
    The within-imputation standard error of μ ^ s , denoted by SE ( μ ^ s ) , was calculated and summarized using the mean and standard deviation across repetitions.
  • Coverage of the 95% Confidence Interval:
    To assess interval performance, we evaluated whether 10 fell within the 95% confidence interval for μ ^ s . Specifically, we checked whether μ ^ s 10 / SE ( μ ^ s ) < t 0.975 , ν where t 0.975 , ν denotes the 97.5th percentile of the t-distribution with pooled degrees of freedom according to Barnard and Rubin [44]. Coverage was then defined as the proportion of simulations in which this criterion was met.
Uncertainty in the estimated 95% coverage rate was quantified using a 90% Wilson confidence interval for binomial proportions [45]. More specifically, for a coverage rate p ^ based on N sim simulation replications, the interval is computed as
p ^ + z 2 2 N sim ± z p ^ ( 1 p ^ ) N sim + z 2 4 N sim 2 1 + z 2 N sim
where z = z 1 α / 2 denotes the corresponding quantile of the standard normal distribution. Finally, for each method and condition, we recorded the average computation time required to complete the imputation and pooling process.

3.4. Secondary Simulation: Sensitivity to the Number of Imputations (m)

To assess the sensitivity of our findings to the number of imputations, we conducted a secondary simulation study varying the number of imputed datasets (m). Specifically, we re-ran the simulation scenarios with a non-additive outcome structure at the smallest sample size ( n = 100 ) using m = 100 imputations and compared the results to those obtained with the default choice of m = 20 . All missingness mechanisms (constant, linear, additive, and non-additive) and noise variables ( p = 10 , 100 ) considered in the primary simulation study were retained.
This specific scenario was chosen since the metric λ = ( B + B / m ) / T [ 0 , 1 ] was comparatively high (see Figure A1 in the Appendix A), where B denotes the between-imputation variance and T the total variance of the MI estimator. λ can be interpreted as the proportion of the variation attributable to the missing data and thus reflects the difficulty of the imputation problem. Values close to zero indicate that the contribution of missingness to the overall variance is negligible, a situation that can only arise if the missing values can be recovered almost perfectly. Conversely, values near one indicate that uncertainty is dominated by missing data, corresponding to situations in which the observed data contain little information about the missing values ([16] Chapter 2.3.5). Increasing the number of imputations m provides an additional lever for stabilizing inference and is therefore sometimes used to guide the choice of m in practice ([16] Chapter 2.8).
Across all scenarios, estimates of the performance metrics were very similar for m = 20 and m = 100 with no qualitative changes in method ranking or conclusions, suggesting that m = 20 imputations are sufficient for the scenarios considered here. Detailed results are provided in Figure A2 in the Appendix A.

4. Results

We conducted a simulation study, broadly following the principles outlined in Little and An [41] and Sepin [27], to compare the performance of TabPFN-based MI with established MI methods. All MI procedures produced complete datasets with no remaining missing values.
Figure 1 shows the bias of estimates, with error bars representing the 25th, 50th (median), and 75th percentiles, with mean and standard deviation reported in parentheses. Bias was defined as the difference between each method’s pooled estimate and the true parameter value of 10. Complete-case (CC) analysis, which omits all observations with missing values, was included as a baseline. Across most experimental settings, CC produced the largest deviations from the true value (except under MCAR), underscoring the inefficiency of discarding incomplete cases. In contrast, MI methods yielded pooled estimates closer to the full-data estimand computed before introducing missingness, substantially reducing bias induced by missingness. Among the MI approaches considered, TabPFN-based MI exhibited among the smallest biases across simulation scenarios. Increasing the sample size generally reduced bias, while reducing the number of noise variables had a smaller but still favorable effect. PMM, which relies on linear and additive conditional models, showed limited bias reduction with increasing sample size when these assumptions were violated, whereas the more flexible CART and TabPFN approaches benefited from larger samples in such settings. Similarly, under MCAR (constant missingness), PMM produced notably biased estimates for nonlinear mean structures (additive and non-additive) when many noise variables were present, whereas CART and TabPFN adapted effectively.
Figure 2 shows the standard errors (SEs) of estimates, with error bars representing the 25th, 50th (median), and 75th percentiles, with mean and standard deviation reported in parentheses. The red vertical lines indicate the median standard errors from the full-data analysis (before introducing missingness), serving as reference values. As expected, in most scenarios MI methods had larger SEs than the full-data SE, reflecting the inclusion of both within- and between-imputation variance. In the most challenging scenarios involving non-additive missingness, TabPFN occasionally produced standard errors slightly smaller than the full-data reference, suggesting mild underestimation of uncertainty in finite samples. CART-based imputations tend to underestimate uncertainty and do not satisfy the conditions for proper MI unless additional stochasticity (e.g., bootstrap resampling) is introduced, as discussed in the MI literature (e.g., van Buuren [16] Chapter 3.5). PMM generally yielded larger and more conservative standard errors. Increasing the sample size reduced SEs, while adding noise variables generally increased them. Performance under MCAR (constant missingness) showed similar behavior.
Coverage rates for 95% confidence intervals (CIs) are shown in Figure 3. Ideally, coverage should be close to 95%. Several methods failed to attain nominal coverage in specific scenarios, indicating underestimation of uncertainty even under MCAR missingness. Larger sample sizes generally improved coverage. However, for methods with restrictive modeling assumptions, such as PMM and CC, coverage sometimes deteriorated when these assumptions were violated by the data-generating mechanism. In some cases, inflated SEs produced conservative coverage, though at the expense of precision. The TabPFN-based MI approach maintained coverage close to the nominal 95% level across most scenarios, with undercoverage primarily observed for the combination of non-additive mean structure and non-additive missingness at the smallest sample size ( n = 100 ).
Average computational times to generate m = 20 imputations are given in Table 1. The whole simulation study was executed in Google Colab using an A100 GPU and took approximately 92 h in total to run. TabPFN required substantially longer computation times than competing methods. Moreover, practical use of TabPFN currently depends on access to specialized hardware, such as GPUs, which may limit its applicability in some settings.

5. Case Study: Estimating the Effect of Smoking Cessation on Weight Gain

We revisit a classic benchmark example from What If? [46], which investigates the causal effect of smoking cessation on weight gain using data from the National Health and Nutrition Examination Survey Epidemiologic Follow-Up Study (NHEFS). The data and original analysis code are publicly available from the book’s website (https://miguelhernan.org/whatifbook, accessed on 16 January 2026). The case study is intended to illustrate practical differences between imputation strategies rather than to establish a causal ground truth.
The cohort consists of 1566 adult smokers aged 25–74 who completed both a baseline visit (1971–1975) and a follow-up visit approximately 10 years later. Smoking status was assessed at both visits, and individuals were classified as quitters ( A = 1 ) if they had quit smoking before follow-up, or as non-quitters ( A = 0 ) otherwise. The outcome of interest, weight gain (Y), was defined as the difference in measured weight (kg) between the two visits. For the present analysis, we restrict attention to individuals with complete baseline covariate information, resulting in a smaller analytic sample described below (n = 490), which is comparable in magnitude to the sample sizes considered in the simulation study.
We follow the approach outlined in Westreich et al. [47], leveraging MI approaches to estimate the average causal effect (ACE) of smoking cessation on weight gain,
ACE = E Y a = 1 E Y a = 0 ,
where Y a = 1 and Y a = 0 denote the potential weight gains had all individuals quit or not quit smoking, respectively. Identification of the ACE relies on the standard assumptions of consistency, conditional exchangeability given measured baseline covariates, and positivity. As in Hernán and Robins [46] (Chapter 13), we assume that the available baseline covariates are sufficient to control for confounding.
Because each individual is observed under only one treatment condition, the counterfactual outcome under the alternative condition is unobserved. This structure allows causal inference to be formulated as a missing-data problem in which the missingness mechanism corresponds to treatment assignment. However, valid inference still depends on appropriate modeling of the outcome conditional on covariates and treatment.
Hernán and Robins estimate the ACE using g-computation based on a linear outcome regression model. Specifically, the observed outcome y i = wt 82 _ 71 i is modeled as
y i = β 0 + β 1 a i + β 2 sex i + β 3 race i + β 4 age i + β 5 age i 2 + β 6 education i + β 7 smokeintensity i + β 8 smokeintensity i 2 + β 9 smokeyrs i + β 10 smokeyrs i 2 + β 11 exercise i + β 12 active i + β 13 wt 71 i + β 14 wt 71 i 2 + β 15 a i · smokeintensity i + ε i ,
where a i indicates smoking cessation, demographic variables include sex, race, age, and education, smoking history is summarized by smoking intensity and smoking years, physical activity is captured by exercise and activity indicators, and wt 71 i denotes baseline weight. In addition to the variables used in Equation (3), the NHEFS dataset contains further baseline covariates describing socioeconomic status, health behaviors, and clinical characteristics, which we later exploit in the imputation models (see Table A1 in the Appendix A).
In the original analysis, Equation (3) is used for g-computation. Briefly, g-computation fits the outcome model and predicts each individual’s potential outcomes under both treatment conditions by setting a i = 1 and a i = 0 , while holding all other covariates fixed. This yields predicted potential outcomes Y i a = 1 ^ and Y i a = 0 ^ for each individual. The ACE is computed as the mean difference between these two predictions across individuals ( ACE ^ = 1 n i = 1 n Y i a = 1 ^ Y i a = 0 ^ ), with uncertainty quantified via non-parametric bootstrap (500 resamples). Because both observed and counterfactual outcomes are replaced by model predictions, the resulting estimate depends entirely on the correctness of the specified outcome model.
In contrast, the MI approach leaves the observed outcomes unchanged and imputes only the missing counterfactual outcomes, treating them explicitly as missing data. This distinction means that observed data retain their original variability, and model uncertainty is propagated through MI. To reflect a realistic setting where the true outcome model is unknown, we extend the imputation model to include a richer set of 46 covariates from NHEFS. To isolate the impact of imputation methods on counterfactual outcomes only, we restrict attention to individuals with complete baseline covariate information, yielding a final analytic sample of 490 individuals for all analyses (see Table A1 in the Appendix A).
The dataset is then restructured so that each individual has two records, one per treatment condition, with weight gain observed for the factual treatment and imputed for the counterfactual. We generate 100 imputations and pool results using Rubin’s rules.
Table 2 presents ACE estimates, standard errors, and 95% confidence intervals based on normal approximation. The g-computation estimate was 3.380 kg, closely matching the result reported in Hernán and Robins [46] (Chapter 13). CART yields an ACE estimate of similar magnitude (3.419 kg), whereas TabPFN produces a somewhat lower estimate (2.746 kg), and PMM obtained the smallest effect estimate (2.363 kg).
Because the true causal effect is unknown in observational data, it is not possible to determine which estimate is closest to the truth. All 95% confidence intervals overlapped, indicating no clear contradiction across methods. The differences in point estimates likely reflect the distinct modeling assumptions underlying the approaches. PMM and g-computation rely on analyst-specified linear models and therefore impose stronger functional form constraints: in PMM, the imputation model is linear and additive, with flexibility arising only indirectly through donor selection, whereas g-computation explicitly incorporates selected nonlinear terms and interactions as specified in Equation (3). In contrast, CART and TabPFN adapt more flexibly to complex covariate structures, with CART being fully data-driven and TabPFN incorporating regularization through its pretrained prior, which may shrink extreme counterfactual predictions toward more conservative values.

6. Discussion

Flexible, non-parametric approaches such as CART and random forests have repeatedly been shown to reduce bias under complex dependency structures, particularly when linearity and additivity assumptions are violated [22,23,24], with recent work using XGBoost pushing performance even further [25]. However, these methods typically require moderate to large sample sizes to estimate such structures reliably, posing challenges in applied settings with limited data availability [29,32]. Our findings suggest that pretrained, foundation-model-style predictors such as TabPFN may partially alleviate this gap in small-sample settings by leveraging strong prior information learned from large collections of synthetic datasets. By introducing and benchmarking a TabPFN-based MI algorithm, we show that it achieves performance that is comparable to, and in some scenarios superior to, standard MI methods with minimal model specification effort, potentially mitigating one of the key limitations of highly flexible approaches in low-sample settings.
The strong performance of TabPFN-based MI likely reflects its ability to approximate complex dependency structures without requiring explicit specification of functional forms or interactions in the imputation model. PMM relies on linear predictive models to define donor pools for matching and is therefore more robust than fully parametric imputation with respect to outcome distributional assumptions, while remaining sensitive to misspecification of the conditional mean structure [18,48]. TabPFN, on the other hand, was able to adapt across data structures, similar to the findings for other tree-based or ensemble methods [24]. While CART also benefited from model flexibility, it has a tendency to underestimate uncertainty because CART-based imputation models, as implemented in MICE, are typically fitted without bootstrap resampling. This results in insufficient between-imputation variability and renders the procedure formally improper under standard MI theory, as documented in the MI literature ([16] Chapter 3.5) and also observed in our simulations. Overall, TabPFN achieved lower bias and more reliable coverage than competing methods, even at smaller sample sizes.
Several limitations of the simulation study warrant explicit consideration. First, as with any simulation study, our evaluation was restricted to a specific set of experimental conditions. In particular, we considered univariate missingness in a continuous outcome with fully observed continuous predictors, a limited set of sample sizes and predictor dimensions, and an overall missingness proportion of approximately 50%. We further considered only selected mean and missingness structures and did not vary several other potentially important design features, such as the missingness rate, mixed data types, multivariate missingness patterns, or broader distributional settings for predictors and errors. Second, the comparison was limited to a selected set of benchmark methods, primarily methods available within the MICE framework. Other relevant competitors, including approaches such as MISSRANGER, MIXGB, MISL, or random-forest-based imputation variants, were not included and may have provided additional insight into the relative performance of TabPFN-based imputation [24,25,26,49]. Accordingly, the findings should be interpreted within the scope of the simulation scenarios considered.
TabPFN provides an unsupervised prediction mode, but this mode currently yields deterministic predictions rather than drawing from a well-defined posterior predictive distribution. This may underestimate variability in subsequent analyses, potentially leading to anti-conservative inference and inflated Type I error rates, a well-known consequence of single or insufficiently stochastic imputation ([2,12] Chapter 1.5, Chapter 4). Although MI is designed to propagate uncertainty due to missingness, its validity depends critically on the adequacy of the underlying imputation models. Biased, poorly calibrated, or overly deterministic imputation mechanisms can counteract the theoretical advantages of MI, particularly in complex or low-sample settings. While TabPFN improved performance in many scenarios, it did not achieve nominal properties in all settings, particularly under the most challenging combinations of non-additive mean structure and non-additive missingness. This may reflect underestimation of the uncertainty or standard errors produced by the imputation procedure. One possible explanation is that the predictive distributions generated by the pretrained TabPFN model do not perfectly translate into calibrated between-imputation variability for a given dataset. Further methodological development is therefore required to improve robustness and theoretical foundation across a broader range of settings.
Because TabPFN is pretrained exclusively on synthetically generated datasets, evaluation based solely on synthetic simulation studies may raise concerns about potential over-optimism if the synthetic pretraining distribution overlaps substantially with the data-generating mechanisms used for evaluation. More realistic simulation strategies, such as plasmode or other semi-synthetic designs anchored in empirical data, may therefore provide further insight into the performance of TabPFN-based imputation under practically relevant conditions [50,51]. In line with Murray [52] and Oberman and Vink [42], we advocate for unified evaluations of MI methods combining controlled simulations with results from publicly available benchmark datasets, similar to TabArena in prediction research [34] or RealCause in causal inference from observational data [53], to ensure fair and transparent comparisons.

7. Conclusions

TabPFN-based MI represents a promising addition to the MI toolkit and illustrates the broader potential of foundation-model-based approaches for missing data imputation, particularly in limited-data settings where traditional imputation models struggle and expert knowledge about functional relationships is limited or unavailable. Its ability to combine modeling flexibility with generally robust coverage properties makes it a promising candidate for further application in missing data imputation.
At the same time, practical and methodological challenges remain. TabPFN is computationally more demanding than simpler MI methods such as PMM or CART. While access to GPU resources is increasingly common in research environments and may not be prohibitive relative to the overall cost of large studies, the additional computational and software complexity may nevertheless pose practical barriers for routine use in standard statistical workflows. Moreover, TabPFN’s relative novelty means it has not yet been extensively tested across diverse applied contexts. Even though adoption in the Healthcare and Life Sciences section is strongest [35], uptake in regulatory settings, where transparency, reproducibility, and extensive methodological validation are emphasized, may be more gradual [1,31].
Future research should focus on extending the approach to more general missing-data settings, improving computational efficiency to facilitate broader adoption, and systematically benchmarking TabPFN-based MI against both established MI methods and alternative foundation-model approaches, including related TabPFN-based extensions such as AutoTabPFN, which enables automatic hyperparameter tuning, and alternative tabular foundation models such as TabICLv2 [54]. More generally, the area of foundation-model-based imputation is developing rapidly, and TabPFN should be viewed as one promising example within a broader emerging class of methods.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/stats9020038/s1.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

All material to reproduce the results is available in the Supplementary Materials.

Acknowledgments

I would like to thank Stefan Boes for valuable discussions and feedback.

Conflicts of Interest

The author declares no competing interests.

Appendix A

Table A1. Description of NHEFS cohort with missing observations removed.
Table A1. Description of NHEFS cohort with missing observations removed.
Stratified by SmokingNon-QuittersQuitters
n363127
Weight change [kg] (mean (SD))2.09 (7.43)4.54 (8.68)
Weight in 1971 [kg] (mean (SD))72.30 (14.67)74.41 (14.46)
Sex = Male (%)205 (56.5)80 (63.0)
Race = White (%)304 (83.7)113 (89.0)
Age in 1971 (mean (SD))42.50 (11.95)46.06 (12.67)
Amount of education (%)
        8th grade or less58 (16.0)23 (18.1)
        College dropout22 (6.1)12 (9.4)
        College or more36 (9.9)20 (15.7)
        Highschool161 (44.4)53 (41.7)
        Highschool dropout86 (23.7)19 (15.0)
Number of cigarettes smoked per day (mean (SD))20.73 (11.26)17.86 (12.27)
Years of smoking (mean (SD))23.94 (11.99)25.49 (13.57)
In recreation, how much exercise? (%)
        Little or no exercise142 (39.1)61 (48.0)
        Moderate exercise137 (37.7)46 (36.2)
        Much exercise84 (23.1)20 (15.7)
In your usual day, how active are you? (%)
        Inactive29 (8.0)9 (7.1)
        Moderately active145 (39.9)56 (44.1)
        Very active189 (52.1)62 (48.8)
Total family income [USD] (%)
        <10003 (0.8)2 (1.6)
        1000–199916 (4.4)2 (1.6)
        10,000–14,99990 (24.8)41 (32.3)
        15,000–19,99954 (14.9)17 (13.4)
        2000–299917 (4.7)3 (2.4)
        20,000–24,99926 (7.2)6 (4.7)
        25,000+15 (4.1)7 (5.5)
        3000–399911 (3.0)7 (5.5)
        4000–499915 (4.1)3 (2.4)
        5000–599915 (4.1)3 (2.4)
        6000–699912 (3.3)3 (2.4)
        7000–999989 (24.5)33 (26.0)
Marital status (%)
        Divorced23 (6.3)6 (4.7)
        Married305 (84.0)110 (86.6)
        Never married21 (5.8)8 (6.3)
        Separated14 (3.9)3 (2.4)
Highest grade of regular school ever (mean (SD))11.23 (2.81)11.59 (2.98)
Height [cm] (mean (SD))169.76 (8.84)170.70 (8.77)
Asthma = Never (%)353 (97.2)121 (95.3)
Chronic Bronchitis/Emphysema = Never (%)342 (94.2)120 (94.5)
Tuberculosis = Never (%)360 (99.2)126 (99.2)
Heart failure = Never (%)362 (99.7)127 (100.0)
High blood pressure = Never (%)321 (88.4)105 (82.7)
Peptic ulcer = Never (%)333 (91.7)110 (86.6)
Colitis = Never (%)352 (97.0)123 (96.9)
Hepatitis = Never (%)361 (99.4)127 (100.0)
Chronic cough = Never (%)347 (95.6)124 (97.6)
Hay fever = Never (%)340 (93.7)117 (92.1)
Diabetes = Never (%)361 (99.4)126 (99.2)
Polio = Never (%)359 (98.9)125 (98.4)
Malignant tumor/growth = Never (%)355 (97.8)124 (97.6)
Nervous breakdown = Never (%)356 (98.1)127 (100.0)
Have you had 1 drink past year? = Ever (%)363 (100.0)127 (100.0)
How often do you drink? (%)
        <12 times/year42 (11.6)14 (11.0)
        1–4 times/month152 (41.9)63 (49.6)
        2–3 times/week68 (18.7)16 (12.6)
        Almost every day101 (27.8)34 (26.8)
Which do you most frequently drink? (%)
        Beer194 (53.4)58 (45.7)
        Liquor144 (39.7)59 (46.5)
        Wine25 (6.9)10 (7.9)
When you drink, how much do you drink? (mean (SD))3.28 (3.53)3.03 (2.62)
Do you eat dirt or clay, starch or other non standard food? = Never (%)359 (98.9)126 (99.2)
Use headache medication = Never (%)138 (38.0)59 (46.5)
Use other pains medication = Never (%)282 (77.7)100 (78.7)
Use weak heart medication = Never (%)357 (98.3)124 (97.6)
Use allergies medication = Never (%)348 (95.9)123 (96.9)
Use nerves medication = Never (%)321 (88.4)108 (85.0)
Use lack of pep medication = Never (%)342 (94.2)123 (96.9)
Use high blood pressure medication = Never (%)348 (95.9)119 (93.7)
Use bowel trouble medication = Never (%)325 (89.5)108 (85.0)
Use weight loss medication = Never (%)357 (98.3)126 (99.2)
Use infection medication = Never (%)311 (85.7)111 (87.4)
Serum cholesterol [mg/100 mL] (mean (SD))218.30 (45.21)228.03 (48.67)
Avg tobacco price in state of residence [USD] (mean (SD))2.17 (0.22)2.14 (0.21)
Tobacco tax in state of residence (mean (SD))1.09 (0.21)1.06 (0.21)
Figure A1. Proportion of the variation attributable to the missing data ( λ ). Values close to zero indicate that the contribution of missingness to the overall variance is negligible (perfect recovery of missing observations). Values near one indicate that uncertainty is dominated by missing data, corresponding to situations in which the observed data contain little information about the missing values.
Figure A1. Proportion of the variation attributable to the missing data ( λ ). Values close to zero indicate that the contribution of missingness to the overall variance is negligible (perfect recovery of missing observations). Values near one indicate that uncertainty is dominated by missing data, corresponding to situations in which the observed data contain little information about the missing values.
Stats 09 00038 g0a1
Figure A2. Sensitivity to the number of imputations: performance metrics remain stable when increasing m from 20 to 100.
Figure A2. Sensitivity to the number of imputations: performance metrics remain stable when increasing m from 20 to 100.
Stats 09 00038 g0a2

References

  1. Committee for Medicinal Products for Human Use (CHMP). EMA Guideline on Missing Data in Confirmatory Clinical Trials (EMA/CPMP/EWP/1776/99). 2010. Available online: https://www.ema.europa.eu/en/documents/scientific-guideline/guideline-missing-data-confirmatory-clinical-trials_en.pdf (accessed on 11 October 2025).
  2. Little, R.J.A.; Rubin, D.B. Statistical Analysis with Missing Data, 3rd ed.; John Wiley & Sons, Inc.: Hoboken, NJ, USA, 2019. [Google Scholar]
  3. Alabadla, M.; Sidi, F.; Ishak, I.; Ibrahim, H.; Affendey, L.S.; Che Ani, Z.; Jabar, M.A.; Bukar, U.A.; Devaraj, N.K.; Muda, A.S.; et al. Systematic Review of Using Machine Learning in Imputing Missing Values. IEEE Access 2022, 10, 44483–44502. [Google Scholar] [CrossRef]
  4. Li, J.; Yan, X.S.; Chaudhary, D.; Avula, V.; Mudiganti, S.; Husby, H.; Shahjouei, S.; Afshar, A.; Stewart, W.F.; Yeasin, M.; et al. Imputation of missing values for electronic health record laboratory data. npj Digit. Med. 2021, 4, 147. [Google Scholar] [CrossRef]
  5. Buczak, P.; Chen, J.J.; Pauly, M. Analyzing the Effect of Imputation on Classification Performance under MCAR and MAR Missing Mechanisms. Entropy 2023, 25, 521. [Google Scholar] [CrossRef] [PubMed]
  6. Kazijevs, M.; Samad, M.D. Deep imputation of missing values in time series health data: A review with benchmarking. J. Biomed. Inform. 2023, 144, 104440. [Google Scholar] [CrossRef] [PubMed]
  7. Liu, M.; Li, S.; Yuan, H.; Ong, M.E.H.; Ning, Y.; Xie, F.; Saffari, S.E.; Shang, Y.; Volovici, V.; Chakraborty, B.; et al. Handling missing values in healthcare data: A systematic review of deep learning-based imputation techniques. Artif. Intell. Med. 2023, 142, 102587. [Google Scholar] [CrossRef]
  8. Varol, B.; Omurlu, I.K.; Ture, M. Review of Simulation Studies Evaluating Imputation Methods in High-Dimensional Datasets. WIREs Comput. Stat. 2025, 17, e70044. [Google Scholar] [CrossRef]
  9. Thurow, M.; Dumpert, F.; Ramosaj, B.; Pauly, M. Imputing missings in official statistics for general tasks–our vote for distributional accuracy. Stat. J. IAOS 2021, 37, 1379–1390. [Google Scholar] [CrossRef]
  10. Schwerter, J.; Gurtskaia, K.; Romero, A.; Zeyer-Gliozzo, B.; Pauly, M. Evaluating tree-based imputation methods as an alternative to MICE PMM for drawing inference in empirical studies. arXiv 2024, arXiv:2401.09602. [Google Scholar] [CrossRef]
  11. Grzesiak, K.; Muller, C.; Josse, J.; Näf, J. Do we Need Dozens of Methods for Real World Missing Value Imputation? arXiv 2025, arXiv:2511.04833. [Google Scholar] [CrossRef]
  12. Rubin, D.B. Multiple Imputation for Nonresponse in Surveys; John Wiley & Sons, Inc.: Hoboken, NJ, USA, 1987. [Google Scholar]
  13. Seaman, S.R.; Bartlett, J.W.; White, I.R. Multiple imputation of missing covariates with non-linear effects and interactions: An evaluation of statistical methods. BMC Med. Res. Methodol. 2012, 12, 46. [Google Scholar] [CrossRef] [PubMed]
  14. Curnow, E.; Carpenter, J.R.; Heron, J.E.; Cornish, R.P.; Rach, S.; Didelez, V.; Langeheine, M.; Tilling, K. Multiple imputation of missing data under missing at random: Compatible imputation models are not sufficient to avoid bias if they are mis-specified. J. Clin. Epidemiol. 2023, 160, 100–109. [Google Scholar] [CrossRef] [PubMed]
  15. van Buuren, S.; Groothuis-Oudshoorn, K. mice: Multivariate Imputation by Chained Equations in R. J. Stat. Softw. 2011, 45, 1–67. [Google Scholar] [CrossRef]
  16. van Buuren, S. Flexible Imputation of Missing Data; Chapman & Hall/CRC: Boca Raton, FL, USA, 2018. [Google Scholar]
  17. Meng, X.L. Multiple-imputation inferences with uncongenial sources of input. Stat. Sci. 1994, 9, 538–558. [Google Scholar] [CrossRef]
  18. White, I.R.; Royston, P.; Wood, A.M. Multiple imputation using chained equations: Issues and guidance for practice. Stat. Med. 2010, 30, 377–399. [Google Scholar] [CrossRef] [PubMed]
  19. Breiman, L.; Friedman, J.H.; Olshen, R.A.; Stone, C.J. Classification and Regression Trees; Chapman & Hall/CRC: Boca Raton, FL, USA, 2017. [Google Scholar] [CrossRef]
  20. Breiman, L. Random Forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  21. Chen, T.; Guestrin, C. XGBoost: A Scalable Tree Boosting System. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining; ACM: New York, NY, USA, 2016; KDD ’16. [Google Scholar] [CrossRef]
  22. Burgette, L.F.; Reiter, J.P. Multiple Imputation for Missing Data via Sequential Regression Trees. Am. J. Epidemiol. 2010, 172, 1070–1076. [Google Scholar] [CrossRef]
  23. Shah, A.D.; Bartlett, J.W.; Carpenter, J.; Nicholas, O.; Hemingway, H. Comparison of Random Forest and Parametric Imputation Models for Imputing Missing Data Using MICE: A CALIBER Study. Am. J. Epidemiol. 2014, 179, 764–774. [Google Scholar] [CrossRef]
  24. Doove, L.; Van Buuren, S.; Dusseldorp, E. Recursive partitioning for missing data imputation in the presence of interaction effects. Comput. Stat. Data Anal. 2014, 72, 92–104. [Google Scholar] [CrossRef]
  25. Deng, Y.; Lumley, T. Multiple Imputation Through XGBoost. J. Comput. Graph. Stat. 2023, 33, 352–363. [Google Scholar] [CrossRef]
  26. Carpenito, T.; Manjourides, J. MISL: Multiple imputation by super learning. Stat. Methods Med. Res. 2022, 31, 1904–1915. [Google Scholar] [CrossRef]
  27. Sepin, J. Multiple imputation using multivariate adaptive regression splines. Res. Sq. 2025, 1–15. [Google Scholar] [CrossRef]
  28. van der Ploeg, T.; Austin, P.C.; Steyerberg, E.W. Modern modelling techniques are data hungry: A simulation study for predicting dichotomous endpoints. BMC Med. Res. Methodol. 2014, 14, 137. [Google Scholar] [CrossRef] [PubMed]
  29. Riley, R.D.; Ensor, J.; Snell, K.I.E.; Harrell, F.E.; Martin, G.P.; Reitsma, J.B.; Moons, K.G.M.; Collins, G.; van Smeden, M. Calculating the sample size required for developing a clinical prediction model. BMJ 2020, 368, m441. [Google Scholar] [CrossRef]
  30. Romaniuk, M.; Grzegorzewski, P. Fuzzy data imputation with DIMP and FGAIN. J. Comput. Sci. 2026, 93, 102738. [Google Scholar] [CrossRef]
  31. Little, R.J.A.; D’Agostino, R.; Cohen, M.L.; Dickersin, K.; Emerson, S.S.; Farrar, J.T.; Frangakis, C.; Hogan, J.W.; Molenberghs, G.; Murphy, S.A.; et al. The Prevention and Treatment of Missing Data in Clinical Trials. N. Engl. J. Med. 2012, 367, 1355–1360. [Google Scholar] [CrossRef]
  32. Ye, H.J.; Liu, S.Y.; Chao, W.L. A Closer Look at TabPFN v2: Understanding Its Strengths and Extending Its Capabilities. arXiv 2025, arXiv:2502.17361. [Google Scholar] [CrossRef]
  33. Hollmann, N.; Müller, S.; Purucker, L.; Krishnakumar, A.; Körfer, M.; Hoo, S.B.; Schirrmeister, R.T.; Hutter, F. Accurate predictions on small data with a tabular foundation model. Nature 2025, 637, 319–326. [Google Scholar] [CrossRef] [PubMed]
  34. Erickson, N.; Purucker, L.; Tschalzev, A.; Holzmüller, D.; Desai, P.M.; Salinas, D.; Hutter, F. TabArena: A Living Benchmark for Machine Learning on Tabular Data. arXiv 2025, arXiv:2506.16791. [Google Scholar] [CrossRef]
  35. Grinsztajn, L.; Flöge, K.; Key, O.; Birkel, F.; Jund, P.; Roof, B.; Jäger, B.; Safaric, D.; Alessi, S.; Hayler, A.; et al. TabPFN-2.5: Advancing the State of the Art in Tabular Foundation Models. arXiv 2025, arXiv:2511.08667. [Google Scholar] [CrossRef]
  36. Feitelberg, J.; Saha, D.; Choi, K.; Ahmad, Z.; Agarwal, A.; Dwivedi, R. TabImpute: Accurate and Fast Zero-Shot Missing-Data Imputation with a Pre-Trained Transformer. arXiv 2025, arXiv:2510.02625. [Google Scholar] [CrossRef]
  37. Jakobsen, J.C.; Gluud, C.; Wetterslev, J.; Winkel, P. When and how should multiple imputation be used for handling missing data in randomised clinical trials—A practical guide with flowcharts. BMC Med. Res. Methodol. 2017, 17, 162. [Google Scholar] [CrossRef]
  38. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Volume 30. [Google Scholar]
  39. Hollmann, N.; Müller, S.; Eggensperger, K.; Hutter, F. Tabpfn: A transformer that solves small tabular classification problems in a second. arXiv 2022, arXiv:2207.01848. [Google Scholar] [CrossRef]
  40. Müller, S.; Hollmann, N.; Arango, S.P.; Grabocka, J.; Hutter, F. Transformers can do bayesian inference. arXiv 2021, arXiv:2112.10510. [Google Scholar] [CrossRef]
  41. Little, R.J.A.; An, H. Robust likelihood-based analysis of multivariate data with missing values. Stat. Sin. 2004, 14, 949–968. [Google Scholar]
  42. Oberman, H.I.; Vink, G. Toward a standardized evaluation of imputation methodology. Biom. J. 2023, 66, 2200107. [Google Scholar] [CrossRef] [PubMed]
  43. Little, R.J.A. Missing-Data Adjustments in Large Surveys. J. Bus. Econ. Stat. 1988, 6, 287–296. [Google Scholar] [CrossRef]
  44. Barnard, J.; Rubin, D.B. Miscellanea. Small-sample degrees of freedom with multiple imputation. Biometrika 1999, 86, 948–955. [Google Scholar] [CrossRef]
  45. Wilson, E.B. Probable Inference, the Law of Succession, and Statistical Inference. J. Am. Stat. Assoc. 1927, 22, 209–212. [Google Scholar] [CrossRef]
  46. Hernán, M.A.; Robins, J.M. Causal Inference: What If; Chapman & Hall/CRC: Boca Raton, FL, USA, 2020. [Google Scholar]
  47. Westreich, D.; Edwards, J.K.; Cole, S.R.; Platt, R.W.; Mumford, S.L.; Schisterman, E.F. Imputation approaches for potential outcomes in causal inference. Int. J. Epidemiol. 2015, 44, 1731–1737. [Google Scholar] [CrossRef]
  48. Morris, T.P.; White, I.R.; Royston, P. Tuning multiple imputation by predictive mean matching and local residual draws. BMC Med. Res. Methodol. 2014, 14, 75. [Google Scholar] [CrossRef]
  49. Mayer, M. missRanger: Fast Imputation of Missing Values. CRAN Contrib. Packag. 2017. [Google Scholar] [CrossRef]
  50. Franklin, J.M.; Schneeweiss, S.; Polinski, J.M.; Rassen, J.A. Plasmode simulation for the evaluation of pharmacoepidemiologic methods in complex healthcare databases. Comput. Stat. Data Anal. 2014, 72, 219–226. [Google Scholar] [CrossRef] [PubMed]
  51. Schreck, N.; Slynko, A.; Saadati, M.; Benner, A. Statistical plasmode simulations–Potentials, challenges and recommendations. Stat. Med. 2024, 43, 1804–1825. [Google Scholar] [CrossRef] [PubMed]
  52. Murray, J.S. Multiple Imputation: A Review of Practical and Theoretical Findings. Stat. Sci. 2018, 33, 142–159. [Google Scholar] [CrossRef]
  53. Neal, B.; Huang, C.W.; Raghupathi, S. RealCause: Realistic Causal Inference Benchmarking. arXiv 2020, arXiv:2011.15007. [Google Scholar] [CrossRef]
  54. Qu, J.; Holzmüller, D.; Varoquaux, G.; Morvan, M.L. TabICLv2: A better, faster, scalable, and open tabular foundation model. arXiv 2026, arXiv:2602.11139. [Google Scholar] [CrossRef]
Figure 1. Bias in pooled coefficient estimates across imputation methods. Bias was calculated as the difference between each method’s pooled estimate and the true value of 10 over 250 simulation runs. Error bars indicate the 25th, 50th (median), and 75th percentiles; numbers report the mean and standard deviation in parentheses.
Figure 1. Bias in pooled coefficient estimates across imputation methods. Bias was calculated as the difference between each method’s pooled estimate and the true value of 10 over 250 simulation runs. Error bars indicate the 25th, 50th (median), and 75th percentiles; numbers report the mean and standard deviation in parentheses.
Stats 09 00038 g001
Figure 2. Standard errors (SEs) of pooled estimates across imputation methods. Error bars indicate the 25th, 50th (median), and 75th percentiles; numbers report the mean and standard deviation in parentheses. The red vertical lines denote the median standard errors from the full-data analysis at the indicated sample size.
Figure 2. Standard errors (SEs) of pooled estimates across imputation methods. Error bars indicate the 25th, 50th (median), and 75th percentiles; numbers report the mean and standard deviation in parentheses. The red vertical lines denote the median standard errors from the full-data analysis at the indicated sample size.
Stats 09 00038 g002
Figure 3. Coverage rates of 90% Wilson confidence intervals across imputation methods. Ideally, coverage should be close to the nominal 95% level. The horizontal reference corresponds to nominal 95% coverage.
Figure 3. Coverage rates of 90% Wilson confidence intervals across imputation methods. Ideally, coverage should be close to the nominal 95% level. The horizontal reference corresponds to nominal 95% coverage.
Stats 09 00038 g003
Table 1. Average computational times (in seconds) required to generate m = 20 imputations.
Table 1. Average computational times (in seconds) required to generate m = 20 imputations.
Methodnp = 10p = 100
cart1000.3241.072
cart2000.3731.281
cart4000.4641.769
pmm1000.2080.688
pmm2000.2180.708
pmm4000.2260.738
tabpfn10014.08118.071
tabpfn20014.21918.483
tabpfn40014.51019.607
Table 2. Comparison of g-computation and MI methods for estimating the ACE of smoking cessation on weight gain. MI imputes only the missing counterfactual outcomes, whereas g-computation replaces all outcomes with model predictions.
Table 2. Comparison of g-computation and MI methods for estimating the ACE of smoking cessation on weight gain. MI imputes only the missing counterfactual outcomes, whereas g-computation replaces all outcomes with model predictions.
MethodEstimateStd. ErrorLower 95%Upper 95%
pmm2.3630.8340.7283.998
cart3.4190.5812.2804.558
tabpfn2.7460.5681.6333.859
g-computation3.3800.8531.7085.053
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Sepin, J. Multiple Imputation of a Continuous Outcome with Fully Observed Predictors Using TabPFN. Stats 2026, 9, 38. https://doi.org/10.3390/stats9020038

AMA Style

Sepin J. Multiple Imputation of a Continuous Outcome with Fully Observed Predictors Using TabPFN. Stats. 2026; 9(2):38. https://doi.org/10.3390/stats9020038

Chicago/Turabian Style

Sepin, Jerome. 2026. "Multiple Imputation of a Continuous Outcome with Fully Observed Predictors Using TabPFN" Stats 9, no. 2: 38. https://doi.org/10.3390/stats9020038

APA Style

Sepin, J. (2026). Multiple Imputation of a Continuous Outcome with Fully Observed Predictors Using TabPFN. Stats, 9(2), 38. https://doi.org/10.3390/stats9020038

Article Metrics

Back to TopTop