Next Article in Journal
An Analysis of Implied Volatility, Sensitivity, and Calibration of the Kennedy Model
Previous Article in Journal
𝒜-Joint Numerical Radius Bounds for Operator Pairs in Semi-Hilbert Spaces with Applications
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
This is an early access version, the complete PDF, HTML, and XML versions will be available soon.
Article

Asymptotic Analysis of the Bias–Variance Trade-Off in Subsampling Metropolis–Hastings

Department of Statistics and Data Science, School of Economics, Xiamen University, Xiamen 361005, China
Mathematics 2025, 13(21), 3395; https://doi.org/10.3390/math13213395 (registering DOI)
Submission received: 20 August 2025 / Revised: 12 October 2025 / Accepted: 21 October 2025 / Published: 24 October 2025

Abstract

Markov chain Monte Carlo (MCMC) methods are fundamental to Bayesian inference but are often computationally prohibitive for large datasets, as the full likelihood must be evaluated at each iteration. Subsampling-based approximate Metropolis–Hastings (MH) algorithms offer a popular alternative, trading a manageable bias for a significant reduction in per-iteration cost. While this bias–variance trade-off is empirically understood, a formal theoretical framework for its optimization has been lacking. Our work establishes such a framework by bounding the mean squared error (MSE) as a function of the subsample size (m), the data size (n), and the number of epochs (E). This analysis reveals two optimal asymptotic scaling laws: the optimal subsample size is m=O(E1/2), leading to a minimal MSE that scales as MSE=O(E1/2). Furthermore, leveraging the large-sample asymptotic properties of the posterior, we show that when augmented with a control variate, the approximate MH algorithm can be asymptotically more efficient than the standard MH method under ideal conditions. Experimentally, we first validate the two optimal asymptotic scaling laws. We then use Bayesian logistic regression and Softmax classification models to highlight a key difference in convergence behavior: the exact algorithm starts with a high MSE that gradually decreases as the number of epochs increases. In contrast, the approximate algorithm with a practical control variate maintains a consistently low MSE that is largely insensitive to the number of epochs.
Keywords: approximate MCMC; bias–variance trade-off; scaling laws; subsampling; variance reduction approximate MCMC; bias–variance trade-off; scaling laws; subsampling; variance reduction

Share and Cite

MDPI and ACS Style

Liu, S. Asymptotic Analysis of the Bias–Variance Trade-Off in Subsampling Metropolis–Hastings. Mathematics 2025, 13, 3395. https://doi.org/10.3390/math13213395

AMA Style

Liu S. Asymptotic Analysis of the Bias–Variance Trade-Off in Subsampling Metropolis–Hastings. Mathematics. 2025; 13(21):3395. https://doi.org/10.3390/math13213395

Chicago/Turabian Style

Liu, Shuang. 2025. "Asymptotic Analysis of the Bias–Variance Trade-Off in Subsampling Metropolis–Hastings" Mathematics 13, no. 21: 3395. https://doi.org/10.3390/math13213395

APA Style

Liu, S. (2025). Asymptotic Analysis of the Bias–Variance Trade-Off in Subsampling Metropolis–Hastings. Mathematics, 13(21), 3395. https://doi.org/10.3390/math13213395

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop