The Block-Poisson Estimator for Optimally Tuned Exact Subsampling MCMC
- American Statistical Association
- Publication Type:
- Journal Article
- Journal of Computational and Graphical Statistics, 2021, 30, (4), pp. 877-888
- Issue Date:
Copyright Clearance Process
- Recently Added
- In Progress
- Closed Access
This item is closed access and not available.
Speeding up Markov chain Monte Carlo (MCMC) for datasets with many observations by data subsampling has recently received considerable attention. A pseudo-marginal MCMC method is proposed that estimates the likelihood by data subsampling using a block-Poisson estimator. The estimator is a product of Poisson estimators, allowing us to update a single block of subsample indicators in each MCMC iteration so that a desired correlation is achieved between the logs of successive likelihood estimates. This is important since pseudo-marginal MCMC with positively correlated likelihood estimates can use substantially smaller subsamples without adversely affecting the sampling efficiency. The block-Poisson estimator is unbiased but not necessarily positive, so the algorithm runs the MCMC on the absolute value of the likelihood estimator and uses an importance sampling correction to obtain consistent estimates of the posterior mean of any function of the parameters. Our article derives guidelines to select the optimal tuning parameters for our method and shows that it compares very favorably to regular MCMC without subsampling, and to two other recently proposed exact subsampling approaches in the literature. Supplementary materials for this article are available online.
Please use this identifier to cite or link to this item: