- Bayesian Anal.
- Advance publication (2018), 26 pages.
Bayesian Bootstraps for Massive Data
In this article, we present data-subsetting algorithms that allow for the approximate and scalable implementation of the Bayesian bootstrap. They are analogous to two existing algorithms in the frequentist literature: the bag of little bootstraps (Kleiner et al., 2014) and the subsampled double bootstrap (Sengupta et al., 2016). Our algorithms have appealing theoretical and computational properties that are comparable to those of their frequentist counterparts. Additionally, we provide a strategy for performing lossless inference for a class of functionals of the Bayesian bootstrap and briefly introduce extensions to the Dirichlet Process.
Bayesian Anal., Advance publication (2018), 26 pages.
First available in Project Euclid: 10 May 2019
Permanent link to this document
Digital Object Identifier
Barrientos, Andrés F.; Peña, Víctor. Bayesian Bootstraps for Massive Data. Bayesian Anal., advance publication, 10 May 2019. doi:10.1214/19-BA1155. https://projecteuclid.org/euclid.ba/1557475224
- Supplementary material: Bayesian Bootstraps for Massive Data. The supplementary material has 6 sections: the first provides theoretical results for the processes proposed in Sections 2.1, 2.2, and 2.3; the second has a figure which details the Monte Carlo algorithm for performing lossless inference for the class of functionals described in Section 2.3; the third contains a scheme for lossless simulation for the example in Section 2.3 in Chamberlain and Imbens (2003); the fourth part explains how to perform lossless inference for the Dirichlet-Multinomial process; the fifth includes a table with relative and absolute errors related to the linear regression coefficients estimated from the OPM-2011 dataset in Section 4.1; Finally, the sixth assesses the performance of the BLBB, SDBB, ANS, and AN approximating coefficients of a quantile regression fitted to the OPM-2011 dataset.