- Library Home /
- Search Collections /
- Open Collections /
- Browse Collections /
- UBC Theses and Dissertations /
- Automatic massively parallel Markov chain Monte Carlo...
Open Collections
UBC Theses and Dissertations
UBC Theses and Dissertations
Automatic massively parallel Markov chain Monte Carlo with quantifiable error Biron Lattes, Miguel
Abstract
Simulated Tempering (ST) is an MCMC algorithm for complex target distributions that operates on a path between the target and an amenable reference distribution. Crucially, if the reference enables i.i.d. sampling, ST is regenerative and therefore embarrassingly parallel. However, the difficulty of tuning ST has hindered its widespread adoption. In this work, we develop a simple nonreversible ST (NRST) algorithm, a general theoretical analysis of ST, and an automated tuning procedure for ST. A core contribution that arises from the analysis is a novel performance metric—Tour Effectiveness (TE)—that controls the asymptotic variance of estimates from ST for bounded test functions. We use the TE to show that NRST dominates its reversible counterpart. We then develop an automated tuning procedure for NRST algorithms that targets the TE while minimizing computational cost. This procedure enables straightforward integration of NRST into existing probabilistic programming languages. We provide extensive experimental evidence that our tuning scheme improves the performance and robustness of NRST algorithms on a diverse set of probabilistic models. NRST can be seen as a meta-MCMC algorithm, in that an explorer Markov chain is required to make local moves within distributions in the path, while NRST orchestrates movement along the path. An ideal explorer must not break down when the distribution that it targets is altered. In our experiments, the Slice Sampler proved to be reliable on an array of models; however, it suffers from poor dimensional scaling. In contrast, gradient-based methods like Metropolis-adjusted Langevin algorithm (MALA) scale favorably with dimension. However, MALA depends critically on a step size parameter, and tuning it requires too much work to be useful for NRST. To resolve this issue we introduce autoMALA, an improved version of MALA that automatically sets its step size at each iteration based on the local geometry of the target distribution. We prove that autoMALA preserves the target measure despite continual adjustments of the step size. Our experiments demonstrate that autoMALA is competitive with related state-of-the-art MCMC methods, in terms of the number of density evaluations per effective sample, and it outperforms state-of-the-art samplers on targets with varying geometries.
Item Metadata
Title |
Automatic massively parallel Markov chain Monte Carlo with quantifiable error
|
Creator | |
Supervisor | |
Publisher |
University of British Columbia
|
Date Issued |
2024
|
Description |
Simulated Tempering (ST) is an MCMC algorithm for complex target distributions that operates on a path between the target and an amenable reference distribution. Crucially, if the reference enables i.i.d. sampling, ST is regenerative and therefore embarrassingly parallel. However, the difficulty of tuning ST has hindered its widespread adoption. In this work, we develop a simple nonreversible ST (NRST) algorithm, a general theoretical analysis of ST, and an automated tuning procedure for ST. A core contribution that arises from the analysis is a novel performance metric—Tour Effectiveness (TE)—that controls the asymptotic variance of estimates from ST for bounded test functions. We use the TE to show that NRST dominates its reversible counterpart. We then develop an automated tuning procedure for NRST algorithms that targets the TE while minimizing computational cost. This procedure enables straightforward integration of NRST into existing probabilistic programming languages. We provide extensive experimental evidence that our tuning scheme improves the performance and robustness of NRST algorithms on a diverse set of probabilistic models. NRST can be seen as a meta-MCMC algorithm, in that an explorer Markov chain is required to make local moves within distributions in the path, while NRST orchestrates movement along the path. An ideal explorer must not break down when the distribution that it targets is altered. In our experiments, the Slice Sampler proved to be reliable on an array of models; however, it suffers from poor dimensional scaling. In contrast, gradient-based methods like Metropolis-adjusted Langevin algorithm (MALA) scale favorably with dimension. However, MALA depends critically on a step size parameter, and tuning it requires too much work to be useful for NRST. To resolve this issue we introduce autoMALA, an improved version of MALA that automatically sets its step size at each iteration based on the local geometry of the target distribution. We prove that autoMALA preserves the target measure despite continual adjustments of the step size. Our experiments demonstrate that autoMALA is competitive with related state-of-the-art MCMC methods, in terms of the number of density evaluations per effective sample, and it outperforms state-of-the-art samplers on targets with varying geometries.
|
Genre | |
Type | |
Language |
eng
|
Date Available |
2024-07-17
|
Provider |
Vancouver : University of British Columbia Library
|
Rights |
Attribution-NonCommercial-NoDerivatives 4.0 International
|
DOI |
10.14288/1.0444168
|
URI | |
Degree | |
Program | |
Affiliation | |
Degree Grantor |
University of British Columbia
|
Graduation Date |
2024-11
|
Campus | |
Scholarly Level |
Graduate
|
Rights URI | |
Aggregated Source Repository |
DSpace
|
Item Media
Item Citations and Data
Rights
Attribution-NonCommercial-NoDerivatives 4.0 International