Large-Scale Stochastic Sampling from the Probability Simplex

Baker, Jack and Fearnhead, Paul and Fox, Emily B and Nemeth, Christopher (2018) Large-Scale Stochastic Sampling from the Probability Simplex. In: 32nd Neural Information Processing Systems Conference (NIPS 2018), 2018-12-032018-12-08, Palais des Congrès de Montréal.

[img]
Text (1806.07137v1)
1806.07137v1.pdf - Accepted Version
Available under License Creative Commons Attribution-NonCommercial.

Download (460kB)

Abstract

Stochastic gradient Markov chain Monte Carlo (SGMCMC) has become a popular method for scalable Bayesian inference. These methods are based on sampling a discrete-time approximation to a continuous time process, such as the Langevin diffusion. When applied to distributions defined on a constrained space, such as the simplex, the time-discretisation error can dominate when we are near the boundary of the space. We demonstrate that while current SGMCMC methods for the simplex perform well in certain cases, they struggle with sparse simplex spaces; when many of the components are close to zero. However, most popular large-scale applications of Bayesian inference on simplex spaces, such as network or topic models, are sparse. We argue that this poor performance is due to the biases of SGMCMC caused by the discretization error. To get around this, we propose the stochastic CIR process, which removes all discretization error and we prove that samples from the stochastic CIR process are asymptotically unbiased. Use of the stochastic CIR process within a SGMCMC algorithm is shown to give substantially better performance for a topic model and a Dirichlet process mixture model than existing SGMCMC approaches.

Item Type:
Contribution to Conference (Paper)
Journal or Publication Title:
32nd Neural Information Processing Systems Conference (NIPS 2018)
Subjects:
ID Code:
136201
Deposited By:
Deposited On:
19 Aug 2019 14:35
Refereed?:
Yes
Published?:
Published
Last Modified:
25 Sep 2020 00:04