Abstract
We propose a stochastic gradient Markov chain Monte Carlo (SG-MCMC) algorithm for scalable inference in mixed-membership stochastic blockmodels (MMSB). Our algorithm is based on the stochastic gradient Riemannian Langevin sampler and achieves both faster speed and higher accuracy at every iteration than the current state-of-the-art algorithm based on stochastic variational inference. In addition we develop an approximation that can handle models that entertain a very large number of communities. The experimental results show that SG-MCMC strictly dominates competing algorithms in all cases.
Original language | English (US) |
---|---|
Pages | 723-731 |
Number of pages | 9 |
State | Published - Jan 1 2016 |
Event | 19th International Conference on Artificial Intelligence and Statistics, AISTATS 2016 - Cadiz, Spain Duration: May 9 2016 → May 11 2016 |
Conference
Conference | 19th International Conference on Artificial Intelligence and Statistics, AISTATS 2016 |
---|---|
Country/Territory | Spain |
City | Cadiz |
Period | 5/9/16 → 5/11/16 |
All Science Journal Classification (ASJC) codes
- Artificial Intelligence
- Statistics and Probability