MAD-Bayes: MAP-based Asymptotic Derivations from Bayes

You are here: Home / Submitted Papers / 2015 / MAD-Bayes: MAP-based Asymptotic Derivations from Bayes

Abstract

The classical mixture of Gaussians model is related to K-means via small-variance asymptotics: as the covariances of the Gaus- sians tend to zero, the negative log-likelihood of the mixture of Gaussians model ap- proaches the K-means objective, and the EM algorithm approaches the K-means algo- rithm. Kulis & Jordan (2012) used this ob- servation to obtain a novel K-means-like al- gorithm from a Gibbs sampler for the Dirich- let process (DP) mixture. We instead con- sider applying small-variance asymptotics di- rectly to the posterior in Bayesian nonpara- metric models. This framework is indepen- dent of any specific Bayesian inference algo- rithm, and it has the major advantage that it generalizes immediately to a range of models beyond the DP mixture. To illustrate, we ap- ply our framework to the feature learning set- ting, where the beta process and Indian buf- fet process provide an appropriate Bayesian nonparametric prior. We obtain a novel ob- jective function that goes beyond clustering to learn (and penalize new) groupings for which we relax the mutual exclusivity and exhaustivity assumptions of clustering. We demonstrate several other algorithms, all of which are scalable and simple to implement. Empirical results demonstrate the benefits of the new framework.

Author

Tamara Broderick, Brian Kulis, Michael I. Jordan

Journal

Proceedings of the 30th International Conference on Ma-chine Learning,

Paper Publication Date

2013

Paper Type

Astroinformatics