Learning Continuous Latent Variable Models with Bregman Divergences
Authors: Shaojun Wang and Dale Schuurmans .
Source: Lecture Notes in Artificial Intelligence Vol. 2842, 2003, 190 - 204.
Abstract. We present a class of unsupervised statistical learning algorithms that are formulated in terms of minimizing Bregman divergences- a family of generalized entropy measures defined by convex functions. We obtain novel training algorithms that extract hidden latent structure by minimizing a Bregman divergence on training data, subject to a set of non-linear constraints which consider hidden variables. An alternating minimization procedure with nested iterative scaling is proposed to find feasible solutions for the resulting constrained optimization problem. The convergence of this algorithm along with its information geometric properties are characterized. Index Terms - statistical machine learning, unsupervised learning, Bregman divergence, information geometry, alternating minimization, forward projection, backward projection, iterative scaling.
©Copyright 2003 Springer