PASCAL - Pattern Analysis, Statistical Modelling and Computational Learning

Variational Learning of Inducing Variables in Sparse Gaussian Processes
Michalis Titsias
In: Twelfth International Conference on Artificial Intelligence and Statistics, April 16-18, 2009, Clearwater Beach, Florida USA.


Sparse Gaussian process methods that use inducing variables require the selection of the inducing inputs and the kernel hyperparameters. We introduce a variational formulation for sparse approximations that jointly infers the inducing inputs and the kernel hyperparameters by maximizing a lower bound of the true log marginal likelihood. The key property of this formulation is that the inducing inputs are defined to be variational parameters which are selected by minimizing the Kullback-Leibler divergence between the variational distribution and the exact posterior distribution over the latent function values. We apply this technique to regression and we compare it with other approaches in the literature.

PDF - Requires Adobe Acrobat Reader or other PDF viewer.
EPrint Type:Conference or Workshop Item (Poster)
Project Keyword:Project Keyword UNSPECIFIED
Subjects:Learning/Statistics & Optimisation
Theory & Algorithms
ID Code:6353
Deposited By:Michalis Titsias
Deposited On:08 March 2010