PASCAL - Pattern Analysis, Statistical Modelling and Computational Learning

A Choice Model with Infinitely Many Latent Features
Dilan Gorur, Frank Jäkel and Carl Edward Rasmussen
In: 23rd International Conference on Machine Learning (ICML 2006), 7-11 August 2005, Bonn, Germany.


Elimination by aspects (EBA) is a probabilistic choice model describing how humans decide between several options. The options from which the choice is made are characterized by binary features and associated weights. For instance, when choosing which mobile phone to buy the features to consider may be: long lasting battery, color screen, etc. Existing methods for inferring the parameters of the model assume pre-specified features. However, the features that lead to the observed choices are not always known. Here, we present a non-parametric Bayesian model to infer the features of the options and the corresponding weights from choice data. We use the Indian buffet process (IBP) as a prior over the features. Inference using Markov chain Monte Carlo (MCMC) in conjugate IBP models has been previously described. The main contribution of this paper is an MCMC algorithm for the EBA model that can also be used in inference for other non-conjugate IBP models---this may broaden the use of IBP priors considerably.

EPrint Type:Conference or Workshop Item (Oral)
Project Keyword:Project Keyword UNSPECIFIED
Subjects:Computational, Information-Theoretic Learning with Statistics
ID Code:2633
Deposited By:Carl Edward Rasmussen
Deposited On:22 November 2006