PASCAL - Pattern Analysis, Statistical Modelling and Computational Learning

A Choice Model with Infinitely Many Latent Features
Dilan Gorur, Frank Jäkel and Carl Edward Rasmussen
In: Proceedings of the Twenty-Third International Conference on Machine Learning (2006) ACM , Pennsylvania, USA , pp. 361-368. ISBN 1-59593-383-2

Abstract

Elimination by aspects (EBA) is a probabilistic choice model describing how humans decide between several options. The options from which the choice is made are characterized by binary features and associated weights. For instance, when choosing which mobile phone to buy the features to consider may be: long lasting battery, color screen, etc. Existing methods for inferring the parameters of the model assume pre-specified features. However, the features that lead to the observed choices are not always known. Here, we present a non-parametric Bayesian model to infer the features of the options and the corresponding weights from choice data. We use the Indian buffet process (IBP) as a prior over the features. Inference using Markov chain Monte Carlo (MCMC) in conjugate IBP models has been previously described. The main contribution of this paper is an MCMC algorithm for the EBA model that can also be used in inference for other non-conjugate IBP models---this may broaden the use of IBP priors considerably.

PDF - Requires Adobe Acrobat Reader or other PDF viewer.
EPrint Type:Book Section
Project Keyword:Project Keyword UNSPECIFIED
Subjects:Learning/Statistics & Optimisation
Theory & Algorithms
ID Code:2384
Deposited By:Dilan Gorur
Deposited On:22 November 2006