PASCAL - Pattern Analysis, Statistical Modelling and Computational Learning

Robust Bounds for Classification via Selective Sampling
Nicolò Cesa-Bianchi, Claudio Gentile and Francesco Orabona
In: ICML 2009, 14-18 June 2009, Montreal, Quebec.


We introduce a new algorithm for binary classification in the selective sampling protocol. Our algorithm uses Regularized Least Squares (RLS) as base classifier, and for this reason it can be efficiently run in any RKHS. Unlike previous margin-based semi-supervised algorithms, our sampling condition hinges on a simultaneous upper bound on bias and variance of the RLS estimate under a simple linear label noise model. This fact allows us to prove performance bounds that hold for an arbitrary sequence of in- stances. In particular, we show that our sampling strategy approximates the margin of the Bayes optimal classifier to any desired accuracy ε by asking O(d/ε^2) queries (in the RKHS case d is replaced by a suitable spectral quantity). While these are the standard rates in the fully supervised i.i.d. case, the best previously known result in our harder setting was O(d^3 /ε^4). Preliminary experiments show that some of our algorithms also exhibit a good practical performance.

PDF - Requires Adobe Acrobat Reader or other PDF viewer.
EPrint Type:Conference or Workshop Item (Oral)
Project Keyword:Project Keyword UNSPECIFIED
Subjects:Learning/Statistics & Optimisation
Theory & Algorithms
ID Code:5725
Deposited By:Francesco Orabona
Deposited On:08 March 2010