An invariant large margin nearest neighbour classifier
Mudigonda Pawan Kumar, Philip Torr and Andrew Zisserman
In: ICCV 2007, 14-20 Oct 2007, Rio de Janeiro, Brazil.
The k-nearest neighbour (kNN) rule is a simple and effective method for multi-way classification that is much used in Computer Vision. However, its performance depends heavily on the distance metric being employed. The recently proposed large margin nearest neighbour (LMNN) classifier learns a distance metric for kNN classification and thereby improves its accuracy. Learning involves optimizing a convex problem using semidefinite programming (SDP).
We extend the LMNN framework to incorporate knowledge about invariance of the data. The main contributions of our work are three fold: (i) Invariances to multivariate polynomial transformations are incorporated without explicitly adding more training data during learning - these can approximate common transformations such as rotations and affinities; (ii) the incorporation of different regularizers on the parameters being learnt; and (iii) for all these variations, we show that the distance metric can still be obtained by solving a convex SDP problem. We call the resulting formulation invariant LMNN (ILMNN) classifier.
We test our approach to learn a metric for matching (i) feature vectors from the standard Iris dataset; and (ii) faces obtained from TV video (an episode of `Buffy the Vampire Slayer'). We compare our method with the state of the art classifiers and demonstrate improvements.
|EPrint Type:||Conference or Workshop Item (Poster)|
|Project Keyword:||Project Keyword UNSPECIFIED|
|Deposited By:||Mudigonda Pawan Kumar|
|Deposited On:||21 September 2007|