PASCAL - Pattern Analysis, Statistical Modelling and Computational Learning

A Stochastic Quasi-Newton Method for Online Convex Optimization
Nicol Schraudolph, Jin Yu and Simon Guenter
In: Proceedings of the 11th International Conference on Artificial Intelligence and Statistics (AIstats) (2007) Society for Artificial Intelligence and Statistics , San Juan, Puerto Rico , pp. 433-440.


We develop stochastic variants of the well-known BFGS quasi-Newton optimization method, in both full and memory-limited (LBFGS) forms, for online optimization of convex functions. The resulting algorithm performs comparably to a well-tuned natural gradient descent but is scalable to very high-dimensional problems. On standard benchmarks in natural language processing, it asymptotically outperforms previous stochastic gradient methods for parameter estimation in conditional random fields. We are working on analyzing the convergence of online (L)BFGS, and extending it to non-convex optimization problems.

PDF - Requires Adobe Acrobat Reader or other PDF viewer.
EPrint Type:Book Section
Project Keyword:Project Keyword UNSPECIFIED
Subjects:Learning/Statistics & Optimisation
ID Code:3992
Deposited By:S V N Vishwanathan
Deposited On:25 February 2008