Abstract
This paper proposes a Hessian-free Newton method for solving large-scale convex functions with an L1 regularization term. These problems arise in supervised machine learning models in which it is important to seek a sparse parameter vector. The proposed method operates in a batch setting, which is well suited for parallel computing environments, and employs sub-sampled Hessian information to accelerate progress of the iteration. The method consists of two phases, an active-set prediction phase that employs first-order and second-order information, and subspace phase that performs a Newton-like step. Numerical results on a speech recognition problem illustrate the practical behavior of the method.
Original language | English (US) |
---|---|
Title of host publication | 2012 IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP 2012 - Proceedings |
Pages | 5237-5240 |
Number of pages | 4 |
DOIs | |
State | Published - Oct 23 2012 |
Event | 2012 IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP 2012 - Kyoto, Japan Duration: Mar 25 2012 → Mar 30 2012 |
Other
Other | 2012 IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP 2012 |
---|---|
Country | Japan |
City | Kyoto |
Period | 3/25/12 → 3/30/12 |
Keywords
- Hessian-Free Newton
- Iterative Shrinkage
- L1 Regularization
- Logistic Regression
- Newton Method
ASJC Scopus subject areas
- Software
- Signal Processing
- Electrical and Electronic Engineering