Emergent Mind

Leverage Score Sampling for Faster Accelerated Regression and ERM

(1711.08426)
Published Nov 22, 2017 in stat.ML , cs.LG , and math.OC

Abstract

Given a matrix $\mathbf{A}\in\mathbb{R}{n\times d}$ and a vector $b \in\mathbb{R}{d}$, we show how to compute an $\epsilon$-approximate solution to the regression problem $ \min{x\in\mathbb{R}{d}}\frac{1}{2} |\mathbf{A} x - b|{2}{2} $ in time $ \tilde{O} ((n+\sqrt{d\cdot\kappa{\text{sum}}})\cdot s\cdot\log\epsilon{-1}) $ where $\kappa{\text{sum}}=\mathrm{tr}\left(\mathbf{A}{\top}\mathbf{A}\right)/\lambda_{\min}(\mathbf{A}{T}\mathbf{A})$ and $s$ is the maximum number of non-zero entries in a row of $\mathbf{A}$. Our algorithm improves upon the previous best running time of $ \tilde{O} ((n+\sqrt{n \cdot\kappa_{\text{sum}}})\cdot s\cdot\log\epsilon{-1})$. We achieve our result through a careful combination of leverage score sampling techniques, proximal point methods, and accelerated coordinate descent. Our method not only matches the performance of previous methods, but further improves whenever leverage scores of rows are small (up to polylogarithmic factors). We also provide a non-linear generalization of these results that improves the running time for solving a broader class of ERM problems.

We're not able to analyze this paper right now due to high demand.

Please check back later (sorry!).

Generate a summary of this paper on our Pro plan:

We ran into a problem analyzing this paper.

Newsletter

Get summaries of trending comp sci papers delivered straight to your inbox:

Unsubscribe anytime.