Emergent Mind

Fast Regression for Structured Inputs

(2203.07557)
Published Mar 14, 2022 in cs.DS

Abstract

We study the $\ellp$ regression problem, which requires finding $\mathbf{x}\in\mathbb R{d}$ that minimizes $|\mathbf{A}\mathbf{x}-\mathbf{b}|p$ for a matrix $\mathbf{A}\in\mathbb R{n \times d}$ and response vector $\mathbf{b}\in\mathbb R{n}$. There has been recent interest in developing subsampling methods for this problem that can outperform standard techniques when $n$ is very large. However, all known subsampling approaches have run time that depends exponentially on $p$, typically, $d{\mathcal{O}(p)}$, which can be prohibitively expensive. We improve on this work by showing that for a large class of common \emph{structured matrices}, such as combinations of low-rank matrices, sparse matrices, and Vandermonde matrices, there are subsampling based methods for $\ellp$ regression that depend polynomially on $p$. For example, we give an algorithm for $\ellp$ regression on Vandermonde matrices that runs in time $\mathcal{O}(n\log3 n+(dp2){0.5+\omega}\cdot\text{polylog}\,n)$, where $\omega$ is the exponent of matrix multiplication. The polynomial dependence on $p$ crucially allows our algorithms to extend naturally to efficient algorithms for $\ell\infty$ regression, via approximation of $\ell\infty$ by $\ell{\mathcal{O}(\log n)}$. Of practical interest, we also develop a new subsampling algorithm for $\ellp$ regression for arbitrary matrices, which is simpler than previous approaches for $p \ge 4$.

We're not able to analyze this paper right now due to high demand.

Please check back later (sorry!).

Generate a summary of this paper on our Pro plan:

We ran into a problem analyzing this paper.

Newsletter

Get summaries of trending comp sci papers delivered straight to your inbox:

Unsubscribe anytime.