Emergent Mind

Extractor-Based Time-Space Lower Bounds for Learning

(1708.02639)
Published Aug 8, 2017 in cs.LG and cs.CC

Abstract

A matrix $M: A \times X \rightarrow {-1,1}$ corresponds to the following learning problem: An unknown element $x \in X$ is chosen uniformly at random. A learner tries to learn $x$ from a stream of samples, $(a1, b1), (a2, b2) \ldots$, where for every $i$, $ai \in A$ is chosen uniformly at random and $bi = M(a_i,x)$. Assume that $k,\ell, r$ are such that any submatrix of $M$ of at least $2{-k} \cdot |A|$ rows and at least $2{-\ell} \cdot |X|$ columns, has a bias of at most $2{-r}$. We show that any learning algorithm for the learning problem corresponding to $M$ requires either a memory of size at least $\Omega\left(k \cdot \ell \right)$, or at least $2{\Omega(r)}$ samples. The result holds even if the learner has an exponentially small success probability (of $2{-\Omega(r)}$). In particular, this shows that for a large class of learning problems, any learning algorithm requires either a memory of size at least $\Omega\left((\log |X|) \cdot (\log |A|)\right)$ or an exponential number of samples, achieving a tight $\Omega\left((\log |X|) \cdot (\log |A|)\right)$ lower bound on the size of the memory, rather than a bound of $\Omega\left(\min\left{(\log |X|)2,(\log |A|)2\right}\right)$ obtained in previous works [R17,MM17b]. Moreover, our result implies all previous memory-samples lower bounds, as well as a number of new applications. Our proof builds on [R17] that gave a general technique for proving memory-samples lower bounds.

We're not able to analyze this paper right now due to high demand.

Please check back later (sorry!).

Generate a summary of this paper on our Pro plan:

We ran into a problem analyzing this paper.

Newsletter

Get summaries of trending comp sci papers delivered straight to your inbox:

Unsubscribe anytime.