Emergent Mind

Learning Sums of Independent Random Variables with Sparse Collective Support

(1807.07013)
Published Jul 18, 2018 in cs.DS , cs.LG , math.ST , and stat.TH

Abstract

We study the learnability of sums of independent integer random variables given a bound on the size of the union of their supports. For $\mathcal{A} \subset \mathbf{Z}{+}$, a sum of independent random variables with collective support $\mathcal{A}$} (called an $\mathcal{A}$-sum in this paper) is a distribution $\mathbf{S} = \mathbf{X}1 + \cdots + \mathbf{X}N$ where the $\mathbf{X}i$'s are mutually independent (but not necessarily identically distributed) integer random variables with $\cupi \mathsf{supp}(\mathbf{X}i) \subseteq \mathcal{A}.$ We give two main algorithmic results for learning such distributions: 1. For the case $| \mathcal{A} | = 3$, we give an algorithm for learning $\mathcal{A}$-sums to accuracy $\epsilon$ that uses $\mathsf{poly}(1/\epsilon)$ samples and runs in time $\mathsf{poly}(1/\epsilon)$, independent of $N$ and of the elements of $\mathcal{A}$. 2. For an arbitrary constant $k \geq 4$, if $\mathcal{A} = { a1,...,ak}$ with $0 \leq a1 < ... < ak$, we give an algorithm that uses $\mathsf{poly}(1/\epsilon) \cdot \log \log ak$ samples (independent of $N$) and runs in time $\mathsf{poly}(1/\epsilon, \log ak).$ We prove an essentially matching lower bound: if $|\mathcal{A}| = 4$, then any algorithm must use $\Omega(\log \log a_4) $ samples even for learning to constant accuracy. We also give similar-in-spirit (but quantitatively very different) algorithmic results, and essentially matching lower bounds, for the case in which $\mathcal{A}$ is not known to the learner.

We're not able to analyze this paper right now due to high demand.

Please check back later (sorry!).

Generate a summary of this paper on our Pro plan:

We ran into a problem analyzing this paper.

Newsletter

Get summaries of trending comp sci papers delivered straight to your inbox:

Unsubscribe anytime.