Emergent Mind

Algorithms for Learning Sparse Additive Models with Interactions in High Dimensions

(1605.00609)
Published May 2, 2016 in cs.LG , cs.IT , math.IT , math.NA , and stat.ML

Abstract

A function $f: \mathbb{R}d \rightarrow \mathbb{R}$ is a Sparse Additive Model (SPAM), if it is of the form $f(\mathbf{x}) = \sum{l \in \mathcal{S}}\phi{l}(xl)$ where $\mathcal{S} \subset [d]$, $|\mathcal{S}| \ll d$. Assuming $\phi$'s, $\mathcal{S}$ to be unknown, there exists extensive work for estimating $f$ from its samples. In this work, we consider a generalized version of SPAMs, that also allows for the presence of a sparse number of second order interaction terms. For some $\mathcal{S}1 \subset [d], \mathcal{S}2 \subset {[d] \choose 2}$, with $|\mathcal{S}1| \ll d, |\mathcal{S}2| \ll d2$, the function $f$ is now assumed to be of the form: $\sum{p \in \mathcal{S}1}\phi{p} (xp) + \sum{(l,l{\prime}) \in \mathcal{S}2}\phi{(l,l{\prime})} (xl,x{l{\prime}})$. Assuming we have the freedom to query $f$ anywhere in its domain, we derive efficient algorithms that provably recover $\mathcal{S}1,\mathcal{S}2$ with finite sample bounds. Our analysis covers the noiseless setting where exact samples of $f$ are obtained, and also extends to the noisy setting where the queries are corrupted with noise. For the noisy setting in particular, we consider two noise models namely: i.i.d Gaussian noise and arbitrary but bounded noise. Our main methods for identification of $\mathcal{S}2$ essentially rely on estimation of sparse Hessian matrices, for which we provide two novel compressed sensing based schemes. Once $\mathcal{S}1, \mathcal{S}2$ are known, we show how the individual components $\phip$, $\phi_{(l,l{\prime})}$ can be estimated via additional queries of $f$, with uniform error bounds. Lastly, we provide simulation results on synthetic data that validate our theoretical findings.

We're not able to analyze this paper right now due to high demand.

Please check back later (sorry!).

Generate a summary of this paper on our Pro plan:

We ran into a problem analyzing this paper.

Newsletter

Get summaries of trending comp sci papers delivered straight to your inbox:

Unsubscribe anytime.