Papers
Topics
Authors
Recent
Detailed Answer
Quick Answer
Concise responses based on abstracts only
Detailed Answer
Well-researched responses based on abstracts and relevant paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses
Gemini 2.5 Flash
Gemini 2.5 Flash 65 tok/s
Gemini 2.5 Pro 47 tok/s Pro
GPT-5 Medium 39 tok/s Pro
GPT-5 High 32 tok/s Pro
GPT-4o 97 tok/s Pro
Kimi K2 164 tok/s Pro
GPT OSS 120B 466 tok/s Pro
Claude Sonnet 4 38 tok/s Pro
2000 character limit reached

Exponentially Improved Dimensionality Reduction for $\ell_1$: Subspace Embeddings and Independence Testing (2104.12946v3)

Published 27 Apr 2021 in cs.DS

Abstract: Despite many applications, dimensionality reduction in the $\ell_1$-norm is much less understood than in the Euclidean norm. We give two new oblivious dimensionality reduction techniques for the $\ell_1$-norm which improve exponentially over prior ones: 1. We design a distribution over random matrices $S \in \mathbb{R}{r \times n}$, where $r = 2{\tilde O(d/(\varepsilon \delta))}$, such that given any matrix $A \in \mathbb{R}{n \times d}$, with probability at least $1-\delta$, simultaneously for all $x$, $|SAx|_1 = (1 \pm \varepsilon)|Ax|_1$. Note that $S$ is linear, does not depend on $A$, and maps $\ell_1$ into $\ell_1$. Our distribution provides an exponential improvement on the previous best known map of Wang and Woodruff (SODA, 2019), which required $r = 2{2{\Omega(d)}}$, even for constant $\varepsilon$ and $\delta$. Our bound is optimal, up to a polynomial factor in the exponent, given a known $2{\sqrt d}$ lower bound for constant $\varepsilon$ and $\delta$. 2. We design a distribution over matrices $S \in \mathbb{R}{k \times n}$, where $k = 2{O(q2)}(\varepsilon{-1} q \log d){O(q)}$, such that given any $q$-mode tensor $A \in (\mathbb{R}{d}){\otimes q}$, one can estimate the entrywise $\ell_1$-norm $|A|_1$ from $S(A)$. Moreover, $S = S1 \otimes S2 \otimes \cdots \otimes Sq$ and so given vectors $u_1, \ldots, u_q \in \mathbb{R}d$, one can compute $S(u_1 \otimes u_2 \otimes \cdots \otimes u_q)$ in time $2{O(q2)}(\varepsilon{-1} q \log d){O(q)}$, which is much faster than the $dq$ time required to form $u_1 \otimes u_2 \otimes \cdots \otimes u_q$. Our linear map gives a streaming algorithm for independence testing using space $2{O(q2)}(\varepsilon{-1} q \log d){O(q)}$, improving the previous doubly exponential $(\varepsilon{-1} \log d){q{O(q)}}$ space bound of Braverman and Ostrovsky (STOC, 2010).

Citations (9)
List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Summary

We haven't generated a summary for this paper yet.

Dice Question Streamline Icon: https://streamlinehq.com

Follow-Up Questions

We haven't generated follow-up questions for this paper yet.

Youtube Logo Streamline Icon: https://streamlinehq.com

Don't miss out on important new AI/ML research

See which papers are being discussed right now on X, Reddit, and more:

“Emergent Mind helps me see which AI papers have caught fire online.”

Philip

Philip

Creator, AI Explained on YouTube