Emergent Mind

Hardness of Bounded Distance Decoding on Lattices in $\ell_p$ Norms

(2003.07903)
Published Mar 17, 2020 in cs.CC and cs.DS

Abstract

$ \newcommand{\Z}{\mathbb{Z}} \newcommand{\eps}{\varepsilon} \newcommand{\cc}[1]{\mathsf{#1}} \newcommand{\NP}{\cc{NP}} \newcommand{\problem}[1]{\mathrm{#1}} \newcommand{\BDD}{\problem{BDD}} $Bounded Distance Decoding $\BDD{p,\alpha}$ is the problem of decoding a lattice when the target point is promised to be within an $\alpha$ factor of the minimum distance of the lattice, in the $\ell{p}$ norm. We prove that $\BDD{p, \alpha}$ is $\NP$-hard under randomized reductions where $\alpha \to 1/2$ as $p \to \infty$ (and for $\alpha=1/2$ when $p=\infty$), thereby showing the hardness of decoding for distances approaching the unique-decoding radius for large $p$. We also show fine-grained hardness for $\BDD{p,\alpha}$. For example, we prove that for all $p \in [1,\infty) \setminus 2\Z$ and constants $C > 1, \eps > 0$, there is no $2{(1-\eps)n/C}$-time algorithm for $\BDD{p,\alpha}$ for some constant $\alpha$ (which approaches $1/2$ as $p \to \infty$), assuming the randomized Strong Exponential Time Hypothesis (SETH). Moreover, essentially all of our results also hold (under analogous non-uniform assumptions) for $\BDD$ with preprocessing, in which unbounded precomputation can be applied to the lattice before the target is available. Compared to prior work on the hardness of $\BDD{p,\alpha}$ by Liu, Lyubashevsky, and Micciancio (APPROX-RANDOM 2008), our results improve the values of $\alpha$ for which the problem is known to be $\NP$-hard for all $p > p1 \approx 4.2773$, and give the very first fine-grained hardness for $\BDD$ (in any norm). Our reductions rely on a special family of "locally dense" lattices in $\ell{p}$ norms, which we construct by modifying the integer-lattice sparsification technique of Aggarwal and Stephens-Davidowitz (STOC 2018).

We're not able to analyze this paper right now due to high demand.

Please check back later (sorry!).

Generate a summary of this paper on our Pro plan:

We ran into a problem analyzing this paper.

Newsletter

Get summaries of trending comp sci papers delivered straight to your inbox:

Unsubscribe anytime.