A Fast Approximation Scheme for Low-Dimensional $k$-Means
(1708.07381)Abstract
We consider the popular $k$-means problem in $d$-dimensional Euclidean space. Recently Friggstad, Rezapour, Salavatipour [FOCS'16] and Cohen-Addad, Klein, Mathieu [FOCS'16] showed that the standard local search algorithm yields a $(1+\epsilon)$-approximation in time $(n \cdot k){1/\epsilon{O(d)}}$, giving the first polynomial-time approximation scheme for the problem in low-dimensional Euclidean space. While local search achieves optimal approximation guarantees, it is not competitive with the state-of-the-art heuristics such as the famous $k$-means++ and $D2$-sampling algorithms. In this paper, we aim at bridging the gap between theory and practice by giving a $(1+\epsilon)$-approximation algorithm for low-dimensional $k$-means running in time $n \cdot k \cdot (\log n){(d\epsilon{-1}){O(d)}}$, and so matching the running time of the $k$-means++ and $D2$-sampling heuristics up to polylogarithmic factors. We speed-up the local search approach by making a non-standard use of randomized dissections that allows to find the best local move efficiently using a quite simple dynamic program. We hope that our techniques could help design better local search heuristics for geometric problems. We note that the doubly exponential dependency on $d$ is necessary as $k$-means is APX-hard in dimension $d = \omega(\log n)$.
We're not able to analyze this paper right now due to high demand.
Please check back later (sorry!).
Generate a summary of this paper on our Pro plan:
We ran into a problem analyzing this paper.