A Sub-Quadratic Time Algorithm for Robust Sparse Mean Estimation (2403.04726v1)
Abstract: We study the algorithmic problem of sparse mean estimation in the presence of adversarial outliers. Specifically, the algorithm observes a \emph{corrupted} set of samples from $\mathcal{N}(\mu,\mathbf{I}_d)$, where the unknown mean $\mu \in \mathbb{R}d$ is constrained to be $k$-sparse. A series of prior works has developed efficient algorithms for robust sparse mean estimation with sample complexity $\mathrm{poly}(k,\log d, 1/\epsilon)$ and runtime $d2 \mathrm{poly}(k,\log d,1/\epsilon)$, where $\epsilon$ is the fraction of contamination. In particular, the fastest runtime of existing algorithms is quadratic ($\Omega(d2)$), which can be prohibitive in high dimensions. This quadratic barrier in the runtime stems from the reliance of these algorithms on the sample covariance matrix, which is of size $d2$. Our main contribution is an algorithm for robust sparse mean estimation which runs in \emph{subquadratic} time using $\mathrm{poly}(k,\log d,1/\epsilon)$ samples. We also provide analogous results for robust sparse PCA. Our results build on algorithmic advances in detecting weak correlations, a generalized version of the light-bulb problem by Valiant.
- “Robust Estimates of Location: Survey and Advances” Princeton, NJ, USA: Princeton University Press, 1972
- J. Alman “An Illuminating Algorithm for the Light Bulb Problem” In Proc. 2nd Symposium on Simplicity in Algorithms (SOSA), 2019
- “Reducibility and Statistical-Computational Gaps from Secret Leakage” In Proc. 33rd Annual Conference on Learning Theory (COLT), 2020
- “Statistical query algorithms and low-degree tests are almost equivalent” In Proc. 34th Annual Conference on Learning Theory (COLT), 2021
- “Computationally Efficient Robust Sparse Estimation in High Dimensions” In Proc. 30th Annual Conference on Learning Theory (COLT), 2017
- “Optimal Robust Linear Regression in Nearly Linear Time” In arXiv abs/2007.08137, 2020
- Y. Cheng, I. Diakonikolas and R. Ge “High-Dimensional Robust Mean Estimation in Nearly-Linear Time” In Proc. 30th Annual Symposium on Discrete Algorithms (SODA), 2019 DOI: 10.1137/1.9781611975482.171
- “Faster Algorithms for High-Dimensional Robust Covariance Estimation” In Proc. 32nd Annual Conference on Learning Theory (COLT), 2019
- “Outlier-Robust Sparse Estimation via Non-Convex Optimization” In Advances in Neural Information Processing Systems 35 (NeurIPS), 2022
- Y. Cheng “High-Dimensional Robust Statistics: Faster Algorithms and Optimization Landscape” See timestamp 22:00 in the talk, Robustness in High-dimensional Statistics and Machine Learning at IDEAL Institute, 2021
- Y. Cherapanamjeri, S. Mohanty and M. Yau “List decodable mean estimation in nearly linear time” In Proc. 61st IEEE Symposium on Foundations of Computer Science (FOCS), 2020
- “A Direct Formulation for Sparse PCA Using Semidefinite Programming” In SIAM Review 49.3, 2007, pp. 434–448
- Y. Dong, S.B. Hopkins and J. Li “Quantum Entropy Scoring for Fast Robust Mean Estimation and Improved Outlier Detection” In Advances in Neural Information Processing Systems 32 (NeurIPS), 2019
- I. Diakonikolas “Computational-Statistical Tradeoffs and Open Problems” See page number 36 in http://www.iliasdiakonikolas.org/stoc19-tutorial/Tradeoffs-and-Open-Problems.pdf, STOC 2019 Tutorial: Recent Advances in High-Dimensional Robust Statistics, 2019
- I. Diakonikolas “Algorithmic Robust Statistics” Available online at https://youtu.be/HKm0L2Cy69Y?t=3527, Statistical thinking in the age of AI : robustness, fairness and privacy (Meeting in Mathematical Statistics), 2023
- “Algorithmic High-Dimensional Robust Statistics” Cambridge University Press, 2023
- “Robust Estimators in High Dimensions without the Computational Intractability” In Proc. 57th IEEE Symposium on Foundations of Computer Science (FOCS), 2016 DOI: 10.1109/FOCS.2016.85
- “Clustering Mixture Models in Almost-Linear Time via List-Decodable Mean Estimation” In Proc. 54th Annual ACM Symposium on Theory of Computing (STOC), 2022
- “Robust Sparse Mean Estimation via Sum of Squares” In Proc. 35th Annual Conference on Learning Theory (COLT), 2022
- “Outlier-Robust High-Dimensional Sparse Estimation via Iterative Filtering” In Advances in Neural Information Processing Systems 32 (NeurIPS), 2019
- “Outlier-Robust Sparse Mean Estimation for Heavy-Tailed Distributions” In Advances in Neural Information Processing Systems 35 (NeurIPS), 2022
- “Streaming Algorithms for High-Dimensional Robust Statistics” In Proc. 39th International Conference on Machine Learning (ICML), 2022
- “Near-Optimal Algorithms for Gaussians with Huber Contamination: Mean Estimation and Linear Regression” In Advances in Neural Information Processing Systems 36 (NeurIPS), 2023
- “Nearly-Linear Time and Streaming Algorithms for Outlier-Robust PCA” In Proc. 40th International Conference on Machine Learning (ICML), 2023
- I. Diakonikolas, D.M. Kane and A. Stewart “Statistical Query Lower Bounds for Robust Estimation of High-Dimensional Gaussians and Gaussian Mixtures” In Proc. 58th IEEE Symposium on Foundations of Computer Science (FOCS), 2017 DOI: 10.1109/FOCS.2017.16
- “Robust Subgaussian Estimation of a Mean Vector in Nearly Linear Time” In The Annals of Statistics 50.1 Institute of Mathematical Statistics, 2022, pp. 511–536
- “Detecting correlations with little memory and communication” In Proc. 31st Annual Conference on Learning Theory (COLT), 2018
- “Compressed sensing: theory and applications” Cambridge University Press, 2012
- “Robust Statistics” John Wiley & Sons, 2009
- T. Hastie, R. Tibshirani and M. Wainwright “Statistical Learning with Sparsity: The Lasso and Generalizations”, 2015
- P.J. Huber “Robust Estimation of a Location Parameter” In The Annals of Mathematical Statistics 35.1, 1964, pp. 73–101 DOI: 10.1214/aoms/1177703732
- “A Faster Interior Point Method for Semidefinite Programming” In Proc. 61st IEEE Symposium on Foundations of Computer Science (FOCS), 2020
- A. Jambulapati, J. Li and K. Tian “Robust sub-gaussian principal component analysis and width-independent schatten packing” In Advances in Neural Information Processing Systems 33 (NeurIPS), 2020
- M. Karppa, P. Kaski and J. Kohonen “A Faster Subquadratic Algorithm for Finding Outlier Correlations” In ACM Trans. Algorithms 14.3, 2018
- “Explicit Correlation Amplifiers for Finding Outlier Correlations in Deterministic Subquadratic Time” In Algorithmica 82.11, 2020
- “A Fast Spectral Algorithm for Mean Estimation with Sub-Gaussian Rates” In Proc. 33rd Annual Conference on Learning Theory (COLT), 2020
- K.A. Lai, A.B. Rao and S. Vempala “Agnostic Estimation of Mean and Covariance” In Proc. 57th IEEE Symposium on Foundations of Computer Science (FOCS), 2016, pp. 665–674 DOI: 10.1109/FOCS.2016.76
- “Information Theory: From Coding to Learning” Cambridge University Press, 2023
- “Faster Algorithms via Approximation Theory” In Foundations and Trends® in Theoretical Computer Science 9, 2014, pp. 125–210
- G. Valiant “Finding Correlations in Subquadratic Time, with Applications to Learning Parities and the Closest Pair Problem” In Journal of the ACM 62.2, 2015 DOI: 10.1145/2728167
- L.G. Valiant “Functionality in Neural Nets” In Proc. of the Seventh AAAI National Conference on Artificial Intelligence AAAI Press, 1988, pp. 629–634
- S. van de Geer “Estimation and Testing Under Sparsity”, École d’Été de Probabilités de Saint-Flour Springer, 2016
- R. Vershynin “High-Dimensional Probability: An Introduction with Applications in Data Science” Cambridge University Press, 2018
- B. Zhu, J. Jiao and J. Steinhardt “Robust Estimation via Generalized Quasi-Gradients” In Information and Inference: A Journal of the IMA, 2022, pp. 581–636 DOI: 10.1093/imaiai/iaab018
Collections
Sign up for free to add this paper to one or more collections.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.