Sparse PCA with Oracle Property (2312.16793v1)
Abstract: In this paper, we study the estimation of the $k$-dimensional sparse principal subspace of covariance matrix $\Sigma$ in the high-dimensional setting. We aim to recover the oracle principal subspace solution, i.e., the principal subspace estimator obtained assuming the true support is known a priori. To this end, we propose a family of estimators based on the semidefinite relaxation of sparse PCA with novel regularizations. In particular, under a weak assumption on the magnitude of the population projection matrix, one estimator within this family exactly recovers the true support with high probability, has exact rank-$k$, and attains a $\sqrt{s/n}$ statistical rate of convergence with $s$ being the subspace sparsity level and $n$ the sample size. Compared to existing support recovery results for sparse PCA, our approach does not hinge on the spiked covariance model or the limited correlation condition. As a complement to the first estimator that enjoys the oracle property, we prove that, another estimator within the family achieves a sharper statistical rate of convergence than the standard semidefinite relaxation of sparse PCA, even when the previous assumption on the magnitude of the projection matrix is violated. We validate the theoretical results by numerical experiments on synthetic datasets.
- A. Amini and M. Wainwright. High-dimensional analysis of semidefinite relaxations for sparse principal components. The Annals of Statistics, 37(5B):2877–2921, 2009.
- Q. Berthet and P. Rigollet. Computational lower bounds for sparse PCA. arXiv preprint arXiv:1304.0828, 2013.
- Q. Berthet and P. Rigollet. Optimal detection of sparse principal components in high dimension. The Annals of Statistics, 41(4):1780–1815, 2013.
- Minimax bounds for sparse PCA with noisy high-dimensional data. The Annals of Statistics, 41(3):1055–1084, 2013.
- P. Breheny and J. Huang. Coordinate descent algorithms for nonconvex penalized regression, with applications to biological feature selection. The Annals of Applied Statistics, 5(1):232–253, 03 2011.
- Sparse PCA: Optimal rates and adaptive estimation. The Annals of Statistics, 41(6):3074–3110, 2013.
- Optimal solutions for sparse principal component analysis. The Journal of Machine Learning Research, 9:1269–1294, 2008.
- A direct formulation for sparse PCA using semidefinite programming. SIAM Review, pages 434–448, 2007.
- J. Dattorro. Convex Optimization & Euclidean Distance Geometry. Meboo Publishing USA, 2011.
- J. Fan and R. Li. Variable selection via nonconcave penalized likelihood and its oracle properties. Journal of the American Statistical Association, 96(456):1348–1360, 2001.
- A strictly contractive peaceman–rachford splitting method for convex programming. SIAM Journal on Optimization, 24(3):1011–1040, 2014.
- I. Johnstone and A. Lu. On consistency and sparsity for principal components analysis in high dimensions. Journal of the American Statistical Association, 104(486):682–693, 2009.
- A modified principal component technique based on the lasso. Journal of Computational and Graphical Statistics, 12(3):531–547, 2003.
- Generalized power method for sparse principal component analysis. The Journal of Machine Learning Research, 11:517–553, 2010.
- T. Kato. Perturbation theory for linear operators, volume 132. springer, 1995.
- Do semidefinite relaxations really solve sparse PCA? arXiv preprint arXiv:1306.3690, 2013.
- J. Lei and V. Q. Vu. Sparsistency and agnostic inference in sparse PCA. arXiv preprint arXiv:1401.6978, 2014.
- Regularized m𝑚mitalic_m-estimators with nonconvexity: Statistical and algorithmic theory for local optima. arXiv preprint arXiv:1305.2436, 2013.
- K. Lounici. Sparse principal component analysis with missing observations. In High Dimensional Probability VI, pages 327–356. Springer, 2013.
- Z. Ma. Sparse principal component analysis and iterative thresholding. The Annals of Statistics, 41(2):772–801, 2013.
- D. Paul and I. M. Johnstone. Augmented sparse principal component analysis for high dimensional data. arXiv preprint arXiv:1202.1242, 2012.
- Consistency of sparse PCA in high dimension, low sample size contexts. Journal of Multivariate Analysis, 115:317–333, 2013.
- H. Shen and J. Huang. Sparse principal component analysis via regularized low rank matrix approximation. Journal of multivariate analysis, 99(6):1015–1034, 2008.
- D. E. Tyler et al. Asymptotic inference for eigenvectors. The Annals of Statistics, 9(4):725–736, 1981.
- R. Vershynin. How close is the sample covariance matrix to the actual covariance matrix? arXiv/1004.3484, 2010.
- Fantope projection and selection: A near-optimal convex relaxation of sparse pca. In NIPS, pages 2670–2678, 2013.
- V. Q. Vu and J. Lei. Minimax rates of estimation for sparse PCA in high dimensions. In International Conference on Artificial Intelligence and Statistics, pages 1278–1286, 2012.
- V. Q. Vu and J. Lei. Minimax sparse principal subspace estimation in high dimensions. The Annals of Statistics, 41(6):2905–2947, 2013.
- Optimal computational and statistical rates of convergence for sparse nonconvex learning problems. The Annals of Statistics, 42(6):2164–2201, 12 2014.
- Nonconvex statistical optimization: minimax-optimal sparse pca in polynomial time. arXiv preprint arXiv:1408.5352, 2014.
- X.-T. Yuan and T. Zhang. Truncated power method for sparse eigenvalue problems. The Journal of Machine Learning Research, 14(1):899–925, 2013.
- C.-H. Zhang. Nearly unbiased variable selection under minimax concave penalty. Ann. Statist., 38(2):894–942, 2010.
- Sparse principal component analysis. Journal of computational and graphical statistics, 15(2):265–286, 2006.
- Quanquan Gu (198 papers)
- Zhaoran Wang (164 papers)
- Han Liu (340 papers)