Optimality of Least-squares for Classification in Gaussian-Mixture Models

作者: Hossein Taheri , Ramtin Pedarsani , Christos Thrampoulidis

DOI: 10.1109/ISIT44484.2020.9174267

关键词:

摘要: We consider the problem of learning coefficients a linear classifier through Empirical Risk Minimization with convex loss function in high-dimensional setting. In particular, we introduce an approach to characterize best achievable classification risk among losses, when data points follow standard Gaussian-mixture model. Importantly, prove that square achieves minimum for this Our numerical illustrations verify theoretical results and show they are accurate even relatively small dimensions.

参考文章(18)
Shai Shalev-Shwartz, Shai Ben-David, Understanding Machine Learning: From Theory to Algorithms ,(2015)
Babak Hassibi, Christos Thrampoulidis, Samet Oymak, Regularized Linear Regression: A Precise Analysis of the Estimation Error conference on learning theory. pp. 1683- 1709 ,(2015)
Mihailo Stojnic, A framework to characterize performance of LASSO algorithms arXiv: Information Theory. ,(2013)
Mohsen Bayati, Andrea Montanari, The LASSO Risk for Gaussian Matrices IEEE Transactions on Information Theory. ,vol. 58, pp. 1997- 2017 ,(2012) , 10.1109/TIT.2011.2174612
Derek Bean, Peter J. Bickel, Noureddine El Karoui, Bin Yu, Optimal M-estimation in high-dimensional regression Proceedings of the National Academy of Sciences of the United States of America. ,vol. 110, pp. 14563- 14568 ,(2013) , 10.1073/PNAS.1307845110
Andrea Montanari, Michael Celentano, Fundamental Barriers to High-Dimensional Regression with Convex Penalties arXiv: Statistics Theory. ,(2019)
Xiaoyi Mai, Zhenyu Liao, Romain Couillet, A Large Scale Analysis of Logistic Regression: Asymptotic Performance and New Insights ICASSP 2019 - 2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). pp. 3357- 3361 ,(2019) , 10.1109/ICASSP.2019.8683376
Christos Thrampoulidis, Ehsan Abbasi, Babak Hassibi, Precise Error Analysis of Regularized $M$ -Estimators in High Dimensions IEEE Transactions on Information Theory. ,vol. 64, pp. 5592- 5628 ,(2018) , 10.1109/TIT.2018.2840720
Pragya Sur, Yuxin Chen, Emmanuel J. Candès, The likelihood ratio test in high-dimensional logistic regression is asymptotically a rescaled Chi-square Probability Theory and Related Fields. ,vol. 175, pp. 487- 558 ,(2019) , 10.1007/S00440-018-00896-9