Efficient Smoothed Concomitant Lasso Estimation for High Dimensional Regression

作者: Eugene Ndiaye , Olivier Fercoq , Alexandre Gramfort , Vincent Leclère , Joseph Salmon

DOI: 10.1088/1742-6596/904/1/012006

关键词: Uncertainty quantificationElastic net regularizationAlgorithmArtificial intelligenceMachine learningCurve fittingSolverCoordinate descentLeverage (statistics)MathematicsComputationNumerical stability

摘要: In high dimensional settings, sparse structures are crucial for efficiency, both in term of memory, computation and performance. It is customary to consider $\ell_1$ penalty enforce sparsity such scenarios. Sparsity enforcing methods, the Lasso being a canonical example, popular candidates address dimension. For they rely on tuning parameter trading data fitting versus sparsity. theory hold this should be proportional noise level, yet latter often unknown practice. A possible remedy jointly optimize over regression as well level. This has been considered under several names literature: Scaled-Lasso, Square-root Lasso, Concomitant estimation instance, could interest confidence sets or uncertainty quantification. work, after illustrating numerical difficulties Smoothed formulation, we propose modification coined aimed at increasing stability. We an efficient accurate solver leading computational cost no more expansive than one Lasso. leverage standard ingredients behind success fast solvers: coordinate descent algorithm, combined with safe screening rules achieve speed by eliminating early irrelevant features.

参考文章(33)
L. H. Dicker, Variance estimation in high-dimensional linear models Biometrika. ,vol. 101, pp. 269- 284 ,(2014) , 10.1093/BIOMET/AST065
Stephen Reid, Robert Tibshirani, Jerome Friedman, None, A Study of Error Variance Estimation in Lasso Regression arXiv: Methodology. ,(2013)
Robert Tibshirani, Regression Shrinkage and Selection Via the Lasso Journal of the Royal Statistical Society: Series B (Methodological). ,vol. 58, pp. 267- 288 ,(1996) , 10.1111/J.2517-6161.1996.TB02080.X
Peter Bühlmann, Sara van de Geer, Nicolas Städler, ℓ1-penalization for mixture regression models Test. ,vol. 19, pp. 209- 256 ,(2010) , 10.1007/S11749-010-0197-Z
T. Sun, C.-H. Zhang, Scaled sparse linear regression Biometrika. ,vol. 99, pp. 879- 898 ,(2012) , 10.1093/BIOMET/ASS043
Laurent El Ghaoui, Vu Pham, Arturo Fernandez, Robust sketching for multiple square-root LASSO problems arXiv: Optimization and Control. ,(2014)
Stephane Chretien, Sebastien Darses, Sparse Recovery With Unknown Variance: A LASSO-Type Approach IEEE Transactions on Information Theory. ,vol. 60, pp. 3970- 3988 ,(2014) , 10.1109/TIT.2014.2301162
Yu. Nesterov, Smooth minimization of non-smooth functions Mathematical Programming. ,vol. 103, pp. 127- 152 ,(2005) , 10.1007/S10107-004-0552-5
Stephen Boyd, Steven Diamond, CVXPY: a python-embedded modeling language for convex optimization Journal of Machine Learning Research. ,vol. 17, pp. 2909- 2913 ,(2016)
Andrew Chang, Robust Statistics Wiley Series in Probability and Statistics. ,(1981) , 10.1002/0471725250