Abstract
This paper consider the penalized least squares estimators with convex penalties or regularization norms. We provide sparsity oracle inequalities for the prediction error for a general convex penalty and for the particular cases of Lasso and Group Lasso estimators in a regression setting. The main contribution is that our oracle inequalities are established for the more general case where the observations noise is issued from probability measures that satisfy a weak spectral gap (or Poincaré) inequality instead of Gaussian distributions. We illustrate our results on a heavy tailed example and a sub Gaussian one; we especially give the explicit bounds of the oracle inequalities for these two special examples.
Article PDF
Similar content being viewed by others
Avoid common mistakes on your manuscript.
References
Tsybakov A, Bellec A, Lecué G. Towards the study of least squares estimators with convex penalty. Arxiv, 2017
Barthe F, Cattiaux P, Roberto C. Concentration for independent random variables with heavy tails. Applied Mathematics Research eXpress, 2005, 2005(2): 39–60
Bellec P, Lecué G, Tsybakov A. Slope meets lasso: improved oracle bounds and optimality. Annals of Statistics, 2018, 46(6B): 3603–3642
Bellec P, Tsybakov A. Bounds on the prediction error of penalized least squares estimators with convex penalty//International Conference on Modern Problems of Stochastic Analysis and Statistics. Springer, 2016: 315–333
Bickel P J, Ritov Y, Tsybakov A. Simultaneous analysis of Lasso and Dantzig selector. Annals of Statistics, 2009, 37(4): 1705–1732
Bühlmann P, Van De Geer S. Statistics for High-dimensional Data: Methods, Theory and Applications. Springer Science & Business Media, 2011
Cattiaux P, Guillin A. On the Poincaré constant of log-concave measures//Geometric Aspects of Functional Analysis. Springer, 2020: 171–217
van de Geer S. Estimation and Testing under Sparsity. Springer, 2016
Giraud C. Graphical models//Introduction to High-Dimensional Statistics. Chapman and Hall/CRC, 2014: 157–180
Koltchinskii V, Mendelson S. Bounding the smallest singular value of a random matrix without concentration. International Mathematics Research Notices, 2015, 2015(23): 12991–13008
Lounici K, et al. Oracle inequalities and optimal inference under group sparsity. Annals of Statistics, 2011, 39(4): 2164–2204
Mendelson S. Learning without concentration//Conference on Learning Theory. PMLR, 2014: 25–39
Negahban S, et al. A unified framework for high-dimensional analysis of M-estimators with decomposable regularizers. Statistical Science, 2012, 27(4): 538–557
Selesnick I. Sparse regularization via convex analysis. IEEE Transactions on Signal Processing, 2017, 65(17): 4481–4494
Taylor J. The geometry of least squares in the 21st century. Bernoulli, 2013, 19(4): 1449–1464
Author information
Authors and Affiliations
Corresponding author
Additional information
Dedicated to the memory of Professor Jiarong YU
This work has been (partially) supported by the Project EFI ANR-17-CE40-0030 of the French National Research Agency.
Rights and permissions
About this article
Cite this article
Abdillahi-Ali, D., Azzaoui, N., Guillin, A. et al. Penalized Least Square in Sparse Setting with Convex Penalty and Non Gaussian Errors. Acta Math Sci 41, 2198–2216 (2021). https://doi.org/10.1007/s10473-021-0624-0
Received:
Revised:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10473-021-0624-0