Prediction consistency of lasso regression does not need normal errors
Sourav Chatterjee in 2014 proved consistency of any estimator using orthogonal least squares (OLS) together with Lasso penalty under the conditions the observations are upper bounded, with normal errors, and being independent of observations, with a zero mean and a finite variance. Reviewing his elegant proof, we come to the conclusion that the prediction consistency of OLS with Lasso can be proven even with fewer assumptions, i.e., without assuming normality of the errors, knowing only they have a finite variance and zero mean. We give an upper bound on the convergence rate of OLS-Lasso estimator for these errors. This upper bound is not asymptotic and depends both on the number of regressors and on the size of the data set. Knowing the number of regressors in a regression problem, one can estimate how large data set is needed, to achieve a prediction error under a given value, and this in comparison to the cited work, without solving the parameter estimation problem for fitting the errors to a normal distribution. The result can encourage practitioners to use OLS Lasso as a convergent algorithm for prediction with other than normal errors satisfying these milder conditions.
Top- Hlavackova-Schindler, Katerina
Category |
Journal Paper |
Divisions |
Data Mining and Machine Learning |
Subjects |
Informatik Sonstiges |
Journal or Publication Title |
British Journal of Mathematics and Computer Science |
ISSN |
2231-0851 |
Publisher |
SCIENCEDOMAIN international |
Page Range |
pp. 1-7 |
Number |
4 |
Volume |
19 |
Date |
2016 |
Export |