Prediction consistency of lasso regression does not need normal errors

Prediction consistency of lasso regression does not need normal errors

Abstract

Sourav Chatterjee in 2014 proved consistency of any estimator using orthogonal least squares (OLS) together with Lasso penalty under the conditions the observations are upper bounded, with normal errors, and being independent of observations, with a zero mean and a finite variance. Reviewing his elegant proof, we come to the conclusion that the prediction consistency of OLS with Lasso can be proven even with fewer assumptions, i.e., without assuming normality of the errors, knowing only they have a finite variance and zero mean. We give an upper bound on the convergence rate of OLS-Lasso estimator for these errors. This upper bound is not asymptotic and depends both on the number of regressors and on the size of the data set. Knowing the number of regressors in a regression problem, one can estimate how large data set is needed, to achieve a prediction error under a given value, and this in comparison to the cited work, without solving the parameter estimation problem for fitting the errors to a normal distribution. The result can encourage practitioners to use OLS Lasso as a convergent algorithm for prediction with other than normal errors satisfying these milder conditions.

Grafik Top
Authors
  • Hlavackova-Schindler, Katerina
Grafik Top
Shortfacts
Category
Journal Paper
Divisions
Data Mining and Machine Learning
Subjects
Informatik Sonstiges
Journal or Publication Title
British Journal of Mathematics and Computer Science
ISSN
2231-0851
Publisher
SCIENCEDOMAIN international
Page Range
pp. 1-7
Number
4
Volume
19
Date
2016
Export
Grafik Top