Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Regression Shrinkage and Selection Via the Lasso
50.555
Zitationen
1
Autoren
1996
Jahr
Abstract
SUMMARY We propose a new method for estimation in linear models. The ‘lasso’ minimizes the residual sum of squares subject to the sum of the absolute value of the coefficients being less than a constant. Because of the nature of this constraint it tends to produce some coefficients that are exactly 0 and hence gives interpretable models. Our simulation studies suggest that the lasso enjoys some of the favourable properties of both subset selection and ridge regression. It produces interpretable models like subset selection and exhibits the stability of ridge regression. There is also an interesting relationship with recent work in adaptive function estimation by Donoho and Johnstone. The lasso idea is quite general and can be applied in a variety of statistical models: extensions to generalized regression models and tree-based models are briefly described.
Ähnliche Arbeiten
Maximum Likelihood from Incomplete Data Via the <i>EM</i> Algorithm
1977 · 49.238 Zit.
Nonparametric Estimation from Incomplete Observations
1992 · 45.534 Zit.
An Introduction to the Bootstrap
1994 · 39.345 Zit.
Nonparametric Estimation from Incomplete Observations
1958 · 38.749 Zit.
The central role of the propensity score in observational studies for causal effects
1983 · 30.320 Zit.