Abstract
Regression procedures are not only hindered by large p and small n, but can also suffer in cases when outliers are present or the data generating mechanisms are heavy tailed. Since the penalized estimates like the least absolute shrinkage and selection operator (LASSO) are equipped to deal with the large p small n by encouraging sparsity, we combine a LASSO type penalty with the absolute deviation loss function, instead of the standard least squares loss, to handle the presence of outliers and heavy tails. The model is cast in a Bayesian setting and a Gibbs sampler is derived to efficiently sample from the posterior distribution. We compare our method to existing methods in a simulation study as well as on a prostate cancer data set and a base deficit data set from trauma patients.
Original language | English |
---|---|
Pages (from-to) | 1115-1132 |
Number of pages | 18 |
Journal | Journal of Statistical Computation and Simulation |
Volume | 86 |
Issue number | 6 |
DOIs | |
State | Published - Apr 12 2016 |
Keywords
- Lasso
- heavy tail
- loss function
- outlier
- regression
- sparsity