Abstract
Regression procedures are not only hindered by large p and small n, but can also suffer in cases when outliers are present or the data generating mechanisms are heavy tailed. Since the penalized estimates like the least absolute shrinkage and selection operator (LASSO) are equipped to deal with the large p small n by encouraging sparsity, we combine a LASSO type penalty with the absolute deviation loss function, instead of the standard least squares loss, to handle the presence of outliers and heavy tails. The model is cast in a Bayesian setting and a Gibbs sampler is derived to efficiently sample from the posterior distribution. We compare our method to existing methods in a simulation study as well as on a prostate cancer data set and a base deficit data set from trauma patients.
| Original language | English |
|---|---|
| Pages (from-to) | 1115-1132 |
| Number of pages | 18 |
| Journal | Journal of Statistical Computation and Simulation |
| Volume | 86 |
| Issue number | 6 |
| DOIs | |
| State | Published - Apr 12 2016 |
UN SDGs
This output contributes to the following UN Sustainable Development Goals (SDGs)
-
SDG 3 Good Health and Well-being
Scopus Subject Areas
- Statistics and Probability
- Modeling and Simulation
- Statistics, Probability and Uncertainty
- Applied Mathematics
Keywords
- Lasso
- heavy tail
- loss function
- outlier
- regression
- sparsity
Fingerprint
Dive into the research topics of 'Balanced Bayesian LASSO for heavy tails'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver