TY - JOUR
T1 - A Tuning-free Robust and Efficient Approach to High-dimensional Regression
AU - Wang, Lan
AU - Peng, Bo
AU - Bradic, Jelena
AU - Li, Runze
AU - Wu, Yunan
N1 - Funding Information:
Wang and Wu’s research was supported by NSF DMS-1712706, NSF OAC-1940160 and FRGMS-1952373. Bradic’s research was supported by NSF DMS-1712481. Li’s research was supported by NSF DMS 1820702, DMS 1953196 and DMS 2015539, NIDA grant P50 DA039838, R01CA229542 and R01 ES019672. The content is solely the responsibility of the authors and does not necessarily represent the official views of the NIDA, the NIH or the NSF. The authors are indebted to the referees, the associate editor and the Co-editors for their valuable comments, which have significantly improved the paper.
Publisher Copyright:
© 2020 American Statistical Association.
PY - 2020
Y1 - 2020
N2 - We introduce a novel approach for high-dimensional regression with theoretical guarantees. The new procedure overcomes the challenge of tuning parameter selection of Lasso and possesses several appealing properties. It uses an easily simulated tuning parameter that automatically adapts to both the unknown random error distribution and the correlation structure of the design matrix. It is robust with substantial efficiency gain for heavy-tailed random errors while maintaining high efficiency for normal random errors. Comparing with other alternative robust regression procedures, it also enjoys the property of being equivariant when the response variable undergoes a scale transformation. Computationally, it can be efficiently solved via linear programming. Theoretically, under weak conditions on the random error distribution, we establish a finite-sample error bound with a near-oracle rate for the new estimator with the simulated tuning parameter. Our results make useful contributions to mending the gap between the practice and theory of Lasso and its variants. We also prove that further improvement in efficiency can be achieved by a second-stage enhancement with some light tuning. Our simulation results demonstrate that the proposed methods often outperform cross-validated Lasso in various settings.
AB - We introduce a novel approach for high-dimensional regression with theoretical guarantees. The new procedure overcomes the challenge of tuning parameter selection of Lasso and possesses several appealing properties. It uses an easily simulated tuning parameter that automatically adapts to both the unknown random error distribution and the correlation structure of the design matrix. It is robust with substantial efficiency gain for heavy-tailed random errors while maintaining high efficiency for normal random errors. Comparing with other alternative robust regression procedures, it also enjoys the property of being equivariant when the response variable undergoes a scale transformation. Computationally, it can be efficiently solved via linear programming. Theoretically, under weak conditions on the random error distribution, we establish a finite-sample error bound with a near-oracle rate for the new estimator with the simulated tuning parameter. Our results make useful contributions to mending the gap between the practice and theory of Lasso and its variants. We also prove that further improvement in efficiency can be achieved by a second-stage enhancement with some light tuning. Our simulation results demonstrate that the proposed methods often outperform cross-validated Lasso in various settings.
UR - http://www.scopus.com/inward/record.url?scp=85097763077&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85097763077&partnerID=8YFLogxK
U2 - 10.1080/01621459.2020.1840989
DO - 10.1080/01621459.2020.1840989
M3 - Article
AN - SCOPUS:85097763077
SN - 0162-1459
VL - 115
SP - 1700
EP - 1714
JO - Journal of the American Statistical Association
JF - Journal of the American Statistical Association
IS - 532
ER -