IDEAS home Printed from https://ideas.repec.org/a/eee/apmaco/v386y2020ics0096300320304586.html
   My bibliography  Save this article

Safe feature screening rules for the regularized Huber regression

Author

Listed:
  • Chen, Huangyue
  • Kong, Lingchen
  • Shang, Pan
  • Pan, Shanshan

Abstract

With the dramatic development of data collection and storage techniques, we often encounter massive high-dimensional data sets which contain outliers and heavy-tailed errors. Recently, the regularized Huber regression has been extensively developed to deal with such complex data sets. Although there are dozens of papers devoted to developing efficient solvers for the regularized Huber regression, it remains challenging when the number of features is extremely large. In this paper, we propose safe feature screening rules for the regularized Huber regression based on duality theory. These rules can remarkably accelerate the existing solvers for the regularized Huber regression by quickly reducing the number of features. To be specific, the proposed safe feature screening rules enable to identify and eliminate inactive features before starting the solver, then the computational effort can be saved significantly. Moreover, the proposed screening rules are safe in theory and practice. Finally, the experimental results on both synthetic and real data sets illustrate that the proposed screening rules can accelerate the speed of solving the regularized Huber regression and maintain its accuracy. In particular, when the number of features is large, the speedup obtained by our rules can be orders of magnitude.

Suggested Citation

  • Chen, Huangyue & Kong, Lingchen & Shang, Pan & Pan, Shanshan, 2020. "Safe feature screening rules for the regularized Huber regression," Applied Mathematics and Computation, Elsevier, vol. 386(C).
  • Handle: RePEc:eee:apmaco:v:386:y:2020:i:c:s0096300320304586
    DOI: 10.1016/j.amc.2020.125500
    as

    Download full text from publisher

    File URL: http://www.sciencedirect.com/science/article/pii/S0096300320304586
    Download Restriction: Full text for ScienceDirect subscribers only

    File URL: https://libkey.io/10.1016/j.amc.2020.125500?utm_source=ideas
    LibKey link: if access is restricted and if your library uses this service, LibKey will redirect you to where you can use your library subscription to access this item
    ---><---

    As the access to this document is restricted, you may want to search for a different version of it.

    References listed on IDEAS

    as
    1. Wang, Hansheng & Li, Guodong & Jiang, Guohua, 2007. "Robust Regression Shrinkage and Consistent Variable Selection Through the LAD-Lasso," Journal of Business & Economic Statistics, American Statistical Association, vol. 25, pages 347-355, July.
    2. Jianqing Fan & Jinchi Lv, 2008. "Sure independence screening for ultrahigh dimensional feature space," Journal of the Royal Statistical Society Series B, Royal Statistical Society, vol. 70(5), pages 849-911, November.
    3. Li, Mei & Kong, Lingchen, 2019. "Double fused Lasso penalized LAD for matrix regression," Applied Mathematics and Computation, Elsevier, vol. 357(C), pages 119-138.
    4. Robert Tibshirani & Jacob Bien & Jerome Friedman & Trevor Hastie & Noah Simon & Jonathan Taylor & Ryan J. Tibshirani, 2012. "Strong rules for discarding predictors in lasso‐type problems," Journal of the Royal Statistical Society Series B, Royal Statistical Society, vol. 74(2), pages 245-266, March.
    5. Qiang Sun & Wen-Xin Zhou & Jianqing Fan, 2020. "Adaptive Huber Regression," Journal of the American Statistical Association, Taylor & Francis Journals, vol. 115(529), pages 254-265, January.
    6. Jianqing Fan & Quefeng Li & Yuyan Wang, 2017. "Estimation of high dimensional mean regression in the absence of symmetry and light tail assumptions," Journal of the Royal Statistical Society Series B, Royal Statistical Society, vol. 79(1), pages 247-265, January.
    Full references (including those not matched with items on IDEAS)

    Most related items

    These are the items that most often cite the same works as this one and are cited by the same works as this one.
    1. Han, Dongxiao & Huang, Jian & Lin, Yuanyuan & Shen, Guohao, 2022. "Robust post-selection inference of high-dimensional mean regression with heavy-tailed asymmetric or heteroskedastic errors," Journal of Econometrics, Elsevier, vol. 230(2), pages 416-431.
    2. Yuyang Liu & Pengfei Pi & Shan Luo, 2023. "A semi-parametric approach to feature selection in high-dimensional linear regression models," Computational Statistics, Springer, vol. 38(2), pages 979-1000, June.
    3. Can Wu & Ying Cui & Donghui Li & Defeng Sun, 2023. "Convex and Nonconvex Risk-Based Linear Regression at Scale," INFORMS Journal on Computing, INFORMS, vol. 35(4), pages 797-816, July.
    4. Umberto Amato & Anestis Antoniadis & Italia De Feis & Irene Gijbels, 2021. "Penalised robust estimators for sparse and high-dimensional linear models," Statistical Methods & Applications, Springer;Società Italiana di Statistica, vol. 30(1), pages 1-48, March.
    5. Zeng, Yaohui & Yang, Tianbao & Breheny, Patrick, 2021. "Hybrid safe–strong rules for efficient optimization in lasso-type problems," Computational Statistics & Data Analysis, Elsevier, vol. 153(C).
    6. Guang Cheng & Hao Zhang & Zuofeng Shang, 2015. "Sparse and efficient estimation for partial spline models with increasing dimension," Annals of the Institute of Statistical Mathematics, Springer;The Institute of Statistical Mathematics, vol. 67(1), pages 93-127, February.
    7. Xing, Li-Min & Zhang, Yue-Jun, 2022. "Forecasting crude oil prices with shrinkage methods: Can nonconvex penalty and Huber loss help?," Energy Economics, Elsevier, vol. 110(C).
    8. Hu Yang & Ning Li & Jing Yang, 2020. "A robust and efficient estimation and variable selection method for partially linear models with large-dimensional covariates," Statistical Papers, Springer, vol. 61(5), pages 1911-1937, October.
    9. Pan Shang & Lingchen Kong, 2021. "Regularization Parameter Selection for the Low Rank Matrix Recovery," Journal of Optimization Theory and Applications, Springer, vol. 189(3), pages 772-792, June.
    10. Junlong Zhao & Chao Liu & Lu Niu & Chenlei Leng, 2019. "Multiple influential point detection in high dimensional regression spaces," Journal of the Royal Statistical Society Series B, Royal Statistical Society, vol. 81(2), pages 385-408, April.
    11. Aneiros, Germán & Novo, Silvia & Vieu, Philippe, 2022. "Variable selection in functional regression models: A review," Journal of Multivariate Analysis, Elsevier, vol. 188(C).
    12. N. Neykov & P. Filzmoser & P. Neytchev, 2014. "Ultrahigh dimensional variable selection through the penalized maximum trimmed likelihood estimator," Statistical Papers, Springer, vol. 55(1), pages 187-207, February.
    13. Xiao, Xuan & Xu, Xingbai & Zhong, Wei, 2023. "Huber estimation for the network autoregressive model," Statistics & Probability Letters, Elsevier, vol. 203(C).
    14. Khan, Faridoon & Muhammadullah, Sara & Sharif, Arshian & Lee, Chien-Chiang, 2024. "The role of green energy stock market in forecasting China's crude oil market: An application of IIS approach and sparse regression models," Energy Economics, Elsevier, vol. 130(C).
    15. Man, Rebeka & Tan, Kean Ming & Wang, Zian & Zhou, Wen-Xin, 2024. "Retire: Robust expectile regression in high dimensions," Journal of Econometrics, Elsevier, vol. 239(2).
    16. Christis Katsouris, 2023. "High Dimensional Time Series Regression Models: Applications to Statistical Learning Methods," Papers 2308.16192, arXiv.org.
    17. Qiang Li & Liming Wang, 2020. "Robust change point detection method via adaptive LAD-LASSO," Statistical Papers, Springer, vol. 61(1), pages 109-121, February.
    18. Gabriel E Hoffman & Benjamin A Logsdon & Jason G Mezey, 2013. "PUMA: A Unified Framework for Penalized Multiple Regression Analysis of GWAS Data," PLOS Computational Biology, Public Library of Science, vol. 9(6), pages 1-19, June.
    19. Kean Ming Tan & Lan Wang & Wen‐Xin Zhou, 2022. "High‐dimensional quantile regression: Convolution smoothing and concave regularization," Journal of the Royal Statistical Society Series B, Royal Statistical Society, vol. 84(1), pages 205-233, February.
    20. Wang, Yibo & Karunamuni, Rohana J., 2022. "High-dimensional robust regression with Lq-loss functions," Computational Statistics & Data Analysis, Elsevier, vol. 176(C).

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:eee:apmaco:v:386:y:2020:i:c:s0096300320304586. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    If CitEc recognized a bibliographic reference but did not link an item in RePEc to it, you can help with this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: Catherine Liu (email available below). General contact details of provider: https://www.journals.elsevier.com/applied-mathematics-and-computation .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.