IDEAS home Printed from https://ideas.repec.org/a/eee/csdana/v173y2022ics0167947322000871.html
   My bibliography  Save this article

Thresholding tests based on affine LASSO to achieve non-asymptotic nominal level and high power under sparse and dense alternatives in high dimension

Author

Listed:
  • Sardy, Sylvain
  • Diaz-Rodriguez, Jairo
  • Giacobino, Caroline

Abstract

Thresholding estimators such as the existing square-root and LAD LASSO, and the new affine and GLM LASSO with new link functions, have the ability to set coefficients to zero. They will yield new pivotal statistics which enjoy high power under sparse or dense alternative hypotheses. Under a general formalism, thresholding tests not only recover existing tests such as Rao score test and Fisher nonparametric sign test, but also unveil new tests, for the global/omnibus hypothesis in high dimension in particular. Although pivotal, the new statistics do not have a known distribution, so the critical value of the test is calculated by straightforward Monte Carlo, which yields exact level and high power as illustrated on simulated data.

Suggested Citation

  • Sardy, Sylvain & Diaz-Rodriguez, Jairo & Giacobino, Caroline, 2022. "Thresholding tests based on affine LASSO to achieve non-asymptotic nominal level and high power under sparse and dense alternatives in high dimension," Computational Statistics & Data Analysis, Elsevier, vol. 173(C).
  • Handle: RePEc:eee:csdana:v:173:y:2022:i:c:s0167947322000871
    DOI: 10.1016/j.csda.2022.107507
    as

    Download full text from publisher

    File URL: http://www.sciencedirect.com/science/article/pii/S0167947322000871
    Download Restriction: Full text for ScienceDirect subscribers only.

    File URL: https://libkey.io/10.1016/j.csda.2022.107507?utm_source=ideas
    LibKey link: if access is restricted and if your library uses this service, LibKey will redirect you to where you can use your library subscription to access this item
    ---><---

    As the access to this document is restricted, you may want to search for a different version of it.

    References listed on IDEAS

    as
    1. Meinshausen, Nicolai & Meier, Lukas & Bühlmann, Peter, 2009. "p-Values for High-Dimensional Regression," Journal of the American Statistical Association, American Statistical Association, vol. 104(488), pages 1671-1681.
    2. Yinchu Zhu & Jelena Bradic, 2018. "Linear Hypothesis Testing in Dense High-Dimensional Linear Models," Journal of the American Statistical Association, Taylor & Francis Journals, vol. 113(524), pages 1583-1600, October.
    3. Bin Guo & Song Xi Chen, 2016. "Tests for high dimensional generalized linear models," Journal of the Royal Statistical Society Series B, Royal Statistical Society, vol. 78(5), pages 1079-1102, November.
    4. Jelle J. Goeman & Hans C. van Houwelingen & Livio Finos, 2011. "Testing against a high-dimensional alternative in the generalized linear model: asymptotic type I error control," Biometrika, Biometrika Trust, vol. 98(2), pages 381-390.
    5. Mee Young Park & Trevor Hastie, 2007. "L1‐regularization path algorithm for generalized linear models," Journal of the Royal Statistical Society Series B, Royal Statistical Society, vol. 69(4), pages 659-677, September.
    6. Sylvain Sardy, 2008. "On the Practice of Rescaling Covariates," International Statistical Review, International Statistical Institute, vol. 76(2), pages 285-297, August.
    7. Fan J. & Li R., 2001. "Variable Selection via Nonconcave Penalized Likelihood and its Oracle Properties," Journal of the American Statistical Association, American Statistical Association, vol. 96, pages 1348-1360, December.
    8. Yu I. Ingster & Alexandre B. Tsybakov & N. Verzelzn, 2010. "Detection Boundary in Sparse Regression," Working Papers 2010-28, Center for Research in Economics and Statistics.
    9. A. Belloni & V. Chernozhukov & L. Wang, 2011. "Square-root lasso: pivotal recovery of sparse signals via conic programming," Biometrika, Biometrika Trust, vol. 98(4), pages 791-806.
    10. Wang, Hansheng & Li, Guodong & Jiang, Guohua, 2007. "Robust Regression Shrinkage and Consistent Variable Selection Through the LAD-Lasso," Journal of Business & Economic Statistics, American Statistical Association, vol. 25, pages 347-355, July.
    11. Sylvain Sardy, 2009. "Adaptive Posterior Mode Estimation of a Sparse Sequence for Model Selection," Scandinavian Journal of Statistics, Danish Society for Theoretical Statistics;Finnish Statistical Society;Norwegian Statistical Association;Swedish Statistical Association, vol. 36(4), pages 577-601, December.
    12. Ming Yuan & Yi Lin, 2006. "Model selection and estimation in regression with grouped variables," Journal of the Royal Statistical Society Series B, Royal Statistical Society, vol. 68(1), pages 49-67, February.
    13. Cun-Hui Zhang & Stephanie S. Zhang, 2014. "Confidence intervals for low dimensional parameters in high dimensional linear models," Journal of the Royal Statistical Society Series B, Royal Statistical Society, vol. 76(1), pages 217-242, January.
    Full references (including those not matched with items on IDEAS)

    Most related items

    These are the items that most often cite the same works as this one and are cited by the same works as this one.
    1. Lan, Wei & Zhong, Ping-Shou & Li, Runze & Wang, Hansheng & Tsai, Chih-Ling, 2016. "Testing a single regression coefficient in high dimensional linear models," Journal of Econometrics, Elsevier, vol. 195(1), pages 154-168.
    2. He, Yi & Jaidee, Sombut & Gao, Jiti, 2023. "Most powerful test against a sequence of high dimensional local alternatives," Journal of Econometrics, Elsevier, vol. 234(1), pages 151-177.
    3. Tianxi Cai & T. Tony Cai & Zijian Guo, 2021. "Optimal statistical inference for individualized treatment effects in high‐dimensional models," Journal of the Royal Statistical Society Series B, Royal Statistical Society, vol. 83(4), pages 669-719, September.
    4. Alexandre Belloni & Victor Chernozhukov & Denis Chetverikov & Christian Hansen & Kengo Kato, 2018. "High-Dimensional Econometrics and Regularized GMM," Papers 1806.01888, arXiv.org, revised Jun 2018.
    5. Xue Wu & Chixiang Chen & Zheng Li & Lijun Zhang & Vernon M. Chinchilli & Ming Wang, 2024. "A three-stage approach to identify biomarker signatures for cancer genetic data with survival endpoints," Statistical Methods & Applications, Springer;Società Italiana di Statistica, vol. 33(3), pages 863-883, July.
    6. Zemin Zheng & Jinchi Lv & Wei Lin, 2021. "Nonsparse Learning with Latent Variables," Operations Research, INFORMS, vol. 69(1), pages 346-359, January.
    7. Umberto Amato & Anestis Antoniadis & Italia De Feis & Irene Gijbels, 2021. "Penalised robust estimators for sparse and high-dimensional linear models," Statistical Methods & Applications, Springer;Società Italiana di Statistica, vol. 30(1), pages 1-48, March.
    8. Toshio Honda, 2021. "The de-biased group Lasso estimation for varying coefficient models," Annals of the Institute of Statistical Mathematics, Springer;The Institute of Statistical Mathematics, vol. 73(1), pages 3-29, February.
    9. Ian W. McKeague & Min Qian, 2015. "An Adaptive Resampling Test for Detecting the Presence of Significant Predictors," Journal of the American Statistical Association, Taylor & Francis Journals, vol. 110(512), pages 1422-1433, December.
    10. Victor Chernozhukov & Christian Hansen & Yuan Liao, 2015. "A lava attack on the recovery of sums of dense and sparse signals," CeMMAP working papers CWP56/15, Centre for Microdata Methods and Practice, Institute for Fiscal Studies.
    11. Zanhua Yin, 2020. "Variable selection for sparse logistic regression," Metrika: International Journal for Theoretical and Applied Statistics, Springer, vol. 83(7), pages 821-836, October.
    12. Luke Mosley & Idris A. Eckley & Alex Gibberd, 2022. "Sparse temporal disaggregation," Journal of the Royal Statistical Society Series A, Royal Statistical Society, vol. 185(4), pages 2203-2233, October.
    13. Wang, Hansheng & Leng, Chenlei, 2008. "A note on adaptive group lasso," Computational Statistics & Data Analysis, Elsevier, vol. 52(12), pages 5277-5286, August.
    14. Xiaofei Wu & Rongmei Liang & Hu Yang, 2022. "Penalized and constrained LAD estimation in fixed and high dimension," Statistical Papers, Springer, vol. 63(1), pages 53-95, February.
    15. Bang, Sungwan & Jhun, Myoungshic, 2012. "Simultaneous estimation and factor selection in quantile regression via adaptive sup-norm regularization," Computational Statistics & Data Analysis, Elsevier, vol. 56(4), pages 813-826.
    16. Tanin Sirimongkolkasem & Reza Drikvandi, 2019. "On Regularisation Methods for Analysis of High Dimensional Data," Annals of Data Science, Springer, vol. 6(4), pages 737-763, December.
    17. Holter, Julia C. & Stallrich, Jonathan W., 2023. "Tuning parameter selection for penalized estimation via R2," Computational Statistics & Data Analysis, Elsevier, vol. 183(C).
    18. Ismail Shah & Hina Naz & Sajid Ali & Amani Almohaimeed & Showkat Ahmad Lone, 2023. "A New Quantile-Based Approach for LASSO Estimation," Mathematics, MDPI, vol. 11(6), pages 1-13, March.
    19. Qing Zhou & Seunghyun Min, 2017. "Uncertainty quantification under group sparsity," Biometrika, Biometrika Trust, vol. 104(3), pages 613-632.
    20. Gabriel E Hoffman & Benjamin A Logsdon & Jason G Mezey, 2013. "PUMA: A Unified Framework for Penalized Multiple Regression Analysis of GWAS Data," PLOS Computational Biology, Public Library of Science, vol. 9(6), pages 1-19, June.

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:eee:csdana:v:173:y:2022:i:c:s0167947322000871. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    If CitEc recognized a bibliographic reference but did not link an item in RePEc to it, you can help with this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: Catherine Liu (email available below). General contact details of provider: http://www.elsevier.com/locate/csda .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.