IDEAS home Printed from https://ideas.repec.org/p/osf/metaar/eqhjd_v1.html
   My bibliography  Save this paper

Dissertation R.C.M. van Aert

Author

Listed:
  • van Aert, Robbie Cornelis Maria

Abstract

More and more scientific research gets published nowadays, asking for statistical methods that enable researchers to get an overview of the literature in a particular research field. For that purpose, meta-analysis methods were developed that can be used for statistically combining the effect sizes from independent primary studies on the same topic. My dissertation focuses on two issues that are crucial when conducting a meta-analysis: publication bias and heterogeneity in primary studies’ true effect sizes. Accurate estimation of both the meta-analytic effect size as well as the between-study variance in true effect size is crucial since the results of meta-analyses are often used for policy making. Publication bias distorts the results of a meta-analysis since it refers to situations where publication of a primary study depends on its results. We developed new meta-analysis methods, p-uniform and p-uniform*, which estimate effect sizes corrected for publication bias and also test for publication bias. Although the methods perform well in many conditions, these and the other existing methods are shown not to perform well when researchers use questionable research practices. Additionally, when publication bias is absent or limited, traditional methods that do not correct for publication bias outperform p¬-uniform and p-uniform*. Surprisingly, we found no strong evidence for the presence of publication bias in our pre-registered study on the presence of publication bias in a large-scale data set consisting of 83 meta-analyses and 499 systematic reviews published in the fields of psychology and medicine. We also developed two methods for meta-analyzing a statistically significant published original study and a replication of that study, which reflects a situation often encountered by researchers. One method is a frequentist whereas the other method is a Bayesian statistical method. Both methods are shown to perform better than traditional meta-analytic methods that do not take the statistical significance of the original study into account. Analytical studies of both methods also show that sometimes the original study is better discarded for optimal estimation of the true effect size. Finally, we developed a program for determining the required sample size in a replication analogous to power analysis in null hypothesis testing. Computing the required sample size with the method revealed that large sample sizes (approximately 650 participants) are required to be able to distinguish a zero from a small true effect. Finally, in the last two chapters we derived a new multi-step estimator for the between-study variance in primary studies’ true effect sizes, and examined the statistical properties of two methods (Q-profile and generalized Q-statistic method) to compute the confidence interval of the between-study variance in true effect size. We proved that the multi-step estimator converges to the Paule-Mandel estimator which is nowadays one of the recommended methods to estimate the between-study variance in true effect sizes. Two Monte-Carlo simulation studies showed that the coverage probabilities of Q-profile and generalized Q-statistic method can be substantially below the nominal coverage rate if the assumptions underlying the random-effects meta-analysis model were violated.

Suggested Citation

  • van Aert, Robbie Cornelis Maria, 2018. "Dissertation R.C.M. van Aert," MetaArXiv eqhjd_v1, Center for Open Science.
  • Handle: RePEc:osf:metaar:eqhjd_v1
    DOI: 10.31219/osf.io/eqhjd_v1
    as

    Download full text from publisher

    File URL: https://osf.io/download/5b1686d6a291c4000d3ac3e0/
    Download Restriction: no

    File URL: https://libkey.io/10.31219/osf.io/eqhjd_v1?utm_source=ideas
    LibKey link: if access is restricted and if your library uses this service, LibKey will redirect you to where you can use your library subscription to access this item
    ---><---

    References listed on IDEAS

    as
    1. Marcel A L M van Assen & Robbie C M van Aert & Michèle B Nuijten & Jelte M Wicherts, 2014. "Why Publishing Everything Is More Effective than Selective Publishing of Statistically Significant Results," PLOS ONE, Public Library of Science, vol. 9(1), pages 1-5, January.
    2. J. Copas, 1999. "What works?: selectivity models and meta‐analysis," Journal of the Royal Statistical Society Series A, Royal Statistical Society, vol. 162(1), pages 95-109.
    3. Nosek, BA & Alter, G & Banks, GC & Borsboom, D & Bowman, SD & Breckler, SJ & Buck, S & Chambers, CD & Chin, G & Christensen, G & Contestabile, M & Dafoe, A & Eich, E & Freese, J & Glennerster, R & Gor, 2015. "Promoting an open research culture," Department of Economics, Working Paper Series qt7wh1000s, Department of Economics, Institute for Business and Economic Research, UC Berkeley.
    4. Gerber, Alan S. & Green, Donald P. & Nickerson, David, 2001. "Testing for Publication Bias in Political Science," Political Analysis, Cambridge University Press, vol. 9(4), pages 385-392, January.
    Full references (including those not matched with items on IDEAS)

    Most related items

    These are the items that most often cite the same works as this one and are cited by the same works as this one.
    1. van Aert, Robbie Cornelis Maria, 2018. "Dissertation R.C.M. van Aert," MetaArXiv eqhjd, Center for Open Science.
    2. Stanley, T. D. & Doucouliagos, Hristos, 2011. "Meta-regression approximations to reduce publication selection bias," Working Papers eco_2011_4, Deakin University, Department of Economics.
    3. Christopher H. Schmid, 2018. "Discussion of “quantifying publication bias in meta‐analysis” by Lin et al," Biometrics, The International Biometric Society, vol. 74(3), pages 797-799, September.
    4. Stanley, T. D. & Doucouliagos, Hristos, 2013. "Better than random: weighted least squares meta-regression analysis," Working Papers eco_2013_2, Deakin University, Department of Economics.
    5. van Aert, Robbie Cornelis Maria & van Assen, Marcel A. L. M., 2018. "P-uniform," MetaArXiv zqjr9, Center for Open Science.
    6. Ash, Elliott & Asher, Sam & Bhowmick, Aditi & Bhupatiraju, Sandeep & Chen, Daniel L. & Devi, Tatanya & Goessmann, Christoph & Novosad, Paul & Siddiqi, Bilal, 2022. "Measuring Gender and Religious Bias in the Indian Judiciary," TSE Working Papers 22-1395, Toulouse School of Economics (TSE).
    7. Jian Qing Shi & John Copas, 2002. "Publication bias and meta‐analysis for 2×2 tables: an average Markov chain Monte Carlo EM algorithm," Journal of the Royal Statistical Society Series B, Royal Statistical Society, vol. 64(2), pages 221-236, May.
    8. Lui, P. Priscilla & Gobrial, Sarah & Pham, Savannah & Giadolor, Westley & Adams, Niki & Rollock, David, 2021. "Open Science and Multicultural Research: Some Data, Considerations, and Recommendations," OSF Preprints em9ua_v1, Center for Open Science.
    9. Neves, Kleber & Amaral, Olavo Bohrer, 2019. "Addressing selective reporting of experiments – the case for predefined exclusion criteria," MetaArXiv a8gu5_v1, Center for Open Science.
    10. Marie Juanchich & Miroslav Sirota, 2016. "How to improve people's interpretation of probabilities of precipitation," Journal of Risk Research, Taylor & Francis Journals, vol. 19(3), pages 388-404, March.
    11. Lohse, Johannes & Rahal, Rima-Maria & Schulte-Mecklenbeck, Michael & Sofianos, Andis & Wollbrant, Conny, 2024. "Investigations of decision processes at the intersection of psychology and economics," Journal of Economic Psychology, Elsevier, vol. 103(C).
    12. Rao, Vijayendra & Ananthpur, Kripa & Malik, Kabir, 2017. "The Anatomy of Failure: An Ethnography of a Randomized Trial to Deepen Democracy in Rural India," World Development, Elsevier, vol. 99(C), pages 481-497.
    13. Chris Hartgerink, 2019. "Verified, Shared, Modular, and Provenance Based Research Communication with the Dat Protocol," Publications, MDPI, vol. 7(2), pages 1-19, June.
    14. Stommes, Drew & Aronow, P. M. & Sävje, Fredrik, 2023. "On the Reliability of Published Findings Using the Regression Discontinuity Design in Political Science," I4R Discussion Paper Series 22, The Institute for Replication (I4R).
    15. Nickerson, David W. & Friedrichs, Ryan D. & King, David, 2004. "Mobilizing the Party Faithful: Results from a Statewide Turnout Experiment in Michigan," Working Paper Series rwp04-018, Harvard University, John F. Kennedy School of Government.
    16. Askarov, Zohid & Doucouliagos, Hristos & Paldam, Martin & Stanley, T.D., 2022. "Rewarding good political behavior: US aid, democracy, and human rights," European Journal of Political Economy, Elsevier, vol. 71(C).
    17. Weiß Bernd & Wagner Michael, 2011. "The Identification and Prevention of Publication Bias in the Social Sciences and Economics," Journal of Economics and Statistics (Jahrbuecher fuer Nationaloekonomie und Statistik), De Gruyter, vol. 231(5-6), pages 661-684, October.
    18. Irsova, Zuzana & Bom, Pedro Ricardo Duarte & Havranek, Tomas & Rachinger, Heiko, 2023. "Spurious Precision in Meta-Analysis," MetaArXiv 3qp2w, Center for Open Science.
    19. Hamilton, Daniel George & Fraser, Hannah & Hoekstra, Rink & Fidler, Fiona, 2020. "Journal policies and editors’ opinions on peer review," MetaArXiv qkjy4_v1, Center for Open Science.

    More about this item

    Statistics

    Access and download statistics

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:osf:metaar:eqhjd_v1. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    If CitEc recognized a bibliographic reference but did not link an item in RePEc to it, you can help with this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: OSF (email available below). General contact details of provider: https://osf.io/preprints/metaarxiv .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.