Estimation of effect sizes in the presence of publication bias: a comparison of meta-analysis methods

Similar documents
Tilburg University. Conducting meta-analyses based on p values van Aert, Robbie; Wicherts, Jelte; van Assen, Marcus

C2 Training: August 2010

Meta-Analysis and Publication Bias: How Well Does the FAT-PET-PEESE Procedure Work?

Overstated Evidence for Short-Term Effects of Violent Games on Affect and Behavior: A Reanalysis of Anderson et al. (2010)

Meta-Analysis. Zifei Liu. Biological and Agricultural Engineering

Examining publication bias A simulation-based evaluation of statistical tests on publication bias

Performance of the Trim and Fill Method in Adjusting for the Publication Bias in Meta-Analysis of Continuous Data

Treatment effect estimates adjusted for small-study effects via a limit meta-analysis

Research Methods in Forest Sciences: Learning Diary. Yoko Lu December Research process

Recent Developments: Should We WAAP Economics Low Power?

Meta-analysis: Basic concepts and analysis

A Short Primer on Power Calculations for Meta-analysis

Meta-analysis: Advanced methods using the STATA software

Bias-correction techniques alone cannot determine whether ego depletion is different from

How to Conduct a Meta-Analysis

An Empirical Assessment of Bivariate Methods for Meta-analysis of Test Accuracy

Cochrane Pregnancy and Childbirth Group Methodological Guidelines

Meta Analysis. David R Urbach MD MSc Outcomes Research Course December 4, 2014

Running head: META-ANALYSIS OF SELF-CONTROL TRAINING

Incorporating Within-Study Correlations in Multivariate Meta-analysis: Multilevel Versus Traditional Models

Two-Way Independent Samples ANOVA with SPSS

Unit 1 Exploring and Understanding Data

Meta-Analysis of Correlation Coefficients: A Monte Carlo Comparison of Fixed- and Random-Effects Methods

Search settings MaxQuant

Bayesian Mixture Modeling of Significant P Values: A Meta-Analytic Method to Estimate the Degree of Contamination from H 0

Examining Relationships Least-squares regression. Sections 2.3

Evidence Based Medicine

Beyond RevMan: Meta-Regression. Analysis in Context

ECONOMICS SERIES SWP 2013/2. Updated July Better than Random: Weighted Least Squares Meta-Regression Analysis

Comparison of Estimates From An Address-Based Mail Survey And A RDD Telephone Survey

Methods Research Report. An Empirical Assessment of Bivariate Methods for Meta-Analysis of Test Accuracy

Meta-Econometrics. Tom Stanley, Professor of Economics, Hendrix College USA:

Comparison of Different Methods of Detecting Publication Bias

RAG Rating Indicator Values

Lec 02: Estimation & Hypothesis Testing in Animal Ecology

Method Comparison for Interrater Reliability of an Image Processing Technique in Epilepsy Subjects

Index. Springer International Publishing Switzerland 2017 T.J. Cleophas, A.H. Zwinderman, Modern Meta-Analysis, DOI /

Impact of Microfinance on Poverty and Microenterprises: A Meta-analysis

UC Davis UC Davis Previously Published Works

Meta-analysis of few small studies in small populations and rare diseases

Consequences of effect size heterogeneity for meta-analysis: a Monte Carlo study

The Meta on Meta-Analysis. Presented by Endia J. Lindo, Ph.D. University of North Texas

Fixing the replicability crisis in science. Jelte M. Wicherts

Analysis of left-censored multiplex immunoassay data: A unified approach

Statistical techniques to evaluate the agreement degree of medicine measurements

African Economic Conference, Johannesburg, South Africa October 2013

Models for potentially biased evidence in meta-analysis using empirically based priors

Live WebEx meeting agenda

Understandable Statistics

Statistical Techniques. Meta-Stat provides a wealth of statistical tools to help you examine your data. Overview

f WILEY ANOVA and ANCOVA A GLM Approach Second Edition ANDREW RUTHERFORD Staffordshire, United Kingdom Keele University School of Psychology

How to do a meta-analysis. Orestis Efthimiou Dpt. Of Hygiene and Epidemiology, School of Medicine University of Ioannina, Greece

The Impact of Cellphone Sample Representation on Variance Estimates in a Dual-Frame Telephone Survey

STATISTICS AND RESEARCH DESIGN

Methods of Reducing Bias in Time Series Designs: A Within Study Comparison

Fixed Effect Combining

Georgina Salas. Topics EDCI Intro to Research Dr. A.J. Herrera

Introduction to Statistical Data Analysis I

Simple Linear Regression the model, estimation and testing

REPRODUCTIVE ENDOCRINOLOGY

2.75: 84% 2.5: 80% 2.25: 78% 2: 74% 1.75: 70% 1.5: 66% 1.25: 64% 1.0: 60% 0.5: 50% 0.25: 25% 0: 0%

Power Posing: P-Curving the Evidence

Evaluating the results of a Systematic Review/Meta- Analysis

A NEW APPROACH TO THERAPEUTIC DRUG MONITORING OF TOBRAMYCIN IN CYSTIC FIBROSIS

Bias in regression coefficient estimates when assumptions for handling missing data are violated: a simulation study

The Essential Role of Pair Matching in. Cluster-Randomized Experiments. with Application to the Mexican Universal Health Insurance Evaluation

Received: 14 April 2016, Accepted: 28 October 2016 Published online 1 December 2016 in Wiley Online Library

Discontinuous Traits. Chapter 22. Quantitative Traits. Types of Quantitative Traits. Few, distinct phenotypes. Also called discrete characters

DEPARTMENT OF ECONOMICS AND FINANCE COLLEGE OF BUSINESS AND ECONOMICS UNIVERSITY OF CANTERBURY CHRISTCHURCH, NEW ZEALAND

Analyzing diastolic and systolic blood pressure individually or jointly?

How much does education improve intelligence? A meta-analysis. UNREVIEWED PREPRINT v2, 09 NOV 2017

Choice of axis, tests for funnel plot asymmetry, and methods to adjust for publication bias

Overview. Survey Methods & Design in Psychology. Readings. Significance Testing. Significance Testing. The Logic of Significance Testing

Detection of Unknown Confounders. by Bayesian Confirmatory Factor Analysis

Explaining heterogeneity

The validity of inferences about the correlation (covariation) between treatment and outcome.

Sample size and power calculations in Mendelian randomization with a single instrumental variable and a binary outcome

Running Head: UNAPPRECIATED HETEROGENEITY OF EFFECTS. The Unappreciated Heterogeneity of Effect Sizes:

The Essential Role of Pair Matching in. Cluster-Randomized Experiments. with Application to the Mexican Universal Health Insurance Evaluation

Week 17 and 21 Comparing two assays and Measurement of Uncertainty Explain tools used to compare the performance of two assays, including

How Much Does Education Improve Intelligence? A Meta-Analysis

Study Guide for the Final Exam

2. How do different moderators (in particular, modality and orientation) affect the results of psychosocial treatment?

Fundamental Clinical Trial Design

Problem solving therapy

26:010:557 / 26:620:557 Social Science Research Methods

Differential Item Functioning

PLS 506 Mark T. Imperial, Ph.D. Lecture Notes: Reliability & Validity

Designing a Bayesian randomised controlled trial in osteosarcoma. How to incorporate historical data?

Chapter 5: Field experimental designs in agriculture

JSM Survey Research Methods Section

Case Studies in Bayesian Augmented Control Design. Nathan Enas Ji Lin Eli Lilly and Company

Supplemental material: Interference between number magnitude and parity: Discrete representation in number processing

Appendix 1: Supplementary tables [posted as supplied by author] Table A. Details of MeSH Search Terms

The moderating impact of temporal separation on the association between intention and physical activity: a meta-analysis

Estimation of delay to diagnosis and incidence in HIV using indirect evidence of infection dates

T-Statistic-based Up&Down Design for Dose-Finding Competes Favorably with Bayesian 4-parameter Logistic Design

6.3.5 Uncertainty Assessment

VALIDITY OF QUANTITATIVE RESEARCH

Preventive physiotherapy interventions for back care in children and adolescents: a meta-analysis

Transcription:

Estimation of effect sizes in the presence of publication bias: a comparison of meta-analysis methods Hilde Augusteijn M.A.L.M. van Assen R. C. M. van Aert APS May 29, 2016

Today s presentation Estimation of effect sizes in the presence of publication bias: a comparison of meta-analysis methods: PET-PEESE (Stanley & Doucouliagos, 2013) P-uniform (van Assen, van Aert & Wicherts, 2015)

Key messages Lot of evidence of publication bias Traditional meta-analysis methods cannot assess effect sizes accurately when there is publication bias P-uniform currently only accurately assesses effect sizes when there is no heterogeneity PET-PEESE does not work when sample sizes are homogeneous Methods need to be improved

Overview presentation Publication bias Meta-analysis PET-PEESE P-uniform Simulation study Results Recommendations

Publication bias Impact on meta-analytical results Overestimated effect sizes Inaccurate estimates of heterogeneity (τ2)

Meta-analysis Objectives of meta-analysis Estimating... 1. True effect 2. Heterogeneity 3. Moderator effects

Meta-analysis methods Traditional methods Fixed effects meta-analysis Random effects meta-analysis New methods: PET-PEESE P-uniform

Meta-analysis: traditional methods Fixed effects: assumes one underlying effect size Weighted average Weight: precision of study: within-study variance (σi2) Random effects: assumes multiple true effect sizes Weighted average Weight: precision & heterogeneity: between- study variance (σi2 +τ2)

Methods dealing with publication bias A lot of methods have been developed. So far, none of them estimates accurately under all circumstances

Methods dealing with publication bias New methods: PET-PEESE (Stanley & Doucouliagos, 2013) P-uniform (van Assen, van Aert & Wicherts, 2015) & P-curve (Simonsohn, Nelson & Simmons, 2014 )

PET-PEESE When there is no publication bias, there is no relation between effect sizes and precision

PET-PEESE 0.1 0.3 PEESE PET No publication bias 0.1 0.3 0.1 0.3 Plot of 300 randomly generated studies, true ES = 0.5, sample sizes differ from 10 to 150.

PET-PEESE When there is no publication bias, there is no relation between effect sizes and precision Publication bias smaller studies report larger effect sizes

PET-PEESE No publication bias 0.1 0.3 PEESE PET 0.1 0.1 0.3 0.3 Standard Error Standard Error 0.1 0.3 Plot of 300 randomly generated studies, true ES = 0.5, sample sizes differ from 10 to 150.

PET-PEESE When there is no publication bias, there is no relation between effect sizes and precision Publication bias smaller studies report larger effect sizes PET-PEESE uses the intercept of a regression line to estimate the effect size

PET-PEESE No publication bias 0.1 0.3 PEESE PET 0.1 0.3 0.1 0.3 Plot of 300 randomly generated studies, true ES = 0.5, sample sizes differ from 10 to 150.

PET-PEESE: homogeneous sample sizes X-axis: Standard error: related to sample size

PET-PEESE: homogeneous sample sizes 0.10 0.15 0 0 5 0.10 0.15 0 0 5 0.10 No publication bias PET PEESE 0.15 0 0.15 0 0.10 0 5 0.15 0 5 0 PEESE PET No publication bias 0 5 0.10 0.15 0 0 5 0.10 Plots of 100 randomly generated studies, true ES= 0.5 with sample sizes differing between 30 and 50.

PET-PEESE: homogeneous sample sizes 0.10 0.15 0 0 5 0.10 0.15 0 0 5 0.10 No publication bias PET PEESE 0.15 0 0.15 0 0.10 0 5 0.15 0 5 0 PEESE PET No publication bias 0 5 0.10 0.15 0 0 5 0.10 Plots of 100 randomly generated studies, true ES= 0.5 with sample sizes differing between 30 and 50.

P-uniform: Distribution of (conditional) p-values is uniform Uses only statistically significant studies Assumption: All significant studies are equally likely to be published Distribution of (conditional) p-values is uniform under the true effect Currently assumes a fixed effect: Does not provide accurate estimates when there is heterogeneity

Simulation study

Simulation study Independent variables True effect size (4 levels: Cohen s d of 0,, 0.5, ) Level of publication bias (4 levels: 0%, 75%, 95%, 100%) Sample size (2 levels: homogeneous & heterogeneous, average sample size = 100) Amount of heterogeneity (3 levels: τ2= 0, 1, 9; I2=0%, 50%, 90% when N = 100) Number of studies (1 level: Number of significant studies = 5) Total: 96 conditions, 10,000 iterations

Results: expectations Performance of fixed- and random-effects suffer from publication bias Performance of PET-PEESE suffers from homogeneous sample sizes Performance of P-uniform suffers from heterogeneity in effect sizes

Simulation study Dependent variables Bias in estimated effect size (Type I error rate) Power Coverage RMSE

Results: estimation of effect sizes Fixed & random Effects: overestimates effect sizes when there is publication bias

Results: estimation of effect sizes Fixed & random Effects ES d=0.5 d= level of publication bias d= d= ES RE: median estimates of ES FE: median estimates of ES level of publication bias

Results: estimation of effect sizes Fixed & random Effects: overestimates effect sizes when there is publication bias PET-PEESE: underestimates effect size when sample sizes are homogeneous

Results: estimation of effect sizes PET PEESE median estimates of ES Heterogeneous sample sizes level of publication bias ES d= d= d=0.5 d= ES Homogeneous sample sizes level of publication bias

Results: estimation of effect sizes Fixed & random Effects: overestimates effect sizes when there is publication bias PET-PEESE: underestimates effect size when sample sizes are homogeneous P-uniform: overestimates effect sizes when there is heterogeneity

Results: estimation of effect sizes P uniform median estimates of ES ES level of publication bias d= d= d=0.5 d= level of publication bias ES d= d= d=0.5 d= ES Large amount of heterogeneity Medium amount of heterogeneity No heterogeneity level of publication bias

Results: Power Ideal situation: power is at least.80 Fixed & Random effects: power is almost always close to 0 PET-PEESE: when sample sizes are homogeneous, power is almost always equal to.05 P-uniform: power is too low when the true effect size is small. Power increases when there is heterogeneity: due to overestimation of the effect.

Results: Power PET PEESE: Power Heterogeneous sample sizes level of publication bias power d= d= d=0.5 d= power Homogeneous sample sizes level of publication bias

Results: Power Ideal situation: power is at least.80 Fixed & Random effects: power is almost always close to 0 PET-PEESE: when sample sizes are homogeneous, power is almost always equal to.05 P-uniform: power is too low when the true effect size is small. Power increases when there is heterogeneity: due to overestimation of the effect.

Results: Coverage Ideal situation: 95% coverage FE/RE: coverage decreases when publication bias increases (up to 0% coverage) PET-PEESE: coverage 95% when sample sizes are homogeneous (large CI), lower when sample sizes are heterogeneous P-uniform: coverage = 95% when there is no heterogeneity in effect sizes, decreases when heterogeneity increases.

Results: RMSE Root-mean-squared-error: magnitude of error FE/RE: RMSE increases as publication bias increases PET-PEESE: Large RMSE values when sample sizes are homogeneous P-uniform: RMSE is only influenced by δ, not by heterogeneity and publication bias

Key messages Lot of evidence of publication bias Traditional meta-analysis methods cannot assess effect sizes accurately when there is publication bias P-uniform currently only accurately assesses effect sizes when there is no heterogeneity PET-PEESE does not work when sample sizes are homogeneous Methods need to be improved

Recommendations If you know there is no publication bias: Use fixed/random effects Do not use PET-PEESE if the sample sizes are homogeneous (or when there is a large amount of heterogeneity) Only use P-uniform if you know there is no or a small amount of heterogeneity Pay attention to new methods being published!

Questions/Remarks? H.E.M.Augusteijn@tilburguniversity.edu www.metaresearch.nl