Statistical methods for reliably updating meta-analyses

Similar documents
False statistically significant findings in cumulative metaanalyses and the ability of Trial Sequential Analysis (TSA) to identify them.

Should Cochrane apply error-adjustment methods when conducting repeated meta-analyses?

Trial Sequential Analysis (TSA)

heterogeneity in meta-analysis stats methodologists meeting 3 September 2015

APPLYING THE LAW OF ITERATED LOGARITHM TO CUMULATIVE META-ANALYSIS OF A CONTINUOUS ENDPOINT

Int. J. Epidemiol. Advance Access published September 29, 2008

Cochrane Scientific Committee AGENDA

Thresholds for statistical and clinical significance in systematic reviews with meta-analytic methods

Sample size and power considerations in network meta-analysis

CINeMA. Georgia Salanti & Theodore Papakonstantinou Institute of Social and Preventive Medicine University of Bern Switzerland

The emerging evidence synthesis tools: Actively Living Network Meta- Analysis

Using Statistical Principles to Implement FDA Guidance on Cardiovascular Risk Assessment for Diabetes Drugs

Statistical considerations in indirect comparisons and network meta-analysis

Models for potentially biased evidence in meta-analysis using empirically based priors

Statistical considerations in indirect comparisons and network meta-analysis

Statistical considerations in indirect comparisons and network meta-analysis

Assessing imprecision in Cochrane systematic reviews: a comparison of GRADE and Trial Sequential Analysis

Meta-analyses triggered by previous (false-)significant findings: problems and solutions

How to Conduct a Meta-Analysis

Methods for assessing consistency in Mixed Treatment Comparison Meta-analysis

Dear Dr. Villanueva,

Meta analysis to integrate effect sizes within a paper: possible misuse and Type 1 error inflation

Study Protocol. Copenhagen Trial Unit, Centre for Clinical Intervention Research, Rigshospitalet, Copenhagen, Denmark 3

Supplementary Online Content

18/11/2013. An Introduction to Meta-analysis. In this session: What is meta-analysis? Some Background Clinical Trials. What questions are addressed?

Bayesian Dose Escalation Study Design with Consideration of Late Onset Toxicity. Li Liu, Glen Laird, Lei Gao Biostatistics Sanofi

Actively Living Network Meta-Analysis

BORROWING EXTERNAL CONTROLS FOR AN EVENT-DRIVEN PEDIATRIC TRIAL IN PAH: A CASE STUDY

Evidence synthesis with limited studies

Systematic review with multiple treatment comparison metaanalysis. on interventions for hepatic encephalopathy

Bayesian meta-analysis of Papanicolaou smear accuracy

Empirical evidence on sources of bias in randomised controlled trials: methods of and results from the BRANDO study

Practical Bayesian Optimization of Machine Learning Algorithms. Jasper Snoek, Ryan Adams, Hugo LaRochelle NIPS 2012

Indirect Comparisons: heroism or heresy. Neil Hawkins With acknowledgements to Sarah DeWilde

A Spreadsheet for Deriving a Confidence Interval, Mechanistic Inference and Clinical Inference from a P Value

Methods for evidence synthesis in the case of very few studies. Guido Skipka IQWiG, Cologne, Germany

MS&E 226: Small Data

Choice of axis, tests for funnel plot asymmetry, and methods to adjust for publication bias

Live WebEx meeting agenda

How to do a meta-analysis. Orestis Efthimiou Dpt. Of Hygiene and Epidemiology, School of Medicine University of Ioannina, Greece

Checking the counterarguments confirms that publication bias contaminated studies relating social class and unethical behavior

Applications of Bayesian methods in health technology assessment

Editorial considerations for reviews that compare multiple interventions

The role of MTM in Overviews of Reviews

Meta-analysis: Advanced methods using the STATA software

Allowing for uncertainty due to missing outcome data in meta-analysis

Quality and Reporting Characteristics of Network Meta-analyses: A Scoping Review

Performance of the Trim and Fill Method in Adjusting for the Publication Bias in Meta-Analysis of Continuous Data

Integrating Effectiveness and Safety Outcomes in the Assessment of Treatments

Papers. Validity of indirect comparison for estimating efficacy of competing interventions: empirical evidence from published meta-analyses.

Protocol Development: The Guiding Light of Any Clinical Study

Logistic regression. Department of Statistics, University of South Carolina. Stat 205: Elementary Statistics for the Biological and Life Sciences

Comparing treatments evaluated in studies forming disconnected networks of evidence: A review of methods

Living network meta-analysis compared with pairwise metaanalysis in comparative effectiveness research: empirical study

Meta-analysis of few small studies in small populations and rare diseases

Do the sample size assumptions for a trial. addressing the following question: Among couples with unexplained infertility does

Evaluating the Quality of Evidence from a Network Meta- Analysis

On the Use of Local Assessments for Monitoring Centrally Reviewed Endpoints with Missing Data in Clinical Trials*

Effective Implementation of Bayesian Adaptive Randomization in Early Phase Clinical Development. Pantelis Vlachos.

An investigation of the impact of using different methods for network metaanalysis: a protocol for an empirical evaluation

Systematic Reviews and meta-analyses of Diagnostic Test Accuracy. Mariska Leeflang

Frequentist Evaluation of Group Sequential Clinical Trial Designs

5/10/2010. ISPOR Good Research Practices for Comparative Effectiveness Research: Indirect Treatment Comparisons Task Force

What is indirect comparison?

Statistics for Clinical Trials: Basics of Phase III Trial Design

Adaptive Treatment Arm Selection in Multivariate Bioequivalence Trials

Treatment effect estimates adjusted for small-study effects via a limit meta-analysis

Advanced Issues in Multiple Treatment Comparison Meta-analysis

A Belief-Based Account of Decision under Uncertainty. Craig R. Fox, Amos Tversky

Bayesian hierarchical modelling

Meta-Analysis. Zifei Liu. Biological and Agricultural Engineering

Between-trial heterogeneity in meta-analyses may be partially explained by reported design characteristics

A COMPARISON OF IMPUTATION METHODS FOR MISSING DATA IN A MULTI-CENTER RANDOMIZED CLINICAL TRIAL: THE IMPACT STUDY

Simulation evaluation of statistical properties of methods for indirect and mixed treatment comparisons

Meta-analysis of two studies in the presence of heterogeneity with applications in rare diseases

SWITCH Trial. A Sequential Multiple Adaptive Randomization Trial

How to use prior knowledge and still give new data a chance?

Imputation methods for missing outcome data in meta-analysis of clinical trials

Comparative Effectiveness Research Collaborative Initiative (CER-CI) PART 1: INTERPRETING OUTCOMES RESEARCH STUDIES FOR HEALTH CARE DECISION MAKERS

Power & Sample Size. Dr. Andrea Benedetti

Full title: A likelihood-based approach to early stopping in single arm phase II cancer clinical trials

Cochrane Pregnancy and Childbirth Group Methodological Guidelines

Bayesian Nonparametric Methods for Precision Medicine

Modelling the effects of drug dose in hierarchical network meta-analysis models to account for variability in treatment classifications

Statistical considerations in indirect comparisons and network meta-analysis

Comparison of Futility Monitoring Methods Using RTOG Clinical Trials. Q. Ed Zhang, PhD

University of Bristol - Explore Bristol Research

Analyzing binary outcomes, going beyond logistic regression

Comparison of Meta-Analytic Results of Indirect, Direct, and Combined Comparisons of Drugs for Chronic Insomnia in Adults: A Case Study

research methods & reporting

University of Bristol - Explore Bristol Research. Peer reviewed version. Link to published version (if available): /jrsm.

Estimating the Power of Indirect Comparisons: A Simulation Study

Advanced IPD meta-analysis methods for observational studies

RESEARCH. Sripal Bangalore, 1 Robert Fakheri, 1 Simon Wandel, 3 Bora Toklu, 2 Jasmin Wandel, 4 Franz H Messerli 5,6,7. open access

eappendix A. Simulation study parameters

Cognitive Science and Bayes. The main question. Several views. Introduction. The main question Several views Examples. The heuristics view

Using historical data for Bayesian sample size determination

Sample Size and Power. Objectives. Take Away Message. Laura Lee Johnson, Ph.D. Statistician

Discussion Meeting for MCP-Mod Qualification Opinion Request. Novartis 10 July 2013 EMA, London, UK

Designing a Bayesian randomised controlled trial in osteosarcoma. How to incorporate historical data?

Transcription:

Statistical methods for reliably updating meta-analyses Mark Simmonds University of York, UK With: Julian Elliott, Joanne McKenzie, Georgia Salanti, Adriani Nikolakopoulou, Julian Higgins

Updating meta-analyses When should we update a meta-analysis? When new studies emerge? When new data might alter our conclusions? Updating is time-consuming

Some issues When can we stop updating? Which meta-analyses should have priority for updating? Conclusions can change over time Risk of error if we stop too soon

Type I error Assuming an intervention is effective when it isn t Usually set at 5% Increases the more updates we perform Can we accept a conventionally statistically significant meta-analysis?

Cumulative meta-analysis It works! OK, maybe not It s a failure! OK, maybe not

Type I error in meta-analyses

Type II error Assuming an intervention isn t effective when it is Not controlled in a meta-analysis When can we stop updating non-significant meta-analyses?

Cumulative meta-analysis Doesn t look promising Give up now? Definitely stop now Oh wait

A caveat The summary effect estimates (and confidence intervals?) are valid at each update Decisions made on the basis of the results may not be Particularly decisions about whether to update

Parallels with sequential trial design Aim to stop a trial as soon as possible review Select a desired Type I and II error rate And desired clinical effect Perform interim analyses throughout trial Meta- review

Key differences Meta-analysis is not controlled No control over timing of studies Size of studies Heterogeneity Studies have different protocols Estimated effects may not be consistent

Controlling error Control Type I and Type II error Sequential meta-analysis Trial sequential analysis Control Type I error Law of Iterated Logarithm Shuster-Pocock method Other methods Fully Bayesian analysis Robustness or stability of analysis Consequences of adding new studies Power gains from adding new studies

Example from Cochrane I 2 = 95%

Cumulative meta-analysis

Sequential meta-analysis (SMA) Higgins, Simmonds, Whitehead 2010 Calculate cumulative Z score and cumulative Information for each updated meta-analysis Stop when a pre-specified boundary is crossed Boundary designed to control type I and II error

Accounting for heterogeneity Select a prior estimate of heterogeneity Generally assuming high heterogeneity Use Bayesian methods to calculate posterior heterogeneity estimate at each update Use this Bayesian estimate in the updated meta-analysis

Sequential meta-analysis

Trial sequential analysis (TSA) Wetterslev, Thorlund, Brok, Gluud 2008 Select a required sample size for the metaanalysis Calculate alpha-spending boundaries Stop if Z score exceeds the boundary Or if sample size is reached Sample size must be adjusted for heterogeneity

Example

Law of Iterated Logarithm (LIL) Uses an adjusted Z statistic Z = Z λ log log N This is bounded as N So controls Type I error Lan, Hu, Cappelleri 2007 Commonly sets λ = 2

Example

Shuster-Pocock method Shuster, Neu 2013 Compares the Z statistic to a t distribution Parameters of t distribution are based on Pocock s group sequential boundaries Must specify number of meta-analyses performed

Example

76 Cochrane Reviews 76 Reviews: 286 meta-analyses Binary outcome 194 (68%) Continuous outcome 92 Stat. sig. 178 (62%) Not stat. sig. 108 Trials per MA Median 9 IQR: 6 to 14 Max: 200 Effect size * Median 0.47 If stat sig. 0.69 If not 0.25 I 2 I 2 = 0: 32% I 2 > 90%: 7.0% * Log odds ratio or standardised mean difference If stat sig. 46% If not: 13%

Applying meta-analysis updating methods Apply to all 286 meta-analyses: Naïve cumulative meta-analysis Trial sequential analysis (heterogeneity adjusted) Sequential meta-analysis With no prior, 50% I 2 and 90% I 2 priors Law of iterated logarithm Shuster-Pocock

Conclusions of analyses

Extra trials required to reach a conclusion

Realistic review updating Have assumed a new meta-analysis after each new trial In reality updates are less frequent First analysis will have good proportion of total trials Re-analyse assuming updates once 50%, 70%, 90% and 100% of trials are available

Conclusions using realistic updating

Simulation study Simulated meta-analyses varying: True treatment effect: 0 or 0.1 Number of studies: 5 to 50 Heterogeneity: I 2 0 to 90% Fixed total sample size of 9000 90% power to detect effect of 0.1 if I 2 = 50%

Methods applied Naïve analysis (standard cumulative MA) Trial Sequential Analysis (TSA) Sequential Meta-Analysis (SMA) No prior heterogeneity Prior I 2 of 50% or 90% Law of Iterated Logarithm (LIL) Shuster method

False positive rates Type I error 20 trials / updates, I 2 = 25%

False positive rates Type I error

Cumulative power 20 trials / updates, I 2 = 25%

Cumulative power

Conventional Naïve analysis Too many inappropriate positive conclusions Elevated Type I error rate But not vastly elevated for most updated reviews? Biased estimates of effect Half of all analyses showing significant results are based on too little evidence?

Trial Sequential Analysis Controls for Type I and II error Need to set desired effect Complex to run Required sample size varies with time Can lead to inconsistent updates

Sequential Meta-Analysis Controls for Type I and II error Need to set desired effect Complex to run Statistical information not intuitive Limited choice of boundaries Bayesian heterogeneity too conservative? Not needed in practice?

Law of Iterated Logarithm Controls for Type I error Easy to implement Biased estimates of effect at stopping? Over-conservative: low-power Uncertainty over parameter

Shuster-Pocock Controls for Type I error Fairly easy to implement Needs more trials before stopping Need to pre-specify number of updates? Needs many studies to have adequate power

Do we need these methods? Is the problem with naïve analysis serious enough in real Cochrane reviews? Do the methods needlessly delay a statistically significant result? Too much focus on decision making over estimation? More complex than necessary?

When should they be implemented? At protocol stage in all reviews? At first update? Only once a statistically significant result is found? Only when evidence is limited? E.g. small total sample size

What are Cochrane reviews for? To present the best evidence at the current time? To aid in making medical decisions or guiding future trials?