What is indirect comparison?

Similar documents
Papers. Validity of indirect comparison for estimating efficacy of competing interventions: empirical evidence from published meta-analyses.

Agenda. Frequent Scenario WS I: Treatment Comparisons and Network Meta-Analysis: Learning the Basics

PROTOCOL. Francesco Brigo, Luigi Giuseppe Bongiovanni

Statistical considerations in indirect comparisons and network meta-analysis

How to Conduct a Meta-Analysis

Statistical considerations in indirect comparisons and network meta-analysis

Network Meta-Analysis for Comparative Effectiveness Research

GUIDELINE COMPARATORS & COMPARISONS:

Meta-Analysis. Zifei Liu. Biological and Agricultural Engineering

Estimating the Power of Indirect Comparisons: A Simulation Study

Simulation evaluation of statistical properties of methods for indirect and mixed treatment comparisons

The role of MTM in Overviews of Reviews

Problem solving therapy

Addressing multiple treatments II: multiple-treatments meta-analysis basic methods

Results. NeuRA Hypnosis June 2016

Comparison of Meta-Analytic Results of Indirect, Direct, and Combined Comparisons of Drugs for Chronic Insomnia in Adults: A Case Study

Methods for assessing consistency in Mixed Treatment Comparison Meta-analysis

Animal-assisted therapy

Distraction techniques

Results. NeuRA Treatments for dual diagnosis August 2016

Models for potentially biased evidence in meta-analysis using empirically based priors

Meta-analyses: analyses:

ISPOR Task Force Report: ITC & NMA Study Questionnaire

Current Issues with Meta-analysis in Medical Research

Traumatic brain injury

Evidence-Based Medicine and Publication Bias Desmond Thompson Merck & Co.

Role of evidence from observational studies in the process of health care decision making

Cochrane Pregnancy and Childbirth Group Methodological Guidelines

Results. NeuRA Treatments for internalised stigma December 2017

What is meta-analysis?

Meta-analysis of safety thoughts from CIOMS X

Sample size and power considerations in network meta-analysis

Results. NeuRA Family relationships May 2017

Evaluating the results of a Systematic Review/Meta- Analysis

Results. NeuRA Forensic settings April 2016

Results. NeuRA Herbal medicines August 2016

Results. NeuRA Mindfulness and acceptance therapies August 2018

Evidence based urology in practice: heterogeneity in a systematic review meta-analysis. Health Services Research Unit, University of Aberdeen, UK

Indirect Comparisons: heroism or heresy. Neil Hawkins With acknowledgements to Sarah DeWilde

Choice of axis, tests for funnel plot asymmetry, and methods to adjust for publication bias

Statistical considerations in indirect comparisons and network meta-analysis

Comparing treatments evaluated in studies forming disconnected networks of evidence: A review of methods

NeuRA Sleep disturbance April 2016

A protocol for a systematic review on the impact of unpublished studies and studies published in the gray literature in meta-analyses

Empirical evidence on sources of bias in randomised controlled trials: methods of and results from the BRANDO study

Introduction to meta-analysis

Systematic Reviews. Simon Gates 8 March 2007

The effects of clinical and statistical heterogeneity on the predictive values of results from meta-analyses

Critical appraisal: Systematic Review & Meta-analysis

Applications of Bayesian methods in health technology assessment

Method. NeuRA Biofeedback May 2016

Evaluating the Quality of Evidence from a Network Meta- Analysis

Transparency and accuracy in reporting health research

A COMMON REFERENCE-BASED INDIRECT COMPARISON META-ANALYSIS OF ESLICARBAZEPINE VERSUS LACOSAMIDE AS ADD ON TREATMENTS FOR FOCAL EPILEPSY PROTOCOL

Results. NeuRA Worldwide incidence April 2016

How to do a meta-analysis. Orestis Efthimiou Dpt. Of Hygiene and Epidemiology, School of Medicine University of Ioannina, Greece

NeuRA Decision making April 2016

NeuRA Obsessive-compulsive disorders October 2017

The moderating impact of temporal separation on the association between intention and physical activity: a meta-analysis

Controlled Trials. Spyros Kitsiou, PhD

Transcranial Direct-Current Stimulation

Lack of association between IL-6-174G>C polymorphism and lung cancer: a metaanalysis

Agomelatine versus placebo: A meta-analysis of published and unpublished trials

School of Dentistry. What is a systematic review?

Types of Data. Systematic Reviews: Data Synthesis Professor Jodie Dodd 4/12/2014. Acknowledgements: Emily Bain Australasian Cochrane Centre

CHAMP: CHecklist for the Appraisal of Moderators and Predictors

Why published medical research may not be good for your health

Papers. Abstract. Methods. Introduction. David Gunnell, Julia Saperia, Deborah Ashby

Clinical research in AKI Timing of initiation of dialysis in AKI

GLOSSARY OF GENERAL TERMS

Method. NeuRA Paliperidone August 2016

Between-trial heterogeneity in meta-analyses may be partially explained by reported design characteristics

GRADE: Applicazione alle network meta-analisi

Editorial considerations for reviews that compare multiple interventions

Statistical considerations in indirect comparisons and network meta-analysis

Live WebEx meeting agenda

Your submission states that the adjusted indirect comparison methodology was based largely on the publication by Glenny et al

Results. NeuRA Motor dysfunction April 2016

Results. NeuRA Essential fatty acids August 2016

Introduction to systematic reviews/metaanalysis

Safinamide (Addendum to Commission A15-18) 1

Using Indirect Comparisons to Support a Health Technology Assessment (HTA)

Trials and Tribulations of Systematic Reviews and Meta-Analyses

Meta-analysis: Basic concepts and analysis

Surveillance report Published: 13 April 2017 nice.org.uk. NICE All rights reserved.

Indirect comparison of the efficacy and safety of gefitinib and cetuximab based therapy in patients with advanced non small cell lung cancer

Principles of meta-analysis

GRADE. Grading of Recommendations Assessment, Development and Evaluation. British Association of Dermatologists April 2014

S160 #15. Comparing Two Proportions, Part 3 Odds Ratio. JC Wang. March 15, 2016

Cochrane Breast Cancer Group

Recent developments for combining evidence within evidence streams: bias-adjusted meta-analysis

GRADE. Grading of Recommendations Assessment, Development and Evaluation. British Association of Dermatologists April 2018

SYSTEMATIC REVIEW: AN APPROACH FOR TRANSPARENT RESEARCH SYNTHESIS

Capture-recapture method for assessing publication bias

A note on the graphical presentation of prediction intervals in random-effects meta-analyses

The Royal College of Pathologists Journal article evaluation questions

Comparative Effectiveness Research Collaborative Initiative (CER-CI) PART 1: INTERPRETING OUTCOMES RESEARCH STUDIES FOR HEALTH CARE DECISION MAKERS

Cardiovascular disease and varenicline (Champix)

Performance of the Trim and Fill Method in Adjusting for the Publication Bias in Meta-Analysis of Continuous Data

Transcription:

...? series New title Statistics Supported by sanofi-aventis What is indirect comparison? Fujian Song BMed MMed PhD Reader in Research Synthesis, Faculty of Health, University of East Anglia Indirect comparison refers to a comparison of different healthcare interventions using data from separate studies, in contrast to a direct comparison within randomised controlled trials. Indirect comparison is often used because of a lack of, or insufficient, evidence from head-to-head comparative trials. Naive indirect comparison is a comparison of the results of individual arms from different trials as if they were from the same randomised trials. This method provides evidence equivalent to that of observational studies and should be avoided in the analysis of data from randomised trials. Adjusted indirect comparison (including mixed treatment comparison) is an indirect comparison of different treatments adjusted according to the results of their direct comparison with a common control, so that the strength of the randomised trials is preserved. Empirical evidence indicates that results of adjusted indirect comparison are usually, but not always, consistent with the results of direct comparison. For further titles in the series, visit: www.whatisseries.co.uk Basic assumptions underlying indirect comparisons include a homogeneity assumption for standard meta-analysis, similarity assumption for adjusted indirect comparison and consistency assumption for the combination of direct and indirect evidence. It is essential to fully understand and appreciate these basic assumptions in order to use adjusted indirect and mixed treatment comparisons appropriately. 1

What is Why use indirect comparison? For many clinical indications, available treatments constantly increase over time. For example, a large number of different old and new drugs are available for the treatment of depressive and other mental disorders. Clinicians, patients and health policy-makers often need to decide which treatment is the most cost-effective. Well-conducted randomised controlled trials provide the most valid estimates of the relative efficacy of competing healthcare interventions. However, many interventions have not been directly compared in randomised controlled trials. When there is no, or insufficient, evidence from direct comparison trials, it may be possible to use results of different trials to estimate the relative effects of different treatments. 1 In contrast to direct within-trial comparison, indirect comparison means a between-study comparison of different interventions (Figure 1). Similarity } Trial A 1 vs C 1 Indirect A 1 B vs 2 C 1 C 2 Direct vs A3 B 3 Trial A 3 vs B 3 Trial B 2 vs C 2 } Consistency Figure 1. Direct and adjusted indirect comparison of treatments A and B Methods for indirect comparison Methods for indirect comparison can be classified into several categories: naive indirect comparison, informal indirect comparison, and formal adjusted indirect comparison. 1 To illustrate these methods, we use an example of risperidone versus haloperidol for schizophrenia, which includes three placebo controlled trials of risperidone, nine placebo controlled trials of haloperidol, and ten trials that directly compared risperidone and haloperidol (Table 1 and Figure 2). 2 In a naive indirect comparison, results of individual arms from different trials are compared as if they were from the same randomised controlled trial. Because the advantage of the randomised trials is completely disregarded, evidence from this naive or unadjusted indirect comparison is equivalent to evidence from observational studies and has increased susceptibility to bias. The effect of a treatment may be over- or underestimated because of bias, resulting in flawed recommendations for clinical practice. Therefore, naive indirect comparison should not be used to analyse data from randomised controlled trials. 1 In informal indirect comparison, neither relative effects nor statistical significance are formally calculated. For example, both risperidone and haloperidol were more effective than placebo in terms of clinical improvement (Table 1 and Figure 2). 2 An informal indirect comparison can be made by examining the point estimates and corresponding 95% confidence intervals (CIs) of the two odds ratios. The odds ratio of haloperidol versus placebo (0.18; 95% CI: 0.10, 0.34) suggested a greater treatment effect than the odds ratio of risperidone versus placebo (0.40; 95% CI: 0.26, 0.62), despite some overlap of the two CIs. 2 Formal adjusted indirect comparison is an indirect comparison of competing interventions adjusted according to the results of their direct comparison with a common 2

Table 1. Meta-analyses of risperidone versus haloperidol for schizophrenia: number of patients without clinical improvement 2 Comparison Number Log odds Odds ratio I 2 % of trials ratio (SE) (95% CI) Placebo controlled trials Risperidone vs placebo 3 0.909 (0.218) 0.40 (0.26, 0.62) 37% Haloperidol vs placebo 9 1.707 (0.318) 0.18 (0.10, 0.34) 11% Risperidone vs haloperidol Direct comparison 10 0.262 (0.142) 0.77 (0.58, 1.02) 14% Adjusted indirect 3/9 0.798 (0.386) 2.22 (1.04, 4.72) comparison Combination of direct 10+(3/9) 0.207 (0.527) 1.23 (0.44, 3.45) 85% and indirect estimates NB Random-effects model was used in meta-analyses of trials and for the combination of the direct and indirect estimates. Odds ratio = EXP(log odds ratio). CI: confidence interval; SE: standard error Treatment 1 vs treatment 2 Risperidone vs placebo Haloperidol vs placebo Risperidone vs haloperidol Adjusted indirect estimate Odds ratio (95% CI) 0.1 1 10 Bucher and colleagues suggested a simple method of adjusted indirect comparison, in which the indirect comparison of A and B is adjusted according to the results of their direct comparisons with a common intervention C. 3 This is illustrated in Figure 1. Let lnor AC denote log odds ratio of A 1 versus C 1 in trial 1, and lnor BC denote log odds ratio of B 2 versus C 2 in trial 2. The log odds ratio of the adjusted indirect comparison of A and B (lnor AB ) can be estimated by: Direct estimate lnor AB = lnor AC lnor BC Combination of direct and indirect CI = confidence interval Favours treatment 1 Favours treatment 2 Figure 2. Results of different methods of comparing risperidone and haloperidol for schizophrenia. Outcome: not clinically improved 2 control. This means that the advantage of the randomised controlled trials can be at least partly preserved in the indirect comparison. Suppose that interventions A and C were compared in randomised trial 1 (with groups A 1 and C 1 ), and that interventions B and C were compared in trial 2 (with groups B 2 and C 2 ) (Figure 1). To overcome the potential problem of different prognostic factors between study participants in different trials, and its standard error is: SE(lnOR AB ) = [SE(lnOR AC ) 2 + SE(lnOR BC ) 2 ] Although an odds ratio is used in the above equations, this adjusted method of indirect comparison may also be used when the relative efficacy is measured by relative risk, risk difference and mean difference. Empirical evidence indicates that results of adjusted indirect comparison are usually, but not always, similar to those of direct comparison trials. 4 More complex methods for adjusted indirect comparison, such as network metaanalysis and mixed treatment comparison, have also been developed to conduct adjusted indirect comparison of multiple healthcare 3

interventions, and to combine evidence from direct and indirect comparisons. 5 Table 1 and Figure 2 show the results of an adjusted indirect comparison of risperidone and haloperidol, using the results of the two meta-analyses of placebo controlled trials. 2 The point estimate of log odds ratio by adjusted indirect comparison is: lnor = 0.909 ( 1.707) = 0.798 and its standard error is: SE(lnOR ) = (0.218 2 + 0.318 2 ) = 0.386 The results of the adjusted indirect comparison suggest that risperidone was less effective than haloperidol for schizophrenia (odds ratio 2.22; 95% CI: 1.04, 4.72). 2 Risperidone and haloperidol were also compared in ten direct comparison trials (Table 1 and Figure 2). 2 According to data from head-to-head comparative trials, risperidone tended to be more effective than haloperidol (odds ratio 0.77; 95% CI: 0.58, 1.02), which conflicts with the results of the adjusted indirect comparison. 2 The discrepancy between the direct and indirect estimates is statistically significant (p=0.01). The issue of inconsistent evidence will be discussed further in the section on basic assumptions. The results of the direct comparison can be quantitatively combined with the indirect estimate (combined odds ratio 1.23; 95% CI: 0.44, 3.45) (Table 1 and Figure 2). 2 However, it is not clear whether the combined estimate is valid, given the significant discrepancy or inconsistency between the direct and indirect estimates. Basic assumptions underlying indirect comparisons Homogeneity assumption for standard meta-analysis When multiple trials are available for a given comparison, the results from multiple trials can be pooled in meta-analyses before an adjusted indirect comparison is conducted. In standard meta-analysis, it is assumed that different trials estimate the same single effect (fixed-effects model) or different effects are distributed around a typical value (randomeffects model); that is, for a meta-analysis to be valid, results from different trials should be sufficiently homogeneous. Heterogeneity in results across studies can be statistically tested using the Chi 2 test and quantified using I 2. 6 For example, the results of three placebo controlled trials of risperidone and the results of nine placebo controlled trials of haloperidol were combined in Table 1. 2 The underlying assumption was that these trials were sufficiently homogenous to be quantitatively combined. Similarity assumption for adjusted indirect comparison For an adjusted indirect comparison to be valid, a similarity assumption is required in terms of moderators of relative treatment effect; that is, patients included should be sufficiently similar in the two sets of placebo controlled trials, so that the relative effect estimated by trials of A versus C is generalisable to patients in trials of B versus C, and the relative effect estimated by trials of B versus C is generalisable to patients in trials of A versus C. Relative effects estimated in trials may also be associated with the methodological quality of the trials. Empirical evidence indicates that trials with poor quality may report greater treatment effects than good quality trials, particularly when outcomes are subjectively measured. It can be mathematically proven that the results of adjusted indirect comparison will be unbiased when the two sets of trials are similarly biased. 2 Therefore, both clinical similarity and methodological similarity should be considered in adjusted indirect comparison. If the trial similarity assumption is not fulfilled, estimates from adjusted indirect comparisons will be invalid and misleading. For the adjusted indirect comparison of risperidone and haloperidol (Table 1), 2 patient characteristics, dose of drug and treatment duration were similar between trials. However, the clinical improvement was defined differently between the two sets of placebo controlled trials. In placebo controlled trials of risperidone, it was predefined as a 20% or greater reduction in 4

the total score on the Positive and Negative Syndrome Scale or Brief Psychiatric Rating Scale, but in placebo controlled trials of haloperidol, it was rated by clinicians using the Clinical Global Impression or other scales. 2 Therefore, the results of adjusted indirect comparison should be interpreted cautiously. Consistency assumption for combining direct and indirect estimates When both direct and indirect evidence is available, an assumption of evidence consistency is required to quantitatively combine the direct and indirect estimates. We can compare the direct estimate with the indirect estimate to examine whether the two estimates are consistent. For example, adjusted indirect comparison of risperidone and haloperidol provided inconsistent results compared with the results of direct comparison (Table 1 and Figure 2). 2 Although we can quantitatively combine the two estimates, the pooled estimate (odds ratio 1.23; 95% CI: 0.44, 3.45) may be invalid and misleading. It is important to investigate possible causes of discrepancy between the direct and indirect evidence, such as the play of chance, invalid indirect comparison, bias in head-to-head comparative trials, and clinically meaningful heterogeneity across trials. When the direct comparison differs from the adjusted indirect comparison, we should usually give more credibility to evidence from head-to-head comparative trials. However, evidence from direct comparative trials may not always be valid. Some case studies indicated that adjusted indirect comparison may provide less biased results than head-tohead comparative trials under certain circumstances. 2 Appropriate use of indirect comparisons A recent survey of published systematic reviews found that methodological problems in the use of indirect comparisons for evaluating healthcare interventions include the use of inappropriate or flawed methods, the lack of objective or validated methods to assess trial similarity, and inadequate comparison or inappropriate combination of direct and indirect evidence. 7 In particular, mixed treatment comparison has often been used to combine direct and indirect evidence without explicit assessment of evidence consistency. To use indirect and mixed treatment comparison appropriately, we need to adequately understand and fully appreciate the basic assumptions underlying valid indirect approaches. The assessment of the similarity of trials involved in adjusted indirect comparison and the assessment of evidence consistency in mixed treatment comparison should be more explicit and systematic. References 1. Glenny AM, Altman DG, Song F et al. Indirect comparisons of competing interventions. Health Technol Assess 2005; 9: 1 134, iii iv. 2. Song F, Harvey I, Lilford R. Adjusted indirect comparison may be less biased than direct comparison for evaluating new pharmaceutical interventions. J Clin Epidemiol 2008; 61: 455 463. 3. Bucher HC, Guyatt GH, Griffith LE, Walter SD. The results of direct and indirect treatment comparisons in meta-analysis of randomized controlled trials. J Clin Epidemiol 1997; 50: 683 691. 4. Song F, Altman DG, Glenny AM, Deeks JJ. Validity of indirect comparison for estimating efficacy of competing interventions: empirical evidence from published metaanalyses. BMJ 2003; 326: 472. 5. Caldwell DM, Ades AE, Higgins JP. Simultaneous comparison of multiple treatments: combining direct and indirect evidence. BMJ 2005; 331: 897 900. 6. Higgins JP, Thompson SG. Quantifying heterogeneity in a meta-analysis. Stat Med 2002; 21: 1539 1558. 7. Song F, Loke YK, Walsh T et al. Methodological problems in the use of indirect comparisons for evaluating healthcare interventions: a survey of published systematic reviews. BMJ 2009: In press. Further reading The references listed above are suggested as useful further reading. 5

...? series What is This publication, along with the others in the series, is available on the internet at www.whatisseries.co.uk The data, opinions and statements appearing in the article(s) herein are those of the contributor(s) concerned. Accordingly, the sponsor and publisher, and their respective employees, officers and agents, accept no liability for the consequences of any such inaccurate or misleading data, opinion or statement. Supported by sanofi-aventis Published by Hayward Medical Communications, a division of Hayward Group Ltd. Copyright 2009 Hayward Group Ltd. All rights reserved. 6