Population-adjusted treatment comparisons Overview and recommendations from the NICE Decision Support Unit

Similar documents
University of Bristol - Explore Bristol Research. Publisher's PDF, also known as Version of record

Workshop 21: INDIRECT TREATMENT COMPARISONS: AN INTERACTIVE WORKSHOP ON CHOOSING THE RIGHT TOOL FOR THE AVAILABLE DATA

Partitioned survival analysis for decision modelling in health care: a critical review

Meta-Analysis. Zifei Liu. Biological and Agricultural Engineering

Comparing treatments evaluated in studies forming disconnected networks of evidence: A review of methods

Indirect Treatment Comparison Without Network Meta-Analysis: Overview of Novel Techniques

Recent developments for combining evidence within evidence streams: bias-adjusted meta-analysis

Daratumumab for treating relapsed and refractory multiple myeloma after a proteasome inhibitor and an immunomodulatory agent [ID933]

GUIDELINE COMPARATORS & COMPARISONS:

Methods for assessing consistency in Mixed Treatment Comparison Meta-analysis

Models for potentially biased evidence in meta-analysis using empirically based priors

The Late Pretest Problem in Randomized Control Trials of Education Interventions

TREATMENT EFFECTIVENESS IN TECHNOLOGY APPRAISAL: May 2015

Using Indirect Comparisons to Support a Health Technology Assessment (HTA)

NICE DSU TECHNICAL SUPPORT DOCUMENT 8: AN INTRODUCTION TO THE MEASUREMENT AND VALUATION OF HEALTH FOR NICE SUBMISSIONS

Indirect comparisons for single-arm trials or trials without common comparator arms

Cochrane Pregnancy and Childbirth Group Methodological Guidelines

Instrumental Variables Estimation: An Introduction

A NEW TRIAL DESIGN FULLY INTEGRATING BIOMARKER INFORMATION FOR THE EVALUATION OF TREATMENT-EFFECT MECHANISMS IN PERSONALISED MEDICINE

Technology appraisal guidance Published: 24 January 2018 nice.org.uk/guidance/ta500

Funnelling Used to describe a process of narrowing down of focus within a literature review. So, the writer begins with a broad discussion providing b

11 questions to help you make sense of a case control study

Discussion Leaders. Workshop Outline. David J. Vanness, PhD. Ágnes Benedict, MA, MS. Background

CRITICAL APPRAISAL SKILLS PROGRAMME Making sense of evidence about clinical effectiveness. 11 questions to help you make sense of case control study

Econometric analysis and counterfactual studies in the context of IA practices

Strategies for handling missing data in randomised trials

What is indirect comparison?

Epidemiologic Methods I & II Epidem 201AB Winter & Spring 2002

Comparing Treatments By Combining Data From Various Randomized And Observational Studies: Introduction To Concept, Methods, And Application

Secukinumab Versus Adalimumab for Psoriatic Arthritis: Comparative Effectiveness up to 48 Weeks Using a Matching-Adjusted Indirect Comparison

Erlotinib for the first-line treatment of EGFR-TK mutation positive non-small cell lung cancer

Guidelines for reviewers

Copyright GRADE ING THE QUALITY OF EVIDENCE AND STRENGTH OF RECOMMENDATIONS NANCY SANTESSO, RD, PHD

The ROBINS-I tool is reproduced from riskofbias.info with the permission of the authors. The tool should not be modified for use.

University of Bristol - Explore Bristol Research. Publisher's PDF, also known as Version of record

Implementing scientific evidence into clinical practice guidelines

ISPOR Task Force Report: ITC & NMA Study Questionnaire

Technology appraisal guidance Published: 26 July 2017 nice.org.uk/guidance/ta459

Selection and estimation in exploratory subgroup analyses a proposal

COMMITTEE FOR PROPRIETARY MEDICINAL PRODUCTS (CPMP) POINTS TO CONSIDER ON MISSING DATA

Introduction to research methods and critical appraisal WEEK 11

Adjusting overall survival for treatment switch

Cost-effectiveness of nivolumab with ipilimumab (Opdivo with Yervoy ) for the treatment of advanced (unresectable or metastatic) melanoma.

Using Statistical Principles to Implement FDA Guidance on Cardiovascular Risk Assessment for Diabetes Drugs

Dealing with uncertainty in the economic evaluation of health care technologies

NICE DSU TECHNICAL SUPPORT DOCUMENT 16: ADJUSTING SURVIVAL TIME ESTIMATES IN THE PRESENCE OF TREATMENT SWITCHING

1. Draft checklist for judging on quality of animal studies (Van der Worp et al., 2010)

Confounding by indication developments in matching, and instrumental variable methods. Richard Grieve London School of Hygiene and Tropical Medicine

RCTs: what s the problem?

ICH E9(R1) Technical Document. Estimands and Sensitivity Analysis in Clinical Trials STEP I TECHNICAL DOCUMENT TABLE OF CONTENTS

CHAMP: CHecklist for the Appraisal of Moderators and Predictors

Mostly Harmless Simulations? On the Internal Validity of Empirical Monte Carlo Studies

Sequence balance minimisation: minimising with unequal treatment allocations

Propensity scores: what, why and why not?

NICE DSU TECHNICAL SUPPORT DOCUMENT 9: THE IDENTIFICATION, REVIEW AND SYNTHESIS OF HEALTH STATE UTILITY VALUES FROM THE LITERATURE

Systematic Reviews. Simon Gates 8 March 2007

Design and Analysis Plan Quantitative Synthesis of Federally-Funded Teen Pregnancy Prevention Programs HHS Contract #HHSP I 5/2/2016

Citation for published version (APA): Ebbes, P. (2004). Latent instrumental variables: a new approach to solve for endogeneity s.n.

Quantitative Methods. Lonnie Berger. Research Training Policy Practice

Alan S. Gerber Donald P. Green Yale University. January 4, 2003

Supplementary Appendix

5/10/2010. ISPOR Good Research Practices for Comparative Effectiveness Research: Indirect Treatment Comparisons Task Force

Dr. Allen Back. Oct. 7, 2016

GRADE. Grading of Recommendations Assessment, Development and Evaluation. British Association of Dermatologists April 2014

Empirical assessment of univariate and bivariate meta-analyses for comparing the accuracy of diagnostic tests

Critical Thinking Rubric. 1. The student will demonstrate the ability to interpret information. Apprentice Level 5 6. graphics, questions, etc.

CRITICAL APPRAISAL WORKSHEET 1

MEA DISCUSSION PAPERS

Estimands and Sensitivity Analysis in Clinical Trials E9(R1)

Explaining heterogeneity

Module 14: Missing Data Concepts

How to use this appraisal tool: Three broad issues need to be considered when appraising a case control study:

Critical Appraisal of Scientific Literature. André Valdez, PhD Stanford Health Care Stanford University School of Medicine

Small-area estimation of mental illness prevalence for schools

Propensity score methods to adjust for confounding in assessing treatment effects: bias and precision

Cost-effectiveness of ixazomib (Ninlaro ) for the Treatment of Adult Patients with Multiple Myeloma who have Received at Least One Prior Therapy

SYSTEMATIC REVIEWS OF TEST ACCURACY STUDIES

Treatment effect estimates adjusted for small-study effects via a limit meta-analysis

Recent advances in non-experimental comparison group designs

Clinical trial design issues and options for the study of rare diseases

CONSORT: missing missing data guidelines, the effects on HTA monograph reporting Yvonne Sylvestre

Incorporating Clinical Information into the Label

Risk Benefit Assessment. Cristina E. Torres, Ph.D. UP-NIH Faculty and FERCAP Coordinator

Background Comparative effectiveness of nivolumab

Programme Name: Climate Schools: Alcohol and drug education courses

NICE DSU TECHNICAL SUPPORT DOCUMENT 19: PARTITIONED SURVIVAL ANALYSIS FOR DECISION MODELLING IN REPORT BY THE DECISION SUPPORT UNIT

Technology appraisal guidance Published: 16 May 2018 nice.org.uk/guidance/ta520

Examining Relationships Least-squares regression. Sections 2.3

Advanced IPD meta-analysis methods for observational studies

Downloaded from:

Ponatinib for treating chronic myeloid leukaemia and acute lymphoblastic leukaemia

Your submission states that the adjusted indirect comparison methodology was based largely on the publication by Glenny et al

Lec 02: Estimation & Hypothesis Testing in Animal Ecology

Choice of axis, tests for funnel plot asymmetry, and methods to adjust for publication bias

Executive Summary. Positron emission tomography (PET and PET/CT) in malignant lymphoma 1. IQWiG Reports Commission No. D06-01A

Technology appraisal guidance Published: 7 February 2018 nice.org.uk/guidance/ta505

Cost-effectiveness of almotriptan and rizatriptan in the treatment of acute migraine Williams P, Reeder C E

Transmission to CHMP July Adoption by CHMP for release for consultation 20 July Start of consultation 31 August 2017

The Effects of Autocorrelated Noise and Biased HRF in fmri Analysis Error Rates

Transcription:

Population-adjusted treatment comparisons Overview and recommendations from the NICE Decision Support Unit David M Phillippo, University of Bristol 3

Available from www.nicedsu.org.uk

Outline Background Standard indirect comparisons Population adjustment MAIC and STC Assumptions and properties Recommendations 5

Background: Indirect Comparisons Wish to compare two treatments B and C Not studied in the same trial Instead, each compared with a common comparator A through AB and AC trials. A B C 6

Background: Indirect Comparisons Standard indirect comparisons: d BC = d AC d AB Biased if there are imbalances in effect modifiers (EMs) between AB and AC; d AB AB d AB AC A d AB d AC B C 7

Background: Population Adjustment Standard indirect comparisons assume constancy of relative effects Population adjustment methods seek to adjust for imbalance in EMs Relaxed constancy assumption Create a fair comparison in a specific target population 8

Background Ideal scenario: full individual patient data (IPD) Gold standard IPD meta-regression AB trial: IPD Y i T i X 1i X 2i A AC trial: IPD Y i T i X 1i X 2i B C 9

Background Common scenario: limited IPD Several recent methods make use of mixed data AB trial: IPD Y i T i X 1i X 2i A AC trial: aggregate data തY A, ഥY C, തX 1, തX 2, σ A, σ C, f X B C 10

Population adjustment: MAIC and STC Matching-Adjusted Indirect Comparison Simulated Treatment Comparison Population reweighting method Outcome regression method Weight AB individuals to balance covariate distribution with AC trial Estimate outcomes on A and B in AC trial using weights Check distribution of weights, effective sample size Fit regression model in AB trial Estimate outcomes on A and B in AC trial using regression model Standard model checking, AIC/DIC, examine residuals AB and AC population must have sufficient overlap Compare covariate distributions, inclusion/exclusion criteria Not the only approaches, but at present the most popular 11

Population adjustment Two possible forms of indirect comparison Anchored Unanchored A B C B C 12

Population adjustment Two possible forms of indirect comparison Anchored Unanchored ˆ g Y g Y g Y ˆ g Y ˆ BC( AC) C( AC) A( AC) B( AC) A( AC) ˆ g Y g Y ˆ BC ( C) C( C) B( C) Comparison is on a given transformed scale The latter requires much stronger assumptions, but doesn t need a common comparator arm 13

Assumptions and properties: constancy Anchored A Unanchored B C B C Form of comparison Standard indirect comparison Anchored population-adjusted indirect comparison Unanchored populationadjusted indirect comparison Constancy assumption Constancy of relative effects Conditional constancy of relative effects Conditional constancy of absolute effects d AB AB = d AB AC Predict d AB AC from AB trial Predict Y B C from B trial Valid only if No effect modifiers in imbalance All effect modifiers known and adjusted for All effect modifiers and prognostic variables known and adjusted for Data Only requires aggregate data Requires IPD on at least one trial Requires IPD on at least one trial

Assumptions and properties MAIC and STC produce estimates of relative treatment effect that are specific to the AC population This is unlikely to be representative of the decision target population If so, population-adjusted estimates are irrelevant for the decision Can make use of the shared EM assumption, if justified Further research ongoing 15

Recommendations for use in HTA 1. Anchored vs. unanchored 2. Justifying anchored comparisons 3. Justifying unanchored comparisons 4. Variables to adjust for 5. Scale of comparison 6. Target population Reporting guidelines and example R code available online 16

Recommendation 1 When connected evidence with a common comparator is available, a populationadjusted anchored indirect comparison may be considered. Unanchored indirect comparisons may only be considered in the absence of a connected network of randomised evidence, or where there are single-arm studies involved. Anchored comparisons are always preferred to unanchored comparisons Unanchored comparisons require much stronger assumptions 17

Recommendation 2 Submissions using population-adjusted analyses in a connected network need to provide evidence that they are likely to produce less biased estimates of treatment differences than could be achieved through standard methods. Justification for moving away from standard methods required Altered decision scenario Consistency between appraisals See the NICE Methods Guide 18

NICE Methods Guide Treatment effect modifiers 5.2.7 Many factors can affect the overall estimate of relative treatment effects obtained from a systematic review. Some differences between studies occur by chance, others from differences in the characteristics of patients (such as age, sex, severity of disease, choice and measurement of outcomes), care setting, additional routine care and the year of the study. Such potential treatment effect modifiers should be identified before data analysis, either by a thorough review of the subject area or discussion with experts in the clinical discipline. NICE (2013) 19

Recommendation 2 (continued) a) Evidence must be presented that there are grounds for considering one or more variables as effect modifiers on the appropriate transformed scale. This can be empirical evidence, or an argument based on biological plausibility. b) Quantitative evidence must be presented that population adjustment would have a material impact on relative effect estimates due to the removal of substantial bias. Anchored comparisons should be justified with evidence for effect modification prior to analysis Judge possible magnitude of bias in relation to relative treatment effect, clinical importance 20

Recommendation 3 Submissions using population-adjusted analyses in an unconnected network need to provide evidence that absolute outcomes can be predicted with sufficient accuracy in relation to the relative treatment effects, and present an estimate of the likely range of residual systematic error in the adjusted unanchored comparison. For unanchored comparisons, need to justify that we are doing any better than a naïve comparison of arms Otherwise amount of bias is unknown, likely substantial, and could exceed size of treatment effect 21

Recommendation 4 a) For an anchored indirect comparison, propensity score weighting methods should adjust for all effect modifiers (in imbalance or not), but no prognostic variables. Outcome regression methods should adjust for all effect modifiers in imbalance, and any other prognostic variables and effect modifiers that improve model fit. For anchored comparisons, only adjustment for EMs is necessary to minimise bias Adjusting for other variables may unnecessarily reduce precision 22

Recommendation 4 b) For an unanchored indirect comparison, both propensity score weighting and outcome regression methods should adjust for all effect modifiers and prognostic variables, in order to reliably predict absolute outcomes. For unanchored comparisons all covariates must be adjusted for, as predictions of absolute outcomes are required 23

Recommendation 5 Indirect comparisons should be carried out on the transformed linear scale, with the same link functions that are usually employed for those outcomes. Effect modification defined with respect to this scale MAIC is not scale-free Consistency between appraisals 24

Recommendation 6 The target population for any treatment comparison must be explicitly stated, and population-adjusted estimates of the relative treatment effects must be generated for this target population. If there are effect modifiers, then the target population is crucial An unbiased comparison is not good enough for decision making, must also be in the correct population Can use the shared EM assumption, if justified 25

Key issues Performance and robustness of methods not known need thorough simulation study Decision target population must be defined, and estimates produced for this population Analysis from different perspective will give different results Evidence for effect modification is required for HTA Unanchored comparisons are very hard to justify 26

Thank you The TSD was commissioned and funded by the Decision Support Unit at the National Institute for Health and Care Excellence. Ongoing work is supported by the Medical Research Council grant number MR/P015298/1. Phillippo, D.M., Ades, A.E., Dias, S., Palmer, S., Abrams, K.R., Welton, N.J. NICE DSU Technical Support Document 18: Methods for population-adjusted indirect comparisons in submission to NICE. 2016. Available from http://www.nicedsu.org.uk Phillippo, D.M., Ades, A.E., Dias, S., Palmer, S., Abrams, K.R., Welton, N.J. Methods for populationadjusted indirect comparisons in health technology appraisal. 2017. Medical Decision Making. 27