Comparative Effectiveness Research Collaborative Initiative (CER-CI) PART 1: INTERPRETING OUTCOMES RESEARCH STUDIES FOR HEALTH CARE DECISION MAKERS

Similar documents
Comparative Effectiveness Research Collaborative Initiative (CER-CI) PART 1: INTERPRETING OUTCOMES RESEARCH STUDIES FOR HEALTH CARE DECISION MAKERS

ISPOR Task Force Report: ITC & NMA Study Questionnaire

Interpreting Prospective Studies

TOOLS YOU CAN USE April 8, 2014

5/10/2010. ISPOR Good Research Practices for Comparative Effectiveness Research: Indirect Treatment Comparisons Task Force

'Summary of findings' tables in network meta-analysis (NMA)

Alcohol interventions in secondary and further education

Agenda. Frequent Scenario WS I: Treatment Comparisons and Network Meta-Analysis: Learning the Basics

Systematic reviews and meta-analyses of observational studies (MOOSE): Checklist.

Standards for the reporting of new Cochrane Intervention Reviews

Quality and Reporting Characteristics of Network Meta-analyses: A Scoping Review

The Joanna Briggs Institute Reviewers Manual 2014

Combination therapy compared to monotherapy for moderate to severe Alzheimer's Disease. Summary

The QUOROM Statement: revised recommendations for improving the quality of reports of systematic reviews

Systematic Review & Course outline. Lecture (20%) Class discussion & tutorial (30%)

Workshop: Cochrane Rehabilitation 05th May Trusted evidence. Informed decisions. Better health.

CRITICAL APPRAISAL OF CLINICAL PRACTICE GUIDELINE (CPG)

The Cochrane Collaboration

SUPPLEMENTARY DATA. Supplementary Figure S1. Search terms*

What is a Special Interest Group (SIG)?

Evaluating the Quality of Evidence from a Network Meta- Analysis

Network Meta-Analysis for Comparative Effectiveness Research

Outcomes and GRADE Summary of Findings Tables: old and new

VALUE IN HEALTH 14 (2011) available at journal homepage:

Overview and Comparisons of Risk of Bias and Strength of Evidence Assessment Tools: Opportunities and Challenges of Application in Developing DRIs

Applications of Bayesian methods in health technology assessment

The Cochrane Collaboration, the US Cochrane Center, and The Cochrane Library

Learning from Systematic Review and Meta analysis

Seeing the Forest & Trees: novel analyses of complex interventions to guide health system decision making

Critical Appraisal Skills. Professor Dyfrig Hughes Health Economist AWMSG

DRAFT REPORT FOR REVIEW VERSION AUGUST 22, 2013

EPF s response to the European Commission s public consultation on the "Summary of Clinical Trial Results for Laypersons"

Issue Panel 18 HOW TO DEAL WITH MODEL UNCERTAINTY AT THE PLANNING STAGE OF COST-EFFECTIVENESS ANALYSIS?: TIPS FROM GLOBAL EXPERIENCES

Uses and misuses of the STROBE statement: bibliographic study

Meta-Analysis. Zifei Liu. Biological and Agricultural Engineering

Recent developments for combining evidence within evidence streams: bias-adjusted meta-analysis

Standards for reporting Plain Language Summaries (PLS) for Cochrane Diagnostic Test Accuracy Reviews (interim guidance adapted from Methodological

Systematic Reviews and Meta- Analysis in Kidney Transplantation

Washington, DC, November 9, 2009 Institute of Medicine

Statistical considerations in indirect comparisons and network meta-analysis

Achieving Operational Excellence in Prospective Observational Research

Critical appraisal: Systematic Review & Meta-analysis

Dementia Priority Setting Partnership. PROTOCOL March 2012

Disclosures. KVN serves as a consultant for the life sciences industry. MW is employed by GSK

Introduzione al metodo GRADE

Development of restricted mean survival time difference in network metaanalysis based on data from MACNPC update.

The role of MTM in Overviews of Reviews

Evidence-based Laboratory Medicine. Prof AE Zemlin Chemical Pathology Tygerberg Hospital

What is indirect comparison?

Statistical considerations in indirect comparisons and network meta-analysis

Live WebEx meeting agenda

Applied Pharmacoeconomic and Outcomes Research Forum. Spring 2007

Objectives. Information proliferation. Guidelines: Evidence or Expert opinion or???? 21/01/2017. Evidence-based clinical decisions

Controlled Trials. Spyros Kitsiou, PhD

U.S. Preventive Services Task Force Methods and Processes. Alex R. Kemper, MD, MPH, MS June 16, 2014

Grading the Evidence Developing the Typhoid Statement. Manitoba 10 th Annual Travel Conference April 26, 2012

GUIDELINE COMPARATORS & COMPARISONS:

The role of Randomized Controlled Trials

Title: Intention-to-treat and transparency of related practices in randomized, controlled trials of anti-infectives

Network Meta-Analysis of the Efficacy of Acupuncture, Alpha-blockers and Antibiotics on

Clinical Epidemiology for the uninitiated

Copyright GRADE ING THE QUALITY OF EVIDENCE AND STRENGTH OF RECOMMENDATIONS NANCY SANTESSO, RD, PHD

CEU screening programme: Overview of common errors & good practice in Cochrane intervention reviews

GRADE. Grading of Recommendations Assessment, Development and Evaluation. British Association of Dermatologists April 2018

Appendix A: Literature search strategy

Clinical research in AKI Timing of initiation of dialysis in AKI

Progress from the Patient-Centered Outcomes Research Institute (PCORI)

Introduction. Rare Disease Research, Health Technology Assessment and Evidence for Reimbursement FORUM

ACR OA Guideline Development Process Knee and Hip

Downloaded from:

Assessment of methodological quality and QUADAS-2

Workshop 21: INDIRECT TREATMENT COMPARISONS: AN INTERACTIVE WORKSHOP ON CHOOSING THE RIGHT TOOL FOR THE AVAILABLE DATA

The detection and management of pain in patients with dementia in acute care settings: development of a decision tool: Research protocol.

Background: Traditional rehabilitation after total joint replacement aims to improve the muscle strength of lower limbs,

Prospective Observational Study Assessment Questionnaire

The EU PIP - a step in Pediatric Drug Development. Thomas Severin Bonn,

Incorporating qualitative research into guideline development: the way forward

NICE Guidelines for HTA Issues of Controversy

Should Cochrane apply error-adjustment methods when conducting repeated meta-analyses?

Component of CPG development ILAE Recommendation Document Identifying topic and developing clinical. S3 research question

Models for potentially biased evidence in meta-analysis using empirically based priors

Surveillance report Published: 6 April 2016 nice.org.uk. NICE All rights reserved.

Healthcare outcomes assessed with observational study designs compared with those assessed in randomized trials(review)

ADDING VALUE AND EXPERTISE FOR SUCCESSFUL MARKET ACCESS. Per Sørensen, Lundbeck A/S

GRADE. Grading of Recommendations Assessment, Development and Evaluation. British Association of Dermatologists April 2014

IAASB Main Agenda (March 2005) Page Agenda Item. DRAFT EXPLANATORY MEMORANDUM ISA 701 and ISA 702 Exposure Drafts February 2005

Prostate Cancer Priority Setting Partnership. PROTOCOL June 2009

Statistical considerations in indirect comparisons and network meta-analysis

Designing Pharmacy Practice Research

EER Assurance Criteria & Assertions IAASB Main Agenda (June 2018)

Meta-analyses: analyses:

Research Synthesis and meta-analysis: themes. Graham A. Colditz, MD, DrPH Method Tuuli, MD, MPH

Introduction to systematic reviews/metaanalysis

Integrating the Patient Perspective Into Value Frameworks

Essential Skills for Evidence-based Practice Understanding and Using Systematic Reviews

This is the publisher s version. This version is defined in the NISO recommended practice RP

Nonpharmacologic Interventions for Treatment-Resistant Depression. Public Meeting December 9, 2011

2. Objectives a) Determine the proportion of drug therapy RCTs that include patients 80 years.

AHRQ Effective Health Care Program Clinical and Comparative Effectiveness Research Methods: II

Transcription:

Comparative Effectiveness Research Collaborative Initiative (CER-CI) PART 1: INTERPRETING OUTCOMES RESEARCH STUDIES FOR HEALTH CARE DECISION MAKERS HEALTH EVIDENCE FOR DECISION MAKING: ASSESSMENT TOOL FOR NETWORK META ANALYSIS STUDIES Jeroen P Jansen, PhD Interpreting Indirect Treatment Comparison Studies For Health Care Decision Makers Task Force Chair VP- Health Economics & Outcomes Research, MAPI Consultancy, Boston, MA, USA Forum: Monday November 5 th ISPOR 15 th EUROPEAN CONGRESS, BERLIN, GERMANY AMCP/ISPOR/NPC Agenda Background Objective Process Draft Assessment Tool Next steps Q&A Forum- November 5, 2012 1

AMCP, NPC, ISPOR Comparative Effectiveness Research Collaborative Initiative (CER-CI) The aim is enhancing usefulness of CER to improve patient health outcomes: Guidance and practical tools to help P&T members critically appraise CER studies to inform decision making Guidance to industry on what kinds of evidence payers want to see and how evidence will be considered in decision making Provide greater uniformity and transparency in the use and evaluation of CER for coverage and decision making Joint AMCP, ISPOR, NPC Meeting Oct 22 2010 Comparative Effectiveness Research Collaborative Initiative (CER-CI), PART 1 Create a set of Assessment Tools to evaluate the credibility, and relevance of non experimental studies Prospective observational studies Retrospective observational studies Modeling studies Network meta-analysis Develop specific questions pertinent to the type of study Promote consistency across the different tools Forum- November 5, 2012 2

Why are these tool potentially useful to decision makers? Provides a critical appraisal tool to help identify credible evidence to help inform decisionmaking Promote uniformity when appraising the quality of evidence Can be used as an educational tool to increase sophistication in the critical review of CE information Network Meta-Analysis: Background Clinicians, patients, and health-policy makers often need to decide which treatment is best. Growing interest in measures/methodologies to help ensure healthcare decision making is better informed by results of relevant evidence. Available treatments tend to increase over time. Unfortunately, robustly designed RCTs that simultaneously compare all interventions of interest are almost never available. Network meta-analysis is an extension of standard pairwise metaanalysis by including multiple pairwise comparisons across a range of interventions. Network meta-analysis provides indirect comparisons for interventions not studied in a head-to-head fashion. Forum- November 5, 2012 3

Evidence Networks Source: Jansen JP, Fleurence R, Devine B, et al. Interpreting indirect treatment comparisons & network meta-analysis for health care decision-making: Report of the ISPOR Task Force on indirect treatment comparisons good research practices Part 1. Value Health 2011;14: 417-28 Network Meta-Analysis: Background There is an increase in the number of indirect treatment comparisons and network meta-analysis published. Given the relevance of network meta-analysis to inform healthcare and coverage decision-making, it is of great interest to improve understanding of these studies by decision-makers and increase their use. Forum- November 5, 2012 4

Objective To develop a measurement tool for the assessment of credibility and relevance of a network meta-analysis. Development Stages INITIATION OF THE TASK FORCE Board Approval Selection of members Based on ISPOR Good Research Practices CONTENT DEVELOPMENT Develop End User Tool Draft Assessment Tool FINAL PRODUCT Testing Modification Members review ROLL OUT OF THE PRODUCT Manuscript to Value in Health Web-based tool Educational Materials Forum- November 5, 2012 5

Interpreting Indirect Treatment Comparison Studies For Health Care Decision Makers Task Force Chair: Jeroen Jansen, PhD VP- Health Economics & Outcomes Research, MAPI Consultancy, Boston, MA, USA Members: AMCP NPC ISPOR Sherry Andes, BSPharm, PharmD, BCPS, BCPP, PAHM Clinical Pharmacist - Drug Information Services, informedrx, an SXC Company Louisville, Kentucky, USA Vijay Joish BS Pharm, MS, PhD Deputy Director HEOR, Bayer Pharmaceuticals, Wayne, New Jersey, USA Thomas Trikalinos MD, PhD Research Associate in Health Services Policy & Practice, Public Health-Health Services Policy & Practice, Brown University, Providence, Rhode Island, USA Jessica Daw, PharmD, MBA Director, Clinical Pharmacy UPMC Health Plan Pittsburgh, PA, USA Joseph C. Cappelleri PhD, MS, MPH Senior Director Pfizer Inc New London, Connecticut, USA Georgia Salanti PhD Assistant Professor, Department of Hygiene and Epidemiology University of Ioannina, School of Medicine University Campus, Ioannina, Greece Steps for Development of Tool 1. Outlining requirements ( Desiderata ) Relevant for decision-making Total evidence base Appropriate statistical methods Transparency Understandable and easy to use 2. Draft items/questions along with glossary 3. Self and user testing 4. Modification (if needed) Forum- November 5, 2012 6

Key Decisions Level of expertise needed Granularity Organization Scoring Level of Expertise Required of Assessor Basic understanding of study design and elementary epidemiological concepts is required. Some jargon in tool To gain acceptance for a wide range of users, including experts To allow assessors to become familiar with the language of ITC/NMA Expected that after some practice, non-experts will understand the key terms used in the instrument Glossary and supporting document explaining the concepts in a non-technical manner Forum- November 5, 2012 7

Granularity The tool needs to be sufficiently specific to help decision-makers judge validity and relevance. More specific questions will help non-experts in their assessment. Too many questions will compromise the practicality of the tool. Organization Divided into domains Questions organized by journal article format Sub-questions that can be utilized depending on reviewer expertise Web-based with hyperlinked explanation of terms Forum- November 5, 2012 8

Two Main Domains Relevance Relevance addresses the extent to which the results of the study, if accurate, apply to the setting of interest to the decision-maker. Addresses issues of population, interventions, endpoints and policy-relevant differences. Credibility Internal validity The extent to which the study appropriately answers the question it is designed or intended to answer. The extent to which the results are affected by bias due to design of the study and conduct of analysis. Addresses issue of reporting comprehensiveness Instrument 6 Categories, 26 Questions Relevance Credibility Evidence base Analysis Reporting Quality & Transparency Interpretation Conflict of interest Forum- November 5, 2012 9

Scoring Credibility and Relevance scored separately Does the study have sufficient credibility or relevance based on the answers to the questions in each domain? Questions are grouped by domain and each question is worded as a yes/no. After answering the items in a domain, it is then rated by the user as Strength Weakness Neutral When some questions are answered in a particular way, they are considered fatal flaws Forum- November 5, 2012 10

Relevance 1. Is the population relevant? 2. Are any critical interventions missing? 3. Are any relevant outcomes missing? 4. Is the context (e.g. settings and circumstances) applicable to your population? Forum- November 5, 2012 11

Credibility: Evidence base 5. Was the body of randomized controlled trials concerning relevant interventions identified? Was the literature search strategy defined in such a way that available randomized controlled trials concerning the relevant interventions could be identified? Were multiple databases searched ( e.g. MEDLINE, EMBASE, Cochrane)? Were study selection criteria defined in such a way that relevant randomized controlled trials were included if identified with the literature search? 6. Are the treatments that are being compared with the network meta-analysis all part of one connected network of only randomized controlled trials? Decision Set vs. All Evidence Set Forum- November 5, 2012 12

Credibility: Evidence base (cont d) 7. Is it likely that bias was induced by including poor quality studies? 8. Is it likely that bias was induced by selective reporting of outcomes in the studies? 9. Are there systematic differences in treatment effect modifiers across the different treatment comparisons in the network? 10. If yes, were imbalances in effect modifiers across the different treatment comparisons identified prior to reviewing individual study results? Imbalance in Treatment Effect Modifiers Cause Bias in Indirect Comparisons Source: Jansen JP, Schmid CH, Salanti G. Direct Acyclic graphs can help understand bias in indirect and mixed treatment comparisons. J Clin Epidemiology (in press) 26 Forum- November 5, 2012 13

Credibility: Analysis 11. Were statistical methods used that synthesize results of trials without breaking randomization? (No naïve comparisons) 12. If direct and indirect comparisons are available for pairwise contrasts (i.e. closed loops), was agreement (i.e. consistency) evaluated or discus 13. In the presence of consistency between direct and indirect evidence for pairwise contrasts, were both direct and indirect evidence included in the network meta-analysis? 14. Were any systematic differences in potential treatment effectmodifiers across comparisons and (unexplained) inconsistency taken into account in the network meta-analysis? Imbalance in Treatment Effect Modifiers Cause Inconsistency in Mixed Treatment Comparisons Forum- November 5, 2012 14

Adjustment for Imbalance in Treatment Effect Modifiers in Indirect Comparisons Source: Jansen JP. Network meta-analysis of individual and aggregate level data. (under review) 29 Credibility: Analysis (cont d) 15. Was a valid rationale provided for the use of random effects or fixed effects models? 16. If a random effects model was used, were assumptions about heterogeneity explored or discussed? 17. If there are indications of heterogeneity, were subgroup analyses or meta-regression analysis with pre-specified covariates performed? Forum- November 5, 2012 15

Credibility: Reporting Quality & Transparency 18. Is a graphical or tabular representation of the evidence network provided with information on the number of RCTs per direct comparison? 19. Are the individual study results reported 20. Are direct results reported separately? Forum- November 5, 2012 16

Credibility: Reporting Quality & Transparency (cont d) 21. Are all pairwise contrasts between interventions as obtained with network meta-analysis reported along with measures of uncertainty? 22. Is a ranking of interventions provided given the reported treatment effects and its uncertainty by outcome? 23. Is the impact of important patient characteristics on treatment effects reported? Probabilistic Interpretation of the Uncertainty Source: Salanti G, Ades AE, Ioannidis JP. Graphical methods and numerical summaries for presenting results from multipletreatment meta-analysis: an overview and tutorial. 34 J Clin Epidemiol. 2011;64:163-71. Forum- November 5, 2012 17

Credibility: Interpretation 24. Are the conclusions fair and balanced? Credibility: Conflict of interest 25. Were there any potential conflicts of interest? 26. If yes, were steps taken to minimize these? Timelines Tool first draft developed June July 2012 Develop Questionnaires Based on ISPOR GRPs September 2011 June 2012 Self testing & modification Aug Sep 2012 User testing & modification Sep - Oct 2012 Members review November 2012 Final product December 2012 AMCP Meeting April 2013 Web- based tools Educational materials Forum- November 5, 2012 18

Thank you http://www.ispor.org/taskforces/interpretingindirecttr eatmentcomparisonstudiestf.asp Questions and comments welcome! Forum- November 5, 2012 19