Evidence- and Value-based Solutions for Health Care Clinical Improvement Consults, Content Development, Training & Seminars, Tools

Similar documents
Delfini Evidence Tool Kit

Controlled Trials. Spyros Kitsiou, PhD

Evidence Informed Practice Online Learning Module Glossary

Learning objectives. Examining the reliability of published research findings

Essential Skills for Evidence-based Practice Understanding and Using Systematic Reviews

GRADE. Grading of Recommendations Assessment, Development and Evaluation. British Association of Dermatologists April 2014

Empirical evidence on sources of bias in randomised controlled trials: methods of and results from the BRANDO study

Module 5. The Epidemiological Basis of Randomised Controlled Trials. Landon Myer School of Public Health & Family Medicine, University of Cape Town

Structural Approach to Bias in Meta-analyses

Evidence Based Practice

CRITICAL APPRAISAL OF MEDICAL LITERATURE. Samuel Iff ISPM Bern

ACR OA Guideline Development Process Knee and Hip

Rapid appraisal of the literature: Identifying study biases

Study Design STUDY DESIGN CASE SERIES AND CROSS-SECTIONAL STUDY DESIGN

GRADE. Grading of Recommendations Assessment, Development and Evaluation. British Association of Dermatologists April 2018

04/12/2014. Research Methods in Psychology. Chapter 6: Independent Groups Designs. What is your ideas? Testing

ISPOR Task Force Report: ITC & NMA Study Questionnaire

Cochrane Pregnancy and Childbirth Group Methodological Guidelines

Mixed Methods Study Design

Title: Reporting and Methodologic Quality of Cochrane Neonatal Review Group Systematic Reviews

RATING OF A RESEARCH PAPER. By: Neti Juniarti, S.Kp., M.Kes., MNurs

School of Dentistry. What is a systematic review?

WHY AN EVIDENCE-BASED FORMULARY SYSTEM?

In this second module in the clinical trials series, we will focus on design considerations for Phase III clinical trials. Phase III clinical trials

Experimental and Quasi-Experimental designs

State of the art pharmacoepidemiological study designs for post-approval risk assessment

Delfini Quality Improvement Project Tool Kit

User s guide to the checklist of items assessing the quality of randomized controlled trials of nonpharmacological treatment

Designing Pharmacy Practice Research

Recent developments for combining evidence within evidence streams: bias-adjusted meta-analysis

Clinical research in AKI Timing of initiation of dialysis in AKI

Appraising the Literature Overview of Study Designs

Randomized Controlled Trial

10/21/2014. Considerations in the Selection of Research Participants. Reasons to think about participant selection

Sampling Controlled experiments Summary. Study design. Patrick Breheny. January 22. Patrick Breheny Introduction to Biostatistics (BIOS 4120) 1/34

Definition. Synonyms 9/23/2010. Trials for which the hypothesis and study design are formulated based on information needed to make a decision

AOTA S EVIDENCE EXCHANGE CRITICALLY APPRAISED PAPER (CAP) GUIDELINES Annual AOTA Conference Poster Submissions Critically Appraised Papers (CAPs) are

Glossary of Practical Epidemiology Concepts

Primary Research Question and Definition of Endpoints. Mario Chen The Fundamentals of Biostatistics in Clinical Research Workshop India, March 2007

Systematic Reviews. Simon Gates 8 March 2007

Role of evidence from observational studies in the process of health care decision making

Quantitative research Methods. Tiny Jaarsma

Clinical Evidence: Asking the Question and Understanding the Answer. Keeping Up to Date. Keeping Up to Date

Guidelines for Writing and Reviewing an Informed Consent Manuscript From the Editors of Clinical Research in Practice: The Journal of Team Hippocrates

Overview and Comparisons of Risk of Bias and Strength of Evidence Assessment Tools: Opportunities and Challenges of Application in Developing DRIs

Updates in Therapeutics 2015: The Pharmacotherapy Preparatory Review & Recertification Course

The role of Randomized Controlled Trials

95% 2.5% 2.5% +2SD 95% of data will 95% be within of data will 1.96 be within standard deviations 1.96 of sample mean

Study Design. Study design. Patrick Breheny. January 23. Patrick Breheny Introduction to Biostatistics (171:161) 1/34

Title: Reporting and Methodologic Quality of Cochrane Neonatal Review Group Systematic Reviews

11 questions to help you make sense of a case control study

1. Draft checklist for judging on quality of animal studies (Van der Worp et al., 2010)

Drug Class Literature Scan: Pancreatic Enzymes

The Healthy User Effect: Ubiquitous and Uncontrollable S. R. Majumdar, MD MPH FRCPC FACP

Types of Biomedical Research

Interpretation of Epidemiologic Studies

Introduction to meta-analysis

EBM: Therapy. Thunyarat Anothaisintawee, M.D., Ph.D. Department of Family Medicine, Ramathibodi Hospital, Mahidol University

The RoB 2.0 tool (individually randomized, cross-over trials)

Alcohol interventions in secondary and further education

Protocol Development: The Guiding Light of Any Clinical Study

Online Supplementary Material

Systematic reviews and meta-analyses of observational studies (MOOSE): Checklist.

Overview of Study Designs in Clinical Research

Statistical considerations in indirect comparisons and network meta-analysis

Determinants of quality: Factors that lower or increase the quality of evidence

Embedding pragmatic trials within databases of electronic health records / disease registries Tjeerd van Staa

Results. NeuRA Treatments for internalised stigma December 2017

The who, what, why, where, and when of Clinical Practice Guidelines (CPGs)

Guidelines for Reporting Non-Randomised Studies

Real World Evidence in the Treatment of Ovarian Cancer. Elizabeth Eisenhauer MD FRCPC Queen s University

CRITICAL APPRAISAL OF CLINICAL PRACTICE GUIDELINE (CPG)

9: 3 TABLE OF CONTENTS P&T

exposure/intervention

Perspectives on Large Simple Trials

Should Cochrane apply error-adjustment methods when conducting repeated meta-analyses?

Instrument for the assessment of systematic reviews and meta-analysis

Observational Study Designs. Review. Today. Measures of disease occurrence. Cohort Studies

Meta-analyses: analyses:

Other potential bias. Isabelle Boutron French Cochrane Centre Bias Method Group University Paris Descartes

BIOSTATISTICAL METHODS

Blood Pressure and Complications in Individuals with Type 2 Diabetes and No Previous Cardiovascular Disease. ID BMJ

Purpose. Study Designs. Objectives. Observational Studies. Analytic Studies

Epidemiologic Methods and Counting Infections: The Basics of Surveillance

To evaluate a single epidemiological article we need to know and discuss the methods used in the underlying study.

Reporting guidelines

the standard deviation (SD) is a measure of how much dispersion exists from the mean SD = square root (variance)

Quality of Clinical Practice Guidelines

Applicable or non-applicable: investigations of clinical heterogeneity in systematic reviews

12/26/2013. Types of Biomedical Research. Clinical Research. 7Steps to do research INTRODUCTION & MEASUREMENT IN CLINICAL RESEARCH S T A T I S T I C

Pros and Cons of Clinical Trials vs. Observational Studies. Beth Devine PCORP Summer Institute July 14, 2015

GLOSSARY OF GENERAL TERMS

A Systematic Review of the Efficacy and Clinical Effectiveness of Group Analysis and Analytic/Dynamic Group Psychotherapy

Checking the counterarguments confirms that publication bias contaminated studies relating social class and unethical behavior

Bias. Zuber D. Mulla

Standard Methods for Quality Assessment of Evidence

Clinical problems and choice of study designs

Section Editor Steven T DeKosky, MD, FAAN Kenneth E Schmader, MD

CAN EFFECTIVENESS BE MEASURED OUTSIDE A CLINICAL TRIAL?

2 Critical thinking guidelines

Transcription:

Definition Key Points Key Problems Bias Choice Lack of Control Chance Observational Study Defined Epidemiological study in which observations are made, but investigators do not control the exposure or intervention and other factors. Changes or differences in one characteristic are studied in relation to changes or differences in others, without the intervention of the investigator. Observational studies are highly prone to selection and observation bias and confounding. Tip: If an intervention is "assigned" through the research, it is an experiment. If it is chosen, then the study type is observational. While observational studies have been used by many for drawing cause and effect conclusions for interventions, frequently the conclusions have been wrong, such as with hormone replacement studies as treatment for preventing subsequent cardiac events in women with previous myocardial infarction. This is due to special opportunities for bias, confounding and chance which are reduced or eliminated by an experimental design, especially if randomization is used. If evidence regarding efficacy from valid, relevant randomized controlled trials is not available, evidence-based decisions are rarely possible. Exceptions may occur in situations where specific criteria have been met (see last entry below), giving preference to study designs or information in the following order: 1) experiments, 2) cohort designs, 3) case/control studies, 4) database studies, and 5) case series. Bottom Line: The positive predictive value for true results with a high quality RCT is approximately 85%; the positive predictive value for a high quality observational study is approximately 20%. Ioannidis JPA. Why Most Published Research Findings are False. PLoS Med 2005; 2(8):696-701). Baseline differences in groups, choice and lack of control create special challenges in observational studies and are likely to affect the observed outcomes. These problems are amplified below (Epi-speak included.) There are numerous instances where observational studies have been in agreement with RCTs or have provided helpful insights and solutions to health care problems. However, there are also numerous instances in which observational studies have not been in agreement with RCTs and have misled end users. For example, based on numerous observational studies reporting benefits, clinicians and patients accepted hormone replacement therapy (HRT) as an effective therapy for prevention of second cardiac events in women who had experienced a myocardial infarction. However RCTs demonstrated that subsequent cardiac events were not prevented by HRT; in fact RCTs revealed significant adverse events associated with HRT. Similarly, observational studies reported decreased mortality with vitamin A, beta-carotene and vitamin E, but high quality RCTs demonstrated increased mortality with these antioxidants. In a review of 18 meta-analyses (1211 clinical trials), outcomes of non-randomized trials varied from outcomes of RCTs (8 studies) from 76% underestimation to 160% overestimation. (Kunz R, Oxman A. BMJ 1998. Vol317:1185-90.) Differing Baseline Characteristics, Choices and Lack of Control You need two equal groups treated in the same way (except for the intervention being studied and its comparator) to understand if the intervention and not some baseline difference or difference in experience or treatment between the groups caused the difference in outcomes. In observational studies, investigators attempt to minimize baseline differences between groups through statistical adjustments for the differences in baseline variables such as sex, age, disease severity, co-morbidities, etc.; however, confounders (both known and unknown) are not adequately managed through this strategy. The actual factor responsible for the outcomes would have to be adjusted, which would require prior knowledge of the answer to the question that the research is designed to answer. Differing baseline characteristics (selection bias) remain a major threat to validity in all observational studies because adjustment models cannot reliably equalize differences between www.delfini.org 2002-2013 Delfini Group, LLC. All Rights Reserved Worldwide. Page 1 of 5

groups. In addition to baseline differences between groups, observational studies are flawed by the many different exposures and choices likely to be made by subjects (or their health care clinicians). Differences in experiences (e.g., co-interventions) are likely to represent important confounders which, in some cases, represent the true reason for the differences in outcomes. These differences in exposures and choices could be due to Health status of patients Provider skills Provision of care (e.g., affordability which can link to socioeconomic factors or patient demand, etc. which could drive provision of care and link to other confounding factors) Patient perceptions (e.g., risk aversion) and personal characteristics (e.g., more well-read, healthyuser effect) Other unknown confounders (e.g., genetic issues, environmental exposures, risk factors) It should be emphasized that in observational studies baseline differences, interventions, cointerventions and other exposures cannot be adequately controlled. This lack of control is likely to result in selection bias, performance bias, differences in follow-up (attrition bias) and differences in how outcomes are evaluated (assessment bias). Examples of How Baseline Differences in Patients and Choice May Affect Outcomes Physician directs patients who are more likely to have favorable outcomes (i.e., those with better prognostic characteristics) to one treatment over another Physician favors one treatment over another due to training or skill level A more demanding patient may receive more monitoring or additional care A patient who is more educated, intelligent, information-seeking or affluent might engage in other behaviors that contribute to the outcomes Risk aversion may be associated with other personal factors which may be confounders Examples of Methods Used in RCTs to Control or Minimize The Above-Mentioned Confounders Found in Observational Studies Factors that may be controlled in experiments, but not in the natural world of care or observational studies (where patients choose their interventions) include Achieving comparability of baseline risks through randomization Blinding Controlling uniformity in procedures, follow-up length and treatment methods such as dosing or specifications on which and how procedures are done Controlling allowed and disallowed medications and wash-out Adherence measurements and concomitant medication monitoring Prevention of contamination or migration to other arm of the study Quality control for standardization such as in training, measurement and reporting Just-in-time monitoring for pre-specified definitions of treatment success or failure or other endpoints Examples of How Lack of Control in Observational Studies May Affect Outcomes Adherence is more likely to be poorer outside of a controlled trial Patients may be taking other medications (there are no disallowed medications in observational studies and you really can t be sure what a patient is doing) Other procedures may be used which could be the actual explanation for the observed results Monitoring, measurement and reporting of outcomes are likely to be driven by the choice of www.delfini.org 2002-2013 Delfini Group, LLC. All Rights Reserved Worldwide. Page 2 of 5

treatment and are likely to vary between individuals People tend to root for the intervention you will notice what you are primed to notice and there is no blinding in observational studies The Play of Chance in Observational Studies: When multiple exposures are examined in observational studies (e.g., in case-control studies evaluating associations between exposures (independent variables) such as diet, weight, socioeconomic status, exercise, etc. and cancer (dependent variable) there is an increased probability of finding a statistically significant differences in outcomes between groups in some variables because of chance. You Can t Blind in an Observational Study Blinding matters. This has been established in RCTs by comparing results of blinded RCTs with results of non-blinded RCTs. Lack of blinding in RCTs has been reported to exaggerate benefit by up to a relative 48% (Schulz, PMID: 7823387, Kjaergard PMID: 11730399, Moher PMID: 9746022) In observational studies patients choose their interventions, care-givers are not blinded and outcomes assessors are frequently not blinded. Theorized as Possibly Reliable Indicators of Efficacy in Observational Studies (Grade U [Uncertain] Evidence) Even if the Observational Study Leads us in the Right Direction With rare exceptions (see below) the only way to know that results from an observational study are likely to be true is to know that a valid RCT has reported similar results (confirmed the findings of an observational study). Furthermore, even if the results of an observational study suggest benefit from an intervention, observational studies are more prone to bias and are thus likely to falsely inflate the results favoring the intervention. We have mentioned that there are rare exceptions to the points made above. There is general agreement that causality MAY be present when there is dramatic change following application of an intervention or technology that is unlikely to be due to confounding (i.e., close to all-or-none results: example, before treatment all died and following treatment, high survival rate). Examples of all-or-none results are antibiotics for meningitis and electrical cardioversion for ventricular fibrillation Public health interventions represent another area where valid RCTs may not be available, yet efficacy of interventions has frequently been accepted even when outcomes are less dramatic than all-or-none results. For example, purification of water, the use of smoke alarms and seatbelts, many immunizations and some screening initiatives have been accepted by most as providing beneficial effects on health outcomes. In some instances studies of public health initiatives may allow cause and effect conclusions. Confounding while present, may be more limited in some public health interventions than in many observational studies because of the extremely broad application of the public health interventions. Below we list what we believe to be other key considerations for concluding that the effects reported in an observational study are due to the interventions, (i.e., that causality may exist). Primary Considerations for Determining Efficacy in Observational Studies: 1. There is extremely low likelihood of improvement without some intervention; and, 2. Any effects are highly likely to be attributable to the intervention (e.g., a single intervention or technology was utilized and the likelihood of patients utilizing co-interventions is low or co-interventions were equivalent in compared groups [note: equivalence should take into account considerations such as administration, duration, dosing, etc.]); and, 3. Convincing sustained improvement is documented following use of the intervention or technology. www.delfini.org 2002-2013 Delfini Group, LLC. All Rights Reserved Worldwide. Page 3 of 5

OR 1. Illness or condition has a predictable course or outcome (e.g., symptoms, diagnostic signs, progression or resolution of the condition are predictable); and, 2. Any effects in this population are likely to be attributable to the intervention (e.g., plausible confounders have been considered, a single intervention or technology was utilized and the likelihood of patients utilizing co-interventions is low or co-interventions were equivalent in groups being compared [note: equivalence should take into account considerations such as administration, duration, dosing, etc.]); and, 3. Convincing sustained improvement in this particular population is documented following use of the intervention or technology with results demonstrably and significantly earlier than what is predicted based on reliable knowledge of natural history. Uses of Observational Studies Secondary Considerations Lending Support for Concluding Efficacy: Factoring in the following conditions may decrease uncertainty: a. Immediacy of effect which may increase confidence in the effectiveness of the intervention or technology not only because of temporal proximity, but also because a delay in effect provides more opportunity for application of a co-intervention which could be a confounder. When considering immediacy, caution is urged because the effect could be due to potential of confounding. b. Dramatic effects (e.g., approximately 80% or greater relative risk reduction). When considering dramatic effects, caution is urged because the effects could be due to the strength of a confounder. For example, The Cochrane Handbook states that observational evidence with large effect sizes where a dose-response gradient exists and all plausible confounding would most likely demonstrate a reduced effect size, may be of moderate or even high quality. Higgins JPT, Green S (2008) Cochrane Handbook for Systematic Reviews of Interventions Version 5.0.0 [updated February 2008]. The Cochrane Collaboration. Available from www.cochrane-handbook.org accessed 08/07/2008. Section 12.2.3]. However, we remain cautious about this conclusion as confounding may result in large differences between groups. Example: in one study of adolescents with migraine headache the placebo effect was ~ 60% (Winner P, Linder SL, Lipton RB, Almas M, Parsons B et al. (2007) Eletriptan for the acute treatment of migraine in adolescents: results of a double-blind, placebo-controlled trial. Headache 47:511-8). c. Dose/response relationship. When considering dose/response relationship, caution is urged because patient expectation bias may result in increased perception of benefit or increased placebo effect if patient is aware of the greater dosing through increase in side effects, for example. d. Consistent changes in multiple outcomes. When considering consistent change, caution is urged because the change in multiple outcomes may be due to confounding (e.g., a causal association between an unknown confounder and each of the outcomes). Observational Study Types Cohort (best overall observational study type) Cross-sectional (useful in diagnostic testing, though RCTs used also) Case-Control (weakest design) Observations are best used for Generating hypotheses and ideas for study topics Understanding the natural history of diseases and prognosis Studies of diagnostic testing Studies of quality of life Satisfaction (caution as this frequently confounded) Understanding the effects of clinical improvement activities via process measurement and success of implementation, etc. www.delfini.org 2002-2013 Delfini Group, LLC. All Rights Reserved Worldwide. Page 4 of 5

Some other uses include the following (special thanks to Jeff White, PharmD, WellPoint Rx) Provide supportive information beyond efficacy of the generalizability of an RCT as it is risky to assume that efficacy evidence in a RCT will result in similar benefits to patients in a "real-world" environment. For example, what percent of members appear to benefit from an intervention given the circumstances of the real-world environment, recognizing that there may explanations other than causality for the observed outcomes (e.g., prognostic variables, biases and chance). Provide insight into real-world practice patterns, given the impact of non-compliance, switching medications, comorbidities, marketing, patient preferences, etc. Understand the degree of off-label use by linking individual drug to an ICD-9 diagnosis code, for example. An assessment of the total cost of care (pharmacy and medical) for members on various therapies recognizing that the cost estimate may not apply to other locations or populations and that the cost estimate may have been distorted by prognostic variables, bias and chance present in the data. Safety assessment recognizing that bias, confounding and chance may explain or distort the findings. www.delfini.org 2002-2013 Delfini Group, LLC. All Rights Reserved Worldwide. Page 5 of 5