exposure/intervention

Similar documents
Hazards and Benefits of Postnatal Steroids. David J. Burchfield, MD Professor and Chief, Neonatology University of Florida

How to Interpret a Clinical Trial Result

Module 5. The Epidemiological Basis of Randomised Controlled Trials. Landon Myer School of Public Health & Family Medicine, University of Cape Town

Experimental Design. Terminology. Chusak Okascharoen, MD, PhD September 19 th, Experimental study Clinical trial Randomized controlled trial

EBM: Therapy. Thunyarat Anothaisintawee, M.D., Ph.D. Department of Family Medicine, Ramathibodi Hospital, Mahidol University

Evidence Informed Practice Online Learning Module Glossary

Screening for Disease

RANDOMIZED CLINICAL TRIALS. Study Designs. Randomized Clinical Trials (RCT) ASSIGN EXPOSURE Follow up Check for OUTCOME. Experimental studies

The Cardiac Arrythmia Suppression Trial

Making comparisons. Previous sessions looked at how to describe a single group of subjects However, we are often interested in comparing two groups

Assessing risk of bias

GS3. Understanding How to Use Statistics to Evaluate an Article. Session Summary. Session Objectives. References. Session Outline

No relevant financial relationships

Rapid appraisal of the literature: Identifying study biases

How to CRITICALLY APPRAISE

Common Statistical Issues in Biomedical Research

Placebo-Controlled Statin Trials Prevention Of CVD in Women"

Placebo-Controlled Statin Trials MANAGEMENT OF HIGH BLOOD CHOLESTEROL MANAGEMENT OF HIGH BLOOD CHOLESTEROL: IMPLICATIONS OF THE NEW GUIDELINES

Primary Research Question and Definition of Endpoints. Mario Chen The Fundamentals of Biostatistics in Clinical Research Workshop India, March 2007

MAT Mathematics in Today's World

Critical Appraisal Practicum. Fabio Di Bello Medical Implementation Manager

Overview of Study Designs

Critical Appraisal Series

Evaluating Scientific Journal Articles. Tufts CTSI s Mission & Purpose. Tufts Clinical and Translational Science Institute

Is There An Association?

Critical Appraisal Istanbul 2011

Outline. What is Evidence-Based Practice? EVIDENCE-BASED PRACTICE. What EBP is Not:

CLINICAL DECISION USING AN ARTICLE ABOUT TREATMENT JOSEFINA S. ISIDRO LAPENA MD, MFM, FPAFP PROFESSOR, UPCM

Issues to Consider in the Design of Randomized Controlled Trials

New Hypertension Guidelines: Why the change? Neil Brummond, M.D. Avera Medical Group Internal Medicine Sioux Falls, SD

The comparison or control group may be allocated a placebo intervention, an alternative real intervention or no intervention at all.

CHL 5225 H Advanced Statistical Methods for Clinical Trials. CHL 5225 H The Language of Clinical Trials

95% 2.5% 2.5% +2SD 95% of data will 95% be within of data will 1.96 be within standard deviations 1.96 of sample mean

RATING OF A RESEARCH PAPER. By: Neti Juniarti, S.Kp., M.Kes., MNurs

Anesthetic-Induced Neuronal Brain Injury in Infants: Finding an Answer with Clinical Studies

GATE CAT Intervention RCT/Cohort Studies

Sampling Controlled experiments Summary. Study design. Patrick Breheny. January 22. Patrick Breheny Introduction to Biostatistics (BIOS 4120) 1/34

Study Designs. Randomized Clinical Trials (RCT) RCT: Example 1. RCT: Two by Two Table. Outcome. Exposure. Yes a b No c d

AVOIDING BIAS AND RANDOM ERROR IN DATA ANALYSIS

EFFECTIVE MEDICAL WRITING Michelle Biros, MS, MD Editor-in -Chief Academic Emergency Medicine

Journal Club September 29, Vanessa AKIKI PGYlII Internal Medicine

Controlled Trials. Spyros Kitsiou, PhD

Disclosure. No relevant financial relationships. Placebo-Controlled Statin Trials

Teaching critical appraisal of randomised controlled trials

Placebo-Controlled Statin Trials

TOO COOL OR NOT TOO COOL- THERAPEUTIC HYPOTHERMIA IN THE ICU SCCM TX 2017 TED WU MD PEDIATRIC CRITICAL CARE UNIVERSITY OF TEXAS HEALTH SAN ANTONIO

Critical Appraisal. Dave Abbott Senior Medicines Information Pharmacist

Study Designs in Epidemiology

Guidelines for Writing and Reviewing an Informed Consent Manuscript From the Editors of Clinical Research in Practice: The Journal of Team Hippocrates

Optimizing Communication of Emergency Response Adaptive Randomization Clinical Trials to Potential Participants

Purpose. Study Designs. Objectives. Observational Studies. Analytic Studies

Bias. A systematic error (caused by the investigator or the subjects) that causes an incorrect (overor under-) estimate of an association.

Evaluating and Interpreting Clinical Trials

Study Design. Study design. Patrick Breheny. January 23. Patrick Breheny Introduction to Biostatistics (171:161) 1/34

Definitions Disease Illness The relationship between disease and illness Disease usually causes Illness Disease without Illness

Learning objectives. Examining the reliability of published research findings

Evidence- and Value-based Solutions for Health Care Clinical Improvement Consults, Content Development, Training & Seminars, Tools

Delfini Evidence Tool Kit

Polypharmacy - arrhythmic risks in patients with heart failure

Beyond the intention-to treat effect: Per-protocol effects in randomized trials

JUPITER NEJM Poll. Panel Discussion: Literature that Should Have an Impact on our Practice: The JUPITER Study

Biases in clinical research. Seungho Ryu, MD, PhD Kanguk Samsung Hospital, Sungkyunkwan University

The Wellbeing Course. Resource: Mental Skills. The Wellbeing Course was written by Professor Nick Titov and Dr Blake Dear

Physiology to Improve RCTs

A Case Study: Two-sample categorical data

Strategies for handling missing data in randomised trials

Trial: Take-Home Message: Executive Summary: Guidelines:

Randomized Controlled Trial

Top 5 (Topics) Papers In GIM Rocky Mountain ACP Internal Medicine Meeting Raj Padwal November 13, 2008

In this second module in the clinical trials series, we will focus on design considerations for Phase III clinical trials. Phase III clinical trials

CRITICAL APPRAISAL OF MEDICAL LITERATURE. Samuel Iff ISPM Bern

Presentation Outline. Introduction to Biomedical Research Designs. EBM: A Practical Definition. Grade the Evidence (Example McMaster Grading System)

Population. population. parameter. Census versus Sample. Statistic. sample. statistic. Parameter. Population. Example: Census.

Incorporating Clinical Information into the Label

Clinical Trials. Susan G. Fisher, Ph.D. Dept. of Clinical Sciences

GLOSSARY OF GENERAL TERMS

Sheila Barron Statistics Outreach Center 2/8/2011

Conduct an Experiment to Investigate a Situation

A Guide to Reading a Clinical or Research Publication

GATE: Graphic Appraisal Tool for Epidemiology picture, 2 formulas & 3 acronyms

Clinical Research Design and Conduction

Generalizing the right question, which is?

Challenges of Observational and Retrospective Studies

Department of OUTCOMES RESEARCH

Observational Study Designs. Review. Today. Measures of disease occurrence. Cohort Studies

Disclosure. No relevant financial relationships. Placebo-Controlled Statin Trials

Metabolism is All About Burning Calories. Most people talk about your metabolism like it s all about burning calories.

Correlation and Causation. Phil 12: Logic and Decision Making Winter 2010 UC San Diego 2/17/2010

Truth Versus Truthiness in Clinical Data

The role of Randomized Controlled Trials

Dr. Engle has received an unrestricted educational grant for the MD Anderson Pain Medicine Fellowship.

Critical Appraisal of Evidence

Protocol Development: The Guiding Light of Any Clinical Study

THERAPEUTIC MISCONCEPTION: Objectives. Definitions. What is it, Why it matters, and How to minimize it

Let s look a minute at the evidence supporting current cancer screening recommendations.

2016 Update in Geriatrics Elizabeth Eckstrom, MD, MPH Oregon Health & Science University Oregon Geriatrics Society October 7, 2016

SAMPLING AND SAMPLE SIZE

Determinants of quality: Factors that lower or increase the quality of evidence

HYPOTHESIS TESTING 1/4/18. Hypothesis. Hypothesis. Potential hypotheses?

No relevant financial relationships

Transcription:

Kathleen A. Kennedy, MD, MPH University of Texas Medical School at Houston Medical School Most clinical research evaluates an association between exposure/intervention outcome. 1

Does the investigator decide who gets the intervention/exposure? Yes Experimental/Interventional Study Usually Randomized No Observational Study Comparison Between Groups Yes No Analytical Study Descriptive Study Cohort Study Usually Prospective Case Control Study Retrospective Cross sectional Study Cohort Design Exposure + - Observer Time? Disease? Disease Case Control Design? Exposure? Exposure Time + - Observer Disease 2

Reduces symptoms of menopause, but should it be used to reduce/prevent cardiovascular disease? Large number of observational studies (case control and cohort) concluded that hormone replacement therapy reduced the risk of cardiovascular disease in post menopausal women. Women s Health Initiative randomized controlled trial relative risk of coronary heart disease 1.29 (1.02 1.63) with use of hormones What s the likely explanation? Women s Health Initiative Group. JAMA 2002 Postmenopausal Women Take hormone replacements Do not take hormone replacements More affluent More Caucasian Better health care access More healthy behaviors Less affluent Less Caucasian Poorer health care access Less healthy behaviors Less cardiac disease More cardiac disease 3

Postmenopausal Women Randomly Assigned Equivalent risk of cardiac disease Equivalent risk of cardiac disease Hormone replacements Placebo More cardiac disease Less cardiac disease Most rigorous design for assessing the effects of interventions Groups are comparable at study entry should differ only by intervention under investigation, even if there are unrecognized important variables But 4

Part of the problem is our slavish devotion to the p value. And the common practice of saying, based on a p value, that there is or isn t a difference. What does the p value really mean? The likelihood that a difference this large could have been observed between the study groups, just by chance, when there really is no difference in the underlying populations (conditions). 5

We never measure outcomes in the entire group we re interested in. We study a sample from each group. If you just pick a few people at random from one group and a few people at random from another group and you measure something (blood sugar, BMI, whatever), the mean in one sample will probably be a little different than the mean in the other sample, just because of chance variation in sampling. BMI BMI Blood Sugar Blood Sugar First Grade Fourth Grade First Grade Fourth Grade The World My Sample 6

Blood Sugar Blood Sugar p>0.05 BMI BMI p>0.05 First Grade Fourth Grade First Grade Fourth Grade The World My Sample Blood Sugar Blood Sugar p>0.05 BMI BMI p<0.05 First Grade Fourth Grade First Grade Fourth Grade The World My Sample 7

True effects of the exposure/intervention/treatment Sampling variation or chance (5% of the time when p value of 0.05 is used) Inherent differences between the groups apart from the exposure or treatment of interest Differences between handling or evaluation of the groups There truly is no difference between the two populations (exposures/treatments). Inherent differences in the study groups or in their handling or evaluation. Because of a small sample size, there appears to be no difference, based on the statistical test. 8

How does that work? Ability to identify a significant difference depends on the size of the difference between the two groups the prevalence of the outcome (if it s a yes/no outcome like death) the variability in the outcome (if it s a continuous outcome like BP) and the number of patients studied. Blood Sugar Blood Sugar p>0.05 BMI BMI p>0.05 First Grade Fourth Grade First Grade Fourth Grade The World My Sample 9

Calculated before the study is started Power = likelihood that you will be able to identify a statistically significant difference if there really is a difference between the populations BUT It depends on how big a difference you re willing to miss or fail to identify Depends on the SD of your measurement if your outcome is continuous, like BP Depends on the prevalence of the outcome if it s yes/no or categorical, like death Hypothermia vs Control for infants with perinatal asphyxia Calculated sample size of 208 to have 80% power to detect a reduction in death or mod severe disability from 50% to 30% (RR=0.60). What does that mean? If the real truth (we can t know that, especially before doing the study) is that 50% of control infants will have death or disability and hypothermia will reduce this to 30%, we have an 80% chance of finding a statistically significant difference if we enroll 208 subjects. 10

Hypothermia (n=102) Control (n=103) RR (95%C)) Death or Disability 45 (44%) 64 (62%) 0.72 (0.54, 0.95) 0.01 If outcome had been reduced from 50% to 40%, we would have concluded there was no significant difference. RR 0.80 (0,59, 1.09), p=0.16. p If you were the parent, you might think that was an important benefit. It s even worse if the outcome is less common and worse still if there are more than 2 treatment groups. Proposed study of lactoferrin and lactobacillus for prevention of NEC Lactoferrin No Lactoferrin Lactobacillus 13% 18% 15% No Lactobacillus 18% 23% 20% 15% 20% For 80% power, 1828 subjects will be needed. 11

At best, it s an educated guess. People designing a study should calculate it because they should have some clue about how big or small a difference they will be able to identify. There may be other reasons, eg feasibility pilot, for doing a small study, but it needs to be justified. If it s not big enough to answer the questions, what s the justification for putting patients through the hassle or risk? In traditional statistical analyses, when investigator is trying to show a difference, the deck is more or less stacked against concluding that there is a difference. We require that the p value (likelihood that the observed difference was due to chance) is less than 5%. At most, we expect as power of 80 90% to identify a difference if there is one. We typically accept a greater likelihood of falsely concluding that there is no difference. 12

Non inferiority study is one type of equivalence study (investigator wants to conclude that the treatments are the same). The statistical analysis is flipped the other way so that it s not so easy to falsely conclude that there s no difference. Investigator must specify the (usually fairly small) difference that they will accept as equivalent. Then the sample size required will usually be very large. Let s say you want to compare a new antihypertensive drug to placebo, and you want to have 80% power to detect a difference of 5 mm Hg in BP (SD = 15)? You need a sample size of 142/group Let s say you want to compare a new drug to an old drug and have 80% power to conclude that the BP is the same within 2 mm Hg (SD = 15)? You need a sample size of 964/group 13

Benefits may be overestimated because of unmasked allocation failure to mask caregivers or assessors use of surrogate outcomes analysis does not include all enrolled subjects ( intention to treat ) The person(s) making/influencing the decision about whether the subject is enrolled are aware of which treatment the next subject will get. Not the same as unmasked treatment or unmasked outcome assessment. 14

Postmenopausal Women Randomly Assigned Equivalent risk of cardiac disease Equivalent risk of cardiac disease Postmenopausal Women Randomly Assigned Next patient gets hormones Next patient gets placebo Symptoms No symptoms Symptoms No symptoms Enroll Don t enroll Don t enroll Enroll Sicker, higher risk Less sick, lower risk RCT of steroids to facilitate weaning and extubation in premature infants: I think steroids will be beneficial so I wean the ventilator and attempt extubation more aggressively when I know the infants have received steroids steroids appear to be beneficial only because I treated those infants differently. I think steroids will cause glucose intolerance so I give fewer calories to babies when I know they re on steroids steroids appear to have a harmful effect on growth only because I treated those infants differently. 15

Laboratory test or physical finding that is believed to predict an outcome that is clinically important. Previous randomized trial showed that anti arrhythmic drugs reduced EKG abnormalities in post MI patients. Arrhythmias can be fatal. In a larger RCT, encainide or flecainide, as compared to placebo, increased post MI mortality. Echt DS et al. New Engl J Med 1991 Analyzing all subjects according to their group assignment, regardless of whether they received or completed the assigned treatment. 16

Postmenopausal Women Randomly Assigned Hormones Equivalent risk of cardiac disease Placebo Equivalent risk of cardiac disease High risk patients have more side effects, stop study drug More deaths High risk patients continue taking placebo Fewer deaths Survivors who continue study drug are relatively low risk Survivors who continue study drug are relatively high risk Benefits or hazards may be underestimated because of poorly implemented intervention inadequate follow up 17

Benefits may be overestimated for the target population (the ones where the treatment will be applied in clinical practice) highly selected patients (highly sensitive to treatment, highly compliant, low risk of complications) overestimates beneficial effects and underestimates adverse effects of use in a more diverse population high risk of primary outcome Eligibility criteria for hospitalized infants < 35 weeks gestation at birth No reporting of eligible non enrolled subjects Average gestational age was <32 weeks Impact RSV Group. Pediatrics 1998 18

Risk Difference = RD = 53/500 (10.6%) 48/1002 (4.8%) = 4.8% Is it worth it? NNT: You need to treat 21 infants (5 injections each at $1500/injection) to prevent 1 RSV hospitalization Palivizumab was FDA approved and recommended for infants < 35 weeks gestation Risk of RSV hospitalization is ~ 2% for 33 34 week infants If you decrease it by half (2% 1%), RD = 1% You need to treat 100 infants (5 injections each at $1500/injection) to prevent 1 RSV hospitalization 19

Schulz KF et al. Ann Int Med 2010 Hypothermia study had FU for all but 3 subjects in the control group (98.6% follow up). Loss to follow up isn t a big problem if it s random (nondifferential loss. If you lost 10% of the patients from each cell below, the RR would be the same; your p value would be higher but it would still be significant. Hypothermia (n=102) Control (n=103) Death or Disability 45 64 No death/disability 57 39 RR (95%C)) p 0.72 (0.54, 0.95) 0.01 But what if it s not random? 20

What if we had 0% loss of cooled infants and 0% loss of good outcomes and 15% loss of control infants with bad outcomes? Hypothermia (n=102) Control (n=103) Death or Disability 45 64 No death/disability 57 39 RR (95%C)) p 0.72 (0.54, 0.95) 0.01 Hypothermia (n=102) Control (n=93) Death or Disability 45 54 No death/disability 57 39 RR (95%C)) p 0.76 (0.58, 1.00) 0.052 Randomized allocation Masked allocation Masked intervention (or other measures to ensure similar treatment of intervention and controls) Masked outcome assessment (or other measures to standardize assessment) Minimal loss to follow up Intention to treat analysis Reporting of enrolled vs non enrolled subjects 21

? 22