Levels of Evaluation. Question. Example of Method Does the program. Focus groups with make sense? target population. implemented as intended?

Similar documents
Threats to validity in intervention studies. Potential problems Issues to consider in planning

The following are questions that students had difficulty with on the first three exams.

Experimental and Quasi-Experimental designs

Formative and Impact Evaluation. Formative Evaluation. Impact Evaluation

OBSERVATION METHODS: EXPERIMENTS

In this chapter we discuss validity issues for quantitative research and for qualitative research.

Experimental Design Part II

HPS301 Exam Notes- Contents

Research Approach & Design. Awatif Alam MBBS, Msc (Toronto),ABCM Professor Community Medicine Vice Provost Girls Section

STATISTICAL CONCLUSION VALIDITY

Educational Research. S.Shafiee. Expert PDF Trial

Sample Exam Questions Psychology 3201 Exam 1

AOTA S EVIDENCE EXCHANGE CRITICALLY APPRAISED PAPER (CAP) GUIDELINES Annual AOTA Conference Poster Submissions Critically Appraised Papers (CAPs) are

What We Will Cover in This Section

PSY 250. Designs 8/11/2015. Nonexperimental vs. Quasi- Experimental Strategies. Nonequivalent Between Group Designs

Study Design. Svetlana Yampolskaya, Ph.D. Summer 2013

A situation analysis of health services. Rachel Jewkes, MRC Gender & Health Research Unit, Pretoria, South Africa

Evaluating Social Programs Course: Evaluation Glossary (Sources: 3ie and The World Bank)

Design of Experiments & Introduction to Research

Myths About HIV. Last updated: July 19, Together, we can change the course of the HIV epidemic one woman at a time.

UNIT III: Research Design. In designing a needs assessment the first thing to consider is level for the assessment

CHAPTER LEARNING OUTCOMES

Introduction to Experiments

Choosing the right study design

Experimental Research. Types of Group Comparison Research. Types of Group Comparison Research. Stephen E. Brock, Ph.D.

Define the population Determine appropriate sample size Choose a sampling design Choose an appropriate research design

GUIDE 4: COUNSELING THE UNEMPLOYED

Epidemiologic Methods and Counting Infections: The Basics of Surveillance

What?! Why?! Emergency Contraception. physical consequences. emotional consequences

PLS 506 Mark T. Imperial, Ph.D. Lecture Notes: Reliability & Validity

What We Will Cover in This Section

For People Who Have Been Sexually Assaulted... What You Need To Know about STDs and Emergency Contraception

Review and Wrap-up! ESP 178 Applied Research Methods Calvin Thigpen 3/14/17 Adapted from presentation by Prof. Susan Handy

Experimental Design and the struggle to control threats to validity

I M ENDING HIV PATIENT INFORMATION. endinghiv.org.au/prep

VALIDITY OF QUANTITATIVE RESEARCH

Empirical Knowledge: based on observations. Answer questions why, whom, how, and when.

Internal Validity and Experimental Design

CORE ELEMENTS, KEY CHARACTERISTICS AND LOGIC MODEL

2-Group Multivariate Research & Analyses

Psych 1Chapter 2 Overview

11 Misconceptions & Myths About HIV

Lecture 4: Research Approaches

Georgina Salas. Topics EDCI Intro to Research Dr. A.J. Herrera

11/2/2017. Often, we cannot manipulate a variable of interest

Informed Consent Flipchart. Version 1.0, 30 Jan 2018

EVALUATION TOOLS...123

10/23/2017. Often, we cannot manipulate a variable of interest

Research Methods & Design Outline. Types of research design How to choose a research design Issues in research design

Research Design. Beyond Randomized Control Trials. Jody Worley, Ph.D. College of Arts & Sciences Human Relations

More on Experiments: Confounding and Obscuring Variables (Part 1) Dr. Stefanie Drew

PM12 Validity P R O F. D R. P A S Q U A L E R U G G I E R O D E P A R T M E N T O F B U S I N E S S A N D L A W

As a result of this training, participants will be able to:

Lecture 5 Conducting Interviews and Focus Groups

investigate. educate. inform.

Chapter 10 Quasi-Experimental and Single-Case Designs

Research Process. the Research Loop. Now might be a good time to review the decisions made when conducting any research project.

Threats and Analysis. Bruno Crépon J-PAL

9 research designs likely for PSYC 2100

The Regression-Discontinuity Design

Chapter 9 Experimental Research (Reminder: Don t forget to utilize the concept maps and study questions as you study this and the other chapters.

A Sexual Health Study with Africans in Frankfurt am Main

2013/4/28. Experimental Research

Quantitative OR methodologies I

As a result of this training, participants will be able to:

CRITICALLY APPRAISED PAPER (CAP)

Lecture 9 Internal Validity

August 29, Introduction and Overview

TECH 646 Analysis of Research in Industry and Technology. Experiments

Health Care Professional Questionnaire

Research Methods. Donald H. McBurney. State University of New York Upstate Medical University Le Moyne College

DATA CORE MEETING. Observational studies in dementia and the MELODEM initiative. V. Shane Pankratz, Ph.D.

Chapter 1 Introduction to I/O Psychology

Experiments. ESP178 Research Methods Dillon Fitch 1/26/16. Adapted from lecture by Professor Susan Handy

Conducting Research in the Social Sciences. Rick Balkin, Ph.D., LPC-S, NCC

Biostatistics and Design of Experiments Prof. Mukesh Doble Department of Biotechnology Indian Institute of Technology, Madras

Evaluation and Assessment of Health Policy. Gerard F. Anderson, PhD Johns Hopkins University

Analysis A step in the research process that involves describing and then making inferences based on a set of data.

Chapter 11 Nonexperimental Quantitative Research Steps in Nonexperimental Research

SOCQ121/BIOQ121. Session 8. Methods and Methodology. Department of Social Science. endeavour.edu.au

TECH 646 Analysis of Research in Industry and Technology. Experiments

The Practice of Statistics 1 Week 2: Relationships and Data Collection

Variables Research involves trying to determine the relationship between two or more variables.

26:010:557 / 26:620:557 Social Science Research Methods

04/12/2014. Research Methods in Psychology. Chapter 6: Independent Groups Designs. What is your ideas? Testing

The Logic of Data Analysis Using Statistical Techniques M. E. Swisher, 2016

CHAPTER 8 EXPERIMENTAL DESIGN

Research Methodology. Characteristics of Observations. Variables 10/18/2016. Week Most important know what is being observed.

You can t fix by analysis what you bungled by design. Fancy analysis can t fix a poorly designed study.

Threats to Validity in Experiments. The John Henry Effect

Evaluation: Scientific Studies. Title Text

Lubrication. What?! Why?! But our sexual health is not private! thoughts and discussions

Index. Handbook SCREENING & TREATMENT ENHANCEMENT P A R T STEP. Guidelines and Program Information for First Felony and Misdemeanor Participants

Psychological Science Handout 2 Research Designs Page 2.1 HANDOUT 2 RESEARCH DESIGNS

Critical Appraisal Series

Community Mobilizer s Guide. A step by step guide to help you mobilize your community to prevent HIV from spreading

Lecture 3. Previous lecture. Learning outcomes of lecture 3. Today. Trustworthiness in Fixed Design Research. Class question - Causality

Validity and Quantitative Research. What is Validity? What is Validity Cont. RCS /16/04

Assessing the Impact of a HIV Prevention Program Among Transgender People in Thailand Using Coarsened Exact Matching

Chapter Three Research Methodology

Transcription:

Program Evaluation Jonathan Howland, PhD, MPH 1 Formative Process Level Outcome Levels of Evaluation Question Example of Method Does the program Focus groups with make sense? target population. Was the program implemented as intended? Was the program effective? 2 Examples of Formative Evaluation Focus groups Field observation Interviews with program stakeholders (e.g. health care providers, patients) Public forums (e.g. community meetings) 3 Examples of Process Evaluations Participant attendance/compliance records Direct observation of program implementation Participant satisfaction questionnaires Log book of critical implementation events Scanning media for current events that might affect program implementation of outcomes Interviews with program staff Reasons for Conducting a Process Evaluation Accepting the null hypothesis (program did not result in hypothesized change) Replication Program management Accountability Study Observational Experiment Types of Causal Study Hypothesis Does exposure Coincidence of Correlation X cause health X and Y in time Case-control problem Y? and space CohortCohort Does program X improve health outcome Y? Causal logic Wiggle X to observe change in Y Examples Measure participant compliance with program regimen. Randomized trial or quasi- experiment RCTRCT Cluster Cluster RCT Quasi- experiment 4 5 6 1

Scientific Method for Establishing Causality X causes Y, if no reasonable alternative cause (c) can be identified Example: The diet program was effective because no other explanation for participant weight loss could be identified. Causality in Jurisprudence: Prosecution: Pete killed Stan because no one else could have done it. Defense: Joe might have killed Stan because Joe had a motive and an opportunity. Trial: Court Trial participant weight loss could be identified. 7 8 Causality in Program Evaluation: Hypothesis 1: The new diet program caused weight reduction because no other factors influenced participants. Alternative Explanation: Participants were all infected with gastro-intestinal disease. Trial: Randomized Control Trial (RCT) Method: Peer review of paper reporting of Method: Jury of Pete s s peers. study results. 9 Types of Validity Threats to Internal Validity Testing Type Internal External Question Is the causal relationship between the program and the outcome real or spurious (confounded)? If the program worked in a randomized trial using recruited volunteers, will it also work with another population? Testing History Instrumentation Maturation Differential Attrition Statistical Regression Selection Contamination Hawthorn Effect Chance The pretest measurement by itself has an impact on the posttest measurement Practice effects Self-reflection Awareness of hypothesis Social desirability Reactivity 10 11 12 2

History Instrumentation Maturation There may be unplanned events that occur between the first and second measurements Programs Policies Events The more time that elapses between measurements, the greater the threat from historical effects The measurement instrument, or the criteria for recording behavior, change between the pretest and the posttest Context of survey questions Interviewer skills Archival records Every person changes over time Development Aging Role transition Seasonal effects 13 14 15 Attrition There might be unequal dropout rates among the experimental and comparison groups Those dropping out from one group may be different from those dropping out of another group Statistical Regression If a study sample was chosen due to extreme behavior, then over time that behavior will tend to regress toward the overall group average For anyone with an extreme pretest score, the best prediction is that their posttest score will be less extreme Selection There are unmeasured differences between an experimental group and a nonequivalent comparison group Motivation Family and social support structures Or any number of other things 16 17 18 3

Contamination Hawthorne Effect Chance The control group was inadvertently exposed to the intervention program. The very fact of being in an evaluation can result in changed behavior Similar to a placebo effect in medical research The outcome was random and had nothing to do with the intervention program. 19 20 21 Results Sig. Diff Confounder Drives > Null Hypo < Null Hypo OK Confounded One Group Pre-test Post-test test Design O X O One Group Pre-test Post-test test Design Example: Pre-test Mean Post-test test Mean 150 lbs X 120 lbs No Sig. Diff Confounded OK 22 23 Are there any other factors that may have caused this change from pre- test to post-test? test? 24 4

Nonequivalent Comparison Group Design (Quasi-Experiment) E O X O C O O 25 Nonequivalent Comparison Group Design Example: (Quasi-Experiment) Pre-test Mean Post-test test Mean E 150 lbs X 120 lbs C 143 lbs 147lbs Are there any other factors that may have caused this change from pre- test to post-test? Drawbacks to Nonequivalent Comparison Group Design Group assignment is not random The two groups can differ from one another in many ways, some of which might affect the changes in outcome measures Various pretest measures might show that the groups are very similar but we can t measure everything 26 27 test? The True Experiment Controlled trial Random assignment among experimental and control groups. Protects against many threats to internal validity E O X O R C O O The True Experiment Example: Pre-test Mean Post-test test Mean E 150 lbs X 120 lbs C 150 lbs 147lbs Are there any other factors that may have caused this change from pre- test to post-test? test? Almost Perfect This design controls for almost all threats to validity Absence of selection effects should be verified Did the randomization work? 28 29 30 5

Remaining Threats to Validity True Experiments: Cluster Trials Analysis of Randomized Trial Differential attrition Treatment group exposure to other program Contamination Hawthorn effect Chance Cluster Randomized Trials or Group Randomized Trials (GRTs) Trials designed to evaluate group interventions that assign intact groups (e.g., schools, workplaces) to experimental conditions Individually Randomized Group Treatment (IRGTSs)Trials Trials that assign individuals to study conditions when interventions are delivered to groups (e.g., exercise groups Both designs involve potential correlation among observations within treatment groups that must be adjusted for in power calculations Outcome Measure (higher score= improvement) A, B C Treatment Compare post-test scores C and D D Control 31 32 Pre-test Post-test 33 Analysis of Nonequivalent Groups Outcome Measure (higher score= improvement) (Quasi-experiment) C-A= 1 D-B= 2 A C Treatment Compare 1 and 2 Types of Errors in Hypothesis Testing Types of error in hypothesis testing Difference Reality No difference EVALUATING AN INTERVENTION TO REDUCE NEW STIs B Pre-test Post-test D Control 34 Study Finding Difference 1 No difference 2 OK Type II 1 Reject null hypothesis 2 Accept null hypothesis Type I OK 35 36 6

WHAT IS THE PUBLIC HEALTH PROBLEM? About 25% of STD clinic patients return with a new infection This group may constitute a core that serves as a reservoir of disease in the community STIs that involve lesions can enhance transmission of HIV CAN ANYTHING BE DONE TO REDUCE THE RATE OF NEW INFECTIONS? Several studies have shown that intensive counseling of clinic patients can reduce rates of new infections IS THE STD CLINIC A GOOD PLACE TO INTERVENE FOR THIS PROBLEM? 37 38 39 WHAT SHOULD THE BEHAVIORAL OBJECTIVE BE FOR THE INTERVENTION? WHAT ARE THE BARRIERS TO CONDOM USE IN THE TARGET POPULATION? WHAT MEDIUM SHOULD BE USED FOR THE INTERVENTION? 40 41 42 7

VIGNETTE # 1 2 3 4 5 6 7 8 9 STI INTERVENTION STUDY: VIDEO VIGNETTE TOPICS VIGNETTE TOPIC Shows how a woman can negotiate condom use with her primary partner On the street interviews about how to make condom use more fun A rap song with visuals showing the proper way to use a condom Shows how to tell a partner, whom you may have infected, that they have to go to the STD clinic for an exam Shows how to talk to a casual partner about using condoms Shows how to respond to excuses for not wearing condoms A monologue about not allowing alcohol or drugs to inhibit condom use Shows how a man can talk to his primary partner about using condoms An infomercial showing the different types of condoms available WHAT IS THE BEST RESEARCH DESIGN TO USE FOR EVALUATING THIS INTERVENTION? WHAT SHOULD THE OUTCOMES BE? 10 A monologue about how the HIV virus is spread from person to 43 person 44 45 HYPOTHESIS THAT INTERACTIVE VIDEO IS MORE EFFECTIVE THAN LINEAR VIDEO? WHAT SHOULD THE CONTROLS RECEIVE? WHAT IS THE REAL RESEARCH QUESTION? 46 47 48 8

ROUTINE CARE CHECKLIST ROUTINE CARE PATIENT EDUCATION PROTOCOL CHECK LIST AS MANDATED BY STATE CONTRACT Counseled on medical instructions Administration of medications Counseled on medication side effects Counseled on complications of health problem Patient verbalizes understanding of treatment instructions STI literature provided STI education provided Counseled on disease symptoms Counseled on disease transmission Counseled on disease prevention Counseled on disease symptoms Counseled on response to future STIs Need for follow-up visit discussed Asked to return to clinic for follow-up visit Partner referral HIV counseling HIV test offered HIV literature provided Condoms provided 49 Assignment R STI INTERVENTION STUDY DESIGN Groups E1 N=500 E2 N=500 C N=500 Baseline O 1 O 1 O 1 Interventio n X Interactive Video X Linear Video Routine Care 6 wk O 2 O 2 O 2 Follow-Up 6 mo 0 3 0 3 0 3 1 yr 0 4 0 4 0 4 50 THREATS TO VALIDITY? TESTING HISTORY INSTRUMENTATION SELECTION BIAS MATURATION ATTRITION REGRESSION CONTAMINATION HAWTHORN EFFECT 51 WHAT WOULD YOU MEASURE FOR PROCESS EVALUATION? Event log book News media scans Enrollment/refusal rates Follow-up rates Monthly reports PROCESS EVALUATION Direct observation of study implementation Review routine care checklists Participant satisfaction interviews WHAT WERE THE STUDY RESULTS? 52 53 54 9

POSITIVE AT ENROLLMENT MALES AND FEMALES 6 WEEKS POSITIVE AT ENROLLMENT MALES 6 WEEKS POSITIVE AT ENROLLMENT FEMALES 6 WEEKS GROUP POSITIVE NEGATIVE % POSITIVE GROUP POSITIVE NEGATIVE % POSITIVE GROUP POSITIVE NEGATIVE % POSITIVE TREATMEN T CONTROL 54 46 134 87 188 133 29% 35% TREATMENT CONTROL 48 38 89 55 137 93 35% 41% TREATMEN T CONTROL 6 8 45 32 51 40 12% 20% 100 221 321 31% 86 144 230 37% 14 77 91 15% OR = 1.3, p= 0.26 OR = 1.3, p= 0.37 OR = 1.9, p= 0.28 55 56 57 CONCLUSIONS Could have been a type 2 error Should have done better formative research An intervention with a small effect (6 weeks) to which many are exposed because it is cheap could yield a big effect for the community 58 10