Formative and Impact Evaluation. Formative Evaluation. Impact Evaluation

Similar documents
Chapter 10 Quasi-Experimental and Single-Case Designs

Experimental and Quasi-Experimental designs

CHAPTER EIGHT EXPERIMENTAL RESEARCH: THE BASICS of BETWEEN GROUP DESIGNS

Research Design & Protocol Development

Lecture 4: Research Approaches

Research Design. Beyond Randomized Control Trials. Jody Worley, Ph.D. College of Arts & Sciences Human Relations

Research Methods. Donald H. McBurney. State University of New York Upstate Medical University Le Moyne College

PSY 250. Designs 8/11/2015. Nonexperimental vs. Quasi- Experimental Strategies. Nonequivalent Between Group Designs

HPS301 Exam Notes- Contents

26:010:557 / 26:620:557 Social Science Research Methods

Educational Research. S.Shafiee. Expert PDF Trial

Research Design. Source: John W. Creswell RESEARCH DESIGN. Qualitative, Quantitative, and Mixed Methods Approaches Third Edition

Chapter 9 Experimental Research (Reminder: Don t forget to utilize the concept maps and study questions as you study this and the other chapters.

PLS 506 Mark T. Imperial, Ph.D. Lecture Notes: Reliability & Validity

EXPERIMENTAL RESEARCH DESIGNS

Define the population Determine appropriate sample size Choose a sampling design Choose an appropriate research design

Levels of Evaluation. Question. Example of Method Does the program. Focus groups with make sense? target population. implemented as intended?

EXPERIMENTAL AND EX POST FACTO DESIGNS M O N A R A H I M I

Quasi-Experimental and Single Case Experimental Designs. Experimental Designs vs. Quasi-Experimental Designs

Georgina Salas. Topics EDCI Intro to Research Dr. A.J. Herrera

Research Approach & Design. Awatif Alam MBBS, Msc (Toronto),ABCM Professor Community Medicine Vice Provost Girls Section

TECH 646 Analysis of Research in Industry and Technology. Experiments

RESEARCH METHODS. Winfred, research methods,

Experimental Design Part II

WWC Standards for Regression Discontinuity and Single Case Study Designs

11/2/2017. Often, we cannot manipulate a variable of interest

M6728. Goals. Depression Scores. Research Designs

RESEARCH METHODS. A Process of Inquiry. tm HarperCollinsPublishers ANTHONY M. GRAZIANO MICHAEL L RAULIN

RESEARCH METHODS. Winfred, research methods, ; rv ; rv

The Regression-Discontinuity Design

CRITICALLY APPRAISED PAPER (CAP)

PM12 Validity P R O F. D R. P A S Q U A L E R U G G I E R O D E P A R T M E N T O F B U S I N E S S A N D L A W

Quantitative Research Methods FSEHS-ARC

Experimental Research. Types of Group Comparison Research. Types of Group Comparison Research. Stephen E. Brock, Ph.D.

In this chapter we discuss validity issues for quantitative research and for qualitative research.

10/23/2017. Often, we cannot manipulate a variable of interest

MBACATÓLICA JAN/APRIL Marketing Research. Fernando S. Machado. Experimentation

Overview of Perspectives on Causal Inference: Campbell and Rubin. Stephen G. West Arizona State University Freie Universität Berlin, Germany

investigate. educate. inform.

QUASI EXPERIMENTAL DESIGN

One slide on research question Literature review: structured; holes you will fill in Your research design

Causal Validity Considerations for Including High Quality Non-Experimental Evidence in Systematic Reviews

DEPENDENT VARIABLE. TEST UNITS - subjects or entities whose response to the experimental treatment are measured or observed. Dependent Variable

REPEATED MEASURES DESIGNS

TECH 646 Analysis of Research in Industry and Technology. Experiments

CHAPTER LEARNING OUTCOMES

Study Design. Svetlana Yampolskaya, Ph.D. Summer 2013

Research Designs. Inferential Statistics. Two Samples from Two Distinct Populations. Sampling Error (Figure 11.2) Sampling Error

Regression Discontinuity Analysis

Topic #6. Quasi-experimental designs are research studies in which participants are selected for different conditions from pre-existing groups.

Sample Exam Questions Psychology 3201 Exam 1

What We Will Cover in This Section

Lecture 3. Previous lecture. Learning outcomes of lecture 3. Today. Trustworthiness in Fixed Design Research. Class question - Causality

Chapter 7. Marketing Experimental Research. Business Research Methods Verónica Rosendo Ríos Enrique Pérez del Campo Marketing Research

What We Will Cover in This Section

Conducting Strong Quasi-experiments

More on Experiments: Confounding and Obscuring Variables (Part 1) Dr. Stefanie Drew

PYSC 224 Introduction to Experimental Psychology

Version No. 7 Date: July Please send comments or suggestions on this glossary to

QUASI-EXPERIMENTAL HEALTH SERVICE EVALUATION COMPASS 1 APRIL 2016

The following are questions that students had difficulty with on the first three exams.

Checklist for Randomized Controlled Trials. The Joanna Briggs Institute Critical Appraisal tools for use in JBI Systematic Reviews

Experimental Design (7)

Design of Experiments & Introduction to Research

I. Methods of Sociology Because sociology is a science as well as being a theoretical discipline, it is important to know the ways in which

Lecture 9 Internal Validity

Pskov Youth Reproductive Health Project

UNIT III: Research Design. In designing a needs assessment the first thing to consider is level for the assessment

1. Which type of research is constrained by ethical considerations? C. The events in the scientific world are unpredictable

Conditional Average Treatment Effects

CRITICALLY APPRAISED PAPER (CAP)

Research Designs For Evaluating Disease Management Program Effectiveness

WWC STUDY REVIEW STANDARDS

Research Approaches Quantitative Approach. Research Methods vs Research Design

UNIT 7 EXPERIMENTAL RESEARCH-I1

STATISTICAL CONCLUSION VALIDITY

PYSC 224 Introduction to Experimental Psychology

Which Comparison-Group ( Quasi-Experimental ) Study Designs Are Most Likely to Produce Valid Estimates of a Program s Impact?:

Quantitative OR methodologies I

Recent developments for combining evidence within evidence streams: bias-adjusted meta-analysis

OBSERVATION METHODS: EXPERIMENTS

Experimental Validity

EXPERIMENTAL AND QUASI-EXPERIMENTAL EVALUATIONS

FCE 3900 EDUCATIONAL RESEARCH E X P E R I M E N T A L A N D Q U A S I E X P E R I M E N T A L R E S E A R C H

2013/4/28. Experimental Research

Working Paper: Designs of Empirical Evaluations of Non-Experimental Methods in Field Settings. Vivian C. Wong 1 & Peter M.

Threats to validity in intervention studies. Potential problems Issues to consider in planning

Evaluating Social Programs Course: Evaluation Glossary (Sources: 3ie and The World Bank)

Can Quasi Experiments Yield Causal Inferences? Sample. Intervention 2/20/2012. Matthew L. Maciejewski, PhD Durham VA HSR&D and Duke University

Measures of Dispersion. Range. Variance. Standard deviation. Measures of Relationship. Range. Variance. Standard deviation.

Methods of Reducing Bias in Time Series Designs: A Within Study Comparison

The use of quasi-experiments in the social sciences: a content analysis

Introduction to Experiments

Just what research design am I using? Shona J. Kelly Professor of Interdisciplinary Research

Using natural experiments to evaluate population health interventions

CHAPTER TEN SINGLE-SUBJECTS (QUASI-EXPERIMENTAL) RESEARCH

chapter thirteen QUASI-EXPERIMENTAL AND SINGLE-CASE EXPERIMENTAL DESIGNS

Systematic reviews and meta-analyses of observational studies (MOOSE): Checklist.

Checklist for Case Control Studies. The Joanna Briggs Institute Critical Appraisal tools for use in JBI Systematic Reviews

Group 3: Individuals and Society. Options: HL History and HL Psychology

Transcription:

Formative and Impact Evaluation Formative Evaluation 2 An evaluation designed to produce qualitative and quantitative data and insight during the early developmental phase of an intervention, including assessment of: Feasibility of program implementation Appropriateness of content, methods, materials, media, and instruments Immediate or short-term cognitive, psychosocial, psychomotor (skill), and/or behavioral impact of an intervention for a well-define population at risk. Impact Evaluation Can the observed change in the behavioral impact rate, if statistically significant, be attributed to the health promotion program? 3 1

External Validity The extent to which an observed, significant change in a behavioral or health impact rate, attributed to a feasible, replicable intervention and documented by multiple evaluations with high internal validity, can be generalized to other program settings and to comparable populations at risk 4 Internal Validity 5 The extent to which an observed impact, a significant improvement in cognitive, skill, behavioral, health status, or economic indicator rate, can be attributed to a planned intervention Did a planned, replicable intervention (X) produce the significant increase in the impact rate (O)? Is X the only or the most plausible explanation for the changes in O, or can all or part of the observed change be attributed to other factors? Primary Sources of Bias Measurement Core sources of bias: poor validity and poor reliability Must be eliminated in order to determine how much change in an impact rate occurred Randomization does not control for this Combines testing and instrumentation threats 6 2

Primary Sources of Bias 7 Selection Sources: participation rate and attrition rate A program must describe eligibility criteria for the defined population at risk Ability to generalize results to a defined population may be severely limited or impossible if not addressed Primary questions: How representative are the results? Can the results of an evaluation be applied to a defined population in this setting? Combines maturation, statistical regression, selection, and participant attrition threats Primary Sources of Bias Historical Primary sources: Transient or enduring external historical events Internal programmatic, historical events Specific intervention procedures delivered or not delivered The extent to which planned exposure to the program was documented and the degree of standardization, replicability, and stability of intervention procedures by staff Combines history and maturation threats 8 Evaluation Designs Start with the most rigorous design possible and then, if necessary, modify it to fit the setting Assess each design to determine the degree to which a program will be able to attribute an observed impact to your program Purpose: To create E vs. C group equivalence at baseline and follow-up 9 3

Evaluation Design 10 Categories Nonexperimental Includes one E group but no C group Quasi-experimental Includes an E group and C group created by methods other than randomization and includes observations of both groups prior to and after the intervention Experimental Includes random assignment of an E group and C group, and observations of both groups, prior to and after the intervention One-Group Pretest and Posttest Most basis design Should not be used to assess program impact Can be used for pilot study Primary weakness: Extent to which participants selected for the program are similar to typical users 11 Nonequivalent Comparison Group Adds an experimental group to the onegroup pretest and posttest design May improve a program s ability to rule out some alternative explanations of impact 12 4

Time Series 13 Can be used if a program can: Establish the periodicity and pattern of the impact rate being examined for a well-defined population at risk Observe multiple monthly/quarterly data points before and after the new intervention Establish the validity, reliability, and completeness of measurement Collect data unobtrusively as a part of an existing monitoring system Document introduction of an intervention at a specific time and, if planned, withdrawal of the intervention abruptly at a specific time Time Series 14 Requires adequate number of observations to document impact rate trends To document the significance of a trend in an observed rate, the treatment must be powerful enough to produce shifts in an impact rate considerably beyond the normally observed variations in the rate External and internal historical biases increase significantly with extended duration Multiple Time Series Impact rates are studied at different times over several years for an experimental group and comparison group Can only be applied in situations where retrospective and prospective databases exist and are accessible or where an organization can periodically observe rates from program participants Major threat: selection bias 15 5

Randomized Pretest and Posttest with an E and C Group Randomization is used to establish two groups, E and C, not significantly different at baseline for any independent or dependent variables that are predictors of impact Random selection Random assignment Should control the 3 major biases to internal validity, assuming no major implementation problems 16 Quasi-Experimental Designs Recommendations to reduce bias: Select internal rather than external comparison subjects Avoid participants self selection into E or C conditions Consider selecting comparison groups on stable pretest measures Include pretests on the dependent variable and use to adjust the posttest Use sensitivity analyses to explore possible effects of hidden bias and different selection assumptions Significantly reduce or prevent attrition by pilot testing 17 Establishing a Comparison (C) Group Matching with external units Matching with internals units participants Apply the one-group pretest and posttest in the same site with standardized methods and procedures Participant- or peer-generated C groups Participants match themselves with a nonparticipant friend 18 6

Establishing a Control Group 19 Random assignment Delayed treatment Randomly assign participants to either immediate (E) or delayed (C) treatment, with each having an equal opportunity to participate Multiple-component program Determine the differential impact of multiple components of a program by exposing subsamples or participants to different components Randomized, factorial study New Program Comparison of new program to existing or standard program 7