Measuring Impact. Conceptual Issues in Program and Policy Evaluation. Daniel L. Millimet. Southern Methodist University.

Similar documents
Measuring Impact. Program and Policy Evaluation with Observational Data. Daniel L. Millimet. Southern Methodist University.

Quasi-experimental analysis Notes for "Structural modelling".

Quantitative Methods. Lonnie Berger. Research Training Policy Practice

Experimental Design Notes

Introduction to Program Evaluation

August 29, Introduction and Overview

Dopamine and Reward Prediction Error

Subjects Motivations

A Brief Introduction to Bayesian Statistics

Coordination in Sensory Integration

Reassessing Quasi-Experiments: Policy Evaluation, Induction, and SUTVA. Tom Boesche

Unlike standard economics, BE is (most of the time) not based on rst principles, but on observed behavior of real people.

Potential Surprise Theory as a theoretical foundation for scenario planning

Manitoba Centre for Health Policy. Inverse Propensity Score Weights or IPTWs

Impact Evaluation Methods: Why Randomize? Meghan Mahoney Policy Manager, J-PAL Global

ICPSR Causal Inference in the Social Sciences. Course Syllabus

Assessing Studies Based on Multiple Regression. Chapter 7. Michael Ash CPPA

Econometric analysis and counterfactual studies in the context of IA practices

TRANSLATING RESEARCH INTO ACTION. Why randomize? Dan Levy. Harvard Kennedy School

Propensity Score Analysis Shenyang Guo, Ph.D.

Book review of Herbert I. Weisberg: Bias and Causation, Models and Judgment for Valid Comparisons Reviewed by Judea Pearl

Introduction to Observational Studies. Jane Pinelis

Executive Summary. Future Research in the NSF Social, Behavioral and Economic Sciences Submitted by Population Association of America October 15, 2010

Pros. University of Chicago and NORC at the University of Chicago, USA, and IZA, Germany

The Practice and Problems of Social Research

Behaviorism: An essential survival tool for practitioners in autism

May 23, 2013 SMU / SMU. Designing Randomized Experiments. Tim Salmon and Danila Serra. Introduction. Benefits of. Experiments.

State-of-the-art Strategies for Addressing Selection Bias When Comparing Two or More Treatment Groups. Beth Ann Griffin Daniel McCaffrey

A note on evaluating Supplemental Instruction

Designed Experiments have developed their own terminology. The individuals in an experiment are often called subjects.

CAUSAL EFFECT HETEROGENEITY IN OBSERVATIONAL DATA

Supplementary Web Appendix Transactional Sex as a Response to Risk in Western Kenya American Economic Journal: Applied Economics

Carrying out an Empirical Project

This module illustrates SEM via a contrast with multiple regression. The module on Mediation describes a study of post-fire vegetation recovery in

Impact Evaluation Toolbox

Introduction to Applied Research in Economics Kamiljon T. Akramov, Ph.D. IFPRI, Washington, DC, USA

Causal Inference in Observational Settings

EXPERIMENTAL DESIGN Page 1 of 11. relationships between certain events in the environment and the occurrence of particular

Causality and Statistical Learning

ACCTG 533, Section 1: Module 2: Causality and Measurement: Lecture 1: Performance Measurement and Causality

An Economic Analysis of Identity and Career Choice

I. Introduction. Armin Falk IZA and University of Bonn April Falk: Behavioral Labor Economics: Psychology of Incentives 1/18

Folland et al Chapter 4

Child health and adult income. Worms at Work: Deworming and schooling in Kenya. Child health and adult income. Presentation by: Edward Miguel

SMALL n IMPACT EVALUATION. Howard White and Daniel Phillips 3ie

Propensity Score Methods for Estimating Causality in the Absence of Random Assignment: Applications for Child Care Policy Research

Key questions when starting an econometric project (Angrist & Pischke, 2009):

EC Macroeconomic Theory Supplement 13 Professor Sanjay Chugh

Biases in clinical research. Seungho Ryu, MD, PhD Kanguk Samsung Hospital, Sungkyunkwan University

Marno Verbeek Erasmus University, the Netherlands. Cons. Pros

Causality and Statistical Learning

Identifying Mechanisms behind Policy Interventions via Causal Mediation Analysis

CSC2130: Empirical Research Methods for Software Engineering

ECON Microeconomics III

CASE STUDY 2: VOCATIONAL TRAINING FOR DISADVANTAGED YOUTH

Keywords: causal channels, causal mechanisms, mediation analysis, direct and indirect effects. Indirect causal channel

Determinants of Health

The scope of perceptual content, II: properties

Lecture 1: Introduction to Personalized Medicine. Donglin Zeng, Department of Biostatistics, University of North Carolina

Evaluating Social Programs Course: Evaluation Glossary (Sources: 3ie and The World Bank)

Theory Building and Hypothesis Testing. POLI 205 Doing Research in Politics. Theory. Building. Hypotheses. Testing. Fall 2015

Health, Mental Health and Labor Productivity: The Role of Self-Reporting Bias

Appendix III Individual-level analysis

Recognizing Ambiguity

Do not copy, post, or distribute

SRDC Technical Paper Series How Random Must Random Assignment Be in Random Assignment Experiments?

Introduction to Applied Research in Economics

Professional Development: proposals for assuring the continuing fitness to practise of osteopaths. draft Peer Discussion Review Guidelines

Chapter 1 Review Questions

Module 14: Missing Data Concepts

THE WORLD MEDICAL ASSOCIATION, INC. WMA STATEMENT ON ORGAN AND TISSUE DONATION

Dimensions of relevance

Overview of Perspectives on Causal Inference: Campbell and Rubin. Stephen G. West Arizona State University Freie Universität Berlin, Germany

Autonomy as a Positive Value Some Conceptual Prerequisites Niklas Juth Dept. of Philosophy, Göteborg University

THE DETERRENT EFFECTS OF CALIFORNIA S PROPOSITION 8: WEIGHING THE EVIDENCE

Causality and counterfactual analysis

RESEARCH METHODS. Winfred, research methods,

Geographic Data Science - Lecture IX

Causality and Treatment Effects

Causal Inference Course Syllabus

- Triangulation - Member checks - Peer review - Researcher identity statement

GE Standard V: Scientific Standard

How to Randomise? (Randomisation Design)

The Holistic Defense Toolkit

Epidemiologic Methods I & II Epidem 201AB Winter & Spring 2002

UNESCO EOLSS. This article deals with risk-defusing behavior. It is argued that this forms a central part in decision processes.

Careful with Causal Inference

Methods of Randomization Lupe Bedoya. Development Impact Evaluation Field Coordinator Training Washington, DC April 22-25, 2013

UNCORRECTED PROOF. Daniel Steel / Across the Boundaries 01-DanielSteel-chap01 Page Proof page :48pm ACROSS THE BOUNDARIES

Methodological requirements for realworld cost-effectiveness assessment

You must answer question 1.

Mixed Methods Study Design

Glossary From Running Randomized Evaluations: A Practical Guide, by Rachel Glennerster and Kudzai Takavarasha

About Reading Scientific Studies

Study Designs in Epidemiology

Problem Set 5 ECN 140 Econometrics Professor Oscar Jorda. DUE: June 6, Name

An Experimental Investigation of Self-Serving Biases in an Auditing Trust Game: The Effect of Group Affiliation: Discussion

Clicker quiz: Should the cocaine trade be legalized? (either answer will tell us if you are here or not) 1. yes 2. no

Why randomize? Rohini Pande Harvard University and J-PAL.

Speaker Notes: Qualitative Comparative Analysis (QCA) in Implementation Studies

Transcription:

Measuring mpact Conceptual ssues in Program and Policy Evaluation Daniel L. Millimet Southern Methodist University 23 May 2013 DL Millimet (SMU) Measuring mpact May 2013 1 / 25

ntroduction Primary concern to economists, policymakers, organizations, etc. is to measures the causal e ect of a program or policy on outcomes of importance Statistical and econometric literature analyzing causation has seen tremendous growth over the past several decades The more we learn, the more we realize how complex the world can be and how di cult measuring impacts can be Our objective is to provide 1 A non-technical overview of issues and concepts that arise when seeking to measure causal impacts 2 A brief introduction to some of the methods used by researchers By the end, you will hopefully agree with Mark Twain: Education: the path from cocky ignorance to miserable uncertainty. DL Millimet (SMU) Measuring mpact May 2013 2 / 25

Outline 1 De nition of causation 2 Types of questions one may ask 3 Types of data 4 Data issues 5 Proper interpretation of statistical analyses 6 Roadmap to remaining sessions DL Millimet (SMU) Measuring mpact May 2013 3 / 25

Causation Philosophy of causality Rich literature in analytic philosophy on causality Two main approaches to de ning causality: Regularity approaches: Hume: We may de ne a cause to be an object followed by another, and where all the objects, similar to the rst, are followed by objects similar to the second. (from An Enquiry Concerning Human Understanding, section V) Counterfactual approaches: Hume: Or, in other words, where, if the rst object had not been, the second never had existed. (from An Enquiry Concerning Human Understanding, section V) DL Millimet (SMU) Measuring mpact May 2013 4 / 25

Regularity approach: a minimal conjunction between two objects Suppes: a probabilistic association between the two objects, which cannot be explained away by other factors dea behind Granger causality undamental notion underlying Granger causality is whether one object helps predict the occurrence of another Prediction, association, and correlation are not what most intend when speaking of causation While predictive ability answers questions that may be of interest, it should not be used to make policy decisions DL Millimet (SMU) Measuring mpact May 2013 5 / 25

Correlation (or predictive ability) and causation are not synonymous Examples abound where correlation ; causation Example: SBP and childhood obesity Examples abound where a lack of correlation ; a lack of causation Example: California s class size reduction policy DL Millimet (SMU) Measuring mpact May 2013 6 / 25

Two reasons why associations may not be indicative of causation 1 Confounding factors ) correlation is driven by other factors SBP & childhood obesity: family background California s class size reduction policy: teacher quality 2 Reverse causation ) correlation is driven by causation in the other direction Public housing and crime? Marital wage premium? Children and female labor supply? DL Millimet (SMU) Measuring mpact May 2013 7 / 25

Counterfactual approach: imagining a range of possible worlds n my view, this view is most consistent with what most intend when speaking of causation Most relevant for evaluating programs and policies Basic idea is to imagine alternative worlds where one (and only one) object is changed and to assess the di erences ) Sliding Doors Dominant view for measuring impacts in microeconomics today Greiner & Rubin (2011): or analysts from a variety of elds, the intensely practical goal of causal inference is to discover what would happen if we changed the world in some way. DL Millimet (SMU) Measuring mpact May 2013 8 / 25

Typically referred to as the Rubin Causal Model (Neyman 1923, 1935; isher 1935; Roy 1951; Quandt 1972, 1988; Rubin 1974) Crucial underpinning is the notion of potential outcomes DL Millimet (SMU) Measuring mpact May 2013 9 / 25

Potential outcomes refer to the outcome that would be realized under di erent states of nature Example: A sick individual may receive either Treatment 0 or 1. The outcome is either Recovery or Death. Thus, there are two possible states of nature (Treatment 0 or 1) and there is an outcome that would be realized in each state of nature. Under the counterfactual approach, the causal e ect of Treatment 1 relative to Treatment 0 would be the di erence in outcomes across these two states of nature DL Millimet (SMU) Measuring mpact May 2013 10 / 25

ormally, Let D = 0, 1 indicate the treatment received by the individual Let Y (0) indicate the outcome (Recovery, Death) the individual would experience if s/he receives Treatment 0 (D = 0) Let Y (1) indicate the outcome (Recovery, Death) the individual would experience if s/he receives Treatment 1 (D = 1) Y (0), Y (1) are potential outcomes as only one will actually be realized and observed in the world The causal e ect of Treatment 1 relative to Treatment 0 on the individual is given by τ = Y (1) Y (0) which is the di erence in outcomes in two alternative, but plausible, worlds DL Millimet (SMU) Measuring mpact May 2013 11 / 25

The counterfactual approach is a very simple yet powerful framework in which to think about causation mmediately leads to three salient points 1 Causal impacts of a treatment/intervention/program/policy are only de ned with respect to a well-de ned alternative Typically the alternative is the absence of treatment Not always obvious and must be made explicit 2 Causal impacts are individual-speci c Each individual potentially has he or her own potential outcomes and hence treatment e ect Referred to as constant vs. heterogeneous treatment e ects Has important implications for thinking about how to interpret the results of data analyses 3 Only one state of nature is actually realized at a point in time We can observe at most one potential outcome for any individual, remaining are missing The causal e ect of a treatment is not observable for any individual Estimating causal e ects must overcome this missing data problem To do so, requires assumptions and these assumptions must be credible DL Millimet (SMU) Measuring mpact May 2013 12 / 25

Types of Questions One Might Ask Granger Causality Granger causality answers questions concerning the ability of one object to predict or forecast another The underlying reason why it is a good predictor is not of (primary) importance However, often we do not care about the why and predictive ability is su cient or example: 1 State governments forecasting welfare caseloads or new applications for unemployment insurance or new applications for public housing 2 ood banks forecasting demand for food 3 Non-pro ts forecasting charitable donations 4 Police forecasting 911 calls at di erent times DL Millimet (SMU) Measuring mpact May 2013 13 / 25

Types of Questions One Might Ask Causal E ects Counterfactual approach answers questions concerning the expected e ect of a treatment (relative to a well-de ned alternative) Understanding the mechanisms behind why a causal e ect may arise is important, but requires additional analysis Examples: 1 Expected e ect of a job training program on employment and earnings 2 Expected e ect of a weekend food program on child nutrition and school achievement 3 Expected e ect of di erent nancial literacy training schemes (class, one-on-one, online, none) on decisionmaking 4 Expected e ect of di erent incentive schemes on charitable donations (e.g., matching, ra e, gift) Examples 1, 2 have binary treatments ) alternative is the absence of treatment Examples 3, 4 have non-binary treatments ) what is the alternative? DL Millimet (SMU) Measuring mpact May 2013 14 / 25

undamental ssues in Measurement of mpacts Estimation of causal e ects is di cult due to missing counterfactual problem With a binary treatment, either Y (1) or Y (0) is observed for any given individual The unobserved potential outcome is the missing counterfactual Holland (1986) refers to this as the fundamental problem of causal inference While the missing counterfactual makes life di cult, it is not impossible The implication is that we must estimate the missing counterfactual Such estimates are only as valid as the assumptions that underlie them and the data used to derive the estimates Discussion of the assumptions and statistical methods will be the focus of the later sections or the rest of this section, focus on understanding data DL Millimet (SMU) Measuring mpact May 2013 15 / 25

undamental ssues in Measurement of mpacts Types of Data Two types of data are employed to estimate causal e ects 1 Experimental ( eld or lab) 2 Non-experimental or observational Experimental data refers to data collected by researcher after randomizing treatments across subjects (as opposed to allowing subjects to self-select into the treatment or control groups) Observational data refers to typical survey data collected by researchers after subjects self-select into di erent treatments While experimental data may be preferable often referred to as the gold standard experiments are often di cult 1 Expensive 2 Unethical 3 Politically unpopular DL Millimet (SMU) Measuring mpact May 2013 16 / 25

undamental ssues in Measurement of mpacts Data ssues Regardless of the data type (experimental or observational), several issues can arise that impact data quality or a ect the interpretation of statistical results These include 1 Sample size 2 Sample selection 3 Sample attrition 4 Measurement issues DL Millimet (SMU) Measuring mpact May 2013 17 / 25

Sample size Refers to the number of subjects used in the analysis Sample size a ects the precision of estimates of causal e ects Precision of estimates is not always discussed outside of academic circles DL Millimet (SMU) Measuring mpact May 2013 18 / 25

Sample selection Refers to how the subjects used in the analysis are chosen Sample selection a ects the representativeness of the sample A ects the interpretation of the results by a ecting generalizability DL Millimet (SMU) Measuring mpact May 2013 19 / 25

Sample attrition When analysis entails following subjects over time, nal sample may be nonrepresentative even if the initial sample is representative Need to think about whether attrition is random or not May occur for many reasons 1 Subjects voluntarily decide to leave study, move away, etc. 2 Subjects involuntarily decide to leave study (e.g., death, jail, etc.) DL Millimet (SMU) Measuring mpact May 2013 20 / 25

Measurement issues De nitions of variables, particularly outcomes Short-run vs. long-run Comprehensive vs. speci c Measurement error Self-reported vs. veri ed DL Millimet (SMU) Measuring mpact May 2013 21 / 25

undamental ssues in Measurement of mpacts nterpretation nternal vs. external validity nternal validity refers to whether an estimate of the causal e ects is valid for the sample used to obtain the estimate External validity refers to whether an estimate of the causal e ects is generalizable to individuals outside the sample Treatment e ect parameters When causal e ects are heterogeneous, several parameters may be of interest 1 Average Treatment E ect (ATE) ) expected treatment e ect for an individual chosen at random from the population 2 Average Treatment E ect on the Treated (ATT) ) expected treatment e ect for an individual chosen at random from the treated population 3 Average Treatment E ect on the Untreated (ATU) ) expected treatment e ect for an individual chosen at random from the untreated population 4 Distributional Treatment E ects Note: These answer di erent policy questions! DL Millimet (SMU) Measuring mpact May 2013 22 / 25

undamental ssues in Measurement of mpacts nterpretation or any treatment e ect parameter, there is an important di erence between ceteris paribus e ects and policy e ects Ceteris paribus ( all else constant ) e ects answer questions like: What is the causal e ect of the treatment on an individual holding all else xed? Policy e ects answer questions like: What is the causal e ect of the treatment on an individual? Answers are not always the same, and both are important Example 1: California s class size reduction policy Example 2: nfra-marginal nutrition assistance DL Millimet (SMU) Measuring mpact May 2013 23 / 25

Conclusion Critical questions to ask of causal statements 1 What is the counterfactual? What is the alternative to the treatment? 2 What is the nature of the data? Randomization or observational? 3 What is the proper interpretation of the results? To whom do the results apply? What is the outcome? 4 Are the results internally valid? What assumptions are made to circumvent the missing counterfactual? 5 Are the results externally valid? Sample selection? DL Millimet (SMU) Measuring mpact May 2013 24 / 25

Roadmap Assumptions used to circumvent the problem of the missing counterfactual 1 Randomization ) Tim, Danila 2 Observational methods ) me orecasting and prediction ) Tom 1 Data mining 2 Big data DL Millimet (SMU) Measuring mpact May 2013 25 / 25