Program Evaluations and Randomization. Lecture 5 HSE, Dagmara Celik Katreniak

Similar documents
Threats and Analysis. Shawn Cole. Harvard Business School

Threats and Analysis. Bruno Crépon J-PAL

Randomization as a Tool for Development Economists. Esther Duflo Sendhil Mullainathan BREAD-BIRS Summer school

What can go wrong.and how to fix it!

Group Work Instructions

Glossary From Running Randomized Evaluations: A Practical Guide, by Rachel Glennerster and Kudzai Takavarasha

How to Randomise? (Randomisation Design)

NBER TECHNICAL WORKING PAPER SERIES USING RAMDOMIZATION IN DEVELOPMENT ECONOMICS RESEARCH: A TOOLKIT. Esther Duflo Rachel Glennerster Michael Kremer

Complier Average Causal Effect (CACE)

DISCUSSION PAPER SERIES. No USING RANDOMIZATION IN DEVELOPMENT ECONOMICS RESEARCH: A TOOLKIT

Evaluating Social Programs Course: Evaluation Glossary (Sources: 3ie and The World Bank)

MIT LIBRARIES. MiltmilWv I! ''.'''' ;i ill:i!.(;!!:i i'.ir-il:>!\! ItllillPlli. ' iilpbilisil HB31 .M415. no AHHH

EMPIRICAL STRATEGIES IN LABOUR ECONOMICS

Instrumental Variables I (cont.)

What is: regression discontinuity design?

THREATS TO VALIDITY. Presentation by: Raul Sanchez de la Sierra

G , G , G MHRN

Applied Econometrics for Development: Experiments II

Brief introduction to instrumental variables. IV Workshop, Bristol, Miguel A. Hernán Department of Epidemiology Harvard School of Public Health

Instrumental Variables Estimation: An Introduction

Planning Sample Size for Randomized Evaluations.

Version No. 7 Date: July Please send comments or suggestions on this glossary to

Does AIDS Treatment Stimulate Negative Behavioral Response? A Field Experiment in South Africa

Demystifying causal inference in randomised trials

Evaluating the Impact of Health Programmes

Statistical Power Sampling Design and sample Size Determination

Supplementary Appendix

Methods for Addressing Selection Bias in Observational Studies

Title:Bounding the Per-Protocol Effect in Randomized Trials: An Application to Colorectal Cancer Screening

The Stable Unit Treatment Value Assumption (SUTVA) and Its Implications for Social Science RCTs

A NEW TRIAL DESIGN FULLY INTEGRATING BIOMARKER INFORMATION FOR THE EVALUATION OF TREATMENT-EFFECT MECHANISMS IN PERSONALISED MEDICINE

Chapter 3. Producing Data

CHL 5225 H Advanced Statistical Methods for Clinical Trials. CHL 5225 H The Language of Clinical Trials

Section 1: The Science of Clinical Investigation

Randomized Experiments with Noncompliance. David Madigan

1. Introduction Consider a government contemplating the implementation of a training (or other social assistance) program. The decision to implement t

Greg Pond Ph.D., P.Stat.

SAMPLING AND SAMPLE SIZE

Economics 270c. Development Economics. Lecture 9 March 14, 2007

Introduction to Program Evaluation

14.11 Lecture 2: Public Health Messages How to Convince Others to Adopt Healthy Behaviors?

Workshop on Experiments in Political Economy May Columbia Center for the Study of Development Strategies & the Harriman Institute

Causal Validity Considerations for Including High Quality Non-Experimental Evidence in Systematic Reviews

Trial designs fully integrating biomarker information for the evaluation of treatment-effect mechanisms in stratified medicine

Lecture II: Difference in Difference. Causality is difficult to Show from cross

Introduction to Observational Studies. Jane Pinelis

Analysis of TB prevalence surveys

TRACER STUDIES ASSESSMENTS AND EVALUATIONS

Methods of Randomization Lupe Bedoya. Development Impact Evaluation Field Coordinator Training Washington, DC April 22-25, 2013

AP Statistics Exam Review: Strand 2: Sampling and Experimentation Date:

GUIDE 4: COUNSELING THE UNEMPLOYED

Causal Interaction and External Validity: Obstacles to the Policy Relevance of Randomized Evaluations

Empirical Strategies 2012: IV and RD Go to School

Beyond the intention-to treat effect: Per-protocol effects in randomized trials

TRANSLATING RESEARCH INTO ACTION. Why randomize? Dan Levy. Harvard Kennedy School

Higher Psychology RESEARCH REVISION

Dichotomizing partial compliance and increased participant burden in factorial designs: the performance of four noncompliance methods

Reflection Questions for Math 58B

Incentives versus sorting in tournaments: Evidence from a field experiment forthcoming Journal of Labor Economics

Estimating average treatment effects from observational data using teffects

Technical Track Session IV Instrumental Variables

Sampling. (James Madison University) January 9, / 13

Empirical Strategies

Lecture Slides. Elementary Statistics Eleventh Edition. by Mario F. Triola. and the Triola Statistics Series 1.1-1

Impact Evaluation Methods: Why Randomize? Meghan Mahoney Policy Manager, J-PAL Global

Issues of Non-Compliance and Their Effect on Validity in Field Experiments - A case study of the field experiment Taxis and Contracts

14.74 Foundations of Development Policy Spring 2009

Practitioner s Guide To Stratified Random Sampling: Part 1

Prediction, Causation, and Interpretation in Social Science. Duncan Watts Microsoft Research

Vocabulary. Bias. Blinding. Block. Cluster sample

University student sexual assault and sexual harassment survey. Notes on reading institutional-level data

Quasi-experimental analysis Notes for "Structural modelling".

Randomized experiments are increasingly used in development economics.

Risk of bias assessment for. statistical methods

Empirical Strategies

Estimating drug effects in the presence of placebo response: Causal inference using growth mixture modeling

Regression Discontinuity Designs: An Approach to Causal Inference Using Observational Data

Randomized Controlled Trial

Chapter 13 Summary Experiments and Observational Studies

Title: Intention-to-treat and transparency of related practices in randomized, controlled trials of anti-infectives

Impact Evaluation Toolbox

Inferential Statistics

School Autonomy and Regression Discontinuity Imbalance

Introduction to NHANES and NAMCS

Chapter 13. Experiments and Observational Studies. Copyright 2012, 2008, 2005 Pearson Education, Inc.

The Fallacy of Taking Random Supplements

WORKING PAPER SERIES

TRANSLATING RESEARCH INTO ACTION

Applied Quantitative Methods II

Welcome to this series focused on sources of bias in epidemiologic studies. In this first module, I will provide a general overview of bias.

Why do Psychologists Perform Research?

Economics 270c Graduate Development Economics. Professor Ted Miguel Department of Economics University of California, Berkeley

EPI 200C Final, June 4 th, 2009 This exam includes 24 questions.

This article appeared in a journal published by Elsevier. The attached copy is furnished to the author for internal non-commercial research and

MATH-134. Experimental Design

Public Policy & Evidence:

Causal Effect Heterogeneity

Confounding by indication developments in matching, and instrumental variable methods. Richard Grieve London School of Hygiene and Tropical Medicine

Transcription:

Program Evaluations and Randomization Lecture 5 HSE, 10.11.2014 Dagmara Celik Katreniak

Overview Treatment Effect under perfect conditions Average Treatment Effect Treatment effect under imperfect conditions Intention To Treat (ITT) Treatment On the Treated (TOT) Local Average Treatment Effect (LATE) Other Externality Effect

Average Treatment Effect (ATE) SUTVA assumption Non-interference Being in the treatment or control group does not influence the outcome of anyone No variation in treatment All subjects are subject to the same treatment Unconfoundedness/Ignorability Assignment to treatment not connected with outcome

Average Treatment Effect (ATE) Randomly assign subjects to T or C Calculate ATE ATE = measures the effect of treatment on a randomly selected person Randomization not conditional on observables Perfect compliance No attrition SUTVA and unconfoundedness

Example Blimpo (AEJ: Applied Microeconomics, 2014) All takeovers, no attrition Randomized at school level T1= Individual target group T2= Team target group T3= Team Tournament group C = Control group sssss i,s = β 0 + 3 k=1 β k T i,s k + ε i,s

Randomization balance

The conditional probability of selection Stratify your sample By stratas, randomize into T/C The probability to be picked to Treatment depends on observables (stratification) The allocation of T unequal across stratas Possible for example if the number of people in T and C not equal but fixed number of awards in T

Stratified sample Checking randomization balance Stratum level But then no overall balance check Problem of too many tests at once Possibly low power cause of small strata size Overall/general level But may not be balanced by stratas F-test by Firpo, Foguel and Borges Jales (conference paper) What if imbalances exist?

Example Voucher program in Colombia Students picked to schools by lottery Given number of winners in each city The ratio of winners differs by town Lottery random within each city conditional on whether households have an access to a phone Causal effect of the program on voucher s applicants with access to a telephone in surveyed cities (Angrist et al., 2001)

The conditional probability of selection Weighted average over all strata Weight = proportion of treated subjects within strata Regression Control for all strata variables Include all dummies plus their interactions Be careful about the degrees of freedom

The conditional probability of selection Conditional Randomization implies: E Y i C X, T E Y i C X, C = 0 E x E Y T i X, T E Y C i X, T = E Y T i x, T E Y C i x, C P X = x T dd If discrete comparison of means using proportion of treatment within cells as weights

(Im)Perfect Compliance Subjects from Treatment/Control group without/with treatment Spillover effects Low take up rate Switching of subjects during the experiment What is the question of interest? The effect of the intervention itself? The effect of a diet rich in iron The effect of the instrument? Common for policy relevant research Do we need all participants to participate or we want to see the effect on those who choose to participate?

(Im)Perfect Compliance Perfect compliance of interest Prepare design accordingly Example: Thomas, Frankenberg, Friedman, Habicht, & Al (2003) on iron supplementation Partial compliance of interest Esp. encouragement designs Duflo and Saez (2003) Partial compliance as the only option Deworming program (Miguel & Glewwe, 2004) Mostly because tracking not possible

TOT=? ITT=?

Intention to Treat Estimate (ITT) The effect of offered intervention Differ from the effect of the treatment in case of imperfect compliance Based on the initial division into treatment and control group Perfect compliance ITT=ATE

Average Treatment Effect on Treated Estimate (TOT) Y(T) = A PPPPPPPPPPP TTTTTTT T) + B (1 PPPPPPPPPPP TTTTTTT T)) Y C = A PPPPPPPPPPP TTTTTTT C) + B (1 PPPPPPPPPPP TTTTTTT C)) TTT = Y T Y C PPPP TTTTTTT T) PPPP TTTTTTT C)

Scenario 1: Example Researcher chooses 100 people into treatment group (he gives them $100) and 100 people into control group (gives them 0) Collects information on rewards All people in T got money No person from C got money ITT=? TOT=? ITT=TOT=$100

Example Scenario 2: Researcher tells 100 students if they come to his office, he gives them $100 (= treatment), he says nothing to the control group Cca 20% of students do not trust him/lazy to come Collects information on rewards On average treatment group received $80, control $0 ITT=?, TOT=? ITT=$80 TOT=$100

Example Scenario 3: Researcher tells subjects to come at midnight to his office to pick up $100 (treatment), he says nothing to the control group Only 50% of treatment group show up Some 10% of control group students show up Collects information on rewards ITT=?, TOT=? ITT=$40 TOT=$100

Example Scenario 4: Researcher gives $25, $50, $75, $100 to four groups of students (each 1/4 in size), $0 to control 30% of treated pick up money, no control Collects information on rewards ITT=?, TOT=? ITT=$21 If everyone came, TOT=$62.5 When 30% came, TOT=$21/0.3=$70

ITT versus ATE Two identification assumptions (Imbens and Angrist, 1994): Independence Outcomes are not directly affected by the instrument (Duflo, Glennerster & Kremer, 2007, p.52) Y i C, Y i T, T i 1, T i 0 Z Monotonicity The monotonicity assumption requires that the instrument makes every person either weakly more or less likely to actually participate in the treatment (Duflo, Glennerster & Kremer, 2007, p.52) T i 1 T i 0 fff aaa i or T i 1 T i 0 fff aaa i

ITT versus ATE Wald estimate local average treatment effect (LATE) on compliers, i.e. on those whose treatment status was affected by the instrument IV Estimate of β (using Z as an instrument) can be interpreted as the ATE for a well-defined group of individuals, namely those who are induced by the instrument Z to take advantage of the treatment (Duflo, Glennerster & Kremer, 2007, p.53)

ITT versus ATE β W = E Y i Z i =1 E Y i Z i =0 E T i Z i =1 E T i Z i =0 β W = E Y i T Y i C T i 1 T i 0 = 1

IV estimate (LATE) No person from control group got treatment LATE = TOT All people from treatment group got treatment Oversubscription design Interpretation! Causal effect for compliers Not necessarily the effect for the entire population

Example Voucher program in Colombia PACES = governmental program, scholarships for private secondary schools assigned by lottery Lottery winners and losers picked from PACES program applicants Stratification by localities, conditional to access to a phone What effects do they measure? The effect of winning the lottery The effect of scholarship

Angrist et al. (2002)

Estimates of the effects The effect of winning the lottery on scholarship use, school choice and schooling 6-7% more likely to begin the 6 th grade 15-16% more likely to be in private school Decision to be in private school seems to be sensitive to price, decision to attend the school not Lower repetition rate for winners Higher likelihood to be in 8 th grade for winners Angrist et al. (2002)

Estimates of the effects The effect of winning the lottery on test scores Subsample of children invited by phone to be tested + by letter More than one test opportunity if failed Refreshments, after each test other rewards picked by lottery (e.g.bicycle) Travel costs covered NO SIGNIFICANT relationship between being tested and voucher status Angrist et al. (2002)

Angrist et al. (2002)

Estimates of the effects The effect of receiving scholarship Using voucher win/loss status as IV Initial randomization = IV versus Actual treatment 6% (24%) of lottery users used voucher 90% of winners used scholarship Two stage least squares estimation The 2SLS estimates based on this difference are necessarily larger than the reduced form effects of winning the lottery since winning the lottery is only imperfectly correlated with receiving a scholarship. Angrist et al. (2002)

Spillovers How would you measure them? Vary the exposure to a treatment Duflo and Saez (2003) Use variation in exposure across groups that arises from randomization Duflo, Kremer and Robinson (2006) Random assignment to peer groups

Attrition May or may not be a problem Problem if there is correlation between attrition and treatment Biases results Random attrition Lowers statistical power Non-random attrition No independence of outcomes Try to limit ex-ante

Attrition Parametric techniques Heckman correction Survival model and attrition Non-parametric techniques Manski-Lee bounds Matching matching lottery winners and losers Angrist et al (2006) Imputation methods

Case Study Please, work on the case study

Summary Randomization Experimental Designs Sample size and power calculations Budget, CBA and CEA Data Analysis under perfect and imperfect randomization *** Further use of randomization Topics Presentations