Conditional Average Treatment Effects

Similar documents
Regression Discontinuity Design

Chapter 10 Quasi-Experimental and Single-Case Designs

GUIDE 4: COUNSELING THE UNEMPLOYED

Causal inference nuts and bolts

Research Design. Beyond Randomized Control Trials. Jody Worley, Ph.D. College of Arts & Sciences Human Relations

Causal inference: Nuts and bolts

EXPERIMENTAL RESEARCH DESIGNS

The Regression-Discontinuity Design

Endogeneity is a fancy word for a simple problem. So fancy, in fact, that the Microsoft Word spell-checker does not recognize it.

Quasi-Experimental and Single Case Experimental Designs. Experimental Designs vs. Quasi-Experimental Designs

Experimental and Quasi-Experimental designs

Overview of Perspectives on Causal Inference: Campbell and Rubin. Stephen G. West Arizona State University Freie Universität Berlin, Germany

Formative and Impact Evaluation. Formative Evaluation. Impact Evaluation

In this chapter we discuss validity issues for quantitative research and for qualitative research.

Presentation Preparation

8 people got 100 (I rounded down to 100 for those who went over) Two had perfect exams

Lecture 4: Research Approaches

Chapter 11: Experiments and Observational Studies p 318

Georgina Salas. Topics EDCI Intro to Research Dr. A.J. Herrera

CHAPTER LEARNING OUTCOMES

Experimental Research. Types of Group Comparison Research. Types of Group Comparison Research. Stephen E. Brock, Ph.D.

One slide on research question Literature review: structured; holes you will fill in Your research design

School-Based Mental Health: Supports, Strategies and Recommendations

Validity and Quantitative Research. What is Validity? What is Validity Cont. RCS /16/04

11/2/2017. Often, we cannot manipulate a variable of interest

Research Approach & Design. Awatif Alam MBBS, Msc (Toronto),ABCM Professor Community Medicine Vice Provost Girls Section

Instrumental Variables Estimation: An Introduction

Chapter 9 Experimental Research (Reminder: Don t forget to utilize the concept maps and study questions as you study this and the other chapters.

Previous Example. New. Tradition

QUASI EXPERIMENTAL DESIGN

PSY 250. Designs 8/11/2015. Nonexperimental vs. Quasi- Experimental Strategies. Nonequivalent Between Group Designs

Lecture II: Difference in Difference. Causality is difficult to Show from cross

Study Design. Svetlana Yampolskaya, Ph.D. Summer 2013

Basic Flow diagram of Life Coaching

VALIDITY OF QUANTITATIVE RESEARCH

HPS301 Exam Notes- Contents

EXPERIMENTAL AND EX POST FACTO DESIGNS M O N A R A H I M I

CASE STUDY 2: VOCATIONAL TRAINING FOR DISADVANTAGED YOUTH

Emotional Intelligence

At Merck, ethics are our first priority. We are looking for scientists, managers, administrators, and workers who share this priority.

STAFF REPORT City of Lancaster NB 2

Chapter 3 Introducing a Control Variable (Multivariate Analysis)

10/23/2017. Often, we cannot manipulate a variable of interest

STATISTICAL CONCLUSION VALIDITY

Chapter 11. Experimental Design: One-Way Independent Samples Design

SAMPLING AND SAMPLE SIZE

Subjective Well-Being and Adjustment

Spring Survey 2014 Report - ADULTS

OBSERVATION METHODS: EXPERIMENTS

The Logic of Data Analysis Using Statistical Techniques M. E. Swisher, 2016

investigate. educate. inform.

Asking and answering research questions. What s it about?

Research Methods II, Spring Term Logic of repeated measures designs

A teaching presentation to help general psychology students overcome the common misconception that correlation equals causation

Lecture II: Difference in Difference and Regression Discontinuity

Chapter 13 Summary Experiments and Observational Studies

UNIT. Experiments and the Common Cold. Biology. Unit Description. Unit Requirements

Chapter 13. Experiments and Observational Studies. Copyright 2012, 2008, 2005 Pearson Education, Inc.

Levels of Evaluation. Question. Example of Method Does the program. Focus groups with make sense? target population. implemented as intended?

QUASI-EXPERIMENTAL HEALTH SERVICE EVALUATION COMPASS 1 APRIL 2016

We re going to talk about a class of designs which generally are known as quasiexperiments. They re very important in evaluating educational programs

Module 3 - How Pain Affects You

If you would like to find out more about this service:

Randomized Evaluations

(CORRELATIONAL DESIGN AND COMPARATIVE DESIGN)

Your Voices Amplified

FGM, FORCED MARRIAGE AND HONOUR-BASED ABUSE THE LEGAL FRAMEWORK

Chapter Eight: Multivariate Analysis

TECH 646 Analysis of Research in Industry and Technology. Experiments

Educational Research. S.Shafiee. Expert PDF Trial

The first step to managing stress is to understand its nature

Choose an approach for your research problem

5: Family, children and friends

M.O.D.E.R.N. Voice-Hearer

The Advocacy Charter Action for Advocacy

2013/4/28. Experimental Research

Assessment of the Child s Experience of Stuttering (ACES) (DRAFT VERSION September 27, 2006)

Internal Validity and Experimental Design

Chapter 7. Marketing Experimental Research. Business Research Methods Verónica Rosendo Ríos Enrique Pérez del Campo Marketing Research

15.301/310, Managerial Psychology Prof. Dan Ariely Recitation 8: T test and ANOVA

Topic #6. Quasi-experimental designs are research studies in which participants are selected for different conditions from pre-existing groups.

Head Up, Bounce Back

DEPENDENT VARIABLE. TEST UNITS - subjects or entities whose response to the experimental treatment are measured or observed. Dependent Variable

Self Esteem and Purchasing Behavior Part Two.

Human Experimentation

For young people living with someone s excessive drinking, drug use or addiction

Thinking Like a Researcher

COMMUNICATION ISSUES IN PALLIATIVE CARE

PM12 Validity P R O F. D R. P A S Q U A L E R U G G I E R O D E P A R T M E N T O F B U S I N E S S A N D L A W

For tonight s talk, I could throw a bunch of numbers at you, like these.

Responding to HIV in the Workplace

The Helping Orientations Inventory Open Materials 1

Werkstraffen: succes verzekerd?

M6728. Goals. Depression Scores. Research Designs

Reducing Liability by Hiring The Right Employ lo ee What Type of Screenings to Conduct Why Conduct Background Checks?

Communication Research Practice Questions

Lying Liars and The Lies They Lie About.

Matching: Observational research

THEORY U. A Way to Change Services for People with Intellectual Disabilities John O Brien Illustrations by Ester Ortega

A FRAMEWORK FOR EMPOWERMENT

Transcription:

CAUSALITY Again

Conditional Average Treatment Effects Average treatment effect (ATE) Average treatment effect for the treated (ATT) Average treatment effect for the controls (ATC) ATT = average treatment effect for those that typically are (choose to be) treated based on counterfactual comparison. In a well designed experiment ATE should (over many replications) be the same for those that just happen to be randomized to the treatment and those that just happen to be randomized to the controls. Q. Is there any reason to expect this in observational data?

Sources of bias in the estimation of the ATE Details of estimation from sample data skipped here (see Morgan & Winship, 2007: 44-46) Turns out that: The true average treatment effect A baseline bias (Where π is the proportion receiving the treatment) A differential treatment effect bias So, much depends on whether the second and third component can be either shown or assumed to = 0.

Numerical example discussed in Morgan & Winship pp 47 Table 2.3: An Example of the Bias of the the Naïve Estimator of the ATE Group E[Y 1.] E[Y 0.] Treatment group (D=1) 10 6 Control group (D=0) 8 5 Effect of college degree on a labour market outcome. Assume π = 0.3 Average PO under treatment for treated = 10 and average PO under control for controls = 5. This is what is observed. Q. Is the ATE = 5? Consider what would have happened in the control state. Those in D=1 would have done better, 6 versus 5. Baseline difference. Consider what would have happened in the treatment state. Those in D=0 would have done worse, 8 versus 10. ATT = 4; ATC = 3 therefore ATE = 0.3(10-6) + 0.7(8-5) = 3.3 NOT 5!

Example: Church Schools We observe (hypothetically) that children attending religious schools do better in exams Why? 1) Religious schools do a better job of teaching kids (there is a causal impact). 2) Kids that entered religious schools were different (smarter, came from more advantaged homes) right from the start. 3) Kids that actually entered religious schools flourish more in religious schools than would the kids whose parents (actually) chose a secular school for them. What we want to know about is 1). But 2) & 3) get in the way. 2) is a problem of heterogeneity; 3) could be a problem of self-selection on the basis of the anticipated outcome (parents select schools they think will suit their kids).

Example 2 Random assignment to treatment and control solves (on average) 2) and 3) But for problems like this randomization is, except under special circumstances, likely to be difficult (or impossible). Many of the problems sociologists are interested in share these characteristics: Treatment and control groups are heterogeneous. Normal move is to try to deal with heterogeneity by : 1) Matching on observables 2) Conditioning on observables (ie introducing relevant control variables) 3) Or both. No guarantee this will work! Units self-select themselves into or are selected into treatment and control on the basis of anticipated outcomes. If we notice that kids who read for an hour a day for pleasure do better in school, would we expect the same result if we forced reading on kids that wouldn t normally choose it as a leisure pursuit?

These thoughts also have implications for what it makes sense to target Causal effect of treatment on the treated? Effect of Catholic school on those kids that would choose a Catholic school Effect of a training programme on those that would choose to take it Average causal effect? Effect of Catholic school on Catholics and others? Causal effect of treatment on those that would choose the control? Effect of Catholic school on Northern Ireland Protestants Effect of marriage on those who prefer cohabitation

Internal Validity Internally valid designs are resistant to rival explanations in terms of factor(s) other than the applied treatment is (are) the cause(s) of any observed effect Good experimental design seeks to maximize internal validity against a number of particular threats, in other words weaknesses of design that call into doubt the unique attribution of changes in values of the response to the experimental manipulation.

Internal Validity Threats to Validity Maturation Selection History Testing Instrumentation Mortality Regression to the mean

External Validity Generalisability beyond the experimental setting The Hawthorne Effect The extent to which the effects discovered in the experiment can be extended to the target population.

Pretest-Postest Nonequivalent Control Group Design GROUP 1 Y 1 X Y 2 GROUP 2 Y 3 Y 4 = non random allocation to groups

Participative decision making A large management consultancy organizes work on the basis of project teams with a team leader Senior partners are worried about morale level (as reflected in absenteeism rate) Team leaders are allowed to adopt (if they wish) more participatory ways of making decisions Some do and others don t Do participatory teams do better than others?

PDM Outcome 1 High Y 3 Y 4 Project teams without PDM Y 1 Absenteeism Project teams with PDM Y 2 Low Pretest January 2001 Posttest January 2002

PDM Outcome 1 continued PDM adopters already have lower absenteeism rates Difference becomes bigger after adoption of PDM But PDM adopters might already be on a downward trend (perhaps team leaders are more easy going) aka selection by maturation threat Can t be ruled out without more pretest observations Conclusion depends on the plausibility of the selection by maturation threat in this particular case

PDM Outcome 2 Imagine a different set up PDM is imposed by the senior partners on the project teams with the highest absenteeism rates

PDM Outcome 2 continued High Y 1 Absenteeism Y 3 Y 4 Project teams without PDM Low Y 2 Project teams with PDM Pretest January 2001 Posttest January 2002

PDM Outcome 2 continued Genuine treatment effect looks more plausible Why should differential maturation lead to a crossover? Why should regression to the mean lead to crossover? Effect of PDM still not proven but case looks stronger

Regression-Discontinuity Design A variant of the PPNCGD If selection into treatment is based in a known way on pretest score then R-D design possible Say top 10% of sales force are given a bonus over and above their commision Does it affect their performance?

May 2002 sales figures Regression-Discontinuity Design continued Plausible effect No effect Regression to the mean? May 2001 sales figures Top 10%

Interrupted Time-Series Design If you can t control who is exposed to the treatment Try to control when observations are made GROUP 1 Y 1 Y 2 Y 3 Y 4 X Y 5 Y 6 Y 7 Y 8 Even spacing is nice Can rule out maturation by looking at trends in the pretests

Possible outcomes of the interrupted time series H H H Possible treatment effect Lagged treatment effect? No treatment effect L L L Y 1 Y 2 Y 3 Y 4 X Y 5 Y 6 Y 7 Y 8 Y 1 Y 2 Y 3 Y 4 X Y 5 Y 6 Y 7 Y 8 Y 1 Y 2 Y 3 Y 4 X Y 5 Y 6 Y 7 Y 8

If the process has no memory Apply treatment; Remove treatment; Apply again GROUP 1 Y 1 Y 2 Y 3 X Y 4 X Y 5 X Y 6 Y 7 Y 8 X Y 9 X Y 10 X Y 11 Is the effect reversible? Harder to take away a benefit than to introduce it!

Let the data speak for itself? Contract to empty parking meters in NYC given to Brinks Inc Brinks employees found guilty in criminal court of skimming NYC launched civil action against Brinks for negligence and breach of contract Contract given to CDC In first 10 months CDC collected $1millon more than in any 10 month period when Brinks held the contract

Revenue Evidence of theft? Monthly revenue per meter-day per month 1 0.95 CDC 0.9 Brinks Inc. 0.85 0 5 10 15 20 25 Month From Fairley, W. B. and Glenn, J.E. A question of theft in DeGroot, M. H., Fienberg, S. E. and Kadane, J. B., eds., Statistics and the Law. New York: Wiley, 1986.

Dollars per collection day in thousands Or just growth? Revenue over 36 months and smoothed trend 100 95 90 85 80 75 70 65 60 Brinks CDC 0 3 6 9 12 15 18 21 24 27 30 33 36 From Levin, B. Comment in DeGroot, M. H., Fienberg, S. E. and Kadane, J. B., eds., Statistics and the Law. New York: Wiley, 1986.