Causal Inference Course Syllabus

Similar documents
ICPSR Causal Inference in the Social Sciences. Course Syllabus

Empirical Strategies

Propensity Score Analysis Shenyang Guo, Ph.D.

Practical propensity score matching: a reply to Smith and Todd

Advanced Quantitative Analysis: Experiments and Causal Inference

EMPIRICAL STRATEGIES IN LABOUR ECONOMICS

Jake Bowers Wednesdays, 2-4pm 6648 Haven Hall ( ) CPS Phone is

Causal Validity Considerations for Including High Quality Non-Experimental Evidence in Systematic Reviews

POL 574: Quantitative Analysis IV

What is: regression discontinuity design?

Supplement 2. Use of Directed Acyclic Graphs (DAGs)

Using register data to estimate causal effects of interventions: An ex post synthetic control-group approach

Introduction to Observational Studies. Jane Pinelis

Overview of Perspectives on Causal Inference: Campbell and Rubin. Stephen G. West Arizona State University Freie Universität Berlin, Germany

Research Design. Beyond Randomized Control Trials. Jody Worley, Ph.D. College of Arts & Sciences Human Relations

Introduction to Applied Research in Economics Kamiljon T. Akramov, Ph.D. IFPRI, Washington, DC, USA

Empirical Strategies

The Logic of Causal Inference

Empirical Strategies 2012: IV and RD Go to School

Recent advances in non-experimental comparison group designs

Ec331: Research in Applied Economics Spring term, Panel Data: brief outlines

Causal Methods for Observational Data Amanda Stevenson, University of Texas at Austin Population Research Center, Austin, TX

PharmaSUG Paper HA-04 Two Roads Diverged in a Narrow Dataset...When Coarsened Exact Matching is More Appropriate than Propensity Score Matching

Causality and Statistical Learning 1. Andrew Gelman Feb 2010

Measuring Impact. Program and Policy Evaluation with Observational Data. Daniel L. Millimet. Southern Methodist University.

Estimating average treatment effects from observational data using teffects

Confounding by indication developments in matching, and instrumental variable methods. Richard Grieve London School of Hygiene and Tropical Medicine

Causal Thinking in the Twilight Zone

Lecture II: Difference in Difference and Regression Discontinuity

Propensity Score Methods for Estimating Causality in the Absence of Random Assignment: Applications for Child Care Policy Research

Version No. 7 Date: July Please send comments or suggestions on this glossary to

Syllabus.

Reassessing Quasi-Experiments: Policy Evaluation, Induction, and SUTVA. Tom Boesche

Title: New Perspectives on the Synthetic Control Method. Authors: Eli Ben-Michael, UC Berkeley,

Careful with Causal Inference

COMMENTS ON: UNDERSTANDING AND MISUNDERSTANDING RANDOMIZED CONTROLLED TRIALS BY C ARTWRIGHT AND DEATON GUIDO IMBENS - STANFORD UNIVERSITY

A Potential Outcomes View of Value-Added Assessment in Education

Measuring Impact. Conceptual Issues in Program and Policy Evaluation. Daniel L. Millimet. Southern Methodist University.

Causal Thinking in the Twilight Zone

Should Causality Be Defined in Terms of Experiments? University of Connecticut University of Colorado

1. INTRODUCTION. Lalonde estimates the impact of the National Supported Work (NSW) Demonstration, a labor

Causal Inference in Observational Settings

Effects of propensity score overlap on the estimates of treatment effects. Yating Zheng & Laura Stapleton

Introduction to Applied Research in Economics

Pros. University of Chicago and NORC at the University of Chicago, USA, and IZA, Germany

Propensity Score Matching with Limited Overlap. Abstract

Mastering Metrics: An Empirical Strategies Workshop. Master Joshway

Using ASPES (Analysis of Symmetrically- Predicted Endogenous Subgroups) to understand variation in program impacts. Laura R. Peck.

Regression Discontinuity Suppose assignment into treated state depends on a continuously distributed observable variable

Quasi-experimental analysis Notes for "Structural modelling".

Understanding Regression Discontinuity Designs As Observational Studies

Key questions when starting an econometric project (Angrist & Pischke, 2009):

A note on evaluating Supplemental Instruction

Complier Average Causal Effect (CACE)

Improving Causal Claims in Observational Research: An Investigation of Propensity Score Methods in Applied Educational Research

Statistical Causality

Lecture II: Difference in Difference. Causality is difficult to Show from cross

Propensity Score Analysis and Strategies for Its Application to Services Training Evaluation

Quantitative Methods. Lonnie Berger. Research Training Policy Practice

The Regression-Discontinuity Design

The Prevalence of HIV in Botswana

T. Kushnir & A. Gopnik (2005 ). Young children infer causal strength from probabilities and interventions. Psychological Science 16 (9):

Identifying Peer Influence Effects in Observational Social Network Data: An Evaluation of Propensity Score Methods

Introduction to Program Evaluation

Dylan Small Department of Statistics, Wharton School, University of Pennsylvania. Based on joint work with Paul Rosenbaum

Write your identification number on each paper and cover sheet (the number stated in the upper right hand corner on your exam cover).

Using directed acyclic graphs to guide analyses of neighbourhood health effects: an introduction

PARTIAL IDENTIFICATION OF PROBABILITY DISTRIBUTIONS. Charles F. Manski. Springer-Verlag, 2003

Causal Inference in Statistics and the Quantitative Sciences

Problem Set 3 ECN Econometrics Professor Oscar Jorda. Name. ESSAY. Write your answer in the space provided.

Early Release from Prison and Recidivism: A Regression Discontinuity Approach *

Marno Verbeek Erasmus University, the Netherlands. Cons. Pros

Predicting the efficacy of future training programs using past experiences at other locations

From Description to Causation

Do not copy, post, or distribute

Biostatistics II

Methods of Reducing Bias in Time Series Designs: A Within Study Comparison

P(HYP)=h. P(CON HYP)=p REL. P(CON HYP)=q REP

EXPERIMENTAL RESEARCH DESIGNS

T. E. Raghunathan Department of Biostatistics and Research Professor, Institute for Social Research, University of Michigan, Ann Arbor, USA

Comparing Experimental and Matching Methods using a Large-Scale Field Experiment on Voter Mobilization

Doing After Seeing. Seeing vs. Doing in Causal Bayes Nets

Causal Effect Heterogeneity

Systematic Reviews in Education Research: When Do Effect Studies Provide Evidence?

Causality and Treatment Effects

INTRODUCTION TO ECONOMETRICS (EC212)

Econometrics II - Time Series Analysis

You must answer question 1.

Understanding the Causal Logic of Confounds

Citation for published version (APA): Ebbes, P. (2004). Latent instrumental variables: a new approach to solve for endogeneity s.n.

Compared to What? The Effectiveness of Synthetic Control Methods for Causal Inference in Educational Assessment

Identification with Models and Exogenous Data Variation

School Autonomy and Regression Discontinuity Imbalance

Brief introduction to instrumental variables. IV Workshop, Bristol, Miguel A. Hernán Department of Epidemiology Harvard School of Public Health

Book review of Herbert I. Weisberg: Bias and Causation, Models and Judgment for Valid Comparisons Reviewed by Judea Pearl

Logistic regression: Why we often can do what we think we can do 1.

Prediction, Causation, and Interpretation in Social Science. Duncan Watts Microsoft Research

Identifying Mechanisms behind Policy Interventions via Causal Mediation Analysis

How Causal Heterogeneity Can Influence Statistical Significance in Clinical Trials

Transcription:

ICPSR Summer Program Causal Inference Course Syllabus Instructors: Dominik Hangartner Marco R. Steenbergen Summer 2011

Contact Information: Marco Steenbergen, Department of Social Sciences, University of Bern. Email: marco.steenbergen@ipw.unibe.ch. Dominik Hangartner, Institute for Governmental Studies, University of California, Berkeley. Email: hangartner@berkeley.edu. Course Information: Meeting times: June 27 July 1, 2011; 9am-5pm. Short Course Description This course provides an introduction to statistical methods used in causal inference and is geared specifically toward students in the social sciences. Using the potential outcomes framework of causality, topics covered include research designs such as randomized experiments and observational studies. We explore the impact of noncompliance in randomized experiments, as well as nonignorable treatment assignment in observational studies. To analyze these research designs, the methods covered include matching, instrumental variables, difference-in-difference, and regression discontinuity. Examples are drawn from economics, political science, public health and sociology. Prerequisites for the course are knowledge of multiple regression using linear algebra and some familiarity with generalized linear models. The course will rely on R for computation. Overview and Objectives Causal inference from observational studies or ``broken'' randomized experiments (due to e.g. noncompliance) historically has been viewed as problematic, or even illegitimate, by most statisticians. This has led to the paradoxical situation that statisticians provided countless sophisticated methods to model conditional expectations for association and prediction, while being almost completely silent about how to validly estimate the effect of a cause---although that is exactly what most of the social and behavioral sciences is all about. While randomized experiments are rightfully considered to be the gold standard for causal inference, they are quite often not a realistic option because of ethical and / or practical limitations. Then, one typically has two choices: to give up and miss the opportunity to learn anything about the problem at hand or to turn to observational studies. By taking this class, you opt for the second option. However, causal inference from observational studies or broken randomized experiments is notoriously difficult, since (for point identification) it necessarily involves unidentifiable and hence untestable assumptions about the magnitude and direction of confounding. Hence, the plausibility of causal inferences from observational data always depends on the amount and strength of substance-matter knowledge and the quality of the research

design. This course aims at providing you with the tools to engage in proper causal inference. The specific goals are twofold: (1) to sensitize you to the many mines that one could step on while making causal claims and (2) to provide a set of research designs and statistical techniques that can help you avoid some of these mines. We pursue these goals within the broad framework of the so called Rubin Causal Model (RCM) which is based on Neyman's potential outcomes notation. The hope is that, by the end of the class, you will be sensitive to the assumptions that causal inference entails and that you will be able to work through research designs and methodologies that, as much as possible, avoid implausible assumptions. Organization Many problems of causal inference from observational studies revolve around the concept of confounders, i.e. true causes of an effect that may render the impact of a putative cause spurious. There are different ways of handling confounders, depending on whether they are observed or not. After proving a general introduction to causation and causal inference (Module I), we begin by considering research designs in which the confounders are unobserved but rendered impotent through randomization (Module II). Since randomization is not always feasible, we may have to rely on other methods of controlling for confounders. In Module III, we consider designs in which the confounders are observed and can be controlled statistically. In Module IV, we focus on cases in which at least some of the confounders are unobserved. Prerequisites Prerequisites for the course are knowledge of multiple regression using linear algebra and some familiarity with generalized linear models. The course will rely on R for computation. If you need to review material, please consult this excellent textbook: Freedman, David. 2005. Statistical Models: Theory and Practice. Cambridge University Press. If you need to review some R basics, you may want to have a look at Fox, John. 2002. An R and S-PLUS Companion to Applied Regression. Sage Publications. Materials The following textbooks are required:

Angrist, Joshua D., and Jorn-Steffen Pischke. 2009. Mostly Harmless Econometrics. Princeton: Princeton University Press. Rosenbaum, Paul. 2009. Design of Observational Studies. New York: Springer. In addition to the required books, you may wish to obtain a copy of the textbook: Morgan, Stephen L., and Christopher Winship. 2007. Counterfactuals and Causal Inference. New York: Cambridge University Press. Evaluation Your course grade is determined on the basis of your homeworks. Most of these exercises will require that you use the statistical package R, which may be downloaded free of charge from http://www.r-project.org/. There will be no final exam. Schedule Module I Causality and Causal Inference Session 1 Causality I: Causal Inference Using Potential Outcomes Today we will introduce the topic of causal inference. We will define causal effects based on the potential outcomes framework of Neyman and Rubin, encounter the fundamental problem of causal inference, and discuss confounding as what separates association from causation, and observational studies from randomized experiments. Freedman, David A. 1991. Statistical Models and Shoe Leather. Sociological Methodology 2: 291-313. Further Splawa-Neyman, Jerzy, [Dabrowska, D. M., and T.P. Speed]. 1923 [1990]. On the Application of Probability Theory to Agricultural Experiments. Essay on Principles. Section 9. Statistical Science 5: 465-472. Rubin, Donald B. 1990. Comment: Neyman (1923) and Causal Inference in Experiments and Observational Studies. Statistical Science 5: 472-480.

Session 2 Causality II: Different Models of Causality We discuss the logic of the Neyman-Rubin Causal Model. We relate the model to alternative and complementary conceptions of causation in philosophy of social science, with special emphasis on counterfactual theories based on the ``closest possible worlds'' (Lewis 1973) and Woodward's (2003) ``manipulability theory'' of causation and causal explanation. Going further back in time, we review Mill's method of difference and contrast it with Fisher's randomized experiment as the ``reasoned basis'' of inference. In addition, we revisit the controversy about ``No Causation Without Manipulation'' (Holland 1986) and discuss the importance of well-defined interventions for meaningful causal inference. Morgan and Winship. 2007. Chapter 1: 1-24. Holland, Paul W. 1986. Statistics and Causal Inference (with discussion). Journal of the American Statistical Association 81: 945-970. Further Woodward, James. 2003. Making Things Happen: a Theory of Causal Explanation. Oxford University Press. Module II Randomized Experiments Session 3 Randomized Experiments We quickly review the logic of randomized experimentation, a research design that is widely believed to maximize internal validity andthat is becoming ever more popular in the social sciences. We pay special attention to Fisher's randomization inference, in which randomization is the ``sole and reasoned basis for inference''. We discuss factorial and randomized-block designs, paying attention to how these designs function. Rosenbaum, Paul. 2009. Design of Observational Studies. New York: Springer, Chapter 1 and 2. Further

Bloom, Howard S. 2006. The Core Analytics of Randomized Experiments for Social Research. MDRC Working Paper. (available online) Session 4 Causality III: Different Approaches to Causal Inference We will discuss the graphical approach to causal inference as introduced by Pearl (e.g. 2000) and Spirtes, Glymour, and Scheines (1993). We discuss some distinctive advantages and drawbacks of directed acyclical graphs (DAGs) and show how they complement the potential outcomes framework as a different, but equivalent language (Lauritzen 2004). Based on DAGs, we revisit and old debate about the (im-) possibility to draw causal inference without substance-matter knowledge. Morgan and Winship. Chapter 1.6, pp. 24-30. Further Pearl, Judea. 2009. Causality. Models, Reasoning, and Inference. 2nd Edition. Cambridge University Press. Spirtes, Peter, Glymour, Clark and Richard Scheines. 1993. Causation, Prediction, and Search, New York: Springer-Verlag. 2nd Edition, MIT Press Lauritzen, Steven L. 2004. ``Discussion on Causality''. Scandinavian Journal of Statistics. 31: 189-193. Robins, James M., Scheines, Richard, Spirtes, Peter and Larry Wasserman (2003). ``Uniform Consistency in Causal Inference''. Biometrika. 90(3):491-515. Homework 1: Assessing the effect of watching more Sesame Street on child cognitive development; program effect of watching TV vs. program effect of encouragement to watch. Module III Causal Inference Under Exogeneity Session 5 Subclassification and Matching I The advantage of randomized experiments is that potential confounders can be safely ignored since they will not bias estimates of treatment effects. But randomization is not always practical, nor is it always ethical. How can one ensure valid causal inference in a world without randomization? Today, we discuss designs which assume that selection into the treatment groups is based on observables. We start by considering two very intuitive methods, subclassification and exact

matching techniques. Morgan and Winship. 2007. Chapter 4: 87-122. Further Cochran, W. G. 1968. The Effectiveness of Adjustment by Subclassification in Removing Bias in Observational Studies. Biometrics 24: 295-313. Session 6 Matching II Next, we discuss matching techniques that are based the propensity score and on Euclidean distance. We also consider some practical issues with matching such as severe sample size reduction, outliers/common support, and estimating standard errors. With respect to the latter topic, we consider the (somewhat surprising) failing of bootstrapping methods for valid standard errors. Dehejia, R. H. and S. Wahba. 1999. Causal Effects in Non-Experimental Studies: Re- Evaluating the Evaluation of Training Programs. Journal of the American Statistical Association 94: 1053-1062. Further readings: Rosenbaum, P. R., and D. B. Rubin. 1983. The Central Role of the Propensity Score in Observational Studies for Causal Effects. Biometrika 70: 41-55. Abadie, A. and G. Imbens. 2005. Large Sample Properties of Matching Estimators for Average Treatment Effects. Econometrica 74: 235-267. Session 7 Matching III Today, we consider weighting as a nifty alternative to matching. In addition, we compare the OLS regression with matching estimators. Angrist and Pischke. Chapter 3.3.1-3.3.3, pp.69-91. Session 8

Matching VI Review / Lab Session: Introduction to the R package ``Matching'' (Sekhon 2007). Homework 2: Re-analysis of Dehejia and Wahba (1999). Module IV Causal Inference When the Confounders Are Not All Observed Session 9 Difference-in-Differences Confounders cannot always be observed and if they cannot, then alternatives to matching have to be found. One such alternative arises in the context of panel data or repeated cross-sections. Here one can take the difference between pre- and post-tests and then compare those across groups. To the extent that the differences in the confounders have remained constant over time, then this estimator can produce valid causal inferences. Angrist and Pischke. Chapter 5, pp. 221-246. Further Card, David and Alan B. Krueger. 1994. Minimum Wages and Employment: A Case Study of the Fast-Food Industry in New Jersey and Pennsylvania. The American Economic Review 84: 772-793. Card, David. 1990. The Impact of the Mariel Boatlift on the Miami Labor Market. Industrial and Labor Relations Review 43: 245-257. Session 10 Synthetic Control If you would like to use Difference-in-Difference but don't have a good control unit: synthesize one (Abadie and Gardeazabal 2003, Abadie, Diamond, Hainmueller 2007). This estimator is a simple but potentially wildely applicable generalization of the Difference-in-Differences estimator. Abadie, Alberto, Diamon, Alexis and Jens Hainmueller. 2007. Synthetic Control Methods for Comparative Case Studies: Estimating the Effect of California's Tobacco Control Program. NBER Working Paper.

Homework 3: Replication of Card and Krueger (1994). Session 11 Instrumental Variables I Instrumental variable (IV) methods can be used to address unobserved confounders in the context of cross-sectional data. Today, we discuss the basic logic of IV-techniques, focusing in particular on the LATE estimator. Angrist and Pischke. Chapter 4.1-4.4.4 Further readings: Angrist, Joshua D., Guido W. Imbens and Donald B. Rubin. 1996. Identification of Causal Effects Using Instrumental Variables. Journal of the American Statistical Association 9: 444-455. Session 12 Review / Lab Session: Illustration of Synthetic Control methods. Session 13 Instrumental Variables II In many settings, an instrument is only valid conditional on some covariates. As we will show, the LATE estimator is only able to identify average treatment effects conditional on covariates under very restrictive assumptions. Abadie's LARF estimator develops an ingenuious weighting scheme to get unbiased treatment effects under heterogeneity for the subsample of compliers only. Angrist and Pischke. Chapter 4.5-4.7 Session 14 Lab Session: Illustration of LATE and LARF estimators. Homework 4: LATE and LARF. Session 15 Regression Discontinuity Designs I

RDDs arise when selection into the treatment group depends on a covariate score that creates some discontinuity in the probability of receiving the treatment. Capturing the treatment effect in these designs can be thought of as a special case of IVs. Although RDDs are not yet widely used in the social sciences, they appear to have a great deal of potential for application. Angrist and Piscke. Chapter 6. Further Lee, David S. 2005. Randomized Experiments from Non-random Selection in U.S. House Elections. Working Paper Pettersson-Lidbom, Per and Bjorn Tyrefors. 2007. The Policy Consequences of Direct versus Representative Democracy: A Regression-Discontinuity Approach. Working Paper. (available at: http://people.su.se/pepet/directdem.pdf) Hahn, Jinyong, Petra Todd and Wilbert Van der Klaauw. 2001. Identification and Estimation of Treatment Effects with a Regression-Discontinuity Design. Econometrica 69: 201-209. Eggers, Andy, and Jens Hainmueller. 2008. MPs For Sale? Estimating Returns to Office in Post-War British Politics. Working Paper