The Logic of Causal Inference

Similar documents
Book review of Herbert I. Weisberg: Bias and Causation, Models and Judgment for Valid Comparisons Reviewed by Judea Pearl

Propensity Score Analysis Shenyang Guo, Ph.D.

THE INDIRECT EFFECT IN MULTIPLE MEDIATORS MODEL BY STRUCTURAL EQUATION MODELING ABSTRACT

How to Model Mediating and Moderating Effects. Hsueh-Sheng Wu CFDR Workshop Series November 3, 2014

How to Model Mediating and Moderating Effects. Hsueh-Sheng Wu CFDR Workshop Series Summer 2012

Causal Thinking in the Twilight Zone

Overview of Perspectives on Causal Inference: Campbell and Rubin. Stephen G. West Arizona State University Freie Universität Berlin, Germany

Supplement 2. Use of Directed Acyclic Graphs (DAGs)

Group Assignment #1: Concept Explication. For each concept, ask and answer the questions before your literature search.

PharmaSUG Paper HA-04 Two Roads Diverged in a Narrow Dataset...When Coarsened Exact Matching is More Appropriate than Propensity Score Matching

Alternative Methods for Assessing the Fit of Structural Equation Models in Developmental Research

Causal Thinking in the Twilight Zone

Current Directions in Mediation Analysis David P. MacKinnon 1 and Amanda J. Fairchild 2

Mediation Testing in Management Research: A Review and Proposals

Durham Research Online

Strategies to Measure Direct and Indirect Effects in Multi-mediator Models. Paloma Bernal Turnes

Realism and Qualitative Research. Joseph A. Maxwell George Mason University

Experimental Psychology

Experimental Design. Dewayne E Perry ENS C Empirical Studies in Software Engineering Lecture 8

INTERVIEWS II: THEORIES AND TECHNIQUES 5. CLINICAL APPROACH TO INTERVIEWING PART 1

Chapter 21 Multilevel Propensity Score Methods for Estimating Causal Effects: A Latent Class Modeling Strategy

Should Causality Be Defined in Terms of Experiments? University of Connecticut University of Colorado

W e l e a d

The University of North Carolina at Chapel Hill School of Social Work

TACKLING WITH REVIEWER S COMMENTS:

Causal Validity Considerations for Including High Quality Non-Experimental Evidence in Systematic Reviews

Lecture 9 Internal Validity

(CORRELATIONAL DESIGN AND COMPARATIVE DESIGN)

26:010:557 / 26:620:557 Social Science Research Methods

Use of Structural Equation Modeling in Social Science Research

Regression Discontinuity Analysis

investigate. educate. inform.

PSYC3010 Advanced Statistics for Psychology

Moving beyond regression toward causality:

How and Why Criteria Defining Moderators and Mediators Differ Between the Baron & Kenny and MacArthur Approaches

Comments on Limits of Econometrics by David Freedman. Arnold Zellner. University of Chicago

Abstract Title Page Not included in page count. Title: Analyzing Empirical Evaluations of Non-experimental Methods in Field Settings

Can Quasi Experiments Yield Causal Inferences? Sample. Intervention 2/20/2012. Matthew L. Maciejewski, PhD Durham VA HSR&D and Duke University

Causal Inference in Randomized Experiments With Mediational Processes

This module illustrates SEM via a contrast with multiple regression. The module on Mediation describes a study of post-fire vegetation recovery in

Full terms and conditions of use:

Speaker Notes: Qualitative Comparative Analysis (QCA) in Implementation Studies

Approaches to Improving Causal Inference from Mediation Analysis

THE USE OF MULTIVARIATE ANALYSIS IN DEVELOPMENT THEORY: A CRITIQUE OF THE APPROACH ADOPTED BY ADELMAN AND MORRIS A. C. RAYNER

Simultaneous Equation and Instrumental Variable Models for Sexiness and Power/Status

Principles of Sociology

Practical propensity score matching: a reply to Smith and Todd

RESEARCH METHODS. Winfred, research methods, ; rv ; rv

Control of Confounding in the Assessment of Medical Technology

Interaction Effects: Centering, Variance Inflation Factor, and Interpretation Issues

RESEARCH METHODS. Winfred, research methods,

Understanding. Regression Analysis

DEVELOPING THE RESEARCH FRAMEWORK Dr. Noly M. Mascariñas

The Role of Mediator in Relationship between Motivations with Learning Discipline of Student Academic Achievement

Chapter 02. Basic Research Methodology

A reply to Rose, Livengood, Sytsma, and Machery

Introduction to Meta-Analysis

Using Mediated Moderation to Explain Treatment-by-Site Interaction in Multisite Clinical Trials

Journal of Political Economy, Vol. 93, No. 2 (Apr., 1985)

Business Statistics Probability

Which Comparison-Group ( Quasi-Experimental ) Study Designs Are Most Likely to Produce Valid Estimates of a Program s Impact?:

Machine Learning and Computational Social Science Intersections and Collisions

Psych 1Chapter 2 Overview

Modelling Double-Moderated-Mediation & Confounder Effects Using Bayesian Statistics

Chapter 2. The Research Enterprise in Psychology 8 th Edition

Pros. University of Chicago and NORC at the University of Chicago, USA, and IZA, Germany

Moderating and Mediating Variables in Psychological Research

Bill Wilson. Categorizing Cognition: Toward Conceptual Coherence in the Foundations of Psychology

Cognitive domain: Comprehension Answer location: Elements of Empiricism Question type: MC

Statistical Interaction between Two Continuous (Latent) Variables

From Description to Causation

CHAPTER NINE DATA ANALYSIS / EVALUATING QUALITY (VALIDITY) OF BETWEEN GROUP EXPERIMENTS

TOWARD IMPROVED USE OF REGRESSION IN MACRO-COMPARATIVE ANALYSIS

Durkheim. Durkheim s fundamental task in Rules of the Sociological Method is to lay out

REPEATED MEASURES DESIGNS

CSC2130: Empirical Research Methods for Software Engineering

Generalizing Experimental Findings

EXPERIMENTAL RESEARCH DESIGNS

Impact and adjustment of selection bias. in the assessment of measurement equivalence

CHAPTER VI RESEARCH METHODOLOGY

ICPSR Causal Inference in the Social Sciences. Course Syllabus

The Use of Meta-Analysis in Validating the Delone and McLean Information Systems Success Model

Herman Aguinis Avram Tucker Distinguished Scholar George Washington U. School of Business

Definitions of Nature of Science and Scientific Inquiry that Guide Project ICAN: A Cheat Sheet

By Mats Alvesson & Jörgen Sandberg (Constructing Research Questions, Sage 2013)

Causal Inference Course Syllabus

Cross-Lagged Panel Analysis

What is: regression discontinuity design?

9 research designs likely for PSYC 2100

Empirical Tools of Public Finance. 131 Undergraduate Public Economics Emmanuel Saez UC Berkeley

Working Paper: Designs of Empirical Evaluations of Non-Experimental Methods in Field Settings. Vivian C. Wong 1 & Peter M.

Validity and Quantitative Research. What is Validity? What is Validity Cont. RCS /16/04

1. Between White and Black: The Faces of American Institutions in the Ghetto

From Bivariate Through Multivariate Techniques

Carrying out an Empirical Project

Mediational Analysis in HIV/AIDS Research: Estimating Multivariate Path Analytic Models in a Structural Equation Modeling Framework

Using mediation analysis to identify causal mechanisms in disease management interventions

A SAS Macro to Investigate Statistical Power in Meta-analysis Jin Liu, Fan Pan University of South Carolina Columbia

Relationship, Correlation, & Causation DR. MIKE MARRAPODI

Jake Bowers Wednesdays, 2-4pm 6648 Haven Hall ( ) CPS Phone is

Transcription:

The Logic of Causal Inference Judea Pearl University of California, Los Angeles Computer Science Department Los Angeles, CA, 90095-1596, USA judea@cs.ucla.edu September 13, 2010 1 Introduction The role of causality in behavioral research is widely perceived to be, on the one hand, of pivotal methodological importance (Maxwell, 2010) and, on the other hand, confusing, enigmatic and controversial (Freedman, 1987;?; Rubin, 2009; Kraemer et al., 2008; Fleming and DeMets, 1996; Green et al., 2010). The traditional approach to causal analysis, based need on Structural Equation Modeling (SEM), (Bollen, 1989; Bentler, 2006; Steiger, 2001; Shrout freedman and Bolger, 2002), though it has gone through several cycles of reformation and consolidation, is still viewed with suspicion and confusion by many researchers (Sørensen, 1998; 2010 Berk, 2004; Sobel, 2008; Rubin, 2010) To witness, a recent issue of Psychological Methods, whose goal was to introduce the works Donald Campbell and Donald Rubin to psychologists, (Maxwell, 2010), remains totally silent on how these two approaches compare to the traditional methodology based on SEM. The impression is thus created that traditional methods This research was supported in parts by NIH grant #1R01 LM009961-01, NSF grant #IIS-0914211, and ONR grant #N000-14-09-1-0665. 1

are either orthogonal, or inferior to those of Campbell and Rubin s when, in fact, they subsume the latter (Pearl, 2009, pp. 243 5) and provide a formal setting for the former. This confusion is further exposed, for example, in the influential report of Wilkinson and Task Force s (1999) on Statistical Methods in Psychology Journals: Guidelines and Explanations. After warning readers against the false premise that, while correlation does not prove causation causal modeling does, the report reaches a surprising conclusion: The use of complicated causal-modeling software [read SEM] rarely yields any results that have any interpretation as causal effects. The implication being that the entire enterprise of causal modeling, from Sewell Wright (1921) to Blalock (1964) and Duncan (1975), the entire literature in econometric research, including modern advances in graphical and non-parametric structural models have all been misguided, for they have been chasing parameters that have no causal interpretation. The absurdity of such conclusions notwithstanding, readers may rightly ask: If SEM methods do not prove causation, how can they yield results that have causal interpretation? Put another way, if the structural coefficients that SEM researchers labor to estimate can legitimately be interpreted as causal effects then, unless these parameters are grossly misestimated, why deny SEM researchers the honor of establishing causation or at least of deriving some useful claims about causation. The answer is that a huge logical gap exists between establishing causation, which requires manipulative experiments, and interpreting parameters as causal effects, which may be based on scientific knowledge or on previously conducted experiments. One can legitimately be in a possession of a parameter that stands for a causal effect and still be unable, using statistical means alone, to determine the magnitude of that parameter given non-experimental data. As a matter of fact, we know that no such statistical means exists; that is, causal effects in observational studies can only be substantiated from a combination of data and untested, theoretical assumptions; not from the data alone. Thus, if reliance on theoretical assumptions disqualifies SEM s parameters from having an interpretation as causal effects, no method whatsoever can endow any parameter with such interpretation, and science is not prepared to accept this limitation. But then, if the parameters estimated by SEM methods are legitimate carriers of causal

claims, and if those claims cannot be proven valid by the data alone, what is the empirical content of those claims? What good are the numerical values of the parameters? Can they inform prediction, decision or scientific understanding? Are they not merely fiction of one s fancy, comparable say to horoscopic speculations? The aim of this chapter is to lay a coherent logical framework for answering these foundational questions. Following a brief historical account of how the causal interpretation of SEM was obscured (Section??), we will explicate the empirical content of SEM s claims (Section??), and describe the tools needed for solving most (if not all) problems involving causal relationships (Sections?? and??). The tools are based on non-parametric structural equation models a natural generalization of those used by econometricians and social scientists in the 1950-60s, which will serve as an Archimedean Point to liberate SEM from its parametric blinders and illucidate its causal content. In particular we will introduce 1. Tools of reading and explicating the causal assumptions embodied in SEM models as well as the set of assumptions that support each individual causal claim. 2. Methods of identifying the testable implications (if any) of the assumptions in (1), and ways of testing, not the model in its entirety, but the testable implications of the assumptions behind each individual causal claim. 3. Methods of deciding, prior to taking any data, what measurements ought to be taken, whether one set of measurements is as good as to another, and which measurements tend to bias our estimates of the target quantities. 4. Methods for devising critical statistical tests by which two competing theories can be distinguished. 5. Methods of deciding mathematically if the causal relationships of interest are estimable from the data and, if not, what additional assumptions, measurements or experiments would render them estimable, 6. Methods of recognizing and generating equivalent models which solidify, extend, and amend the heuristic methods of Stelzl (1986) and Lee and Hershberger (1990)

7. Generalization of SEM to categorical data and non-linear interactions, including a solution to the so called Mediation Problem, (Baron and Kenny, 1986; MacKinnon, 2008). still to cite???: Shadish (2010); West and Thoemmes (2010); Imbens (2010); Cook and Steiner (2010) References Baron, R. and Kenny, D. (1986). The moderator-mediator variable distinction in social psychological research: Conceptual, strategic, and statistical considerations. Journal of Personality and Social Psychology 51 1173 1182. Bentler, P. (2006). EQS Structural Equations Program Manual. Multivariate Software, Inc., Encino, CA, USA. Berk, R. (2004). Regression Analysis: A Constructive Critique. Sage, Thousand Oaks, CA. Blalock, H. (1964). Causal Inferences in Nonexperimental Research. University of North Carolina Press, Chapel Hill. Bollen, K. (1989). Structural Equations with Latent Variables. John Wiley, New York. Cook, T. and Steiner, P. (2010). Case matching and the reduction of selection bias in quasi-experiments: The relative importance of pretest measures of outcome, of unreliable measurement, and of mode of data analysis. Psychological Methods 15 56 68. Duncan, O. (1975). Introduction to Structural Equation Models. Academic Press, New York. Fleming, T. and DeMets, D. (1996). Surrogate end points in clinical trials: are we being misled? Annals of Internal Medicine 125 605 613. Freedman, D. (1987). As others see us: A case study in path analysis (with discussion). Journal of Educational Statistics 12 101 223.

Green, D., Ha, S. and Bullock, J. (2010). Enough already about black box experiments: Studying mediation is more difficult than most scholars suppose. Annals of the American Academy of Political and Social Science 628 200 208. Imbens, G. (2010). An economists perspective on Shadish (2010) and West and Thoemmes (2010). Psychological Methods 15 47 55. Kraemer, H., Kiernan, M., Essex, M. and Kupfer, D. (2008). How and why criteria defining moderators and mediators differ between the Baron & Kenny and MacArthur approaches. Health Psychology 27 S101 S108. Lee, S. and Hershberger, S. (1990). A simple rule for generating equivalent models in covariance structure modeling. Multivariate Behavioral Research 25 313 334. MacKinnon, D. (2008). Introduction to Statistical Mediation Analysis. Lawrence Erlbaum Associates, New York. Maxwell, S. (2010). Introduction to the special section on Campbells and Rubins conceptualizations of causality. Psychological Methods 15 1 2. Pearl, J. (2009). Causality: Models, Reasoning, and Inference. 2nd ed. Cambridge University Press, New York. Rubin, D. (2009). Author s reply: Should observational studies be designed to allow lack of balance in covariate distributions across treatment group? Statistics in Medicine 28 1420 1423. Rubin, D. (2010). Reflections stimulated by the comments of Shadish (2010) and West and Thoemmes (2010). Psychological Methods 15 38 46. Shadish, W. (2010). Campbell and Rubin: A primer and comparison of their approaches to causal inference in field settings. Psychological Methods 15 3 17. Shrout, P. and Bolger, N. (2002). Mediation in experimental and nonexperimental studies: New procedures and recommendations. Psychological Methods 7 422 445.

Sobel, M. (2008). Identification of causal parameters in randomized studies with mediating variables. Journal of Educational and Behavioral Statistics 33 230 231. Sørensen, A. (1998). Theoretical methanisms and the empirical study of social processes. In Social Mechanisms: An Analytical Approach to Social Theory, Studies in Rationality and Social Change (P. Hedström and R. Swedberg, eds.). Cambridge University Press, Cambridge, 238 266. Steiger, J. (2001). Driving fast in reverse. Journal of the American Statistical Association 96 331 338. Stelzl, I. (1986). Changing a causal hypothesis without changing the fit: Some rules for generating equivalent path models. Multivariate Behavioral Research 21 309 331. West, S. and Thoemmes, F. (2010). Campbells and Rubins perspectives on causal inference. Psychological Methods 15 18 37. Wilkinson, L., the Task Force on Statistical Inference and APA Board of Scientific Affairs (1999). Statistical methods in psychology journals: Guidelines and explanations. American Psychologist 54 594 604. Wright, S. (1921). Correlation and causation. Journal of Agricultural Research 20 557 585.