Causal Thinking in the Twilight Zone

Similar documents
Causal Thinking in the Twilight Zone

Book review of Herbert I. Weisberg: Bias and Causation, Models and Judgment for Valid Comparisons Reviewed by Judea Pearl

Propensity Score Analysis Shenyang Guo, Ph.D.

Generalizing Experimental Findings

The Logic of Causal Inference

Generalizing Experimental Findings

Supplement 2. Use of Directed Acyclic Graphs (DAGs)

Structural Approach to Bias in Meta-analyses

Generalization and Theory-Building in Software Engineering Research

ICPSR Causal Inference in the Social Sciences. Course Syllabus

Careful with Causal Inference

1 Transportability and selection bias

Principal Stratification A goal or a tool?

Introduction to Observational Studies. Jane Pinelis

Causal Methods for Observational Data Amanda Stevenson, University of Texas at Austin Population Research Center, Austin, TX

How Does Analysis of Competing Hypotheses (ACH) Improve Intelligence Analysis?

Questions and Answers

Evaluation of Medication-Mediated Effects in Pharmacoepidemiology. By EJ Tchetgen Tchetgen and K Phiri

August 29, Introduction and Overview

PharmaSUG Paper HA-04 Two Roads Diverged in a Narrow Dataset...When Coarsened Exact Matching is More Appropriate than Propensity Score Matching

Wason's Cards: What is Wrong?

Jake Bowers Wednesdays, 2-4pm 6648 Haven Hall ( ) CPS Phone is

Lecture 9 Internal Validity

Judea Pearl. Cognitive Systems Laboratory. Computer Science Department. University of California, Los Angeles, CA

Overview of Perspectives on Causal Inference: Campbell and Rubin. Stephen G. West Arizona State University Freie Universität Berlin, Germany

Causal Inference Course Syllabus

PARTIAL IDENTIFICATION OF PROBABILITY DISTRIBUTIONS. Charles F. Manski. Springer-Verlag, 2003

Bradford Hill Criteria for Causal Inference Based on a presentation at the 2015 ANZEA Conference

Impact and adjustment of selection bias. in the assessment of measurement equivalence

Summary. 20 May 2014 EMA/CHMP/SAWP/298348/2014 Procedure No.: EMEA/H/SAB/037/1/Q/2013/SME Product Development Scientific Support Department

Lab 2: The Scientific Method. Summary

STATISTICAL INFERENCE 1 Richard A. Johnson Professor Emeritus Department of Statistics University of Wisconsin

Practical propensity score matching: a reply to Smith and Todd

Propensity Score Methods for Estimating Causality in the Absence of Random Assignment: Applications for Child Care Policy Research

(CORRELATIONAL DESIGN AND COMPARATIVE DESIGN)

P(HYP)=h. P(CON HYP)=p REL. P(CON HYP)=q REP

Logistic regression: Why we often can do what we think we can do 1.

Chapter 3 Tools for Practical Theorizing: Theoretical Maps and Ecosystem Maps

Cross-Lagged Panel Analysis

Research Approach & Design. Awatif Alam MBBS, Msc (Toronto),ABCM Professor Community Medicine Vice Provost Girls Section

Causal Inference in Statistics and the Quantitative Sciences

Kepler tried to record the paths of planets in the sky, Harvey to measure the flow of blood in the circulatory system, and chemists tried to produce

Quantitative Methods. Lonnie Berger. Research Training Policy Practice

Underlying Theory & Basic Issues

Pros. University of Chicago and NORC at the University of Chicago, USA, and IZA, Germany

INTERVIEWS II: THEORIES AND TECHNIQUES 5. CLINICAL APPROACH TO INTERVIEWING PART 1

Epidemiologic Causation: Jerome Cornfield s Argument for a Causal Connection between Smoking and Lung Cancer

METHODOLOGY FOR DISSERTATION

II. The Behavioral Approach to Understanding Cognition

Comments on Limits of Econometrics by David Freedman. Arnold Zellner. University of Chicago

Modelling Causation in Evidence-Based Legal Reasoning ICAIL Doctoral Consortium

A Brief Introduction to Bayesian Statistics

VALIDITY OF QUANTITATIVE RESEARCH

[1] provides a philosophical introduction to the subject. Simon [21] discusses numerous topics in economics; see [2] for a broad economic survey.

Checking the counterarguments confirms that publication bias contaminated studies relating social class and unethical behavior

Chapter 21 Multilevel Propensity Score Methods for Estimating Causal Effects: A Latent Class Modeling Strategy

Modelling Double-Moderated-Mediation & Confounder Effects Using Bayesian Statistics

Research Design. Beyond Randomized Control Trials. Jody Worley, Ph.D. College of Arts & Sciences Human Relations

Propensity Score Methods for Causal Inference with the PSMATCH Procedure

Response to the ASA s statement on p-values: context, process, and purpose

Stepwise Knowledge Acquisition in a Fuzzy Knowledge Representation Framework

Should Causality Be Defined in Terms of Experiments? University of Connecticut University of Colorado

Representation Theorems for Multiple Belief Changes

Competency Rubric Bank for the Sciences (CRBS)

Identifying Peer Influence Effects in Observational Social Network Data: An Evaluation of Propensity Score Methods

Identifying Mechanisms behind Policy Interventions via Causal Mediation Analysis

Evolutionary Approach to Investigations of Cognitive Systems

Durkheim. Durkheim s fundamental task in Rules of the Sociological Method is to lay out

Samples, Sample Size And Sample Error. Research Methodology. How Big Is Big? Estimating Sample Size. Variables. Variables 2/25/2018

Evaluating health management programmes over time: application of propensity score-based weighting to longitudinal datajep_

Meta-Analysis. Zifei Liu. Biological and Agricultural Engineering

Measuring Impact. Conceptual Issues in Program and Policy Evaluation. Daniel L. Millimet. Southern Methodist University.

Prediction, Causation, and Interpretation in Social Science. Duncan Watts Microsoft Research

Understanding the Causal Logic of Confounds

Design of Experiments & Introduction to Research

Statistical Causality

Effects of propensity score overlap on the estimates of treatment effects. Yating Zheng & Laura Stapleton

Comments on David Rosenthal s Consciousness, Content, and Metacognitive Judgments

Recent developments for combining evidence within evidence streams: bias-adjusted meta-analysis

The Development and Application of Bayesian Networks Used in Data Mining Under Big Data

On the Representation of Nonmonotonic Relations in the Theory of Evidence

Support system for breast cancer treatment

Principles of Sociology

Climate-Related Decision Making Under Uncertainty. Chris Weaver February 2016

Temporalization In Causal Modeling. Jonathan Livengood & Karen Zwier TaCitS 06/07/2017

Assignment 4: True or Quasi-Experiment

Propensity Score Analysis and Strategies for Its Application to Services Training Evaluation

Funnelling Used to describe a process of narrowing down of focus within a literature review. So, the writer begins with a broad discussion providing b

Models of Information Retrieval

School of Nursing, University of British Columbia Vancouver, British Columbia, Canada

How to interpret results of metaanalysis

EXPERIMENTAL RESEARCH DESIGNS

The use of quasi-experiments in the social sciences: a content analysis

CSC2130: Empirical Research Methods for Software Engineering

Bill Wilson. Categorizing Cognition: Toward Conceptual Coherence in the Foundations of Psychology

Doing After Seeing. Seeing vs. Doing in Causal Bayes Nets

Simpson s Paradox and the implications for medical trials

Recent advances in non-experimental comparison group designs

Methodological Thinking on Disciplinary Research Fields of Theories of Sports Training from the Philosophical Perspective Bin LONG

Title:Continuity of GP care is associated with lower use of complementary and alternative medical providers A population-based cross-sectional survey

Transcription:

Forthcoming: Observational Studies Special Issue on William Cochran's contributions. TECHNICAL REPORT R-456 July 2015 Causal Thinking in the Twilight Zone Judea Pearl University of California, Los Angeles Computer Science Department Los Angeles, CA, 90095-1596, USA (310) 825-3243 / judea@cs.ucla.edu To students of causality, the writings of William Cochran provide an excellent and intriguing vantage point for studying how statistics, lacking the necessary mathematical tools, managed nevertheless to cope with increasing demands for policy evaluation from observational studies. Cochran met this challenge in the years 1955-1980, when statistics was preparing for a profound, albeit tortuous transition from a science of data, to a science of data generating processes. The former, governed by Fisher s dictum (1922) the object of statistical methods is the reduction of data was served well by the traditional language of probability theory. The latter, on the other hand, seeking causal effects and policy recommendations, required an extension of probability theory to facilitate mathematical representations of generating processes. No such representation was allowed into respectable statistical circles in the 1950-60s, when Cochran started looking into the social effects of public housing in Baltimore. While data showed improvement in health and well-being of families that moved from slums to public housing, it soon became obvious that the estimated improvement was strongly biased; Cochran reasoned that in order to become eligible for public housing the parent of a family may have to possess both initiative and some determination in dealing with the bureaucracy, thus making their families more likely to obtain better healthcare than non-eligible families. 1 This led him to suggest adjustment for covariates for the explicit purpose of reducing this causal effect bias. While there were others before Cochran who applied adjustment for various purposes, Cochran is credited for introducing this technique to statistics (Salsburg, 2002) primarily because he popularized the method and taxonomized it by purpose of usage. Unlike most of his contemporaries, who considered cause-effect relationships ill-defined outside the confines of Fisherian experiments, Cochran had no qualm admitting that he sought such relationships in observational studies. He in fact went as far as defining the objective of an observational study: to elucidate causal-and-effect relationships in situations where controlled experiments are infeasible (Cochran, 1965). Indeed, in the paper before us, the word cause is used fairly freely, and other causal terms such as effect, influence, and explanation are almost as frequent as regression or variance. Still, Cochran was well aware that he was dealing with unchartered extra-statistical territory and cautioned us: 1 Narrated in Cochran (1983, p. 24). 1

Claim of proof of cause and effect must carry with it an explanation of the mechanism by which this effect is produced. Today, when an analyst declares that a claim depends on the mechanism by which an effect is produced we expect the analyst to specify what features of the mechanism would make the claim valid. For example, when Rosenbaum and Rubin (1983) claimed that propensity score methods may lead to unbiased estimates of causal effects, they conditioned the claim on a counterfactual assumption named strong ignorability. Such identifying assumptions, though cognitively formidable, provided a formal instrument for proving that some adjustments can yield unbiased estimates. Similarly, when a structural analyst makes the claim that an indirect effect is estimable from observational studies, the claim must follow assumptions about the structure of the underlying graph which, again, assures us of zero-bias estimates (see Pearl (2014b)). Things were quite different in Cochran s era; an appeal to a mechanism, like an appeal to subject matter information stood literally for a confession of helplessness, since mechanisms and causal relationships had no representation in statistics. Structural equation models (SEM), the language used by economists to represent mechanisms, were deeply mistrusted by statisticians, who could not bring themselves to distinguish structural from regression models (Guttman, 1977; Freedman, 1987; Cliff, 1983; Wermuth, 1992; Holland, 1995). 2 Counterfactuals, on the other hand, were still in the embryonic state that Neyman left them in symbols with no model, no formal connection to realizable variables, and no inferential machinery with which to support or refute claims. 3 Fisher s celebrated advice: make your theories elaborate was no help in this transitional era of pre-formal causation; there is no way to elaborate on a theory that cannot be represented in some language. It is not surprising, therefore, that Cochran s conclusions are quite gloomy: It is well known that evidence of a relationship between x and y is no proof that x causes y. The scientific philosophers to whom we might turn for expert guidance on this tricky issue are a disappointment. Almost unanimously and with evident delight they throw the idea of cause and effect overboard. As the statistical study of relationships has become more sophisticated, the statistician might admit, however, that his point of view is not very different, even if he wishes to retain the terms cause and effect. It is likewise not surprising that in the present article, Cochran does not offer readers any advice on which covariates are likely to reduce bias and which would amplify bias. Any such advice, as we know today, requires a picture of reality, which Cochran understood to be both needed and lacking at his time. 4 On the positive side, though, he did have the vision to anticipate the emergence of a new type of research paradigm within statistics, a paradigm centered on mechanisms: 2 This mistrust persists to some degree even in our century, see Berk (2004) or Sobel (2008). 3 These had to wait for Rubin (1974), Robins (1986), and the structural semantics of Balke and Pearl (1994). 4 To the best of my knowledge, the only adjustment-related advice in the entire statistics literature prior to 1980 was Cox s warning that the concomitant observations be quite unaffected by the treatments (Cox, 1958, p. 48); an exceptional defiance of an unwritten taboo against the use of data-generating models. 2

A claim of proof of cause and effect must carry with it an explanation of the mechanism by which the effect is produced. Except in cases where the mechanism is obvious and undisputed, this may require a completely different type of research from the observational study that is being summarized. I believe the type of research we see flourishing today, based on a symbiosis between the graphical and counterfactual languages (Morgan and Winship, 2014; VanderWeele, 2015; Bareinboim and Pearl, 2015) would perfectly meet Cochran s vision of a completely different type of research. This research differs fundamentally from the type of research conducted in Cochran s generation. First, it commences with a commitment to understanding what reality must be like for a statistical routine to succeed and, second, it represents reality in terms of data-generating models (read: mechanisms ), rather than probability distributions. Encoded as nonparametric structural equations, these models have led to a fruitful symbiosis between graphs and counterfactuals and have unified the potential outcome framework of Neyman, Rubin, and Robins with the econometric tradition of Haavelmo, Marschak, and Heckman. In this symbiosis, counterfactuals (potential outcomes) emerge as natural byproducts of structural equations and serve to formally articulate research questions of interest. Graphical models, on the other hand, are used to encode scientific assumptions in a qualitative (i.e., nonparametric) and transparent language and to identify the logical ramifications of these assumptions, in particular their testable implications. 5 A summary of results emerging from this symbiotic methodology is given in Pearl (2014a) and includes complete solutions 6 to several long-standing problem areas, ranging from policy evaluation (Tian and Shpitser, 2010) and selection bias (Bareinboim et al., 2014) to external validity (Bareinboim and Pearl, 2015; Pearl and Bareinboim, 2014) and missing data (Mohan et al., 2013). This development has not met with universal acceptance. Cox and Wermuth (2015), for example, are still reluctant to endorse the tools that this symbiosis has spawned, questioning in essence whether interventions cannot be mathematized. 7 Others regard the symbiosis as unscientific (Rubin, 2008) or less than helpful (Imbens and Rubin, 2015, p. 22), insisting for example that investigators should handle ignorability judgments by unaided intuition. I strongly believe, however, and I say it with a deep sense of responsibility, that future explorations of observational studies will rise above these inertial barriers and take full advantage of the tools that the graphical-counterfactual symbiosis now offers. 5 Note that the potential outcome framework alone does not meet these qualifications. Scientific assumptions must be converted to conditional ignorability statements (Rosenbaum and Rubin, 1983; Imbens and Rubin, 2015) which, being cognitively formidable, escape the scrutiny of plausibility judgment and impede the search for their testable implications. 6 By complete solution I mean a method of producing consistent estimates of (causal) parameters of interests, applicable to any hypothesized model, and accompanied by a proof that no other method can do better except by strengthening the model assumptions. 7 Unwittingly, the very calculus that they reject happens to resolve the problem that they pose ( indirect confounding ) in just four steps (Pearl, 2015a, http://www.mii.ucla.edu/causality/; Pearl, 2015b). 3

References Balke, A. and Pearl, J. (1994). Counterfactual probabilities: Computational methods, bounds, and applications. In Uncertainty in Artificial Intelligence 10 (R. L. de Mantaras and D. Poole, eds.). Morgan Kaufmann, San Mateo, CA, 46 54. Bareinboim, E. and Pearl, J. (2015). Causal inference from big data: Theoretical foundations and the data-fusion problem. Tech. Rep. R-450, <http://ftp.cs.ucla.edu/pub/stat ser/r450.pdf>, Department of Computer Science, University of California, Los Angeles, CA. Forthcoming, Proceedings of the National Academy of Sciences. Bareinboim, E., Tian, J. and Pearl, J. (2014). Recovering from selection bias in causal and statistical inference. In Proceedings of the Twenty-eighth AAAI Conference on Artificial Intelligence (C. E. Brodley and P. Stone, eds.). AAAI Press, Palo Alto, CA. Best Paper Award, <http://ftp.cs.ucla.edu/pub/stat ser/r425.pdf>. Berk, R. (2004). Regression Analysis: A Constructive Critique. Sage, Thousand Oaks, CA. Cliff, N. (1983). Some cautions concerning the application of causal modeling methods. Multivariate Behavioral Research 18 115 126. Cochran, W. (1965). The planning of observational studies of human population. Journal of the Royal Statistical Society (Series A) 128 234 255. Cochran, W. G. (1983). Planning and Analysis of Observational Studies. Wiley, New York. Cox, D. (1958). The Planning of Experiments. John Wiley and Sons, NY. Cox, D. and Wermuth, N. (2015). Design and interpretation of studies: Relevant concepts from the past and some extensions. Observational Studies, to appear. URL http://arxiv.org/pdf/1505.02452v1.pdf Fisher, R. (1922). On the mathematical foundations of theoretical statistics. Philosophical Transactions of the Royal Society of London, Series A 222 311. Freedman, D. (1987). As others see us: A case study in path analysis (with discussion). Journal of Educational Statistics 12 101 223. Guttman (1977). What is not what in statistics. The Statistician 26 81 107. Holland, P. (1995). Some reflections on Freedman s critiques. Foundations of Science 1 50 57. Imbens, G. W. and Rubin, D. B. (2015). Causal Inference for Statistics, Social, and Biomedical Sciences: An Introduction. Cambridge University Press, New York, NY. 4

Mohan, K., Pearl, J. and Tian, J. (2013). Graphical models for inference with missing data. In Advances in Neural Information Processing Systems 26 (C. Burges, L. Bottou, M. Welling, Z. Ghahramani and K. Weinberger, eds.). Curran Associates, Inc., 1277 1285. URL http://papers.nips.cc/paper/4899-graphical-models-for-inference-with-missing-data. Morgan, S. L. and Winship, C. (2014). Counterfactuals and Causal Inference: Methods and Principles for Social Research (Analytical Methods for Social Research. 2nd ed. Cambridge University Press, New York, NY. Pearl, J. (2014a). The deductive approach to causal inference. Journal of Causal Inference 2 115 129. Pearl, J. (2014b). Interpretation and identification of causal mediation. Psychological Methods 19 459 481. Pearl, J. (2015a). Indirect Confounding and Causal Calculus (On three papers by Cox and Wermuth). Blog entry: http://www.mii.ucla.edu/causality/. Pearl, J. (2015b). Indirect Confounding and Causal Calculus (On three papers by Cox and Wermuth). Tech. Rep. R-457, <http://ftp.cs.ucla.edu/pub/stat ser/r457.pdf>, Department of Computer Science, University of California, Los Angeles, CA. Pearl, J. and Bareinboim, E. (2014). External validity: From do-calculus to transportability across populations. Statistical Science 29 579 595. Robins, J. (1986). A new approach to causal inference in mortality studies with a sustained exposure period applications to control of the healthy workers survivor effect. Mathematical Modeling 7 1393 1512. Rosenbaum, P. and Rubin, D. (1983). The central role of propensity score in observational studies for causal effects. Biometrika 70 41 55. Rubin, D. (1974). Estimating causal effects of treatments in randomized and nonrandomized studies. Journal of Educational Psychology 66 688 701. Rubin, D. (2008). Author s reply (to Ian Shrier s Letter to the Editor). Statistics in Medicine 27 2741 2742. Salsburg, D. (2002). The Lady Tasting Tea: How Statistics Revolutionized Science in the Twentieth Century. Henry Holt and Company, LLC, New York, NY. Sobel, M. (2008). Identification of causal parameters in randomized studies with mediating variables. Journal of Educational and Behavioral Statistics 33 230 231. Tian, J. and Shpitser, I. (2010). On identifying causal effects. In Heuristics, Probability and Causality: A Tribute to Judea Pearl (R. Dechter, H. Geffner and J. Halpern, eds.). College Publications, UK, 415 444. VanderWeele, T. (2015). Explanation in Causal Inference: Methods for Mediation and Interaction. Oxford University Press, New York, NY. 5

Wermuth, N. (1992). On block-recursive regression equations. Brazilian Journal of Probability and Statistics (with discussion) 6 1 56. 6