Comments on Limits of Econometrics by David Freedman. Arnold Zellner. University of Chicago

Similar documents
Philosophy and objectives of econometrics $

Response to the ASA s statement on p-values: context, process, and purpose

Audio: In this lecture we are going to address psychology as a science. Slide #2

Bayesian and Non-Bayesian Approaches to Scientific Modeling and Inference in Economics and Econometrics by Arnold Zellner* U. of Chicago.

Macroeconometric Analysis. Chapter 1. Introduction

Bayesian Inference Bayes Laplace

Milton Friedman: A Bayesian?* Gerald P. Dwyer Clemson University University of Carlos III, Madrid

Professor Deborah G. Mayo

Principles of Sociology

Chapter 1 Social Science and Its Methods

BAYESIAN ESTIMATORS OF THE LOCATION PARAMETER OF THE NORMAL DISTRIBUTION WITH UNKNOWN VARIANCE

2 Research Methodology 1.2 OBJECTIVES OF RESEARCH The purpose of research is to discover answers to questions through the application of scientific pr

EC Macroeconomic Theory Supplement 13 Professor Sanjay Chugh

Introduction to Bayesian Analysis 1

Thoughts on Social Design

The Logic of Causal Inference

DEVELOPING THE RESEARCH FRAMEWORK Dr. Noly M. Mascariñas

COMP 516 Research Methods in Computer Science. COMP 516 Research Methods in Computer Science. Research Process Models: Sequential (1)

Nature of Science and Scientific Method Guided Notes

The Century of Bayes

Daniel Little University of Michigan-Dearborn www-personal.umd.umich.edu/~delittle/

Cognitive domain: Comprehension Answer location: Elements of Empiricism Question type: MC

Research and science: Qualitative methods

Cult of Statistical Significance 8/03/

6. A theory that has been substantially verified is sometimes called a a. law. b. model.

Table of Contents. Chapter 1 Theoretical Criminology: An Introductory Overview [page 79] Chapter 3 Biosocial Theories of Crime [page 99]

Definitions of Nature of Science and Scientific Inquiry that Guide Project ICAN: A Cheat Sheet

Propensity Score Analysis Shenyang Guo, Ph.D.

Who, among the following, is the writer of Business Research Methods? Which of the folowing is the basis of the Scientific Method?

Evolutionary Approach to Investigations of Cognitive Systems

Modeling and Environmental Science: In Conclusion

Game Theoretic Statistics

INTERVIEWS II: THEORIES AND TECHNIQUES 5. CLINICAL APPROACH TO INTERVIEWING PART 1

HPSC30023 Science and Society

Book Information Jakob Hohwy, The Predictive Mind, Oxford: Oxford University Press, 2013, ix+288, 60.00,

Bill Wilson. Categorizing Cognition: Toward Conceptual Coherence in the Foundations of Psychology

CONCEPTUAL FRAMEWORK, EPISTEMOLOGY, PARADIGM, &THEORETICAL FRAMEWORK

How to use the Lafayette ESS Report to obtain a probability of deception or truth-telling

Social Change in the 21st Century

METHODOLOGY FOR DISSERTATION

The Scientific Enterprise. The Scientific Enterprise,

Author's response to reviews

Bayesian and Frequentist Approaches

Chapter 02 Developing and Evaluating Theories of Behavior

Part 1 Three Stylized Facts on UWSEs Modernization: Depoliticization, Resilience and Sustainability

Beyond Subjective and Objective in Statistics

MBA SEMESTER III. MB0050 Research Methodology- 4 Credits. (Book ID: B1206 ) Assignment Set- 1 (60 Marks)

Econ 270: Theoretical Modeling 1

Methodological Thinking on Disciplinary Research Fields of Theories of Sports Training from the Philosophical Perspective Bin LONG

Choose an approach for your research problem

The role of theory in construction management: a call for debate

UNLOCKING VALUE WITH DATA SCIENCE BAYES APPROACH: MAKING DATA WORK HARDER

Glossary of Research Terms Compiled by Dr Emma Rowden and David Litting (UTS Library)

Speaker Notes: Qualitative Comparative Analysis (QCA) in Implementation Studies

Bayesian Confidence Intervals for Means and Variances of Lognormal and Bivariate Lognormal Distributions

In addition, a working knowledge of Matlab programming language is required to perform well in the course.

Science is a way of learning about the natural world by observing things, asking questions, proposing answers, and testing those answers.

PLANNING THE RESEARCH PROJECT

Qualitative Data Analysis. Richard Boateng, PhD. Arguments with Qualitative Data. Office: UGBS RT18 (rooftop)

This exam consists of three parts. Provide answers to ALL THREE sections.

PSYCHOLOGISM VS. CONVENTIONAL VALUES by L.A. Boland

INTRODUCTION. Evidence standards for justifiable evidence claims, June 2016

Chapter-2 RESEARCH DESIGN

Whose psychological concepts?

Book review of Herbert I. Weisberg: Bias and Causation, Models and Judgment for Valid Comparisons Reviewed by Judea Pearl

Behavioural Economics University of Oxford Vincent P. Crawford Michaelmas Term 2012

[1] provides a philosophical introduction to the subject. Simon [21] discusses numerous topics in economics; see [2] for a broad economic survey.

Lecture 9 Internal Validity

Christ-Centered Critical Thinking. Thinking in a Discipline

Empirical Validation in Agent-Based Models

The Importance of Jeffreys s Legacy (Comment on Robert, Chopin, and Rousseau)

TEACHING BAYESIAN METHODS FOR EXPERIMENTAL DATA ANALYSIS

SYMSYS 130: Research Methods in the Cognitive and Information Sciences (Spring 2013)

T. E. Raghunathan Department of Biostatistics and Research Professor, Institute for Social Research, University of Michigan, Ann Arbor, USA

PSYCHOLOGICAL CONSCIOUSNESS AND PHENOMENAL CONSCIOUSNESS. Overview

Cognitive & Linguistic Sciences. What is cognitive science anyway? Why is it interdisciplinary? Why do we need to learn about information processors?

1. Between White and Black: The Faces of American Institutions in the Ghetto

What is analytical sociology? And is it the future of sociology?

Durkheim. Durkheim s fundamental task in Rules of the Sociological Method is to lay out

Models of Information Retrieval

Do Human Science. Yutaka Saeki

Do Old Fallacies Ever Die? by Milton Friedman Journal of Economic Literature 30, December 1992, pp American Economic Association

For general queries, contact

Jake Bowers Wednesdays, 2-4pm 6648 Haven Hall ( ) CPS Phone is

Critical dilemmas in the methodology of economics facing the crisis

Write your identification number on each paper and cover sheet (the number stated in the upper right hand corner on your exam cover).

August 29, Introduction and Overview

Formulation of Research Design

Famous Monsters (in Statistics) Erum Marfani Frederick Community College Frederick, MD

III. WHAT ANSWERS DO YOU EXPECT?

Crossing boundaries between disciplines: A perspective on Basil Bernstein s legacy

Critical review (Newsletter for Center for Qualitative Methodology) concerning:

PARADIGMS, THEORY AND SOCIAL RESEARCH

Models of Causality and Explanation in Economics

Chapter 2 Epidemiology

Handout on Perfect Bayesian Equilibrium

Causation, the structural engineer, and the expert witness

Social Studies 4 8 (118)

Transcription:

Comments on Limits of Econometrics by David Freedman Arnold Zellner University of Chicago David Freedman s impressive paper reveals well his deep understanding of not only statistical techniques and their uses but also of scientific methodology and its philosophy. On visits to Berkeley over the years, I have had the opportunity to talk with David and to learn much from our friendly and constructive conversations. He was a great member of the intellectual community and his passing is a huge loss to us all. In particular, I would have enjoyed discussing his paper, Limits of Econometrics with him, especially his final sentence, The goal of empirical research is---or should be---to increase our understanding of the phenomena, rather than displaying our mastery of technique. Also, in his wide ranging discussion, he has identified crucial obstacles to progress in several areas of social science research, including econometrics. The overall message conveyed reflects his fruitful and productive life. It questions commonly accepted methodologies and makes a plea for the development of alternatives more solidly grounded in data, appropriate statistical methods and scientific methodology. In particular, Freedman has stressed the difficulties of verifying many assumptions made in regression analyses and the impact that incorrect assumptions can have on the validity of conclusions, one general important theme of his research. In this regard, I can identify deeply with his point of view since I too have been engaged in developing sound alternatives to conventional methods that are often based on weak foundations and lead to unsatisfactory results. In what follows, I shall attempt to respond to some of the fundamental issues raised by David by discussing a number of principles and procedures that I believe are important in promoting progress in understanding, predicting and possibly controlling phenomena in all areas of research. First, note that Karl Pearson s (1938) Unity of Science Principle, namely that any area of study, be it physics, astronomy, economics, political science, etc., can be scientific if scientific methods are employed in producing, analyzing and learning from data. Now it may be asked, Where can one learn about the methodology of science? In this regard, I, and many others, recommend highly Sir Harold Jeffreys (1998) classic book, Theory of Probability, that first appeared in 1939 with many subsequent editions, the last appearing in the Oxford U. Press Classics Series, 1998. Note that the well known mathematician and statistician, I. J. Good wrote, In my review (Good, 1962) I said that Jeffreys book on probability is of greater importance for the philosophy of science and obviously of greater immediate practical importance, than nearly all the books on probability written by professional philosophers lumped together. I believe that this is still true, though more professional philosophers have woken up. (quote from Zellner (1989:32); see also reviews of Jeffreys work by S. Geisser (1989) and D.V. Lindley (1989), in this volume). Also, the recent paper, Harold Jeffreys Theory of Probability Revisited, by Robert et al. (2008) to appear in Statistical Science with discussion presents a chapter by chapter review of Jeffreys book and concludes that it is indeed a classic work. Note that Jeffreys, along with many other physical scientist use the term probability theory rather than statistics to denote the area of work in which measurements are made and used to describe phenomena and to evaluate old and new models that may be useful in description, explanation, prediction, control, etc. 28

International Econometric Review (IER) because they do not like the word statistics according to some because Lord Rutherford, a famous physicist, is remembered as saying, if you need statistics to analyze your data, redesign your experiment. Briefly, Jeffreys work is fundamentally important because it was undertaken to provide a fruitful framework for scientific research not only for the physical sciences but also for all areas of study, including the biological sciences and social sciences, about which David Freedman is so concerned. Some points that emerge from Jeffreys and others works that appear very important are: 1. Formal Learning Model: In analyzing and learning from data, be it regression or other data, it is important to use a good, explicit learning model. As has been pointed out many times in the literature, many researchers do not use a formal learning model and thus learn informally. Bayesians use, and sometimes misuse, Bayes Theorem as a formal learning model that has been logically justified in several different ways and shown to work well in many estimation, testing, prediction and other studies; see for example the many empirical examples in Jeffreys book and many Bayesian texts (a listing of texts is available at http://www.bayesian.org, the homepage of the International Society for Bayesian Analysis). 2. Simplicity Postulate: Simpler models will probably work better in explanation and prediction. See Jeffreys (1998:4 and 47) and Zellner et al. (2001) for discussions of the concept of simplicity, measures of simplicity, and the results of a poll of Nobel Economic Prize winners that showed that almost all of them favored a sophisticatedly simple approach in modeling rather than complicated approaches. Note that it is important the models be sophisticatedly simple because some simple models are stupid. As regards examples of relatively simple models that have worked well in economics and econometrics, there are Marshallian demand and supply models that have been widely used world-wide for many years to analyze the operation of product and factor markets. Also there is the recent formulation of Marshallian Macroeconomic Models, implemented with disaggregated data, that are a synthesis of relatively simple disaggregated Marshallian industrial models and elements of macroeconomic theory to form large, but relatively simple, models that explain and predict fairly well and can be used for analyzing alternative policies; see, e.g. Zellner and Chen (2001), Zellner and Israilevich (2005) and Ngoie and Zellner (2008). 3. Predictive testing of models is central in science, as is well known but not implemented enough in empirical work. Also, as noted in my invited review paper on concepts of causality, published in Zellner (1979), eds. K. Brunner and A.H. Meltzer, the famous philosopher of science, Herbert Feigl (1953) presented a critical review of concepts of causality in the literature and concluded that the following definition is most appropriate and useful: The clarified (purified) concept of causation is defined in terms of predictability according to a law (or more adequately according to a set of laws). (p. 408). In view of the central role of prediction in the definition of causation and in many other areas, it is regrettable that many statistical analysts do not check the predictive performance of their models. Indeed in terms of the exchange between Keynes and Tinbergen on econometric models performance that David Freedman discussed in his paper, it is noteworthy that Christ (1951) did compare and discuss the predictive performance of one of Klein s macro econometric models vis a vis that of a 29

benchmark random walk model and found that it did not perform very well and thus, as Milton Friedman noted in his discussion, needed reformulation. 4. Since laws are fundamental in defining causality and in other respects, as David Freedman recognizes, a fundamental issue is how do we produce laws, e.g., Newton s and Einstein s laws, Marshall s laws of demand and supply, etc.? In my work over the years, I have suggested, in accord with Hadamard s (1945) findings, that generation of more unusual and surprising facts, some contradicting old laws and beliefs, will probably inspire many to generate new theories to explain the unusual facts that can then be tested with much new data. For further discussion of this process and information about eight ways to produce unusual facts, see Zellner (1984:9-10). 5. One class of unusual facts that Franz Palm and I have emphasized in our theoretical and empirical work, summarized in Zellner and Palm (2004) is that often some purely statistical univariate and multivariate time series models, e.g. Box-Jenkins ARIMA or transfer function models or variants of statistical multivariate time series models forecast future outcomes, e.g. the next period real GDP growth rate, reasonably well. In our structural econometric, time series analysis (SEMTSA) approach, we use economic theory to formulate causal multivariate dynamic economic models that mathematically imply the statistical model that forecasts reasonably well. Thus, using our methodology, we produce causal models, e.g. our Marshallian Model, mentioned above. Or if we have a causal time series econometric model, we check it using time series model identification and diagnostic testing procedures in efforts to improve its performance in explanation, prediction and policy-making. See Zellner and Palm (2004) for further discussion of these procedures and applications. 6. Last, there is the issue of which statistical inference procedures to use in estimation, testing, prediction, model combining, decision-making, etc. Over the years, much of my research has been directed at evaluating the comparative performance of Bayesian and non-bayesian procedures and I have concluded that Bayesian procedures have performed better. This is not to say that Bayesian procedures are perfect and indeed in Zellner (1988) a new information theoretic derivation of Bayes theorem and discussions of it by Jaynes, Kullback, Hill and Bernardo, are provided. In more recent research, using the approach in Zellner (1988) that produced Bayes theorem, by varying the conditions of the optimization problem, new, optimal learning models have been derived. By having a broader range of optimal learning models on the self, some that do not require use of a prior distribution, statisticians and other scientists will become more effective in learning from data and solving inference and decision problems. 7. In summary, thanks to David Freedman for giving us such a stimulating paper, his life s work and being a wonderful person. REFERENCES Christ, C.F. (1951). A Test of an Econometric Model for the United States, 1921-1947. In Conference on Business Cycles with comments by Milton Friedman. New York: National Bureau of Economic Research, 35-107. 30

International Econometric Review (IER) Feigl, H. (1953), Notes on Causality. In Readings in the Philosophy of Science, ed. H. Feigl and M. Brodbeck. New York: Appleton-Century-Crofts, 408-418. Geisser, S. (1989), The Contributions of Sir Harold Jeffreys to Bayesian Inference. In Bayesian Analysis in Econometrics and Statistics: Essays in Honor of Harold Jeffreys, ed. A. Zellner, reprint of 1980 edition. North-Holland: Amsterdam, 13-20. Good, I.J. (1962). Theory of Probability by Harold Jeffreys. Journal of the Royal Statistical Society A, 125, 487 489. Hadamard, J. (1945). The psychology of invention in the mathematical field. New York, NY: Dover Publications Jeffreys, H. (1998). Theory of Probability. In Oxford Classic Texts in the Physical Sciences, first published in 1939 with subsequent editions in 1948, 1961, 1981 and 1983. Oxford: Oxford U. Press. Lindley, D.V. (1989), Jeffreys s Contribution to Modern Statistical Thought. In Bayesian Analysis in Econometrics and Statistics: Essays in Honor of Harold Jeffreys, ed. A. Zellner, reprint of 1980 edition. North-Holland: Amsterdam, 35-39. Ngoie, J.K. and A. Zellner (2008). The Effects of Freedom Reforms on the Growth Rate of the South African Economy. Working Paper, H.G.B. Alexander Research Foundation, U. of Chicago. Pearson, K. (1938). The Grammar of Science. London: Everyman Edition. Robert, C.P., Chopin, N. and J. Rousseau (2008). Harold Jeffreys Theory of Probability Revisited. Forthcoming in Statistical Science with invited discussion arxiv:0804.3173v6. Zellner, A. (1979). Causality and Econometrics. In Three Aspects of Policy and Policymaking: Knowledge, Data and Institutions, eds. K. Brunner and A.H. Meltzer. Amsterdam: North-Holland Publishing Co, 9-54. Zellner, A. (1984). Basic Issues in Econometrics. Chicago: U. of Chicago Press. Zellner, A. (1988). Optimal Information Processing and Bayes Theorem. American Statistician, 42, 278-284 with discussion by E.T. Jaynes, B.M. Hill, J.M. Bernardo and S. Kullback and the author s response. Zellner, A. (1989). Bayesian Analysis in Econometrics and Statistics: Essays in Honor of Harold Jeffreys. Reprint of 1980 edition, North-Holland: Amsterdam. Zellner, A., Kuezenkamp, H. and M. McAleer (2001). Simplicity, Inference and Modeling: Keeping it Sophisticatedly Simple. Cambridge: Cambridge U. Press. Zellner, A. and Chen, B. (2001), Bayesian Modeling of Economies and Data Requirements. Macroeconomic Dynamics, 5, 673-700. 31

Zellner, A. and F.C. Palm (2004). The Structural Econometric Modeling. Time Series Analysis (SEMTSA) Approach, Cambridge: Cambridge U. Press. Zellner, A. and G. Israilevich (2005). The Marshallian Macroeconomic Model: A Progress Report. Macroeconomic Dynamics, 9, 220-243 and reprinted in International Journal of Forecasting, 21, 627-645 with discussion by A. Espasa, Carlos III U. of Madrid. 32