Practical Regression: Convincing Empirical Research in Ten Steps

Similar documents
Carrying out an Empirical Project

Session 3: Dealing with Reverse Causality

August 29, Introduction and Overview

Session 1: Dealing with Endogeneity

Economics 2010a. Fall Lecture 11. Edward L. Glaeser

Technical Track Session IV Instrumental Variables

Introduction to Econometrics

Causality and Statistical Learning

Assignment 4: True or Quasi-Experiment

The Logic of Data Analysis Using Statistical Techniques M. E. Swisher, 2016

7 Statistical Issues that Researchers Shouldn t Worry (So Much) About

Chapter 23. Inference About Means. Copyright 2010 Pearson Education, Inc.

What Constitutes a Good Contribution to the Literature (Body of Knowledge)?

UNIT 4 ALGEBRA II TEMPLATE CREATED BY REGION 1 ESA UNIT 4

Endogeneity is a fancy word for a simple problem. So fancy, in fact, that the Microsoft Word spell-checker does not recognize it.

Data and Statistics 101: Key Concepts in the Collection, Analysis, and Application of Child Welfare Data

Citation for published version (APA): Ebbes, P. (2004). Latent instrumental variables: a new approach to solve for endogeneity s.n.

Econ 270: Theoretical Modeling 1

Objectives. Quantifying the quality of hypothesis tests. Type I and II errors. Power of a test. Cautions about significance tests

15.301/310, Managerial Psychology Prof. Dan Ariely Recitation 8: T test and ANOVA

multilevel modeling for social and personality psychology

Chapter 11. Experimental Design: One-Way Independent Samples Design

1 Simple and Multiple Linear Regression Assumptions

Empirical Tools of Public Finance. 131 Undergraduate Public Economics Emmanuel Saez UC Berkeley

Regression Discontinuity Design (RDD)

WELCOME! Lecture 11 Thommy Perlinger

How to Communicate Effectively and Handle Difficult People, 2 nd Edition. Ten Tips for Presentation Confidence and Reducing Nervousness

Political Science 15, Winter 2014 Final Review

Unit 1 Exploring and Understanding Data

A NON-TECHNICAL INTRODUCTION TO REGRESSIONS. David Romer. University of California, Berkeley. January Copyright 2018 by David Romer

Chapter 1 Introduction to Educational Research

Stepwise method Modern Model Selection Methods Quantile-Quantile plot and tests for normality

Chapter 18: Categorical data

Statistical reports Regression, 2010

The Pretest! Pretest! Pretest! Assignment (Example 2)

Confidence Intervals On Subsets May Be Misleading

Write your identification number on each paper and cover sheet (the number stated in the upper right hand corner on your exam cover).

How is ethics like logistic regression? Ethics decisions, like statistical inferences, are informative only if they re not too easy or too hard 1

Instrumental Variables Estimation: An Introduction

ISC- GRADE XI HUMANITIES ( ) PSYCHOLOGY. Chapter 2- Methods of Psychology

Qualitative Data Analysis. Richard Boateng, PhD. Arguments with Qualitative Data. Office: UGBS RT18 (rooftop)

Propensity Score Methods for Estimating Causality in the Absence of Random Assignment: Applications for Child Care Policy Research

Planning for a time when you cannot make decisions for yourself

Module 14: Missing Data Concepts

Final Exam - section 2. Thursday, December hours, 30 minutes

Biostatistics and Design of Experiments Prof. Mukesh Doble Department of Biotechnology Indian Institute of Technology, Madras

ORIENTATION SAN FRANCISCO STOP SMOKING PROGRAM

CHAPTER - 6 STATISTICAL ANALYSIS. This chapter discusses inferential statistics, which use sample data to

STA Module 9 Confidence Intervals for One Population Mean

The Wellbeing Course. Resource: Mental Skills. The Wellbeing Course was written by Professor Nick Titov and Dr Blake Dear

EC352 Econometric Methods: Week 07

Clever Hans the horse could do simple math and spell out the answers to simple questions. He wasn t always correct, but he was most of the time.

The Assisted Decision-Making (Capacity) Act 2015 and the Decision Support Service

Pitfalls in Linear Regression Analysis

Introduction to Behavioral Economics Like the subject matter of behavioral economics, this course is divided into two parts:

Chapter 5: Field experimental designs in agriculture

Measuring and Assessing Study Quality

plural noun 1. a system of moral principles: the ethics of a culture. 2. the rules of conduct recognized in respect to a particular group, culture,

Student Performance Q&A:

Correlation and Regression

Handout on Perfect Bayesian Equilibrium

Theory Building and Hypothesis Testing. POLI 205 Doing Research in Politics. Theory. Building. Hypotheses. Testing. Fall 2015

Preliminary Report on Simple Statistical Tests (t-tests and bivariate correlations)

PEER REVIEW HISTORY ARTICLE DETAILS VERSION 1 - REVIEW. Ball State University

Chapter Eight: Multivariate Analysis

Causal Research Design- Experimentation

I ll Do it Tomorrow. READTHEORY Name Date

Public Opinion Survey on Tobacco Use in Outdoor Dining Areas Survey Specifications and Training Guide

Selected Topics in Biostatistics Seminar Series. Missing Data. Sponsored by: Center For Clinical Investigation and Cleveland CTSC

MULTIPLE LINEAR REGRESSION 24.1 INTRODUCTION AND OBJECTIVES OBJECTIVES

t-test for r Copyright 2000 Tom Malloy. All rights reserved

handouts for women 1. Self-test for depression symptoms in pregnancy and postpartum Edinburgh postnatal depression scale (epds) 2

Understanding. Regression Analysis

Interviewer: Tell us about the workshops you taught on Self-Determination.

Problem Set 3 ECN Econometrics Professor Oscar Jorda. Name. ESSAY. Write your answer in the space provided.

Health Insight: A Consumer s Guide to Taking Charge of Health Information

observational studies Descriptive studies

Chapter 11: Advanced Remedial Measures. Weighted Least Squares (WLS)

MULTIPLE CHOICE. Choose the one alternative that best completes the statement or answers the question.

How Does Analysis of Competing Hypotheses (ACH) Improve Intelligence Analysis?

AP STATISTICS 2014 SCORING GUIDELINES

These methods have been explained in partial or complex ways in the original Mind Reading book and our rare Risk Assessment book.

We admitted that we were powerless over alcohol that our lives had become unmanageable.

Why Coaching Clients Give Up

We admitted that we were powerless over alcohol that our lives had become unmanageable. Alcoholics Anonymous (AA) (2001, p. 59)

The Five Types of Fear

NEXT. Top 12 Myths about OEE

5 $3 billion per disease

The Perils of Empirical Work on Institutions

HEPATITIS C LESSONS PART 4

How to Conduct an Unemployment Benefits Hearing

A Brief Guide to Writing

Sawtooth Software. The Number of Levels Effect in Conjoint: Where Does It Come From and Can It Be Eliminated? RESEARCH PAPER SERIES

Chapter 19: Categorical outcomes: chi-square and loglinear analysis

Regression Discontinuity Analysis

Doing High Quality Field Research. Kim Elsbach University of California, Davis

You re listening to an audio module from BMJ Learning. Hallo. I'm Anna Sayburn, Senior Editor with the BMJ Group s Consumer Health Team.

Chapter 02 Developing and Evaluating Theories of Behavior

WRITTEN PRELIMINARY Ph.D. EXAMINATION. Department of Applied Economics. January 17, Consumer Behavior and Household Economics.

Transcription:

DAVID DRANOVE 7-112-001 Practical Regression: Convincing Empirical Research in Ten Steps This is one in a series of notes entitled Practical Regression. These notes supplement the theoretical content of most statistics texts with practical advice on solving real world empirical problems through regression analysis. Most of the Practical Regression series centers on regression analysis and related tools. The focus on regression is entirely appropriate. Most interesting strategy questions involve relational analysis, that is, the exploration of how two or more factors relate to each other. The most important tool for relational analysis is regression. Regression encompasses almost every major form of data analysis. Here is a short list of analytic tools that are closely related to regression: Correlation (a simple analysis of the relationship between two variables) Comparisons of means across two different populations (basically a regression with dummy variables ) ANOVA (the decomposition of the sources of variation in a variable) Conjoint analysis (a tool used in marketing research) Regression analysis has its own language. Here are some terms to remember: The variable on the left hand side of a regression is called the dependent variable, Y variable, or LHS variable. The variables on the right hand side of a regression are called the predictors, independent variables, X variables, or RHS variables. Each row of data is an observation. 2012 by the Kellogg School of Management, Northwestern University. This technical note was prepared by Professor David Dranove. Technical notes are developed solely as the basis for class discussion. Technical notes are not intended to serve as endorsements, sources of primary data, or illustrations of effective or ineffective management. To order copies or request permission to reproduce materials, call 847-491-5400 or e-mail cases@kellogg.northwestern.edu. No part of this publication may be reproduced, stored in a retrieval system, used in a spreadsheet, or transmitted in any form or by any means electronic, mechanical, photocopying, recording, or otherwise without the permission of the Kellogg School of Management.

TECHNICAL NOTE: EMPIRICAL STEPS 7-112-001 The Gold Standard for Regression: Convincing Results The highest praise one can give to an empirical research project is to say that the findings are convincing. Convincing research has at least three characteristics: The reported effects (i.e., the regression coefficients) are unbiased. There should be no reason to suspect that the reported effects are systematically larger or smaller than the real world effects under investigation. Unbiasedness depends on model and variable selection. Important note about bias: If an estimate is free from bias, then it is correct on average. In other words, if we repeat our experiments and statistical analyses many times, the average of our estimates will converge toward the true effect we are studying. But the estimate we obtain from the one regression we do perform can still be wildly inaccurate much too large or much too small. This is why freedom from bias is only one of our goals. Precision is another, which we evaluate by examining standard errors. The standard errors are computed correctly and are small enough to generate precise estimates of the parameters of interest. Computation of standard errors under ordinary least squares requires several strong assumptions. Among these, we assume that the errors are independent across observations. We also assume that the variance of the errors is completely random (i.e., the errors are homoscedastic ). If these assumptions are violated, you will need to modify how you compute the standard errors. The findings are robust. Researchers are often uncertain about the best way to specify their models. Your findings are robust if they hold up under any plausible specification of the model. The Ten Steps The following steps can help prevent the most serious problems that emerge in empirical research and go a long way toward ensuring that your findings are convincing. 1. THINK ABOUT A MODEL. A well-structured question involves some hypothesized relationships between a dependent variable and causal factors. State the hypothesized relationships as specifically as possible. Are they theoretically plausible? Your statistical model may be in the form of a simple question, as the following examples illustrate. In each case, there is a research question (q) and a corresponding statistical model (m). (1q) Does a change in the value of the Euro affect the prices charged for imported beers in the United States? (1m) Price of imported beer = f(value of Euro, other predictors) (2q) Do larger firms have higher healthcare costs per employee than do smaller firms? (2m) Healthcare expenditures per employee = f(firm size, other predictors) 2 KELLOGG SCHOOL OF MANAGEMENT

7-112-001 TECHNICAL NOTE: EMPIRICAL STEPS (3q) Do late adopters of technology tend to use the same brands as early adopters? (3m) Choice of brand = f(time of adoption of technology, other predictors) In each case, the model is expressed in the form Y = f(x), where Y is the dependent variable and X represents a vector of independent variables (including the main predictor variables of interest and control variables). Each model has a testable hypothesis, namely that one or more of the variables in X has a posited effect on Y. (We do not have to posit the effects of control variables included in X; they are included to improve precision and help avoid omitted variable bias. ) 2. THINK ABOUT CAUSALITY AND MAKE SURE THAT IT IS UNAMBIGUOUS. Your goal is to determine how changing a factor under management control X will affect some outcome Y. A regression coefficient β x tells you if X is correlated with Y, holding constant all other variables in the regression. But it is not sufficient to learn the value of β x. Your model is not identified unless you can unambiguously determine the direction of causality. There are several reasons why two variables X and Y may be correlated in the data: X causes Y, as implicitly hypothesized by the model. If so, then changing X would definitely cause Y to change. Y causes X (reverse causality). Some third variable Z causes both X and Y to move in lockstep. In all three cases, the regression coefficient is positive. But only in the first case would changing X cause Y to change. If the second or third cases are possible, the model is not identified and you cannot determine if and how changing X will affect Y. Assuring that you can state the direction of causality is known as the identification problem. As this discussion suggests, there are two types of identification problems. Confusion about whether X causes Y or Y causes X is called simultaneity bias. If some factor Z causes both X and Y to move in lockstep, then failure to include Z in the regression induces an omitted variable bias. These problems are closely related (the underlying mathematics is nearly identical) and econometricians often use the term endogeneity bias to refer to either one of them. Other notes in the Practical Regression series discuss endogeneity bias in more detail and offer potential solutions. The Importance of Storytelling The famous econometrician W. G. Cochran once wrote: When summarizing the results of a study that shows an association consistent with a causal hypothesis, the investigator should always list and discuss all alternative KELLOGG SCHOOL OF MANAGEMENT 3

TECHNICAL NOTE: EMPIRICAL STEPS 7-112-001 explanations of his results (including different hypotheses and biases in the results) that occur to him. 1 Cochran is discussing the identification problem. We call the process of listing and discussing alternative explanations storytelling. This is an apt term because solving the identification problem often begins by telling stories that could explain the association between X and Y. For example, suppose you want to evaluate whether firms should increase CEO compensation as a way to promote performance. You observe that when CEOs are paid more, their firms enjoy higher returns on assets (ROAs). In other words, ρ(roa, CEO pay) > 0. 2 This correlation supports the view that higher compensation causes better performance, but it is not definitive. To determine whether the direction of causality is identified, you should try to tell another story that might explain the observed correlation. Here are two possibilities: Many corporate boards reward CEOs when their firms do well. In this case, causality runs from performance to compensation, not vice versa. (Alternatively, you could note that many firms have incentive-laden CEO compensation contracts and that the use of such incentives causes both high pay and good performance.) In some industries, large firms tend to outperform smaller firms. Large firms also tend to pay their CEOs more (perhaps because it is harder work). Thus, firm size causes both high pay and good performance. It is not necessary to be definitive about these alternatives. Any or all of them are enough to cause doubts about causality; sorting them out requires similar empirical approaches. Remember, good storytellers have a grasp of the institutional facts about the phenomena they are studying as well as economic theories. They understand the possible relationships among variables and can clearly articulate possible directions of causality. Causal Inference Econometrician Richard Berk makes the following critically important observation: Credible causal inference cannot be drawn from regression alone. He goes on: Coefficients do not represent the causal importance of a variable. A good overall fit does not demonstrate that a causal model is correct. Moreover, there are no regression diagnostics through which causal effects can be demonstrated with confidence... no mathematical formalisms through which causal effects can be demonstrated.... In the absence of a well-supported model (i.e., hypothesized relationships based on prior research and theory), the potential problems with regression-based causal models outweigh existing remedial strategies. 3 1 W. G. Cochran, Summarizing the Results of a Series of Experiments, 80-2, 21-33. Durham, NC, Proceedings of the 25th Conference on the Design of Experiments in Army Research Development and Testing, U.S. Army Research Office, 1980. 2 The symbol ρ stands for correlation. 3 Richard A. Berk, Regression Analysis: A Constructive Critique (Thousand Oaks, CA: Sage Publications, 2004). 4 KELLOGG SCHOOL OF MANAGEMENT

7-112-001 TECHNICAL NOTE: EMPIRICAL STEPS Mastery of regression requires learning a lot of quantitative methods. But your most important task as a statistical researcher is not to memorize the mathematics (though that is important). The most important task is to sort out causality through prior research, theory (storytelling), and learning what methods are most appropriate for the identification problem that you face. 3. OBTAIN, ORGANIZE, AND CLEAN THE DATA. It is ideal to think about the underlying model before you obtain your data. Otherwise you may convince yourself to rule out certain theoretical possibilities merely because you lack the data to test them. The result may be confirmation bias, as you convince yourself that whatever analysis you end up running is the correct analysis. (For example, say you are unsure whether income or education affects test scores. You only have data on income, so you run your model without education. Income proves to be a significant predictor, so you convince yourself that income matters without ever testing the alternative.) Data come from many sources government websites, consulting firms selling proprietary surveys, market research firms selling scanner data, and so on. Most statistical software packages permit you to merge data from different sources. It is essential that you can link the data through a variable common to both data sets. It helps (but is not essential) if the link is numerical, such as the CUSIP code for firms or census codes for U.S. geographic regions. Once you have the data, you need to clean it. This means doing the best job you can to assure that there are no errors in the data or that data that made sense for someone else to use is ready for you to use. For example, one of my data sets includes the zip codes of consumers. U.S. zip codes range from 00001 to 99999. I was dismayed to find zip codes in my data set as large as 998750135. After giving the matter some thought, I realized that the bad zip codes were actually nine-digit zips. I recovered the five-digit zip codes by lopping off the last four digits. This is an example of dirty data. Dirty data is not ready for analysis because the values are incorrect or inappropriate. You must judge whether your data is too dirty to analyze and decide on the best way to clean it. So get to know your data. Here are some steps you can follow. (1) Print out a distribution; look at the minimum, maximum, and modal values. Do these make sense? (2) Examine outliers. (Outliers are extreme values of variables, or in the case of regressions, observations that have extreme values of residual.) Determine if the outliers are plausible. For example, a distribution of population by U.S. county should have an enormous range; a distribution of per capita income by county should not. (3) Pick some variables that you know from theory should be correlated with each other. Are they? If not, try to determine what went wrong. Perhaps one enormous outlier is wreaking havoc with the correlation coefficient. If the data look dirty, feel dirty, or even smell dirty, they probably are dirty. Clean them. Cleaning data may require judgment calls. If you have reason to believe that a value is incorrect and you think you know the correct value, then correct it. The zip code example is a nobrainer. Here is an example of an obvious typo. You have pricing data for ice cream that shows the prices range from $2.50 to $5.00, except for one firm, which has prices of $350 to $400. It KELLOGG SCHOOL OF MANAGEMENT 5

TECHNICAL NOTE: EMPIRICAL STEPS 7-112-001 seems certain that the outlier prices have been entered in pennies rather than dollars. Once again, it is okay to rely on your judgment, change the values, and move on. But it isn t always so easy to clean dirty data. If the price of the ice cream was $375732, you would be hard pressed to figure out the correct price. What should you do when you are uncertain about an outlier? One thing to bear in mind is that ordinary least squares (OLS) regression minimizes the squared residuals. An outlier dependent variable can have a very large squared residual, so the computer will work extra hard to try to explain this outlier. It follows that one bad value of a dependent variable can profoundly influence your analysis. The same is true to a lesser extent of predictor variables. As important as it is to weed out incorrect outliers, it is equally important not to weed out correct values that happen to be out of line with the rest of your data. One good but unusual data point may be important to your analysis. To summarize: It is sometimes easy to recognize data entry errors. By all means, go ahead and make the changes when you can. Sometimes you will be unable to figure out whether the data is correct. You should delete observations when the dependent variable is clearly in error. If a predictor variable for an observation is in error, you can stop short of deleting the entire observation. But it is worth spending time trying to fix the error. If a predictor variable has a missing value for one or several observations, you can replace the missing value through a process known as imputation. 4 Reminder: Cleaning data can be time consuming, but it is a vital step in your analysis. 4. MATCH THE DATA TO THE THEORY. Data can be classified as either primary or secondary. Researchers collect primary data for specific projects. Secondary data is collected by one researcher but used by another. The advantage of primary data is that the researcher can choose what to measure. But even with primary data and especially with secondary data it may take a leap of faith to accept that the variable for which you have data matches the theoretical variable of interest. For example, you may hypothesize that new product introductions are a function of a firm s technical know-how. How would you define a product introduction? How would you measure technical know-how? When assessing the appropriateness of your variables, be sure to tell stories. You might convince yourself that higher levels of R&D spending would be a good measure of technical know-how. But what if firms spend more on R&D when they are trying to catch up with more technically savvy competitors? In that case, higher expenditures could indicate lesser know-how. (Bonus question: Do you see that R&D spending might be endogenous?) Another measure of know-how might be accumulated patents. This is an intuitively plausible measure that does not seem to suffer from the causality problem described above. Thus, 4 Imputation is essentially an informed guess. You could replace the missing value with the average value for other observations. More sophisticated imputation methods involve regressing the dependent variable on a set of predictors and using the resulting model to predict the values of the missing observations. Stata has a good imputation package called mi impute. 6 KELLOGG SCHOOL OF MANAGEMENT

7-112-001 TECHNICAL NOTE: EMPIRICAL STEPS accumulated patents might be a fine proxy for know-how. It is easy to measure, correlated with know-how (we hope), and appropriate for use in regression analysis. 5. HAVING WORKED THROUGH THE THEORY, ONCE AGAIN DELVE DEEPLY INTO YOUR DATA. Look for correlations, trends, and so on. Find the action in the data do the predictor variables move around a lot? Does the dependent variable show any patterns? Most students learn about multicollinearity in a basic statistics course. When it is a problem, standard errors can get very large. Sometimes the coefficients on the correlated variables may flip signs. Unfortunately, most students have an overblown fear of multicollinearity, which is rarely considered a major problem in real world empirical research. Moreover, it is very easy to detect and correct for multicollinearity; these methods are described at length in a later technical note in this series. 5 6. BEGIN YOUR EMPIRICAL WORK BY BRIEFLY EXAMINING A SIMPLE MODEL, BUT THEN SETTLE ON A RICHER CORE MODEL. Start with a stripped-down approach to your analysis. For example, you could assess the claim that size is an important determinant of profitability by dividing the sample of firms into two groups, big and small, and comparing mean profits across the two groups. If this simple comparison fails, then odds are good that a richer comparison will fail as well. Simple comparisons are essential to learning about your data. But do not rely too much on simple analyses because you may be omitting key control variables, which could seriously bias your results. Most of your analysis should center on a core model that includes those variables that theory says matters most. 7. AFTER AN INITIAL LOOK AT THE CORE MODEL IN OLS, SETTLE ON AN APPROPRIATE EMPIRICAL METHOD. OLS regression is usually a good starting point for analysis, but depending on the problem at hand, it can generate misleading results and imprecise estimates. Remember that OLS assumes that the dependent variable Y is generated by some model along the lines of Y = B 0 + B 1 X 1 + B 2 X 2 + ε, where ε is a normally distributed error term. This is not always the best way to describe the relationship between the predictor variables and the dependent variable. There are many issues to address when selecting an empirical method. Some key questions that will guide your choice of methods include: Is the dependent variable discrete (e.g., 0 or 1) rather than continuous? Is the dependent variable categorical (e.g., the choice among brands of automobiles)? Is the dependent variable truncated at 0? If you answered yes to any of these questions, then a method other than OLS may be indicated, or you may need to transform the values of your variables to generate a more 5 David Dranove, Practical Regression: Building Your Model, Case #7-112-003 (Kellogg School of Management, 2012). KELLOGG SCHOOL OF MANAGEMENT 7

TECHNICAL NOTE: EMPIRICAL STEPS 7-112-001 appropriate empirical specification. Some key questions that will guide your variable specification include: Are any of the independent variables categorical? If so, OLS might be suitable, but you might need to create dummy variables. Do you need to include interaction terms? Do you need to take logs, include polynomial terms, and so forth? 8. ASSESS THE ERROR STRUCTURE FOR VIOLATIONS OF OLS ASSUMPTIONS. OLS generates unbiased standard errors only when certain assumptions about the error terms are satisfied. Two important assumptions are: The errors are independent. The errors are homoscedastic, that is, they have identical distributions. Violation of this assumption is known as heteroskedasticity. Non-Independence Are your data naturally grouped, with multiple observations for a given firm, state, individual, and so on? If you have such groups, then the intragroup errors (e.g., the errors for each division within the firm) are unlikely to be independent. This means you do not have as many degrees of freedom as suggested by the sheer number of rows of data. You must account for the grouping of the data or again you will obtain the wrong standard errors. Heteroskedasticity Are the magnitudes of your errors systematically correlated with any RHS variables or any other variables you can think of? If so, you have heteroskedasticity. Your coefficients are probably unbiased, but the standard errors may be incorrect. 9. DETERMINE WHETHER YOUR RESULTS ARE ROBUST. As you analyze your data, you may decide that there are several equally defensible models. Your results are robust if you obtain similar key results from a number of defensible models. Your goal when performing regression is to find robust results, not to find a mythical best model. If your results are not robust, then they are not convincing. Robustness can be a confusing concept, so remember this critical robustness question: Are your key findings immune to defensible changes in the regression model? Some secondary robustness questions include: Do coefficients vary much as you add potentially important control variables to what seems to be a rich model? 8 KELLOGG SCHOOL OF MANAGEMENT

7-112-001 TECHNICAL NOTE: EMPIRICAL STEPS Are the results of fancy methods different from the results from OLS? Are the results driven by outliers? If you answered yes to any of the above questions, then you may have a robustness problem. At the very least, you should try to gain a better understanding of relationships in the data that are causing these problems. If your results are not robust, then perhaps you have done a poor job of specifying the model. More likely, the patterns in the data are not consistent enough to warrant a strong conclusion. 10. INTERPRET YOUR RESULTS. No statistical analysis is complete until you interpret the results. Your interpretation should largely relate back to the question you posed at the start of your analysis, focusing on two elements: What is the direction of the observed relationship between the predictor variable of interest and the dependent variable? Can you reject the null hypothesis of no relationship? What is the magnitude of the relationship? It is not enough to state that when a CEO dies unexpectedly, firms tend to lose value. You must go further and state that on average, these firms lose X percent of their value. This way, you move from theory to practicality. Always give a real world interpretation of your results. Summary 1. Write down a model. 2. Think about causality. 3. Obtain, organize, and clean the data. 4. Match the data to the theory. 5. Get to know your data. 6. Briefly examine a simple model and then focus on a core model. 7. Implement the appropriate empirical method. 8. Assess the error structure and make necessary corrections to the model. 9. Test for robustness. 10. Interpret the results. KELLOGG SCHOOL OF MANAGEMENT 9