Regression Including the Interaction Between Quantitative Variables

Similar documents
Example of Interpreting and Applying a Multiple Regression Model

Simple Linear Regression

Study Guide #2: MULTIPLE REGRESSION in education

Simple Linear Regression One Categorical Independent Variable with Several Categories

One-Way ANOVAs t-test two statistically significant Type I error alpha null hypothesis dependant variable Independent variable three levels;

CRITERIA FOR USE. A GRAPHICAL EXPLANATION OF BI-VARIATE (2 VARIABLE) REGRESSION ANALYSISSys

SPSS output for 420 midterm study

CHAPTER TWO REGRESSION

MULTIPLE LINEAR REGRESSION 24.1 INTRODUCTION AND OBJECTIVES OBJECTIVES

Problem #1 Neurological signs and symptoms of ciguatera poisoning as the start of treatment and 2.5 hours after treatment with mannitol.

SPSS output for 420 midterm study

ANOVA in SPSS (Practical)

Complex Regression Models with Coded, Centered & Quadratic Terms

Simple Linear Regression the model, estimation and testing

ANOVA. Thomas Elliott. January 29, 2013

2 Assumptions of simple linear regression

MULTIPLE OLS REGRESSION RESEARCH QUESTION ONE:

Business Statistics Probability

Regression. Page 1. Variables Entered/Removed b Variables. Variables Removed. Enter. Method. Psycho_Dum

Research paper. Split-plot ANOVA. Split-plot design. Split-plot design. SPSS output: between effects. SPSS output: within effects

bivariate analysis: The statistical analysis of the relationship between two variables.

12/31/2016. PSY 512: Advanced Statistics for Psychological and Behavioral Research 2

Dr. Kelly Bradley Final Exam Summer {2 points} Name

Chapter 11 Multiple Regression

3 CONCEPTUAL FOUNDATIONS OF STATISTICS

BIOL 458 BIOMETRY Lab 7 Multi-Factor ANOVA

IAPT: Regression. Regression analyses

Multiple Regression Models

Analysis of Variance: repeated measures

NORTH SOUTH UNIVERSITY TUTORIAL 2

Choosing a Significance Test. Student Resource Sheet

Advanced ANOVA Procedures

Regression CHAPTER SIXTEEN NOTE TO INSTRUCTORS OUTLINE OF RESOURCES

Subescala D CULTURA ORGANIZACIONAL. Factor Analysis

Statistics for Psychology

MMI 409 Spring 2009 Final Examination Gordon Bleil. 1. Is there a difference in depression as a function of group and drug?

Two-Way Independent Samples ANOVA with SPSS

Multiple Linear Regression Analysis

Chapter 3 CORRELATION AND REGRESSION

Item-Total Statistics

Unit 1 Exploring and Understanding Data

Biology 345: Biometry Fall 2005 SONOMA STATE UNIVERSITY Lab Exercise 5 Residuals and multiple regression Introduction

Overview of Lecture. Survey Methods & Design in Psychology. Correlational statistics vs tests of differences between groups

Business Research Methods. Introduction to Data Analysis

1. You want to find out what factors predict achievement in English. Develop a model that

11/18/2013. Correlational Research. Correlational Designs. Why Use a Correlational Design? CORRELATIONAL RESEARCH STUDIES

Subescala B Compromisso com a organização escolar. Factor Analysis

Daniel Boduszek University of Huddersfield

Analysis of Variance (ANOVA)

Lesson 9: Two Factor ANOVAS

WELCOME! Lecture 11 Thommy Perlinger

SUMMER 2011 RE-EXAM PSYF11STAT - STATISTIK

Stepwise method Modern Model Selection Methods Quantile-Quantile plot and tests for normality

Chapter 10: Moderation, mediation and more regression

11/24/2017. Do not imply a cause-and-effect relationship

Two-Way Independent ANOVA

Correlation and Regression

Correlation and regression

Results & Statistics: Description and Correlation. I. Scales of Measurement A Review

Stat Wk 9: Hypothesis Tests and Analysis

MULTIPLE REGRESSION OF CPS DATA

kxk BG Factorial Designs kxk BG Factorial Designs Basic and Expanded Factorial Designs

Statistics 2. RCBD Review. Agriculture Innovation Program

Statistical Techniques. Meta-Stat provides a wealth of statistical tools to help you examine your data. Overview

Addendum: Multiple Regression Analysis (DRAFT 8/2/07)

STATISTICS INFORMED DECISIONS USING DATA

CLINICAL RESEARCH METHODS VISP356. MODULE LEADER: PROF A TOMLINSON B.Sc./B.Sc.(HONS) OPTOMETRY

1. Below is the output of a 2 (gender) x 3(music type) completely between subjects factorial ANOVA on stress ratings

THE STATSWHISPERER. Introduction to this Issue. Doing Your Data Analysis INSIDE THIS ISSUE

APÊNDICE 6. Análise fatorial e análise de consistência interna

LAMPIRAN A KUISIONER

Sociology 593 Exam 2 March 28, 2003

AP Statistics Practice Test Ch. 3 and Previous

3.2 Least- Squares Regression

Lab 4 (M13) Objective: This lab will give you more practice exploring the shape of data, and in particular in breaking the data into two groups.

Statistical questions for statistical methods

Describe what is meant by a placebo Contrast the double-blind procedure with the single-blind procedure Review the structure for organizing a memo

Problem Set 3 ECN Econometrics Professor Oscar Jorda. Name. ESSAY. Write your answer in the space provided.

Statistics as a Tool. A set of tools for collecting, organizing, presenting and analyzing numerical facts or observations.

STAT 135 Introduction to Statistics via Modeling: Midterm II Thursday November 16th, Name:

Section 3.2 Least-Squares Regression

Research paper. One-way Analysis of Variance (ANOVA) Research paper. SPSS output. Learning objectives. Alcohol and driving ability

Problem 1) Match the terms to their definitions. Every term is used exactly once. (In the real midterm, there are fewer terms).

Chapter 13: Factorial ANOVA

Chapter 1: Exploring Data

Interactions between predictors and two-factor ANOVA

Data and Statistics 101: Key Concepts in the Collection, Analysis, and Application of Child Welfare Data

0= Perempuan, 1= Laki-Laki

THE UNIVERSITY OF SUSSEX. BSc Second Year Examination DISCOVERING STATISTICS SAMPLE PAPER INSTRUCTIONS

Chapter 12: Introduction to Analysis of Variance

Still important ideas

One-Way Independent ANOVA

appstats26.notebook April 17, 2015

CHAPTER ONE CORRELATION

Readings Assumed knowledge

Small Group Presentations

Skala Stress. Putaran 1 Reliability. Case Processing Summary. N % Excluded a 0.0 Total

Manuscript Presentation: Writing up APIM Results

Introduction to Quantitative Methods (SR8511) Project Report

MODULE S1 DESCRIPTIVE STATISTICS

Transcription:

Regression Including the Interaction Between Quantitative Variables The purpose of the study was to examine the inter-relationships among social skills, the complexity of the social situation, and performance in a social situation. Each participant considered their most recent interaction in a group of 0 or larger that included at least 50% strangers, and rated their social performance (perf) and the complexity of the situation (sitcom), Then, each participant completed a social skills inventory that provided a single index of this construct (soskil). The researcher wanted to determine the contribution of the person and situation variables to social performance, as well as to consider their interaction. Descriptive Statistics SITCOM SOSKIL Valid N (listwise N Minimum Maximum Mean Std. Deviation 60.00 37.00 9.7000 8.5300 60 27.00 65.00 49.4700 8.2600 60 As before, we will want to center our quantitative variables by subtracting the mean from each person s scores. We also need to compute an interaction term as the product of the two centered variables. Some prefer using a full model approach, others a hierarchical model approach remember they produce the same results. Using the hierarchical approach, the centered quantitative variables (main effects) are entered on the first step and the interaction is added on the second step Be sure to check the R-squared change on the Statistics window

SPSS Output: is the main effects model and 2 is the full model. Summary 2 Change Statistics Adjusted Std. Error of R Square R R Square R Square the Estimate Change F Change df df2 Sig. F Change.374 a.40.0.3406.374 4.636 2 57.04.526 b.277.238 0.4899.37 0.620 3 56.002 a. Predictors: (Constant), C_SITCOM, C_SOSKIL b. Predictors: (Constant), C_SITCOM, C_SOSKIL, INT The R Square Change for 2 is the contribution of the interaction. The F Change for 2 is the test of the interaction HO:. 2 Regression Residual Total Regression Residual Total ANOVA c Sum of Squares df Mean Square F Sig. 92.355 2 596.78 4.636.04 a 7330.742 57 28.60 8523.097 59 2360.96 3 786.986 7.52.000 b 662.37 56 0.038 8523.098 59 a. Predictors: (Constant), C_SITCOM, C_SOSKIL b. Predictors: (Constant), C_SITCOM, C_SOSKIL, INT c. Dependent Variable: PERF 2 (Constant) C_SOSKIL C_SITCOM (Constant) C_SOSKIL C_SITCOM INT a. Dependent Variable: PERF Unstandardized Coefficients Coefficients a Standardi zed Coefficien ts B Std. Error Beta t Sig. 29.02.464 9.878.000.524.80.360 2.94.005.20.74.49.204.234 29.65.365 2.728.000.378.72.260 2.96.032.47.62.04.907.368 7.20E-02.022.385 3.259.002 Inspection of the full model tells us that we have an interaction the R² change =.37, which is nearly ½ of the variance accounted for by the model, and the R² change test tells us that increase is larger than would be expected by chance (the same information is provided by the test of the interaction b weight in the full model t² - F-change). Note: Raw regression weights for interactions are often numerically small. Why? Remember the interaction term is computed as the product of the centered main effect term. It is likely to have a very large standard deviation, and so, a much smaller regression weight than the main effects. Always use ß weights to consider the relative contribution of the model terms. We can also see from the full model that there is a main effect for social skill, but not for the complexity of the social situation. We should be careful not to say that situational complexity does not contribute to this model!! Even though there is no main effect for this variable, it is involved in the interaction. Also, remember that because there is a significant interaction, we need to check carefully to determine if the main effects (significant or not) are descriptive or misleading. Now that we have determined that there is an interaction between these variables, we need to identify the pattern of that interaction, as a basis for interpreting it. To do this we will plot the model, interpret it and tell the story of how these variables relate to the criterion.

One Small Difference when Plotting the Interaction of 2 Quantitative Variables When we plotted the model for the interaction between a quantitiative variable and a categorical variable, we plotted the separate Y-X regression line for each of the different Z groups. Now, however, we have no groups but we need some lines! The usual convention is to plot the Y-X regression line for the mean of Z, for + std above the mean of Z and - std below the mean of Z. This roughly corresponds to plotting the his roughly corresponds to plotting the Y-X relationship for small, medium and large values of Z. You can also plot the Y-X lines for any specific value of Z that you like. We will use the 2 quant predictors w/ +/- Std Z portion of the IntPlot computator to obtain the plotting information. I have chosen to examine the interaction by looking at the relationship between social performance and situational complexity, for different amounts of social skill, so social complexity is X social skill is Z enter the regression weights from the full model and the standard deviation for each variable The output incudes the specific Y-X regression line for each value of Z It aso gives ploting points for each of these lines Finally, it gives the cross over point for where the three Y-X lines converge, which can make plotting the separate lines much easier

Plotting these simple regression lines, we get the following figure (ploting values are marked as -, o, and +, respectively). Perf. + std on social skills 40 + + mean on social skills 30 o o - - 20 - std on social skills -6.52-8.26 0 8.26 6.52 Situational Complexity (centered) Interpreting Regression Weights Like with factorial designs, it is a good idea to start by interpreting the interaction (because whether or not it is significant is important do decide how to interpret the main effects) the interaction b weight tells the direction and extent of the change in the slope of the Y-X regression line for a -unit increase in Z (or the direction and extent of the change in slope of the Y-Z regression line for a -unit change in X) a + weight means that as z gets larger, the Y-X slope gets more positive a 0 weight (no interaction) means that the Y-X slope is the same for all values of Z a weight means that as Z gets larger, the Y-X slope gets less positive the b weight of X tells the slope of the Y-X line at the mean of Z the constant (a) tells the y-intercept for the Y-X line at the mean of Z (which was centered) the b weight of Z tells how the direction and extent of the change in the height of the Y-X regression line for a -unit increase in Z The crossover of the separate regression lines precludes having descriptive main effects! Since the regression lines vary from + to -, the Y-X regression line for the mean of Z isn t descriptive for all values of Z. Since the regressions line cross, differences between the height of the different lines at the 0-point of X isn t descriptive for al values of X

Using SPSS to get the Separate Regression Lines The draw-back to the approach we just took is that we do not have significance tests of whether the simple regression slopes are different from zero. When we want that formal test, we must apply the following procedure. Obtaining the simple regression line for complexity std ABOVE the mean of social skills -- we need to recenter the scores around a point one standard deviation above the mean -- this means, in effect, that we need to lower all the scores by one standard deviation -- then we need to compute an interaction term specific to these recentered values -- these new variables are used to obtain the simple regression line SPSS: Transform Compute Compute the variable abvskil by subtracting the standard deviation (8.26) from c_soskil. Then compute the recentered interaction term abvint by multiplying abvskil by c_sitcom. SPSS: Analyze Regression Linear Enter perf as the dependent variable and c_sitcom, abvskil and abvint as the Independents.

SPSS Output: Notice that the R² is the same as the original model! - the recentering re-partitions the variance in the full model Summary Adjusted Std. Error of R R Square R Square the Estimate.526 a.277.238 0.4899 a. Predictors: (Constant), ABVINT, ABVSKIL, C_SITCOM Regression Residual Total ANOVA b Sum of Squares df Mean Square F Sig. 2360.960 3 786.986 7.52.000 a 662.37 56 0.038 8523.090 59 a. Predictors: (Constant), ABVINT, ABVSKIL, C_SITCOM b. Dependent Variable: PERF (Constant) C_SITCOM ABVSKIL ABVINT a. Dependent Variable: PERF Unstandardized Coefficients Coefficients a Standardi zed Coefficien ts B Std. Error Beta t Sig. 32.775.930 6.896.000.735.228.522 3.225.002.378.72.260 2.96.032 7.20E-02.022.53 3.259.002 The b-weight and its t-test tells us that there is a significant, positive relationship between performance and social complexity for those std above the mean on social skills (same value as was obtained by hand-calculation above). One could interpret this as meaning that, for those with particularly good social skills, increased social complexity brings out their best.

Obtaining the simple regression line for complexity std BELOW the mean of social skills -- we need to recenter the scores around a point one standard deviation below the mean -- this means, in effect, that we need to raise all the scores by one standard deviation -- then we need to compute an interaction term specific to these recentered values -- these new variables are used to obtain the simple regression line SPSS: Transform Compute Compute the variable belskil by adding the standard deviation (8.26) to c_soskil. Then compute the recentered interaction term belint by multiplying belskil by c_sitcom. SPSS: Analyze Regression Linear Enter perf as the dependent variable and c_sitcom, belskil and belint as the Independents.

SPSS Output: Summary Adjusted Std. Error of R R Square R Square the Estimate.526 a.277.238 0.4899 a. Predictors: (Constant), BELINT, BELSKIL, C_SITCOM Regression Residual Total ANOVA b Sum of Squares df Mean Square F Sig. 2360.960 3 786.987 7.52.000 a 662.37 56 0.038 8523.097 59 a. Predictors: (Constant), BELINT, BELSKIL, C_SITCOM b. Dependent Variable: PERF (Constant) C_SITCOM BELSKIL BELINT a. Dependent Variable: PERF Coefficients a Unstandardized Coefficients Standardi zed Coefficien ts B Std. Error Beta t Sig. 26.529 2.002 3.250.000 -.44.257 -.33 -.79.09.378.72.260 2.96.032 7.20E-02.022.592 3.259.002 The b-weight tells us that there is a negative relationship between performance and social complexity for those std below the mean on social skills (same value as was obtained by hand-calculation above). One could interpret this as meaning that, for those with particularly poor social skills, do not handle increased social complexity well, and their performance suffers. You will probably notice that the significance test of the simple regression weight we have just interpreted does not reject H0:. What to do? Remember that the choice of +/- standard deviation is arbitrary, so many folks who do this sort of analysis suggest re-computing the simple slopes at +/-.5 standard deviations. It is likely that this more extreme set of simple slopes will be more different from the slope for the mean of social skills.