Chapter 12: Introduction to Analysis of Variance

Similar documents
Analysis of Variance (ANOVA)

One-Way ANOVAs t-test two statistically significant Type I error alpha null hypothesis dependant variable Independent variable three levels;

Testing Means. Related-Samples t Test With Confidence Intervals. 6. Compute a related-samples t test and interpret the results.

Analysis of Variance (ANOVA) Program Transcript

One-Way Independent ANOVA

Analysis of Variance: repeated measures

Still important ideas

Applied Statistical Analysis EDUC 6050 Week 4

Still important ideas

Readings: Textbook readings: OpenStax - Chapters 1 13 (emphasis on Chapter 12) Online readings: Appendix D, E & F

8/28/2017. If the experiment is successful, then the model will explain more variance than it can t SS M will be greater than SS R

Two-Way Independent ANOVA

Readings: Textbook readings: OpenStax - Chapters 1 11 Online readings: Appendix D, E & F Plous Chapters 10, 11, 12 and 14

Two-Way Independent Samples ANOVA with SPSS

Final Exam PS 217, Fall 2010

ANOVA. Thomas Elliott. January 29, 2013

Business Statistics Probability

Describe what is meant by a placebo Contrast the double-blind procedure with the single-blind procedure Review the structure for organizing a memo

Comparing 3 Means- ANOVA

Math Section MW 1-2:30pm SR 117. Bekki George 206 PGH

Describe what is meant by a placebo Contrast the double-blind procedure with the single-blind procedure Review the structure for organizing a memo

Chapter 11. Experimental Design: One-Way Independent Samples Design

Chapter 14 B Section D

PSY 216: Elementary Statistics Exam 4

Chapter 13: Introduction to Analysis of Variance

Stat Wk 9: Hypothesis Tests and Analysis

CHAPTER ONE CORRELATION

Hypothesis Testing. Richard S. Balkin, Ph.D., LPC-S, NCC

Basic Statistics and Data Analysis in Work psychology: Statistical Examples

PSYCHOLOGY 320L Problem Set #4: Estimating Sample Size, Post Hoc Tests, and Two-Factor ANOVA

Analysis of Variance ANOVA, Part 2. What We Will Cover in This Section. Factorial ANOVA, Two-way Design

Describe what is meant by a placebo Contrast the double-blind procedure with the single-blind procedure Review the structure for organizing a memo

Lesson 11.1: The Alpha Value

Research Methods II, Spring Term Logic of repeated measures designs

1) What is the independent variable? What is our Dependent Variable?

Research Methods 1 Handouts, Graham Hole,COGS - version 1.0, September 2000: Page 1:

ANALYSIS OF VARIANCE (ANOVA): TESTING DIFFERENCES INVOLVING THREE OR MORE MEANS

Chapter 9. Factorial ANOVA with Two Between-Group Factors 10/22/ Factorial ANOVA with Two Between-Group Factors

Midterm Exam MMI 409 Spring 2009 Gordon Bleil

STA 3024 Spring 2013 EXAM 3 Test Form Code A UF ID #

The t-test: Answers the question: is the difference between the two conditions in my experiment "real" or due to chance?

Chapter 13: Introduction to Analysis of Variance

Statistics for EES Factorial analysis of variance

Comparing Two Means using SPSS (T-Test)

EPS 625 INTERMEDIATE STATISTICS TWO-WAY ANOVA IN-CLASS EXAMPLE (FLEXIBILITY)

Statisticians deal with groups of numbers. They often find it helpful to use

Research Analysis MICHAEL BERNSTEIN CS 376

Dr. Kelly Bradley Final Exam Summer {2 points} Name

SPSS output for 420 midterm study

Lesson 9: Two Factor ANOVAS

Lessons in biostatistics

CHAPTER - 6 STATISTICAL ANALYSIS. This chapter discusses inferential statistics, which use sample data to

Exam 3 PS 217, Spring 2011

The Single-Sample t Test and the Paired-Samples t Test

Biology 345: Biometry Fall 2005 SONOMA STATE UNIVERSITY Lab Exercise 8 One Way ANOVA and comparisons among means Introduction

One way Analysis of Variance (ANOVA)

Psychology Research Process

Statistical Significance, Effect Size, and Practical Significance Eva Lawrence Guilford College October, 2017

CHAPTER THIRTEEN. Data Analysis and Interpretation: Part II.Tests of Statistical Significance and the Analysis Story CHAPTER OUTLINE

Simple Sensitivity Analyses for Matched Samples Thomas E. Love, Ph.D. ASA Course Atlanta Georgia

1. Below is the output of a 2 (gender) x 3(music type) completely between subjects factorial ANOVA on stress ratings

CHAPTER OBJECTIVES - STUDENTS SHOULD BE ABLE TO:

Final Exam Practice Test

Chapter 11 Nonexperimental Quantitative Research Steps in Nonexperimental Research

A Brief (very brief) Overview of Biostatistics. Jody Kreiman, PhD Bureau of Glottal Affairs

12.1 Inference for Linear Regression. Introduction

Chapter 12: Analysis of covariance, ANCOVA

Variability. After reading this chapter, you should be able to do the following:

ANOVA in SPSS (Practical)

25. Two-way ANOVA. 25. Two-way ANOVA 371

Sheila Barron Statistics Outreach Center 2/8/2011

Research paper. One-way Analysis of Variance (ANOVA) Research paper. SPSS output. Learning objectives. Alcohol and driving ability

Psychology Research Process

Research Questions, Variables, and Hypotheses: Part 2. Review. Hypotheses RCS /7/04. What are research questions? What are variables?

HS Exam 1 -- March 9, 2006

CHAPTER NINE DATA ANALYSIS / EVALUATING QUALITY (VALIDITY) OF BETWEEN GROUP EXPERIMENTS

Interactions between predictors and two-factor ANOVA

Chapter 23. Inference About Means. Copyright 2010 Pearson Education, Inc.

Statistical inference provides methods for drawing conclusions about a population from sample data.

Problem #1 Neurological signs and symptoms of ciguatera poisoning as the start of treatment and 2.5 hours after treatment with mannitol.

Overview of Lecture. Survey Methods & Design in Psychology. Correlational statistics vs tests of differences between groups

STAT 200. Guided Exercise 4

Regression Including the Interaction Between Quantitative Variables

3 CONCEPTUAL FOUNDATIONS OF STATISTICS

MULTIPLE LINEAR REGRESSION 24.1 INTRODUCTION AND OBJECTIVES OBJECTIVES

Day 11: Measures of Association and ANOVA

PSYCHOLOGY 300B (A01) One-sample t test. n = d = ρ 1 ρ 0 δ = d (n 1) d

Study Guide for the Final Exam

Fixed-Effect Versus Random-Effects Models

EXERCISE: HOW TO DO POWER CALCULATIONS IN OPTIMAL DESIGN SOFTWARE

Steps in Inferential Analyses. Inferential Statistics. t-test

Library of Congress Cataloging-in-Publication Data. QA U dc

Levene Statistic df1 df2 Sig

THIS PROBLEM HAS BEEN SOLVED BY USING THE CALCULATOR. A 90% CONFIDENCE INTERVAL IS ALSO SHOWN. ALL QUESTIONS ARE LISTED BELOW THE RESULTS.

ID# Exam 3 PS 217, Spring 2009 (You must use your official student ID)

Inferential Statistics

t-test for r Copyright 2000 Tom Malloy. All rights reserved

Overview. Survey Methods & Design in Psychology. Readings. Significance Testing. Significance Testing. The Logic of Significance Testing

Instructions for doing two-sample t-test in Excel

CHAPTER TWO REGRESSION

Transcription:

Chapter 12: Introduction to Analysis of Variance

of Variance Chapter 12 presents the general logic and basic formulas for the hypothesis testing procedure known as analysis of variance (ANOVA). The purpose of ANOVA is much the same as for t tests: the goal of t tests is to compare distances; between sample means and between distances computed for the standard error. The goal of ANOVA is to determine whether the variances (measures of squared distance) between means are sufficiently large to justify a conclusion that there are mean differences between the populations from which the samples were obtained.

of Variance (cont'd.) The difference between ANOVA and the t tests is that ANOVA can be used in situations where there are two or more means being compared, whereas the t tests are limited to situations where only two means are involved. When only two means are involved, either test can be used, since these two techniques always result in the same statistical decision (p. 420). ANOVA is necessary to protect researchers from excessive risk of a Type I error (rejecting a null hypothesis that is actually true) in situations where a study is comparing more than two population means.

of Variance (cont'd.) These situations would require a series of several t tests to evaluate all of the mean differences. (Remember, a t test can compare only two means at a time.) Although each t test can be done with a specific α-level (risk of Type I error), the α-levels accumulate over a series of tests so that the final experiment-wide α-level can be quite large. For this reason, ANOVA is the most practical method for comparing multiple sample means. ANOVA allows researchers to evaluate all of the mean differences in a single hypothesis test using a single α-level, and thus keeps the risk of a Type I error under control no matter how many different means are being compared.

The Logic and the Process of Analysis Although ANOVA can be used in a variety of different research situations, this chapter presents only independentmeasures designs involving only one independent variable. Because ANOVA requires so many different formulas, it can be problematic to keep track of formulas and numbers, let alone calculating all of them. For this reason, we are learning how ANOVA is organized, but not requiring you to actually calculate all the formulas. In other words, you need to know about ANOVA, and how to interpret its results. Those of you who want to actually work with ANOVA can do so using the problems given in your text and your SPSS software.

of Variance (cont'd.) Thus, the F-ratio has the same basic structure as the independent-measures t statistic presented in Chapter 10. obtained mean differences (including treatment effects) MS between F = = differences expected by chance (without treatment effects) MS within

A large value for the F-ratio indicates that the obtained sample mean differences are greater than would be expected if the treatments had no effect. Each of the sample variances, MS values, in the F-ratio is computed using the basic formula for sample variance: SS sample variance = S 2 = df

To obtain the SS and df values, you must go through an analysis that separates the total variability for the entire set of data into two basic components: within-treatment variability (which will be the denominator) and between-treatment variability (which will become the numerator of the F-ratio).

The two components of the F-ratio can be described as follows: Within-Treatments Variability: MS within measures the size of the differences that exist inside each of the samples. Because all the individuals in a sample receive exactly the same treatment, any differences (or variance) within a sample cannot be caused by different treatments.

Thus, these differences are caused by only one source: Chance or Error: The unpredictable differences that exist between individual scores are not caused by any systematic factors and are simply considered to be random chance or error.

Between-Treatments Variability: MS between measures the size of the differences between the sample means. For example, suppose that three treatments, each with a sample of n = 5 subjects, have means of M 1 = 1, M 2 = 2, and M 3 = 3. Notice that the three means are different; that is, they are variable.

By computing the variance for the three means we can measure the size of the differences. Although it is possible to compute a variance for the set of sample means, it usually is easier to use the total, T, for each sample instead of the mean, and compute variance for the set of T values.

Logically, the differences (or variance) between means can be caused by two sources: Treatment Effects: If the treatments have different effects, this could cause the mean for one treatment to be higher (or lower) than the mean for another treatment. Chance or Sampling Error: If there is no treatment effect at all, you would still expect some differences between samples. Mean differences from one sample to another are an example of random, unsystematic sampling error.

Considering these sources of variability, the structure of the F-ratio becomes: treatment effect + random differences F = random differences

When the null hypothesis is true and there are no differences between treatments, the F-ratio is balanced. That is, when the "treatment effect" is zero, the top and bottom of the F-ratio are measuring the same variance. In this case, you should expect an F-ratio near 1.00. When the sample data produce an F-ratio near 1.00, we will conclude that there is no significant treatment effect.

of Variance (cont'd.) On the other hand, a large treatment effect will produce a large value for the F-ratio. Thus, when the sample data produce a large F-ratio we will reject the null hypothesis and conclude that there are significant differences between treatments. To determine whether an F-ratio is large enough to be significant, you must select an α-level, find the df values for the numerator and denominator of the F-ratio, and consult the F-distribution table to find the critical value. (see appendix B for statistical tables)

Analysis of Variance and Post Tests The null hypothesis for ANOVA states that for the general population there are no mean differences among the treatments being compared; H 0 : μ 1 = μ 2 = μ 3 =... When the null hypothesis is rejected, the conclusion is that there are significant mean differences. However, the ANOVA simply establishes that differences exist, it does not indicate exactly which treatments are different.

Measuring Effect Size for an Analysis of Variance As with other hypothesis tests, an ANOVA evaluates the significance of the sample mean differences; that is, are the differences bigger than would be reasonable to expect just by chance. With large samples, however, it is possible for relatively small mean differences to be statistically significant. Thus, the hypothesis test does not necessarily provide information about the actual size of the mean differences.

Measuring Effect Size for an Analysis To supplement the hypothesis test, it is recommended that you calculate a measure of effect size. For an analysis of variance the common technique for measuring effect size is to compute the percentage of variance that is accounted for by the treatment effects.

Measuring Effect Size for an Analysis For the t statistics, this percentage was identified as r 2, but in the context of ANOVA the percentage is identified as η 2 (the Greek letter eta, squared). The formula for computing effect size is: SS between treatments η 2 = SS total

Post Hoc Tests With more than two treatments, this creates a problem. Specifically, you must follow the ANOVA with additional tests, called post hoc tests, to determine exactly which treatments are different and which are not. The Tukey s HSD and Scheffé test are examples of post hoc tests. These tests are done after an ANOVA where H 0 is rejected with more than two treatment conditions. The tests compare the treatments, two at a time, to test the significance of the mean differences.