Sheila Barron Statistics Outreach Center 2/8/2011

Similar documents
OCW Epidemiology and Biostatistics, 2010 David Tybor, MS, MPH and Kenneth Chui, PhD Tufts University School of Medicine October 27, 2010

Applied Statistical Analysis EDUC 6050 Week 4

Standard Deviation and Standard Error Tutorial. This is significantly important. Get your AP Equations and Formulas sheet

Biostatistics 3. Developed by Pfizer. March 2018

15.301/310, Managerial Psychology Prof. Dan Ariely Recitation 8: T test and ANOVA

APPENDIX N. Summary Statistics: The "Big 5" Statistical Tools for School Counselors

Inferential Statistics

Psychology Research Process

PSYCHOLOGY 300B (A01) One-sample t test. n = d = ρ 1 ρ 0 δ = d (n 1) d

Statistical inference provides methods for drawing conclusions about a population from sample data.

10 Intraclass Correlations under the Mixed Factorial Design

Psychology Research Process

Power & Sample Size. Dr. Andrea Benedetti

Objectives. Quantifying the quality of hypothesis tests. Type I and II errors. Power of a test. Cautions about significance tests

Statistical Significance, Effect Size, and Practical Significance Eva Lawrence Guilford College October, 2017

1 The conceptual underpinnings of statistical power

One-Way ANOVAs t-test two statistically significant Type I error alpha null hypothesis dependant variable Independent variable three levels;

Hypothesis Testing. Richard S. Balkin, Ph.D., LPC-S, NCC

Checking the counterarguments confirms that publication bias contaminated studies relating social class and unethical behavior

Methods for Determining Random Sample Size

Chapter 23. Inference About Means. Copyright 2010 Pearson Education, Inc.

SAMPLING AND SAMPLE SIZE

The t-test: Answers the question: is the difference between the two conditions in my experiment "real" or due to chance?

Stat 13, Intro. to Statistical Methods for the Life and Health Sciences.

Creative Commons Attribution-NonCommercial-Share Alike License

appstats26.notebook April 17, 2015

Reflection Questions for Math 58B

Chapter 12. The One- Sample

Chi Square Goodness of Fit

Chapter 19. Confidence Intervals for Proportions. Copyright 2010 Pearson Education, Inc.

A Brief (very brief) Overview of Biostatistics. Jody Kreiman, PhD Bureau of Glottal Affairs

AP Statistics TOPIC A - Unit 2 MULTIPLE CHOICE

t-test for r Copyright 2000 Tom Malloy. All rights reserved

Economics 345 Applied Econometrics

Measuring the User Experience

Business Statistics Probability

Data and Statistics 101: Key Concepts in the Collection, Analysis, and Application of Child Welfare Data

Glossary From Running Randomized Evaluations: A Practical Guide, by Rachel Glennerster and Kudzai Takavarasha

Designing Psychology Experiments: Data Analysis and Presentation

Lec 02: Estimation & Hypothesis Testing in Animal Ecology

Power of a Clinical Study

Student Performance Q&A:

Lesson 11.1: The Alpha Value

Chapter 19. Confidence Intervals for Proportions. Copyright 2010, 2007, 2004 Pearson Education, Inc.

5.3: Associations in Categorical Variables

Statistics for Psychology

An Introduction to Bayesian Statistics

Hypothesis testing: comparig means

STATISTICS - CLUTCH CH.11: HYPOTHESIS TESTING: PART 1.

A Case Study: Two-sample categorical data

Psy201 Module 3 Study and Assignment Guide. Using Excel to Calculate Descriptive and Inferential Statistics

Chapter 8 Estimating with Confidence

CHAPTER 3 DATA ANALYSIS: DESCRIBING DATA

Risk Aversion in Games of Chance

Research Questions, Variables, and Hypotheses: Part 2. Review. Hypotheses RCS /7/04. What are research questions? What are variables?

Chapter 11. Experimental Design: One-Way Independent Samples Design

EXERCISE: HOW TO DO POWER CALCULATIONS IN OPTIMAL DESIGN SOFTWARE

Title:Mixed-strain Housing for Female C57BL/6, DBA/2, and BALB/c Mice: Validating a Split-plot Design that promotes Refinement and Reduction

HYPOTHESIS TESTING 1/4/18. Hypothesis. Hypothesis. Potential hypotheses?

Chapter 7: Descriptive Statistics

The Single-Sample t Test and the Paired-Samples t Test

Understanding Statistics for Research Staff!

Research Methods 1 Handouts, Graham Hole,COGS - version 1.0, September 2000: Page 1:

Quantitative Evaluation

Statistics: Bar Graphs and Standard Error

Sampling for Impact Evaluation. Maria Jones 24 June 2015 ieconnect Impact Evaluation Workshop Rio de Janeiro, Brazil June 22-25, 2015

Midterm Exam MMI 409 Spring 2009 Gordon Bleil

STAT 200. Guided Exercise 4

Villarreal Rm. 170 Handout (4.3)/(4.4) - 1 Designing Experiments I

Utilizing t-test and One-Way Analysis of Variance to Examine Group Differences August 31, 2016

Why do Psychologists Perform Research?

Statistical Power Sampling Design and sample Size Determination

THIS PROBLEM HAS BEEN SOLVED BY USING THE CALCULATOR. A 90% CONFIDENCE INTERVAL IS ALSO SHOWN. ALL QUESTIONS ARE LISTED BELOW THE RESULTS.

I. Introduction and Data Collection B. Sampling. 1. Bias. In this section Bias Random Sampling Sampling Error

Testing Means. Related-Samples t Test With Confidence Intervals. 6. Compute a related-samples t test and interpret the results.

Describe what is meant by a placebo Contrast the double-blind procedure with the single-blind procedure Review the structure for organizing a memo

Final Exam Practice Test

Do not copy, post, or distribute

Exam 4 Review Exercises

Statistical Methods Exam I Review

Inferential Statistics: An Introduction. What We Will Cover in This Section. General Model. Population. Sample

Study Guide for the Final Exam

CHAPTER NINE DATA ANALYSIS / EVALUATING QUALITY (VALIDITY) OF BETWEEN GROUP EXPERIMENTS

d =.20 which means females earn 2/10 a standard deviation more than males

Probability and Sample space

ANOVA. Thomas Elliott. January 29, 2013

Chapter 12: Introduction to Analysis of Variance

12.1 Inference for Linear Regression. Introduction

Never P alone: The value of estimates and confidence intervals

STAT 113: PAIRED SAMPLES (MEAN OF DIFFERENCES)

Unit 6.2: Mendelian Inheritance

BIOSTATISTICAL METHODS

Confidence Intervals On Subsets May Be Misleading

Analysis of Variance (ANOVA)

Review: Conditional Probability. Using tests to improve decisions: Cutting scores & base rates

3 CONCEPTUAL FOUNDATIONS OF STATISTICS

Designing a Questionnaire

SAMPLE SIZE IN CLINICAL RESEARCH, THE NUMBER WE NEED

How to Conduct On-Farm Trials. Dr. Jim Walworth Dept. of Soil, Water & Environmental Sci. University of Arizona

Assignment 4: True or Quasi-Experiment

Transcription:

Sheila Barron Statistics Outreach Center 2/8/2011

What is Power? When conducting a research study using a statistical hypothesis test, power is the probability of getting statistical significance when the null hypothesis is false. Key Points Hypothesis testing Power is a probability Statistical significance Null hypothesis is false 2/8/2011 2

Hypothesis Testing Procedure Set up a null (or no effect) hypothesis (this is the hypothesis that the researcher would like to disprove). Conduct the study -- Draw one or more random samples (depending on your design) from the population, possibly apply a treatment, and measure each individual on the construct(s) of interest, compute any statistics necessary. Compare the results from the sample(s) to the null hypothesis. The sample result is based on a statistic calculated using the data from the sample(s). This statistic is often based on the sample mean, but other statistics are also used. If the sample result is highly unlikely assuming the null hypothesis is true, reject the null hypothesis. If the sample result is likely or just somewhat unlikely assuming the null hypothesis is true, retain the null hypothesis as a possibility. 2/8/2011 3

Statistical Inference When we conduct a hypothesis test, the purpose is to draw a conclusion about a population based on data from a sample. Population all of the individual units that are of interest (they might be people, schools, mice, or anything that can be counted). Sample a subset of the population. We generally assume that we have a random sample which means each individual in the population has an equal chance of being selected for the sample. 2/8/2011 4

How do we decide if the sample result is highly unlikely? Define highly unlikely in terms of probability. This is called alpha (α). It is also called the Type I error rate the probability of rejecting a true null hypothesis. If the likelihood of the sample result is less than alpha then we feel confident saying that the null hypothesis must be false. 2/8/2011 5

Type I Error Incorrectly rejecting the null hypothesis. In other words, claiming that you found something in your study when in fact there was no effect. Researchers long ago decided that it was advantageous to minimize the probability of making a Type I error. That way you wouldn t have lots of people running around making claims that they had discovered something when in fact the result was just due to chance. 2/8/2011 7

Type I Error (cont.) In order to allow the researcher to have strict control over Type I errors, the hypothesis testing procedure is set up so that the researcher picks the value for the Type I error rate that s/he is willing to tolerate. The probability of a Type I error is called alpha (α). By convention alpha is often set at.05. This means that the researcher is willing to accept a 5% chance of falsely rejecting the null hypothesis. 2/8/2011 8

Type II Error There is another kind of error that can occur in hypothesis testing: Incorrectly retaining the null hypothesis. In other words, failing to find an effect when there actually is one. The probability of the Type II error is called beta (β). This type of error is not fixed in a typical study. Although if the researcher can pick the N to use, power analysis will allow you to plan ahead so that you can keep this type of error low. 2/8/2011 9

Power Power is the probability of correctly rejecting the null hypothesis. In other words, power is the probability of correctly detecting an effect. Power = 1 β Power can be calculated if you know the size of the effect. Power can be estimated if you are willing to hypothesize a value for the effect. 2/8/2011 10

Sampling Distributions Knowing the value you want to use for alpha is not enough. You need to translate that into a decision rule that identifies which sample results would qualify as sufficiently unlikely so as to allow you to reject the null hypothesis. To do this you need to know the probability distribution for the sample result. This is called the sampling distribution. 2/8/2011 11

A Simple Example Suppose I roll a 6-sided die twice, record the number that is facing up each time, and calculate the average of the two values. So if I got a 1 on the first roll and a 4 on the second roll, the mean would be 2.5. This is a simple little experiment with a sample size (N) of 2. 2/8/2011 12

What is the probability that the mean is 1? If the die is a fair die, the probability would be 1/36. (1/6*1/6) which equals.028 or 2.8%. This tells us that if we were to repeat this experiment 100 times, we would only expect to get a sample mean of 1 about three times. 2/8/2011 13

Sampling Distribution N=2 Sample Mean Percent 1 2.8% 1.5 5.6% 2 8.3% 2.5 11.1% 3 13.9% 3.5 16.7% 4 13.9% 4.5 11.1% 5 8.3% 5.5 5.6% 6 2.8% 2/8/2011 14

18.0% Sampling Distribution (N=2) 16.0% 14.0% 12.0% 10.0% 8.0% 6.0% 4.0% 2.0% 0.0% 1 1.5 2 2.5 3 3.5 4 4.5 5 5.5 6 2/8/2011 15

What if N=3? Suppose I roll a 6-sided die three times, record the number that is facing up each time, and calculate the average of the three values. So if I got a 2 on the first roll, a 6 on the second roll, and a 5 on the third role, the mean would be 4.33. Now what is the probability that the mean is 1? If the die is a fair die, the probability would be 1/216. (1/6*1/6*1/6) which equals.005 or 0.5%. This tells us that if we were to repeat this experiment 100 times, we would expect to get a sample mean of 1 either not at all or maybe just once. 2/8/2011 16

Sampling Distribution N=3 Sample Mean Percent 1 0.5% 1.33 1.4% 1.67 2.8% 2 4.6% 2.33 6.9% 2.67 9.7% 3 11.6% 3.33 12.5% 3.67 12.5% 4 11.6% 4.33 9.7% 4.67 6.9% 5 4.6% 5.33 2.8% 5.67 1.4% 6 0.5% 2/8/2011 17

14% Sampling Distribution (N=3) 12% 10% 8% 6% 4% 2% 0% 1 1.33 1.67 2 2.33 2.67 3 3.33 3.67 4 4.33 4.67 5 5.33 5.67 6 2/8/2011 18

What do you notice? N= 2 N=3 p(mean = 1) 2.8%.5% p(mean >= 5) 16.7% 9.3% The probability of an extreme result is smaller when the result is based on more information (in this case more rolls of the die) 2/8/2011 19

Experiment Say we re interested in testing the null hypothesis, This is a fair die against an alternative hypothesis that High numbers are more likely on this die than low numbers. We decide to use the mean to test this hypothesis (there are other ways to test the same hypothesis, but for our purpose, this is the way we want to test it). Let s also suppose we go with the conventional Type I error rate of.05 (α=.05). 2/8/2011 20

Experiment (cont.) If the die is fair and all numbers are equally likely then the mean of the distribution would be 3.5 and we would expect the sample mean to be close to 3.5. What would be our critical value be for rejecting the null hypothesis? If N= 3, the sample mean would have to be 5.33 or higher for us to be able to say that the probability of falsely rejecting the null hypothesis is less than.05. 2/8/2011 21

What would the Type II error rate be? This depends on the characteristics of the alternative distribution. If high numbers are more likely then the mean will be something higher than 3.5. How much higher? Let s just suppose that if the die is loaded, the true mean is 4.0 (This is based on assuming that 4, 5, & 6 are twice as likely to occur as 1, 2, and 3). Assuming the alternative hypothesis is true and the mean really is 4, the probability of getting a sample mean lower than 5.33 when in fact the alternative is true would be.889. In other words 89% of the time when this study was done with a loaded die (specifically with a die that has a mean of 4), the result would be incorrectly retaining the null hypothesis. 2/8/2011 22

Experiment (cont.) Thus what would the power be? 1 minus the probability of a Type II error 1-.89 =.11 Thus only 11% of the time would we expect to correctly find that the die is loaded. 2/8/2011 23

Experiment (cont.) Obviously this is a terrible study, but what can you expect with an N of 3. Luckily we didn t have much invested in it, it just takes a few seconds to roll a die 3 times. But what if the study had been more involved, it cost money, or took a lot of the researcher s time. Would you really want to invest those resources if there was only an 11 percent chance of actually detecting the effect? What if there was a 50% chance? 80%? At what point would you consider the probability high enough that you would want to devote your time and resources to the project? 2/8/2011 24

What is Power Analysis? Power analysis is a process where a researcher specifies the details of a study and uses those details to either: Determine the power of the study (the probability of finding a statistically significant result assuming that the null hypothesis is false). Determine the sample size that would be needed to obtain a specified level of power. 2/8/2011 25

Statistical Power In order to determine statistical power, the research must know four things about the study. Design Alpha (including one- or two-tailed) Sample size Effect size 2/8/2011 26

What is an effect size? There are many different measures of effect size, depending on the design of the study. The one that is easiest to understand is probably the correlation coefficient. The correlation coefficient is a measure of the degree of relationship between two variables. Another simple one is the mean difference. If your study involves comparing two groups, the difference in the means for the two groups is a measure of the effect size. An even more useful measure of the effect size is the difference between the means divided by the pooled standard deviation (This was first introduced by Cohen in 1969 and is typically referred to as Cohen s d). 2/8/2011 27

Cohen s d d = m 1 m 2 /s pooled Thus d reports the mean difference in standard deviation units. d = 1.0 means that one group performed a standard deviation above the other group. d =.2 means that one group performed two-tenths of an SD above the other group. 2/8/2011 28

The Big Problem with Power Analysis In order for a researcher to do a power analysis, the researcher must pick an effect size on which to base the calculation. However, typically the reason a researcher conducts a study is because s/he wants to find out if there is an effect and how big it is. So how can you do a power analysis prior to conducting a study, when the power analysis relies on a parameter that you aren t going to have an estimate of until after you have conducted the study. 2/8/2011 29

What to do? Some researchers will base the effect size on a rule of thumb. For Cohen s d the rule of thumb goes this way: Small effect d =.2 Medium effect d =.5 Large effect d =.8 Cohen advised (as paraphrased by Bausell & Li), when in doubt, hypothesize an effect size of.5. I strongly urge you not to rely too heavily on the rules of thumb. The average effect size over a great many studies may in fact be close to.5, but only someone with in-depth understanding of what is being studied can decide what a reasonable effect size is for a particular study. 2/8/2011 30

How to pick an effect size? Way #1 Conduct a pilot study and estimate the effect size from the pilot data. This is an excellent idea it allows you to work out the kinks in the study, possibly improve the instruments. I help a lot of people with their research and I can say that a pilot study is one of the best ways to improve your chances of conducting a successful study. And it gives you a little data on which to calculate an estimate of the effect size. 2/8/2011 31

How to pick an effect size? (cont.) Way #2 Find the effect size in studies that involved similar or related populations and variables to the planned study. This is another thing that should always be done. Good studies are almost always based on a thorough review of the literature. It is getting more and more common for studies to report effect sizes, but even when they don t, the information needed for you to calculate an estimate of the effect size is generally reported. 2/8/2011 32

How to pick an effect size? (cont.) Way #3 Determine what the effect size would need to be for anyone to care about your results. Imagine there is a drug to treat a childhood disorder. You suspect that this drug is causing children to not grow as tall as they otherwise would. In this case, you could decide on the effect size to use in the power analysis based on the minimum decrease in height that you think anyone would care about. Would people care if kids taking the medication were on average.1 inch shorter than those not taking it? What about if they were a half inch shorter? A full inch? At what point would the finding be something of interest? 2/8/2011 33

How to go about doing a power analysis? First of all you need to decide what you want to calculate. For any design, there are four quantities that go into the power analysis: Alpha Effect size N Power If you specify three of these, you can calculate the fourth. 2/8/2011 34

How To Estimate Power You can calculate power by hand for simple designs, and I have students in Intermediate statistics class do that for a very simple test. But it is much easier to have a computer do it and there are now a number of programs that will calculate power for a variety of designs. I am going to focus on a free program that you can download from the Internet called G*Power. 2/8/2011 35

G*Power http://www.psycho.uniduesseldorf.de/abteilungen/aap/gpower3 Covers a variety of designs. It s free There is a user s guide, but it is a little patchy in terms of details. The easiest way to find it is to search on G*Power. 2/8/2011 36

Back to the Simple Example Let s start by using G*Power to calculate the power of my little experiment to decide if the die was fair or loaded. First we need to calculate the effect size. d = m H1 m H0 /s = 4-3.5 /1.71 =.29 Now, the underlying distributions here are not normal it is in fact uniform under the null hypothesis but we will see how close we get if normality is assumed. 2/8/2011 37

Notice, power equals.10 which is remarkably close to the value we calculated. The difference is due to using a t-test rather than an exact test. 2/8/2011 38

If I were to roll the die 10 times, rather than 3, power would increase to.22. Still not very good. G*Power can also be used to find out what sample size I would need to use to have the power I want to have. 2/8/2011 39

Power=.80 is typically considered a lower bound on what is acceptable. The power analysis indicates that I would need an N of 74. If I conduct the study with the recommended N of 74, I would have an 80% chance of getting a statistically significant result (assuming of course that the effect size of.29 is accurate). 2/8/2011 40

A More Interesting Example We are interested in whether or not a new drug is better for managing chronic back pain than the drug that is traditionally used. Suppose there was a special pain rating scale developed for doctors to use to rate the level of pain a patient is experiencing. In this rating scale higher scores mean the patient has more pain. Say we know that after a month of drug treatment, people who have the traditional drug have a mean score on the rating scale on 6.2. 2/8/2011 41

Example (cont.) Suppose we plan to recruit a random sample of 20 people with chronic back pain and gave them a month s worth of the new drug. After 1 month on the new drug, doctors will rate each patient on their level of pain. We are going to submit a proposal to the National Institutes of Health (NIH) for the money to conduct this study. But they aren t likely to fund the study unless we can convince them that the study is likely to be successful. Success is defined as correctly determining that the new drug works. Thus we need a power analysis. Based on a small pilot study, we estimate that the mean on the new drug is 5.2 and the standard deviation is 1.8. Putting these values in G*Power, we find that our power is.77. Do we want to invest the resources needed for a 77% chance of a significant result? And, even if we are, is NIH going to be willing to give us the money to conduct the study if the chance of success is.77. 2/8/2011 42

Study Design One sample tests are rarely optimal for many reasons so let s redesign the study but keep the fundamental scenario. We are interested in whether or not a new drug is better for managing chronic back pain than the drug that is traditionally used. We have 2 random samples, in one the subjects take the standard medication, in the other the subjects take the new drug. We want to know how many people we should include in the study in order to have a reasonable amount of power. 2/8/2011 43

Two dependent Groups T-test Suppose that we match the patients prior to assigning them to the standard or new drug. The two patients with the most severe pain are a pair, the next two in terms of severity of pain are a pair, and so on. For each pair, one subject is randomly assigned the new drug, and the other gets the standard drug. Now we are doing a dependent groups t-test. This is similar to an independent groups t-test but we need to take into account the correlation between the scores for the two groups. A very important thing to keep in mind is that the effect size is now measured using a different formula. d z is still a standardized group difference, but now the SD that we are using is the standard deviation of the difference scores. This means that the values that you may have in your head for what is a small, medium, or large effect size, have to be thrown out. 2/8/2011 44

Dependent Groups Effect Size The effect size index d z is defined as: d z = μ z /σ z = μ x μ y / (σ x 2 + σy 2 2ρxy σ x σ y ) I haven t found any rules of thumb for what is a small, medium, or large effect size using d z, but the values tend to be much larger than effect sizes using Cohen s d. So, don t use the Cohen s d rules of thumb. 2/8/2011 45

One-way ANOVA What if we had 3 independent groups? Standard Drug, New Drug (Dose 1), New Drug (Dose 2). For an ANOVA, the effect size used by G*Power is Cohen s f which is a ratio of the standard deviation among the population means and the standard deviation of the scores within the groups. 2/8/2011 46

Cohen s f The effect size f is defined as: f = σ m /σ. In this equation σ m is the standard deviation of the group means µ i and σ the common standard deviation within each of the k groups. The total variance is then σ 2 t = σ 2 m + σ 2. Cohen (1969, p.348) defined the following effect size conventions: small f = 0.10 medium f = 0.25 large f = 0.40 2/8/2011 47

Effect Sizes From http://www.psycho.uniduesseldorf.de/aap/projects/gpower/user_manual/use r_manual_02.html#effect_size Index small medium large T-test on Means d 0.20 0.50 0.80 T-test on Correlations r 0.10 0.30 0.50 F-test (ANOVA) f 0.10 0.25 0.40 F-test (MCR) f 2 0.02 0.15 0.35 Chi-square Test w 0.10 0.30 0.50 2/8/2011 48

Last Notes Whenever possible, base your effect size estimate on in-depth knowledge of the construct being studied. Rules of thumb are really rough guidelines, don t rely solely on them. Different measures of effect size (e.g., d, f, r) use different metrics. What is a small effect using one measure of effect size might be a large effect using another measure. 2/8/2011 49