Role of Statistics in Research

Similar documents
CHAPTER - 6 STATISTICAL ANALYSIS. This chapter discusses inferential statistics, which use sample data to

Chapter 8 Statistical Principles of Design. Fall 2010

Interpretation of Data and Statistical Fallacies

Hypothesis-Driven Research

Chapter 5: Field experimental designs in agriculture

One-Way ANOVAs t-test two statistically significant Type I error alpha null hypothesis dependant variable Independent variable three levels;

Survey Research Methodology

CRITICAL EVALUATION OF BIOMEDICAL LITERATURE

9 research designs likely for PSYC 2100

Designing Research and Demonstration Tests for Farmers Fields

Understanding and Managing Variation in Nutrient Composition

2 Critical thinking guidelines

SOME PRINCIPLES OF FIELD EXPERlMENTS WITH SHEEP By P. G. SCHINCICEL *, and G. R. MOULE *

CHECKLIST FOR EVALUATING A RESEARCH REPORT Provided by Dr. Blevins

Assignment #6. Chapter 10: 14, 15 Chapter 11: 14, 18. Due tomorrow Nov. 6 th by 2pm in your TA s homework box

Concepts Underlying the Design of Experiments

You can t fix by analysis what you bungled by design. Fancy analysis can t fix a poorly designed study.

EXPERIMENTAL RESEARCH DESIGNS

The Logic of Data Analysis Using Statistical Techniques M. E. Swisher, 2016

What Is Science? Lesson Overview. Lesson Overview. 1.1 What Is Science?

How to Conduct On-Farm Trials. Dr. Jim Walworth Dept. of Soil, Water & Environmental Sci. University of Arizona

Leaning objectives. Outline. 13-Jun-16 BASIC EXPERIMENTAL DESIGNS

Sequence balance minimisation: minimising with unequal treatment allocations

UNIVERSITY OF THE FREE STATE DEPARTMENT OF COMPUTER SCIENCE AND INFORMATICS CSIS6813 MODULE TEST 2

Evaluating Social Programs Course: Evaluation Glossary (Sources: 3ie and The World Bank)

Between Fact and Fiction: Artifacts and Ethics in Social Research. Toh Wah Seng ABSTRACT

9.0 L '- ---'- ---'- --' X

Chapter 2 Methodology: How Social Psychologists Do Research

Student Performance Q&A:

Investigative Biology (Advanced Higher)

Incorporating Experimental Research Designs in Business Communication Research

Response to the ASA s statement on p-values: context, process, and purpose

JSM Survey Research Methods Section

CHAPTER NINE DATA ANALYSIS / EVALUATING QUALITY (VALIDITY) OF BETWEEN GROUP EXPERIMENTS

Chapter 11. Experimental Design: One-Way Independent Samples Design

What Case Study means? Case Studies. Case Study in SE. Key Characteristics. Flexibility of Case Studies. Sources of Evidence

Enumerative and Analytic Studies. Description versus prediction

Experimental and survey design

USE AND MISUSE OF MIXED MODEL ANALYSIS VARIANCE IN ECOLOGICAL STUDIES1

Scoping and Planning. Experiment Planning. Planning Phase Overview. Planning Steps. Context Selection. Context Selection

DesCartes (Combined) Subject: Concepts and Processes Goal: Processes of Scientific Inquiry

Experimental Psychology

Sociological Research Methods and Techniques Alan S.Berger 1

lab exam lab exam Experimental Design Experimental Design when: Nov 27 - Dec 1 format: length = 1 hour each lab section divided in two

CONSORT 2010 checklist of information to include when reporting a randomised trial*

(77, 72, 74, 75, and 81).

CHAPTER 3 DATA ANALYSIS: DESCRIBING DATA

To conclude, a theory of error must be a theory of the interaction between human performance variability and the situational constraints.

MBA SEMESTER III. MB0050 Research Methodology- 4 Credits. (Book ID: B1206 ) Assignment Set- 1 (60 Marks)

SEMINAR ON SERVICE MARKETING

CHAPTER 5: PRODUCING DATA

Experimental Methods. Anna Fahlgren, Phd Associate professor in Experimental Orthopaedics

25. Two-way ANOVA. 25. Two-way ANOVA 371

MULTIPLE LINEAR REGRESSION 24.1 INTRODUCTION AND OBJECTIVES OBJECTIVES

GE Standard V: Scientific Standard

Lecture Notes Module 2

The Practice of Statistics 1 Week 2: Relationships and Data Collection

specific references is intentional: it is better to get expert personal

Expanding the Toolkit: The Potential for Bayesian Methods in Education Research (Symposium)

Vocabulary. Bias. Blinding. Block. Cluster sample

Psychology Research Process

Other Designs. What Aids Our Causal Arguments? Time Designs Cohort. Time Designs Cross-Sectional

04/12/2014. Research Methods in Psychology. Chapter 6: Independent Groups Designs. What is your ideas? Testing

Chapter 13 Summary Experiments and Observational Studies

In the broadest sense of the word, the definition of research includes any gathering of data, information, and facts for the advancement of knowledge.

Data and Statistics 101: Key Concepts in the Collection, Analysis, and Application of Child Welfare Data

1 The conceptual underpinnings of statistical power

A Brief Guide to Writing

Chapter 13. Experiments and Observational Studies. Copyright 2012, 2008, 2005 Pearson Education, Inc.

INTRODUCTION TO STATISTICS SORANA D. BOLBOACĂ

Definitions of Nature of Science and Scientific Inquiry that Guide Project ICAN: A Cheat Sheet

Agreement Coefficients and Statistical Inference

Political Science 15, Winter 2014 Final Review

Analysis of Environmental Data Conceptual Foundations: En viro n m e n tal Data

Assessing Limitations and Uses of Convenience Samples: A Guide for Graduate Students

Principles of Experimental Design

Principles of Experimental Design

RNA-seq. Differential analysis

What Science Is and Is Not

Research Methods 1 Handouts, Graham Hole,COGS - version 1.0, September 2000: Page 1:

Chapter 5: Research Language. Published Examples of Research Concepts

Summative Assessment For Twin Traits

PSYC3010 Advanced Statistics for Psychology

Psychology Research Process

Research Approaches Quantitative Approach. Research Methods vs Research Design

Designing Experiments... Or how many times and ways can I screw that up?!?

2.2 Psychologists Use Descriptive, Correlational, and Experimental Research Designs to Understand Behavior LEARNING OBJECTIVES

The t-test: Answers the question: is the difference between the two conditions in my experiment "real" or due to chance?

To open a CMA file > Download and Save file Start CMA Open file from within CMA

Science is a way of learning about the natural world by observing things, asking questions, proposing answers, and testing those answers.

Lecture 4: Research Approaches

SAMPLING AND SAMPLE SIZE

A. Indicate the best answer to each the following multiple-choice questions (20 points)

UNIT 5 - Association Causation, Effect Modification and Validity

Making Inferences from Experiments

Assignment 4: True or Quasi-Experiment

To open a CMA file > Download and Save file Start CMA Open file from within CMA

Quantitative survey methods

THE SCIENTIFIC WORLD VIEW

Overview of the Scientific Process

Transcription:

Role of Statistics in Research

Role of Statistics in research Validity Will this study help answer the research question? Analysis What analysis, & how should this be interpreted and reported? Efficiency Is the experiment the correct size, making best use of resources?

Validity Will the study answer the research question? Surveys select a sample from a population describe, but can t explain can identify relationships, but can t establish causality

Surveys & Causality In a survey: farm income increased by 10% for each increase in fertiliser of 30 kg/ha Is this relationship causal?

Surveys & Causality In a survey: farm income increased by 10% for each increase in fertiliser of 30 kg/ha Is this relationship causal? Not necessarily, other factors are involved: Managerial ability Farm size Educational level of farmer Fertiliser level may be related to these other possible causes, and may (or may not) be a cause itself

Survey Unit Example: In an survey to assess whether Herefords have a higher level of calving difficulty than Friesians, the individual cow is the survey unit.

Survey Unit Example: In a survey to assess the height of Irish males vs English males, the unit is the individual male in that one would sample a number of males of each country and take their heights rather than measure one male from each country many times.

Designed Experiments

Comparing treatment effect Effect = difference between treatments A well designed experiment leads to conclusion: or Either the treatments have produced the observed effect An improbable (chance < 1:20, 1:100 etc) event has occurred Technically we calculate a p-value of the data: i.e. the probability of obtaining an effect as large as that observed when in fact the average effect is zero

Essential elements of a designed experiment 1. COMPARATIVE The objective is to compare a number (>1) of treatments 2. REPLICATION Each treatment is tested on more than one experimental unit 3. RANDOMISATION experimental units are allocated to treatments at random

Replication Each treatment is tested on more than one experimental unit (the population item that receives the treatment) To compare treatments we need to know the inherent variability of units receiving the same treatment background noise this might be a sufficient explanation for the observed differences between treatments

Replication: 2 facts Our faith in treatment means will: Increase with greater replication Decrease when noise increases In particular the standard error of difference (SED) between 2 treatment means where: r = (common) replication; s = typical difference between observations from same treatment: SED is the typical difference between 2 treatment means where the treatments don t differ

Validity & Efficiency Validity: The first requirement of an experiment is that it be valid. Otherwise it is at best a waste of time and resources and at worst it is misleading. Efficiency: the use of experimental resources to get the most precise answer to the question being asked, is not an absolute requirement but is certainly desirable because cost is an important aspect of any experiment.

Pseudoreplication - how to invalidate your experiment! Treating multiple measurements on the same unit as if they were measurements on independent units Example: In an experiment testing the effect of a hormone treatment on follicle development, the cow is the experimental unit, not the follicle.

Example: In an experiment to compare three cultivars of grass, a rectangular tray was assigned at random to each treatment. Trays were filled with John Innes Number 2 compost and 54 seedlings of the appropriate cultivar were planted in a rectangular pattern in each tray. After ten weeks the 28 central plants were harvested, dried and weighed and the 84 plant weights recorded. What was the experimental unit?

Example: In an experiment to compare three cultivars of grass, 7 square pots were assigned at random to each treatment. Pots were filled with John Innes number 2 compost and 16 seedlings of the appropriate cultivar planted in a square pattern in each pot. After ten weeks the 4 central plants were harvested, dried and weighed. Thus 84 plant weights were recorded. What is the experimental unit and what should be analysed?

Randomisation - allocating treatments to units Ensures the only systematic force working on experimental units is that produced by the treatments All other factor that might affect the outcome are randomly allocated across the treatments

Randomisation - how it works What do we mean by In a randomised experiment any difference between the mean response on different treatments is due to treatment difference or random variation or both?

The estimated treatment effect is the difference 6.70-5.13 = 1.57 between these two means. It is partly influenced by the treatment effect (2 units) and partly by the variation between experimental units, the background noise. Example: Suppose 8 experimental units, allocated at random to two treatments. Unit 1 2 3 4 5 6 7 8 Response if treated the same 4.1 5.3 7.2 2.6 3.5 6.4 5.5 4.7 Allocated at random to treatment T1 T1 T2 T2 T2 T1 T2 T1 Treatment effect 0 0 2 2 2 0 2 0 Experimental response 4.1 5.3 9.2 4.6 5.5 6.4 7.5 4.7 Mean response T1 5.13 T2 6.70

Now suppose the most extreme allocation, with the poorest experimental units receiving T2. Unit 1 2 3 4 5 6 7 8 Response if treated the same 4.1 5.3 7.2 2.6 3.5 6.4 5.5 4.7 Allocated at random to treatment T2 T1 T1 T2 T2 T1 T1 T2 Treatment effect 2 0 0 2 2 0 0 2 Experimental response 6.1 5.3 7.2 4.6 5.5 6.4 5.5 6.7 Mean response T1 6.10 T2 5.73 The estimated treatment effect is 5.73-6.10 = -0.37. Again it is partly influenced by the treatment effect (+2) and partly by the variation between experimental units, the background noise. The treatment effect is swamped by the extreme allocation.

The estimated treatment effect is the difference 13.73-6.10 = 7.63. Again consider the same extreme allocation but with a larger treatment effect. Unit 1 2 3 4 5 6 7 8 Response if treated the same 4.1 5.3 7.2 2.6 3.5 6.4 5.5 4.7 Allocated at random to treatment T2 T1 T1 T2 T2 T1 T1 T2 Treatment effect 10 0 0 10 10 0 0 10 Experimental response 14.1 5.3 7.2 12.6 13.5 6.4 5.5 14.7 Mean response T1 6.10 T2 13.73

Three points: The observed treatment difference is due only to treatment effect and variation. If the treatment effect is large relative to the background noise then even an extreme allocation will not obscure the treatment effect. (Signal/Noise ratio). If the number of experimental units is large then a treatment effect will usually be more obvious, since an extreme allocation of experimental units is less likely. With 20 experimental units, unlikely that the 10 worst and the 10 best allocated to different treatments.

Tests of Hypotheses - Tests of Significance Survey: Are the observed differences between groups compatible with a view that there are no differences between the populations from which the samples of values are drawn? Designed experiments: Are observed differences between treatment means compatible with a view that there are no differences between treatments?

Tests of Hypotheses - Tests of Significance Designed experiment -only two explanations for a negative answer, difference is due to the applied treatments or a chance effect Survey is silent in distinguishing between various possible causes for the difference, merely noting that it exists.

Example An experiment on artificially raised salmon compared two treatments and 20 fish per treatment. Average gains (g) over the experimental period were 1210 and 1320. Variation between fish within a group was RSE = 135g Did treatment improve growth rate?

Procedure a) NULL HYPOTHESIS Treatments have no effect and any difference observed between groups treated differently is due to chance (variation in the experimental material)' b) Measure -the variation between groups treated differently -the variation expected if due solely to chance c) TEST STATISTIC Compare the two measures of variation. Do treatments produce a 'large' effect?

d) The observed difference could have occurred by chance. Statistical theory gives rules to determine how likely a given difference in variation is liable to be by chance. e) SIGNIFICANCE TEST Face the choice. -This difference in variation could have occurred by chance with probability? (5%, 1%, etc) OR -There is a real difference (produced by treatment). f) GOOD EXPERIMENTAL PROCEDURE makes sure in experiments that there is no other possible explanation.

Example: - The t test An experiment on artificially raised salmon compared two treatments and 20 fish per treatment. Average gains (g) over the experimental period were 1210 and 1320. Variation between fish within a group was RSE = 135g Did treatment improve growth rate?

Example a) NULL HYPOTHESIS - Treatment does not affect salmon growth rate b) Observed difference between groups 1320-1210 = 110 Variation expected solely from chance 135 x (2/20).5 = 42.7 c) Test Statistic t = 110/42.7 = 2.58 d) Statistical theory (t tables) shows that the chance of a value as large as 2.58 is about 1 in 100 e) Make the choice f) Are there other possible explanations?

Responsibility of the Researcher and Statistician; PLANNING PHASE Researcher Seek statistical training Seek statistical advice Use minimum experiment size Select experimental material properly Develop detailed research plan Exercise proper protocols for human or animal experiments Statistician Keep up to date with statistical technology Teach principles Provide statistical input to plan Give researcher different alternatives

EXECUTION PHASE Research Carry out study as planned Log important dates related to data Statistician Give road map/for execution Point out weak links in research chain

ANALYSIS PHASE Researcher Study data patterns Statistician Assist in studying data patterns Keep integrity of data set Choose proper statistical analytical procedure Choose probability levels, contrasts to make, etc. Avoid result guided procedures Assist in choosing analytical procedure Assist in choosing probability levels, contrasts, etc. Keep data, but not statistical methods used, confidential

INTERPRETATION AND REPORTING Researcher Provide description of statistical methods Present results in such a way that reader can evaluate the interpretation Report variability estimates such an standard deviation or variance Statistician Assist in writing statistical methods used Review interpretation. Modification if necessary

Responsibility of Researcher and Statistician Promote high standards of scientific inquiry and professionalism Involve appropriate techniques for research Honor the rights of other researchers give credit to other researcher where due Consider interdependence of natural, social and technological systems Give objectivity a major role Guard against misinterpretation and misuse of data

Good Practices Checklist Planning is very important in experimentation Statistician can assist in planning Planning does not ensure success but avoids built-in disasters Statistics cannot compensate for negative impacts of persisting in a faulty line of research

Good Planning Can Prevent: Costly waste of resources Difficult statistical analysis Data for which interpretation is controversial An experiment which is precise but which answers the wrong questions

Setting Up Original Hypothesis Objectively 2 Rules: 1. Hypothesis should be clearly related to original problem 2. The hypothesis should be stated as simply as possible

IV. Discipline Specific Ethical Issues Flexibility needed: Ethics vary Among Different Application Areas Business Application: Withholding Negative Results Problem Formulation Important Involve Statistician Design Considerations: Costs, Definition of Population, Sampling Frame

Statistician provides report but does not make decisions for management Company should have same responsibilities to a salaried statistician as to a consulting one (and conversely). See: Deming, W.W., Sample Designs in Business Research. Wiley NY 1960.

Medical Application Medical review boards Informed consent Methods of selecting subjects Withholding a treatment to a control group Access to data Confidentiality of identity of subjects

V. Ethical Issues in Interpretation and Reporting Insufficient statistical methods description Statistical significance vs. practical significance Access to data Kinds of means in the factorial experiment reporting Reporting of measures of dispersion Proper decimal reporting Bonafide scientific conclusions vs. speculation Clarity of reporting Indication that results are not final word Enumeration of new study questions

VI. Case Studies Skagerrak Case Precautionary Principle 2 Highly respected scientists interpret their results differently Case emphasized 2 critical aspects of research 1. The actual statistical analysis 2. How and when to disseminate the information from research Elton s Withholding of Anomalous Data US Census: Use of sample survey methods to adjust census counts