Where does "analysis" enter the experimental process?

Similar documents
SAMPLING AND SAMPLE SIZE

QA 605 WINTER QUARTER ACADEMIC YEAR

Response to the ASA s statement on p-values: context, process, and purpose

Assignment #6. Chapter 10: 14, 15 Chapter 11: 14, 18. Due tomorrow Nov. 6 th by 2pm in your TA s homework box

Survival Skills for Researchers. Study Design

Chapter 23. Inference About Means. Copyright 2010 Pearson Education, Inc.

Using The Scientific method in Psychology

04/12/2014. Research Methods in Psychology. Chapter 6: Independent Groups Designs. What is your ideas? Testing

Design of Experiments & Introduction to Research

Chapter 5: Field experimental designs in agriculture

Research Questions, Variables, and Hypotheses: Part 2. Review. Hypotheses RCS /7/04. What are research questions? What are variables?

Fundamental Clinical Trial Design

Lec 02: Estimation & Hypothesis Testing in Animal Ecology

Audio: In this lecture we are going to address psychology as a science. Slide #2

Statistics 2. RCBD Review. Agriculture Innovation Program

What Is Science? Lesson Overview. Lesson Overview. 1.1 What Is Science?

Chapter 19. Confidence Intervals for Proportions. Copyright 2010 Pearson Education, Inc.

Psychology Research Process

Randomized Block Designs 1

VALIDITY OF QUANTITATIVE RESEARCH

STATISTICAL CONCLUSION VALIDITY

Chapter 02. Basic Research Methodology

EXERCISE: HOW TO DO POWER CALCULATIONS IN OPTIMAL DESIGN SOFTWARE

Sample Size, Power and Sampling Methods

Lab 2: The Scientific Method. Summary

Analysis of Environmental Data Conceptual Foundations: En viro n m e n tal Data

Evaluating Social Programs Course: Evaluation Glossary (Sources: 3ie and The World Bank)

Hypothesis Testing. Richard S. Balkin, Ph.D., LPC-S, NCC

Chapter 19. Confidence Intervals for Proportions. Copyright 2010, 2007, 2004 Pearson Education, Inc.

Objectives. Quantifying the quality of hypothesis tests. Type I and II errors. Power of a test. Cautions about significance tests

NATURE OF SCIENCE. Professor Andrea Garrison Biology 3A

Bayesian and Frequentist Approaches

2 Critical thinking guidelines

EXPERIMENTAL RESEARCH DESIGNS

Variables Research involves trying to determine the relationship between two or more variables.

LAB 7: THE SCIENTIFIC METHOD

Lecture 4: Research Approaches

The Scientific Method the process of doing science

Unit 1 Exploring and Understanding Data

Disposition. Quantitative Research Methods. Science what it is. Basic assumptions of science. Inductive and deductive logic

Research Methodology. Characteristics of Observations. Variables 10/18/2016. Week Most important know what is being observed.

Experimental Research in HCI. Alma Leora Culén University of Oslo, Department of Informatics, Design

BIOL 458 BIOMETRY Lab 7 Multi-Factor ANOVA

PSYCHOLOGY 300B (A01) One-sample t test. n = d = ρ 1 ρ 0 δ = d (n 1) d

9 research designs likely for PSYC 2100

The Research Process: Coming to Terms

Overview. Survey Methods & Design in Psychology. Readings. Significance Testing. Significance Testing. The Logic of Significance Testing

Statistical Power Sampling Design and sample Size Determination

Instructions for doing two-sample t-test in Excel

The essential focus of an experiment is to show that variance can be produced in a DV by manipulation of an IV.

Lecture Notes Module 2

Simple Linear Regression the model, estimation and testing

LAW RESEARCH METHODOLOGY HYPOTHESIS

Applied Statistical Analysis EDUC 6050 Week 4

Lecture 01 Analysis of Animal Populations: Theory and Scientific Process

A Brief (very brief) Overview of Biostatistics. Jody Kreiman, PhD Bureau of Glottal Affairs

Describe what is meant by a placebo Contrast the double-blind procedure with the single-blind procedure Review the structure for organizing a memo

Lecture 21. RNA-seq: Advanced analysis

Hypothesis-Driven Research

Psychology Research Process

Inferential Statistics

Lessons in biostatistics

AIM #4: DEFINE SCIENCE AND EXPLAIN THE PROCESS OF THE SCIENTIFIC METHOD

RNA-seq. Design of experiments

Creative Commons Attribution-NonCommercial-Share Alike License

Reliability of Ordination Analyses

The Research Roadmap Checklist

Introduction & Basics

An Introduction to Bayesian Statistics

What you should know before you collect data. BAE 815 (Fall 2017) Dr. Zifei Liu

MEA DISCUSSION PAPERS

CHAPTER OBJECTIVES - STUDENTS SHOULD BE ABLE TO:

MS&E 226: Small Data

Sheila Barron Statistics Outreach Center 2/8/2011

Example 7.2. Autocorrelation. Pilar González and Susan Orbe. Dpt. Applied Economics III (Econometrics and Statistics)

Midterm Exam MMI 409 Spring 2009 Gordon Bleil

Experimental Design Part II

Sociological Research Methods and Techniques Alan S.Berger 1

BIOSTATISTICAL METHODS

Profile Analysis. Intro and Assumptions Psy 524 Andrew Ainsworth

Karl Popper is probably the most influential philosopher

12.1 Inference for Linear Regression. Introduction

Study Guide for the Final Exam

III. WHAT ANSWERS DO YOU EXPECT?

Abdul Latif Jameel Poverty Action Lab Executive Training: Evaluating Social Programs Spring 2009

Lesson 9: Two Factor ANOVAS

Confidence Intervals On Subsets May Be Misleading

Inference with Difference-in-Differences Revisited

Catherine A. Welch 1*, Séverine Sabia 1,2, Eric Brunner 1, Mika Kivimäki 1 and Martin J. Shipley 1

The Role and Importance of Research

Methods for Determining Random Sample Size

Psych 1Chapter 2 Overview

THE SCIENTIFIC WORLD VIEW

Multilevel analysis quantifies variation in the experimental effect while optimizing power and preventing false positives

Chapter 02 Developing and Evaluating Theories of Behavior

How many speakers? How many tokens?:

Lecture Outline. Biost 590: Statistical Consulting. Stages of Scientific Studies. Scientific Method

Business Statistics Probability

10 Intraclass Correlations under the Mixed Factorial Design

Transcription:

Lecture Topic : ntroduction to the Principles of Experimental Design Experiment: An exercise designed to determine the effects of one or more variables (treatments) on one or more characteristics (response variables) of some well-defined system (experimental unit). Where does "analysis" enter the experimental process? The "Scientific Method" Formulate a hypothesis Plan an experiment Conduct the experiment Analyze and interpret the results Steps in experimentation (adapted from Little and Hills 978) Define the question Determine the intended scope of conclusions -- Select the experimental material Select the treatments Select the experimental unit and number of replications (choose level of risk) Ensure proper randomization and layout Ensure proper means of data collection Conduct the experiment Analyze the data -- nterpret the results Effectively communicate the results Experimental Design the logical structure of an experiment

Hypothesis, scope, and experimental design The selection of an experimental design depends on your objective (hypothesis + scope). Variety screening trial: Maximize accession number (less replication). Variety release trial: Appropriate replication for high precision. The intended scope of conclusions is a major determinant of experimental design. Variety release trial for northern New Hampshire. Variety release trial for New England. "Clearly the instruction, 'Observe!' is absurd.observation is always selective. t needs a chosen object, a definite task, an interest, a point of view, a problem." Karl R. Popper Conjectures and Refutations: The Growth of Scientific Knowledge Concepts about hypotheses A hypothesis is a statement that can be tested and falsified. A null hypothesis (H 0 ) can never be proven correct. t can only be rejected with known (chosen) risks of being wrong. A good experimental design allows the quantification of uncertainty. The experiment is designed so that it one can calculate the possibility of obtaining the observed results by chance alone.

Testing hypotheses The significance level (α) is the probability that one rejects H 0 when it is true (Type error rate). For many fields, default α 0.05 ( error in 0). Lowering α protects you from false positives but increases the Type error rate (β), the probability of failing to reject H 0 when it is false. H 0 is rejected is not rejected is true Type error Correct decision is false Correct decision Type error The only way to reduce both types of error is to increase the number of replications or to improve the experimental design. "The purpose of statistical science is to provide an objective basis for the analysis of problems in which the data depart from the laws of exact causality." D.J. Finney Experimental Design: The logical structure of an experiment.. Treatment structure The set of treatments used and how they relate to one another.. Treatment replication The number of experimental units to be subjected to each treatment. 3. Design structure The manner in which treatments are assigned to experimental units. 4. Response structure The set of response variables to be measured and the sampling plan that specifies when, where, and with what components of the experimental unit one will measure those response variables. 5. Error control "Noise" reduction through the strategic use of blocking techniques, covariables, or environmental controls (e.g. growth chambers, greenhouses, lab studies). Among the above considerations, replication and randomization are the most important basic principles in designing experiments. 3

Replication The functions of replication. To provide an estimate of the experimental error. The experimental error is the variation which exists among experimental units that are treated alike.. To improve the precision of an experiment. Replication reduces the standard error and thus improves precision. 3. To increase the scope of inference. Replication refers to the number of experimental units that are treated alike. Experimental unit: The smallest system or unit of experimental material to which a single treatment (or treatment combination) is assigned and which is dealt with independently of other such systems under that treatment at all stages in the experiment in which important variation may enter. (Hurlbert 006) Each replication is independent from every other. 4

Example: A field trial comparing three different fertilizers (A, B, C). Response variable: Yield (kg). 3 4 5 6 7 8 9 0 Case : Four plots are selected at random to receive each fertilizer. The total yield of each plot is measured at the end of the season. B A C B C A B A C B A C This is a completely randomized design (CRD) with four replications per treatment. f the experiment were repeated the following year, with new random assignments of treatments, this could provide a second set of replications. 5

Case : Same as Case, but now the crop is a perennial. The total yield of each plot is measured at the end of each season. n this case, the plots in the second year are not replications; they are repeated measurements. Case 3: Same as Case, except each of the plots is further divided into three subplots and yield is measured separately for each subplot. B B B A A A C C C B B B C C C A A A B B B A A A C C C B B B A A A C C C The experimental unit is still the plot. The subplots are not replications, they are subsamples. Case 4: The three treatment levels are randomly assigned to the three rows in the field. Yield is measured on each plot. B B B B A A A A C C C C n this case, the experimental unit is the row. Each plot is a subsample. This experiment has no replication. WHAT WAS RANDOMZED? 6

Type and number of measurements The theoretical consideration: σ Y ( n) σ n The standard error determines the lengths of confidence intervals and the powers of tests. Decreasing the standard error increases the precision of the experiment. Precision has to do with the concept of random errors, and the precision of an average can always be improved by increasing the sample size (n). A good experimental design has sufficient precision, meaning there is a high probability that the experiment is able to detect expected differences. The practical consideration: Finite resources. Subsampling: Multiple measurements on the same experimental unit can yield a more precise value for that experimental unit. This can reduce the apparent variation among experimental units subjected to the same treatment and thereby reduce the standard error. σ Y ( n) σ n A strategic combination of replications and subsamples can be used to achieve the necessary precision, given limited resources. 7

Other forms of error control: Homogeneous experimental materials Good equipment Careful observation Well-trained assistants Blocking Covariables Another practical consideration: Statistical significance Biological significance Too many measurements without real purpose can lead to the statistical declaration of significance, even in the absence of meaningful biological effect: With a 5% Type error rate and 00 comparisons, one can expect about 0 false positives (Type errors). Thus, it is important to define biological significance and then design experiments to detect this amount, no more and no less. With sufficient replication, a difference can always be found: With intense sampling, one could show that resource-intensive mitigation efforts reduces active N in surface waters by ppm. But who cares? 8

Design structure (the assignment of treatments to experimental units) The functions of randomization:. To neutralize systematic biases. Proper randomization helps provide valid estimates of experimental error and treatment means.. To ensure independence of errors, an assumption of many statistical tests. Not accurate, not precise Accurate, but not precise Precise, but not accurate Accurate and precise A good experimental design is characterized by the absence of systematic error. Experimental units should not differ in any systematic way from one another. Treatment structure The type and number of treatment levels are important considerations, particularly when the treatments are quantitative:. Separation of levels at similar intervals can facilitate comparisons and interpretation.. The number of levels sets the limit on the detectable complexity of the response. A final aesthetic consideration A good experimental design is as simple as possible for the desired objective. 9

Relative precision of designs involving few treatments Precision, sensitivity, or amount of information is measured as the reciprocal of the variance of the means. f we let represent the amount of information contained in a sample mean, then: Y (n) Y ( n) σ Y ( n) n σ Thus the amount of information per observation is: σ f s is used to estimate σ there is a correction to this formula: + ) + 3) s Note that when n à, then the correction factor (df+)/(df+3) à. The amount of information provided by each experimental unit (i.e. independent observation) in a given experiment is: e e + ) + 3) MSE The relative efficiency of Design relative to Design is calculated as the ratio of the amount of information in the two designs: RE to e+ ) e+ 3) MSE e+ ) + 3) MSE e e e + )e + ) e + 3) MSE + 3) MSE f this ratio is greater than, Design provides more information and is more efficient than Design. 0

SO, when designing your own experiments, when reviewing the experiments of others, when considering whether or not to contribute your efforts to a pre-conceived experiment, there are some basic questions you should answer: FUNDAMENTAL. What exactly is your question? What falsifiable hypothesis is genuinely put at risk by this experiment? Where does this hypothesis lie in the larger logical tree of hypotheses?. What is the intended scope for your conclusions? 3. What constitutes a "valid" result? NET. Are the methods appropriate to the question and the intended scope for conclusions (i.e. the Objective)? What are the treatments? What is the experimental unit? What is the unit of observation? How were each selected, manipulated, and observed?. Do the data meet the assumptions of the analysis? 3. s there sufficient replication to "falsify" the hypothesis? "Get it right or let it alone. The conclusion you jump to may be your own." James Thurber Further Fables for Our Time