A Short DOE Glossary. blocking. comparative experiment. confounding or aliasing. design constraint
|
|
- Elfreda Lang
- 5 years ago
- Views:
Transcription
1 A Short DOE Glossary blocking An experimental technique that allows the possible effects of known but uncontrolled variables to be completely eliminated from an experiment. Here is a simple example. Suppose you wanted to do an experiment comparing basketball-shooting accuracy with the dominant hand (e.g. right hand for righthanders) vs. the nondominant hand. One way to do the experiment would be to randomly select 10 people and the hand they shoot with, and then compare the overall right hand results with the overall left hand results. In this way of doing things, it is quite possible that any difference between the hands would get washed out by the large overall difference in individual basketball shooting ability. A better way to do the experiment would be to randomly select 5 people and have each shoot both with their right and left hands (perhaps varying which hand they use in random order). One can then look at the difference for each person and average these differences as an overall measure of right vs/ left ability. Because the variation between people cancel out when things are done this way, this variation is completely eliminated from the experiment -- it as if it didn't even exist. Done this way, the experiment has been blocked on people. Of course, this is the simplest sort of blocking that can occur, but it does illustrate the idea. comparative experiment Any experiment whose purpose is to determine the quantitative effect of input(s) that are deliberately changed (the experimental "variables" or "factors")on measured output(s) (the "response(s)"). confounding or aliasing Main and/or interaction effects are said to be confounded or aliased if only their combined effect, not their separate individual effects, can be determined from the experimental design. As a very simple example, suppose that one ran two experimental trials to determine the effect of teaching method and instructor on student performance. A group of 40 students are randomly split into two groups of 20. Half are assigned to Teacher A using method 1 to teach, say, factoring in a math class. The other half are assigned to Teacher B using method 2. After the factoring unit is covered, the performance of the two groups is assessed by comparing student scores on a common exam. Clearly, any systematic difference between the groups can only be ascribed to a combined effect of different teacher and different method, as both teacher and method change together. In DOE terminology, the effect of teacher and teaching method are fully confounded or aliased with one another. Note that no amount of data analysis can determine the separate effects. The aliasing is inherent to the design. Although it may seem that one would always want to avoid such aliasing, it turns out that this is not the case -- and is essentially unavoidable anyway. Indeed, proper control of aliasing turns out to be one of the keys in the sequential design strategy. Note, also, that one may also partially confound effects in a design. This is an advanced (but quite useful) technique. design constraint A mathematical or physical limitation that restricts the possible combinations of the factors that can be tried. For example, in a mixtures experiment (an experiment in which the factors are proportions of the mixture ingredients), any given combination of proportions must always add up to 100% (of course!). In a 1
2 chemical experiment in which the factors are concentrations of various chemicals, certain concentrations may be explosive and so must be avoided. continuous or measurement-type factor An experimental factor that can, in principle, be set anywhere within its experimental range for an experimental trial. Examples are temperature, time, ph, amount of fertilizer added to the soil, height, weight, and so forth. D-optimal design criterion D-optimality is a mathematical technique that is sometimes useful in producing experimental designs, especially in nonstandard and irregular (e.g., not hpercubes or hyperspheres) design spaces. Special purpose computer software is required to use this method as the computations are far too extensive to be done by hand. For those who might care, finding D-optimal designs is an NP-complete problem, so that such designs are only approximated by the software. design resolution The degree of confounding present in a design. Design resolution refers to the amount of detail -- separate identification of factor effects and interactions -- that the design supports. This is only relevant for multifactor, not OFAT, experiments. discrete or categorical factor An experimental factor that can be set only at distinct, separate levels. For example, male and female (in an experiment on fish behavior); metal, glass, or plastic stirrer in a chemical experiment; type of soil -- sandy, clay, loam, gravel (in an experiment on plant growth). Note that categorical factors can be either unordered (male/female) or ordered (number of times the rat previously traversed the maze). efficiency of an experimental design The amount of information generated by an experimental design. Equivalently, the precision in the fitted coefficients of the response surface. Although a complete explanation of this is rather technical, what it comes down to is a way of defining the amount of averaging that the design can achieve. A more efficient design is equivalent to saying that it generates more information which is equivalent to saying that the response surface is known with greater precision which is equivalent to saying that there is less uncertainty in the conclusions.the important idea is that for a fixed amount of experimental effort, usually the more efficient the design, the better. experimental precision The amount of experimental variability that exists, usually determined from the variability of replicated trials. The greater the precision, the less variability there is and the less uncertainty there is in the results, including the fitted response surface. 2
3 Experimental bias The tendency of an experiment to produce results that systematically differ from the true results. For example, a result may be "biased high" if an instrument is improperly operated. A biased measurement is a measurement that is either higher or lower on average than it should be. Experimental variability, error, or noise These words are used synonymously to refer to the fact that when experimental trials are repeated without changing the settings of the factors, the response varies rather than remaining constant. This is due, of course, to the hopefully small effects of changes in many uncontrolled factors that exist in any experiment or measurement. It is never possible to exactly repeat anything. In order to quantify such variability -- which is necessary in order to properly assess how the response depends on the experimental factors -- statistical methods must be used. (experimental) factor or variable Variables that are deliberately manipulated in an experiment in order to assess their effect on the response. For example, in an experiment to assess the effect of various lengths, diameters, and materials on voltage drop across a length of wire, the experimental factors are length, diameter, and material of which the wire is made. The response is the measured voltage drop. factor level The setting of an experimental factor. Typically, in DOE, continuous factors are "standardized" to the range from -1 to +1. For example, if temperature is an experimental factor that is to be varied between 30 and 60 C., then convert 30 to -1, 60 to +1 and linearly interpolate any value between (e.g., 50 interpolates to +1/3). This is equivalent to making a simple linear scale change (like Fahrenheit to centigrade, pounds to kilograms, and so forth). Note, however, that with categorical factors, the ±1 standardization can only be done when there are exactly two categories. When there are more, it makes no sense because it would convert a non-ordered identifying label (which variety of 3 seed varieties) to an ordered scale (-1,0,+1). For this reason, advanced methods must be used to design experiments with categorical factors having more than two categories. Fisher, Sir Ronald A. ( ) The famous British geneticist and statistician who originated and developed the foundations of experimental design. His books, Statistical Methods for Research Workers and The Design of Experiments are classic. Much standard statistical terminology -- like anova and randomization test -- derives from his work. The "F" of the statistical F distribution (upon which the F test is based) is named after him. foldover A sequential design technique that produces a "mirror image" of a given design in order to separate confounded interactions. Generally, this converts designs of resolution III to designs of resolution IV. 3
4 fractional factorial design A fractional factorial design is a design in which only a selected fraction of all the possible combinations of the design factors are run. For the two level hypercube designs, this means only a subset of all the hypercube corners are actually run. hidden replication An experimental design is said to permit hidden replication when, if some of the factors can be safely ignored, there is replication in the remaining factors. For example, suppose one did a 2-factor experiment in which runs at the the four different combinations (-,-), (-,+), (+,-), and (+,+) were conducted. If the second of the two factors had essentially no effect, then, so far as Mother Nature was concerned, a one factor experiment in which the + and - settings was replicated twice was actually done. This replication was "hidden" until it became clear that the second factor could be safely ignored. Hidden replication is most commonly used in "screening" experiments with many (more than 4, say) factors. hypercube The equivalent of a cube in an arbitrary number of dimensions. In DOE, hypercubes are usually stanardized so that all coordinate entries are ±1. Hence a 2-d hypercube is just the ordinary square with 4 corners at (-1,-1), (-1,+1), (+1,-1), and (+1,+1). To save writing, the 1's are usually omitted. Hence, we would give the corners simply as (-,-), (-,+), (+,-), and (+,+). Using this convention, a 3-d hypercube is just an ordinary cube with 8 corners at (-,-,-), (+,-,-), (-,+,-), (+,+,-), (-,-,+), (+,-,+), (-,+,+), and (+,+,+). And so on with 4,5, and more dimensions. Note that in 2 dimensions, there are 2^2=4 corners; in 3, there are 2^3=8; in 4, there are 2^4=16; and, in general, in n dimensions, a hypercube has 2^n corners. interaction A 2-factor interaction(2 fi) is the difference in the response that occurs when both factors are changed simultaneously from what was expected to occur based on the effect of changing the factors individually. When the combined effect is significantly greater than the sum of the individual effects, it is often called symbiosis; when it is significantly less, it is often called interference. Algebraically, a 2-factor interaction is represented by the presence of a cross product term (factor_1 * factor_2) in the model. Graphically, 2 fi's are indicated by significant non-parallelism of the two lines in an interaction plot. An example of such a plot is provided in the herbicide example. Higher order interactions -- that is 3 or more factor interactions -- also rarely may be important. However, these require more complicated designs with more experimental trials than are typically used. So in the basic approach followed in the DOE project, they are not considered. multifactor design An experiment with several experimental factors in which more than one factor at a time changed. 4
5 parsimony, Occam's razor, or the Pareto Principle All of these terms are used equivalently here and refer to the "vital few; trivial many" principle. That is, in any real experiment in which many factors are considered, almost always, only a very small proportion of them will have most of the effect. The rest should be treated as the "trivial many" and considered to be indistinguishable from experimental noise. So in trying to build a model(=fit a simple algebraic equation in basic DOE) to describe the experimental results, one should try to find one that uses as few factors (= parameters =coefficients) as possible. That is, one should be as parsimonious in using coefficients as possible. Occam's Razor refers to the idea that if several models (theoretical or experimental) do equally well in explaining what is observed, than the simplest one (fewest parameters) should be chosen. randomization Running the experimental trials in a random order. This is done to protect against the systematic effects of unknown non-experimental variables (like environment) that might bias the experimental results. There are also other ways in which random assignment is used. For example, in doing clinical trials to determine efficacy and safety of new drugs, patients are almost always assigned randomly to the treatment (receive the drug) vs. the control (receive an inactive placebo) group. This prevents unconscious biases (for example, assigning sicker people to receive the drug) from influencing the experimental results. replication Repeating an experimental trial at constant factor settings in order to determine the amount of experimental variability. Since the settings of the experimental factors do not change, observed variability in the response must be due to the effects of other, uncontrolled factors that are present throughout the experiment. This includes measurement factors. It is important when doing replicates NOT to do them close together in time under nearly identical circumstances. Rather, the replicates should be done over the same range on conditions in which the entire experiment is performed. This allows all the experimental variability that is actually present to be observed and quantified. residual analysis Residual are what's "left over" from the data after a model has been fit. That is, the residuals are defined as: residual = actual data value - value predicted by fitted model If the model fits well, then all systematic behavior is predicted by the model and the residuals should look like random noise. When residuals depart from this behavior and exhibit systematic trends or dependencies, the model may need to be modified. This, in turn, may require appropriate design changes at the next stage of the experimental process. response surface 5
6 The higher dimensional "surface" of true responses (that is, absent all extraneous experimental variatibility) obtained from all possible combinations of settings of the experimental factors (over their allowable experimental ranges). Knowledge of this surface is equivalent to a complete understanding of how the response depends on the experimental factors. If some or all the factors are categorical, the "surface" may actually be isolated points, curves, or other lower dimensional structures. response variable A measured experimental outcome. Response variables may come in many forms. For example, the response in a physics experiment exploring the effects of different numbers of windings and currents on the performance of an electromagnet could be the magnetic force generated. In an industrial experiment on a chemical process, the response might be the yield of the product. In an experiment to develop a new cake mix, the response variables might be taste and texture as rated by a panel of raters on a 1 to 10 scale. In an experiment to see what effect height, sex, and distance from the basket have on foul shooting accuracy, the response could be the number of baskets made out of ten tries. A single experiment might have several response variables that characterize different aspects of the outcome. The key idea is that there must be some kind of "reliable" measurement that is made that can be used for analysis of the results. What is meant by "reliable" is, itself, a complex statistical issue. response surface methods A broad category of experimental design and analysis methods based on fitting models which are linear and quadratic equations in the experimental factors (this includes cross-terms for interactions). Such purely empirical models are useful for describing systems behavior, process improvement, and often increasing understanding so that more detailed conceptual (mechanistic) models can be developed. screening design A screening design is one in which relatively few experimental runs are used to efficiently study a large number of experimental factors to "screen out" those few that are most active from the remainder that are relatively inactive over the ranges being considered. Such designs are very useful in the early stages of sequential experimentation in order to conserve resources and identify the most influential experimental factors for more detailed study. Other essentially synonomous terms for this are "Resolution III," "Plackett-Burman," and "Saturated" design. sequential experimental strategy Sequential experimentation means investigations that are carried out in stages so that each successive experiment can be designed and executed in the light of information gained from previous ones. This is really a description of a scientific learning strategy that encourages the efficient expenditure of limited experimental resources. Although most experimenters intuitively try to do things this way, there are specific design and analytical tools in DOE that have been rigorously developed for this purpose. Some of these procedures are: Residual analysis, fractional factorial designs, design resolution, sequential assembly, foldover, 6
7 response surface methods, D-optimality design criteria, and steepest ascent/gradient optimization It is important to emphasize that this provides experimenters a systematic framework -- not merely an artful philosophy -- in which to execute the strategy. This gives greater control and improved likelihood of success. sequential assembly of designs Building and performing complex experimental designs one stage at a time. Later stages are added only when and if needed. This conserves experimental resources while yielding the maximal information at each stage of the assembly process. Split-plotting A method for running experiments in non-random fashion when not all experimental factors cannot be completely randomized. This is an advanced topic that requires the use of nested ANOVA. statistical model A statistical model is an algebraic equation that expresses how a response of interest is related to the experimental factors and the experimental variability. For example, (I): Resp = K + A*Factor_1 + B*Factor_2 + C*Factor_1*Factor_2 + random_variability is such a model. "K", "A", "B", and "B" are unknown "parameters" or "coefficients" that must be estimated from the experimental data. Factor_1 and Factor_2 are the (known)settings of the the two experimental factors at which the response is actually measured in the experiment. This model is said to be linear because the response is a linear function of the unknown coefficients. This can be a bit confusing, because roles of unknown coefficient and known variable setting reverse what we are accustomed to in such equations. For example, the model (II): Resp = K + A*[Factor_1]^2 + B*Factor_1 + C*Factor_2 + random_variability is also linear for the same reason, even though Factor_1 now also appears as a squared term. In fact, in basic DOE only models of type (I) are usually considered. These suffice for many applications. steepest ascent/gradient methods Methods for improvement based on experiments and analysis that model the response surface as a "mountain" (in n dimensions). The fastest way to climb such a mountain -- that is, the path of steepest ascent -- is to go straight up the sides. By mathematically determining this direction, one can determine how to change the experimental factors to effect the greatest possible change in the response. 7
8 The Scientific Method Some of the fundamental ways in which science is different than philosophy or art or literature surely must include: 1. Science is about predicting observable phenomena. Merely giving explanations after the fact is not good enough. You must predict what will be observed before it is observed. The concept of observable phenomena is also central. This means that given instructions on how to construct measurement equipment, anyone who produces the equipment should be able to measure the "same" results (within experimental variability). Science is democratic and replicable. That is, observation should not depend on who we are, what beliefs we hold, or what salary we make. Of course, the predictions may be probabilistic and involve a level of uncertainty: we only know that the likelihood of thunderstorms is higher under some conditions than others; or that a major earthquake will almost certainly occur along the San Andreas fault within 100 years. Although such predictions involve uncertainty, they are just as legitimate science as, say, the prediction of a space shuttle's orbit. 2. Predictions are made by the development of scientific models. Science is not about discovering eternal truths; rather, it is about developing models from which precise and accurate predictions can be made. On the most fundamental level, science does not discuss truth or the underlying nature of reality at all -- this is the realm of philosophy. All scientific models, can be flawed or incomplete in some respects but still be useful to make predictions within certain defined realms. Another way of saying this, is that all scientific models are falsifiable, but none can be proven (unlike mathematics). There may always be another consequence that observation will contradict. 3. Models are usually, but not always, quantitative and expressed mathematically. One can broadly distinguish two overlapping kinds of scientific models: mechanistic or conceptual models, in which some kind of theoretical construct is used to develop the model; and empirical models which are based exclusively on observed data (and use statistical analysis to develop predictions of what will be observed in the future under other conditions). Overlap occurs, because extended observation usually motivates the development of conceptual models, and conceptual models must always be criticized (i.e. put to the test) by real data. 4. Because all science involves observable phenomena, the inevitable presence of some uncontrolled variability in all observations means that no scientific observation is exactly known -- "we see through a glass darkly." All scientific observation therefore involves uncertainty, and the uncertainty must explicitly and quantitatively be dealt with as part of the process of scientific learning. Contrast this with philosophy or religion or literature, for example. 8
9.0 L '- ---'- ---'- --' X
352 C hap te r Ten 11.0 10.5 Y 10.0 9.5 9.0 L...- ----'- ---'- ---'- --' 0.0 0.5 1.0 X 1.5 2.0 FIGURE 10.23 Interpreting r = 0 for curvilinear data. Establishing causation requires solid scientific understanding.
More information9 research designs likely for PSYC 2100
9 research designs likely for PSYC 2100 1) 1 factor, 2 levels, 1 group (one group gets both treatment levels) related samples t-test (compare means of 2 levels only) 2) 1 factor, 2 levels, 2 groups (one
More information1.4 - Linear Regression and MS Excel
1.4 - Linear Regression and MS Excel Regression is an analytic technique for determining the relationship between a dependent variable and an independent variable. When the two variables have a linear
More informationINVESTIGATING FIT WITH THE RASCH MODEL. Benjamin Wright and Ronald Mead (1979?) Most disturbances in the measurement process can be considered a form
INVESTIGATING FIT WITH THE RASCH MODEL Benjamin Wright and Ronald Mead (1979?) Most disturbances in the measurement process can be considered a form of multidimensionality. The settings in which measurement
More information3 CONCEPTUAL FOUNDATIONS OF STATISTICS
3 CONCEPTUAL FOUNDATIONS OF STATISTICS In this chapter, we examine the conceptual foundations of statistics. The goal is to give you an appreciation and conceptual understanding of some basic statistical
More informationChapter 3: Examining Relationships
Name Date Per Key Vocabulary: response variable explanatory variable independent variable dependent variable scatterplot positive association negative association linear correlation r-value regression
More informationMULTIPLE LINEAR REGRESSION 24.1 INTRODUCTION AND OBJECTIVES OBJECTIVES
24 MULTIPLE LINEAR REGRESSION 24.1 INTRODUCTION AND OBJECTIVES In the previous chapter, simple linear regression was used when you have one independent variable and one dependent variable. This chapter
More informationReliability, validity, and all that jazz
Reliability, validity, and all that jazz Dylan Wiliam King s College London Introduction No measuring instrument is perfect. The most obvious problems relate to reliability. If we use a thermometer to
More information(CORRELATIONAL DESIGN AND COMPARATIVE DESIGN)
UNIT 4 OTHER DESIGNS (CORRELATIONAL DESIGN AND COMPARATIVE DESIGN) Quasi Experimental Design Structure 4.0 Introduction 4.1 Objectives 4.2 Definition of Correlational Research Design 4.3 Types of Correlational
More informationSection 3.2 Least-Squares Regression
Section 3.2 Least-Squares Regression Linear relationships between two quantitative variables are pretty common and easy to understand. Correlation measures the direction and strength of these relationships.
More informationUnit 1 Exploring and Understanding Data
Unit 1 Exploring and Understanding Data Area Principle Bar Chart Boxplot Conditional Distribution Dotplot Empirical Rule Five Number Summary Frequency Distribution Frequency Polygon Histogram Interquartile
More informationBayesian and Frequentist Approaches
Bayesian and Frequentist Approaches G. Jogesh Babu Penn State University http://sites.stat.psu.edu/ babu http://astrostatistics.psu.edu All models are wrong But some are useful George E. P. Box (son-in-law
More informationOptimization and Experimentation. The rest of the story
Quality Digest Daily, May 2, 2016 Manuscript 294 Optimization and Experimentation The rest of the story Experimental designs that result in orthogonal data structures allow us to get the most out of both
More informationKepler tried to record the paths of planets in the sky, Harvey to measure the flow of blood in the circulatory system, and chemists tried to produce
Stats 95 Kepler tried to record the paths of planets in the sky, Harvey to measure the flow of blood in the circulatory system, and chemists tried to produce pure gold knowing it was an element, though
More informationChapter 5: Field experimental designs in agriculture
Chapter 5: Field experimental designs in agriculture Jose Crossa Biometrics and Statistics Unit Crop Research Informatics Lab (CRIL) CIMMYT. Int. Apdo. Postal 6-641, 06600 Mexico, DF, Mexico Introduction
More informationThe Pretest! Pretest! Pretest! Assignment (Example 2)
The Pretest! Pretest! Pretest! Assignment (Example 2) May 19, 2003 1 Statement of Purpose and Description of Pretest Procedure When one designs a Math 10 exam one hopes to measure whether a student s ability
More informationHow Does Analysis of Competing Hypotheses (ACH) Improve Intelligence Analysis?
How Does Analysis of Competing Hypotheses (ACH) Improve Intelligence Analysis? Richards J. Heuer, Jr. Version 1.2, October 16, 2005 This document is from a collection of works by Richards J. Heuer, Jr.
More informationLesson 9: Two Factor ANOVAS
Published on Agron 513 (https://courses.agron.iastate.edu/agron513) Home > Lesson 9 Lesson 9: Two Factor ANOVAS Developed by: Ron Mowers, Marin Harbur, and Ken Moore Completion Time: 1 week Introduction
More informationAudio: In this lecture we are going to address psychology as a science. Slide #2
Psychology 312: Lecture 2 Psychology as a Science Slide #1 Psychology As A Science In this lecture we are going to address psychology as a science. Slide #2 Outline Psychology is an empirical science.
More informationQuality Digest Daily, March 3, 2014 Manuscript 266. Statistics and SPC. Two things sharing a common name can still be different. Donald J.
Quality Digest Daily, March 3, 2014 Manuscript 266 Statistics and SPC Two things sharing a common name can still be different Donald J. Wheeler Students typically encounter many obstacles while learning
More informationQA 605 WINTER QUARTER ACADEMIC YEAR
Instructor: Office: James J. Cochran 117A CAB Telephone: (318) 257-3445 Hours: e-mail: URL: QA 605 WINTER QUARTER 2006-2007 ACADEMIC YEAR Tuesday & Thursday 8:00 a.m. 10:00 a.m. Wednesday 8:00 a.m. noon
More informationChapter 1: Explaining Behavior
Chapter 1: Explaining Behavior GOAL OF SCIENCE is to generate explanations for various puzzling natural phenomenon. - Generate general laws of behavior (psychology) RESEARCH: principle method for acquiring
More informationCHAPTER ONE CORRELATION
CHAPTER ONE CORRELATION 1.0 Introduction The first chapter focuses on the nature of statistical data of correlation. The aim of the series of exercises is to ensure the students are able to use SPSS to
More informationApplied Analysis of Variance and Experimental Design. Lukas Meier, Seminar für Statistik
Applied Analysis of Variance and Experimental Design Lukas Meier, Seminar für Statistik About Me Studied mathematics at ETH. Worked at the statistical consulting service and did a PhD in statistics (at
More informationAbout the Article. About the Article. Jane N. Buchwald Viewpoint Editor
Do Bariatric Surgeons and Co-Workers Have An Adequate Working Knowledge of Basic Statistics? Is It Really Important? Really Important? George Cowan, Jr., M.D., Professor Emeritus, Department of Surgery,
More informationAnalysis of Environmental Data Conceptual Foundations: En viro n m e n tal Data
Analysis of Environmental Data Conceptual Foundations: En viro n m e n tal Data 1. Purpose of data collection...................................................... 2 2. Samples and populations.......................................................
More informationRegression Discontinuity Analysis
Regression Discontinuity Analysis A researcher wants to determine whether tutoring underachieving middle school students improves their math grades. Another wonders whether providing financial aid to low-income
More informationClever Hans the horse could do simple math and spell out the answers to simple questions. He wasn t always correct, but he was most of the time.
Clever Hans the horse could do simple math and spell out the answers to simple questions. He wasn t always correct, but he was most of the time. While a team of scientists, veterinarians, zoologists and
More informationEXPERIMENTAL DESIGN Page 1 of 11. relationships between certain events in the environment and the occurrence of particular
EXPERIMENTAL DESIGN Page 1 of 11 I. Introduction to Experimentation 1. The experiment is the primary means by which we are able to establish cause-effect relationships between certain events in the environment
More information1 The conceptual underpinnings of statistical power
1 The conceptual underpinnings of statistical power The importance of statistical power As currently practiced in the social and health sciences, inferential statistics rest solidly upon two pillars: statistical
More informationStatistics 2. RCBD Review. Agriculture Innovation Program
Statistics 2. RCBD Review 2014. Prepared by Lauren Pincus With input from Mark Bell and Richard Plant Agriculture Innovation Program 1 Table of Contents Questions for review... 3 Answers... 3 Materials
More informationReliability, validity, and all that jazz
Reliability, validity, and all that jazz Dylan Wiliam King s College London Published in Education 3-13, 29 (3) pp. 17-21 (2001) Introduction No measuring instrument is perfect. If we use a thermometer
More informationChapter 11 Nonexperimental Quantitative Research Steps in Nonexperimental Research
Chapter 11 Nonexperimental Quantitative Research (Reminder: Don t forget to utilize the concept maps and study questions as you study this and the other chapters.) Nonexperimental research is needed because
More informationTHE USE OF MULTIVARIATE ANALYSIS IN DEVELOPMENT THEORY: A CRITIQUE OF THE APPROACH ADOPTED BY ADELMAN AND MORRIS A. C. RAYNER
THE USE OF MULTIVARIATE ANALYSIS IN DEVELOPMENT THEORY: A CRITIQUE OF THE APPROACH ADOPTED BY ADELMAN AND MORRIS A. C. RAYNER Introduction, 639. Factor analysis, 639. Discriminant analysis, 644. INTRODUCTION
More informationC-1: Variables which are measured on a continuous scale are described in terms of three key characteristics central tendency, variability, and shape.
MODULE 02: DESCRIBING DT SECTION C: KEY POINTS C-1: Variables which are measured on a continuous scale are described in terms of three key characteristics central tendency, variability, and shape. C-2:
More informationHuman intuition is remarkably accurate and free from error.
Human intuition is remarkably accurate and free from error. 3 Most people seem to lack confidence in the accuracy of their beliefs. 4 Case studies are particularly useful because of the similarities we
More informationOverview of the Logic and Language of Psychology Research
CHAPTER W1 Overview of the Logic and Language of Psychology Research Chapter Outline The Traditionally Ideal Research Approach Equivalence of Participants in Experimental and Control Groups Equivalence
More information12/31/2016. PSY 512: Advanced Statistics for Psychological and Behavioral Research 2
PSY 512: Advanced Statistics for Psychological and Behavioral Research 2 Introduce moderated multiple regression Continuous predictor continuous predictor Continuous predictor categorical predictor Understand
More informationChapter 4 DESIGN OF EXPERIMENTS
Chapter 4 DESIGN OF EXPERIMENTS 4.1 STRATEGY OF EXPERIMENTATION Experimentation is an integral part of any human investigation, be it engineering, agriculture, medicine or industry. An experiment can be
More informationCRITERIA FOR USE. A GRAPHICAL EXPLANATION OF BI-VARIATE (2 VARIABLE) REGRESSION ANALYSISSys
Multiple Regression Analysis 1 CRITERIA FOR USE Multiple regression analysis is used to test the effects of n independent (predictor) variables on a single dependent (criterion) variable. Regression tests
More informationThe Practice of Statistics 1 Week 2: Relationships and Data Collection
The Practice of Statistics 1 Week 2: Relationships and Data Collection Video 12: Data Collection - Experiments Experiments are the gold standard since they allow us to make causal conclusions. example,
More informationRESPONSE SURFACE MODELING AND OPTIMIZATION TO ELUCIDATE THE DIFFERENTIAL EFFECTS OF DEMOGRAPHIC CHARACTERISTICS ON HIV PREVALENCE IN SOUTH AFRICA
RESPONSE SURFACE MODELING AND OPTIMIZATION TO ELUCIDATE THE DIFFERENTIAL EFFECTS OF DEMOGRAPHIC CHARACTERISTICS ON HIV PREVALENCE IN SOUTH AFRICA W. Sibanda 1* and P. Pretorius 2 1 DST/NWU Pre-clinical
More informationIAPT: Regression. Regression analyses
Regression analyses IAPT: Regression Regression is the rather strange name given to a set of methods for predicting one variable from another. The data shown in Table 1 and come from a student project
More informationmultilevel modeling for social and personality psychology
1 Introduction Once you know that hierarchies exist, you see them everywhere. I have used this quote by Kreft and de Leeuw (1998) frequently when writing about why, when, and how to use multilevel models
More informationTo conclude, a theory of error must be a theory of the interaction between human performance variability and the situational constraints.
The organisers have provided us with a both stimulating and irritating list of questions relating to the topic of the conference: Human Error. My first intention was to try to answer the questions one
More informationExperimental and survey design
Friday, October 12, 2001 Page: 1 Experimental and survey design 1. There is a positive association between the number of drownings and ice cream sales. This is an example of an association likely caused
More informationVariable Data univariate data set bivariate data set multivariate data set categorical qualitative numerical quantitative
The Data Analysis Process and Collecting Data Sensibly Important Terms Variable A variable is any characteristic whose value may change from one individual to another Examples: Brand of television Height
More informationMeasurement and meaningfulness in Decision Modeling
Measurement and meaningfulness in Decision Modeling Brice Mayag University Paris Dauphine LAMSADE FRANCE Chapter 2 Brice Mayag (LAMSADE) Measurement theory and meaningfulness Chapter 2 1 / 47 Outline 1
More informationScientific Inquiry Section 1: Length & Measurement ruler or meter stick: equipment used in the lab to measure length in millimeters, centimeters or
Scientific Inquiry Section 1: Length & Measurement ruler or meter stick: equipment used in the lab to measure length in millimeters, centimeters or meters. meter: metric unit for length -Scientists use
More informationScientific Research. The Scientific Method. Scientific Explanation
Scientific Research The Scientific Method Make systematic observations. Develop a testable explanation. Submit the explanation to empirical test. If explanation fails the test, then Revise the explanation
More informationThe Scientific Method
Course "Empirical Evaluation in Informatics" The Scientific Method Prof. Dr. Lutz Prechelt Freie Universität Berlin, Institut für Informatik http://www.inf.fu-berlin.de/inst/ag-se/ Science and insight
More informationFinal Exam: PSYC 300. Multiple Choice Items (1 point each)
Final Exam: PSYC 300 Multiple Choice Items (1 point each) 1. Which of the following is NOT one of the three fundamental features of science? a. empirical questions b. public knowledge c. mathematical equations
More informationUnit 1 History and Methods Chapter 1 Thinking Critically with Psychological Science
Myers PSYCHOLOGY (7th Ed) Unit 1 History and Methods Chapter 1 Thinking Critically with James A. McCubbin, PhD Clemson University Worth Publishers Fact vs. Falsehood 1. Human intuition is remarkably accurate
More informationCompletely randomized designs, Factors, Factorials, and Blocking
Completely randomized designs, Factors, Factorials, and Blocking STAT:5201 Week 2: Lecture 1 1 / 35 Completely Randomized Design (CRD) Simplest design set-up Treatments are randomly assigned to EUs Easiest
More informationHandout on Perfect Bayesian Equilibrium
Handout on Perfect Bayesian Equilibrium Fudong Zhang April 19, 2013 Understanding the concept Motivation In general, the Perfect Bayesian Equilibrium (PBE) is the concept we are using when solving dynamic
More informationEmpowered by Psychometrics The Fundamentals of Psychometrics. Jim Wollack University of Wisconsin Madison
Empowered by Psychometrics The Fundamentals of Psychometrics Jim Wollack University of Wisconsin Madison Psycho-what? Psychometrics is the field of study concerned with the measurement of mental and psychological
More informationSix Sigma Glossary Lean 6 Society
Six Sigma Glossary Lean 6 Society ABSCISSA ACCEPTANCE REGION ALPHA RISK ALTERNATIVE HYPOTHESIS ASSIGNABLE CAUSE ASSIGNABLE VARIATIONS The horizontal axis of a graph The region of values for which the null
More information4. Model evaluation & selection
Foundations of Machine Learning CentraleSupélec Fall 2017 4. Model evaluation & selection Chloé-Agathe Azencot Centre for Computational Biology, Mines ParisTech chloe-agathe.azencott@mines-paristech.fr
More informationTechnical Specifications
Technical Specifications In order to provide summary information across a set of exercises, all tests must employ some form of scoring models. The most familiar of these scoring models is the one typically
More informationLec 02: Estimation & Hypothesis Testing in Animal Ecology
Lec 02: Estimation & Hypothesis Testing in Animal Ecology Parameter Estimation from Samples Samples We typically observe systems incompletely, i.e., we sample according to a designed protocol. We then
More informationMeasurement. 500 Research Methods Mike Kroelinger
Measurement 500 Research Methods Mike Kroelinger Levels of Measurement Nominal Lowest level -- used to classify variables into two or more categories. Cases placed in the same category must be equivalent.
More informationChapter 11: Experiments and Observational Studies p 318
Chapter 11: Experiments and Observational Studies p 318 Observation vs Experiment An observational study observes individuals and measures variables of interest but does not attempt to influence the response.
More informationThe Role of Feedback in Categorisation
The Role of in Categorisation Mark Suret (m.suret@psychol.cam.ac.uk) Department of Experimental Psychology; Downing Street Cambridge, CB2 3EB UK I.P.L. McLaren (iplm2@cus.cam.ac.uk) Department of Experimental
More informationHUMAN-COMPUTER INTERACTION EXPERIMENTAL DESIGN
HUMAN-COMPUTER INTERACTION EXPERIMENTAL DESIGN Professor Bilge Mutlu Computer Sciences, Psychology, & Industrial and Systems Engineering University of Wisconsin Madison CS/Psych-770 Human-Computer Interaction
More informationEnumerative and Analytic Studies. Description versus prediction
Quality Digest, July 9, 2018 Manuscript 334 Description versus prediction The ultimate purpose for collecting data is to take action. In some cases the action taken will depend upon a description of what
More informationMath 1680 Class Notes. Chapters: 1, 2, 3, 4, 5, 6
Math 1680 Class Notes Chapters: 1, 2, 3, 4, 5, 6 Chapter 1. Controlled Experiments Salk vaccine field trial: a randomized controlled double-blind design 1. Suppose they gave the vaccine to everybody, and
More informationLecture 01 Analysis of Animal Populations: Theory and Scientific Process
1 of 1 Lecture 01 Analysis of Animal Populations: Theory and Scientific Process Motivation 1. Review the basic theory of animal population dynamics 2. Lay the foundation for the analysis of animal populations
More informationPubH 7470: STATISTICS FOR TRANSLATIONAL & CLINICAL RESEARCH
PubH 7470: STATISTICS FOR TRANSLATIONAL & CLINICAL RESEARCH Instructor: Chap T. Le, Ph.D. Distinguished Professor of Biostatistics Basic Issues: COURSE INTRODUCTION BIOSTATISTICS BIOSTATISTICS is the Biomedical
More informationTHE ROLE OF THE COMPUTER IN DATA ANALYSIS
CHAPTER ONE Introduction Welcome to the study of statistics! It has been our experience that many students face the prospect of taking a course in statistics with a great deal of anxiety, apprehension,
More informationChapter 4: More about Relationships between Two-Variables Review Sheet
Review Sheet 4. Which of the following is true? A) log(ab) = log A log B. D) log(a/b) = log A log B. B) log(a + B) = log A + log B. C) log A B = log A log B. 5. Suppose we measure a response variable Y
More informationInterpretation of Data and Statistical Fallacies
ISSN: 2349-7637 (Online) RESEARCH HUB International Multidisciplinary Research Journal Research Paper Available online at: www.rhimrj.com Interpretation of Data and Statistical Fallacies Prof. Usha Jogi
More informationStudent Performance Q&A:
Student Performance Q&A: 2009 AP Statistics Free-Response Questions The following comments on the 2009 free-response questions for AP Statistics were written by the Chief Reader, Christine Franklin of
More informationChapter 02 Developing and Evaluating Theories of Behavior
Chapter 02 Developing and Evaluating Theories of Behavior Multiple Choice Questions 1. A theory is a(n): A. plausible or scientifically acceptable, well-substantiated explanation of some aspect of the
More informationResearch and science: Qualitative methods
Research and science: Qualitative methods Urban Bilstrup (E327) Urban.Bilstrup@hh.se 140922 2 INTRODUCTION TO RESEARCH AND SCIENTIFIC METHODS Outline Definitions Problem formulation? Aim and goal operational
More informationQualitative and Quantitative Approaches Workshop. Comm 151i San Jose State U Dr. T.M. Coopman Okay for non-commercial use with attribution
Qualitative and Quantitative Approaches Workshop Comm 151i San Jose State U Dr. T.M. Coopman Okay for non-commercial use with attribution This Workshop This is a research skill workshop. This workshop
More informationReinforcement Learning : Theory and Practice - Programming Assignment 1
Reinforcement Learning : Theory and Practice - Programming Assignment 1 August 2016 Background It is well known in Game Theory that the game of Rock, Paper, Scissors has one and only one Nash Equilibrium.
More informationMark J. Anderson, Patrick J. Whitcomb Stat-Ease, Inc., Minneapolis, MN USA
Journal of Statistical Science and Application (014) 85-9 D DAV I D PUBLISHING Practical Aspects for Designing Statistically Optimal Experiments Mark J. Anderson, Patrick J. Whitcomb Stat-Ease, Inc., Minneapolis,
More informationBetween Micro and Macro: Individual and Social Structure, Content and Form in Georg Simmel
Michela Bowman Amy LeClair Michelle Lynn Robert Weide Classical Sociological Theory November 11, 2003 Between Micro and Macro: Individual and Social Structure, Content and Form in Georg Simmel Through
More informationWELCOME! Lecture 11 Thommy Perlinger
Quantitative Methods II WELCOME! Lecture 11 Thommy Perlinger Regression based on violated assumptions If any of the assumptions are violated, potential inaccuracies may be present in the estimated regression
More informationStructural Equation Modeling (SEM)
Structural Equation Modeling (SEM) Today s topics The Big Picture of SEM What to do (and what NOT to do) when SEM breaks for you Single indicator (ASU) models Parceling indicators Using single factor scores
More informationIndiana Academic Standards Addressed By Zoo Program WINGED WONDERS: SEED DROP!
Indiana Academic Standards Addressed By Zoo Program WINGED WONDERS: SEED DROP! Program description: Discover how whether all seeds fall at the same rate. Do small or big seeds fall more slowly? Students
More informationSTA630 Research Methods Solved MCQs By
STA630 Research Methods Solved MCQs By http://vustudents.ning.com 31-07-2010 Quiz # 1: Question # 1 of 10: A one tailed hypothesis predicts----------- The future The lottery result The frequency of the
More informationSOME PRINCIPLES OF FIELD EXPERlMENTS WITH SHEEP By P. G. SCHINCICEL *, and G. R. MOULE *
SOME PRINCIPLES OF FIELD EXPERlMENTS WITH SHEEP By P. G. SCHINCICEL *, and G. R. MOULE * Summary The principles of scientific method, with particular reference to the role of hypotheses and experiments
More informationLecture 4: Research Approaches
Lecture 4: Research Approaches Lecture Objectives Theories in research Research design approaches ú Experimental vs. non-experimental ú Cross-sectional and longitudinal ú Descriptive approaches How to
More informationChapter 2. The Data Analysis Process and Collecting Data Sensibly. Copyright 2005 Brooks/Cole, a division of Thomson Learning, Inc.
Chapter 2 The Data Analysis Process and Collecting Data Sensibly Important Terms Variable A variable is any characteristic whose value may change from one individual to another Examples: Brand of television
More informationResearch. how we figure stuff out. Methods
Research how we figure stuff out Methods Penny in the Glass Activity Let s Make A Deal! One volunteer is needed for a chance to win $1,334,499! Let s Make A Deal Shows Us That: Human Intuition is highly
More informationThe essential focus of an experiment is to show that variance can be produced in a DV by manipulation of an IV.
EXPERIMENTAL DESIGNS I: Between-Groups Designs There are many experimental designs. We begin this week with the most basic, where there is a single IV and where participants are divided into two or more
More informationReliability of Ordination Analyses
Reliability of Ordination Analyses Objectives: Discuss Reliability Define Consistency and Accuracy Discuss Validation Methods Opening Thoughts Inference Space: What is it? Inference space can be defined
More informationWhat is Science 2009 What is science?
What is science? The question we want to address is seemingly simple, but turns out to be quite difficult to answer: what is science? It is reasonable to ask such a question since this is a book/course
More informationLearning with Rare Cases and Small Disjuncts
Appears in Proceedings of the 12 th International Conference on Machine Learning, Morgan Kaufmann, 1995, 558-565. Learning with Rare Cases and Small Disjuncts Gary M. Weiss Rutgers University/AT&T Bell
More informationEvaluation Models STUDIES OF DIAGNOSTIC EFFICIENCY
2. Evaluation Model 2 Evaluation Models To understand the strengths and weaknesses of evaluation, one must keep in mind its fundamental purpose: to inform those who make decisions. The inferences drawn
More informationStudy on perceptually-based fitting line-segments
Regeo. Geometric Reconstruction Group www.regeo.uji.es Technical Reports. Ref. 08/2014 Study on perceptually-based fitting line-segments Raquel Plumed, Pedro Company, Peter A.C. Varley Department of Mechanical
More informationMeasuring and Assessing Study Quality
Measuring and Assessing Study Quality Jeff Valentine, PhD Co-Chair, Campbell Collaboration Training Group & Associate Professor, College of Education and Human Development, University of Louisville Why
More informationChapter 4: Defining and Measuring Variables
Chapter 4: Defining and Measuring Variables A. LEARNING OUTCOMES. After studying this chapter students should be able to: Distinguish between qualitative and quantitative, discrete and continuous, and
More informationChapter 13. Experiments and Observational Studies
Chapter 13 Experiments and Observational Studies 1 /36 Homework Read Chpt 13 Do p312 1, 7, 9, 11, 17, 20, 25, 27, 29, 33, 40, 41 2 /36 Observational Studies In an observational study, researchers do not
More informationOriginal content Copyright by Holt, Rinehart and Winston. Additions and changes to the original content are the responsibility of the instructor.
Answer Key Directed Reading A 1. life science 2. diversity 3. Answers may vary. Sample answer: Where does it live? 4. anyone 5. anywhere in a laboratory, on farms, in forests, on the ocean floor, in space,
More informationPlacebo and Belief Effects: Optimal Design for Randomized Trials
Placebo and Belief Effects: Optimal Design for Randomized Trials Scott Ogawa & Ken Onishi 2 Department of Economics Northwestern University Abstract The mere possibility of receiving a placebo during a
More informationAgreement Coefficients and Statistical Inference
CHAPTER Agreement Coefficients and Statistical Inference OBJECTIVE This chapter describes several approaches for evaluating the precision associated with the inter-rater reliability coefficients of the
More information2.75: 84% 2.5: 80% 2.25: 78% 2: 74% 1.75: 70% 1.5: 66% 1.25: 64% 1.0: 60% 0.5: 50% 0.25: 25% 0: 0%
Capstone Test (will consist of FOUR quizzes and the FINAL test grade will be an average of the four quizzes). Capstone #1: Review of Chapters 1-3 Capstone #2: Review of Chapter 4 Capstone #3: Review of
More information