Experimental design (continued)

Similar documents
HUMAN-COMPUTER INTERACTION EXPERIMENTAL DESIGN

Analysis A step in the research process that involves describing and then making inferences based on a set of data.

Instrumental Variables I (cont.)

Political Science 15, Winter 2014 Final Review

Designing Experiments... Or how many times and ways can I screw that up?!?

Propensity Score Methods for Estimating Causality in the Absence of Random Assignment: Applications for Child Care Policy Research

Science, Society, and Social Research (1) Benjamin Graham

Lecture 3. Previous lecture. Learning outcomes of lecture 3. Today. Trustworthiness in Fixed Design Research. Class question - Causality

Review: Logistic regression, Gaussian naïve Bayes, linear regression, and their connections

Lecture 11: Measurement to Hypotheses. Benjamin Graham

Research in Real-World Settings: PCORI s Model for Comparative Clinical Effectiveness Research

15.301/310, Managerial Psychology Prof. Dan Ariely Recitation 8: T test and ANOVA

The degree to which a measure is free from error. (See page 65) Accuracy

Why randomize? Rohini Pande Harvard University and J-PAL.

Carrying out an Empirical Project

Experimental Research. Types of Group Comparison Research. Types of Group Comparison Research. Stephen E. Brock, Ph.D.

Introduction to Path Analysis

PSY 250. Designs 8/11/2015. Nonexperimental vs. Quasi- Experimental Strategies. Nonequivalent Between Group Designs

investigate. educate. inform.

CHAPTER 8 EXPERIMENTAL DESIGN

Observational Study Designs. Review. Today. Measures of disease occurrence. Cohort Studies

Methods for Addressing Selection Bias in Observational Studies

Challenges of Observational and Retrospective Studies

Recent advances in non-experimental comparison group designs

The Logic of Data Analysis Using Statistical Techniques M. E. Swisher, 2016

Fitting the Method to the Question

CSC2130: Empirical Research Methods for Software Engineering

Introduction to the Scientific Method. Knowledge and Methods. Methods for gathering knowledge. method of obstinacy

PLS 506 Mark T. Imperial, Ph.D. Lecture Notes: Reliability & Validity

Reflection Questions for Math 58B

Study Design. Svetlana Yampolskaya, Ph.D. Summer 2013

Fitting the Method to the Question

VALIDITY OF QUANTITATIVE RESEARCH

Experimental and Quasi-Experimental designs

Lecture 5 Conducting Interviews and Focus Groups

UNIT 3 & 4 PSYCHOLOGY RESEARCH METHODS TOOLKIT

Experiments. ESP178 Research Methods Dillon Fitch 1/26/16. Adapted from lecture by Professor Susan Handy

Handout 1: Introduction to the Research Process and Study Design STAT 335 Fall 2016

PubH 7470: STATISTICS FOR TRANSLATIONAL & CLINICAL RESEARCH

Exploring Antimalarial Drugs

Lec 02: Estimation & Hypothesis Testing in Animal Ecology

Overview of Study Designs

Assignment 4: True or Quasi-Experiment

Qualitative Research Methods

INTERNAL VALIDITY, BIAS AND CONFOUNDING

TRANSLATING RESEARCH INTO ACTION. Why randomize? Dan Levy. Harvard Kennedy School

You must answer question 1.

MS&E 226: Small Data

Theory, Models, Variables

Descriptive Statistics

Topics. Experiment Terminology (Part 1)

The essential focus of an experiment is to show that variance can be produced in a DV by manipulation of an IV.

Overview of the Logic and Language of Psychology Research

Instrumental Variables Estimation: An Introduction

QUASI-EXPERIMENTAL HEALTH SERVICE EVALUATION COMPASS 1 APRIL 2016

Chapter 17 Sensitivity Analysis and Model Validation

Recent developments for combining evidence within evidence streams: bias-adjusted meta-analysis

Lecture 4: Research Approaches

Übung zur Vorlesung Informationsvisualisierung

HPS301 Exam Notes- Contents

Quantitative Analysis and Empirical Methods

Threats to validity in intervention studies. Potential problems Issues to consider in planning

Overview of Experimentation

SMALL n IMPACT EVALUATION. Howard White and Daniel Phillips 3ie

Chapter 9 Experimental Research (Reminder: Don t forget to utilize the concept maps and study questions as you study this and the other chapters.

INTERPRETATION OF STUDY FINDINGS: PART I. Julie E. Buring, ScD Harvard School of Public Health Boston, MA

Today s reading concerned what method? A. Randomized control trials. B. Regression discontinuity. C. Latent variable analysis. D.

Work, Employment, and Industrial Relations Theory Spring 2008

Quantitative Methods. Lonnie Berger. Research Training Policy Practice

Marno Verbeek Erasmus University, the Netherlands. Cons. Pros

OBSERVATION METHODS: EXPERIMENTS

Validity and Quantitative Research. What is Validity? What is Validity Cont. RCS /16/04

Quantitative Evaluation

Research Synthesis and meta-analysis: themes. Graham A. Colditz, MD, DrPH Method Tuuli, MD, MPH

CASE STUDY 2: VOCATIONAL TRAINING FOR DISADVANTAGED YOUTH

Doing Quantitative Research 26E02900, 6 ECTS Lecture 6: Structural Equations Modeling. Olli-Pekka Kauppila Daria Kautto

Impact Evaluation Methods: Why Randomize? Meghan Mahoney Policy Manager, J-PAL Global

Oak Meadow Autonomy Survey

Patrick Breheny. January 28

Lecture II: Difference in Difference. Causality is difficult to Show from cross

Geographic Data Science - Lecture IX

PTHP 7101 Research 1 Chapter Assignments

Key questions when starting an econometric project (Angrist & Pischke, 2009):

Experimental Design and the struggle to control threats to validity

Reduce Tension by Making the Desired Choice Easier

PSYC1024 Clinical Perspectives on Anxiety, Mood and Stress

Chapter 4: Defining and Measuring Variables

1. The Role of Sample Survey Design

26:010:557 / 26:620:557 Social Science Research Methods

Measuring the User Experience

The Process of Scientific Inquiry Curiosity. Method. The Process of Scientific Inquiry. Is there only one scientific method?

Seminar 3: Basics of Empirical Research

Examining Relationships Least-squares regression. Sections 2.3

Research Designs. Internal Validity

T. Kushnir & A. Gopnik (2005 ). Young children infer causal strength from probabilities and interventions. Psychological Science 16 (9):

Biostatistics II

Regression Discontinuity Analysis

Psychology Research Process

Experimental Design Part II

Transcription:

Experimental design (continued) Spring 2017 Michelle Mazurek Some content adapted from Bilge Mutlu, Vibha Sazawal, Howard Seltman 1

Administrative No class Tuesday Homework 1 Plug for Tanu Mitra grad student session 2

Today s class Finish threats to validity Experimental design / choices Alternatives to experiments 3

Quick review Internal validity: causality Isolate variable of interest Randomized assignment External validity Representative sample Representative environment/task/analysis Valid constructs Measure something meaningful Reliable 4

Know what you re measuring Especially when dealing with large-scale data from the internet What are you missing? What is duplicated? What is the precision and accuracy of the data? Are you capturing what you think you re capturing? *Vantage point* Representativeness / diversity 5

Calibrating constructs Examine outliers and spikes Check for self-consistency Compare multiple measures Multiple datasets Multiple ways of calculating a value Test with synthetic data Check longitudinal data periodically! 6

Mis-measurements, now what? Discard? (Why might this be bad?) Discard outliers? Definition? Use an explicit adjustment? 7

Other measurement notes (Don t really fit here, but from Paxson paper) Metadata and good analysis logging is critical! Be clear about unknowns and limitations 8

4. Power Power: Likelihood that if there s a real effect, you will find it. Why might you not find it? Sample size Effect size Missing explanatory variables Variability 9

Promote power Covariates: Measure possible confounds, include in analysis Use reliable measurements Control the environment Potential tradeoff: Generalizability for power E.g., limit variability between subjects http://4.bp.blogspot.com/-fuha1- JXxto/T_ssNrNODtI/AAAAAAAAAo0/LXcl0Pxzg40/s1 10

EXPERIMENTAL DESIGN 11

Some important decisions What is the hypothesis? Between or within subjects? What treatment levels / conditions? What dependent variables to measure? 12

Good hypothesis design Predicted relationship between (at least) 2 vars Testable, falsifiable Operational Vars are clearly defined Relationship / how you measure it clearly defined 13

Good hypothesis design (cont.) Justified Exploratory results Theory in related area Well justified intuition? Parsimonious 14

Between vs. Within Between: Each participant belongs to exactly one condition Within: Each participant belongs to multiple 15

Between vs. Within More participants Cleaner/less bias More time each More power (less variability subj-subj) 16

Improving on between-subjects Matching: Get like participants for each condition Pro: reduces variability Con: Hard to find; what do you match on? In general, be very cautious 17

Improving on within-subjects Ordering effects can be HUGE Learning, fatigue Range effects: learn most for closest conditions Mitigate via counterbalancing All possible orders Balanced latin square A B C D C A D B B D A C D C B A 18

Counterbalancing doesn t fix: Range effects (most average treatment) Context effects (what most participants are more familiar with) 19

Mixed models are also possible Everyone gets the same three tasks Order of tasks varies Tool with which to execute tasks varies 20

Selecting conditions How many IVs? Password meter example How many / which levels for each? Cannot infer anything about levels you didn t test 21

Full-factorial (or not) Full-factorial: All possible combinations of all Ivs And all orderings? Not: Only a subset Selected how? Recall: Vary at most one thing each time! Planned comparisons! 22

Why multivariate? What is different between running one experiment with two IVs vs. two experiments with one IV each? Interaction effects! 23

Dependent variables What and how to measure? Construct validity, again! Performance (time, errors, FP/FN, etc.) Opinions/attitude Audio recording, screen capture, keystrokes, copypasting behavior, etc. Demographics Multiple measures toward higher-level construct? 24

NOT JUST EXPERIMENTS 25

Kinds of measurement studies Experimental Observational/correlational Quasi-experimental 26

Observational/correlational Observe that X and Y (don t) increase and decrease together / in opposition Research doesn t apply any control or treatment: just measure incidence Does lead exposure correlate with crime rate? Directionality and third-variable both issues 27

Quasi-experiments Subset of observational studies Can t randomize assignment But, experimenter controls something Group 1 Group 1 Treatment Group 2 Group 2 28

Observational examples Cohort study Regression discontinuity BIBIFI example 29

Pluses and minuses Can measure things that simply can t be done with true experiments In general, association at best causality very hard to establish Some statistical techniques to help exist Low internal validity can you maximize it within the available constraints? 30