The validity of inferences about the correlation (covariation) between treatment and outcome.

Similar documents
PLS 506 Mark T. Imperial, Ph.D. Lecture Notes: Reliability & Validity

26:010:557 / 26:620:557 Social Science Research Methods

Internal Validity and Experimental Design

STATISTICAL CONCLUSION VALIDITY

Study Design. Svetlana Yampolskaya, Ph.D. Summer 2013

VALIDITY OF QUANTITATIVE RESEARCH

Experimental and Quasi-Experimental designs

Research Design. Source: John W. Creswell RESEARCH DESIGN. Qualitative, Quantitative, and Mixed Methods Approaches Third Edition

Rival Plausible Explanations

PSY 250. Experimental Design: The Basic Building Blocks. Simple between subjects design. The Two-Group Design 7/25/2015. Experimental design

CHAPTER NINE DATA ANALYSIS / EVALUATING QUALITY (VALIDITY) OF BETWEEN GROUP EXPERIMENTS

Threats to Validity in Experiments. The John Henry Effect

TECH 646 Analysis of Research in Industry and Technology. Experiments

Research Approach & Design. Awatif Alam MBBS, Msc (Toronto),ABCM Professor Community Medicine Vice Provost Girls Section

Topic #2. A key criterion in evaluating any test, measure, or piece of research is validity.

Experimental Research. Types of Group Comparison Research. Types of Group Comparison Research. Stephen E. Brock, Ph.D.

RESEARCH METHODS. Winfred, research methods, ; rv ; rv

RESEARCH METHODS. Winfred, research methods,

Validity and Quantitative Research. What is Validity? What is Validity Cont. RCS /16/04

Lecture 3. Previous lecture. Learning outcomes of lecture 3. Today. Trustworthiness in Fixed Design Research. Class question - Causality

Threats to validity in intervention studies. Potential problems Issues to consider in planning

Experimental Design Part II

HPS301 Exam Notes- Contents

Educational Research. S.Shafiee. Expert PDF Trial

Experimental Validity

Measurement. Reliability vs. Validity. Reliability vs. Validity

EXPERIMENTAL RESEARCH DESIGNS

Designing and Analyzing RCTs. David L. Streiner, Ph.D.

Propensity Score Analysis Shenyang Guo, Ph.D.

investigate. educate. inform.

Chapter 1: Explaining Behavior

TECH 646 Analysis of Research in Industry and Technology. Experiments

Lecture 9 Internal Validity

Mixed Methods Study Design

AP LATIN 2014 SCORING GUIDELINES

Introduction to Applied Research in Economics

CHAPTER LEARNING OUTCOMES

Sociology 201: Social Research Design

In this chapter we discuss validity issues for quantitative research and for qualitative research.

Introduction to Applied Research in Economics

RESEARCH METHODS. A Process of Inquiry. tm HarperCollinsPublishers ANTHONY M. GRAZIANO MICHAEL L RAULIN

Chapter 9 Experimental Research (Reminder: Don t forget to utilize the concept maps and study questions as you study this and the other chapters.

Formative and Impact Evaluation. Formative Evaluation. Impact Evaluation

Research Questions, Variables, and Hypotheses: Part 2. Review. Hypotheses RCS /7/04. What are research questions? What are variables?

Research Designs. Inferential Statistics. Two Samples from Two Distinct Populations. Sampling Error (Figure 11.2) Sampling Error

Measures of Dispersion. Range. Variance. Standard deviation. Measures of Relationship. Range. Variance. Standard deviation.

Research Methods & Design Outline. Types of research design How to choose a research design Issues in research design

LAPORAN AKHIR PROJEK PENYELIDIKAN TAHUN AKHIR (PENILAI) FINAL REPORT OF FINAL YEAR PROJECT (EXAMINER)

Design of Experiments & Introduction to Research

FSA Training Papers Grade 4 Exemplars. Rationales

Analysis A step in the research process that involves describing and then making inferences based on a set of data.

Contents. Acknowledgments

Student Performance Q&A:

WWC STUDY REVIEW STANDARDS

Experimental Design. Dewayne E Perry ENS C Empirical Studies in Software Engineering Lecture 8

MBACATÓLICA JAN/APRIL Marketing Research. Fernando S. Machado. Experimentation

Measurement is the process of observing and recording the observations. Two important issues:

Research Methodologies

Sample Exam Questions Psychology 3201 Exam 1

Identifying Peer Influence Effects in Observational Social Network Data: An Evaluation of Propensity Score Methods

Incorporating Clinical Information into the Label

Epidemiologic Methods I & II Epidem 201AB Winter & Spring 2002

Experiments. ESP178 Research Methods Dillon Fitch 1/26/16. Adapted from lecture by Professor Susan Handy

Checklist for Randomized Controlled Trials. The Joanna Briggs Institute Critical Appraisal tools for use in JBI Systematic Reviews

9/30/2017. personality traits: Sheltered, Confident, Team- Oriented, Conventional, Pressured, Achieving

Clinical research in AKI Timing of initiation of dialysis in AKI

Key questions when starting an econometric project (Angrist & Pischke, 2009):

AP STATISTICS 2006 SCORING GUIDELINES (Form B) Question 5

Use of the Quantitative-Methods Approach in Scientific Inquiry. Du Feng, Ph.D. Professor School of Nursing University of Nevada, Las Vegas

Dichotomizing partial compliance and increased participant burden in factorial designs: the performance of four noncompliance methods

Adverse Outcomes After Hospitalization and Delirium in Persons With Alzheimer Disease

Introduction to Applied Research in Economics Kamiljon T. Akramov, Ph.D. IFPRI, Washington, DC, USA

Research Designs For Evaluating Disease Management Program Effectiveness

Brief introduction to instrumental variables. IV Workshop, Bristol, Miguel A. Hernán Department of Epidemiology Harvard School of Public Health

Confounded: Causal Inference and the Requirement of Independence

Checklist for Randomized Controlled Trials. The Joanna Briggs Institute Critical Appraisal tools for use in JBI Systematic Reviews

Chapter 02 Developing and Evaluating Theories of Behavior

Research Design. Beyond Randomized Control Trials. Jody Worley, Ph.D. College of Arts & Sciences Human Relations

Lecture 4: Research Approaches

UNIT III: Research Design. In designing a needs assessment the first thing to consider is level for the assessment

The essential focus of an experiment is to show that variance can be produced in a DV by manipulation of an IV.

Biological Models of Anxiety

Today s reading concerned what method? A. Randomized control trials. B. Regression discontinuity. C. Latent variable analysis. D.

Estimation of effect sizes in the presence of publication bias: a comparison of meta-analysis methods

Durham Research Online

STEP II Conceptualising a Research Design

Topic #4 CONTROL. 1. What are the threats to the validity of a contemplated piece of research?

What We Will Cover in This Section

WWC Standards for Regression Discontinuity and Single Case Study Designs

PÄIVI KARHU THE THEORY OF MEASUREMENT

Internal & External Validity

ANCOVA with Regression Homogeneity

Topic #4 CONTROL. 1. What are the threats to the validity of a contemplated piece of research?

2013/4/28. Experimental Research

Non-Randomized Trials

Chapter 9: Experiments

NATIONAL QUALITY FORUM

Handout 1: Introduction to the Research Process and Study Design STAT 335 Fall 2016

One slide on research question Literature review: structured; holes you will fill in Your research design

Transcription:

Threats Summary (MGS 9940, Fall, 2004-Courtesy of Amit Mehta) Statistical Conclusion Validity The validity of inferences about the correlation (covariation) between treatment and outcome. Threats to Statistical Conclusion Validity Low Statistical Power Violated Assumptions of Statistical Test Fishing and Error Rate Problem Unreliability of Measures Restriction of Range Unreliability of Treatment Implementation Extraneous Variance in the Experimental Setting Heterogeneity of Units Inaccurate Effect Size Estimation An insufficiently powered experiment may incorrectly conclude that the relationship between treatment and outcome is not significant. Violations of statistical test assumptions can lead to either overestimating or underestimating the size and significance. Repeated tests for significant relationships, if uncorrected for the number of tests, can artifactually inflate statistical significance. Measurement error weakens the relationship between two variables and strengthens or weakens the relationship between three or more variables Reduced range on a variable usually weakens the relationship between it and another variable. The effects of partial implementation may be underestimated compared to full implementation. Some features of an experimental setting may inflate error, making decision of an effect more difficult. Increased variability on the outcome variable within conditions increases error variance, making detection of a relationship more difficult. Some statistics symmetrically overestimate or underestimate the size of an effect. 1

Internal Validity The validity of inferences about whether the observed co-variation between the treatment and outcome reflects a causal relationship. Threats to Internal Validity Ambiguous Temporal Precedence Selection History Maturation Regression Attrition Testing Instrumentation Additive and Interactive Effects of Threats to Internal Lack of clarity about which variable occurred first, may yield confusion about which variable is the cause and which is the effect. Systematic differences over conditions in respondent characteristics that could also cause the observed effect. Events occurring concurrently with treatment could cause the observed effect. Naturally occurring changes over time could be confused with a treatment. When units are selected for their extreme scores, they will often have less extreme scores on the other variables, which in turn can be confused with a treatment. Loss of respondents to treatment or to measurement can produce artifactual effects if that loss is systematically correlated with conditions. Exposure to a test can affect scores on subsequent exposures to that test, an occurrence that can be confused with a treatment effect. The nature of a measure may change over time or conditions in a way that could be confused with a treatment effect. The impact of a threat can be added to that of another threat or may depend on the level of another threat. 2

Validity Construct Validity The validity of inferences about the higher order constructs that represent sampling particulars. Threats to Construct Validity Inadequate Explication of Constructs Construct Confounding Mono-Operation Bias Mono-Method Bias Confounding Constructs with Levels of Constructs Treatment Sensitive Factorial Structure Reactive Self-Report Changes Reactivity to the Experimental Situation Failure to adequately explicate a construct may lead to incorrect inferences about the relationship between operation and construct. Operations usually involve more than one construct, and failure to describe all the constructs may result in incomplete construct inferences. Any one operationalization of a construct both underrepresents the construct of interest and measures irrelevant constructs, complicating inference. When all operationalizations use the same method, that method is part of the construct actually studied. Inferences about constructs that best represent study operations may fail to describe the limited levels of the construct studied. The structure of a measure may change as a result of treatment, change that may be hidden if the same scoring is always used. Self-reports can be affected by participant motivation to be in a treatment condition, motivation that can change after assignment is made. Participant responses reflect not just treatments and measures but also participants; perceptions of the experimental situation, and those perceptions are part of the treatment construct actually tested. 3

Experimenter Expectancies Novelty and Disruption Effect Compensatory Equalization Compensatory Rivalry Resentful Demoralization Treatment Diffusion External Validity The experimenter can influence participant responses by conveying expectations about desirable responses, and those expectations are part of the treatment construct as actually tested. Participants may respond unusually well to a novel innovation or unusually poorly to one that disrupts their routine, a response that must then be included as a part of the treatment construct description. When treatment provides desirable goods or services, administrators, staff, or constituents may provide compensatory goods or services to those not receiving treatment, and this action must then be included as a part of the treatment construct description. Participants not receiving treatment may be motivated to show they can do as well as those receiving treatment, and this compensatory rivalry must then be included as part of the treatment construct description. Participants not receiving treatment may be so resentful or demoralized that they may respond more negatively than otherwise and this resentful demoralization must then be included as a part of the treatment construct description. Participants may receive services from a condition to which they were not assigned, making construct descriptions of both conditions more difficult. The validity of inferences about whether the cause-andeffect relationship holds over variation in persons, settings, treatments and measurements. Threats to External Validity 4

Interaction of the Causal Relationship with Units Interaction of the Causal Relationship over Treatment Variations Interaction of the Causal Relationship with Outcomes Interaction of the Causal Relationship with Settings Context-Dependent Mediation An effect found with certain kinds of units might not hold if other kinds of units had been studied. An effect found with one treatment is combined with other treatments, or when only partial treatment is used. An effect found on one kind of outcome observation may not hold if other outcome observations were to be used. An effect found in one kind of setting may not hold if other kinds of settings were to be used. An explanatory mediator of a causal relationship in one context may not mediate in another context. Table 3.1 Threats to Construct Validity: Reasons Why Inferences about the Constructs that Characterize Study Operations may be Incorrect (DSC 8820, Fall, 2003-Courtesy of Robert Jones and Julie Petherbridge) Inadequate Explication of Constructs Construct Confounding Mono-Operation Bias Mono-Method Bias Confounding Constructs with Levels of Constructs Failure to adequately explicate a construct may lead to incorrect inferences about the relationship between operation and construct. Operations usually involve more than on construct, and failure to describe all the constructs may result in incomplete construct inferences. Any one operationalization of a construct both underrepresents the construct of interest and measures irrelevant constructs, complicating inference. When all operationalizations use the same method that method is part of the construct actually studied. Inferences about the constructs that best represent study operations may fail to descry be the limited levels of the construct that were actually studied. Treatment Sensitive Factorial The structure of a measure may change as a result of treatment, 5

Structure Reactive Self-Report Changes Reactivity to the Experimental Situation Experimenter Expectancies Novelty and Disruption Effects Compensatory Equalization Compensatory Rivalry Resentful Demoralization Treatment Diffusion change that may be hidden is the same scoring is always used. Self-reports can be affected by participant motivation to be in a treatment condition, motivation that can change after assignment is made. Participant responses reflect not just treatments and measures but also participants perceptions of the experimental situation, and those perceptions are part of the treatment construct actually tested. The experimenter can influence participant responses by conveying expectations about desirable responses, and those expectations are part of the treatment construct as actually tested. Participants may respond unusually well to a novel innovation or unusually poorly to one that disrupts their routine, a response that must then be included as part of the treatment construct description. When treatment provides desirable goods or services, administrators, staff, or constituents may provide compensatory goods or services to those not receiving treatment, and this action must then be included as part of the treatment construct description. Participants not receiving treatment may be motivated to show they can do as well as those receiving treatment, and this compensatory rivalry must then be included as part of the treatment construct description. Participants not receiving a desirable treatment may be so resentful or demoralized that they may respond more negatively than otherwise, and this resentful demoralization must then be included as part of the treatment construct description. Participants may receive services from a condition to which they were not assigned, making construct descriptions of both conditions more difficult. 6