Construct Validation of Direct Behavior Ratings: A Multitrait Multimethod Analysis

Similar documents
Using Direct Behavior Ratings in a Middle School Setting

PÄIVI KARHU THE THEORY OF MEASUREMENT

Measurement is the process of observing and recording the observations. Two important issues:

ADMS Sampling Technique and Survey Studies

INTERPARENTAL CONFLICT, ADOLESCENT BEHAVIORAL PROBLEMS, AND ADOLESCENT COMPETENCE: CONVERGENT AND DISCRIMINANT VALIDITY

2 Types of psychological tests and their validity, precision and standards

Chapter 3. Psychometric Properties

Family Income (SES) Age and Grade 4/22/2014. Center for Adolescent Research in the Schools (CARS) Participants n=647

Revisiting the Halo Effect Within a Multitrait, Multimethod Framework

Adult Substance Use and Driving Survey-Revised (ASUDS-R) Psychometric Properties and Construct Validity

Review of Various Instruments Used with an Adolescent Population. Michael J. Lambert

Special Section QUANTITATIVE METHODS IN THE STUDY OF DRUG USE. Edited by George J. Huba ABSTRACT

Buy full version here - for $ 7.00

No part of this page may be reproduced without written permission from the publisher. (

Reliability. Internal Reliability

Measurement of Commitment to Role Identities

The Psychometric Principles Maximizing the quality of assessment

Reliability of the Ghiselli Self-Description Inventory as a Measure of Self-Esteem


Research Questions and Survey Development

PSYCHOMETRICS OF THE SSIS IN AN AUSTRALIAN POPULATION A DISSERTATION SUBMITTED TO THE FACULTY

In 2013, a group of researchers published a paper evaluating the construct

CareerCruising. MatchMaker Reliability and Validity Analysis

Helping the Student With ADHD in the Classroom

Factors Influencing Undergraduate Students Motivation to Study Science

A Cross-validation of easycbm Mathematics Cut Scores in. Oregon: Technical Report # Daniel Anderson. Julie Alonzo.

Variable Measurement, Norms & Differences

Process of a neuropsychological assessment

sample of 85 graduate students at The University of Michigan s influenza, benefits provided by a flu shot, and the barriers or costs associated

Remembering the Way it Was : Development and Validation of the Historical Nostalgia Scale

Table S1. Search terms applied to electronic databases. The African Journal Archive African Journals Online. depression OR distress

LONGITUDINAL MULTI-TRAIT-STATE-METHOD MODEL USING ORDINAL DATA. Richard Shane Hutton

Evaluating Reading Diagnostic Tests:

VARIABLES AND MEASUREMENT

Reliability of Data Derived from Time Sampling Methods with Multiple Observation Targets

ACDI. An Inventory of Scientific Findings. (ACDI, ACDI-Corrections Version and ACDI-Corrections Version II) Provided by:

REVIEW OF METHODS FOR FORMATIVE ASSESSMENT OF CHILD SOCIAL BEHAVIOR

Early Childhood Measurement and Evaluation Tool Review

Test Validity. What is validity? Types of validity IOP 301-T. Content validity. Content-description Criterion-description Construct-identification

Excellence in Prevention descriptions of the prevention

Keywords HAART HIV infection HRQoL QoL Pediatric PedsQL TM 4.0 ThQLC

It was hypothesized that the male subjects would show a low level of emotional intelligence as compared to the female subjects.

ESTABLISHING VALIDITY AND RELIABILITY OF ACHIEVEMENT TEST IN BIOLOGY FOR STD. IX STUDENTS

Cultural Intelligence: A Predictor of Ethnic Minority College Students Psychological Wellbeing

STRATEGIES FOR STUDENTS WITH ADHD. Classroom Management Strategies for Students with ADHD Mary Johnson Seattle Pacific University 2016

Construction of an Attitude Scale towards Teaching Profession: A Study among Secondary School Teachers in Mizoram

SAQ-Short Form Reliability and Validity Study in a Large Sample of Offenders

Behavioral Intervention Rating Rubric. Group Design

ISSN X Journal of Educational and Social Research Vol. 2 (8) October 2012

PLS 506 Mark T. Imperial, Ph.D. Lecture Notes: Reliability & Validity

T. Rene Jamison * and Jessica Oeth Schuttler

Sikha Naik Mark Vosvick, Ph.D, Chwee-Lye Chng, Ph.D, and John Ridings, A.A. Center for Psychosocial Health

The Effect of Extremes in Small Sample Size on Simple Mixed Models: A Comparison of Level-1 and Level-2 Size

Development of a Scale to Measure Impression Management in Job Interviews

Supplementary Material. other ethnic backgrounds. All but six of the yoked pairs were matched on ethnicity. Results

Cross-validation of easycbm Reading Cut Scores in Washington:

Emotional Intelligence Assessment Technical Report

Structural Equation Modeling of Multiple- Indicator Multimethod-Multioccasion Data: A Primer

Validity. Ch. 5: Validity. Griggs v. Duke Power - 2. Griggs v. Duke Power (1971)

Training, Inclusion, and Behaviour: Effect on Student Teacher and Student SEA Relationships for Students with Autism Spectrum Disorders

Measurement and Descriptive Statistics. Katie Rommel-Esham Education 604

DISCUSSION: PHASE ONE THE RELIABILITY AND VALIDITY OF THE MODIFIED SCALES

CRITICALLY APPRAISED PAPER (CAP) FOCUSED QUESTION

DISCPP (DISC Personality Profile) Psychometric Report

ANN ARBOR PUBLIC SCHOOLS 2555 S. State Street Ann Arbor, MI www. a2schools.org Pioneer High School Annual Education Report (AER)!

11 Validity. Content Validity and BIMAS Scale Structure. Overview of Results

STRESS LEVELS AND ALCOHOL USE AMONG UNDERGRADUATE STUDENTS: A QUANTITATIVE STUDY. Noemi Alsup California State University, Long Beach May 20, 2014

S.O.S. Suicide Prevention Program

Prediction of academic achievement using the School Motivation Analysis Test.

Conners 3. Conners 3rd Edition

Georgina Salas. Topics EDCI Intro to Research Dr. A.J. Herrera

Hubley Depression Scale for Older Adults (HDS-OA): Reliability, Validity, and a Comparison to the Geriatric Depression Scale

Validation of Scales

Excellence in Prevention descriptions of the prevention programs and strategies with the greatest evidence of success

SECONDARY SCHOOL STUDENTS TEST ANXIETY AND ACHIEVEMENT IN ENGLISH

N Utilization of Nursing Research in Advanced Practice, Summer 2008

Investigating the Reliability of Classroom Observation Protocols: The Case of PLATO. M. Ken Cor Stanford University School of Education.

Using the WHO 5 Well-Being Index to Identify College Students at Risk for Mental Health Problems

26:010:557 / 26:620:557 Social Science Research Methods

Personal Listening Profile Research Report

August 10, School Name Reason(s) for not making AYP Key actions underway to address the issues McKinley Adult and Alternative Education Center

Behavioral Intervention Rating Rubric. Group Design

Conners CPT 3, Conners CATA, and Conners K-CPT 2 : Introduction and Application

An International Study of the Reliability and Validity of Leadership/Impact (L/I)

Warren Consolidated Schools

LEADS: For Youth (Linking Education and Awareness of Depression and Suicide)

Associate Prof. Dr Anne Yee. Dr Mahmoud Danaee

The Youth Experience Survey 2.0: Instrument Revisions and Validity Testing* David M. Hansen 1 University of Illinois, Urbana-Champaign

Validity and reliability of measurements

Running head: CPPS REVIEW 1

Comparing Multiple-Choice, Essay and Over-Claiming Formats as Efficient Measures of Knowledge

Development and Validation of Implicit Measures of Organizational Citizenship Motives. Tonielle Fiscus and Donald Fischer Missouri State University

Manual for the ASEBA Brief Problem Monitor for Ages 6-18 (BPM/6-18)

Examining the efficacy of the Theory of Planned Behavior (TPB) to understand pre-service teachers intention to use technology*

Education Options for Children with Autism

GENERALIZABILITY AND RELIABILITY: APPROACHES FOR THROUGH-COURSE ASSESSMENTS

Reliability & Validity Dr. Sudip Chaudhuri

CHAPTER 4: FINDINGS 4.1 Introduction This chapter includes five major sections. The first section reports descriptive statistics and discusses the

THE APPLICATION OF ORDINAL LOGISTIC HEIRARCHICAL LINEAR MODELING IN ITEM RESPONSE THEORY FOR THE PURPOSES OF DIFFERENTIAL ITEM FUNCTIONING DETECTION

Transcription:

Construct Validation of Direct Behavior Ratings: A Multitrait Multimethod Analysis NASP Annual Convention 2014 Presenters: Dr. Faith Miller, NCSP Research Associate, University of Connecticut Daniel Cohen, M.P.H. Research Assistant, University of Missouri Wesley Sims, CAGS, NCSP Research Assistant, University of Missouri Contributors: Dr. Megan Welsh University of Connecticut Dr. Sandra Chafouleas University of Connecticut Dr. Chris Riley-Tillman University of Missouri Dr. Gregory Fabiano University at Buffalo

2 Purpose: To discuss the importance of understanding the psychometric properties of assessments To review the development of Direct Behavior Ratings Single Item Scales To review results from a multitrait multimethod (MTMM) investigation of DBR To discuss implications for practice

3

4 The importance of the assessment process: We need reliable and valid data in order to support students Nearly all of our decisions depend on it Understanding the strengths and limitations of our assessments is essential Different assessments provide us with different information Data-based decisionmaking Intervention Assessment

5 Purpose of Assessment Screening Who needs help? Diagnosis Why is the problem occurring? Progress Monitoring Is intervention working? Evaluation How well are we doing overall? Emphasized within a Multi- Tiered Service Delivery Framework (RTI) Within each category, we can assess different traits using different methods: what are we measuring and how are we measuring it?

6 Behavioral Assessment Rating Scales Teacher Report Parental Report Student Report Inattention Screening Observations Event recording Time sampling Hyperactivity Social Skills Evaluation Diagnosis Motivation Disruptive Behavior Interviews Unstructured Semi-structured Structured Progress Monitoring Internalizing Problems Extant Data ODRs Attendance

7 School-based behavior assessment within RTI Current methods of behavior assessment were not built for multi-tiered assessment New options must possess four desirable characteristics Defensible Efficient Flexible Repeatable Desirable Features (Chafouleas, 2011; Chafouleas, Christ, & Riley-Tillman, 2009; Chafouleas, Volpe, Gresham, & Cook, 2010)

8 Direct Behavior Rating

9 What is DBR? An emerging alternative to systematic direct observation and behavior rating scales which involves brief ratings of target behaviors following a specified observation period

10 A little background Other Names for DBR-like Tools: Home-School Note Behavior Report Card Daily Progress Report Good Behavior Note Check-In Check-Out Card Performance-based behavioral recording Contemporary Defining Features: BRS SDO Used repeatedly to represent behavior that occurs over a specified period of time (e.g., 4 weeks) and under specific and similar conditions (e.g., 45 min. morning seat work)

11 Example Scale Formats for DBR Source: Chafouleas, Riley-Tillman, & Christ (2009)

12 DBR-SIS AE DB Core Behavioral Competencies RS

DBR-SIS Target Behaviors 13

14 Development & Validation of DBR-SIS

15 RESEARCH: Project VIABLE (2006-2011) and Project VIABLE II (2011-current) Develop instrumentation and procedures, then evaluate defensibility of DBR in decisionmaking Rating Procedures Method Comparisons Behavior Targets Rater Training Defensibility Scale Design Evaluate defensibility and usability of DBR in decision-making at larger scale Triannual behavioral screening Single-case design studies using DBR DBR Multi-trait multimethod investigation Teacher input regarding usability and perceptions Funding provided by the Institute for Education Sciences, U.S. Department of Education

16 Development & Validation Applications in Screening Developing cut scores to identify students at-risk Concurrent validity with established screeners: SRSS, BESS Examining bias Development & Validation of DBR-SIS Scale development Behavior wording Training Influence of observation duration How teachers assign ratings Perceptions of usability Applications in Progress Monitoring Determining scale sensitivity to change Concurrent validity with SDO

17 Questions Remain Foundational psychometric evidence of DBR-SIS Reliability evidence Accuracy or precision of scores Validity evidence The extent to which it is appropriate to use DBR-SIS for screening and progress monitoring Many different types of validity evidence Here, we focus on construct validity

18 Multitrait Multimethod Analysis

19 Rationale Test developers must accurately define, measure, and rigorously validate the construct(s) of interest Campbell and Fiske (1959) developed an approach to assessing construct validity MTMM analysis permits the examination of: Convergent validity - evidence that scores are consistent with other measures of the same trait Discriminant validity evidence that scores diverge from measures of similar, but distinct traits Examining both convergent and discriminant evidence contributes to validity argument by determining not only whether a measure is consistent with criterion measures of the same construct, but also whether the measure is less strongly associated with measures of different, but related constructs

20 Purpose of MTMM Analysis Provides a way to systematically evaluate the correlations among a set of measures Correlations tell us the degree of association between variables Evaluate construct validity Convergent validity Discriminant validity Evaluate variance attributed to traits vs. methods Behavioral traits & measurement methods Behavioral Data

21 Example MTMM Matrix What are we looking for? High reliability coefficients Correlations between measures of the same trait obtained using different methods should be large Correlations between measures of the same trait obtained through different methods should be stronger than those observed between different traits using the same method The same pattern of trait correlations should hold for all methods and all combinations of methods K. Widaman (2010)

22 Primary Research Questions How are scores obtained from DBR-SIS associated with other measures of school-based behavior? Evidence for convergent validity? Evidence for discriminant validity? Do there appear to be strong methods factors associated with various measures of behavior?

23 Methods Participants and Setting: 993 students 122 teachers Public school settings were located in 4 states: Connecticut, Rhode Island, New York, and Missouri Students were enrolled in a total of 19 different schools, including rural, suburban, and urban districts Participating students were in grades 3-8

24 Student characteristic n % Gender Male 452 45 Female 541 55 Race Caucasian 780 79 African American 154 16 Asian 35 4 Other 12 1 Ethnicity Hispanic 89 9 Non-Hispanic 904 91 Grade Third 210 21 Fourth 204 21 Fifth 206 21 Sixth 166 17 Seventh 124 12 Eighth 83 8

25 Methods: Measures DBR-SIS teacher ratings: AE, DB, RS DBR-SIS student ratings: AE, DB, RS SDO observations: AE, DB, RS Momentary time sampling, 10 second intervals Teacher rating scales Attention Problems Subscale (BASC-2) Hyperactivity Subscale (BASC-2) Communication Subscale (SSIS Rating Scale) Student self-report rating scales Attention Problems Subscale (BASC-2) Hyperactivity Subscale (BASC-2) Communication Subscale (SSIS)

26 Methods: Procedures Data collection occurred in a single assessment period in winter/spring of 2013 Up to 10 students could participate per classroom Teachers and students were asked to complete: a) DBR-SIS scales over 10 occasions (one week) b) Behavior rating scales matched to the target constructs External observers completed SDO observations Goal: 3+ 15 minute observations IOA observations were also conducted Assessment order was counterbalanced in order to control for potential order effects

27 Results 3 (trait) x 5 (method) matrix Reliability coefficients were calculated as follows: DBR-SIS Teacher: derived from intraclass correlation coefficient (ICC) DBR-SIS Student: derived from intraclass correlation coefficient (ICC) SDO: Pearson s product moment correlations (inter-rater reliability) Teacher rating scales: internal consistency (α) Student rating scales: internal consistency (α)

28 Method 1 Method 2 Method 3 Method 4 Method 5 A B C A B C A B C A B C A B C 1. DBR Teacher a. Academic Engagement.90 b. Disruptive Behavior -.87.88 c. Respectful.81 -.91.88 2. DBR- Student a. Academic Engagement.49 -.41.41.82 Rules of thumb for interpreting correlations: <.20 = Weak.20-.69 = Moderate >.69 = Strong b. Disruptive Behavior -.45.48 -.44 -.75.80 c. Respectful.45 -.47.47.96 -.84.81 3. SDO a. Academic Engagement.37 -.39.33.27 -.29.30.93 b. Disruptive Behavior -.29.35 -.30 -.23.23 -.24 -.80.96 c. Respectful.21 -.28.30.16 -.19.23.48 -.61.78 4. Rating Scale Teacher a. Academic Engagement 1 -.75.63 -.55 -.39.41 -.35 -.24.23 -.20.95 b. Disruptive Behavior 2 -.58.71 -.65 -.35.41 -.39 -.28.28 -.27.76.95 c. Respectful 3.55 -.50.48.33 -.31.31.23 -.18.20 -.67 -.55.93 5. Rating Scale - Student a. Academic Engagement 1 -.47.41 -.34 -.53.50 -.53 -.25.23 -.25.48.39 -.40.77 b. Disruptive Behavior 2 -.34.39 -.32 -.38.45 -.42 -.23.24 -.21.36.47 -.24.80.75 c. Respectful 3.14 -.16.11.30 -.29.33.10 -.08.03 -.15 -.15.16 -.42 -.36.71 Note. 1 BASC-2 Attention Problems Subscale, 2 BASC-2 Hyperactivity Subscale, 3 SSIS-RS Communication Subscale

29 Results Reliability coefficients were highest for the teacher rating scales, and lowest for the student rating scales Reliability coefficients across methods were generally high Validity diagonals provide information on convergent validity Coefficients were variable Higher for AE & DB (Moderate to Strong) Lower for RS (Weak to Moderate) Analysis of heterotrait-monomethod triangles suggests method effects Same method, different traits, strong correlations Validity coefficients were often similar in magnitude to those in the heterotrait-heteromethod triangles Are traits distinct? Does the method effect overpower the trait effect?

30 Primary Research Questions How are scores obtained from DBR-SIS associated with other measures of school-based behavior? Evidence for convergent validity? Yes: Teacher DBR and Teacher Rating Scale No: Student Rating Scale and SDO, Student DBR Evidence for discriminant validity? Limited evidence Do there appear to be strong methods factors associated with various measures of behavior? Yes, method seems to matter

31 Next steps Structural Equation Modeling Account for nesting of students within teachers Estimate trait and method related variance Test the amount of trait-related and method-related variance statistically

32 Discussion Implications for practice What are the implications of these findings on assessment selection? Our methods impact our results As school psychologists, should we be surprised when we find varied results using different assessment methods? Do you think these measurement challenges are unique to behavioral assessment?

Website: www.directbehaviorratings.org Contact: faith.miller@uconn.edu 33