Reliability and Validity Plan 2017

Similar documents
A pre-conference should include the following: an introduction, a discussion based on the review of lesson materials, and a summary of next steps.

EXPLORING THE PROCESS OF ASSESSMENT AND OTHER RELATED CONCEPTS

Year 10 Food Technology. Assessment Task 1: Foods for Special Needs. Name: Teacher:

PET FORM Planning and Evaluation Tracking ( Assessment Period)

Assessment Field Activity Collaborative Assessment, Planning, and Support: Safety and Risk in Teams

Improving Surveillance and Monitoring of Self-harm in Irish Prisons

Accounting Assessment Report

Campus Climate Survey

2018 Medical Association Poster Symposium Guidelines

Annual Assembly Abstract Review Process

P02-03 CALA Program Description Proficiency Testing Policy for Accreditation Revision 1.9 July 26, 2017

Frontier School of Innovation District Wellness Policy

Statement of Work for Linked Data Consulting Services

EASTERN ARIZONA COLLEGE Criminal Investigation

Catherine Worthingham Fellows of APTA Instructions for Writing a Letter of Support

New Mexico Striving Toward Excellence Program (NM STEP), The Data Scholars Initiative for Child Welfare

MGPR Training Courses Guide

Structured Assessment using Multiple Patient. Scenarios (StAMPS) Exam Information

Assessment criteria for Primary Health Disciplines Eligibility for Recognition as Credentialled Diabetes Educator. December 2015 ADEA

State Health Improvement Plan Choosing Priorities, Creating a Plan. DHHS DPH - SHIP Priorities (Sept2016) 1

STAKEHOLDER IN-DEPTH INTERVIEW GUIDE

Catherine Worthingham Fellows of the APTA Instructions for Nominators

CONSENT FOR KYBELLA INJECTABLE FAT REDUCTION

Novel methods and approaches for sensing, evaluating, modulating and regulating mood and emotional states.

HSC 106 Personal Health Plan for Learning Activities & Assessment linked to Michigan Teacher Preparation Standards

Programme of Learning. Physical Education. Key Stage 4 Year 10 BTEC Sport

Practicum Evaluation Form - Formative Assessment

Success Criteria: Extend your thinking:

PROVIDER ALERT. Comprehensive Diagnostic Evaluation (CDE) Guidelines to Access the Applied Behavior Analysis (ABA) Benefit.

PROCEDURAL SAFEGUARDS NOTICE PARENTAL RIGHTS FOR PRIVATE SCHOOL SPECIAL EDUCATION STUDENTS

Interpretation. Historical enquiry religious diversity

VCCC Research and Education Lead for Breast Cancer

1/19/2018. Writing Proficiency: Memos Self Leadership Self Awareness and Leadership

Frequently Asked Questions: IS RT-Q-PCR Testing

Completing the NPA online Patient Safety Incident Report form: 2016

1/16/2019. Self Leadership Self Awareness and Leadership. Leader Traits and self-awareness Role management exercise Leadership Philosophy

UNIT 6. DEVELOPING THREAT/HAZARD-SPECIFIC ANNEXES

Breast Cancer Awareness Month 2018 Key Messages (as of June 6, 2018)

FDA Dietary Supplement cgmp

Susan Wortman, Career Development Center

Commissioning Policy: South Warwickshire CCG (SWCCG)

DATA RELEASE: UPDATED PRELIMINARY ANALYSIS ON 2016 HEALTH & LIFESTYLE SURVEY ELECTRONIC CIGARETTE QUESTIONS

ACSQHC National Consensus Statement: Essential Elements for High Quality End-oflife Care in Acute Hospitals.

Dysart Unified School District

Corporate Governance Code for Funds: What Will it Mean?

EDPS 475: Instructional Objectives for Midterm Exam Behaviorism

2017 PEPFAR Data and Systems Applied Learning Summit Day 2: MER Analytics/Available Visualizations, Clinical Cascade Breakout Session TB/HIV EXERCISE

International Integrative Psychotherapy Association IIPA-

True Patient & Partner Engagement How is it done? How can I do it?

Podcast Transcript Title: Common Miscoding of LARC Services Impacting Revenue Speaker Name: Ann Finn Duration: 00:16:10

NYS Common Core ELA & Literacy Curriculum Grade 12 Module 4 Unit 1 Lesson 14

FOUNDATIONS OF DECISION-MAKING...

Forensic Science: Fundamentals & Investigations, 2e Chapter 2. All rights Reserved Cengage/NGL/South-Western 2016

GUIDANCE DOCUMENT FOR ENROLLING SUBJECTS WHO DO NOT SPEAK ENGLISH

Human papillomavirus (HPV) refers to a group of more than 150 related viruses.

Nutrition Care Process Model Tutorials. Nutrition Monitoring & Evaluation: Overview & Definition. By the end of this module, the participant will:

Test Blueprinting II: Creating a Test Blueprint NBME

CSHCN Services Program Benefits to Change for Outpatient Behavioral Health Services Information posted November 10, 2009

Introduction Teaching Interpretation

Position Title Diabetes Educator Program / Funding Stream Primary Health Care

Physical Fitness for the Physically Limited. o Work Experience, General. o Open Entry/Exit. Distance (Hybrid Online) for online supported courses

Organizational Capacity for Change and Patient Safety

HOSA 105 EMERGENCY PREPAREDNESS

FUNCTIONAL MOVEMENT SYSTEMS SCREEN FINDINGS REPORT

Continuous Quality Improvement: Treatment Record Reviews. Third Thursday Provider Call (August 20, 2015) Wendy Bowlin, QM Administrator

US Public Health Service Clinical Practice Guidelines for PrEP

A Phase I Study of CEP-701 in Patients with Refractory Neuroblastoma NANT (01-03) A New Approaches to Neuroblastoma Therapy (NANT) treatment protocol.

The effects of a two-school. school-year. back education program. in elementary schoolchildren

British Sign Language (BSL) Plan October 2018 Scottish Charity Regulator

Module 6: Goal Setting

Annual Principal Investigator Worksheet About Local Context

LEVEL OF CARE GUIDELINES: INTENSIVE BEHAVIORAL THERAPY/APPLIED BEHAVIOR ANALYSIS FOR AUTISM SPECTRUM DISORDER HAWAII MEDICAID QUEST

University of Rochester Course Evaluation Project. Ronald D. Rogge. Associate Professor. Ista Zahn. Doctoral Candidate

Effective date: 15 th January 2017 Review date: 1 st May 2017

PSYCHOSEXUAL ASSESSMENTS for Children and Adolescents with Problematic Sexual Behavior. Who is qualified to conduct a psychosexual evaluation?

The estimator, X, is unbiased and, if one assumes that the variance of X7 is constant from week to week, then the variance of X7 is given by

TRAINING INDIVIDUAL IMAGE INTERPRETERS USING TEAM CONSENSUS FEEDBACK. John T. Cockrell System Development Corporation. and

Implementation of Early retention monitoring of HIV positive pregnant and breastfeeding women; and data use in the EMTCT program MOH-UGANDA

Humanities and Social Sciences Division. o Work Experience, General. o Open Entry/Exit. Distance (Hybrid Online) for online supported courses

Lesson Unit content* Activities Resource checklist Links to other units

Impacts of State Level Dental Hygienist Scope of Practice on Oral Health Outcomes in the U.S. Population

(Please text me on once you have submitted your request online and the cell number you used)

Independent Charitable Patient Assistance Program (IPAP) Code of Ethics

Pain Management Learning Plan

Bariatric Surgery FAQs for Employees in the GRMC Group Health Plan

Health Screening Record: Entry Level Due: August 1st MWF 150 Entry Year

World Confederation for Physical Therapy Congress , May Singapore

NIA Magellan 1 Spine Care Program Interventional Pain Management Frequently Asked Questions (FAQs) For Medicare Advantage HMO and PPO

CNMC Rounds: Can CME save lives? Dave Davis, MD Senior Director, Continuing Education & Performance Improvement (with Nancy Davis, PhD)

Newborn Hearing Screening, Early Identification and Loss to Follow-Up

Q 5: Is relaxation training better (more effective than/as safe as) than treatment as usual in adults with depressive episode/disorder?

GSB of EDA Meeting Minutes

Cardiac Rehabilitation Services

WISCONSIN ORAL HEALTH COALITION ACCESS WORKGROUP FOCUS CHART WORKGROUP SUMMARY CURRENT STATUS ACTION ITEMS TOOLS NECESSARY

Pediatric and adolescent preventive care and HEDIS *

Related Policies None

Clinical Orthopaedic Rehabilitation Spinal Disorders

ICT4LIFE. Final Conference. ICT4Life field work - tailored solutions in diverse regional context Ariane Girault, E-Seniors Association

1100 Marie Mount Hall College Park, Maryland Tel: (301) Fax: (301)

Chapter 6: Impact Indicators

Transcription:

Reliability and Validity Plan 2017 Frm CAEP The principles fr measures used in the CAEP accreditatin prcess include: (a) validity and reliability, (b) relevance, (c) verifiability, (d) representativeness, (e) cumulativeness, (f) fairness, (g) stakehlder interest, (h) benchmarks, (i) vulnerability t manipulatin, and (j) actinability 1. CAEP requires valid and reliable assessments t demnstrate candidate quality and that varius stakehlders must cntribute t the validity f the assessments. Validity and reliability are tw f the mst imprtant criteria when assessing instruments. Reliability means cnsistency and a test is valid if it measures what it is suppsed t measure. Befre we start, here is a reminder f CAEP Standard 5 and cmpnent 5.2. Standard 5: Prvider Quality, Cntinuus Imprvement, and Capacity The prvider maintains a quality assurance system cmprised f valid data frm multiple measures, including evidence f candidates and cmpleters psitive impact n P-12 student learning and develpment. The prvider supprts cntinuus imprvement that is sustained and evidence-based, and that evaluates the effectiveness f its cmpleters. The prvider uses the results f inquiry and data cllectin t establish pririties, enhance prgram elements and capacity, and test innvatins t imprve cmpleters impact n P-12 student learning and develpment. Cmpnents: Quality and Strategic Evaluatin 5.2 The prvider s quality assurance system relies n relevant, verifiable, representative, cumulative and actinable measures, and prduces empirical evidence that interpretatins f data are valid and cnsistent. Accrding t CAEP s Evidence Guide, the respnsibility lies with the EPP t prvide valid (and reliable) evidence. CAEP is cmmitted t strnger preparatin and accreditatin data. The prfessin needs evidence that assessment is intentinal, purpseful, and addresses deliberately psed questins f imprtance. Such reprting entails interpretatin and reflectin; measures need t be integrated and hlistic; and appraches t its assessment can be qualitative and quantitative, and direct and indirect. ALL EPP created assessments used in the CAEP review must meet the Sufficient level n the CAEP Instrument rubric. 1 Ewell, P. (2013). Principles fr measures used in the CAEP accreditatin prcess.

Submissin fr EPP-Created instruments 1. A cpy f the assessment 2. Data chart(s) 3. Instructins t candidates 4. Can include a page n hw yu addressed validity and reliability r yu can simply respnd t the 5 questins 5. (ptinal) Can include an analysis f data fr this instrument r this may be in yur Self Study Reprt Narrative 6. Respnse t five questins in AIMS Hw t address reliability and validity MSU EPP self-studies needs t include evidence related t the reliability and validity f the reprted data. Reliability and validity are frequently measured quantitatively. EPP s quantitative apprach t assess the reliability f instruments can invlve fur facets: 1. Supervisr (e.g., inter-rater reliability, internal cnsistency, bias) 2. Candidate (e.g., distributin f ratings) 3. Item (e.g., variability f items) 4. Time (e.g., variability f candidate perfrmance acrss time) Befre we review quantitative ways f assessing validity, we shuld cnsider that describing the reliability and validity f EPP created instruments has nt been an easy task fr institutins cmpleting their CAEP selfstudies. CAEP has recmmended the inclusin f bth quantitative and qualitative appraches when describing reliability and validity. The infrmatin prvided next cmes frm the Breaux and Ellit s presentatin: Reliability and Validity: Establishing and Cmmunicating Trustwrthy Findings (CAEP Spring Cnference 2015). Althugh the terms validity and reliability are traditinally assciated with quantitative research, CAEP des nt mean t imply that nly quantitative data are expected r valued.

Differences between quantitative and qualitative methds when establishing validity and reliability: Quantitative These methds f establishing validity and reliability are easier t describe briefly. The standards fr judging the results are less subjective. They require statistical literacy, and the results are decntextualized. Qualitative These methds f establishing validity and reliability depend much mre n anticipating and discnfirming a variety f ptential dubts that varius readers culd have. The prcess takes mre effrt, and the reader s judgment is less predictable. They require strng skills in lgical argumentatin and expsitry writing, but the results are mre cntextualized. Strategies Used in CAEP Self-Studies: Reliability 1. Quantitative studies explain hw they manage subjectivity using cmmn terminlgy, standard prcedures, and unifrm frmats fr reprting results. We need t make sure the crrect prcedures are selected, cnducted prperly, interpreted validly, and cmmunicated clearly. Fcus n key reliabilities. Use inter-rater crrelatins fr large. Use rater agreement fr small samples. The instrument s reliability sectin n CAEP s Instrument Rubric addresses the fllwing: Degree in which an assessment prduces stable and cnsistent results Ask the questin - Can the evidence be crrbrated? Criteria A detailed descriptin r plan is prvided Training f scrers and checking n inter-rater reliability are dcumented Steps are described that meet accepted research standards fr establishing interrater reliability Strategies Used in CAEP Self-Studies: Validity It is nt necessary t establish every frm. Sme f the prcesses are qualitative and invlve demnstrating alignment. Others are quantitative and invlve calculating values. Quantitative methds t assess validity; Fcused n key validities: Cntent: all relevant elements f the cnstruct are measured Cnstruct: measures intended attribute

Criterin: measures f attributes predict target behavirs Cncurrent: crrelates with a knwn gd measure Predictive: predicts scre n a significant future measure Cnvergent: measure crrelates with related measures The types f validity that are needed are judgment calls that have t be justified in the self-study s ratinale. Hwever, cntent validity and cnstruct validity shuld be included. Example When cperating teachers and university supervisrs rate a candidate, we need t shw that the assessment is a valid measure f the cnstruct r cnstructs; and that bth raters understand the items and verall intent in the same way. T shw the assessment is a valid measure: Expert judgment: what d university supervisrs and cperating teachers say? Alignment with relevant standards Agreement with lgically-related measures Is there sufficient variance in the evidence? T shw the assessment is a reliable measure: Inter-rater agreement Mre n Cntent Validity Fllwing Dr. Stevie Chepk s view, there are three imprtant cmpnents t establish cntent validity: 1. Determining the bdy f knwledge fr the cnstruct t be measured. The agreement amng experts requires the use f recgnized subject matter experts and it is based n their judgment. It als relies n individuals wh are familiar with the cnstruct such as faculty members, EPP based clinical educatrs, and/r P-12 based clinical educatrs. The key is having them answer the fundamental questin: D the indicatrs assess the cnstruct t be measured? 2. Aligning indicatrs t cnstruct. Indicatrs must assess sme aspects r segment f the cnstruct and indicatrs must align with the cnstruct. 3. Using Lawshe s Cntent Validity Rati

4. Lawshe s Cntent Validity Rati (CVR) Perfrmance dmains: Behavirs that are directly bservable Can be a simple prficiencies Can be higher mental prcess (inductive/deductive reasning) Operatinal definitin Extent t which verlap exists between perfrmance n assessment under investigatin, and ability t functin in the defined jb Attempts t identify the extent f the verlap The Cntent Evaluatin Panel is cmpsed f persns knwledgeable abut the jb, and it is mst successful when it is a cmbinatin f P-12 based clinical educatrs, EPP based clinical educatrs, and faculty. Each panel member is given the list f indicatrs r items independently and are asked t d the fllwing: Rate the item as essential, useful but nt essential, r nt necessary. Items/indicatrs must be aligned with the cnstruct being measured T quantifying cnsensus, any item/indicatr which is perceived as essential by mre than half f the panelists, has sme degree f cntent validity. The mre panelist (beynd 50%) wh perceive the indicatr as essential, the greater the extent r degree f its cntent validity Calculating the cntent validity rati (CVR) n e = number f panelists indicating essential N = ttal number f panelists Cmpare answer with CVR chart t determine CVR value based n the number f panelists. CVR is calculated fr each indicatr, and minimum value f the CVR is based n the number f panelists and is n a CVR Table. Keep r reject individual items based n the table results. CVR values range frm -1.0 t + 1.0. The mre panelists, the lwer the CVR value. Fr example, 5 panelists requires minimum CVR value f.99 15 panelists requires minimum CVR value f.60 40 panelists requires minimum CVR value f.30

Anther Methd f Establishing Cntent Validity Cnduct a jb-task analysis t identify essential jb tasks, knwledge areas, skills and abilities Link jb tasks, knwledge areas r skills t the assciated test cnstruct r cmpnent that it is intended t assess Use subject-matter experts The instrument s validity sectin n CAEP s Instrument Rubric addresses the fllwing: A descriptin r plan is prvided Describes the steps t be used fr determining cntent validity Research was used in the develpment f the plan Pilt was cmpleted prir t administratin O Steps meet accepted research standards/prtcls Lawshe s methd (CVR) Questins t Be Answered fr each Submitted EPP- Created Instrument 1. During which part f the candidate's experience is the assessment used? Is the assessment used just nce r multiple times during the candidate's preparatin? 2. Wh uses the assessment and hw are the individuals trained n the use f the assessment? 3. What is the intended use f the assessment and what is the assessment purprted t measure? 4. Please describe hw validity/trustwrthiness was established fr the assessment. 5. Please describe hw reliability/cnsistency was established fr the assessment. MSU EPP Hmegrwn Instruments We currently use seven EPP-created instruments t assess educatin prgrams at MSU. We need t determine the adequacy f these measures fr the accreditatin prcess. The measures are: 1. Experiential Lg 2. Candidate Prfessinal Dispsitin Traits 3. Missuri Educatr Evaluatin System (MEES) Rubric

4. Diversity Prficiencies 5. Cmprehensive Exam Assessment Rubric fr Advanced Prgrams 6. Research Rubric fr Advanced Prgrams 7. Student teaching exit survey (based n a prprietary survey) 8. EDC345 Multiculturalism Lessn Plan Next Steps 1. Revisit the instruments. Des it address critical elements required by CAEP? Which nes? 2. Make necessary adjustments. Define in specific terms what shuld be addressed and assessed Align with prgram bjectives, CAEP, InTASC and/r state standards Clarify language make sure we include distinguishable and measurable statements in rubrics 3. Use CAEP s Assessment Rubric as a guide 4. Establish cntent validity. Create Cntent Evaluatin Panels fr each instrument cmpsed f P-12 based clinical educatrs, EPP based clinical educatrs, and faculty. Thrugh a survey, ask them t rate the items as essential, useful but nt essential, r nt necessary. Calculate the Lawshe s Cntent Validity Rati (CVR) fr each item f the assessment t determine which items will remain. Fcus grups t discuss the cntent f each instrument 5. Establish inter-rater reliability. Use spring 2017 data as a pilt f the instruments When pssible, instructrs within a prgram shuld scre at least 3 samples independently f ne anther Cllect results and calculate the percentage f agreement n each cmpnent & submissin

If scres vary and yield <80% agreement, meet t discuss each item scre n each submissin Pay attentin t: Discrepancies between/amng scrers Whether discrepancies are due t language r hw items are defined Reslving discrepancies with clarified language, rearranging items, r ther changes Make nte f these changes and revise the assessment as necessary If substantial changes are necessary, each instructr shuld scre at least 2 wrk samples independently f ne anther; until instructrs reach at least 80% agreement When pssible triangulate (cmpare cperating teacher and supervisr results) When pssible, cmpare results ver time (chrt and panel-wise) When summarizing reliability, try t include data n the fllwing: Supervisr (e.g., inter-rater reliability, internal cnsistency, bias) Candidate (e.g., distributin f ratings) Item (e.g., variability f items) Time (e.g., variability f apprentice perfrmance acrss time)