What Counts as Credible Evidence in Evaluation Practice?

Similar documents
What Counts as Credible Evidence in Contemporary Evaluation Practice?

Reviewed by Jason T. Burkhardt and Jan K. Fields Western Michigan University

What is program evaluation?

Realism and Qualitative Research. Joseph A. Maxwell George Mason University

CONFLICT MANAGEMENT Fourth Edition

Student Name: XXXXXXX XXXX. Professor Name: XXXXX XXXXXX. University/College: XXXXXXXXXXXXXXXXXX

Definitions of Nature of Science and Scientific Inquiry that Guide Project ICAN: A Cheat Sheet

Chapter 1 Introduction to Educational Research

Durkheim. Durkheim s fundamental task in Rules of the Sociological Method is to lay out

What is good qualitative research?

CRIMINOLOGY TODAY. AN INTEGRATIVE INTRODUCTION sixth edition. By FRANK SCHMALLEGER. Pearson Education, Inc.

CONCEPTUAL FRAMEWORK, EPISTEMOLOGY, PARADIGM, &THEORETICAL FRAMEWORK

CHAPTER 3. Methodology

Repositioning Validity

The Objects of Social Sciences: Idea, Action, and Outcome [From Ontology to Epistemology and Methodology]

2) Ethical and scientific standards that guide the use of quarantine or other movement restrictions during public health emergencies.

Ideas RESEARCH. Theory, Design Practice. Turning INTO. Barbara Fawcett. Rosalie Pockett

Theoretical Perspectives in the PhD thesis: How many? Dr. Terence Love We-B Centre School of Management Information Systems Edith Cowan University

Qualitative Data Analysis. Richard Boateng, PhD. Arguments with Qualitative Data. Office: UGBS RT18 (rooftop)

The 2004 Claremont Debate: Lipsey vs. Scriven

The debate about the Randomized Controlled Trial (RCT) in impact evaluation:

Developing Program Logic Models B.E.S.T. Collaboration in LA County

Benefits and constraints of qualitative and quantitative research methods in economics and management science

A QUALITY FRAMEWORK FOR CASE STUDY RESEARCH: CONVINCINGNESS

Defining Evidence-Based : Developing a Standard for Judging the Quality of Evidence for Program Effectiveness and Utility

9/5/ Research Hazards Awareness Training

An evidence rating scale for New Zealand

Issues in Conducting Rigorous and Relevant Research in Education

SOCQ121 & BIOQ321. Session 12. Research into Practice. Department: Social Science.

1 Qualitative Research and Its Use in Sport and Physical Activity

Research Methodology in Social Sciences. by Dr. Rina Astini

In this second module, we will focus on features that distinguish quantitative and qualitative research projects.

GENERAL EDUCATION JOURNAL; Vol. 4; 1ssue 2; Pages 1-14; September 2015; Published By Mount Meru University Research Unit

Narrative Analysis Handout

Instructor s Test Bank. Social Research Methods

Research in Education. Tenth Edition

PLANNING THE RESEARCH PROJECT

Positioning for Sustainability: Developing a Logic Model

This article, the last in a 4-part series on philosophical problems

Table of Contents. Chapter 1 Theoretical Criminology: An Introductory Overview [page 79] Chapter 3 Biosocial Theories of Crime [page 99]

CRIMINAL JUSTICE (CJ)

Whose psychological concepts?

Audio: In this lecture we are going to address psychology as a science. Slide #2

Grounded Theory s Contested Family of Methods: Historical and Contemporary Applications

Mommy Wars Redux: A False Conflict

THE UNIVERSITY OF VERMONT SOCIAL AND BEHAVIORAL FOUNDATIONS OF PUBLIC HEALTH PH 395 OL2 CRN: 61108

Qualitative and Quantitative Approaches Workshop. Comm 151i San Jose State U Dr. T.M. Coopman Okay for non-commercial use with attribution

INTERVIEWS II: THEORIES AND TECHNIQUES 5. CLINICAL APPROACH TO INTERVIEWING PART 1

DOING SOCIOLOGICAL RESEARCH C H A P T E R 3

Basics of philosophy of science

Daniel Little University of Michigan-Dearborn www-personal.umd.umich.edu/~delittle/

society. The social perspective is a way of looking at society. It sees society as something over and above the very people who are in that society.

FORUM: QUALITATIVE SOCIAL RESEARCH SOZIALFORSCHUNG

3: (Semi-)Structured Interviews. Tina FREYBURG [Introduction to Qualitative Methods]

Theory and Methods Question Bank

It is the theory that decides what we can observe. -- Albert Einstein

Final Strategic Long-Range Plan

DEFINING THE CASE STUDY Yin, Ch. 1

Qualitative Research. Prof Jan Nieuwenhuis. You can learn a lot just by watching

Building Leaders Worth Following 5 VOICES QUICK START GUIDE. From Steve Cockram & Jeremie Kubicek Authors of 5 Voices Co-Founders, GiANT Worldwide

Insight Assessment Measuring Thinking Worldwide

Conducting Research in the Social Sciences. Rick Balkin, Ph.D., LPC-S, NCC

Checklist of Key Considerations for Development of Program Logic Models [author name removed for anonymity during review] April 2018

Developmental Perspectives on Problem-Solving

The Significance of Mixed Methods Research in Information Systems Research

reward based power have ability to give you what you want. coercive have power to punish

TSE in Animal Populations - Fact and Fiction September 10 11, 2003 Fort Collins, Colorado Conrad G. Brunk, PhD University of Victoria, Canada

Validity Arguments for Alternate Assessment Systems

Discussion on David Olson s book Psychological Theory and Educational Reform, Cambridge University Press, 2003.

INTERVIEWS II: THEORIES AND TECHNIQUES 1. THE HUMANISTIC FRAMEWORK FOR INTERVIEWER SKILLS

NEGOTIATION SEVENTH EDITION

FORUM: QUALITATIVE SOCIAL RESEARCH SOZIALFORSCHUNG

1 of 16 24/05/ :06

Groups. Giovanni Masino - University of Ferrara

Educational Research

Review of Animals and the Economy. Steven McMullen Palgrave, pp., ebook and hardcover. Bob Fischer Texas State University

1 Integrative Multi-Method Research

BOR 3305 PERSPECTIVES ON CRIME IN AMERICA. Eight Week Course TEXTBOOK:

Glossary of Research Terms Compiled by Dr Emma Rowden and David Litting (UTS Library)

Wither Participation and Empowerment?

INTRODUCTION. Evidence standards for justifiable evidence claims, June 2016

Qualitative Research Design

Social Studies Skills and Methods Analyzing the Credibility of Sources

The Nature of Science: What is Science? A Effective Synthesis for Science Instruction. What is Science, Really?

Introduction to the Scientific Method. Knowledge and Methods. Methods for gathering knowledge. method of obstinacy

The 9 Virtues of Exceptional Leaders: Unlocking Your Leadership Potential Discussion Guide

Psy2005: Applied Research Methods & Ethics in Psychology. Week 14: An Introduction to Qualitative Research

Research in Development Studies: Philosophy, Methods and Rigor

Hypothesis testing crime analysis: SUMMARY (1 of 5)

SOCIAL WORK PROGRAM. MSW Degree Program Student Learning Plan

MOOC: February 2017: Qualitative Research Methods: Assignment 1

Paradoxical Theory of Change Arnold Beisser, M.D.

Effective communication can have a significant effect on client satisfaction.

Stephen Madigan PhD madigan.ca Vancouver School for Narrative Therapy

Survey Research. We can learn a lot simply by asking people what we want to know... THE PREVALENCE OF SURVEYS IN COMMUNICATION RESEARCH

OWNING OUR STORIES CHIEF KRISTEN ROMAN UNIVERSITY OF WISCONSIN-MADISON POLICE

Commentary on Constructing New Theory for Identifying Students with Emotional Disturbance

Fairfield Public Schools Social Studies Curriculum Sociology Grades 11-12

Autonomous Family Groups and Networks: A Promising Development

Variation in Theory Use in Qualitative Research

Transcription:

What Counts as Credible Evidence in Evaluation Practice? Stewart I. Donaldson, Ph.D. Claremont Graduate University stewart.donaldson@cgu.edu CDC/OSH Evaluation Net Conference March 28, 2012

"The issue of what constitutes credible evidence isn't about to get resolved. And it isn't going away. This book explains why. The diverse perspectives presented are balanced, insightful, and critical for making up one's own mind about what counts as credible evidence.

And, in the end, everyone must take a position. You simply can't engage in or use research and evaluation without deciding what counts as credible evidence. So read this book carefully, take a position, and enter the fray. (Patton, 2009)

Overview Contemporary Evaluation Practice Evidence-based Practice Debates about What Counts as Evidence Experimental Routes to Credible Evidence Non-Experimental Approaches Moving Beyond the Debates

Patton, 2009

Contemporary Evaluation Practice Booming Global Diverse Contexts Many More Evaluands Multidisciplinary Many New Approaches & Methods More than Traditional Social Science Research Methods

Second Boom in Evaluation Practice 1980s Only 3 National and Regional Evaluation Societies 1990 5 2000 More than 50 2006 More than 70 including a Formal International Cooperation Network 8

9

Evidence-Based Practice Highly Valued Global Multidisciplinary Many Applications

Sample of Applications Evidence-based Medicine Evidence-based Mental Health Evidence-based Management Evidence-based Decision Making Evidence-based Education Evidence-based Coaching

Sample of Applications Evidence-based Social Services Evidence-based Policing Evidence-based Conservation Evidence-based Dentistry Evidence-based Policy Evidence-based Thinking about Health Care

Sample of Applications Evidence-based Evidence-based Evidence-based Evidence-based Evidence-based Evidence-based Evidence-based Evidence-based Occupational Therapy Prevention Science Dermatology Gambling Treatment Sex Education Needle Exchange Programs Prices Education Help Desk

What Counts as Credible Evidence?

Sample of The Debates Qualitative-Quantitative Debate Visions for the Desire Future of Evaluation Practice AEA Statement vs. Not AEA Statement The Lipsey vs. Scriven Debate What Counts a Credible Evidence? EES Statement

Experimental Design: Gold Standard? Random Assignment Experimental Control Ruling Out Threats to Validity

Supreme Courts of Credible Evidence What Works Clearinghouse Campbell Collaboration Cochrane Collaboration

Experimental Approaches Henry: When Getting it Right Matters Bickman & Reich: RCTs - A Gold Standard with a Feet of Clay Gersten & Hitchcock The What Works Clearinghouse Julnes & Rog - Methods for Producing Actionable Evidence

Non-Experimental Approaches Scriven: Demythologizing Causation and Evidence Greene: Evidence as Proof and Evidence as Inkling Rallis: Reasoning With Rigor and Probity: Ethical Premises for Credible Evidence Mathison: Seeing Is Believing: The Credibility of Image- Based Research and Evaluation Schwandt: Toward a Practical Theory of Evidence for Evaluation

Challenges of the Gold Standard AEA Statement vs. Not AEA Statement Theoretical Practical Methodological Ethical Ideological Political Scriven s Summative Conclusion

Scriven, 2009 To insist we use RCTs is simply bigotry not pragmatic and not logical. In short, it is a dogmatic approach that is an affront to scientific method.

AEA Statement vs. Not AEA Statement AEA Opposition to Priority on RCTs Privileging RCTs: Back to the Dark Ages Priority Manifests Fundamental Misunderstandings Causality and Evaluation AEA Members Opposition to AEA Statement Lack of Input from Key AEA Members Unjustified, Speciously Argued, Does Not Represent Norms or Many AEA Members Views AEA Compared to the Flat Earth Society

Diverse Prescriptive Theories of Evaluation Practice Social experimentation Science of valuing Results oriented management Utilization focused evaluation Empowerment evaluation Realist evaluation Theory-driven evaluation Inclusive evaluation Fourth generation evaluation

RCTs Not Practical/Feasible Often Impossible to Implement Well Not Cost Effective Very Limited Range of Applications Chapter Authors Provide Evidence to the Contrary

RCT Ethical Issues Unethical to Withhold Treatment from Control Groups Why Evaluate if Treatment is Better? Delay Treatment Non Evidence-Based Programs are Unethical

Methodological Challenges Zero Blind vs. Double Blind Experimenter Effects Allegiance Effects Unmasked Assignment Misguided Arguments about Causality External Validity Concerns Chapter Authors Claim Recent Methodological Developments to Overcome Some Challenges Noted in the Past

Political Concerns The RCT Gang has hijacked the term evidence-based for political and financial gain Evidence and especially scientific or rigorous evidence have become code for RCTs Focusing evaluation around these particular ideas about scientific evidence, allows social inquiry to become a tool for institutional control and to advance policy in particular directions

Political Concerns It is epistemological politics, not the relative merits of RCTs, that underlie federal directives on methodology choice The demand for evidence advances a master epistemology. The very dangerous claim is that a single epistemology governs all science Privileging the interests of the elite in evaluation is radically undemocratic

Ideological Differences: Paradigm Wars The positivist can t believe their luck, they ve lost all the arguments of the last 30 years and they ve still won the war! The world view underlying the current demand for evidence is generously speaking a form of conservative post-positivism, but in many ways is more like a kind of neo-positivism.

Ideological Differences: Paradigm Wars Many of us thought we d seen the last of this obsolete way of thinking about the causes and meanings of human activity, as it was a consensual casualty of the great quantitativequalitative debate in the latter part of the 20th century. Human action is not like activity in the physical world. Social knowledge is interpreted, contextual, dynamic or even transient, social or communal, and quite complicated. Privilege and honor complexity.

Ideological Differences: Paradigm Wars Evidence-based evaluation concentrates evaluation resources around one small question, does the program work?, and uses but one methodology, despite a considerable richness of options. The result is but one small answer. So what kind of evidence is needed? Not evidence that claims purchase on the truth with but a small answer to a small question, neat and tidy as it may be.

So What Kind of Evidence is Needed? (Greene, 2009) Evidence: that provides a window into the messy complexity of human experience that accounts for history, culture, and context that respects differences in perspective and values about experience in addition to consequences about the responsibilities of government not just responsibilities of its citizens with the potential for democratic inclusion and legitimization of multiple voices - evidence not as proof but as inkling

Changing the terms of the debate Melvin Mark

An attempt to Change the terms of the debate Inputs: Claremont symposium, resulting chapters. Other writings, interactions, etc. Mark s contention: Divergent methods positions rest on differing assumptions Focus on underlying assumptions may lead to more productive debate

Disagreement 1: What s the preferred evaluation question? And evaluation use? (1) Average effect size. For use in program/policy choice. (2) Other. Understanding lived experience. Or complexity. Or. For other uses. Each bolstered by democratic rationale

Alternative debate topics Value of estimating the effect of a given program? Value, relative to addressing other questions? Who decides the above, and how? If program s average effects are of interest, what ancillary methods are needed?

Gold Standards in context Unfortunately, too many people like to do their statistical work [or their evaluation/applied research planning] as they say their prayers merely substitute in a formula found in a highly respected book written a long time ago. Hotelling et al. (1948)

Wide Range of Views about Credible Evidence

CDC Evaluation Framework: 6 Key Steps + 4 Standards 39

CDC: Gathering Credible Evidence Definition: Compiling information that stakeholders perceive as trustworthy and relevant for answering their questions. Such evidence can be experimental or observational, qualitative or quantitative, or it can include a mixture of methods. Adequate data might be available and easily accessed, or it might need to be defined and new data collected. Whether a body of evidence is credible to stakeholders might depend on such factors as how the questions were posed, sources of information, conditions of data collection, reliability of measurement, validity of interpretations, and quality control procedures.

So What Counts as Credible Evidence? It depends: Question(s) of Interest The Context Assumptions of Evaluators & Stakeholders Theory of Practice Practical, Time, & Resource Constraints

Some Guiding Principles for Evaluation Practice Ongoing Discussion of Stakeholder Expectations Secure Buy-in to the Evaluation Design Before Revealing Results Be Aware of Potential Standards of Judgment Be Prepared for Meta-Evaluation Credible Evidence is Key for Influence & Positive Change

From Experimenting Society to Evidence-based Global Society? From RCTs as the Gold Standard to Methodological Appropriateness