Mixed Methods Study Design

Similar documents
Appraising the Literature Overview of Study Designs

Overview of Study Designs in Clinical Research

Objectives. Distinguish between primary and secondary studies. Discuss ways to assess methodological quality. Review limitations of all studies.

Lecture 5 Conducting Interviews and Focus Groups

Clinical problems and choice of study designs

Research Design. Source: John W. Creswell RESEARCH DESIGN. Qualitative, Quantitative, and Mixed Methods Approaches Third Edition

QUALITATIVE RESEARCH. Not everything that counts can be counted, and not everything that can be counted counts (Albert Einstein)

Running head: Investigating the merit of discreetness in health care services targeting young men. Edvard K.N. Karlsen

Biostats Final Project Fall 2002 Dr. Chang Claire Pothier, Michael O'Connor, Carrie Longano, Jodi Zimmerman - CSU

A to Z OF RESEARCH METHODS AND TERMS APPLICABLE WITHIN SOCIAL SCIENCE RESEARCH

Epidemiological study design. Paul Pharoah Department of Public Health and Primary Care

04/12/2014. Research Methods in Psychology. Chapter 6: Independent Groups Designs. What is your ideas? Testing

In this second module, we will focus on features that distinguish quantitative and qualitative research projects.

Study Designs in Epidemiology

Insights and Innovation in Improving Health CRISP Seminar Series.

Analysis A step in the research process that involves describing and then making inferences based on a set of data.

In this chapter we discuss validity issues for quantitative research and for qualitative research.

Design of Qualitative Research

Rapid appraisal of the literature: Identifying study biases

positivist phenomenological

Handout 1: Introduction to the Research Process and Study Design STAT 335 Fall 2016

CONCEPTUALIZING A RESEARCH DESIGN

ISC- GRADE XI HUMANITIES ( ) PSYCHOLOGY. Chapter 2- Methods of Psychology

School of Population and Public Health SPPH 503 Epidemiologic methods II January to April 2019

Intervention vs. Observational Trials:

Qualitative Research Design

BIOSTATISTICS. Dr. Hamza Aduraidi

Chapter 02. Basic Research Methodology

Study Designs in Epidemiology

Evidence- and Value-based Solutions for Health Care Clinical Improvement Consults, Content Development, Training & Seminars, Tools

MULTIPLE CHOICE. Choose the one alternative that best completes the statement or answers the question.

Research Methodologies

Qualitative and quantitative research as complements (Black, 1999)

Lesson 3.1 What is Qualitative Research? Qualitative Research

Evaluation: Scientific Studies. Title Text

INTERPRETATION OF STUDY FINDINGS: PART I. Julie E. Buring, ScD Harvard School of Public Health Boston, MA

Article Critique - Use of Triangulation 1. Running Header: ARTICLE CRITIQUE USE OF TRIANGULATION FOR

Fitting the Method to the Question

Appraising Qualitative Research

Vocabulary. Bias. Blinding. Block. Cluster sample

Social Research (Complete) Agha Zohaib Khan

Version No. 7 Date: July Please send comments or suggestions on this glossary to

HPS301 Exam Notes- Contents

Lecture 4: Research Approaches

What Case Study means? Case Studies. Case Study in SE. Key Characteristics. Flexibility of Case Studies. Sources of Evidence

Live WebEx meeting agenda

Types of Biomedical Research

MODULE 3 APPRAISING EVIDENCE. Evidence-Informed Policy Making Training

Decision Making Process

Causation Assessment and Methodology

12/26/2013. Types of Biomedical Research. Clinical Research. 7Steps to do research INTRODUCTION & MEASUREMENT IN CLINICAL RESEARCH S T A T I S T I C

Qualitative Research. Prof Jan Nieuwenhuis. You can learn a lot just by watching

Study Design. Svetlana Yampolskaya, Ph.D. Summer 2013

Principles of Sociology

Nonresponse Rates and Nonresponse Bias In Household Surveys

Survey Methods in Relationship Research

Chapter Three: Sampling Methods

Tips for adding qualitative data collection to yoga research

Protocol Development: The Guiding Light of Any Clinical Study

UNDERSTANDING EPIDEMIOLOGICAL STUDIES. Csaba P Kovesdy, MD FASN Salem VA Medical Center, Salem VA University of Virginia, Charlottesville VA

Introduction. Activities undertaken 13/10/2015. Qualitative Data Analysis. CSHG Hotel Management School of Galicia

AP Psychology -- Chapter 02 Review Research Methods in Psychology

- Triangulation - Member checks - Peer review - Researcher identity statement

Research Approach & Design. Awatif Alam MBBS, Msc (Toronto),ABCM Professor Community Medicine Vice Provost Girls Section

UNIT I SAMPLING AND EXPERIMENTATION: PLANNING AND CONDUCTING A STUDY (Chapter 4)

The validity of inferences about the correlation (covariation) between treatment and outcome.

Study Design STUDY DESIGN CASE SERIES AND CROSS-SECTIONAL STUDY DESIGN

Choosing and Using Quantitative Research Methods and Tools

Fitting the Method to the Question

Gerene S. Bauldoff, PhD, RN, FAAN Professor of Clinical Nursing EBP Mentor CRITICAL APPRAISAL OF DESCRIPTIVE STUDIES

Educational Research

Chapter 8. Learning Objectives 9/10/2012. Research Principles and Evidence Based Practice

SOCIOLOGICAL RESEARCH PART I. If you've got the truth you can demonstrate it. Talking doesn't prove it. Robert A. Heinlein

CHECKLIST FOR EVALUATING A RESEARCH REPORT Provided by Dr. Blevins

The Regression-Discontinuity Design

Design of Experiments & Introduction to Research

Research Approaches Quantitative Approach. Research Methods vs Research Design

Chapter 1 A Cultural Approach to Child Development

Formative and Impact Evaluation. Formative Evaluation. Impact Evaluation

VALIDITY OF QUANTITATIVE RESEARCH

Glossary From Running Randomized Evaluations: A Practical Guide, by Rachel Glennerster and Kudzai Takavarasha

UNIVERSITY OF THESSALY. MSc Course PSYCHOLOGY OF SPORT & EXERCISE INFORMATION ABOUT THE MODULE

CHAPTER NINE DATA ANALYSIS / EVALUATING QUALITY (VALIDITY) OF BETWEEN GROUP EXPERIMENTS

Epidemiologic Methods I & II Epidem 201AB Winter & Spring 2002

INTRODUCTION. Evidence standards for justifiable evidence claims, June 2016

Online Supplementary Material

Survival Skills for Researchers. Study Design

PubH 7470: STATISTICS FOR TRANSLATIONAL & CLINICAL RESEARCH

Trial Designs. Professor Peter Cameron

26:010:557 / 26:620:557 Social Science Research Methods

Relationships Between the High Impact Indicators and Other Indicators

Internal Validity and Experimental Design

Evidence-Based Medicine Journal Club. A Primer in Statistics, Study Design, and Epidemiology. August, 2013

Other Designs. What Aids Our Causal Arguments? Time Designs Cohort. Time Designs Cross-Sectional

Analysis of the Reliability and Validity of an Edgenuity Algebra I Quiz

Chapter 11. Experimental Design: One-Way Independent Samples Design

RESEARCH METHODS. A Process of Inquiry. tm HarperCollinsPublishers ANTHONY M. GRAZIANO MICHAEL L RAULIN

Are the likely benefits worth the potential harms and costs? From McMaster EBCP Workshop/Duke University Medical Center

Evaluation: Controlled Experiments. Title Text

Evidence Based Medicine

Transcription:

1 Mixed Methods Study Design Kurt C. Stange, MD, PhD Professor of Family Medicine, Epidemiology & Biostatistics, Oncology and Sociology Case Western Reserve University 1. Approaches 1, 2 a. Qualitative first, then quantitative b. Quantitative first, then qualitative c. Concurrent 2. Quantitative designs, strengths & weaknesses 3 Deductive data collection and statistical analysis techniques, such as clinical trials or cohort studies, that seek to examine variables that can be measured numerically. Quantitative studies yield numerical data and typically a priori hypotheses, but may be hard to generalize. 3. Qualitative designs, strengths & weaknesses 4 Inductive data collection and analysis techniques, such as ethnography or focus groups, that seek to understand meaning and context. Qualitative studies yield rich, grounded data and new hypotheses, but may be hard to generalize. 4. Examples 5, 6 a. Qualitative first, then quantitative b. Quantitative first, then qualitative c. Concurrent 5. Developing a mixed methods team 7 6. Publishing mixed methods research 8 References 1. Stange KC, Zyzanski SJ. The integrated use of quantitative and qualitative research methods. Fam Med. 1989;21:448-451. 2. Stange KC, Miller WL, Crabtree BF, O'Connor PJ, Zyzanski SJ. Multimethod research: approaches for integrating qualitative and quantitative methods. J Gen Intern Med. 1994;9:278-282. 3. Hulley SB, Cummings SR, Browner WS, eds. Designing Clinical Research: An Epidemiological Approach, 2nd Edition. Baltimore: Lippincott Williams & Wilkins; 2001. 4. Crabtree BF, Miller WL, eds. Doing Qualitative Research. 2nd ed. Newbury Park, California: Sage Publications; 1999. 5. Borkan JM. Mixed methods studies: a foundation for primary care research. Ann Fam Med. 2004;2(1):4-6. 6. Creswell JW, Fetters MD, Ivankova NV. Designing a mixed methods study in primary care. Ann Fam Med. 2004;2:7-12. 7. The DOPC Writing Group. The Direct Observation of Primary Care Study: Insights from the process of conducting multimethod, transdisciplinary research in community family practice. J Fam Pract. 2001;50:345-352. 8. Stange KC, Crabtree BF. Publishing Mixed Methods Research. Ann Fam Med. 2006;4(4):(in press).

2 QUANTITATIVE STUDY TYPES Quantitative designs are specified at the outset, typically along with a priori hypotheses. Analyses usually involve descriptive and inferential statistics. A. Case reports & case series Detailed descriptions of an individual case or cases, without an explicit control group. May be useful for generating new hypotheses, or for describing unusual or new diseases. Particularly prone to selection bias and confounding. B. Ecologic studies Comparisons of factors, using group data (as opposed to data on individuals). Efficient use of available data for generating hypotheses or a "bottom line" assessment of strong relationships. Any link between exposure and outcome must be inferred, since individual data is not collected. May lead to "ecologic fallacy" of falsely attributing causation to confounding factors. C. Cross-sectional (prevalence) studies The relationship between an outcome and other factors of interest is compared, in a defined population, at one point in time. Efficient in that exposure and outcome data are collected at one time. The argument for a causal relationship is weakened by the lack of a temporal relationship between the predictor and outcome variables. Best design for determining prevalence. D. Case control studies Cases of an outcome and suitable controls are compared with respect to past exposures or other factors of interest. Efficient for a rare disease. The temporal relationship of exposure and outcome is an issue here, too. The control group must be representative of the population from which the cases are chosen. Measurement of exposure must be similar for cases and controls. Watch for ascertainment and recall biases. E. Cohort studies A defined group is followed over time for the development of an outcome thought to be related to a prior exposure. Typically, an exposed and an un-exposed group are then compared for their relative risk of disease. Can provide temporal evidence for a causal relationship, since exposure precedes the outcome. Costly for rare diseases, or diseases with a long latency. Attrition may cause bias. F. Clinical trials Eligible subjects are assigned randomly to receive either treatment or serve as a control, and are followed for the development of the outcome of interest. Most like an experiment. Strongest evidence for causality. Selection factors may decrease generalizability. Expensive. Ethical issues.

3 QUALTITATIVE STUDY TYPES Often qualitative designs involve multiple methods. The use of multiple data sources and multiple analysts is called "triangulation," and helps to assure the trustworthiness of the findings. Typically qualitative study designs evolve over time in an iterative fashion based on the initial findings. Analyses involve a combination of editing, templates or immersion/crystallization. A. Participant observation Immersion of the researcher in the area of study; fieldnotes based on observations. Provides an in-depth 3 rd party viewpoint. Delicate balance between immersion and aloof objectivity. Challenges of gaining entrée. Rapport vs. "going native." B. Interviews (key informant, depth, informal) Talking with informants. Often recorded for later analysis. Provides the perspective of those being studied. Selection of key informants can dramatically alter findings, as can relationships and hidden agendas. C. Focus groups Moderated discussion among homogeneous or heterogeneous groups of 4-12 people. Provides a shared perspective stimulated by prompts and other participants. Selection of participants and the group process affects findings. Seldom good for sensitive topics. D. Text analysis Analysis of printed or electronic materials. Useful if the needed information exists in this form. E. Participatory inquiry Involving the participants in forming the question, designing and conducting the study, interpreting and disseminating the findings. The process develops trust and helps to assure the relevance of the findings to participants. May be challenging to convey the learning outside of the group. F. (Comparative) case studies Integrating multiple data sources to provide an in-depth evaluation of a particular case. Comparative case studies then perform analyses across cases. Useful for generating insight into multifactorial problems or understanding the effects of complex interventions. Selection of cases is key. Challenging to identify over-arching lessons.

4 Integrative Studies A. Review Articles An author summarizes data and experience on a topic, and makes recommendations. Review articles or editorials often incorporate useful clinical insights from the author, but are particularly prone to the biases of the writer. This is particularly true for non-peer reviewed articles. B. Meta Analysis All published original research relevant to a specific question is examined using explicit criteria for quality and validity. Data from studies meeting inclusion criteria are summarized or pooled. This approach is less subject to the biases of the writer than the typical review article, and can be useful in increasing statistical power when multiple small studies have yielded conflicting results. Make sure that you agree with how the question is focused, the inclusion/exclusion criteria, and weighting schemes. C. Decision Analysis A technique for specifying alternate choices for managing a clinical problem and for quantitating and comparing the outcomes from different approaches. Decision analyses can help specify multistep clinical decisions, provide a bottom line recommendation, and demonstrate the sensitivity of the optimal choice of different assumptions. They are dependent on the logic and adequacy of specification of their decision trees, the quality of data upon which they are based, including information on the value or "utility" placed on various outcomes. D. Economic Analyses Cost effectiveness analyses compare the costs and health care outcomes of alternate decisions or health care policies cost benefit analyses go further, in that outcomes are valued in monetary terms. Like decision analyses, economic analyses are highly dependent on the logic and adequacy of specification of the costs and outcomes, the quality of the cost, outcome, and utility data upon which they are based, and the assumptions made in the analyses. E. Clinical Policies or Guidelines A technique for using explicit guidelines for gathering and interpreting available information to develop an approach to a particular clinical problem. Most clinical practice guidelines produce a general recommendation for most cases, along with options for tailoring the recommendation to fit individual circumstances. Guidelines that are developed using explicit rules for examining the scientific evidence for clinical decisions can provide a very useful synthesis of information, and can identify particular criteria for deviating from the policy in individual cases. Policies that are developed by a group of "experts" without explicit specification of their rules are subject to bias.

5 ASSESSMENT OF THE VALIDITY OF THE STUDY Examine potential sources of error and ask: How (in what direction) would the threats to validity affect the findings? Were potential sources of error adequately controlled in the design and analysis? How do validity concerns affect your interpretation or application of the data? POTENTIAL SOURCES OF ERROR A. CHANCE: Random error in sampling from a population. 1. Type I (alpha) error: The probability of falsely saying that there is a difference between two populations. 2. Type II (beta) error: The probability of falsely saying that there is no difference between two populations. B. BIAS: Systematic distortion 1. Selection bias - Systematic slant in how subjects are assembled or retained for study 2. Information bias - Systematic distortion from inaccuracy in measurement or classification of study variables C. CONFOUNDING: Misattribution of cause and effect Cause Effect Confounder REDUCING THE CHANCE OF ERROR In the Design 1. Study design appropriate for the question B,C 2. Adequate power A 3. Complete accounting of eligible patients B1 4. Verification of information B2 5. Blinding B2 6. Multiple control groups B,C 7. Randomization C 8. Restriction C 9. Matching C In the Analysis 1. Best case / worst case scenario B,C 2. Sensitivity analysis B,C 3. Adjustment B1,C 4. Mathematical modeling C A Reduces the role of chance; B1 Reduces selection bias ; B2 Reduces information bias; C Reduces confounding