Design and Analysis Plan Quantitative Synthesis of Federally-Funded Teen Pregnancy Prevention Programs HHS Contract #HHSP I 5/2/2016

Similar documents
Cochrane Pregnancy and Childbirth Group Methodological Guidelines

Meta-Analysis. Zifei Liu. Biological and Agricultural Engineering

Fixed Effect Combining

Introductory: Coding

Instrument for the assessment of systematic reviews and meta-analysis

META-ANALYSIS OF DEPENDENT EFFECT SIZES: A REVIEW AND CONSOLIDATION OF METHODS

How to Conduct a Meta-Analysis

Programme Name: Climate Schools: Alcohol and drug education courses

The Meta on Meta-Analysis. Presented by Endia J. Lindo, Ph.D. University of North Texas

Meta-Analysis De-Mystified: A Step-by-Step Workshop

Meta Analysis. David R Urbach MD MSc Outcomes Research Course December 4, 2014

Evidence Summary Title: Abstinence-plus HIV prevention programs in high-income countries: Evidence and implications for public health

Southern Nevada Health District Teen Pregnancy Prevention Program

Co-Variation in Sexual and Non-Sexual Risk Behaviors Over Time Among U.S. High School Students:

Reproductive Health s Knowledge, Attitudes, and Practices A European Youth Study Protocol October 13, 2009

Received: 14 April 2016, Accepted: 28 October 2016 Published online 1 December 2016 in Wiley Online Library

HIV/AIDS KNOWLEDGE SERIES NO. 46 APRIL 1996

Technical Specifications

Meta-Analysis David Wilson, Ph.D. Upcoming Seminar: October 20-21, 2017, Philadelphia, Pennsylvania

PREGNANCY PREVENTION INTERVENTION IMPLEMENTATION REPORT

Addendum: Multiple Regression Analysis (DRAFT 8/2/07)

The moderating impact of temporal separation on the association between intention and physical activity: a meta-analysis

Steady Ready Go! teady Ready Go. Every day, young people aged years become infected with. Preventing HIV/AIDS in young people

10/21/2015 REAL TALK PROGRAM COLLABORATIVE PROGRAM PARTNERS

heterogeneity in meta-analysis stats methodologists meeting 3 September 2015

Epidemiologic Methods I & II Epidem 201AB Winter & Spring 2002

In this module I provide a few illustrations of options within lavaan for handling various situations.

APPENDIX C YEAR ONE EVALUATION REPORT SOUTHERN NEVADA HEALTH DISTRICT GRANT 1 TP1AH July 1, 2015 June 30, 2016

Running Head: HIV/AIDS EDUCATION AND SAFE SEX PRACTICES

Sexual Health and Pregnancy Prevention among Community College Students: Gaps in Knowledge and Barriers to Health Care Access

Controlled Trials. Spyros Kitsiou, PhD

Small-area estimation of mental illness prevalence for schools

Individual Participant Data (IPD) Meta-analysis of prediction modelling studies

Presentation to the Worcester Board of Health Matilde Castiel MD June 9, 2016

Version No. 7 Date: July Please send comments or suggestions on this glossary to

Defining Evidence-Based : Developing a Standard for Judging the Quality of Evidence for Program Effectiveness and Utility

Propensity Score Methods for Estimating Causality in the Absence of Random Assignment: Applications for Child Care Policy Research

INSTRUCTION BP SEXUAL HEALTH AND HIV/AIDS PREVENTION INSTRUCTION

Donna L. Coffman Joint Prevention Methodology Seminar

Does AIDS Treatment Stimulate Negative Behavioral Response? A Field Experiment in South Africa

Non-medical use of prescription drugs and sexual risk behaviors among depressed adolescents

GUIDELINE COMPARATORS & COMPARISONS:

A Short Primer on Power Calculations for Meta-analysis

Research Synthesis and meta-analysis: themes. Graham A. Colditz, MD, DrPH Method Tuuli, MD, MPH

D. Paul Moberg University of Wisconsin

Live WebEx meeting agenda

How do we combine two treatment arm trials with multiple arms trials in IPD metaanalysis? An Illustration with College Drinking Interventions

Re-Examining the Evidence for Comprehensive Sex Education in Schools

CORE ELEMENTS, KEY CHARACTERISTICS AND LOGIC MODEL

Meta-Analysis and Subgroups

Statistical Techniques. Meta-Stat provides a wealth of statistical tools to help you examine your data. Overview

Assessing the efficacy of mobile phone interventions using randomised controlled trials: issues and their solutions. Dr Emma Beard

University of Bristol - Explore Bristol Research

Index. Springer International Publishing Switzerland 2017 T.J. Cleophas, A.H. Zwinderman, Modern Meta-Analysis, DOI /

Meta-analysis: Advanced methods using the STATA software

PREVENTING PREGNANCY: TALKING ABOUT AND USING CONTRACEPTION

Weight Adjustment Methods using Multilevel Propensity Models and Random Forests

SUPPORT FOR SCHOOL- BASED SEXUALITY EDUCATION: NO LONGER JUST FOR HIGH SCHOOL STUDENTS. Michele J Moore, Elissa Barr, Tammie Johnson

Does it Matter How You Ask? Question Wording and Males' Reporting of Contraceptive Use at Last Sex

CDC Review and Dissemination of Evidence-based HIV Treatment Adherence Interventions

Biostatistics for Med Students. Lecture 1

The Impact of Relative Standards on the Propensity to Disclose. Alessandro Acquisti, Leslie K. John, George Loewenstein WEB APPENDIX

educational assessment and educational measurement

RE-EXAMINING THE EVIDENCE:

Modern Strategies to Handle Missing Data: A Showcase of Research on Foster Children

A COMPARISON OF IMPUTATION METHODS FOR MISSING DATA IN A MULTI-CENTER RANDOMIZED CLINICAL TRIAL: THE IMPACT STUDY

Introduction and Background

HIV/AIDS. National Survey of Teens on PUBLIC KNOWLEDGE AND ATTITUDES ABOUT HIV/AIDS

MEASURES OF ASSOCIATION AND REGRESSION

International technical guidance on sexuality education

PREVENTION STRATEGIES RELATED TO HIV/AIDS Narra Smith Cox, Ph.D., CHES

Services for Men at Publicly Funded Family Planning Agencies,

Objective: To describe a new approach to neighborhood effects studies based on residential mobility and demonstrate this approach in the context of

Gambling attitudes and misconceptions

Clinical Epidemiology for the uninitiated

How effective is comprehensive sexuality education in preventing HIV?

Manuscript Presentation: Writing up APIM Results

OHDSI Tutorial: Design and implementation of a comparative cohort study in observational healthcare data

KNOWLEDGE OF HIV/AIDS AND OTHER SEXUALLY

Annex A: Impact, Outcome and Coverage Indicators (including Glossary of Terms)

Supplementary Online Content

Prentice Hall Health (Pruitt et. al.) 2007 Correlated to: Maryland - Voluntary State Curriculum Health Education (High School)

Multi-level approaches to understanding and preventing obesity: analytical challenges and new directions

Dates to which data relate Cost and effectiveness data were collected between 1995 and The price year was 1998.

CDC s 16 Critical Sexual Health Education Topics, Florida Physical Education and Health Education Benchmarks and Physical Education Courses

Recent developments for combining evidence within evidence streams: bias-adjusted meta-analysis

Program implementation practices among the OAH Teen Pregnancy Prevention Program cohort: Case Study Findings

Adolescent Sexual Risk-Taking Behaviors and The Impact of Programs to Reduce It. Douglas Kirby, Ph.D., ETR Associates April, 2009

Selected Topics in Biostatistics Seminar Series. Missing Data. Sponsored by: Center For Clinical Investigation and Cleveland CTSC

Introduction to diagnostic accuracy meta-analysis. Yemisi Takwoingi October 2015

bivariate analysis: The statistical analysis of the relationship between two variables.

Advanced Handling of Missing Data

Statistical Considerations in Pilot Studies

Nettle, Andrews & Bateson: Food insecurity as a driver of obesity in humans: The insurance hypothesis

Problem solving therapy

Help! Statistics! Missing data. An introduction

MONTEREY COUNTY OFFICE OF EDUCATION. Instruction AR

Violence Prevention: Rethinking the Standard of Care for Family Planning

Types of Data. Systematic Reviews: Data Synthesis Professor Jodie Dodd 4/12/2014. Acknowledgements: Emily Bain Australasian Cochrane Centre

Transcription:

Design and Analysis Plan Quantitative Synthesis of Federally-Funded Teen Pregnancy Prevention Programs HHS Contract #HHSP233201500069I 5/2/2016 Overview The goal of the meta-analysis is to assess the effects of teenage pregnancy prevention (TPP) programs on sexual behavior outcomes, with specific emphasis on examining whether program effects vary according to various aspects of program design, program implementation, participant characteristics, and study methods. The meta-analysis will also examine whether program characteristics affect participant retention in TPP programs. A random-effects metaregression modeling framework will be used to address each of the project s research questions, where effect sizes indexing program effects will be regressed on a range of effect size moderators (i.e., study characteristics) related to program design, program implementation, participant characteristics, and study methods. In the event that multiple, statistically dependent effect sizes are available from each study, the meta-regression models will be fitted using robust variance estimates. Finally, individual participant data will be used (again, under a randomeffects meta-regression modeling framework) to provide more refined examination of how participant characteristics are associated with program effects. Below, we provide more detailed description of the proposed analytic strategies. Statistical Procedures Effect size metric We will compute effect sizes for all reported outcome measures in three domains: sexual activity (e.g., abstinence, unprotected sex, number of sexual partners, number of sexual experiences), sexually transmitted infections (e.g., any STI, number of STIs), and pregnancy/births (e.g., any pregnancy, number of pregnancies, any births). All analyses will be conducted separately for these three outcome domains. 1

Standardized mean difference effect sizes will be used as the effect size metric (Lipsey & Wilson, 2001). All effect sizes will be coded so that positive effect sizes represent better outcomes (e.g., more abstinence, less unprotected sex) for the group receiving the intervention. Standardized mean difference effect sizes (d) are calculated as: where the numerator is the difference in group means for the intervention group ( ) and comparison group ( ), and the denominator is the pooled standard deviation for the intervention and comparison groups. All standardized mean difference effect sizes will be adjusted with the small-sample correction factor to provide unbiased estimates of effect size (g) (Hedges, 1981). The small-sample corrected effect size g and its standard error are calculated as: where N is the total sample size for the intervention and comparison groups, d is the original standardized mean difference effect size, n TX is the sample size for the intervention group, and n CT is the sample size for the comparison group. For binary outcomes, the Cox transformation outlined by Sánchez-Meca and colleagues (2003) will be used to convert log odds ratio effect sizes into standardized mean difference effect sizes: where A is the count of successes in the intervention group (e.g., number of participants who were abstinent), B is the count of failures in the intervention group (e.g., number of participants who were not abstinent), C is the count of successes in the comparison group, and D is the count of failures in the comparison group. Because the Cox transformation assumes 2

the binary outcome measures are from an underlying continuous distribution, all analytic models will include (i.e., adjust for) a dummy variable indicating whether the effect size was calculated via the Cox transformation for dichotomous outcomes. If sensitivity analysis indicates this variable has a substantial impact on model results, alternative strategies will be used to explore whether separate analyses should be conducted on the subset of binary outcomes using log odds ratio effect sizes. In the event that most calculated effect sizes are for binary outcomes, and only a small subset of standardized mean difference effect sizes emerge for continuous outcomes, we will instead use the Cox transformation to convert all standardized mean difference effect sizes into a log odds ratio effect size: Again, all analytic models will include (i.e., adjust for) a dummy variable indicating whether the effect size was calculated via the Cox transformation for continuous outcomes. Again, if sensitivity analysis indicates this variable has a substantial impact on model results, we will explore alternative strategies that separate all analyses for the continuous and binary outcomes. Effect size moderators As detailed in the coding manual for the meta-analysis, a wide range of study characteristics will be coded from the original study reports for descriptive purposes and for use as covariate controls in the final meta-regression models. However, the key effect size moderators of interest will be those related to the program design, program implementation, participant characteristics, and study methods outlined in the project research questions. Each of the research questions for the project is associated with a block of meta-regression models, which will be described in greater detail below. The first research question focuses on whether program design affects the impact of TPP programs. We will examine six key moderators related to program design: program focus, program components, program teaching/communication strategies, frequency of program contact, duration of the program, and group composition. 3

Program focus will be measured with a series of binary dummy variables indicating whether the primary focus of the program was: abstinence, sexual health, youth development, HIV/AIDS prevention, or reproductive health services. Program components will be measured with a series of binary dummy variables indicating whether the program included any of the following components: condom demonstration, service learning, role-plays, games, reflective exercises, mentoring/tutoring, individualized counseling, direct provision of reproductive health or other health services, parent activities, and community outreach. Program strategies will be measured with a series of binary dummy variables indicating the standard format of the program delivery: individual, small group (<10) with provider, large group or whole classrooms with provider, online, or other. Frequency of contact will be measured with a series of binary dummy variables indicating whether the program is delivered daily, 3-4 times per week, 1-2 times per week, less than weekly, or one day only. Contact hours will be measured with a continuous variable indicating the intended length/intensity of the intervention as measured in total hours of contact time. Program duration will be measured with a continuous variable indicating the number of weeks from first to last contact with program participants. Group composition will be measured with a binary dummy variable indicating whether the program was delivered to same-sex groups versus mixed- sex groups. The second research question focuses on whether program implementation affects the impact of TPP programs. We will examine two key moderators related to program implementation: program setting and program delivery personnel. If sufficient data can be collected on the level of preparation or training required for program staff, this research question will be revised to include an indicator of personnel preparation/training as a third moderator. Program setting will be measured with a series of binary dummy variables indicating whether the intervention was typically delivered in: classrooms, health clinics, community, or other. Program delivery personnel will be measured with a series of binary dummy variables indicating whether the intervention was typically delivered by: medical professionals, 4

health educators, classroom teachers, peer educators, other, or mixed (no predominant provider type). The third research question focuses on whether participant characteristics affect the impact of TPP programs. We will examine four key moderators related to participants: sex, race/ethnicity, age, and prior sexual experience. Participant sex will be measured with a continuous variable indicating the proportion of males present in the intervention group. Participant race/ethnicity will be measured with several continuous variables indicating the proportion of White, Black, and Hispanic participants present in the intervention group. Participant age will be measured with a continuous variable indicating the average age of participants in the intervention group. Participant prior sexual experience will be measured with a continuous variable indicating the proportion of participants in the intervention group who reported ever having had sex at baseline. The fourth research question focuses on whether study design, methods, and procedures affect the impact of TPP programs. We will examine six key moderators related to study methods: study design, overall attrition, differential attrition, risk of bias due to sequence generation, intent-to-treat analysis, and handling of missing data. Study design will be measured with a series of binary dummy variables indicating the study design: individual level random assignment, cluster level random assignment, quasi-random assignment, non-random assignment with matching, or non-random assignment with statistical controls. These categories may ultimately be collapsed into randomized and quasi-experimental design categories. Overall attrition will be measured with a continuous variable indicating the overall attrition rate in the sample as a whole. Differential attrition will be measured with a continuous variable indicating the differential attrition rate between the intervention and comparison conditions. 5

Sequence generation risk of bias will be measured with a nominal three-category measure of low risk of bias, high risk of bias, or unclear risk of bias, using the Cochrane collaboration s item on risk of selection bias due to inadequate generation of a randomized sequence. Intent-to-treat analysis will be measured with a binary dummy variable indicating whether the authors explicitly reported using an intent-to-treat analysis in their final outcome analyses. Missing data handling will be measured with a series of binary dummy variables indicating the methods that the authors used to handle missing data: listwise deletion, pairwise deletion, mean/mode imputation, single regression imputation, dummy variable imputation, multiple imputation, FIML, other method, not applicable, or cannot tell. Analytic Strategies All analyses will be conducted in a meta-regression framework with robust variance estimates (RVE), which can be used to synthesize statistically dependent effect sizes. Because we anticipate each study sample to provide multiple (dependent) effect size estimates (e.g., for different operationalizations of the same outcome, for multiple follow-ups), the RVE metaregression model can be used to synthesize all available effect sizes without loss of information. The RVE meta-regression is similar in form to traditional meta-regression, namely: where for i = 1 k j, j = 1 m, y ij is the ith effect size in the jth study; β 0 is the average population effect; u j is the study level random effect such that Var(u j ) = τ 2 is the between-study variance component; and e ij is the residual for the ith effect size in the jth study. In an unconditional metaregression model, the intercept estimates the traditional average effect size. The above regression equation includes main effects of study-level covariates but may also include multiplicative terms (interaction effects). For instance, we can use multiplicative interaction terms to examine whether different program types have larger or smaller effects on different types of outcomes within the same outcome domain (e.g., do abstinence-only programs have larger effects on abstinence relative to unprotected sex). 6

We will use a series of variables (described above) to measure the key intervention, participant, and methodological moderators of interest. In addition to these key moderators, the metaregression models may also include additional covariate controls to address any potential confounding effects. This might include, for instance, information about the strength of the counterfactual condition (e.g., the proportion of controls receiving information within the past year on relationships and dating, reproduction, abstinence, how to say no to sex, STDs, and birth control methods). Because the study samples will span the period of adolescence, the effect of age will be examined closely. Specifically, exploratory analyses will examine whether the effects of the other participant characteristics are moderated by age, and whether there is enough heterogeneity in effects to justify estimating separate models for different age categories. Differences associated with prior sexual experience will also receive attention to determine if separate models are needed for pregnancy prevention interventions that target youth with little or no sexual history and those that target youth with more sexual experience at baseline. For studies with multiple follow-up measurements, we will code the time from end of treatment as a predictor and use techniques that account for the statistical dependencies for effect sizes from the same sample. This allows for formal testing of the extent to which effects were sustained. Before conducting the final analysis, we will conduct an outlier analysis to identify any effect size and/or sample size outliers with the potential to distort the analysis; these will be Winsorized to the corresponding lower/upper fence values of their respective distributions and a sensitivity analysis will be conducted to ensure that such decisions do not have an inordinate effect on the findings. These outlier adjustments ensure that such studies will not exercise a highly disproportionate provided to evaluators, we anticipate that a small number of studies may have missing data on method, participant, or treatment variables used in the final analyses. In such cases we will contact the authors to see if they can provide the missing data; any remaining missing values will be imputed using an expectation-maximization (EM) algorithm. Consistent with standard meta-analysis approaches, we will give more weight to studies whose effect size estimates have greater precision, where precision is primarily a function of study sample size. In the proposed RVE meta-regression approach, the weights include a within-study 7

as well as a between-study component to the variance. The within-study component is the average variance across effect sizes within the study, and the between-study component is calculated using a method of moments estimator (Hedges, Tipton, & Johnson, 2010). Individual Participant Data (IPD) Meta-analysis If raw data on participant-level outcomes can be obtained from some or all evaluators, we will supplement our analyses of the aggregate data (AD) meta-analysis with individual participant data (IPD) meta-analysis approaches. Specifically, we will use a combination of IPD (as available) with AD meta-analysis to examine variability in effects across participant characteristics (i.e., sex, race/ethnicity, age, prior sexual history). Whereas the standard aggregate-data meta-analysis approach is useful for identifying whether study-level moderators (e.g., average age of sample) are associated with larger or smaller program effects, IPD metaanalysis can be useful for examining whether participant-level covariates (e.g., participant age) are associated with program effects. IPD meta-analysis can thus provide more detailed information about program effects for clinically relevant subgroups. If available, we will also examine program effects on other non-behavior outcome measures (e.g., attitudes, intentions). Because we anticipate that some evaluators will not provide IPD for the project, we anticipate having a mixture of IPD and AD available for synthesis. We will therefore use the one-step approach outlined by Riley et al. (2008) to synthesize findings with a combination of IPD and AD. This method uses a multilevel model to examine the effects of individual participant covariates on program effects, but includes a dummy variable to distinguish between IPD trials and AD trials. In this multi-level model, only the IPD trials contribute information to the analysis examining the effect of participant level moderators, but both the IPD and AD trials contribute information to the overall average program effect as well as the between-study variance component. 8