On the Use of Beta Coefficients in Meta-Analysis

Similar documents
Meta-Analysis of Correlation Coefficients: A Monte Carlo Comparison of Fixed- and Random-Effects Methods

A brief history of the Fail Safe Number in Applied Research. Moritz Heene. University of Graz, Austria

Statistical Techniques. Meta-Stat provides a wealth of statistical tools to help you examine your data. Overview

11/18/2013. Correlational Research. Correlational Designs. Why Use a Correlational Design? CORRELATIONAL RESEARCH STUDIES

Preliminary Report on Simple Statistical Tests (t-tests and bivariate correlations)

11/24/2017. Do not imply a cause-and-effect relationship

Unit 1 Exploring and Understanding Data

CHAPTER VI RESEARCH METHODOLOGY

Introduction to Meta-Analysis

Score Tests of Normality in Bivariate Probit Models

bivariate analysis: The statistical analysis of the relationship between two variables.

WELCOME! Lecture 11 Thommy Perlinger

Meta-Analysis and Subgroups

Investigating the robustness of the nonparametric Levene test with more than two groups

The moderating impact of temporal separation on the association between intention and physical activity: a meta-analysis

Standard Errors of Correlations Adjusted for Incidental Selection

Business Statistics Probability

Describe what is meant by a placebo Contrast the double-blind procedure with the single-blind procedure Review the structure for organizing a memo

EPSE 594: Meta-Analysis: Quantitative Research Synthesis

Profile Analysis. Intro and Assumptions Psy 524 Andrew Ainsworth

Meta-Analysis David Wilson, Ph.D. Upcoming Seminar: October 20-21, 2017, Philadelphia, Pennsylvania

CHAPTER 3 RESEARCH METHODOLOGY

Meta-Analysis. Zifei Liu. Biological and Agricultural Engineering

OLS Regression with Clustered Data

Measuring and Assessing Study Quality

Confidence Intervals On Subsets May Be Misleading

How to interpret results of metaanalysis

Identifying Peer Influence Effects in Observational Social Network Data: An Evaluation of Propensity Score Methods

Meta-Analysis and Publication Bias: How Well Does the FAT-PET-PEESE Procedure Work?

Section on Survey Research Methods JSM 2009

6. Unusual and Influential Data

12/31/2016. PSY 512: Advanced Statistics for Psychological and Behavioral Research 2

STATISTICS AND RESEARCH DESIGN

Still important ideas

Understanding Uncertainty in School League Tables*

Bias in regression coefficient estimates when assumptions for handling missing data are violated: a simulation study

Assessing publication bias in genetic association studies: evidence from a recent meta-analysis

Evidence-Based Medicine and Publication Bias Desmond Thompson Merck & Co.

Chapter 11: Advanced Remedial Measures. Weighted Least Squares (WLS)

Meta-Analysis: A Gentle Introduction to Research Synthesis

accuracy (see, e.g., Mislevy & Stocking, 1989; Qualls & Ansley, 1985; Yen, 1987). A general finding of this research is that MML and Bayesian

Adjustments for Rater Effects in

MEA DISCUSSION PAPERS

Conditional spectrum-based ground motion selection. Part II: Intensity-based assessments and evaluation of alternative target spectra

Readings: Textbook readings: OpenStax - Chapters 1 13 (emphasis on Chapter 12) Online readings: Appendix D, E & F

Daniel Boduszek University of Huddersfield

Manifestation Of Differences In Item-Level Characteristics In Scale-Level Measurement Invariance Tests Of Multi-Group Confirmatory Factor Analyses

Biology 345: Biometry Fall 2005 SONOMA STATE UNIVERSITY Lab Exercise 5 Residuals and multiple regression Introduction

Fixed Effect Combining

Mantel-Haenszel Procedures for Detecting Differential Item Functioning

META-ANALYSIS OF DEPENDENT EFFECT SIZES: A REVIEW AND CONSOLIDATION OF METHODS

Marno Verbeek Erasmus University, the Netherlands. Cons. Pros

Version No. 7 Date: July Please send comments or suggestions on this glossary to

Survey research (Lecture 1) Summary & Conclusion. Lecture 10 Survey Research & Design in Psychology James Neill, 2015 Creative Commons Attribution 4.

Survey research (Lecture 1)

INADEQUACIES OF SIGNIFICANCE TESTS IN

JSM Survey Research Methods Section

2.75: 84% 2.5: 80% 2.25: 78% 2: 74% 1.75: 70% 1.5: 66% 1.25: 64% 1.0: 60% 0.5: 50% 0.25: 25% 0: 0%

Cross-Cultural Meta-Analyses

Meta-analysis using HLM 1. Running head: META-ANALYSIS FOR SINGLE-CASE INTERVENTION DESIGNS

Evidence-Based Medicine Journal Club. A Primer in Statistics, Study Design, and Epidemiology. August, 2013

Catherine A. Welch 1*, Séverine Sabia 1,2, Eric Brunner 1, Mika Kivimäki 1 and Martin J. Shipley 1

A SAS Macro to Investigate Statistical Power in Meta-analysis Jin Liu, Fan Pan University of South Carolina Columbia

On the Performance of Maximum Likelihood Versus Means and Variance Adjusted Weighted Least Squares Estimation in CFA

Political Science 15, Winter 2014 Final Review

Minimizing Uncertainty in Property Casualty Loss Reserve Estimates Chris G. Gross, ACAS, MAAA

A Brief Guide to Writing

2. How do different moderators (in particular, modality and orientation) affect the results of psychosocial treatment?

Readings: Textbook readings: OpenStax - Chapters 1 11 Online readings: Appendix D, E & F Plous Chapters 10, 11, 12 and 14

1.4 - Linear Regression and MS Excel

Proof. Revised. Chapter 12 General and Specific Factors in Selection Modeling Introduction. Bengt Muthén

Describe what is meant by a placebo Contrast the double-blind procedure with the single-blind procedure Review the structure for organizing a memo

Small Group Presentations

CHAPTER TWO REGRESSION

Sawtooth Software. The Number of Levels Effect in Conjoint: Where Does It Come From and Can It Be Eliminated? RESEARCH PAPER SERIES

CHAPTER ONE CORRELATION

Checking the counterarguments confirms that publication bias contaminated studies relating social class and unethical behavior

Still important ideas

Individual Participant Data (IPD) Meta-analysis of prediction modelling studies

The Meta on Meta-Analysis. Presented by Endia J. Lindo, Ph.D. University of North Texas

Abstract Title Page Not included in page count. Title: Analyzing Empirical Evaluations of Non-experimental Methods in Field Settings

An Empirical Assessment of Bivariate Methods for Meta-analysis of Test Accuracy

Statistical data preparation: management of missing values and outliers

9 research designs likely for PSYC 2100

Edinburgh Research Explorer

1 Introduction. st0020. The Stata Journal (2002) 2, Number 3, pp

The prevention and handling of the missing data

Dr. Kelly Bradley Final Exam Summer {2 points} Name

Data and Statistics 101: Key Concepts in the Collection, Analysis, and Application of Child Welfare Data

Technical Whitepaper

A Comparison of Robust and Nonparametric Estimators Under the Simple Linear Regression Model

What is Meta-analysis? Why Meta-analysis of Single- Subject Experiments? Levels of Evidence. Current Status of the Debate. Steps in a Meta-analysis

Regression Discontinuity Analysis

Multiple Regression Analysis

Cochrane Pregnancy and Childbirth Group Methodological Guidelines

Exploring the Impact of Missing Data in Multiple Regression

Best-practice recommendations for estimating interaction effects using meta-analysis

Citation for published version (APA): Ebbes, P. (2004). Latent instrumental variables: a new approach to solve for endogeneity s.n.

Transcription:

Journal of Applied Psychology Copyright 2005 by the American Psychological Association 2005, Vol. 90, No. 1, 175 181 0021-9010/05/$12.00 DOI: 10.1037/0021-9010.90.1.175 On the Use of Beta Coefficients in Meta-Analysis Robert A. Peterson The University of Texas at Austin Steven P. Brown University of Houston This research reports an investigation of the use of standardized regression (beta) coefficients in meta-analyses that use correlation coefficients as the effect-size metric. The investigation consisted of analyzing more than 1,700 corresponding beta coefficients and correlation coefficients harvested from published studies. Results indicate that, under certain conditions, using knowledge of corresponding beta coefficients to impute missing correlations (effect sizes) generally produces relatively accurate and precise population effect-size estimates. Potential benefits from applying this knowledge include smaller sampling errors because of increased numbers of effect sizes and smaller nonsampling errors because of the inclusion of a broader array of research designs. When undertaking meta-analyses, researchers sometimes locate studies that report effect sizes for relevant relationships but only in a metric that seemingly cannot be integrated into the analysis. For example, a researcher using r, the Pearson correlation coefficient, as the effect-size metric in a meta-analysis might uncover relevant studies that report only standardized multiple regression coefficients (beta coefficients, beta weights, betas, or simply s). Despite journal editors increasing insistence that authors report correlation matrices, the absence of such matrices poses a data problem for researchers conducting meta-analyses. Although some researchers (e.g., Farley, Lehmann, & Sawyer, 1995; Raju, Fralicx, & Steinhaus,1986; Rosenthal & DiMatteo, 2001) have opined that beta coefficients can serve as effect-size metrics, the received view among meta-analysts (cf. Hunter & Schmidt, 1990) is that beta coefficients should not be used as surrogates for correlation coefficients in meta-analyses. The reason is simple and, on its face, compelling: A beta coefficient is a partial coefficient that reflects the influence of all predictor variables in a multiple regression model. Logic dictates that unless an effect-size metric reflects a simple bivariate or zero-order relationship between two variables, effect sizes cannot be meaningfully combined and aggregated across studies. Therefore, when only beta coefficients are reported, unless correlation coefficients or underlying data can be obtained, relationships reflected by the beta coefficients are generally not included in a meta-analysis. Typically, after an effort to contact authors to solicit missing correlations, a researcher would simply omit from the meta-analysis those studies for which simple pairwise correlations could not be obtained. It is possible, however, that omitting studies in this manner is unnecessary and may cause other problems that are more serious Robert A. Peterson, Department of Marketing, McCombs School of Business and IC 2 Institute, The University of Texas at Austin; Steven P. Brown, Department of Marketing and Entrepreneurship, Bauer College of Business, University of Houston. We appreciate insightful comments from Ed Blair on a previous version of this article. Correspondence concerning this article should be addressed to Robert A. Peterson, Department of Marketing, McCombs School of Business, 1 University Station, The University of Texas at Austin, Austin, TX 78712. E-mail: rap@mail.utexas.edu than the one the practice attempts to circumvent. Although it is indisputably true that beta coefficients are not technically comparable to correlation coefficients except under certain conditions, there is a paucity of evidence documenting the empirical relationship between corresponding beta coefficients and correlation coefficients. Given the discovery of such evidence, which is provided by the present research, it may be possible to derive a very close approximation to r on the basis of knowledge of the corresponding beta coefficient. This approximation is important because estimating population effect sizes and variances only on the basis of available effect sizes poses at least three potentially serious threats to inferences. First, omitting relevant effect sizes fails to make use of available information and ignores studies that may be essential to an accurate understanding of effect sizes and the conditions that generate them. Second, omitting relevant effect sizes increases sampling error, which decreases the precision of meta-analytic estimates of population parameters. Third, omission decreases the extent to which the set of effect sizes included in a meta-analysis collectively represents the universe of designs and contexts from which the population parameters would emerge. In other words, omitting relevant effect sizes potentially increases nonsampling error. The cumulative consequence of these omissions is to limit both the precision and generalizability of meta-analytic estimates of population effect sizes. Inasmuch as precision and generalizability of population effectsize estimates are key objectives in a meta-analysis, the consequences of omitting relevant effect sizes represent serious shortcomings. Therefore, it is worth considering whether and under what conditions beta coefficients might be used to impute correlation coefficients and what the effect of using them in this manner would be on meta-analytic effect-size estimates. Accordingly, the purpose of this research was to (a) estimate empirically the relationship between r and on the basis of a large sample of matched coefficients representing actual data, (b) identify study characteristics that might influence the magnitude of any difference between r and, (c) derive an approach to impute r given knowledge of, and (d) argue that use of this approach or an approximation to it provides a preferable alternative to using only existing data or using other common data imputation approaches (such as substituting zeros or the mean of observed rs for missing values). 175

176 RESEARCH REPORTS Missing Data in Meta-Analysis In view of the previously noted problems, Hedges (1992, p. 292) has termed missing data the most pervasive problem in large scale meta-analyses. Generally, pertinent effect sizes can be missing from a meta-analysis for three broad reasons. First, missing effect sizes can occur because a researcher has not exerted sufficient effort to identify and collect all relevant studies. This type of missing effect size can be addressed by carefully and systematically searching computer databases, building on citation patterns, and so forth (Cooper, 1989). Second, missing effect sizes can occur because certain research findings are never reported in the literature and consequently are not available from literature searches. For example, research findings may never be written up, may be written up but never submitted for publication, or may be written up and submitted for publication but rejected. This type of missing effect size is addressed in two distinct ways. One way consists of extensively and intensively contacting individuals or organizations that might possess relevant effect sizes (or the data to calculate them). A second way to address this type of missing effect size is to analytically estimate their extent and potential impact. Third, even though relevant research findings may exist in the literature, missing effect sizes can occur because they are not reported in a metric convertible to the one used to integrate findings across studies or because insufficient information is reported to permit their calculation. As previously mentioned, although the most common way to address this type of missing effect size is to attempt to acquire appropriate data from the reporting individual or organization, such attempts are often unproductive. This is the problem directly addressed in the present research, which derives and tests a means of imputing missing r values given knowledge of a beta coefficient for the same relationship. The result of doing so suggests that certain studies that do not report correlation coefficients need not constitute missing data, even absent recourse to the original data. The severity of the impact of missing effect sizes on the validity of meta-analytic conclusions depends on the extent to which the missing effect sizes differ systematically from those that are included in the analysis. Meta-analysts sometimes assume that data are missing because the results of some primary studies were not statistically significant and for that reason never published. If this were true, statistical integration of only available effect sizes would systematically overestimate population effect sizes. This is commonly referred to as the file drawer problem (Orwin, 1983; Rosenthal, 1979) and is typically addressed by estimating the number of studies with an effect size of zero that would be necessary to reduce a particular relationship to nonsignificance. Although Orwin s (1983) and Rosenthal s (1979) approaches to the file-drawer problem are widely used, they do not take into account the possibility that file-drawer studies may not simply be producing null results. When effect sizes are missing for reasons such as insufficient researcher effort or lack of a usable effect-size metric, the likelihood that the missing effect sizes will produce biased estimates of population effect sizes is less clear. For example, if the effect sizes not included in a meta-analysis are missing at random (such that the likelihood that any effect size is missing is unrelated to its magnitude or directionality), then the estimates derived from the analysis of available effect sizes will not be biased. Even in this case, however, omission of relevant effect sizes increases sampling error and potentially reduces the representativeness of the included studies and the generalizability of the estimates (Pigott, 1994). However, if omitted effect sizes differ systematically from those that are included in a meta-analysis, estimates of population effect sizes will be biased and the validity of inferences based on only the included effect sizes will be jeopardized. In this case, it is particularly important to minimize the number of missing effect sizes. In many cases, it may not be clear whether missing effect sizes are missing randomly or missing systematically. For example, when effect sizes are calibrated only in beta coefficients (i.e., correlation coefficients are not reported), it may sometimes be reasonable to assume that they are missing at random. The reason for the missing effect sizes may be the reporting standards of a journal, publication format (e.g., dissertation, conference paper), or literature in which the results appeared. It is not clear whether differences in standards, format, or literature would be related to the magnitude or directionality of effect sizes. It is plausible, however, that different journals or literatures could differ in the characteristics and effect sizes of studies that they publish. For example, if the same relationship figures in the literatures of two disciplines, the disciplines may differ in their paradigmatic approaches to studying it and consequently in the magnitudes of the effects that the approaches respectively produce. In this situation, the assumption that data are missing at random (and therefore that estimates of population effect sizes based on available effect sizes are unbiased) may not be tenable. Even when the missing-atrandom assumption holds, the previously noted adverse effects on sampling error and generalizability still apply. Imputing Missing Values The utility of any approach to impute missing effect sizes will be greatly enhanced if it is applicable when the missing-at-random assumption cannot be confidently assumed to hold. Imputing zeros for missing effect sizes on the assumption that nonsignificant effect sizes are more likely to be inaccessible is conservative but generally will likely underestimate population mean effect size and overestimate effect-size variance (Pigott, 1994). Substituting the mean estimated on observed or available effect sizes for missing effect sizes results in unbiased estimates when effect sizes are missing at random but not when they are missing systematically. This approach also underestimates the between-study variance in effect sizes because the variance component associated with any imputed effect size will equal zero (Little & Rubin, 1986; Pigott, 1994). Regression-based data imputation methods (e.g., Buck, 1960), although sometimes useful and effective when sample sizes are sufficient to provide stable estimates (Little & Rubin, 1986; Pigott, 1994), are often not applicable in meta-analysis because sample sizes in integrative studies are usually not large enough to produce stable estimates. The data imputation approach proposed here applies to the specific (albeit common) situation in which beta coefficients are reported for relationships relevant to a metaanalysis in some studies but correlation coefficients (the metric of the meta-analysis) are not. In this situation, the proposed data imputation approach is applicable when missing effect sizes differ systematically from available effect sizes, as well as when effect

RESEARCH REPORTS 177 sizes are missing at random. In either case, the proposed approach allows a meta-analyst to reduce sampling error and increase the representativeness of the effect sizes analyzed and the generality of inferences without sacrificing the precision of population effectsize estimates by including studies and effect sizes that otherwise would be excluded as unusable. Method Relationship Between and r The mathematical relationship between a corresponding and r in a multiple regression model can be easily shown. For a regression model with two predictor variables, and with no loss of generality (cf. Neter, Wasserman, & Kutner, 1990), 1 r y1 r 12 r y2 / 1 r 2 12, where the subscript y represents the dependent variable and the subscripts 1 and 2 represent the predictor variables. Then, by transposing and collecting terms, the following equation emerges: r y1 1 r 12 r y2 1 r 12. As limiting cases, if r 12 equals zero (i.e., there is no correlation between the two predictor variables), r y1 equals 1.Ifr 12 equals 1.0, r y1 equals r y2 ;if r 12 equals 1.0, r y1 equals r y2. Generally, however, the relationship between 1 and r y1 is a function of the magnitudes and signs of correlations involving the remaining predictor variable(s). Data The focus of the present research was on a comparison of corresponding beta coefficients and correlation coefficients. Although such a comparison could be made using synthetic data, synthetic data often do not capture the nuances and relationships that exist in actual data. Therefore, the present comparison was based on corresponding beta coefficients and correlation coefficients calculated on existing behavioral data. Obtaining these behavioral data required searching journals likely to contain articles reporting both s and rs for a particular study. Every issue of 10 behavioral journals and every other issue of 25 additional behavioral journals were searched for data from the period of 1975 2001 (a list of the journals reviewed and the studies providing data can be obtained from the corresponding author). To provide as broad a base of behavioral findings as possible (i.e., to obtain s and rs for the types of variables likely to be investigated in a meta-analysis), we sampled journals from disciplines including psychology (e.g., Journal of Applied Psychology), consumer behavior (e.g., Journal of Consumer Research), management (e.g., Academy of Management Journal), marketing (e.g., Journal of Marketing Research), and sociology (e.g., American Sociological Review). All articles in the sampled journals were examined for the desired s and rs. To qualify for inclusion in the database, the s and rs had to be calculated at the individual level. Moreover, s had to be calculated from single-equation regression models that were linear in their parameters. In addition to and r values, data were collected on sample sizes, number of predictor variables in the regression model, and whether predictor variable interactions were specified in the regression model. Further, to the extent possible, predictor-variable correlation matrices were recorded, and the average correlation among predictor variables was calculated as a surrogate indicator of collinearity. Results A total of 1,504 corresponding and r values were identified and harvested from the sampled literatures. All and r values identified were harvested, even those that might qualify as outliers or might reflect typographical errors (certain typographical errors were detected and corrected during the harvesting process). These values were obtained from 143 articles containing 160 data sets and 270 regression models. The median sample size of the data sets was 191; the median number of predictor variables in a regression model was six. Because the analysis consisted of a comparison of reported and r values, missing data due to researcher lack of effort and reporting biases were not deemed problematic. For example, if reporting biases existed, they were assumed present in both the and r values. Table 1 presents selected characteristics of the respective distributions of beta and correlation coefficients. When interpreting these characteristics, several things should be noted. First, all and r coefficients were rounded to two digits when entered into the database. Second, although values can technically be larger than 1.0 (Deegan, 1978), in the present instance the range was from.82 to.82. Thus, keeping with the spirit of meta-analysis, the two sets of coefficients were deemed numerically comparable. Third, both distributions were platykurtic ( p.01). Fourth, the overlap between the two distributions was approximately 95%, using Tilton s (1937) overlap statistic. Finally, comparing the present distribution of correlation coefficients with distributions of r-based effect sizes observed in several published meta-analyses (e.g., Rhoades & Eisenberger, 2002; Sheeran, Abraham, & Orbell, 1999) revealed that it was similar to those in the published meta-analyses. This correspondence inferentially supports the use of the present data in the context of r-based meta-analyses. Figure 1 contains a plot of the joint distribution of the 1,504 s and rs. Approximately 97% of the beta coefficients, 92% of the correlation coefficients, and 91% of the coefficients jointly were in the range of.50 units. There was no difference between and r in 6% of the comparisons; in an additional 17% of the comparisons, the difference was less than or equal to.02 units. Thus, in nearly one quarter of the comparisons, the difference between and r was within rounding error. The correlation between the s and rs was.84, producing an r 2 of approximately.70 and suggesting that the relationship is relatively linear. Hence, more than two thirds of the variance in corresponding s and rs is directly shared, with the remainder being due to the other predictor variables in the model, nonlinearity, and error. To explore the relationship between corresponding s and rs in more detail, we created two additional variables: (a) the absolute difference between and r and (b) the ratio r/, a measure of the relative difference between and r. The average absolute difference between corresponding beta coefficients and correlation co- Table 1 Characteristics of and r Distributions Characteristic distribution r distribution Range.82.82.86.82 Mean.06 a.09 b Median.05.09 Variance.05.07 Skewness 0.10 0.06 Kurtosis 1.33 0.27 a SD.21. b SD.26.

178 RESEARCH REPORTS Figure 1. Plot of correlation coefficients versus beta coefficients. efficients was.11; the median difference was.08. Approximately 56% of the absolute differences were less than or equal to.10. The distribution of absolute differences was platykurtic and somewhat positively skewed. The number of observations for the r/ ratio was 1,473 since values equal to zero did not permit calculation of the ratio. The average ratio was 1.23; the median ratio was 1.10. Fifteen percent of the ratios were negative, which means that the and r values possessed different signs. Of these 15%, more than half occurred when the and r values were between.10 and.10 and thus were possibly due to rounding and random errors. In general, and as might be expected given the manner in which coefficients are calculated, coefficients were slightly smaller on average than r coefficients. Moreover, the variance of the coefficients was smaller than the variance of the r coefficients. To investigate whether differences between corresponding and r values were systematically related to characteristics of the regression model, we correlated the absolute-difference variable and the ratio variable with sample size, number of predictor variables, a variable representing the ratio of sample size to number of predictor variables, whether there were interactions present, and the average r value of the predictor variable correlation matrix, respectively. The results are presented in Table 2. There was a statistically significant relationship between the average r value of the predictor-variable correlation matrix and the difference between corresponding s and rs. To the extent that the average r is a surrogate indicator of, or relates to, predictor variable collinearity, a significant relationship would be expected. This is because, in general, the more collinearity, the less stable the s, and the more likely the existence of differences between and r. Because only 51 of the 1,504 coefficient relationships reflected an interaction term, the significant correlation for the absolute difference variable must be interpreted with caution. There was not a statistically significant relationship between sample size or number of predictor variables and the absolute-difference or ratio variables. An additional analysis was undertaken to examine the robustness of the relationship between r and. This analysis consisted of dividing the 1,504 observations into four groups corresponding to the r quartiles. For each of the four groups, a linear regression model was estimated in which r was predicted by. Although the four slope coefficients differed in magnitude (Q1.68, Q2.13, Q3.22, Q4.67), as might be expected given the influence of extreme data values in the outer quartiles and the different combinations of r and values in the inner quartiles (i.e., values within a pair were more likely to possess different signs in the inner quartiles than in the outer quartiles), all four coefficients were statistically significant at p.02. Collectively, the results of these analyses speak to the robustness of the relationship between corresponding and r values. Table 2 Correlations Between Sample Characteristics and Relationships Between and r Sample characteristics r r/ Sample size.05.04 No. of predictors variable.04.00 Sample size/no. of predictors.03.04 Interactions present a.11*.04 Average predictor variable intercorrelations b.45*.09* a yes 1, no 0. * p.001. b n 1,119/1,092.

RESEARCH REPORTS 179 Imputing r Given Given the characteristics of the observed s and rs, it is possible to derive a formula for imputing an r value assuming knowledge of a corresponding value. Visual inspection of Figure 1 revealed a relatively tight joint distribution of and r values within the range from.50 to.50. Consequently, the data were restricted to values between.50 when deriving an imputation formula. A number of linear models using the coded characteristics of the regression models as independent variables were fit inductively to identify and select the best fitting model. From this process, the two-parameter least-squares equation of r.98.05 produced the best fit, where is an indicator variable that equals 1 when is nonnegative and 0 when is negative. The purpose of the indicator variable was to take into account the fact that nonnegative values tended to be relatively smaller than their corresponding r values compared with the magnitude of the difference between corresponding negative and r values. Thus, a value of.20 would produce an imputed r of.25, whereas a value of.20 would produce an imputed r of.20. The R 2 for this imputation formula was approximately.69, and the standard error of the estimate was.02. Given that there is no intercept term in the formula, it is not meaningful to compare the R 2 value with the variance common to and r. Because of Hunter and Schmidt s (1990) cautionary note about r being inflated when transformations are used, r and were evaluated directly when deriving the imputation formula. Testing Formula Efficacy Subsequent to deriving the imputation formula, 200 corresponding and r values ranging from.50.50 were obtained from studies published in 2002 and 2003 in the 10 journals that were completely searched previously and used to evaluate the formula s efficacy. As a first test of the formula s efficacy, it was used to impute the 200 rs from their corresponding s. The correlation between the 200 observed and imputed rs was.75, which suggests that the imputation formula is relatively robust. A second test of the formula s efficacy consisted of applying it under naturalistic conditions. In this test, the 200 r values were defined as a population of effect sizes whose mean (.079) was to be estimated in the presence of missing data. Three types of missing effect sizes were created: effect sizes missing at random, missing effect sizes that were systematically close to zero, and missing effect sizes that were systematically far from zero. To create these three types of missing data, we randomly selected samples of 10 and 20 r values representing 5% and 10% missing data, respectively, and deleted them from each of the following three sampling frames: (a) the entire population of 200 r values, (b) the 100 r values closest to zero, and (c) the 100 r values furthest from zero. Samples drawn from the latter two frames respectively reflected situations wherein missing r values were drawn from r values ranging from.12 to.12 (systematically close to zero ) or from r values less than.12 and greater than.12 (systematically far from zero ). Ten random samples of 10 and 20 missing r values were generated for each missing data type, and the population effect size ( ) was estimated using five different imputation approaches. The imputation approaches included (a) replacing missing r values with zeros, (b) replacing missing r values with the mean of the remaining 180 or 190 observed r values, (c) replacing missing r values using the imputation formula of r.98.05, (d) replacing missing r values with corresponding values, and (e) replacing missing r values using a convenience version of the imputation formula such that r.05, where the.98 weight is treated as a unitary weight and is defined as before (i.e., 1 when is nonnegative and 0 when is negative). Approaches (a) and (b) were used to provide benchmark data for comparison purposes. Table 3 presents the estimated mean value for each imputation approach and missing data type for the two missing data sample sizes. In addition to the mean rs, the mean absolute deviation (MAD) of the 10 sample replications from the population mean is reported for each imputation approach, missing data type, and missing data sample size. Because the standard errors of the estimated values derived from imputation approaches (c) (e) were virtually identical and were equal to the standard error of the population mean (SE.014), they are not reported in the table. Thus, all three imputation approaches reproduced the population standard error with a high degree of accuracy. The equivalence of the standard errors associated with imputation approaches (c) (e) and the population standard error is ascribed primarily to the fact that beta coefficients generally possess less variability than their corresponding correlation coefficients, in part because of partialling effects in the beta coefficients. Consequently, even though an effect-size imputation process might be expected to increase variation in the effect sizes to be analyzed, an imputation process based on s will generally (and perhaps counterintuitively) not produce standard error estimates that are inflated relative to the true standard error, especially if the percentage of missing effect sizes is relatively small (i.e., less than 10%). The average intercorrelation among the predictor variables was not used when deriving the imputation formula because this information would not be available for studies that do not report rs. However, in some well-developed subject areas, there may be prior knowledge of the average intercorrelation among predictor variables, or knowledge may be acquired during a meta-analysis from observed data. When either of these situations exists, it may be possible to use such knowledge when imputing rs from s. Following this logic, a three-parameter model was developed to predict r given,, and a parameter representing an indicator variable such that 0 if the average intercorrelation of the predictor variable set was.17 or less, and 1 if the average intercorrelation of the predictor variable set was.18 or greater (where.175 was the median value of the 1,119 observations on which the average intercorrelation of the predictor variable set was calculated). The model so produced yielded the imputation formula of r.99.04.02, with all parameters statistically significant at p.003 and R 2.69. Thus, a value of.20 with an average predictor variable set intercorrelation of.30 would produce an imputed r of.26, whereas a value of.20 with an average predictor variable set intercorrelation of.10 would produce an imputed r of.20. As before, the.99 parameter can be treated as unity when imputing r values with no practical loss in accuracy or precision. The bottom line of Table 3 contains the results of applying this three-parameter imputation formula. Table 3 reveals that imputing missing effect sizes by using zeros or observed means produced less accurate and less precise estimates than using beta coefficients directly or applying one of the

180 RESEARCH REPORTS Table 3 Estimates of (.079) Using Various Imputation Approaches Type of missing effect size (correlation) Imputation approach: Missing rs replaced by Randomly missing Systematically close to 0 (.12 r.12) Systematically far from 0 (r.12 or r.12) r MAD a r MAD r MAD Zeros 5% missing data.075.007.079.008.071.008 10% missing data.074.014.078.014.064.014 Mean of observed rs b 5% missing data.080.003.083.003.075.005 10% missing data.082.004.087.007.071.009 Formula rs c 5% missing data.080.001.081.002.080.003 10% missing data.082.003.083.003.078.002 Corresponding s d 5% missing data.079.001.079.001.075.005 10% missing data.079.012.080.001.075.005 Convenience s e 5% missing data.080.002.081.002.080.002 10% missing data.082.003.083.003.078.003 Formula betas with collinearity f 5% missing data.080.001.081.002.080.003 10% missing data.082.003.083.003.076.003 Note. MAD mean absolute deviation. a Mean absolute deviation of estimates. b Mean of 180 (10% missing data) or 190 (5% missing data) observed rs. c r i.98 i.05. d r i i. e r i i if i is negative; r i i.05 if nonnegative. f r i.99 i.04.02. imputation formulas, especially when effect sizes were systematically missing. Ceteris paribus, the performance of the convenience imputation formula, r.05, was equivalent to that of the other imputation approaches. Hence, in part because of its simplicity of application, its use is recommended. The three-parameter imputation formula did not perform better than the other imputation approaches; its use is not recommended. As would be expected, regardless of imputation approach, more accurate and precise results were obtained when missing data constituted 5% as opposed to 10% of the population. Thus, the results support the utility of the proposed imputation approach, even when it cannot be confidently assumed that missing r values are not systematically related to the magnitudes and directionalities of observed r values. Conclusion This research offers an approach for imputing r-based metaanalytic effect sizes in behavioral studies using knowledge of corresponding beta coefficients. Using the proposed imputation approach should produce (a) more accurate and more precise estimates of population effect sizes than either imputing zeros or observed means for missing effect sizes and (b) lower sampling error (because of increased numbers of effect sizes) and may potentially permit the incorporation of more representative study designs from which to generalize meta-analysis inferences. Thus, even if only coefficients are reported for a study, it is possible to use these coefficients to impute corresponding correlation coefficients if the coefficients reside in the interval.50. Because the relationship between corresponding s and rs appears to be both robust and independent of sample size and number of predictor variables, knowledge of coefficients would seem useful for a wide variety of -to-r imputation situations. Moreover, because effect-size metrics can be easily transformed (e.g., Fern & Monroe, 1996; Rosenthal, 1994), coefficients can be used to impute d-family metrics as well as r-family metrics. However, despite the potential usefulness of the proposed imputation approach, meta-analysts are still encouraged to make every effort to obtain original correlation coefficients. References Buck, S. F. (1960). A method for estimation of missing values in multivariate data suitable for use with an electronic computer. Journal of the Royal Statistical Society, Series B22, 302 303. Cooper, H. M. (1989). Integrating research: A guide for literature reviews (2nd ed.). Newbury Park, CA: Sage. Deegan, J., Jr. (1978). On the occurrence of standardized regression coefficients greater than one. Educational and Psychological Measurement, 38, 873 888. Farley, J. U., Lehmann, D. R., & Sawyer, A. (1995). Empirical marketing generalization using meta-analysis. Marketing Science, 14, G36 G46. Fern, E. F., & Monroe, K. B. (1996). Effect-size estimates: Issues and problems in interpretation. Journal of Consumer Research, 23, 89 105. Hedges, L. V. (1992). Meta-analysis. Journal of Educational Statistics, 17, 279 296.

RESEARCH REPORTS 181 Hunter, J. E., & Schmidt, F. L. (1990). Methods of meta-analysis: Correcting error and bias in research findings. Newbury Park, CA: Sage. Little, R. J. A., & Rubin, D. B. (1986). Statistical analysis with missing data. New York: Wiley. Neter, J., Wasserman, W., & Kutner, M. H. (1990). Applied linear statistical models (3rd ed.). Burr Ridge, IL: Irwin. Orwin, R. G. (1983). A fail-safe N for effect size in meta-analysis. Journal of Educational Statistics, 8, 157 159. Pigott, T. D. (1994). Methods for handling missing data in research synthesis. In H. Cooper & L. V. Hedges (Eds.), The handbook of research synthesis (pp. 163 178). New York: Russell Sage Foundation. Raju, N. S., Fralicx, R., & Steinhaus, S. D. (1986). Covariance and regression slope models for studying validity generalization. Applied Psychological Measurement, 10, 195 211. Rhoades, L., & Eisenberger, R. (2002). Perceived organizational support: A review of the literature. Journal of Applied Psychology, 87, 698 714. Rosenthal, R. (1979). The file drawer problem and tolerance for null results. Psychological Bulletin, 86, 638 641. Rosenthal, R. (1994). Parametric measures of effect size. In H. Cooper & L. V. Hedges (Eds.), The handbook of research synthesis (pp. 231 244). New York: Russell Sage Foundation. Rosenthal, R., & DiMatteo, M. R. (2001). Meta-analysis: Recent developments in quantitative methods for literature reviews. In S. T. Fiske, D. Schacter, & C. Zahn-Waxler (Eds.), Annual review of psychology (Vol. 52, pp. 59 82). Palo Alto, CA: Annual Reviews. Sheeran, P., Abraham, C., & Orbell, S. (1999). Psychosocial correlates of heterosexual condom use: A meta-analysis. Psychological Bulletin, 125, 90 132. Tilton, J. W. (1937). The measurement of overlapping. Journal of Educational Psychology, 28, 656 662. Received October 4, 2002 Revision received September 12, 2003 Accepted October 28, 2003