APPLES TO ORANGES OR GALA VERSUS GOLDEN DELICIOUS? COMPARING DATA QUALITY OF NONPROBABILITY INTERNET SAMPLES TO LOW RESPONSE RATE PROBABILITY SAMPLES

Size: px
Start display at page:

Download "APPLES TO ORANGES OR GALA VERSUS GOLDEN DELICIOUS? COMPARING DATA QUALITY OF NONPROBABILITY INTERNET SAMPLES TO LOW RESPONSE RATE PROBABILITY SAMPLES"

Transcription

1 Public Opinion Quarterly, Vol. 81, Special Issue, 2017, pp APPLES TO ORANGES OR GALA VERSUS GOLDEN DELICIOUS? COMPARING DATA QUALITY OF NONPROBABILITY INTERNET SAMPLES TO LOW RESPONSE RATE PROBABILITY SAMPLES DAVID DUTWIN* TRENT D. BUSKIRK Abstract Nonprobability samples have gained mass popularity and use in many research circles, including market research and some political research. One justification for the use of nonprobability samples is that low response rate probability surveys have nothing significant to offer over and above a well built nonprobability sample. Utilizing an elemental approach, we compare a range of samples, weighting, and modeling procedures in an analysis that evaluates the estimated bias of various cross-tabulations of core demographics. Specifically, we compare a battery of bias related metrics for nonprobability panels, dualframe telephone samples, and a high-quality in-person sample. Results indicate that there is roughly a linear trend, with nonprobability samples attaining the greatest estimated bias, and the in-person sample, the lowest. Results also indicate that the bias estimates vary widely for the nonprobability samples compared to either the telephone or in-person samples, which themselves tend to have consistently smaller amounts of estimated bias. Specifically, both weighted and unweighted dual-frame telephone samples were found to have about half the estimated bias compared to analogous nonprobability samples. Advanced techniques such as propensity weighting and sample matching did not improve these measures, and in some cases made matters worse. Implications for fit for purpose in survey research are discussed given these findings. David Dutwin is executive vice president and chief methodologist for SSRS/Social Science Research Solutions, Media, PA, USA. Trent Buskirk is professor of management science and information systems (data science) and director of the Center for Survey Research at the University of Massachusetts Boston, Boston, MA, USA, and formerly the vice president of statistics and methodology at the Marketing Systems Group (MSG). The authors thank Jeffrey Bareham of MSG for his assistance with data management and analysis, and they thank the anonymous reviewers for their very helpful comments. *Address correspondence to David Dutwin, SSRS, 53 West Baltimore Pike, Media, PA 19063; ddutwin@ssrs.com. doi: /poq/nfw061 The Author Published by Oxford University Press on behalf of the American Association for Public Opinion Research. All rights reserved. For permissions, please journals.permissions@oup.com

2 214 Dutwin and Buskirk Telephone surveys are dead, claim some survey researchers, and should be replaced with nonprobability samples from Internet panels. That Internet panels are less costly is indisputable. At the heart of the question, however, is the resulting data quality from these types of panels. In particular, while the cost is lower, how do the errors from nonprobabilistic Internet panels and samples compare to the errors obtained using common probability-based samples of the day, that is, low response rate RDD samples? Certainly in a fit-for-purpose paradigm one must balance cost with error, but at some point even the cheapest means of data collection will have error properties that are not tolerable, regardless of the price. Probability-based samples are certainly not immune to cost and error considerations either. In recent years we have seen declining response rates for RDD samples, with many such surveys consistently reporting response rates below 10 percent (Pew Research 2012). Such changes have, in fact, led many to now claim that there is no real random sampling anymore (Gelman 2012). Of course, at question is not whether the samples are random (they are), but whether the actual respondents who are acquired from such samples differ in considerable and substantive ways from nonrespondents, and thus whether respondents can themselves still adequately represent the population from which they were sampled. While research of the prior decade finds that telephonic research has been quite resistant to bias stemming from non-response (Keeter et al. 2000; Groves 2006; Keeter et al. 2006; Groves and Peytcheva 2008), response rates for studies in that era were about twice as large as they are for similar studies conducted today. Of interest here is not just the question of the quality of low response rate telephone surveys but also whether and when nonprobabilistic samples, like those obtained through Internet panels, could serve as viable and reliable replacements or alternatives. While the literature on the comparative quality between nonprobability and probability samples is vibrant in European countries (see Callegaro et al. [2014] for an example), the research within the US context is just beginning to emerge. One of the earliest comparisons in the United States was by Malhotra & Krosnick (2007), who explored representation of the opt-in Internet surveys of YouGov and Harris to the American National Election Survey, a randomly sampled face-to-face interview, and the Current Population Survey. Walker and colleagues (2009) extended the comparisons between probability and nonprobability samples to also include comparisons across nonprobability samples obtained from several online panels on behalf of the Online Quality Research Council of the Advertising Research Foundation. Krosnick also collaborated with Chang (2009) and extended the comparisons of nonprobability samples to an online probability sample as well as to an RDD telephone sample. Yeager et al. (2011) expanded the number of nonprobability samples to seven and compared these to both an online probability sample as well as an RDD telephone sample. In all of these studies, a general pattern has been consistently reported namely, that more often than not, estimates derived from

3 Bias in Nonprobability and Probability Samples 215 probability samples were more accurate than those obtained from nonprobability samples across a wide array of political, attitudinal, factual, behavioral, and demographic outcomes. And while the differences in the accuracy in estimates derived using nonprobability and probability samples have been small in some instances, differences as large as 20 percentage points on measures like health behavior and technology use are also not uncommon (see Walker et al. [2009]). Importantly, nearly all of the research reported thus far employed fairly standard weighting procedures to limit the biases found in nonprobability samples, and in most cases the results between weighted and unweighted estimates were roughly equal. Indeed, a major effort looking at the impact of weighting (Tourangeau, Conrad, and Couper 2013) found that while standard iterative proportional fitting (raking) procedures sometimes reduce bias, and often significantly (by up to 60 percent), they also sometimes increase bias. Of great concern is the substantial impact of weighting in nonprobability samples, altering estimates by as much as 20 percentage points, far more than typically seen when similar weighting adjustments are applied to probability samples. Perhaps to improve upon standard weighting methods, the online survey research industry has developed additional methods for reducing biases in nonprobability samples, including propensity modeling and sample matching. Propensity modeling (Rosenbaum and Rubin 1983; Rosenbaum 1987) is a straightforward technique in which estimates are obtained from both a nonprobability sample as well as a probability sample. Typically, a logistic regression model is built to predict the likelihood of belonging to one sample versus the other, with independent variables including demographics and behavioral and attitudinal variables as well. Nonprobability samples can then be weighted using the inverse of the predicted probability derived from these propensity models, as described in Dever and Shook-Sa (2015). Schonlau et al. (2003) made an early effort at propensity modeling, with mixed success in that biases were reduced substantially in about 20 percent of the estimates while the remaining estimates still held substantial biases even after the adjustments were applied. Similar results were found by Duffy et al. (2005), where propensity adjustments were very helpful for vote intention and horserace questions but offered no marked improvement over standard raking adjustments for outcomes related to political participation, immigration policies, and knowledge and attitudes about cholesterol and technology use. The authors also note an additional concern, that, bias aside, propensity modeling can introduce significant variance into the estimates due to unequal weighting effects (UWEs) The unequal weighting effect (or the design effect due to unequal weighting [Kish 1992]) is the component of the design effect that is attributable to the variation in the sampling weights and is not dependent on any particular survey outcome.

4 216 Dutwin and Buskirk In addition to propensity modeling, sample matching is a technique sometimes used to reduce biases in nonprobability samples (Rivers 2007). Simply put, sample matching is a technique in which a probability sample is utilized as a gold standard and nonprobability online panelists are matched on a oneto-one (and in some cases many-to-one) basis to members of the probability sample based on a specified number of variables. Typically, these are demographic measures but can include behavioral or attitudinal metrics as well. Note that this procedure occurs before anyone in the online panel is invited to participate in a specific study, the notion here being to attempt to whittle the pool of nonprobability panelists to a set that match probability respondents on key matching variables. Unlike propensity weighting, sample matching is not an explicit weighting technique, but a method that essentially attempts to balance a nonprobability sample to a probability sample based on key target variables. Rivers and Bailey (2009) reported a great degree of success in using panel matching in estimating the 2008 election. Note, however, that vote intention and horserace have traditionally done well in web panels, using simple raking and propensity modeling, as noted above, and so at least in this respect the success of sample matching to accurately predict election outcomes is perhaps unsurprising. Recent submissions to both the AAPOR and Total Survey Error Conferences have included several papers experimenting with calibration (raking), propensity modeling, and sample matching techniques (Brick et al. 2015; Buskirk and Dutwin 2015a, 2015b; Dever and Shook-Sa 2015; DiSogra et al. 2015; Dutwin and Buskirk 2015; Peters, Driscoll, and Saavedra 2015; ZuWallack et al. 2015). Without reviewing each of these contributions specifically, it is notable that nearly all uniformly found propensity modeling and sample matching were not clear panaceas for reducing error in nonprobability samples, and due to the many inconclusive findings, most authors pointed to the need for further research and development of these techniques. Despite the emerging literature that consistently documents larger errors for nonprobability samples relative to probability samples and varying levels of success or failure with adjustment methods applied to nonprobability samples, some claim that the persistent low response rates attained from phone probability sample surveys render such probabilistic approaches almost dead (Hopper 2014). As a possible replacement, many have touted that online [optin] surveys work really well (Hopper 2014), and such nonprobability samples are capable of generating quality data that mirrors census representative data (Peanut Labs website advertisement on their AAPOR presentation [Petit 2015]). But empirical investigation and support for these claims appears to be relatively absent from the research literature. While there have been a growing number of studies that compare nonprobability and probability samples on their relative merits and data quality (e.g., Callegaro et al. [2014]; Yeager et al. [2011]), there are no research articles, to date, that have explicitly made a point to compare low response rate (which we define as below 10 percent)

5 Bias in Nonprobability and Probability Samples 217 probability samples to nonprobability samples in terms of basic data quality. Moreover, there is relatively little literature directly comparing propensity weighting and sample matching as possible solutions for improving the quality of nonprobability samples within the context of the United States. Responding to the tumultuous change faced by the industry in the past decade, and in part to the general dearth of research on the relative merits of low response rate probability and nonprobabilistic samples and Internet panels, AAPOR organized a task force and mini-conference in 2015 concerning New Developments in Survey Research. Additionally, the call for this Special Issue of Public Opinion Quarterly was for assessing the quality of research in the current environment along with a specific prerequisite to go beyond just an assessment, but to offer as well new methods of measuring representivity, that is, new metrics for evaluating the quality of estimates derived from different samples. This paper is a response to these calls and represents a substantive effort to assess the quality of a number of nonprobability Internet samples in direct comparison to low response rate probability samples using an elemental approach that utilizes the most fundamental survey data available: core demographics. We compare the estimated bias in conditional demographic distributions from low response rate RDD surveys with the estimated bias from nonprobability Internet panel surveys. We further add to the literature by exploring multiple methods for adjusting the nonprobability samples including propensity weighting, and sample matching in addition to raking. Data The data utilized for this study come from five principal sources, with data collection spanning two years, from October 2012 through October 2014, for all sources except the National Health Interview Survey (NHIS). Specifically, the survey data used in this paper were obtained from (1) the nonprobability Internet panel survey interviews from the Centris survey of communication, entertainment, and telephony (referred to as Panel 1) 2 ; (2) a national dualframe telephone RDD omnibus survey (Telephone 1) 3 ; (3) a general population, aged 18 54, dual-frame RDD telephone survey (Telephone 2) 4 ; (4) a companion nonprobability Internet panel sports-tracker survey (referred to as Panel 2) 5 ; and (5) the NHIS of More specific information about these studies and sample composition, target populations, and response rates is provided in table 1. The two nonprobability samples were obtained from two different Internet panels. There is little information on how respondents are specifically attained 2. Administered by Luth. 3. The SSRS omnibus. 4. SportsPoll. 5. Administered by Research Now.

6 218 Dutwin and Buskirk Table 1. Descriptions of the Five Samples, Including Sample Type, Scope, Response Rates (if applicable) and Types of Adjustments Applied to Each Sample Sample name Type Scope of sample Sample size Response rate (RR3) Methods of adjustment applied Telephone 1 a Probability Adults , % None (unweighted); raked Telephone 1: Cell phones Probability Adults , % None (unweighted); raked Telephone 2 b Probability Adults , % None (unweighted); raked National Health Interview Survey (NHIS) c Probability Adults , % None (unweighted); raked Panel 1 Nonprobability Adults ,797 N/A None (unweighted); raked Panel 1 matched Nonprobability Adults 18+ 5,110 N/A Matched; matched & raked Panel 2 Nonprobability Adults ,147 N/A None (unweighted); raked; propensity weighted; propensity weighted & raked Panel 2 matched Nonprobability Adults ,013 N/A Matched; matched & raked a The SSRS Telephone Omnibus survey is a national, dual-frame, bilingual telephone survey of 1,000 adults fielded weekly. While 40 percent of interviews were gathered via cell phones in 2012, the ratio increased to 50 percent in March One-third of Hispanic interviews are gathered in Spanish. The response rate for these two years of data collectively was 8.2 percent. b SportsPoll is a survey of sports, leisure, and other issues, and has run continuously since The telephone survey gathered 40 percent of its interviews via cell phones at the beginning of 2012 but increased to 50 percent in July The study is conducted in English and Spanish. Response rate for the survey was 8 percent. c The National Health Interview Survey is a multistage area probability survey on a range of health status and behavior. Fielded since 1957, it is used here as our high-quality survey for measures of bias and variance given its purpose to produce official governmental estimates on health and its high response rate, 81.7 percent in 2013.

7 Bias in Nonprobability and Probability Samples 219 by each panel, other than the usual notion of website banners, pop-ups, and other means. Because each study is a continuous tracking survey, panelists could in time be re-interviewed (no more frequently than once per year). As such, panelist interviews were de-duplicated so that a single respondent had only a single case of data. Deduplication was conducted at random to attain a single case. Only 14 percent of cases were duplicates and never more than twice, and an analysis of the full sample and the de-duplicated sample did not reveal any significant differences on a range of demographic measures. The 2013 US Current Population Survey (CPS) served as the probability sample source for the sample matching. We used the 2012 US Census Bureau s American Community Survey (ACS) as our gold standard source for population benchmarks in order to assess the estimated biases. 6 The ACS data also serve as the source for control totals for calibration via raking, which we describe in more detail in the next section. Methods While our study brings together several nonprobability samples and probability samples of varying sizes and scopes, demographic variables are common to all. To facilitate comparisons, we wanted to identify a key set of variables that are not likely to be subject to satisficing, social desirability biases, or other measurement errors that could confound the impact of sample type beyond mode which is inherent in the nature of this work (e.g., nonprobability Internet versus low response telephone surveys) (Callegaro et al. 2014). Instead, we explore core demographics as the basis for our analysis. Where one lives in the United States; how old they are; their race/ethnicity; and the level of education are attributes foundational to a vast array of other potential attitudinal and behavioral metrics. They have served as key covariates in myriad research projects across the social sciences. Raking, propensity models, and sample matching all widely employ these variables to reduce errors associated with nonresponse, coverage, and other sources. Of course, therein lies the rub in utilizing these measures to compare different samples. Raking to main effects of key demographics constrains the weighted distribution to match the marginal distribution of each of these key demographics (up to error tolerances), thereby eliminating any utility these variables have for sample comparisons. However, this method in no way constrains the individual cells within a cross-tabulation table nor the row-wise or column-wise conditional 6. As is typical in weighting, the benchmarks lag to principal year of the data collection since benchmarks are typically not available at the time of a survey s completion. However, given the longer time frame of publication, we were able as well to replicate the entire study utilizing 2013 benchmarks. Results are highly similar, with zero directional changes in results but some variance in the size of effects.

8 220 Dutwin and Buskirk distributions of two demographic variables considered together (e.g., the percentage of males who are young or the percentage of young adults who are male). Our elemental approach will look at evaluating conditional estimates computed from the cross-tabulation of pairs of demographic variables. The notion behind our elemental cross-tabulation approach is simple: to quantify the bias of the estimates of one demographic variable within each level of the second demographic variable or, more simply, to examine the conditional distribution of one demographic variable given a category of another. For example, if we look at the cross-tabulation of age group by race/ethnicity (e.g., non-hispanic White, non-hispanic Black, Hispanic, non-hispanic Other), then our key estimates are not the marginal distributions of age or race (which are fixed if calibration/raking is applied to the data using these demographics), but rather the distribution of age group within each of the four race groups and similarly, the distribution of race within each of the four age groups. This is akin to column versus row percentages in a cross-tabulated table of age group by race. In this way, we might expect that the bias from raking to known demographic main effect benchmarks would be minimized to the greatest extent possible by virtue of the strong association between our outcomes (conditional distributions of demographic variables) and these demographic variables. For our analyses, we consider four specific demographic variables: age group (18 34, 35 49, 50 64, 65+); race/ethnicity (non-hispanic White, non-hispanic Black, Hispanic, non-hispanic Other); education (Less than High School, High School, Some College, College or Beyond); and region (Northeast, South, Midwest, West). For each possible pair (A, B) of demographic variables, we will estimate conditional distributions using the crosstabulation table of demographic variable A (rows) and demographic variable B (columns). The column and row percentages computed from this table will be referred to collectively as A within B and B within A, respectively. Using the four demographic variables, there are 12 possible sets of estimates formed by the six cross-tabulations of the demographic variables that in turn result in a total of 216 target percentages that are evaluated from each sample source. 7 We evaluate the accuracy of these estimates using three different statistics, as described by Callegaro et al. (2014), including (a) the mean estimated absolute bias, (b) the standard deviation of the estimated absolute biases, and (c) the maximum estimated absolute error. Every metric is expressed in percentage points (pps). The estimated absolute bias for each of the row and column percentages from the cross-tabulation of demographic variable A with demographic variable B 7. If we consider the 4-by-4 cross tabulation of Region by Race, for example, there are four row percentages computed within each of the four levels of region that describe the distribution of Race within Region. Similarly, there are 16 total column percentages that describe the distribution of Region within Race. Thus, this 4-by-4 table alone contributes a total of 32 total comparisons.

9 Bias in Nonprobability and Probability Samples 221 are computed simply as the absolute value of the difference between the specific row/column percentage computed from the sample data and the corresponding percentage computed from the ACS data. Finally, the overall mean estimated absolute bias for a given sample represents the arithmetic mean of the average estimated absolute biases from the 12 possible sets of row and column percentages over all cross-tabulation tables. The standard deviation of the estimated absolute bias measure for a given collection of column percentages from A within B represents the standard deviation of the estimated absolute bias measures in these statistics contained within this cross-tabulation. Small values of the standard deviation metric for the collection A within B imply that the biases in the column percentages are generally consistent across the table, while large values of this metric imply that the biases in the column percentages are inconsistent in magnitude across the table. Finally, the maximum estimated absolute bias (also referred to as the largest absolute error [Callegaro et al. 2014]) is defined for each set of column percentages from A within B as the largest estimated absolute bias among this set of estimates. This metric is similarly defined for row percentages from B within A. The 216 target percentages were computed and evaluated from each of the six sample sources using both unweighted as well as weighted data. Overall, there are 18 sample/weighting combinations. First, there are the six unweighted samples: Panel 1, Panel 2, Telephone 1, Telephone 1 limited to only cell phones, Telephone 2, and the NHIS. Next, there are two matched samples: Panel 1 matched, and Panel 2 matched. The first panel was matched to a simple random sample of 5,010 anonymized cases (e.g., approximately 3.5 percent sample) selected from the 2013 CPS, while the second panel was matched to a second, independently drawn simple random sample of 5,013 anonymized cases (3.5 percent sample) selected from the 2013 CPS. 8 To obtain the one-to-one matches, a simple matching coefficient index was used to identify the nonprobability sample case with the best match to each probability case based on eight categorical variables, including region, sex, age group, race, education, homeownership, marital status, and current employment status Panel 1 matched thus contains 5,010 matched cases from the larger Panel 1 sample, and similarly, Panel 2 contains 5,013 matched cases from the larger Panel 2 sample. In both cases, we chose the 3.5 percent simple random sample to use as the basis of the match in order to allow the ratio of available nonprobability to probability cases to be large enough to ensure an adequate number of possible matches obtained from the nonprobability samples. This decision was related to the common support conditions for matching across sample types (Rivers 2007; Rivers and Bailey 2009). 9. To determine a match for each probability case, a distance measure was computed between the probability case and all available nonprobability cases using the simple matching coefficient which simply computes the proportion of the eight variables for which two cases have different values. The matched sample case was then that nonprobability sample case with the smallest distance to the probability case. Ties were broken randomly and the matches were processed sequentially based on randomly reordering the probability sample.

10 222 Dutwin and Buskirk Finally, there is the propensity weighted sample. Specifically, in the case of Panel 2, additional webographic variables (see appendix) were available, allowing us to compute propensity models for being in the probability panel based on main effects of both demographic and webographic variables. Such variables are typically used to discriminate nonprobability and probability respondents (Schonlau, van Soest, and Kapteyn 2007). Propensity weights were computed as the inverse of the predicted probabilities derived from the propensity models and mark the ninth possible combination of weighting and sample: Panel 2 propensity weighted. Since there were no webographic questions available, Panel 1 was not propensity weighted. 10 Each of the nine samples was then weighted by calibrating the base weights to identical targets, including education (less than high school, high school diploma, some college, or bachelor s degree, and graduate degree), gender, age (18 29, 30 49, 50 64, 65+), 11 and region, giving us a total of 18 sample types (nine unweighted versions and nine weighted/raked versions). For the nonprobability samples, the base weights were either 1 or the applicable propensity weight. The base sampling weights for all telephone datasets were computed using a single-frame estimator approach (Bankier 1986; Kalton and Anderson 1986; Buskirk and Best 2012) to account for multiple sampling frames and disproportionate probabilities of selection as functions of the number of persons and phones associated with a household. In addition to the four metrics evaluating the estimated biases in the estimates, we also measure the impact of various weighting/adjustment methods for each of the samples using a more traditional approach, that is, with measures of unequal weighting effects (UWEs). While generally defined for probability designs, we apply a simple computation of the UWE (as 1 plus the squared coefficient of variation in the final weights) across all sample types as a means to consistently measure the impact of given adjustments on the effective sample sizes and speak to the question of cost. 12 Results For the first part of our analyses, we will focus on the three metrics across the 12 sets of row/column percentage estimates computed from the demographic 10. We did trim or stratify the propensity weight, since our goal was to create a propensity weight that would reduce as much bias as possible. 11. The tracking studies (Panel 2 and Telephone 2) focused on year-old adults, and as such truncated the age target variable at the third break. 12. For example, if a given sample/weighting approach is inexpensive to obtain, but has a very large UWE, then the true cost of data obtained from that sample is based not on the cost per interview, but the cost per effective interview, which is a combination of the raw cost of all interviews divided by the effective sample size.

11 Bias in Nonprobability and Probability Samples 223 cross-tabulations. In the second part of our analyses, we focus on comparing samples based on aggregated statistics. We start with an overview of the estimated mean absolute bias depicted in figure 1 for each of these 12 sets by all 18 sampling/weighting combinations. A number of patterns emerge in this graph. First, there are four sets of row/column percentages with estimated mean absolute biases in excess of 10 percent, all from the unweighted nonprobability panels. In general, there appears a clear bifurcation in the average estimated absolute biases in the unweighted samples. The nonprobability samples attain relatively high estimated absolute biases, on average, while the unweighted telephone samples and the NHIS attain low estimated bias. Among weighted Figure 1. Estimated Absolute Biases for each of the 12 Sets of Row/ Column Percentages by Sample/Weighting Type.

12 224 Dutwin and Buskirk or matched samples, the pattern appears to be similar, though of course with much lower overall levels of estimated absolute bias, on average, than is seen in the unweighted samples. The estimated mean absolute biases displayed in figure 1 are also provided along with the other two summary metrics, computed for each of the 12 sets of row/column percentage estimates, for the unweighted/matched and propensity weighted nonprobability samples in table 2; the raked/weighted nonprobability samples in table 3; the unweighted probability samples in table 4; and the weighted/raked probability samples in table 5. If we focus on the nonprobability samples, we see, in table 2, that the matched versions of each of the nonprobability samples offered improvements over the unweighted versions for a majority of the 12 sets of row/ column percentage estimates across the three metrics. The degree of improvement for the mean absolute bias metric varied a bit more across the two nonprobability samples (e.g., seven of 12 for Panel 1 and 11 of 12 for Panel 2). But, as we observe comparing table 2 with table 3, matching alone was generally not better than raking alone. In fact, for only two of the 12 sets was the estimated mean absolute bias improved for the Panel 1 matched sample compared to the Panel 1 raked sample (and for five of 12 for the comparable Panel 2 samples). While generally the matched and raked nonprobability samples fared better than matching alone for at least eight of 12 sets across the three metrics for the Panel 1 sample, the results were not as conclusive for Panel 2 (seven of 12 sets had improvements in estimated mean absolute bias and standard deviation, but only four of 12 for maximum estimated absolute bias). The propensity weighted Panel 2 sample did generally fare better than the unweighted sample on most of the row/column percentage estimate sets. However, more generally, the estimates from propensity weighting and raking were generally inferior in terms of the three metrics when compared to the other nonprobability samples/ methods. In general, a closer look across table 3 reveals that there is no single method (e.g., raking, matching + raking, propensity weighting + raking) that performs better consistently across the 12 row/column percentage sets on any of the three metrics for either of the two nonprobability samples. While some of the newer methods (e.g., propensity weighting and matching) offered some improvements for nonprobability samples in terms of absolute biases, these improved estimates generally had larger and more variable errors compared to those from the probability samples. In particular, a comparison of table 2 and table 4 reveals that the nonprobability samples had the largest maximum estimated absolute bias for 10 out of the 12 sets of row/column pps estimates and higher standard deviations of the estimated absolute biases for 11 out of 12 of these sets. A comparison between tables 3 and 5 shows that even after applying weighting/raking to the nonprobability samples (and their adjusted versions), these weighted samples still have the largest maximum estimated

13 Bias in Nonprobability and Probability Samples 225 Table 2. Statistical Metrics for Each of the 12 Sets of Row/Column Percentage Estimates for the Unweighted/Matched/ Propensity Weighted Nonprobability Samples ( = estimated absolute bias) Panel 1 unweighted Panel 2 unweighted Panel 1 matched Panel 2 matched Panel 2 propensity weighted Demographic cross-tabulation Std. dev s Std. dev s Std. dev s Std. dev s Std. dev s Race within education Age within education Region within education Race within age Education within age Region within age Age within race Region within race Education within race Race within region Age within region Education within region

14 226 Dutwin and Buskirk Table 3. Statistical Metrics for Each of the 12 Sets of Row/Column Percentage Estimates for the Unweighted Probability Samples ( = estimated absolute bias) Telephone 1 (cell) unweighted Telephone 1 unweighted Telephone 2 unweighted NHIS unweighted Demographic cross-tabulation Std. dev of s Std. dev of s Std. dev of s Std. dev of s Race within education Age within education Region within education Race within age Education within age Region within age Age within race Region within race Education within race Race within region Age within region Education within region

15 Bias in Nonprobability and Probability Samples 227 Table 4. Statistical Metrics for Each of the 12 Sets of Row/Column Percentage Estimates for the Weighted/Raked Nonprobability Samples ( = estimated absolute bias) Panel 1 raked Panel 2 raked Panel 1 matched & raked Panel 2 matched & raked Panel 2 propensity weighted & raked Demographic cross-tabulation Std. dev of s Std. dev of s Std. dev of s Std. dev of s Std. dev of s Race within education Age within education Region within education Race within age Education within age Region within age Age within race Region within race Education within race Race within region Age within region Education within region

16 228 Dutwin and Buskirk Table 5. Statistical Metrics for Each of the 12 Sets of Row/Column Percentage Estimates for the Weighted/Raked Probability Samples ( = estimated absolute bias) Telephone 1 (cell) raked Telephone 1 raked Telephone 2 raked NHIS weighted Demographic cross-tabulation Std. dev of s Std. dev of s Std. dev of s Std. dev of s Race within education Age within education Region within education Race within age Education within age Region within age Age within race Region within race Education within race Race within region Age within region Education within region

17 Bias in Nonprobability and Probability Samples 229 absolute bias statistics for 11 of the 12 sets of column/row percentage estimates compared to the weighted/raked probability samples. While weighting helped reduce the variability in the estimated absolute biases measured in each of the 12 sets, the weighted/raked nonprobability samples collectively had higher standard deviations for eight of the 12 sets of row/column pps estimates. To further simplify and clarify our analysis, we now focus on measures of overall estimated absolute bias and average standard deviation for each of the 18 sample/weighting combinations obtained by aggregating these metrics across the 12 row/column percentage tables computed for each sample. The overall mean estimated absolute bias for the unweighted and matched/ weighted samples are given in tables 6 and 7, respectively. In table 6, we see that the probability samples overall hold substantially less inherent (that is, unweighted) estimated absolute bias compared to the nonprobability samples. The mean overall average estimated absolute bias for the two web panels is 6.4 pps, compared to 3.5 pps for the dual frame samples, 3.4 pps for the cell-only sample, and 2.2 pps for the NHIS. Put another way, the nonprobability samples hold about 1.8 times more estimated absolute bias, on average, than the dual-frame samples. As measured by percent difference to the ACS, nonprobability samples were about 30 percent different in their estimates, compared to 16 percent different in the dual frame samples. Table 7 summarizes our weighted, modeled, and matched samples. Notably, Panel 1 s (unweighted) overall average estimated absolute bias is reduced from 5.9 to 2.1 percentage points after it is raked. Perhaps unsurprisingly, the NHIS had the smallest overall average estimated absolute bias of any of the samples and there was also no practical difference in the overall average Table 6. Statistical Metrics Averaged over the 12 Sets of Row/Column Percentages for Each of the Unweighted Samples ( = estimated absolute bias) Unweighted sample Overall average for row/ column percentages Overall median for row/ column percentages Overall average standard deviation of s Panel 1 unweighted Panel 2 unweighted Telephone 1 (cell) unweighted Telephone 1 unweighted Telephone unweighted NHIS unweighted

18 230 Dutwin and Buskirk Table 7. Statistical Metrics Averaged over the 12 Sets of Row/Column Percentages for each of the Matched/Weighted Samples Matched/weighted sample Overall average for row/ column percentages Overall median for row/column percentages Overall average standard deviation of s Panel 1 matched Panel 2 matched Panel 2 propensity weighted Panel 1 raked Panel 2 raked Panel 1 matched & raked Panel 2 matched & raked Panel 2 propensity weighted & raked Telephone 1 (cell) raked Telephone 1 raked Telephone 2 raked NHIS weighted estimated absolute bias between the cell-only sample and its corresponding dual-frame sample. Generally speaking, all samples have significantly reduced overall average estimated absolute bias when weighted and the reduction in estimated bias was roughly similar across samples, with some notable exceptions. The overall estimated absolute bias for the two nonprobability samples was, on average, reduced by a factor of 2.5, while the telephone samples saw an average reduction by a factor of 1.9 (and the NHIS by a factor of 2.8). The overall average estimated absolute bias of the two raked nonprobability samples was, on average, greater than that of the probability samples (NHIS excluded) (2.4 versus 1.7 percentage points) In terms of real percentages, there are also notable differences. As can often be the case in raking, cells that are underrepresented in raw data can sometimes become overcorrected once weighted. For example, while 12 percent of African Americans in Panel 1 do not have a high school diploma (ACS = 16 percent), once weighted this cell balloons to 21 percent. Comparatively, Telephone 1 rose from 13 percent unweighted to nail the benchmark at 16 percent, once weighted. This is not to say that there were no examples of the weighted telephone samples doing poorly. As might be expected, the telephone samples over-represent older Americans, and when weighted this can result in imprecision in the estimates of race/ethnicity. So, whereas 12 percent of Hispanics reported being age 65 and older in Telephone 1 (compared to 13 in the ACS), because raking suppresses all persons age 65 and older in the telephone samples, this weighted estimate drops to 6 percent.

19 Bias in Nonprobability and Probability Samples 231 Of course, our expectation for adjustment techniques such as propensity weighting and sample matching is that they would further reduce bias over and above what raking can achieve. If true, this could place the bias in the nonprobability samples on a par with telephone samples. In fact, however, these procedures, while apparently reducing estimated bias compared to the unweighted samples, did not reduce bias as efficaciously as with simple raking. Simple raking reduced the overall average estimated absolute bias of both nonprobability samples to 2.4 pps, while the propensity weight only reduced the overall mean estimated absolute bias to 3.9 pps; propensity weighting and raking to 2.7 pps; matching in both samples to 3.4 pps; and matching and raking in both samples to 2.5 pps. Notably, Panel 1 had the same level of estimated mean absolute bias once matched and raked compared to just the simple rake, although the size of the matched sample was considerably less. As well, the Panel 2 propensity weighted sample attained the same level of estimated bias as the Panel 1 sample matched, 3.9 pps. The overall average of the standard deviations of the estimated absolute biases is also reported in tables 6 and 7 for the unweighted, matched, and weighted samples, respectively. For the unweighted samples (table 6), we find a pattern similar to that of the estimated biases in that the nonprobability samples show substantially greater variability in the estimated absolute biases, on average, compared to the probability samples. In fact, taken together, the unweighted nonprobability panels attain an average standard deviation of 6.4 pps, compared to 3.5 pps for the probability samples (NHIS excluded). A closer look at table 7 reveals that this gap in variability shrinks with various weighting/matching methods. While it is still the case that the probability samples have less variability in estimated absolute biases, on average, compared to nonprobability samples, the differences between the two types of samples are less pronounced. Taken together, the various weighted and matched nonprobability samples attain an average standard deviation of 2.9 pps, compared to 1.7 pps for the probability samples (NHIS excluded). The largest average standard deviation in estimated absolute biases across the samples came from Panel 2 (unweighted). Finally, we observe that the variability in estimated absolute bias measures for the cell phone sample was on par with the full telephone samples and the NHIS had by far the lowest variability of all weighted samples (0.8 pps). As for the potential impact to individual estimates resulting from UWEs, the pattern is much the same as we have seen thus far. Specifically, while the UWEs for the telephone samples are relatively low (1.3), the raking had to, in effect, work harder for the nonprobability samples, with an UWE of 2.8 (see figure 2). We note that the UWEs for dual-frame samples accounts for multiple inclusion probabilities across two sampling frames as well as impacts from the raking. Worse still was any attempt to apply propensity weighting, which resulted in a UWE of 5.9. One advantage of sample matching, however, is that

20 232 Dutwin and Buskirk Figure 2. Unequal Weighting Effects. by stacking the deck before inviting people to respond to surveys, sample matched data requires far less adjustment with raking. As such, the average UWE across the two matched and raked samples was rather low, at 1.2. Discussion and Conclusions Anecdotally and empirically, researchers have found that estimates derived using samples from nonprobability panels can be less accurate than those derived using what we would in the present day consider high response rate probability samples. In our analysis, the unweighted samples from nonprobability Internet panels showed substantially higher estimated bias than did the low response rate probability samples. Beyond that, what is notable is the combination of 1) the overall size of the unweighted nonprobability estimated biases and the gap between probability and nonprobability samples on the estimated mean absolute bias; 2) the variability noted in the estimated absolute biases of the nonprobability samples; and 3) in comparison to probability samples, the substantial biases computed from the nonprobability samples persisted even after applying raking, propensity weighting, or matching. More specifically, the nonprobability samples in our analysis had mean estimated absolute bias measures that were about two times that of the low response rate probability samples, and, at 6.4 pps (unweighted), this average estimated absolute bias often translates into a substantive difference between the survey estimate and the corresponding benchmark value computed from the ACS. Not only were the estimated biases larger for the nonprobability samples, they were also more varied. The standard deviations of the estimated biases for statistics computed across the 12 cross-tabulations of the demographic variables were on average about two times larger for the nonprobability samples, taken together, as compared to the two dual-frame probability samples. The

21 Bias in Nonprobability and Probability Samples 233 magnitude of the estimated biases for a given row or column percentage in the cross-tabulation of the demographic variables for the nonprobability samples appears to fluctuate widely and no systematic pattern seems to emerge, with some variables exhibiting high estimated biases for one nonprobability sample but not for the other. In contrast, the estimated absolute biases computed from the probability samples appear to be consistently low, as evidenced by smaller mean estimated absolute biases and standard deviations. This may suggest that, with regard to nonprobability samples, one is rolling the dice in terms of how accurate any one particular estimate may be. Because the demographic variables in our analysis were used as main effects in all raking applications, our investigation of biases represents a best-case scenario for which raking would expect to work very well in favor of reducing biases. But, even in this best-case scenario, substantial biases persisted for some of the nonprobability samples. In this sense, we believe that our elemental approach illustrates a lower bound on the overall estimated biases one might expect to see in practice when working with any samples whose data are used to estimate more than just conditional demographic distributions. While attempting to focus on the best-case scenario in terms of bias in final estimates, this investigation has also exposed the deficits in each type of sample. The larger errors in the nonprobability samples we investigated were driven in large part by education as well as race, and third by age (too young), given that in terms of main effects these distributions are quite biased in the first place. Only age (too old) is similarly biased in the telephone samples. And, one can attain a basically self-weighting age distribution in dual-frame surveys with a larger percent (70 80 percent) cell phone and the remainder landline sample. The fact that the raked nonprobability samples seemingly reduced biases overall is no surprise given these deficits in nonprobability samples. Notably, the overall average estimated absolute bias in the nonprobability samples was reduced more than for telephone samples. However, despite this difference, nonprobability samples do not catch up to telephone samples once weighted. In exchange for the reduction in absolute bias, higher UWEs were seen for the raked nonprobability samples, even though the base weights were all one to begin with for all but the propensity weighted sample. In the end, the final average estimated absolute bias among nonprobability samples, even after adjustments, was 2.5 pps, which certainly might be good enough for a great deal of research, again noting both the concept of fit for purpose and the reality that nonprobability research is now valued in the billions of dollars (Lorenzetti 2014). But, in some contexts, even this amount of bias may not fall within acceptable limits. While the telephone samples are still biased at twice the rate of the NHIS, the variability of the estimated absolute mean biases for unweighted probability samples is quite similar between the telephone samples and the NHIS, with a relatively low overall average estimated absolute bias of about 1.5 pps. In short, there appears far less risk that

How Errors Cumulate: Two Examples

How Errors Cumulate: Two Examples How Errors Cumulate: Two Examples Roger Tourangeau, Westat Hansen Lecture October 11, 2018 Washington, DC Hansen s Contributions The Total Survey Error model has served as the paradigm for most methodological

More information

Sources of Comparability Between Probability Sample Estimates and Nonprobability Web Sample Estimates

Sources of Comparability Between Probability Sample Estimates and Nonprobability Web Sample Estimates Sources of Comparability Between Probability Sample Estimates and Nonprobability Web Sample Estimates William Riley 1, Ron D. Hays 2, Robert M. Kaplan 1, David Cella 3, 1 National Institutes of Health,

More information

Selection bias and models in nonprobability sampling

Selection bias and models in nonprobability sampling Selection bias and models in nonprobability sampling Andrew Mercer Senior Research Methodologist PhD Candidate, JPSM A BRIEF HISTORY OF BIAS IN NONPROBABILITY SAMPLES September 27, 2017 2 In 2004-2005

More information

Methodology for the VoicesDMV Survey

Methodology for the VoicesDMV Survey M E T R O P O L I T A N H O U S I N G A N D C O M M U N I T I E S P O L I C Y C E N T E R Methodology for the VoicesDMV Survey Timothy Triplett December 2017 Voices of the Community: DC, Maryland, Virginia

More information

Matching an Internet Panel Sample of Health Care Personnel to a Probability Sample

Matching an Internet Panel Sample of Health Care Personnel to a Probability Sample Matching an Internet Panel Sample of Health Care Personnel to a Probability Sample Charles DiSogra Abt SRBI Carla L Black CDC* Stacie M Greby CDC* John Sokolowski Abt SRBI K.P. Srinath Abt SRBI Xin Yue

More information

AnExaminationoftheQualityand UtilityofInterviewerEstimatesof HouseholdCharacteristicsinthe NationalSurveyofFamilyGrowth. BradyWest

AnExaminationoftheQualityand UtilityofInterviewerEstimatesof HouseholdCharacteristicsinthe NationalSurveyofFamilyGrowth. BradyWest AnExaminationoftheQualityand UtilityofInterviewerEstimatesof HouseholdCharacteristicsinthe NationalSurveyofFamilyGrowth BradyWest An Examination of the Quality and Utility of Interviewer Estimates of Household

More information

BLACK RESIDENTS VIEWS ON HIV/AIDS IN THE DISTRICT OF COLUMBIA

BLACK RESIDENTS VIEWS ON HIV/AIDS IN THE DISTRICT OF COLUMBIA PUBLIC OPINION DISPARITIES & PUBLIC OPINION DATA NOTE A joint product of the Disparities Policy Project and Public Opinion and Survey Research October 2011 BLACK RESIDENTS VIEWS ON HIV/AIDS IN THE DISTRICT

More information

Comparing Alternatives for Estimation from Nonprobability Samples

Comparing Alternatives for Estimation from Nonprobability Samples Comparing Alternatives for Estimation from Nonprobability Samples Richard Valliant Universities of Michigan and Maryland FCSM 2018 1 The Inferential Problem 2 Weighting Nonprobability samples Quasi-randomization

More information

Within-Household Selection in Mail Surveys: Explicit Questions Are Better Than Cover Letter Instructions

Within-Household Selection in Mail Surveys: Explicit Questions Are Better Than Cover Letter Instructions University of Nebraska - Lincoln DigitalCommons@University of Nebraska - Lincoln Sociology Department, Faculty Publications Sociology, Department of 2017 Within-Household Selection in Mail Surveys: Explicit

More information

Unit 1 Exploring and Understanding Data

Unit 1 Exploring and Understanding Data Unit 1 Exploring and Understanding Data Area Principle Bar Chart Boxplot Conditional Distribution Dotplot Empirical Rule Five Number Summary Frequency Distribution Frequency Polygon Histogram Interquartile

More information

A critical review of studies investigating the quality of data obtained with online panels based on probability and nonprobability samples 1

A critical review of studies investigating the quality of data obtained with online panels based on probability and nonprobability samples 1 2 A critical review of studies investigating the quality of data obtained with online panels based on probability and nonprobability samples 1 Mario Callegaro a, Ana Villar b, David Yeager c,and Jon A.

More information

BY Lee Rainie and Cary Funk

BY Lee Rainie and Cary Funk NUMBERS, FACTS AND TRENDS SHAPING THE WORLD JULY 8, BY Lee Rainie and Cary Funk FOR MEDIA OR OTHER INQUIRIES: Lee Rainie, Director Internet, Science and Technology Research Cary Funk, Associate Director,

More information

Tips and Tricks for Raking Survey Data with Advanced Weight Trimming

Tips and Tricks for Raking Survey Data with Advanced Weight Trimming SESUG Paper SD-62-2017 Tips and Tricks for Raking Survey Data with Advanced Trimming Michael P. Battaglia, Battaglia Consulting Group, LLC David Izrael, Abt Associates Sarah W. Ball, Abt Associates ABSTRACT

More information

RECOMMENDED CITATION: Pew Research Center, December, 2014, Perceptions of Job News Trend Upward

RECOMMENDED CITATION: Pew Research Center, December, 2014, Perceptions of Job News Trend Upward NUMBERS, FACTS AND TRENDS SHAPING THE WORLD FOR RELEASE DECEMBER 16, 2014 FOR FURTHER INFORMATION ON THIS REPORT: Carroll Doherty, Director of Political Research Seth Motel, Research Analyst Rachel Weisel,

More information

PDRF About Propensity Weighting emma in Australia Adam Hodgson & Andrey Ponomarev Ipsos Connect Australia

PDRF About Propensity Weighting emma in Australia Adam Hodgson & Andrey Ponomarev Ipsos Connect Australia 1. Introduction It is not news for the research industry that over time, we have to face lower response rates from consumer surveys (Cook, 2000, Holbrook, 2008). It is not infrequent these days, especially

More information

Panel Attrition and its Effects on Results from a Longitudinal Study:

Panel Attrition and its Effects on Results from a Longitudinal Study: Panel Attrition and its Effects on Results from a Longitudinal Study: An Examination of Changes in Participants and Attitudes in a Web-based Panel Survey of HIV and AIDS Stigma Attitudes By: Rodney Baxter

More information

Exploring the Methodological Tradeoffs of Mixed-Mode Surveys with an Experimental Design

Exploring the Methodological Tradeoffs of Mixed-Mode Surveys with an Experimental Design Exploring the Methodological Tradeoffs of Mixed-Mode Surveys with an Experimental Design David Sterrett, Dan Malato, Jennifer Benz, Trevor Tompson, Ipek Bilgen, Nadarajasundaram Ganesh, J. Michael Dennis,

More information

UNEQUAL ENFORCEMENT: How policing of drug possession differs by neighborhood in Baton Rouge

UNEQUAL ENFORCEMENT: How policing of drug possession differs by neighborhood in Baton Rouge February 2017 UNEQUAL ENFORCEMENT: How policing of drug possession differs by neighborhood in Baton Rouge Introduction In this report, we analyze neighborhood disparities in the enforcement of drug possession

More information

Geographical Accuracy of Cell Phone Samples and the Effect on Telephone Survey Bias, Variance, and Cost

Geographical Accuracy of Cell Phone Samples and the Effect on Telephone Survey Bias, Variance, and Cost Geographical Accuracy of Cell Phone Samples and the Effect on Telephone Survey Bias, Variance, and Cost Abstract Benjamin Skalland, NORC at the University of Chicago Meena Khare, National Center for Health

More information

DAZED AND CONFUSED: THE CHARACTERISTICS AND BEHAVIOROF TITLE CONFUSED READERS

DAZED AND CONFUSED: THE CHARACTERISTICS AND BEHAVIOROF TITLE CONFUSED READERS Worldwide Readership Research Symposium 2005 Session 5.6 DAZED AND CONFUSED: THE CHARACTERISTICS AND BEHAVIOROF TITLE CONFUSED READERS Martin Frankel, Risa Becker, Julian Baim and Michal Galin, Mediamark

More information

A critical review of studies investigating the quality of data obtained with online panels based on probability and nonprobability samples 1

A critical review of studies investigating the quality of data obtained with online panels based on probability and nonprobability samples 1 Callegaro c02.tex V1-01/16/2014 6:25 P.M. Page 23 2 A critical review of studies investigating the quality of data obtained with online panels based on probability and nonprobability samples 1 Mario Callegaro

More information

The Impact of Relative Standards on the Propensity to Disclose. Alessandro Acquisti, Leslie K. John, George Loewenstein WEB APPENDIX

The Impact of Relative Standards on the Propensity to Disclose. Alessandro Acquisti, Leslie K. John, George Loewenstein WEB APPENDIX The Impact of Relative Standards on the Propensity to Disclose Alessandro Acquisti, Leslie K. John, George Loewenstein WEB APPENDIX 2 Web Appendix A: Panel data estimation approach As noted in the main

More information

The Pennsylvania State University. The Graduate School. The College of the Liberal Arts THE INCONVENIENCE OF HAPHAZARD SAMPLING.

The Pennsylvania State University. The Graduate School. The College of the Liberal Arts THE INCONVENIENCE OF HAPHAZARD SAMPLING. The Pennsylvania State University The Graduate School The College of the Liberal Arts THE INCONVENIENCE OF HAPHAZARD SAMPLING A Thesis In Sociology and Demography by Veronica Leigh Roth Submitted in Partial

More information

Public Attitudes and Knowledge about HIV/AIDS in Georgia Kaiser Family Foundation

Public Attitudes and Knowledge about HIV/AIDS in Georgia Kaiser Family Foundation Public Attitudes and Knowledge about HIV/AIDS in Georgia Kaiser Family Foundation Chart Pack November 2015 Methodology Public Attitudes and Knowledge about HIV/AIDS in Georgia is a representative, statewide

More information

Introduction and Methodology

Introduction and Methodology To: Valerie Briggs, Director, External Communications and Marketing Outreach National Community Pharmacists Association From: Kellyanne Conway, President & CEO Shelley West, Project Director the polling

More information

Attitudes about Opioids among North Carolina Voters

Attitudes about Opioids among North Carolina Voters Attitudes about Opioids among North Carolina Voters Registered Voters in North Carolina November 6-9th, 2017 TOPLINE RESULTS... 1 CROSS TABULATIONS... 4 QUESTION ORDERING... 9 METHODOLOGICAL INFORMATION...

More information

The Harvard community has made this article openly available. Please share how this access benefits you. Your story matters.

The Harvard community has made this article openly available. Please share how this access benefits you. Your story matters. Can Non-full-probability Internet Surveys Yield Useful Data? A Comparison with Fullprobability Face-to-face Surveys in the Domain of Race and Social Inequality Attitudes The Harvard community has made

More information

Analysis and Interpretation of Data Part 1

Analysis and Interpretation of Data Part 1 Analysis and Interpretation of Data Part 1 DATA ANALYSIS: PRELIMINARY STEPS 1. Editing Field Edit Completeness Legibility Comprehensibility Consistency Uniformity Central Office Edit 2. Coding Specifying

More information

Supplementary Online Content

Supplementary Online Content Supplementary Online Content Gupta RS, Warren CM, Smith BM, et al. Prevalence and severity of food allergies among US adults. JAMA Netw Open. 2019;2(1): e185630. doi:10.1001/jamanetworkopen.2018.5630 emethods.

More information

An evaluation of the weighting procedures for an online access panel survey

An evaluation of the weighting procedures for an online access panel survey Survey Research Methods (2008) Vol.2, No.2, pp. 93-105 ISSN 1864-3361 http://www.surveymethods.org c European Survey Research Association An evaluation of the weighting procedures for an online access

More information

Nonresponse Adjustment Methodology for NHIS-Medicare Linked Data

Nonresponse Adjustment Methodology for NHIS-Medicare Linked Data Nonresponse Adjustment Methodology for NHIS-Medicare Linked Data Michael D. Larsen 1, Michelle Roozeboom 2, and Kathy Schneider 2 1 Department of Statistics, The George Washington University, Rockville,

More information

Statistics is the science of collecting, organizing, presenting, analyzing, and interpreting data to assist in making effective decisions

Statistics is the science of collecting, organizing, presenting, analyzing, and interpreting data to assist in making effective decisions Readings: OpenStax Textbook - Chapters 1 5 (online) Appendix D & E (online) Plous - Chapters 1, 5, 6, 13 (online) Introductory comments Describe how familiarity with statistical methods can - be associated

More information

1. The Role of Sample Survey Design

1. The Role of Sample Survey Design Vista's Approach to Sample Survey Design 1978, 1988, 2006, 2007, 2009 Joseph George Caldwell. All Rights Reserved. Posted at Internet website http://www.foundationwebsite.org. Updated 20 March 2009 (two

More information

Sampling Weights, Model Misspecification and Informative Sampling: A Simulation Study

Sampling Weights, Model Misspecification and Informative Sampling: A Simulation Study Sampling Weights, Model Misspecification and Informative Sampling: A Simulation Study Marianne (Marnie) Bertolet Department of Statistics Carnegie Mellon University Abstract Linear mixed-effects (LME)

More information

Appendix I: Methodology

Appendix I: Methodology Appendix I: Methodology METHODOLOGY SSRS conducted a survey of Muslims and Jews for the Institute for Social Policy and Understanding from January 8 to January 24, 2018. The study investigated the opinions

More information

Use of Paradata in a Responsive Design Framework to Manage a Field Data Collection

Use of Paradata in a Responsive Design Framework to Manage a Field Data Collection Journal of Official Statistics, Vol. 28, No. 4, 2012, pp. 477 499 Use of Paradata in a Responsive Design Framework to Manage a Field Data Collection James Wagner 1, Brady T. West 1, Nicole Kirgis 1, James

More information

UPPER MIDWEST MARKETING AREA ANALYSIS OF COMPONENT LEVELS AND SOMATIC CELL COUNT IN INDIVIDUAL HERD MILK AT THE FARM LEVEL 2015

UPPER MIDWEST MARKETING AREA ANALYSIS OF COMPONENT LEVELS AND SOMATIC CELL COUNT IN INDIVIDUAL HERD MILK AT THE FARM LEVEL 2015 UPPER MIDWEST MARKETING AREA ANALYSIS OF COMPONENT LEVELS AND SOMATIC CELL COUNT IN INDIVIDUAL HERD MILK AT THE FARM LEVEL 2015 Staff Paper 16-01 Prepared by: Corey Freije August 2016 Federal Milk Market

More information

Annenberg Science Knowledge Survey: Zika August 18-August 22, 2016 Florida Oversample/Non-Florida (Week 28, N=1,472) Appendix

Annenberg Science Knowledge Survey: Zika August 18-August 22, 2016 Florida Oversample/Non-Florida (Week 28, N=1,472) Appendix Annenberg Science Knowledge Survey: Zika August 18-August 22, 2016 Florida Oversample/Non-Florida (Week 28, N=1,472) Appendix As you may know, some food products, crops, insects and fish are being modified

More information

Comparability Study of Online and Paper and Pencil Tests Using Modified Internally and Externally Matched Criteria

Comparability Study of Online and Paper and Pencil Tests Using Modified Internally and Externally Matched Criteria Comparability Study of Online and Paper and Pencil Tests Using Modified Internally and Externally Matched Criteria Thakur Karkee Measurement Incorporated Dong-In Kim CTB/McGraw-Hill Kevin Fatica CTB/McGraw-Hill

More information

THE LAST-BIRTHDAY SELECTION METHOD & WITHIN-UNIT COVERAGE PROBLEMS

THE LAST-BIRTHDAY SELECTION METHOD & WITHIN-UNIT COVERAGE PROBLEMS THE LAST-BIRTHDAY SELECTION METHOD & WITHIN-UNIT COVERAGE PROBLEMS Paul J. Lavrakas and Sandra L. Bauman, Northwestern University Survey Laboratory and Daniel M. Merkle, D.S. Howard & Associates Paul J.

More information

National Survey of Young Adults on HIV/AIDS

National Survey of Young Adults on HIV/AIDS REPORT National Survey of Young Adults on HIV/AIDS November 2017 Kaiser Family Foundation Introduction More than three and a half decades have passed since the first case of AIDS. An entire generation

More information

Data and Statistics 101: Key Concepts in the Collection, Analysis, and Application of Child Welfare Data

Data and Statistics 101: Key Concepts in the Collection, Analysis, and Application of Child Welfare Data TECHNICAL REPORT Data and Statistics 101: Key Concepts in the Collection, Analysis, and Application of Child Welfare Data CONTENTS Executive Summary...1 Introduction...2 Overview of Data Analysis Concepts...2

More information

An Application of Propensity Modeling: Comparing Unweighted and Weighted Logistic Regression Models for Nonresponse Adjustments

An Application of Propensity Modeling: Comparing Unweighted and Weighted Logistic Regression Models for Nonresponse Adjustments An Application of Propensity Modeling: Comparing Unweighted and Weighted Logistic Regression Models for Nonresponse Adjustments Frank Potter, 1 Eric Grau, 1 Stephen Williams, 1 Nuria Diaz-Tena, 2 and Barbara

More information

Subject index. bootstrap...94 National Maternal and Infant Health Study (NMIHS) example

Subject index. bootstrap...94 National Maternal and Infant Health Study (NMIHS) example Subject index A AAPOR... see American Association of Public Opinion Research American Association of Public Opinion Research margins of error in nonprobability samples... 132 reports on nonprobability

More information

Finding the strength in numbers

Finding the strength in numbers panel research snapshot Finding the strength in numbers The value of consistency auditing of online panels The very nature of survey research requires that online sample sources have robust quality standards.

More information

On the Importance of Representative Online Survey Data : Probability Sampling and Non-Internet Households

On the Importance of Representative Online Survey Data : Probability Sampling and Non-Internet Households On the Importance of Representative Online Survey Data : Probability Sampling and Non-Internet Households Annelies G. Blom Principal Investigator of the German Internet Panel Juniorprofessor for Methods

More information

Huber, Gregory A. and John S. Lapinski "The "Race Card" Revisited: Assessing Racial Priming in

Huber, Gregory A. and John S. Lapinski The Race Card Revisited: Assessing Racial Priming in Huber, Gregory A. and John S. Lapinski. 2006. "The "Race Card" Revisited: Assessing Racial Priming in Policy Contests." American Journal of Political Science 50 (2 April): 421-440. Appendix, Version 1.0

More information

Best on the Left or on the Right in a Likert Scale

Best on the Left or on the Right in a Likert Scale Best on the Left or on the Right in a Likert Scale Overview In an informal poll of 150 educated research professionals attending the 2009 Sawtooth Conference, 100% of those who voted raised their hands

More information

P E R S P E C T I V E S

P E R S P E C T I V E S PHOENIX CENTER FOR ADVANCED LEGAL & ECONOMIC PUBLIC POLICY STUDIES Revisiting Internet Use and Depression Among the Elderly George S. Ford, PhD June 7, 2013 Introduction Four years ago in a paper entitled

More information

NATIONAL SURVEY OF YOUNG ADULTS ON HIV/AIDS

NATIONAL SURVEY OF YOUNG ADULTS ON HIV/AIDS NATIONAL SURVEY OF YOUNG ADULTS ON HIV/AIDS Kaiser Family Foundation November 30, 2017 Introduction More than three and a half decades have passed since the first case of AIDS. An entire generation has

More information

Scientific Surveys Based on Incomplete Sampling Frames and High Rates of Nonresponse

Scientific Surveys Based on Incomplete Sampling Frames and High Rates of Nonresponse Vol. 8, Issue 6, 2015 Scientific Surveys Based on Incomplete Sampling Frames and High Rates of Nonresponse Mansour Fahimi 1, Nicole Buttermore 2, Randall K Thomas 3, Frances M Barlas 4 Survey Practice

More information

Identifying Peer Influence Effects in Observational Social Network Data: An Evaluation of Propensity Score Methods

Identifying Peer Influence Effects in Observational Social Network Data: An Evaluation of Propensity Score Methods Identifying Peer Influence Effects in Observational Social Network Data: An Evaluation of Propensity Score Methods Dean Eckles Department of Communication Stanford University dean@deaneckles.com Abstract

More information

The Impact of Advance Letters on Cellphone Response in a Statewide Dual-Frame Survey

The Impact of Advance Letters on Cellphone Response in a Statewide Dual-Frame Survey Vol. 11, Issue 2, 2018 The Impact of Advance Letters on Cellphone Response in a Statewide Dual-Frame Survey Eva Aizpurua *, Ki H. Park, Mitchell Avery, Jill Wittrock *, Rodney Muilenburg, Mary E. Losch

More information

Weight Adjustment Methods using Multilevel Propensity Models and Random Forests

Weight Adjustment Methods using Multilevel Propensity Models and Random Forests Weight Adjustment Methods using Multilevel Propensity Models and Random Forests Ronaldo Iachan 1, Maria Prosviryakova 1, Kurt Peters 2, Lauren Restivo 1 1 ICF International, 530 Gaither Road Suite 500,

More information

Stafff Paper Prepared by: Corey Freije. December 2018

Stafff Paper Prepared by: Corey Freije. December 2018 UPPER MIDWEST MARKETING AREA ANALYSIS OF COMPONENT LEVELS AND SOMATIC CELLL COUNT IN INDIVIDUAL HERD MILK AT THE FARM LEVEL 2017 Stafff Paper 18-022 Prepared by: Corey Freije December 2018 Federal Milk

More information

Statistics is the science of collecting, organizing, presenting, analyzing, and interpreting data to assist in making effective decisions

Statistics is the science of collecting, organizing, presenting, analyzing, and interpreting data to assist in making effective decisions Readings: OpenStax Textbook - Chapters 1 5 (online) Appendix D & E (online) Plous - Chapters 1, 5, 6, 13 (online) Introductory comments Describe how familiarity with statistical methods can - be associated

More information

UPPER MIDWEST MARKETING AREA ANALYSIS OF COMPONENT LEVELS AND SOMATIC CELL COUNT IN INDIVIDUAL HERD MILK AT THE FARM LEVEL 2010

UPPER MIDWEST MARKETING AREA ANALYSIS OF COMPONENT LEVELS AND SOMATIC CELL COUNT IN INDIVIDUAL HERD MILK AT THE FARM LEVEL 2010 UPPER MIDWEST MARKETING AREA ANALYSIS OF COMPONENT LEVELS AND SOMATIC CELL COUNT IN INDIVIDUAL HERD MILK AT THE FARM LEVEL Staff Paper 12-02 Prepared by: Corey Freije March 2012 Federal Milk Market Administrator

More information

SURVEY TOPIC INVOLVEMENT AND NONRESPONSE BIAS 1

SURVEY TOPIC INVOLVEMENT AND NONRESPONSE BIAS 1 SURVEY TOPIC INVOLVEMENT AND NONRESPONSE BIAS 1 Brian A. Kojetin (BLS), Eugene Borgida and Mark Snyder (University of Minnesota) Brian A. Kojetin, Bureau of Labor Statistics, 2 Massachusetts Ave. N.E.,

More information

Minnesota Postsecondary Institutions Tobacco-use Policies and Changes in Student Tobacco-use Rates ( )

Minnesota Postsecondary Institutions Tobacco-use Policies and Changes in Student Tobacco-use Rates ( ) Minnesota Postsecondary Institutions Tobacco-use Policies and Changes in Student Tobacco-use Rates (2007 2013) Boynton Health Service Minnesota Postsecondary Institutions Tobacco-use Policies and Changes

More information

Methodology for Non-Randomized Clinical Trials: Propensity Score Analysis Dan Conroy, Ph.D., inventiv Health, Burlington, MA

Methodology for Non-Randomized Clinical Trials: Propensity Score Analysis Dan Conroy, Ph.D., inventiv Health, Burlington, MA PharmaSUG 2014 - Paper SP08 Methodology for Non-Randomized Clinical Trials: Propensity Score Analysis Dan Conroy, Ph.D., inventiv Health, Burlington, MA ABSTRACT Randomized clinical trials serve as the

More information

RECOMMENDED CITATION: Pew Research Center, October 2014, Most Are Confident in Government s Ability to Prevent Major Ebola Outbreak in U.S.

RECOMMENDED CITATION: Pew Research Center, October 2014, Most Are Confident in Government s Ability to Prevent Major Ebola Outbreak in U.S. NUMBERS, FACTS AND TRENDS SHAPING THE WORLD FOR RELEASE OCTOBER 6, 2014 FOR FURTHER INFORMATION ON THIS REPORT: Carroll Doherty, Director of Political Research Alec Tyson, Senior Researcher Rachel Weisel,

More information

Panel Bias from Attrition and Conditioning: A Case Study of the Knowledge Networks Panel

Panel Bias from Attrition and Conditioning: A Case Study of the Knowledge Networks Panel Panel Bias from Attrition and Conditioning: A Case Study of the Knowledge Networks Panel Joshua D. Clinton Department of Political Science Stanford University May 25, 2001 The author would like to thank

More information

Center for Advanced Studies in Measurement and Assessment. CASMA Research Report

Center for Advanced Studies in Measurement and Assessment. CASMA Research Report Center for Advanced Studies in Measurement and Assessment CASMA Research Report Number 39 Evaluation of Comparability of Scores and Passing Decisions for Different Item Pools of Computerized Adaptive Examinations

More information

Speaker Notes: Qualitative Comparative Analysis (QCA) in Implementation Studies

Speaker Notes: Qualitative Comparative Analysis (QCA) in Implementation Studies Speaker Notes: Qualitative Comparative Analysis (QCA) in Implementation Studies PART 1: OVERVIEW Slide 1: Overview Welcome to Qualitative Comparative Analysis in Implementation Studies. This narrated powerpoint

More information

The Impact of Cellphone Sample Representation on Variance Estimates in a Dual-Frame Telephone Survey

The Impact of Cellphone Sample Representation on Variance Estimates in a Dual-Frame Telephone Survey The Impact of Cellphone Sample Representation on Variance Estimates in a Dual-Frame Telephone Survey A. Elizabeth Ormson 1, Kennon R. Copeland 1, B. Stephen J. Blumberg 2, and N. Ganesh 1 1 NORC at the

More information

Americans See Weight Problems Everywhere But In the Mirror

Americans See Weight Problems Everywhere But In the Mirror Americans See Weight Problems Everywhere But In the Mirror EMBARGOED FOR RELEASE APRIL 11, 2006 12:01AM EST Paul Taylor, Executive Vice President Cary Funk, Senior Project Director Peyton Craighill, Project

More information

THE EFFECTS OF SELF AND PROXY RESPONSE STATUS ON THE REPORTING OF RACE AND ETHNICITY l

THE EFFECTS OF SELF AND PROXY RESPONSE STATUS ON THE REPORTING OF RACE AND ETHNICITY l THE EFFECTS OF SELF AND PROXY RESPONSE STATUS ON THE REPORTING OF RACE AND ETHNICITY l Brian A. Harris-Kojetin, Arbitron, and Nancy A. Mathiowetz, University of Maryland Brian Harris-Kojetin, The Arbitron

More information

Readings: Textbook readings: OpenStax - Chapters 1 4 Online readings: Appendix D, E & F Online readings: Plous - Chapters 1, 5, 6, 13

Readings: Textbook readings: OpenStax - Chapters 1 4 Online readings: Appendix D, E & F Online readings: Plous - Chapters 1, 5, 6, 13 Readings: Textbook readings: OpenStax - Chapters 1 4 Online readings: Appendix D, E & F Online readings: Plous - Chapters 1, 5, 6, 13 Introductory comments Describe how familiarity with statistical methods

More information

DO NOT OPEN THIS BOOKLET UNTIL YOU ARE TOLD TO DO SO

DO NOT OPEN THIS BOOKLET UNTIL YOU ARE TOLD TO DO SO NATS 1500 Mid-term test A1 Page 1 of 8 Name (PRINT) Student Number Signature Instructions: York University DIVISION OF NATURAL SCIENCE NATS 1500 3.0 Statistics and Reasoning in Modern Society Mid-Term

More information

Department of International Health

Department of International Health JOHNS HOPKINS U N I V E R S I T Y Center for Clinical Trials Department of Biostatistics Department of Epidemiology Department of International Health Memorandum Department of Medicine Department of Ophthalmology

More information

bivariate analysis: The statistical analysis of the relationship between two variables.

bivariate analysis: The statistical analysis of the relationship between two variables. bivariate analysis: The statistical analysis of the relationship between two variables. cell frequency: The number of cases in a cell of a cross-tabulation (contingency table). chi-square (χ 2 ) test for

More information

Using Soft Refusal Status in the Cell-Phone Nonresponse Adjustment in the National Immunization Survey

Using Soft Refusal Status in the Cell-Phone Nonresponse Adjustment in the National Immunization Survey Using Soft Refusal Status in the Cell-Phone Nonresponse Adjustment in the National Immunization Survey Wei Zeng 1, David Yankey 2 Nadarajasundaram Ganesh 1, Vicki Pineau 1, Phil Smith 2 1 NORC at the University

More information

The Effectiveness of Advance Letters for Cell Telephone Samples

The Effectiveness of Advance Letters for Cell Telephone Samples The Effectiveness of Advance Letters for Cell Telephone Samples Benjamin Skalland 1, Zhen Zhao 2, Jenny Jeyarajah 2 1 NORC at the University of Chicago 2 Centers for Disease Control and Prevention Abstract

More information

NEW JERSEY RESIDENTS DON T KNOW OR LIKE MUCH ABOUT COMMON CORE

NEW JERSEY RESIDENTS DON T KNOW OR LIKE MUCH ABOUT COMMON CORE For immediate release Monday, March 9, 2015 9 pages Contact Dan Cassino 973.896.7072 dcassino@fdu.edu @dancassino NEW JERSEY RESIDENTS DON T KNOW OR LIKE MUCH ABOUT COMMON CORE With PARCC (Partnership

More information

Evaluation of Strategies to Improve the Utility of Estimates from a Non-Probability Based Survey

Evaluation of Strategies to Improve the Utility of Estimates from a Non-Probability Based Survey Evaluation of Strategies to Improve the Utility of Estimates from a Non-Probability Based Survey Presenter: Benmei Liu, National Cancer Institute, NIH liub2@mail.nih.gov Authors: (Next Slide) 2015 FCSM

More information

A Probability Puzzler. Statistics, Data and Statistical Thinking. A Probability Puzzler. A Probability Puzzler. Statistics.

A Probability Puzzler. Statistics, Data and Statistical Thinking. A Probability Puzzler. A Probability Puzzler. Statistics. Statistics, Data and Statistical Thinking FREC 408 Dr. Tom Ilvento 213 Townsend Hall Ilvento@udel.edu A Probability Puzzler Pick a number from 2 to 9. It can be 2 or it can be 9, or any number in between.

More information

Sleep & Effectiveness are Linked, but Few Plan Their Sleep

Sleep & Effectiveness are Linked, but Few Plan Their Sleep Sleep in America Poll 2018 Sleep & Effectiveness are Linked, but Few Plan Their Sleep A majority of American adults (65%) think sleep contributes to next day effectiveness. Great sleepers realize the benefit,

More information

Trends in Smoking Prevalence by Race based on the Tobacco Use Supplement to the Current Population Survey

Trends in Smoking Prevalence by Race based on the Tobacco Use Supplement to the Current Population Survey Trends in Smoking Prevalence by Race based on the Tobacco Use Supplement to the Current Population Survey William W. Davis 1, Anne M. Hartman 1, James T. Gibson 2 National Cancer Institute, Bethesda, MD,

More information

Two-in-three are dieting or exercising In the Battle of the Bulge, More Soldiers Than Successes

Two-in-three are dieting or exercising In the Battle of the Bulge, More Soldiers Than Successes Two-in-three are dieting or exercising In the Battle of the Bulge, More Soldiers Than Successes EMBARGOED FOR RELEASE APRIL 25, 2006 Paul Taylor, Executive Vice President Cary Funk, Senior Project Director

More information

Sub-optimal Respondent Behavior and Data Quality in Online Surveys. Randall K. Thomas GfK Custom Research

Sub-optimal Respondent Behavior and Data Quality in Online Surveys. Randall K. Thomas GfK Custom Research Sub-optimal Respondent Behavior and Data Quality in Online Surveys Randall K. Thomas GfK Custom Research Sub-optimal Response in Surveys Survey satisficing occurs when respondents respond in ways that

More information

Introduction: Statistics, Data and Statistical Thinking Part II

Introduction: Statistics, Data and Statistical Thinking Part II Introduction: Statistics, Data and Statistical Thinking Part II FREC/STAT 608 Dr. Tom Ilvento Department of Food and Resource Economics Let s Continue with our introduction We need terms and definitions

More information

Time-Sharing Experiments for the Social Sciences. Impact of response scale direction on survey responses to factual/behavioral questions

Time-Sharing Experiments for the Social Sciences. Impact of response scale direction on survey responses to factual/behavioral questions Impact of response scale direction on survey responses to factual/behavioral questions Journal: Time-Sharing Experiments for the Social Sciences Manuscript ID: TESS-0.R Manuscript Type: Original Article

More information

MULTIPLE LINEAR REGRESSION 24.1 INTRODUCTION AND OBJECTIVES OBJECTIVES

MULTIPLE LINEAR REGRESSION 24.1 INTRODUCTION AND OBJECTIVES OBJECTIVES 24 MULTIPLE LINEAR REGRESSION 24.1 INTRODUCTION AND OBJECTIVES In the previous chapter, simple linear regression was used when you have one independent variable and one dependent variable. This chapter

More information

Tim Johnson Survey Research Laboratory University of Illinois at Chicago September 2011

Tim Johnson Survey Research Laboratory University of Illinois at Chicago September 2011 Tim Johnson Survey Research Laboratory University of Illinois at Chicago September 2011 1 Characteristics of Surveys Collection of standardized data Primarily quantitative Data are consciously provided

More information

Still important ideas

Still important ideas Readings: OpenStax - Chapters 1 11 + 13 & Appendix D & E (online) Plous - Chapters 2, 3, and 4 Chapter 2: Cognitive Dissonance, Chapter 3: Memory and Hindsight Bias, Chapter 4: Context Dependence Still

More information

J. Michael Brick Westat and JPSM December 9, 2005

J. Michael Brick Westat and JPSM December 9, 2005 Nonresponse in the American Time Use Survey: Who is Missing from the Data and How Much Does It Matter by Abraham, Maitland, and Bianchi J. Michael Brick Westat and JPSM December 9, 2005 Philosophy A sensible

More information

Political Science 15, Winter 2014 Final Review

Political Science 15, Winter 2014 Final Review Political Science 15, Winter 2014 Final Review The major topics covered in class are listed below. You should also take a look at the readings listed on the class website. Studying Politics Scientifically

More information

CHAPTER 2. MEASURING AND DESCRIBING VARIABLES

CHAPTER 2. MEASURING AND DESCRIBING VARIABLES 4 Chapter 2 CHAPTER 2. MEASURING AND DESCRIBING VARIABLES 1. A. Age: name/interval; military dictatorship: value/nominal; strongly oppose: value/ ordinal; election year: name/interval; 62 percent: value/interval;

More information

Methods for treating bias in ISTAT mixed mode social surveys

Methods for treating bias in ISTAT mixed mode social surveys Methods for treating bias in ISTAT mixed mode social surveys C. De Vitiis, A. Guandalini, F. Inglese and M.D. Terribili ITACOSM 2017 Bologna, 16th June 2017 Summary 1. The Mixed Mode in ISTAT social surveys

More information

RECOMMENDED CITATION: Pew Research Center, February 2015, 83% Say Measles Vaccine Is Safe for Healthy Children

RECOMMENDED CITATION: Pew Research Center, February 2015, 83% Say Measles Vaccine Is Safe for Healthy Children NUMBERS, FACTS AND TRENDS SHAPING THE WORLD FOR RELEASE FEBRUARY 9, 2015 FOR FURTHER INFORMATION ON THIS REPORT: Carroll Doherty, Director of Political Research Seth Motel, Research Analyst Rachel Weisel,

More information

A Smarter Way to Select Respondents for Surveys

A Smarter Way to Select Respondents for Surveys Title: A Smarter Way to Select Respondents for Surveys Authors: George H. Terhanian and John F. Bremer Date: 26 February 2012 Status: Final Prepared For: CASRO Toluna 21 River Road Wilton Connecticut Contents

More information

Module 14: Missing Data Concepts

Module 14: Missing Data Concepts Module 14: Missing Data Concepts Jonathan Bartlett & James Carpenter London School of Hygiene & Tropical Medicine Supported by ESRC grant RES 189-25-0103 and MRC grant G0900724 Pre-requisites Module 3

More information

A Structured Approach for Reducing Bias in Surveys Using Online Access Panels

A Structured Approach for Reducing Bias in Surveys Using Online Access Panels A Structured Approach for Reducing Bias in Surveys Using Access Panels Yehuda Dayan Survey Methodologist - Ipsos MORI Neil Farrer Technical Director - Ipsos MediaCT Introduction For many years, the gold

More information

UMbRELLA interim report Preparatory work

UMbRELLA interim report Preparatory work UMbRELLA interim report Preparatory work This document is intended to supplement the UMbRELLA Interim Report 2 (January 2016) by providing a summary of the preliminary analyses which influenced the decision

More information

Modest Rise in Percentage Favoring General Legalization BROAD PUBLIC SUPPORT FOR LEGALIZING MEDICAL MARIJUANA

Modest Rise in Percentage Favoring General Legalization BROAD PUBLIC SUPPORT FOR LEGALIZING MEDICAL MARIJUANA NEWS Release 1615 L Street, N.W., Suite 700 Washington, D.C. 20036 Tel (202) 419-4350 Fax (202) 419-4399 FOR IMMEDIATE RELEASE: Thursday April 1, 2010 Modest Rise in Percentage Favoring General Legalization

More information

Catherine A. Welch 1*, Séverine Sabia 1,2, Eric Brunner 1, Mika Kivimäki 1 and Martin J. Shipley 1

Catherine A. Welch 1*, Séverine Sabia 1,2, Eric Brunner 1, Mika Kivimäki 1 and Martin J. Shipley 1 Welch et al. BMC Medical Research Methodology (2018) 18:89 https://doi.org/10.1186/s12874-018-0548-0 RESEARCH ARTICLE Open Access Does pattern mixture modelling reduce bias due to informative attrition

More information

Readings: Textbook readings: OpenStax - Chapters 1 11 Online readings: Appendix D, E & F Plous Chapters 10, 11, 12 and 14

Readings: Textbook readings: OpenStax - Chapters 1 11 Online readings: Appendix D, E & F Plous Chapters 10, 11, 12 and 14 Readings: Textbook readings: OpenStax - Chapters 1 11 Online readings: Appendix D, E & F Plous Chapters 10, 11, 12 and 14 Still important ideas Contrast the measurement of observable actions (and/or characteristics)

More information

AP Statistics Exam Review: Strand 2: Sampling and Experimentation Date:

AP Statistics Exam Review: Strand 2: Sampling and Experimentation Date: AP Statistics NAME: Exam Review: Strand 2: Sampling and Experimentation Date: Block: II. Sampling and Experimentation: Planning and conducting a study (10%-15%) Data must be collected according to a well-developed

More information

American Views on Stem Cell Research Summary of Survey Findings. Results for America

American Views on Stem Cell Research Summary of Survey Findings. Results for America March 2000 American Views on Stem Cell Research Summary of Survey Findings Prepared for: Results for America A Project of Civil Society Institute Prepared by OPINION RESEARCH CORPORATION February 15, 2005

More information