Developments in cluster randomized trials and Statistics in Medicine

Size: px
Start display at page:

Download "Developments in cluster randomized trials and Statistics in Medicine"

Transcription

1 STATISTICS IN MEDICINE Statist. Med. 2007; 26:2 19 Published online in Wiley InterScience ( Paper Celebrating the 25th Anniversary of Statistics in Medicine Developments in cluster randomized trials and Statistics in Medicine M. J. Campbell 1,, A. Donner 2,3 and N. Klar 2 1 Medical Statistics Group, School of Health and Related Research, University of Sheffield, Sheffield, U.K. 2 Department of Epidemiology and Biostatistics, University of Western Ontario, Canada 3 Robarts Clinical Trials, Robarts Research Institute, London, Canada SUMMARY The design and analysis of cluster randomized trials has been a recurrent theme in Statistics in Medicine since the early volumes. In celebration of 25 years of Statistics in Medicine, this paper reviews recent developments, particularly those that featured in the journal. Issues in design such as sample size calculations, matched paired designs, cohort versus cross-sectional designs, and practical design problems are covered. Developments in analysis include modification of robust methods to cope with small numbers of clusters, generalized estimation equations, population averaged and cluster specific models. Finally, issues on presenting data, some other clustering issues and the general problem of evaluating complex interventions are briefly mentioned. Copyright q 2006 John Wiley & Sons, Ltd. KEY WORDS: cluster randomized trial; matched pair design; generalized estimating equations; population averaged models; cluster specific models; Consort statement 1. INTRODUCTION Trials where groups of subjects, rather than individuals, are randomized, commonly called cluster randomized or group randomized trials are a common theme in Statistics in Medicine, witha special issue [1] devoted to this topic and many important developments chronicled. In celebration of 25 years of Statistics in Medicine, this paper will review developments in cluster randomized trials, particularly, but not exclusively, those published in the journal. Correspondence to: M. J. Campbell, Medical Statistics Group, ScHARR, University of Sheffield, Regent Court, 30 Regent St., Sheffield S1 4DA, U.K. m.j.campbell@sheffield.ac.uk Contract/grant sponsor: Royal Society Contract/grant sponsor: Natural Sciences Engineering Research Council of Canada Received 27 July 2006 Copyright q 2006 John Wiley & Sons, Ltd. Accepted 18 September 2006

2 CLUSTER RANDOMIZED TRIALS 3 Cluster randomized trials are now a key tool in the evaluation of interventions in health services research. They are widely used and two main types of trial are as follows: (1) interventions that involve training or education of health professionals, with the aim of finding a benefit for patients; (2) community randomized trials where the clusters are complete communities (some authors call these large field trials ). These are generally characterized by a relatively small number of clusters each enrolling a large number of subjects. A well-known example of this is the COMMIT smoking intervention trial to be discussed in Section 3. It is worth defining a few terms first. The intraclass correlation (ICC) is the ratio of the between cluster variance to the total variance of an outcome variable and is often denoted by ρ. Different designs can lead to different formulas for estimating ρ. Thedesign effect (DE) is the ratio of the variance of an outcome measure when clustering is accounted for to the variance of the outcome measure when clustering is not accounted for. It is often referred to as the variance inflation factor, since DE measures the amount that one should increase a variance estimate obtained by ignoring clustering to allow for the clustering effect. For clusters of equal size m, it can be shown that DE = 1 + (m 1)ρ. Extensions of this formula to the case of variable cluster sizes are given in Section 2.4. There have been a number of reviews of cluster randomized trials. In 2001 and in Statistics in Medicine, Klar and Donner [2], following on from their book [3] published the previous year, reviewed developments up until that time and suggested areas of further research. These areas included: ethical issues; sample size estimation; analytic issues; economic analysis; meta-analysis of cluster randomized trials and recommendations for trial reporting. The analytic issues included population averagedversus cluster specific (CS) models, methods involving a small number of clusters, issues arising from matched paired designs, overlapping cluster membership, and issues involving missing data. Some related issues were also covered more recently by the same authors [4]. Methodological developments for cluster randomized trials have also been reviewed by Murray et al. [5]. This paper will update the earlier reviews, in particular the 2001 review in Statistics in Medicine. This paper is structured as follows. Sections 2 4 concern design issues: Section 2 presents issues concerning power and sample size, Section 3 matched pair designs and Section 4 cohort versus cross-sectional designs. In Section 5 we will discuss issues in the analysis of cluster randomized trials. Section 6 describes issues on presentation and reporting and Section 7 deals with some further issues. 2. ISSUES ARISING IN THE ASSESSMENT OF POWER AND SAMPLE SIZE 2.1. Number of clusters and number of subjects per cluster Often the number of clusters available to be enrolled in a cluster randomized trial is limited. This has led to considerable interest in developing reliable methods of sample size estimation. Two early papers in Statistics in Medicine reviewed sample size issues in clinical trials in general [6] and particularly in cluster randomized trials [7]. Although values of ρ in cluster randomized trials tend to be small (typically around 0.05 for primary care trials [8]) and in community randomized trials even smaller (usually less than 0.01 and often near [3]) the resulting DE may be very substantial when combined with clusters of large size. Thus, the practice of ignoring clustering effects in the design stage of a trial can lead to an elevated type 2 error, while ignoring it at the analysis stage inevitably leads to an elevated type 1 error.

3 4 M. J. CAMPBELL, A. DONNER AND N. KLAR In cluster randomized trials there are often two sample size choices to be made: the number of clusters and the number of subjects per cluster. A useful rule of thumb is that the power does not increase appreciably once the number of subjects per cluster exceeds 1/ρ [4]. Thus,ifitis believed that the ICC is about 0.05 then it is not worth enrolling more than about 20 subjects per cluster. However, if the ICC is near 0.001, which is often typical of community intervention trials, then a sample of 1000 subjects per cluster may be worthwhile, particularly if recruiting new clusters is difficult. The sample size formulas we give here assume that data for sample size estimation are obtained from a single sample of clusters from the population of interest, that is the intervention itself has no effect on the cluster size. Examples where this is not true have been given by Campbell [8], and we will return to this topic in Section Sample sizes for trials involving survival data One of the earliest (and highest cited) papers in Statistics in Medicine is by Freedman [9] on the sample size required in a study when the outcome is survival. The total number of events, e, required for a power β and two-sided significance α to detect an expected hazard ratio θ is given by 2 (1 + θ)2 e = (z α/2 + z β ) (1 θ) 2 (1) where z α/2 and z β are the normal ordinates associated with probabilities α/2 andβ, is the ratio of the sample size in group 2 to group 1 and θ is the expected hazard ratio between groups 2 and 1. Xie and Waksman [10] generalized equation (1) to cluster randomized trials. An important issue in applying this formula is whether one should estimate ρ from the clustered survival data or, alternatively, from the censoring indicators. Xie and Waksman [10] chose to estimate ρ from the indicators, since this approach is widely used in the clustered binary data literature. Thus, in what follows ρ is estimated as a simple correlation of the 0,1 s indicating whether or not an event is censored, although further research on this estimation question would be useful. They found that the number of events e c required is a simple generalization of Freedman s formula: e c = e(1 + (m 1)ρ) where m = K i=1 m i /K, is the average number of subjects per cluster and K is the number of clusters. The number of subjects required is then given by n = (1 + )e c m((1 P 1 ) + (1 P 2 )) where P 1 and P 2 are the proportion of individuals with non-events in groups 1 and Allowing for imprecision in the ICC Much of the sample size literature deals with the difficulty of obtaining accurate estimates of between-community variation, and hence of ρ, that are needed for sample size planning.

4 CLUSTER RANDOMIZED TRIALS 5 In practice estimates of ρ for a given outcome variable are usually derived from previously reported studies using similar randomization units. However, these estimates are frequently based on studies which themselves are of small size, and thus their inherent inaccuracy may lend the investigators a false sense of confidence. Ukoumunne [11] has given a series of methods for calculating confidence intervals for the ICC, including an application of the bootstrap. However, it is difficult to know how to use these for planning purposes. As Turner et al. [12] have pointed out, for planning purposes it may be better to select a design on the basis of power averaged across uncertainty rather than calculate power from an informally assessed ICC value. Moreover, substitution of a conservative ICC such as the upper value from a 95 per cent confidence interval usually reduces the power for a fixed sample size estimate much more than a method which allows for the imprecision of the ICC. They discuss the use of prior distributions for the ICC and in a Bayesian context show how the uncertainty about this parameter can be expressed in the form of a parametric distribution which naturally leads to a distribution of projected power for any particular design. This Bayesian approach towards the determination of sample size could then be followed by a statistical analysis within the traditional frequentist framework. Turner et al. [12] showed that increasing the number of clusters enables the total sample size to be reduced for a given power. With a given number of subjects estimated from an unclustered design, increasing the number of clusters also considerably reduces the lower limit of the posterior distribution of power. In other words, uncertainty in the ICC will produce uncertainty in the actual power of a study, but a design based on a greater number of clusters has less chance of having a very low power. An alternative approach to dealing with uncertainty in observed values of ρ is described by Feng and Grizzle [13], who proposed the use of a method similar in principle to the bootstrap procedure. Their approach requires the simulation of results of studies of the same size that yielded the observed estimate. One then can substitute the values of ρ obtained from each simulation into the appropriate sample size formula to generate a distribution of projected powers, followed by the selection of a point on this distribution, e.g. the 90th percentile, that reflects the degree of conservativeness desired Allowing for varying cluster sizes Variation in cluster size is another source of imprecision and Kerry and Bland [14] following Donner et al. [15] suggest using DE = 1 + (m a 1)ρ where m a = m 2 i / m i (2) The problem of using this formula is that the individual cluster sizes must be known prior to conducting the trial. Eldridge et al. [16] modified (2) to give DE = 1 + (( 2 (k 1) cv + 1 k ) ) m 1 ρ (3) where cv = s m /m and s m is the s.d. of the cluster size. They suggest that formula (3) is more practical than (2) since the value of cv (coefficient of variation) may often be known in advance. Eldridge et al. [16] also provide some examples of the coefficient of variation for sample sizes typically seen in cluster randomized trials. A further issue is that the estimated ICC will have sampling variation. Eldridge et al. [16] give an empirical correction for the DE in this case.

5 6 M. J. CAMPBELL, A. DONNER AND N. KLAR This formula is appropriate for cluster level analyses that weight by cluster size. For the much more common GEE analysis, which we describe in Section 5, this formula is likely to be conservative Sample size re-estimation When prior knowledge of the value of the variance components required to ascertain the sample size is unavailable, we can undertake an internal pilot study to provide these estimates, using a well cited paper in this area by Wittes and Brittain [17], and as extended to cluster randomized trials by Lake et al. [18]. With this approach, several nuisance parameters, including ρ, the mean cluster size and measures of cluster size variation, are estimated at an interim point in the study, followed by re-estimation of the final required trial size. Although this procedure is most suited to trials that randomize a relatively large number of clusters, such as families or households, over an extended period of time, Lake et al. [18] point out that it could also be applied to at least some community intervention trials provided the participating clusters are recruited prospectively. Turner et al. [12] state that imprecision in the estimate of ρ is not accounted for in this application, but their own method could easily be extended to do so. More experience on the application of internal pilot studies to such trials is clearly needed Empirical investigations of design effects Empirical investigations of design effects likely to arise in cluster randomized trials implemented in a specific setting are also helpful to investigators. An example of such an investigation is given by Mickey et al. [19], who presented design effects from a trial designed to intervene on breast cancer mortality. Ukoumunne et al. [20] presented variance components for some common outcomes, as did Gulliford et al. [21]. Several authors have shown that adjusting for covariates either at the community or individuallevel can improve the power of a trial by reducing the magnitude of the between-cluster variation [8, 22]. Additional gains in power may be realized by modelling individual level covariates (e.g. age) and also at the cluster level (e.g. mean cluster age) as described by Klar and Darlington [23]. 3. ISSUES IN THE DESIGN AND ANALYSIS OF MATCHED PAIR COMMUNITY RANDOMIZED TRIALS 3.1. Design of matched pair studies Matched pair studies are not frequently seen in clinical trials randomizing individual subjects to different intervention groups. However, they have proved to be the design of choice for many investigators embarking on a community intervention trial, largely because of the perceived ability of this design to create intervention groups that are comparable at baseline with respect to important prognostic factors, including, for example, community size and geographic area. The relatively small number of communities that can be enrolled in such studies further enhances the attractiveness of pair matching as a method of reducing the probability of substantively important imbalances that may detract from the credibility of the reported results. Statistics in Medicine has been the forum for much of the debate surrounding the advantages and disadvantages of this design.

6 CLUSTER RANDOMIZED TRIALS 7 Freedman et al. [24] were among the first authors to investigate the gain in efficiency obtained from matching in a community intervention trial. This was done in the context of COMMIT (Community Intervention Trial for Smoking Cessation), a community programme designed to encourage citizens to stop smoking. As discussed by Gail et al. [25], 11 pairs of communities were matched on the basis of several factors expected to be related to the smoking quit rates, such as community size, geographical proximity and demographic profile. Within each matched pair, one community was allocated at random to the intervention group, with the other acting as the control. It is also interesting to note that this trial was one of the first large-scale community intervention studies to use formal power considerations at the planning stage, and, perhaps not coincidentally, to be substantially larger in size than its predecessors. Assuming the gain in efficiency due to matching may be quantified by the factor G = 1/(1 ρ m ), where ρ m is the correlation between members of a pair with respect to the outcome variable, Freedman et al. [24] showed that matching can lead to considerable gains in statistical precision when it is based on an effective surrogate for outcome. However, since G is simply the ratio of population variances ignoring or accounting for pair-matching, it does not take into account the difference in degrees of freedom for estimating these variances, a factor which is particularly relevant in trials enrolling a small number of communities. For example, in the COMMIT study, the degrees of freedom available for estimating error from the paired differences in event rates would be only 10, as compared to the 20 degrees of freedom available for an unmatched analysis. This issue was subsequently addressed in detail by Martin et al. [26], who evaluated the relative efficiency of a matched to an unmatched analysis by using percentiles of the non-central t-distribution to account for the differences in degrees of freedom. On the basis of these results they concluded that for studies having no more than 20 pairs, matching should be used for the purpose of increasing power only if the investigators are confident that ρ m exceeds By considering the practical difficulties that often arise in securing good matches, they also concluded more generally that matching may be overused as a design tool in community intervention trials. These considerations suggest that a tempting strategy in practice may be to perform an unmatched analysis of data arising from a matched pair design, particularly when matching is adopted mainly for the purpose of avoiding a bad randomization. The effectiveness of such a strategy was investigated by Diehr et al. [27], who concluded on the basis of an extensive simulation study that breaking the matches can actually result in an increase in power when the number of pairs is less than ten. Thus, the loss in precision identified by Martin et al. [26] in the presence of weak matching correlations can be at least partially regained. A secondary objective of many community intervention trials is to investigate the effect of individual-level risk factors on one or more outcome variables. Focusing on the case of a continuous outcome variable, Donner et al. [28] showed that the practice of performing an unmatched analysis on data arising from a pair-matched design can lead to bias in the estimated regression coefficient, and a corresponding test of significance which is overly liberal. However, for large-scale community intervention trials, which typically recruit a relatively small number of large clusters, such an analysis will generally be both valid and efficient. Perhaps the most commonly adopted matching factors in large-scale community randomized trials have been cluster size and geographic area (e.g. urban versus rural). Matching by cluster size is attractive not only because it protects against large imbalances in the number of subjects per intervention group, an efficiency consideration, but also because cluster size may be associated with other important but unaccounted for baseline variables, such as socioeconomic status and access to health care resources (Lewsey [29]). Matching by categorized levels of baseline versions

7 8 M. J. CAMPBELL, A. DONNER AND N. KLAR of the trial outcome rate would also seem attractive. However, results reported by Feng et al. [22] suggest that if primary interest is changed from the baseline, such matching is not likely to add benefits in power beyond that yielded by an analysis of change scores Limitations of matched pair designs Klar and Donner [30] explored some further limitations of the matched pair designs that are more general in nature. These limitations arise largely from the total confounding of the intervention effect with the natural variation that exists between two members of a matched pair. One consequence of such confounding is that it precludes the use of standard methods for estimating the underlying ICC, which in turn reduces analytic flexibility. For example, a secondary objective of many studies is to estimate the effect of selected individual level risk factors on one or more outcome variables using regression modelling. However, the calculation of appropriate standard errors for the regression coefficients obtained from such a model requiresa valid estimator of ρ. Thus, although it is possible to perform adjustments for the effect of such risk factors using a two-stage residualized approach e.g. Gail et al. [31], the task of testing for their independent relationship with outcome is more difficult (Donner et al. [28]). Moreover the matched pair design frequently does not bring large gains in precision. This led Klar and Donner [30] to recommend that greater attention be paid to the possibility of adopting a stratified design, in which two or more clusters are randomized to each intervention group within the strata. This design may be particularly attractive when investigators find it challenging to create matched pairs that correspond to unique estimates of risk for each pair. Most importantly, the cluster level replication inherent in this less rigid allocation scheme removes many of the analytic limitations associated with pair-matching, while increasing the degrees of freedom available for estimating error. 4. OTHER DESIGN ISSUES 4.1. Cohort versus cross-sectional designs Many community intervention trials are longitudinal in nature, allowing a choice between a cohort design and a cross-sectional design. For a cohort design, clusters are randomly assigned to intervention groups, with or without stratification. Cohorts sampled from each cluster are then measured over two or more time points, with at least the first measurement occurring before randomization. The cross-sectional design is similar, except that samples are selected independently within each cluster at each time point. Because responses within the same subject often have a strong positive correlation, the cohort design is theoretically more efficient. Feldman and McKinlay [32] presented a unified statistical model that embraces both designs as special cases, thus allowing an assessment of how the values of different design parameters affect their relative precision. A principal conclusion from their investigation was that cohort designs have unique disadvantages that may outweigh any advantage in theoretical efficiency. The first of these is related to possible instability in cohorts of large size, with the resulting likelihood of subject loss to follow-up. Although this disadvantage can be compensated for by oversampling at baseline, this might well negate the original reasons for adopting a cohort design. Differential loss to follow-up by intervention group also creates the risk of bias. The second disadvantage is related to the issue of representativeness of the target population, which is invariably hampered by the ageing of the cohort over time. Assuming that changes related

8 CLUSTER RANDOMIZED TRIALS 9 to the ageing process are independent of the intervention assignment, this effect will not invalidate the principal comparison of interest. However, it does imply that a difference observed in a cohort trial with respect to a given outcome variable cannot be directly compared to the corresponding difference between observed cross-sectional samples. Thus, if the primary questions of interest focus on change at the community level rather than at the level of the individual, cohort samples are the less natural choice. This point is discussed by others [33, 34], who describe and compare several approaches that might be taken to the analysis of repeated cross-sectional samples. Additional discussion is provided by Murray et al. [35] Practical design issues There are two useful papers in Statistics in Medicine that look at practical issues for the conduct of cluster randomized trials. Flynn et al. [36] address the issue of whether it is worth recruiting an extra cluster or to recruit more individuals to existing clusters. They show how the use of contour graphs of power by number of clusters per treatment arm and cluster size can be usefully exploited. Moerbeek [37] suggests that often a cheaper strategy than recruiting another cluster would be to measure additional covariates related to the outcome, in order to reduce the variance. This, of course, requires the cost of measuring the covariate to be relatively small and the correlation of the covariate and the outcome to be reasonably high. 5. ISSUES IN THE ANALYSIS OF CLUSTER RANDOMIZED TRIALS 5.1. Cluster specific versus marginal models Consider an outcome variable Y (which may be continuous or binary), with expected value μ and a generalized link of the form g(μ) = η = β 0 + β 1 X + Z (4) where the function g is assumed strictly monotone and differentiable. We further assume that Z is a random variable E(Z) = 0andvar(Z) = σ 2 Z and so is independent of the fixed effect X. Equation (4) may be a suitable model for clustered data. We assume that the clusters are sampled from a larger population and the effect of any particular cluster i is to add a random effect Z i to all the outcomes. For a cluster randomized we could set X = 1 for intervention and X = 0 for control. A CS model measures the effect on Y of changing X, while Z is held constant. This is a common model for longitudinal data, where it is possible to imagine, say in a cross-over trial, a treatment value changing over time. However, in a cluster randomized, everyone in a cluster receives the same treatment, and although a CS model can be fitted, the result can be interpreted only theoretically. The alternative is to fit a model which looks at the average effect of X over the range of Z. Thisistheso-calledpopulation averaged (PA)ormarginal model. Consider a model where we fit only X, sothat η = β 0 + β 1 X (5) Fitting model (5) is equivalent to fitting a PA model, that is we estimate the effect of X on Y as averaged over all the clusters Z. Neuhaus and Jewell [38] approach the contrast between CS models and PA models by observing that model (5) is simply model (4) with the variable Z omitted. If we assume that the coefficients

9 10 M. J. CAMPBELL, A. DONNER AND N. KLAR for X in the two models are related by β 1 Bβ, where B is the bias factor, then they show that for a linear model B = 1, and so the interpretation of CS and marginal models is the same. However, for a logistic link, μ = logit(p(y = 1 X, Z) is a random variable and σ 2 z B = 1 E(μ){1 E(μ)} = 1 ρ where ρ is the correlation of the Y s within clusters assuming β 1 = 0. Since, as discussed earlier, the value of ρ in community randomized trials is usually less than 0.01, this suggests that the bias in incorrectly assuming a marginal model should not be great. In this case 0<B<1, so the general effect of using a PA model is to attenuate the regression coefficient. One can also see that for a logistic link the greater the variability of the random variable Z, and so the greater the intracluster correlation, the greater the attenuation. Neuhaus and Jewell [38] also show that B = 1 for a log link, which would suggest that in prospective studies such as clinical trials it would be advantageous to use a log-linear model which estimates the relative risk rather than a logistic model for odds ratios. Until recently the software for accurately estimating the standard errors of the estimates for cluster log-linear models was not available, but Zou [39] has recently demonstrated how it can be done using Poisson regression and robust standard errors. The fact that log-linear models estimate the same population parameter for CS and marginal models does not appear well known. For example Murray et al. [5] state that in the case of binary data the coefficient from a marginal model is smaller than that from a conditional model Inflating the standard error For a linear model we write y ij = β 0 + β 1 X i + Z i + ε ij (6) where y ij is the outcome of the jth subject ( j = 1,...,K ) in the ith cluster (i = 1,...,m j ),and we assume E(ε ij ) = E(Z i ) = 0, ε ij and Z i are independent, V (ε ij ) = σ 2 and V (Z i ) = σ 2 Z.Here X i is an intervention indicator variable (0, 1) which depends only on the cluster. An exchangeable correlation structure is assumed, which effectively means that one can exchange subjects j and j within a cluster without changing the covariance. This breaks down if the subjects are measured more than once (for example, at baseline and at follow-up) since the correlation of the same subject measured twice will not be the same as the correlation of two different subjects within a cluster. It also means that the ICC must be assumed to be the same within each arm of the trial, an assumption which is guaranteed in a randomized trial under the null hypothesis of no intervention effect. We can rewrite (6) as y ij = β 0 + β 1 X i + η ij (7) where the η ij are random errors which are not independent, but are typically assumed to have an exchangeable structure defined by the clusters. This is the PA model described in Section 5.1. We can fit model (7) but not account for the correlated nature of the errors terms. Provided that Z i is uncorrelated with X i then we get a valid estimate of the treatment effect but the standard error of the treatment effect is underestimated, and should be inflated. A technique originating in sample survey research is to simply multiply the standard error by the DE. For example, the

10 CLUSTER RANDOMIZED TRIALS 11 DE given in Section 1 may be estimated by replacing the ICC with its sample estimate [3]. A simple test of whether a parameter is zero, known as a modified Wald test, is to divide an estimate of the parameter by its modified standard error which is then compared to a standard normal distribution. Donner and Klar [3] give a number of methods for continuous and binary outcomes which modify the standard error associated with either the t-test or the chi-squared test, respectively. It is important to note that the estimate of the treatment effect is unchanged, only the standard error is inflated. An alternative method is to use the so-called sandwich or Huber White estimator which has a long history in econometrics for estimators with continuous data and with heterogeneous variances. This fits a model where the variance of the error term V (η) = σ 2 i is not assumed constant, and is estimated through the residuals of the model. The advantage of the robust standard error is that one does not need to estimate the ICC separately before conducting the analysis. The disadvantage is that it may be less efficient and so would need a greater number of clusters than does use of the model-dependent standard error. The use of robust standard errors for cluster randomized trials was first suggested by Brook et al. [40], and implemented in a number of statistical packages, for example in Stata by Rogers [41]. In general, Donner and Klar s [3] methods which adjust the standard error give an estimate of the treatment effect which, in effect, weights each cluster by the size of the cluster, so that larger clusters have greater weight. An alternative is to use the method of summary measures, as popularized by Matthews et al. [42]. This gives equal weight to each cluster, irrespective of size and is a CS method. This approach has a great deal to recommend it since it simply uses the summary statistics (mean or proportion) for each cluster and is easy to apply without specialist software. However, one cannot adjust for individual level covariates directly using this approach Fitting models to data The generalized estimating equation (GEE) method, developed by Liang and Zeger [43] in the context of longitudinal studies, has proved to very popular for the analysis of data arising from cluster randomized trials. It fits the PA model and uses an iteratively reweighting algorithm to estimate the parameters and a robust method for the standard error. Use of the GEE is a shrinkage approach which is a compromise between no weighting and weighting by the sample size. Assuming an exchangeable correlation structure, it effectively weights each mean by m i /(m i σ 2 z + σ2 ),which weights the means by the sample size when σ 2 Z = 0 and gives equal weight when σ2 = 0. In practice, estimates of the variance components are used. The analysis of continuous data is relatively straightforward and a comparison of maximum likelihood assuming a normal mixed model, GEE, a bootstrap and a 4-stage method have been givenbyfenget al. [44]. The bootstrap used by Feng et al. draws a random sample of size K from the original K clusters. Then one can use ordinary least squares to estimate the β s and repeat a large number of times. The four stage method is non-iterative, where the first step is to estimate the β s by ordinary least squares and obtain the residuals e i = Y i X i ˆβ. Then the ei s are regressed against the Z i s, leading to estimates of σ 2 and σ 2 Z. One can then use these estimates in a weighted least squares regression of Y i versus X i. For small numbers of clusters (<10 per arm) and for near balanced data the bootstrap has been shown to do well, especially if one does not wish to assume normality. For larger numbers of clusters, the maximum likelihood method performed better than GEE. GEE is used widely for hypothesis testing and confidence interval construction because it can control for the influence of potential confounders on outcome without the need to specify

11 12 M. J. CAMPBELL, A. DONNER AND N. KLAR an underlying distribution for the sample observations. The robust variance estimation relies on between-cluster information to assure the validity of the resulting inferences. It is therefore important to be wary of this approach to community intervention trials, where the amount of such information tends to be relatively small. Feng et al. [44] recommended for continuous data that GEE should not be applied to trials having 20 or fewer communities. Recent research has shown that for studies enrolling few clusters one should take account of the fact that the observed residuals, which are used in robust estimation, are biased estimates of the error variable, because of the requirement that they sum to zero. This bias can be removed by dividing the squared error by the diagonal of the hat matrix. It has also been found that using a t-distribution (or a Satterthwaite type degrees of freedom correction) and a jack-knife improves the estimate of the standard error [45 47]. Pan and Wall [47] proposed replacing the GEE Wald test by approximate t- orf-tests. Although the proposed procedures showed type 1 errors closer to nominal than the usual Wald test, they were shown to be strictly applicable only in clusters of small sizes. It is therefore clear that more research is needed on the development of adjusted GEE procedures that can be applied to clusters of the size that typically arise in community intervention trials. Recently Guo et al. [48] discussed not only the use of a robust standard error in the Wald test, but also the robust score test, particularly for cluster randomized trials. They show that the use of the Wald test is too liberal (more likely to give a significant value) and the score test is too conservative and suggest that a simple correction K/(K 1) to the score statistic will give a test size closer to the nominal value. It is worth commenting that the use of the sandwich estimator has been associated with the GEE method so many times that they may appear synonymous. In fact they are not, and one can use robust standard error without GEE as described in Section Analysis of binary data As discussed in Section 5.1 binary models can estimate different population parameters and so deserve special consideration. Let Y ij (0 or 1) be the jth observation ( j = 1,...,m i ) in the ith cluster (i = 1, 2,...,K ). The intervention variable is X i (0 or 1). The CS logistic model is log it(π ij ) = β 0 + βx i + Z i (8) where Z i is the effect of being in cluster i and where π ij = E(Y ij X i, Z i ). This model can be extended to include individual specific covariates X ij. The random variable Z i is assumed independent of X i and is usually assumed to be normally distributed with mean 0 and variance σ 2 Z.GiventheZ i,they ij s are assumed independently distributed with binomial parameter π ij. Recently Heo and Leon [49] compared different methods of analysis under (8): (i) full likelihood; (ii) penalized quasi-likelihood; (iii) GEE and (iv) fixed effects logistic regression. The third method is a marginal method which, following the discussion in Section 5.1, estimates a different population parameter than the regression coefficient in (8). However, it does not require one to assume a normal distribution for the Z i. The last method does not take the ICC into account, and is an invalid method in general for cluster randomized trials. However if σ 2 Z = 0 it may be expected to yield valid tests and efficient estimates.

12 CLUSTER RANDOMIZED TRIALS 13 The full likelihood L is given by L = K mi μ ij (β, Z i ) y ij {1 μ ij (β, Z i )} 1 y ij f (Z i ; σ 2 ) dz i (9) i=1 j=1 where f (.) is the normal distribution function. Equation (9) can be maximized using Gaussian quadrature or other numerical methods to obtain estimates of the regression coefficients. Penalized quasi-likelihood is a method of maximizing the likelihood whilst avoiding the integration, using a Laplace method for approximating the integral (Breslow and Clayton [50]). Heo and Leon [49] found that the full likelihood method and the penalized likelihood methods to be similar and no worse than the fixed effects method even when the within-cluster correlations are zero. As expected the GEE method gave biased estimates of β cs. They did not investigate the effects of estimating β PA. Preisser et al. [51] showed how to apply a PA model to a pre-test post-test cross-sectional design, where the assumption of an exchangeable correlation matrix breaks down. They state that the GEE approach is asymptotically equivalent to the summary measure approach, and quote Mancl and De Rouen [46] to the effect that using bias-corrected variances can yield valid test sizes even with unequal cluster sizes and with as few as 10 clusters Bayesian methods The likelihood given in (9) requires numerical integration to solve. An alternative method is to use simulation via Monte Carlo Markov Chain (MCMC) algorithms. These are normally associated with a Bayesian analysis, but choice of suitable non-informative priors will yield results similar to the conventional likelihood methods. Spiegelhalter [52] described the method for the Bayesian analysis of cluster randomized trials with a continuous response. This was extended to a binary outcome by Turner et al. [53]. They used model (8) and looked at different prior distributions for σ 2 Z.Sinceσ2 Z is closely related to the ICC, they argued that often it is more appropriate to use a prior distribution on the ICC, since prior information on the ICC is now becoming available [21, 22]. They experimented with different prior distributions and showed that the estimate of the treatment effect is not entirely robust to the distributional assumptions of the model, and suggested caution in using the conventional normality assumption. They contrasted their results to an earlier study by Omar and Thompson [54] who used a multilevel model and a penalized quasi-likelihood procedure to obtain the estimates. Turner et al. [53] showed that the variance components tend to be underestimated when using the non-bayesian approach. Thompson et al. [55] argue that often in a clinical trial one is interested in a risk difference rather than an odds ratio, and describe how to carry out a Bayesian analysis in such a situation Modelling in matched pairs designs Thompson et al. [56] replaced standard modelling approaches by techniques borrowed from metaanalysis. Thus, an intervention/control pair replaces an individual clinical trial of a meta-analysis. This is essentially equivalent to relying on between-stratum information to estimate ρ under the assumption of no intervention by stratum interaction. An attractive feature of this is that the forest

13 14 M. J. CAMPBELL, A. DONNER AND N. KLAR plot can show which pairs appear to be outliers. However, this approach requires a large number of strata (pairs) to ensure its validity, and is therefore not applicable to many community intervention studies. The meta-analysis method as applied to the matched pairs design was extended to binary data by Alexander and Emerson [57] using a Bayesian approach. Their technique was illustrated by applying it to incidence rates of trachoma in Africa. Methods that may be applied to data arising from a stratified cluster randomized design are discussed by Dobbins and Simpson [58] and Song and Ahn [59]. However, they restrict themselves to a comparison of the performance of specific tests in terms of the size and power, rather than on the general problem of estimation. The strict lack of applicability of the t-test to binary outcomes in a matched pair design has led some investigators to alternatively recommend non-parametric approaches, such as Fisher s one sample randomization (permutation) test. Simulations performed by Gail et al. [31] showed that inferences for matched pair binary data using permutation procedures will have significance levels near nominal under conditions likely to arise in community intervention trials. Essentially, the same conclusions were reached by Brookmeyer and Chen [60] for person time data arising from matched pair trials. It is useful to note, however, that the one sample permutation test requires a minimum of six pairs to yield a two-sided p-value of less than 0.05, reflecting its relatively weak power. Donner and Donald [61] showed that a weighted paired t-test based on a logistic transformation of the crude event rates tends to be more powerful than both the permutation test and the standard paired t-test in trials having a small number of strata. 6. PRESENTING DATA 6.1. CONSORT statement It is a perennial complaint amongst analysts that cluster randomized trials are poorly reported [62]. The original CONSORT statement was an attempt to improve reporting of individually randomized trials. This statement was subsequently adapted for cluster randomized trials, published in Statistics in Medicine [63], and has been well cited. It has now been revised and published in a medical journal [64, 65]. The most important distinctions from the original are: (i) to give a rationale for adopting a cluster design; (ii) to describe how the effects of clustering were incorporated into the sample size calculations; (iii) to describe how the effects of clustering were incorporated into the analysis; and (iv) to describe the flow of both clusters and participants through the trial, from assignment to analysis. Thus, in a primary care trial, for example, one would like to know how many primary care groups were approached, how many agreed to participate, how many were randomized, and how many dropped out during the trial, as well as the characteristics of the patients in the study. The most up-to-date statement is available at Improvements in reporting A recent review by Bland [66] contrasted the years 1983, 88, 93, 98 and He suggests that cluster randomized trials are becoming more prevalent and the presentation of cluster randomized trials has improved, particularly after He suggests that statistical pressure has contributed to this and it is of note that the two seminal books were published in that period [3, 67].

14 CLUSTER RANDOMIZED TRIALS OTHER CONSIDERATIONS 7.1. Clustering by practitioner Investigators may need to contend with clustering of subjects responses even for trials using individual randomization. For example, patients may be randomized as they enter a trial by whether or not they will receive a new intervention (say acupuncture). However, there may be a limited number of acupuncturists and so a single acupuncturist may treat a number of patients. The outcomes will thus be affected not only by whether the patient received acupuncture but by which acupuncturist. The model is y ijk = β 0 + β 1 X i + γz k + ε ij where the subscript k indicates the effect of different acupuncturists and Z k = 0 for the control group. However, it is important to note that the subjects in the control group are not naturally clustered. Issues in the analysis of these trials have been covered by Roberts and Roberts [68]. Lee and Thompson [69] discuss a Bayesian approach to the analysis of such trials, and point out that taking into account clustering in the analysis can affect the results Compliance and recruitment We now discuss some potential sources of bias that are peculiar to cluster randomization trials. For example, if the subjects in a trial are newly diagnosed patients and the intervention is some new approach to treating a disease, then it is possible that practitioners in the intervention arm, being newly educated about this disease, may be more likely to diagnose it [8]. This can lead to serious problems of selection bias. Also compliance to treatment may well vary over the participating clusters in a cluster randomized trial (Loeys et al. [70]). While this will not affect the conclusions of an intent to treat analysis, it has major implications for any attempt at casual modelling. Loeys et al. [70] demonstrate how to use standard GEE and random effects models to allow for variable compliance among clusters Complex interventions The use of cluster randomized trials is now well appreciated in the medical world, and methods for analysis are increasingly becoming available in standard software. The main area of further advance would seem to be in the design and planning of the interventions. These are often complex, by which we mean that they depend on a number of interventions. Thus, even a relatively simple study, such as a patient education trial in diabetes raises a number of issues: (i) standardizing what to teach the educators; (ii) training the educators; (iii) deciding how many classes to have; (iv) deciding how many students to allow in each class; and (v) tailoring the education to the needs of the participants. Phase III drug trials are usually preceded by Phase I and Phase II studies in order to decide on issues such as timing and dosage. These are often missing in cluster randomized trials of complex interventions. However, methods of designing complex trials have been described in a set of guidelines produced by the British Medical research Council [71]. These has been criticized on two grounds [72, 73]. First,the guidelines imply that planning of complex interventions is a linear process, whereas often there is no clear ordering to the respective planning stages. Second, they suggest that one should attempt to standardize the intervention, whereas effective interventions,

15 16 M. J. CAMPBELL, A. DONNER AND N. KLAR particularly educational ones, may have to be tailored to the participants. This can lead to problems of understanding exactly what is being tested, which components in an intervention are important, and whether the intervention is replicable. Due to the large number of possible inter-relations in complex trials, computer modelling is playing an increasing role in the planning of such studies. We expect to see an increasing role played in such pre-trial modelling to decide on the optimum design of the intervention. 8. DISCUSSION There is now a vast literature on the design and analysis of cluster randomized trials. The level of understanding has increased greatly since the papers published in early volumes of Statistics in Medicine [6, 7, 13, 26]. However, there is still a need for more empirical work to help the user decide which methods might give better results in particular situations. This review has, of necessity been partial, and biased towards papers published in Statistics in Medicine. However, we feel that many of the key papers have appeared in that journal, and hope that this will continue in the future! ACKNOWLEDGEMENTS The authors are grateful to Obi Ukoumunne and Sandra Eldridge for useful comments on a previous draft. The work of Dr Campbell was supported by a visiting scholarship grant from the Royal Society. The work of Drs Donner and Klar has been partially supported by grants from the Natural Sciences Engineering Research Council of Canada. REFERENCES 1. Campbell MJ, Donner A, Elbourne DR. Design and analysis of cluster randomized trials preface. Statistics in Medicine 2001; 20: Klar N, Donner A. Current and future challenges in the design and analysis of cluster randomization trials. Statistics in Medicine 2001; 20: Donner A, Klar N. Design and Analysis of Cluster Randomization Trials. Arnold: London, Donner A, Klar N. Pitfalls of and controversies in cluster randomized trials. American Journal of Public Health 2004; 94: Murray DM, Vernell SP, Blitstein JL. Design and analysis of group-randomized trials: a review of recent methodological developments. American Journal of Public Health 2004; 94: Donner A. Approaches to sample size estimation in the design of clinical trials. Statistics in Medicine 1984; 3: Hsieh FY. Sample size formulae for intervention studies with the cluster as unit of randomization. Statistics in Medicine 1988; 8: Campbell MJ. Cluster randomized trials in general (family) practice research. Statistical Methods in Medical Research 200; 9: Freedman LS. Tables of the number of patients required in clinical trials using the logrank test. Statistics in Medicine 1982; 1: Xie T, Waksman J. Design and sample size estimation in clinical trials with clustered survival times as the primary endpoint. Statistics in Medicine 2003; 22: Ukoumunne O. Confidence intervals for the intraclass correlation coefficient in cluster randomized trials. Ph.D. Thesis, King s College, University of London, Turner RM, Prevost AT, Thompson SG. Allowing for imprecision of the intracluster correlation coefficient in the design of cluster randomized trials. Statistics in Medicine 2004; 23:

Citation for published version (APA): Ebbes, P. (2004). Latent instrumental variables: a new approach to solve for endogeneity s.n.

Citation for published version (APA): Ebbes, P. (2004). Latent instrumental variables: a new approach to solve for endogeneity s.n. University of Groningen Latent instrumental variables Ebbes, P. IMPORTANT NOTE: You are advised to consult the publisher's version (publisher's PDF) if you wish to cite from it. Please check the document

More information

Bayesian and Frequentist Approaches

Bayesian and Frequentist Approaches Bayesian and Frequentist Approaches G. Jogesh Babu Penn State University http://sites.stat.psu.edu/ babu http://astrostatistics.psu.edu All models are wrong But some are useful George E. P. Box (son-in-law

More information

A Brief Introduction to Bayesian Statistics

A Brief Introduction to Bayesian Statistics A Brief Introduction to Statistics David Kaplan Department of Educational Psychology Methods for Social Policy Research and, Washington, DC 2017 1 / 37 The Reverend Thomas Bayes, 1701 1761 2 / 37 Pierre-Simon

More information

Lecture Outline. Biost 590: Statistical Consulting. Stages of Scientific Studies. Scientific Method

Lecture Outline. Biost 590: Statistical Consulting. Stages of Scientific Studies. Scientific Method Biost 590: Statistical Consulting Statistical Classification of Scientific Studies; Approach to Consulting Lecture Outline Statistical Classification of Scientific Studies Statistical Tasks Approach to

More information

Jinhui Ma 1,2,3, Parminder Raina 1,2, Joseph Beyene 1 and Lehana Thabane 1,3,4,5*

Jinhui Ma 1,2,3, Parminder Raina 1,2, Joseph Beyene 1 and Lehana Thabane 1,3,4,5* Ma et al. BMC Medical Research Methodology 2013, 13:9 RESEARCH ARTICLE Open Access Comparison of population-averaged and cluster-specific models for the analysis of cluster randomized trials with missing

More information

Fundamental Clinical Trial Design

Fundamental Clinical Trial Design Design, Monitoring, and Analysis of Clinical Trials Session 1 Overview and Introduction Overview Scott S. Emerson, M.D., Ph.D. Professor of Biostatistics, University of Washington February 17-19, 2003

More information

MEA DISCUSSION PAPERS

MEA DISCUSSION PAPERS Inference Problems under a Special Form of Heteroskedasticity Helmut Farbmacher, Heinrich Kögel 03-2015 MEA DISCUSSION PAPERS mea Amalienstr. 33_D-80799 Munich_Phone+49 89 38602-355_Fax +49 89 38602-390_www.mea.mpisoc.mpg.de

More information

THE UNIVERSITY OF OKLAHOMA HEALTH SCIENCES CENTER GRADUATE COLLEGE A COMPARISON OF STATISTICAL ANALYSIS MODELING APPROACHES FOR STEPPED-

THE UNIVERSITY OF OKLAHOMA HEALTH SCIENCES CENTER GRADUATE COLLEGE A COMPARISON OF STATISTICAL ANALYSIS MODELING APPROACHES FOR STEPPED- THE UNIVERSITY OF OKLAHOMA HEALTH SCIENCES CENTER GRADUATE COLLEGE A COMPARISON OF STATISTICAL ANALYSIS MODELING APPROACHES FOR STEPPED- WEDGE CLUSTER RANDOMIZED TRIALS THAT INCLUDE MULTILEVEL CLUSTERING,

More information

Bayesian Confidence Intervals for Means and Variances of Lognormal and Bivariate Lognormal Distributions

Bayesian Confidence Intervals for Means and Variances of Lognormal and Bivariate Lognormal Distributions Bayesian Confidence Intervals for Means and Variances of Lognormal and Bivariate Lognormal Distributions J. Harvey a,b, & A.J. van der Merwe b a Centre for Statistical Consultation Department of Statistics

More information

Bayesian Logistic Regression Modelling via Markov Chain Monte Carlo Algorithm

Bayesian Logistic Regression Modelling via Markov Chain Monte Carlo Algorithm Journal of Social and Development Sciences Vol. 4, No. 4, pp. 93-97, Apr 203 (ISSN 222-52) Bayesian Logistic Regression Modelling via Markov Chain Monte Carlo Algorithm Henry De-Graft Acquah University

More information

Chapter 11: Advanced Remedial Measures. Weighted Least Squares (WLS)

Chapter 11: Advanced Remedial Measures. Weighted Least Squares (WLS) Chapter : Advanced Remedial Measures Weighted Least Squares (WLS) When the error variance appears nonconstant, a transformation (of Y and/or X) is a quick remedy. But it may not solve the problem, or it

More information

Analysis of TB prevalence surveys

Analysis of TB prevalence surveys Workshop and training course on TB prevalence surveys with a focus on field operations Analysis of TB prevalence surveys Day 8 Thursday, 4 August 2011 Phnom Penh Babis Sismanidis with acknowledgements

More information

Estimating drug effects in the presence of placebo response: Causal inference using growth mixture modeling

Estimating drug effects in the presence of placebo response: Causal inference using growth mixture modeling STATISTICS IN MEDICINE Statist. Med. 2009; 28:3363 3385 Published online 3 September 2009 in Wiley InterScience (www.interscience.wiley.com).3721 Estimating drug effects in the presence of placebo response:

More information

Instrumental Variables Estimation: An Introduction

Instrumental Variables Estimation: An Introduction Instrumental Variables Estimation: An Introduction Susan L. Ettner, Ph.D. Professor Division of General Internal Medicine and Health Services Research, UCLA The Problem The Problem Suppose you wish to

More information

Glossary. Ó 2010 John Wiley & Sons, Ltd

Glossary. Ó 2010 John Wiley & Sons, Ltd Glossary The majority of the definitions within this glossary are based on, but are only a selection from, the comprehensive list provided by Day (2007) in the Dictionary of Clinical Trials. We have added

More information

How to analyze correlated and longitudinal data?

How to analyze correlated and longitudinal data? How to analyze correlated and longitudinal data? Niloofar Ramezani, University of Northern Colorado, Greeley, Colorado ABSTRACT Longitudinal and correlated data are extensively used across disciplines

More information

Cochrane Pregnancy and Childbirth Group Methodological Guidelines

Cochrane Pregnancy and Childbirth Group Methodological Guidelines Cochrane Pregnancy and Childbirth Group Methodological Guidelines [Prepared by Simon Gates: July 2009, updated July 2012] These guidelines are intended to aid quality and consistency across the reviews

More information

Sample size and power calculations in Mendelian randomization with a single instrumental variable and a binary outcome

Sample size and power calculations in Mendelian randomization with a single instrumental variable and a binary outcome Sample size and power calculations in Mendelian randomization with a single instrumental variable and a binary outcome Stephen Burgess July 10, 2013 Abstract Background: Sample size calculations are an

More information

EPI 200C Final, June 4 th, 2009 This exam includes 24 questions.

EPI 200C Final, June 4 th, 2009 This exam includes 24 questions. Greenland/Arah, Epi 200C Sp 2000 1 of 6 EPI 200C Final, June 4 th, 2009 This exam includes 24 questions. INSTRUCTIONS: Write all answers on the answer sheets supplied; PRINT YOUR NAME and STUDENT ID NUMBER

More information

Unit 1 Exploring and Understanding Data

Unit 1 Exploring and Understanding Data Unit 1 Exploring and Understanding Data Area Principle Bar Chart Boxplot Conditional Distribution Dotplot Empirical Rule Five Number Summary Frequency Distribution Frequency Polygon Histogram Interquartile

More information

Biostatistics II

Biostatistics II Biostatistics II 514-5509 Course Description: Modern multivariable statistical analysis based on the concept of generalized linear models. Includes linear, logistic, and Poisson regression, survival analysis,

More information

A re-randomisation design for clinical trials

A re-randomisation design for clinical trials Kahan et al. BMC Medical Research Methodology (2015) 15:96 DOI 10.1186/s12874-015-0082-2 RESEARCH ARTICLE Open Access A re-randomisation design for clinical trials Brennan C Kahan 1*, Andrew B Forbes 2,

More information

You must answer question 1.

You must answer question 1. Research Methods and Statistics Specialty Area Exam October 28, 2015 Part I: Statistics Committee: Richard Williams (Chair), Elizabeth McClintock, Sarah Mustillo You must answer question 1. 1. Suppose

More information

Understanding Uncertainty in School League Tables*

Understanding Uncertainty in School League Tables* FISCAL STUDIES, vol. 32, no. 2, pp. 207 224 (2011) 0143-5671 Understanding Uncertainty in School League Tables* GEORGE LECKIE and HARVEY GOLDSTEIN Centre for Multilevel Modelling, University of Bristol

More information

Section on Survey Research Methods JSM 2009

Section on Survey Research Methods JSM 2009 Missing Data and Complex Samples: The Impact of Listwise Deletion vs. Subpopulation Analysis on Statistical Bias and Hypothesis Test Results when Data are MCAR and MAR Bethany A. Bell, Jeffrey D. Kromrey

More information

Sampling Weights, Model Misspecification and Informative Sampling: A Simulation Study

Sampling Weights, Model Misspecification and Informative Sampling: A Simulation Study Sampling Weights, Model Misspecification and Informative Sampling: A Simulation Study Marianne (Marnie) Bertolet Department of Statistics Carnegie Mellon University Abstract Linear mixed-effects (LME)

More information

Meta-analysis using individual participant data: one-stage and two-stage approaches, and why they may differ

Meta-analysis using individual participant data: one-stage and two-stage approaches, and why they may differ Tutorial in Biostatistics Received: 11 March 2016, Accepted: 13 September 2016 Published online 16 October 2016 in Wiley Online Library (wileyonlinelibrary.com) DOI: 10.1002/sim.7141 Meta-analysis using

More information

Glossary From Running Randomized Evaluations: A Practical Guide, by Rachel Glennerster and Kudzai Takavarasha

Glossary From Running Randomized Evaluations: A Practical Guide, by Rachel Glennerster and Kudzai Takavarasha Glossary From Running Randomized Evaluations: A Practical Guide, by Rachel Glennerster and Kudzai Takavarasha attrition: When data are missing because we are unable to measure the outcomes of some of the

More information

Score Tests of Normality in Bivariate Probit Models

Score Tests of Normality in Bivariate Probit Models Score Tests of Normality in Bivariate Probit Models Anthony Murphy Nuffield College, Oxford OX1 1NF, UK Abstract: A relatively simple and convenient score test of normality in the bivariate probit model

More information

Russian Journal of Agricultural and Socio-Economic Sciences, 3(15)

Russian Journal of Agricultural and Socio-Economic Sciences, 3(15) ON THE COMPARISON OF BAYESIAN INFORMATION CRITERION AND DRAPER S INFORMATION CRITERION IN SELECTION OF AN ASYMMETRIC PRICE RELATIONSHIP: BOOTSTRAP SIMULATION RESULTS Henry de-graft Acquah, Senior Lecturer

More information

CHL 5225 H Advanced Statistical Methods for Clinical Trials. CHL 5225 H The Language of Clinical Trials

CHL 5225 H Advanced Statistical Methods for Clinical Trials. CHL 5225 H The Language of Clinical Trials CHL 5225 H Advanced Statistical Methods for Clinical Trials Two sources for course material 1. Electronic blackboard required readings 2. www.andywillan.com/chl5225h code of conduct course outline schedule

More information

Mediation Analysis With Principal Stratification

Mediation Analysis With Principal Stratification University of Pennsylvania ScholarlyCommons Statistics Papers Wharton Faculty Research 3-30-009 Mediation Analysis With Principal Stratification Robert Gallop Dylan S. Small University of Pennsylvania

More information

Improving ecological inference using individual-level data

Improving ecological inference using individual-level data Improving ecological inference using individual-level data Christopher Jackson, Nicky Best and Sylvia Richardson Department of Epidemiology and Public Health, Imperial College School of Medicine, London,

More information

Combining Risks from Several Tumors Using Markov Chain Monte Carlo

Combining Risks from Several Tumors Using Markov Chain Monte Carlo University of Nebraska - Lincoln DigitalCommons@University of Nebraska - Lincoln U.S. Environmental Protection Agency Papers U.S. Environmental Protection Agency 2009 Combining Risks from Several Tumors

More information

A novel approach to estimation of the time to biomarker threshold: Applications to HIV

A novel approach to estimation of the time to biomarker threshold: Applications to HIV A novel approach to estimation of the time to biomarker threshold: Applications to HIV Pharmaceutical Statistics, Volume 15, Issue 6, Pages 541-549, November/December 2016 PSI Journal Club 22 March 2017

More information

Bayesian Analysis of Between-Group Differences in Variance Components in Hierarchical Generalized Linear Models

Bayesian Analysis of Between-Group Differences in Variance Components in Hierarchical Generalized Linear Models Bayesian Analysis of Between-Group Differences in Variance Components in Hierarchical Generalized Linear Models Brady T. West Michigan Program in Survey Methodology, Institute for Social Research, 46 Thompson

More information

Chapter 5: Field experimental designs in agriculture

Chapter 5: Field experimental designs in agriculture Chapter 5: Field experimental designs in agriculture Jose Crossa Biometrics and Statistics Unit Crop Research Informatics Lab (CRIL) CIMMYT. Int. Apdo. Postal 6-641, 06600 Mexico, DF, Mexico Introduction

More information

Strategies for handling missing data in randomised trials

Strategies for handling missing data in randomised trials Strategies for handling missing data in randomised trials NIHR statistical meeting London, 13th February 2012 Ian White MRC Biostatistics Unit, Cambridge, UK Plan 1. Why do missing data matter? 2. Popular

More information

Two-stage Methods to Implement and Analyze the Biomarker-guided Clinical Trail Designs in the Presence of Biomarker Misclassification

Two-stage Methods to Implement and Analyze the Biomarker-guided Clinical Trail Designs in the Presence of Biomarker Misclassification RESEARCH HIGHLIGHT Two-stage Methods to Implement and Analyze the Biomarker-guided Clinical Trail Designs in the Presence of Biomarker Misclassification Yong Zang 1, Beibei Guo 2 1 Department of Mathematical

More information

Catherine A. Welch 1*, Séverine Sabia 1,2, Eric Brunner 1, Mika Kivimäki 1 and Martin J. Shipley 1

Catherine A. Welch 1*, Séverine Sabia 1,2, Eric Brunner 1, Mika Kivimäki 1 and Martin J. Shipley 1 Welch et al. BMC Medical Research Methodology (2018) 18:89 https://doi.org/10.1186/s12874-018-0548-0 RESEARCH ARTICLE Open Access Does pattern mixture modelling reduce bias due to informative attrition

More information

Methods Research Report. An Empirical Assessment of Bivariate Methods for Meta-Analysis of Test Accuracy

Methods Research Report. An Empirical Assessment of Bivariate Methods for Meta-Analysis of Test Accuracy Methods Research Report An Empirical Assessment of Bivariate Methods for Meta-Analysis of Test Accuracy Methods Research Report An Empirical Assessment of Bivariate Methods for Meta-Analysis of Test Accuracy

More information

A COMPARISON OF IMPUTATION METHODS FOR MISSING DATA IN A MULTI-CENTER RANDOMIZED CLINICAL TRIAL: THE IMPACT STUDY

A COMPARISON OF IMPUTATION METHODS FOR MISSING DATA IN A MULTI-CENTER RANDOMIZED CLINICAL TRIAL: THE IMPACT STUDY A COMPARISON OF IMPUTATION METHODS FOR MISSING DATA IN A MULTI-CENTER RANDOMIZED CLINICAL TRIAL: THE IMPACT STUDY Lingqi Tang 1, Thomas R. Belin 2, and Juwon Song 2 1 Center for Health Services Research,

More information

The Effect of Extremes in Small Sample Size on Simple Mixed Models: A Comparison of Level-1 and Level-2 Size

The Effect of Extremes in Small Sample Size on Simple Mixed Models: A Comparison of Level-1 and Level-2 Size INSTITUTE FOR DEFENSE ANALYSES The Effect of Extremes in Small Sample Size on Simple Mixed Models: A Comparison of Level-1 and Level-2 Size Jane Pinelis, Project Leader February 26, 2018 Approved for public

More information

Sample size calculation for a stepped wedge trial

Sample size calculation for a stepped wedge trial Baio et al. Trials (2015) 16:354 DOI 10.1186/s13063-015-0840-9 TRIALS RESEARCH Sample size calculation for a stepped wedge trial Open Access Gianluca Baio 1*,AndrewCopas 2, Gareth Ambler 1, James Hargreaves

More information

Generalized Estimating Equations for Depression Dose Regimes

Generalized Estimating Equations for Depression Dose Regimes Generalized Estimating Equations for Depression Dose Regimes Karen Walker, Walker Consulting LLC, Menifee CA Generalized Estimating Equations on the average produce consistent estimates of the regression

More information

Multilevel analysis quantifies variation in the experimental effect while optimizing power and preventing false positives

Multilevel analysis quantifies variation in the experimental effect while optimizing power and preventing false positives DOI 10.1186/s12868-015-0228-5 BMC Neuroscience RESEARCH ARTICLE Open Access Multilevel analysis quantifies variation in the experimental effect while optimizing power and preventing false positives Emmeke

More information

Treatment effect estimates adjusted for small-study effects via a limit meta-analysis

Treatment effect estimates adjusted for small-study effects via a limit meta-analysis Treatment effect estimates adjusted for small-study effects via a limit meta-analysis Gerta Rücker 1, James Carpenter 12, Guido Schwarzer 1 1 Institute of Medical Biometry and Medical Informatics, University

More information

Use of GEEs in STATA

Use of GEEs in STATA Use of GEEs in STATA 1. When generalised estimating equations are used and example 2. Stata commands and options for GEEs 3. Results from Stata (and SAS!) 4. Another use of GEEs Use of GEEs GEEs are one

More information

THE INDIRECT EFFECT IN MULTIPLE MEDIATORS MODEL BY STRUCTURAL EQUATION MODELING ABSTRACT

THE INDIRECT EFFECT IN MULTIPLE MEDIATORS MODEL BY STRUCTURAL EQUATION MODELING ABSTRACT European Journal of Business, Economics and Accountancy Vol. 4, No. 3, 016 ISSN 056-6018 THE INDIRECT EFFECT IN MULTIPLE MEDIATORS MODEL BY STRUCTURAL EQUATION MODELING Li-Ju Chen Department of Business

More information

Chapter 13 Estimating the Modified Odds Ratio

Chapter 13 Estimating the Modified Odds Ratio Chapter 13 Estimating the Modified Odds Ratio Modified odds ratio vis-à-vis modified mean difference To a large extent, this chapter replicates the content of Chapter 10 (Estimating the modified mean difference),

More information

Bayesian graphical models for combining multiple data sources, with applications in environmental epidemiology

Bayesian graphical models for combining multiple data sources, with applications in environmental epidemiology Bayesian graphical models for combining multiple data sources, with applications in environmental epidemiology Sylvia Richardson 1 sylvia.richardson@imperial.co.uk Joint work with: Alexina Mason 1, Lawrence

More information

Meta-Analysis. Zifei Liu. Biological and Agricultural Engineering

Meta-Analysis. Zifei Liu. Biological and Agricultural Engineering Meta-Analysis Zifei Liu What is a meta-analysis; why perform a metaanalysis? How a meta-analysis work some basic concepts and principles Steps of Meta-analysis Cautions on meta-analysis 2 What is Meta-analysis

More information

Regression Discontinuity Analysis

Regression Discontinuity Analysis Regression Discontinuity Analysis A researcher wants to determine whether tutoring underachieving middle school students improves their math grades. Another wonders whether providing financial aid to low-income

More information

Lecture Outline. Biost 517 Applied Biostatistics I. Purpose of Descriptive Statistics. Purpose of Descriptive Statistics

Lecture Outline. Biost 517 Applied Biostatistics I. Purpose of Descriptive Statistics. Purpose of Descriptive Statistics Biost 517 Applied Biostatistics I Scott S. Emerson, M.D., Ph.D. Professor of Biostatistics University of Washington Lecture 3: Overview of Descriptive Statistics October 3, 2005 Lecture Outline Purpose

More information

A Case Study: Two-sample categorical data

A Case Study: Two-sample categorical data A Case Study: Two-sample categorical data Patrick Breheny January 31 Patrick Breheny BST 701: Bayesian Modeling in Biostatistics 1/43 Introduction Model specification Continuous vs. mixture priors Choice

More information

An Empirical Assessment of Bivariate Methods for Meta-analysis of Test Accuracy

An Empirical Assessment of Bivariate Methods for Meta-analysis of Test Accuracy Number XX An Empirical Assessment of Bivariate Methods for Meta-analysis of Test Accuracy Prepared for: Agency for Healthcare Research and Quality U.S. Department of Health and Human Services 54 Gaither

More information

The Relative Performance of Full Information Maximum Likelihood Estimation for Missing Data in Structural Equation Models

The Relative Performance of Full Information Maximum Likelihood Estimation for Missing Data in Structural Equation Models University of Nebraska - Lincoln DigitalCommons@University of Nebraska - Lincoln Educational Psychology Papers and Publications Educational Psychology, Department of 7-1-2001 The Relative Performance of

More information

Modelling Research Productivity Using a Generalization of the Ordered Logistic Regression Model

Modelling Research Productivity Using a Generalization of the Ordered Logistic Regression Model Modelling Research Productivity Using a Generalization of the Ordered Logistic Regression Model Delia North Temesgen Zewotir Michael Murray Abstract In South Africa, the Department of Education allocates

More information

Methods for Addressing Selection Bias in Observational Studies

Methods for Addressing Selection Bias in Observational Studies Methods for Addressing Selection Bias in Observational Studies Susan L. Ettner, Ph.D. Professor Division of General Internal Medicine and Health Services Research, UCLA What is Selection Bias? In the regression

More information

Investigating the robustness of the nonparametric Levene test with more than two groups

Investigating the robustness of the nonparametric Levene test with more than two groups Psicológica (2014), 35, 361-383. Investigating the robustness of the nonparametric Levene test with more than two groups David W. Nordstokke * and S. Mitchell Colp University of Calgary, Canada Testing

More information

Statistical Science Issues in HIV Vaccine Trials: Part I

Statistical Science Issues in HIV Vaccine Trials: Part I Statistical Science Issues in HIV Vaccine Trials: Part I 1 2 Outline 1. Study population 2. Criteria for selecting a vaccine for efficacy testing 3. Measuring effects of vaccination - biological markers

More information

Introduction to Bayesian Analysis 1

Introduction to Bayesian Analysis 1 Biostats VHM 801/802 Courses Fall 2005, Atlantic Veterinary College, PEI Henrik Stryhn Introduction to Bayesian Analysis 1 Little known outside the statistical science, there exist two different approaches

More information

Using Statistical Principles to Implement FDA Guidance on Cardiovascular Risk Assessment for Diabetes Drugs

Using Statistical Principles to Implement FDA Guidance on Cardiovascular Risk Assessment for Diabetes Drugs Using Statistical Principles to Implement FDA Guidance on Cardiovascular Risk Assessment for Diabetes Drugs David Manner, Brenda Crowe and Linda Shurzinske BASS XVI November 9-13, 2009 Acknowledgements

More information

Analysis of Vaccine Effects on Post-Infection Endpoints Biostat 578A Lecture 3

Analysis of Vaccine Effects on Post-Infection Endpoints Biostat 578A Lecture 3 Analysis of Vaccine Effects on Post-Infection Endpoints Biostat 578A Lecture 3 Analysis of Vaccine Effects on Post-Infection Endpoints p.1/40 Data Collected in Phase IIb/III Vaccine Trial Longitudinal

More information

Intraclass Correlation Coefficients Typical of Cluster-Randomized Studies: Estimates From the Robert Wood Johnson Prescription for Health Projects

Intraclass Correlation Coefficients Typical of Cluster-Randomized Studies: Estimates From the Robert Wood Johnson Prescription for Health Projects Intraclass Correlation Coefficients Typical of Cluster-Randomized Studies: Estimates From the Robert Wood Johnson Prescription for Health Projects David M. Thompson, PhD 1 Douglas H. Fernald, MA 2,3 James

More information

A NEW TRIAL DESIGN FULLY INTEGRATING BIOMARKER INFORMATION FOR THE EVALUATION OF TREATMENT-EFFECT MECHANISMS IN PERSONALISED MEDICINE

A NEW TRIAL DESIGN FULLY INTEGRATING BIOMARKER INFORMATION FOR THE EVALUATION OF TREATMENT-EFFECT MECHANISMS IN PERSONALISED MEDICINE A NEW TRIAL DESIGN FULLY INTEGRATING BIOMARKER INFORMATION FOR THE EVALUATION OF TREATMENT-EFFECT MECHANISMS IN PERSONALISED MEDICINE Dr Richard Emsley Centre for Biostatistics, Institute of Population

More information

1 Introduction. st0020. The Stata Journal (2002) 2, Number 3, pp

1 Introduction. st0020. The Stata Journal (2002) 2, Number 3, pp The Stata Journal (22) 2, Number 3, pp. 28 289 Comparative assessment of three common algorithms for estimating the variance of the area under the nonparametric receiver operating characteristic curve

More information

Flexible Matching in Case-Control Studies of Gene-Environment Interactions

Flexible Matching in Case-Control Studies of Gene-Environment Interactions American Journal of Epidemiology Copyright 2004 by the Johns Hopkins Bloomberg School of Public Health All rights reserved Vol. 59, No. Printed in U.S.A. DOI: 0.093/aje/kwg250 ORIGINAL CONTRIBUTIONS Flexible

More information

OLS Regression with Clustered Data

OLS Regression with Clustered Data OLS Regression with Clustered Data Analyzing Clustered Data with OLS Regression: The Effect of a Hierarchical Data Structure Daniel M. McNeish University of Maryland, College Park A previous study by Mundfrom

More information

Appropriate Statistical Methods to Account for Similarities in Binary Outcomes Between Fellow Eyes

Appropriate Statistical Methods to Account for Similarities in Binary Outcomes Between Fellow Eyes Appropriate Statistical Methods to Account for Similarities in Binary Outcomes Between Fellow Eyes Joanne Katz,* Scott Zeger,-\ and Kung-Yee Liangf Purpose. Many ocular measurements are more alike between

More information

EPSE 594: Meta-Analysis: Quantitative Research Synthesis

EPSE 594: Meta-Analysis: Quantitative Research Synthesis EPSE 594: Meta-Analysis: Quantitative Research Synthesis Ed Kroc University of British Columbia ed.kroc@ubc.ca March 28, 2019 Ed Kroc (UBC) EPSE 594 March 28, 2019 1 / 32 Last Time Publication bias Funnel

More information

Performance of Median and Least Squares Regression for Slightly Skewed Data

Performance of Median and Least Squares Regression for Slightly Skewed Data World Academy of Science, Engineering and Technology 9 Performance of Median and Least Squares Regression for Slightly Skewed Data Carolina Bancayrin - Baguio Abstract This paper presents the concept of

More information

10 Intraclass Correlations under the Mixed Factorial Design

10 Intraclass Correlations under the Mixed Factorial Design CHAPTER 1 Intraclass Correlations under the Mixed Factorial Design OBJECTIVE This chapter aims at presenting methods for analyzing intraclass correlation coefficients for reliability studies based on a

More information

Applied Medical. Statistics Using SAS. Geoff Der. Brian S. Everitt. CRC Press. Taylor Si Francis Croup. Taylor & Francis Croup, an informa business

Applied Medical. Statistics Using SAS. Geoff Der. Brian S. Everitt. CRC Press. Taylor Si Francis Croup. Taylor & Francis Croup, an informa business Applied Medical Statistics Using SAS Geoff Der Brian S. Everitt CRC Press Taylor Si Francis Croup Boca Raton London New York CRC Press is an imprint of the Taylor & Francis Croup, an informa business A

More information

Bayesian Estimation of a Meta-analysis model using Gibbs sampler

Bayesian Estimation of a Meta-analysis model using Gibbs sampler University of Wollongong Research Online Applied Statistics Education and Research Collaboration (ASEARC) - Conference Papers Faculty of Engineering and Information Sciences 2012 Bayesian Estimation of

More information

Economic evaluation of factorial randomised controlled trials: challenges, methods and recommendations

Economic evaluation of factorial randomised controlled trials: challenges, methods and recommendations Research Article Received: 31 May 2016, Accepted: 4 April 2017 Published online 3 May 2017 in Wiley Online Library (wileyonlinelibrary.com) DOI: 10.1002/sim.7322 Economic evaluation of factorial randomised

More information

Detection of Unknown Confounders. by Bayesian Confirmatory Factor Analysis

Detection of Unknown Confounders. by Bayesian Confirmatory Factor Analysis Advanced Studies in Medical Sciences, Vol. 1, 2013, no. 3, 143-156 HIKARI Ltd, www.m-hikari.com Detection of Unknown Confounders by Bayesian Confirmatory Factor Analysis Emil Kupek Department of Public

More information

Analyzing diastolic and systolic blood pressure individually or jointly?

Analyzing diastolic and systolic blood pressure individually or jointly? Analyzing diastolic and systolic blood pressure individually or jointly? Chenglin Ye a, Gary Foster a, Lisa Dolovich b, Lehana Thabane a,c a. Department of Clinical Epidemiology and Biostatistics, McMaster

More information

2.75: 84% 2.5: 80% 2.25: 78% 2: 74% 1.75: 70% 1.5: 66% 1.25: 64% 1.0: 60% 0.5: 50% 0.25: 25% 0: 0%

2.75: 84% 2.5: 80% 2.25: 78% 2: 74% 1.75: 70% 1.5: 66% 1.25: 64% 1.0: 60% 0.5: 50% 0.25: 25% 0: 0% Capstone Test (will consist of FOUR quizzes and the FINAL test grade will be an average of the four quizzes). Capstone #1: Review of Chapters 1-3 Capstone #2: Review of Chapter 4 Capstone #3: Review of

More information

Student Performance Q&A:

Student Performance Q&A: Student Performance Q&A: 2009 AP Statistics Free-Response Questions The following comments on the 2009 free-response questions for AP Statistics were written by the Chief Reader, Christine Franklin of

More information

A review of statistical methods in the analysis of data arising from observer reliability studies (Part 11) *

A review of statistical methods in the analysis of data arising from observer reliability studies (Part 11) * A review of statistical methods in the analysis of data arising from observer reliability studies (Part 11) * by J. RICHARD LANDIS** and GARY G. KOCH** 4 Methods proposed for nominal and ordinal data Many

More information

STATISTICAL INFERENCE 1 Richard A. Johnson Professor Emeritus Department of Statistics University of Wisconsin

STATISTICAL INFERENCE 1 Richard A. Johnson Professor Emeritus Department of Statistics University of Wisconsin STATISTICAL INFERENCE 1 Richard A. Johnson Professor Emeritus Department of Statistics University of Wisconsin Key words : Bayesian approach, classical approach, confidence interval, estimation, randomization,

More information

Propensity scores: what, why and why not?

Propensity scores: what, why and why not? Propensity scores: what, why and why not? Rhian Daniel, Cardiff University @statnav Joint workshop S3RI & Wessex Institute University of Southampton, 22nd March 2018 Rhian Daniel @statnav/propensity scores:

More information

Models for HSV shedding must account for two levels of overdispersion

Models for HSV shedding must account for two levels of overdispersion UW Biostatistics Working Paper Series 1-20-2016 Models for HSV shedding must account for two levels of overdispersion Amalia Magaret University of Washington - Seattle Campus, amag@uw.edu Suggested Citation

More information

Chapter 1: Exploring Data

Chapter 1: Exploring Data Chapter 1: Exploring Data Key Vocabulary:! individual! variable! frequency table! relative frequency table! distribution! pie chart! bar graph! two-way table! marginal distributions! conditional distributions!

More information

Bias in regression coefficient estimates when assumptions for handling missing data are violated: a simulation study

Bias in regression coefficient estimates when assumptions for handling missing data are violated: a simulation study STATISTICAL METHODS Epidemiology Biostatistics and Public Health - 2016, Volume 13, Number 1 Bias in regression coefficient estimates when assumptions for handling missing data are violated: a simulation

More information

Meta-analysis of two studies in the presence of heterogeneity with applications in rare diseases

Meta-analysis of two studies in the presence of heterogeneity with applications in rare diseases Meta-analysis of two studies in the presence of heterogeneity with applications in rare diseases Christian Röver 1, Tim Friede 1, Simon Wandel 2 and Beat Neuenschwander 2 1 Department of Medical Statistics,

More information

Bayesian methods in health economics

Bayesian methods in health economics Bayesian methods in health economics Gianluca Baio University College London Department of Statistical Science g.baio@ucl.ac.uk Seminar Series of the Master in Advanced Artificial Intelligence Madrid,

More information

Title: A robustness study of parametric and non-parametric tests in Model-Based Multifactor Dimensionality Reduction for epistasis detection

Title: A robustness study of parametric and non-parametric tests in Model-Based Multifactor Dimensionality Reduction for epistasis detection Author's response to reviews Title: A robustness study of parametric and non-parametric tests in Model-Based Multifactor Dimensionality Reduction for epistasis detection Authors: Jestinah M Mahachie John

More information

Ordinal Data Modeling

Ordinal Data Modeling Valen E. Johnson James H. Albert Ordinal Data Modeling With 73 illustrations I ". Springer Contents Preface v 1 Review of Classical and Bayesian Inference 1 1.1 Learning about a binomial proportion 1 1.1.1

More information

BIOSTATISTICAL METHODS AND RESEARCH DESIGNS. Xihong Lin Department of Biostatistics, University of Michigan, Ann Arbor, MI, USA

BIOSTATISTICAL METHODS AND RESEARCH DESIGNS. Xihong Lin Department of Biostatistics, University of Michigan, Ann Arbor, MI, USA BIOSTATISTICAL METHODS AND RESEARCH DESIGNS Xihong Lin Department of Biostatistics, University of Michigan, Ann Arbor, MI, USA Keywords: Case-control study, Cohort study, Cross-Sectional Study, Generalized

More information

IAPT: Regression. Regression analyses

IAPT: Regression. Regression analyses Regression analyses IAPT: Regression Regression is the rather strange name given to a set of methods for predicting one variable from another. The data shown in Table 1 and come from a student project

More information

A Decision Tree for Controlled Trials

A Decision Tree for Controlled Trials SPORTSCIENCE Perspectives / Research Resources A Decision Tree for Controlled Trials Alan M Batterham, Will G Hopkins sportsci.org Sportscience 9, 33-39, 2005 (sportsci.org/jour/05/wghamb.htm) School of

More information

Instrumental Variables I (cont.)

Instrumental Variables I (cont.) Review Instrumental Variables Observational Studies Cross Sectional Regressions Omitted Variables, Reverse causation Randomized Control Trials Difference in Difference Time invariant omitted variables

More information

Dichotomizing partial compliance and increased participant burden in factorial designs: the performance of four noncompliance methods

Dichotomizing partial compliance and increased participant burden in factorial designs: the performance of four noncompliance methods Merrill and McClure Trials (2015) 16:523 DOI 1186/s13063-015-1044-z TRIALS RESEARCH Open Access Dichotomizing partial compliance and increased participant burden in factorial designs: the performance of

More information

Critical Appraisal Series

Critical Appraisal Series Definition for therapeutic study Terms Definitions Study design section Observational descriptive studies Observational analytical studies Experimental studies Pragmatic trial Cluster trial Researcher

More information

Carrying out an Empirical Project

Carrying out an Empirical Project Carrying out an Empirical Project Empirical Analysis & Style Hint Special program: Pre-training 1 Carrying out an Empirical Project 1. Posing a Question 2. Literature Review 3. Data Collection 4. Econometric

More information

Lecture Slides. Elementary Statistics Eleventh Edition. by Mario F. Triola. and the Triola Statistics Series 1.1-1

Lecture Slides. Elementary Statistics Eleventh Edition. by Mario F. Triola. and the Triola Statistics Series 1.1-1 Lecture Slides Elementary Statistics Eleventh Edition and the Triola Statistics Series by Mario F. Triola 1.1-1 Chapter 1 Introduction to Statistics 1-1 Review and Preview 1-2 Statistical Thinking 1-3

More information

A Bayesian Nonparametric Model Fit statistic of Item Response Models

A Bayesian Nonparametric Model Fit statistic of Item Response Models A Bayesian Nonparametric Model Fit statistic of Item Response Models Purpose As more and more states move to use the computer adaptive test for their assessments, item response theory (IRT) has been widely

More information