An empirical comparison of methods to meta-analyze individual patient data of diagnostic accuracy

Size: px
Start display at page:

Download "An empirical comparison of methods to meta-analyze individual patient data of diagnostic accuracy"

Transcription

1 An empirical comparison of methods to meta-analyze individual patient data of diagnostic accuracy Gabrielle Simoneau Master of Science Department of Epidemiology, Biostatistics and Occupational Health McGill University Montreal, Quebec August 2015 A thesis submitted to McGill University in partial fulfillment of the requirements of the degree of Master in Biostatistics Gabrielle Simoneau 2015

2 ACKNOWLEDGEMENTS I acknowledge the CIHR Knowledge Synthesis Grant Improving Depression Screening by Reducing Bias in Accuracy Estimates: An Independent Patient Data Meta-Analysis of the PHQ-9 for funding my year of research. I thank my supervisor Dr. Andrea Benedetti for guiding me through this research and for offering me interesting academic and career opportunities. I thank Dr. Brett Thombs for his thoughtful advice and Brooke Levis for the large amount of work she has previously done to make my research feasible. I thank my parents, my boyfriend and my roommates for their patience, support, encouragement. ii

3 ABSTRACT Individual patient data (IPD) meta-analyses are increasingly common in the literature. In the context of diagnostic accuracy of ordinal or semi-continuous scale tests, sensitivity and specificity are traditionally reported for a given threshold and a meta-analysis is conducted via a bivariate approach to account for the correlation between sensitivity and specificity. With IPD, sensitivity and specificity can be pooled for every possible threshold. One way to analyze these data is to use the bivariate approach separately at every threshold. Another idea is to analyze sensitivity and specificity for all thresholds simultaneously, and thus account for their within-study correlation across thresholds. Among other approaches, the ordinal multivariate random-effects model and the normal multivariate random-effects model have been proposed to meta-analyze diagnostic test data with multiple thresholds. Our aim is to compare these two models to the bivariate approach when IPD of diagnostic accuracy are available, empirically and via simulations. The CAGE IPD dataset and the 9-item Patient Health Questionnaire (PHQ-9) IPD dataset are used to conduct the empirical comparisons. The empirical comparisons showed that the two multivariate methods were difficult to apply and prone to sporadic convergence issues while the bivariate approach was easy to apply but overly simplistic. Notwithstanding, simulations showed that ignoring the within-study correlation of sensitivity and specificity across thresholds does not dramatically affect the inference of the bivariate model. The two multivariate models were not suitable for simulations, which emphasized the complexity of the models. We recommend IPD of diagnostic accuracy to be iii

4 meta-analyzed via the bivariate approach or the more complete ordinal multivariate approach. iv

5 ABRÉGÉ Les méta-analyses sur les données individuelles de patient (IPD) sont de plus en plus communes dans la littérature scientifique. L utilisation de données individuelles de patient a plusieurs avantages. Dans le contexte d études sur la précision du diagnostique de test à échelle ordinale ou semi-continue, la sensibilité et la spécificité du test sont habituellement reportées pour un certain seuil et une méta-analyse est conduite via une approche bivariée qui prend en compte la corrélation entre la sensibilité et la spécificité. Avec des IPD, il est possible d obtenir une paire de sensibilité et spécificité pour tous les seuils possibles. Une façon d analyser ce genre de données est d appliquer le modèle bivarié seuil par seuil. Une autre façon d analyser ces données est de modéliser les résultats pour tous les seuils simultanément. La corrélation entre les différents seuils pour chaque étude est alors prise en compte. Parmi d autres méthodes, les modèles ordinal multivarié avec effets aléatoires et normal multivarié avec effets aléatoires ont été proposés pour méta-analyser des données sur la précision du diagnostic avec plusieurs seuils. Notre objectif est de comparer ces deux méthodes au modèle bivarié, empiriquement et via une étude de simulation, lorsque des données IPD sont disponibles. Les données IPD CAGE et les données IPD sur le Questionnaire sur la Santé du Patient-9 (PHQ-9) sont utilisées pour mener les comparaisons empiriques. Les comparaisons empiriques ont montré que les deux approches multivariées sont difficiles à appliquer et sujettes à d occasionnel problèmes de convergence tandis que l approche bivariée est simple à appliquer, mais exagérément simpliste. Néanmoins, les études de simulation ont montré qu ignorer la corrélation intra-étude v

6 n a pas une effet dramatique sur l inférence du modèle bivarié. Les deux approches multivariées ne convenaient pas aux études de simulation, ce qui met l accent sur la complexité des modèles. Nous recommandons de méta-analyser les données IPD sur la précision du diagnostique via l approche bivariée ou le plus complexe modèle ordinal multivarié. vi

7 TABLE OF CONTENTS ACKNOWLEDGEMENTS ABSTRACT ABRÉGÉ LIST OF TABLES LIST OF FIGURES ii iii v ix x 1 Introduction Background Diagnostic Accuracy Measures Traditional and Individual Patient Data Meta-Analysis Traditional meta-analysis Individual Patient Data meta-analysis Meta-analysis of diagnostic accuracy studies Models Bivariate random-effects model Ordinal multivariate random-effects model Normal multivariate random-effects model Comparison of proposed methods Assumptions Practical interest Methodology Objective Motivating examples Motivating example 1: CAGE questionnaire Motivating example 2: PHQ-9 questionnaire Methods vii

8 4 Results CAGE questionnaire: results PHQ-9 questionnaire: results Implementation in statistical software Simulation Discussion Diagnostic accuracy of the CAGE and the PHQ-9 questionnaires Related work Strengths and limitations Conclusion Appendix A Appendix B Appendix C References viii

9 Table LIST OF TABLES page 2 1 Reporting results of a diagnostic accuracy study for a given threshold, 2 2 table The IPD CAGE dataset Estimated parameters by the three methods: CAGE Estimated sensitivities and specificities by the three methods: CAGE The IPD PHQ-9 dataset Estimated parameters by the two methods: PHQ Estimated sensitivities and specificities by the three methods: PHQ Simulation results for scenarios 1 and Simulation results statistical analysis: bias of the pooled estimates Simulation results for scenarios 3 and Simulation results for scenarios 5 and Simulation results for scenarios 7 and Simulation results for scenarios 9 and Simulation results for scenarios 11 and Simulation results for scenarios 13 and Simulation results for scenarios 15 and ix

10 Figure LIST OF FIGURES page 4 1 Individual ROC curves: CAGE Pooled sroc curves from the three methods: CAGE Pooled sroc curves with their 95% confidence band: CAGE Individual ROC curves: PHQ Pooled sroc curves from the three methods: CAGE Pooled sroc curves with their 95% confidence band: CAGE x

11 CHAPTER 1 Introduction The use of diagnostic 1 tests in patients is of interest both for clinicians and health-policy makers [34]. On one hand, clinicians need to decide whether to use a diagnostic test and how to interpret it. On the other hand, health-policy makers need to assess the value of a diagnostic test compared to other available tests, and decide whether to make it available. The evaluation of the diagnostic accuracy of a test is an important step in this decision-making. All available published information on the accuracy of a test need to be properly accounted for to base decision-making on the overall value of diagnostic accuracy [15]. Considering this setting, systematic reviews and meta-analyses are well-established methods to rigorously combine information from multiple studies. Guidelines on how to conduct a systematic review in general [9, 38, 51] or specifically for diagnostic accuracy are discussed elsewhere [34, 39, 42]. Individual patient data (IPD) meta-analyses are increasingly common in the literature [7, 13, 39, 63, 70]. The use of IPD has many benefits over traditional meta-analyses of published data. In the context of diagnostic accuracy studies, a pair of sensitivity and specificity is usually reported for a given threshold and a 1 The tests we will consider in this work are actually screening tests. 1

12 traditional meta-analysis is conducted to obtain a pooled result for that threshold [33, 49, 52, 58, 67]. However, primary studies might report pairs of sensitivity and specificity for the different sets of thresholds and selective reporting of thresholds might lead to inaccurate conclusions [43]. When IPD are available for diagnostic studies, sensitivity and specificity can be estimated for every possible threshold of a test [37, 62]. One might analyze the data threshold by threshold. However, such approach ignores the correlation between sensitivities and specificities across thresholds within a study [66]. Alternatively, one might analyze data for all thresholds simultaneously and thus take into account the complex structure of the data. Such methods are referred to IPD meta-analysis methods [37, 63]. Few papers have considered analyzing multiple thresholds simultaneously in the context of IPD meta-analyses in the last 15 years [24, 37, 56, 66]. The objective of this work is 1) to understand current statistical challenges in IPD meta-analyses of diagnostic accuracy studies, 2) to explore the potential statistical benefits of IPD meta-analyses of diagnostic accuracy studies, 3) to empirically compare available meta-analytic methods of diagnostic studies when IPD are available and 4) to compare the validity of the inference across meta-analytic methods via simulation studies. In order to achieve these goals, a literature review will highlight the statistical challenges in this area. Two empirical comparisons and a simulation study will be perform to assess the performance of each method. In Chapter 2, important background concepts of diagnostic accuracy studies will be defined. The concepts of meta-analysis and IPD meta-analysis will be explained. Three existing IPD meta-analytic models will be presented and compared 2

13 in terms of assumptions and applicability. In Chapter 3, two motivating examples will be introduced, and our methods will be explained. In Chapter 4, results from the two empirical comparisons will be shown. Simulation results will be presented in Chapter 5 and Chapter 6 will be dedicated to a thoughtful discussion around these comparisons. 3

14 CHAPTER 2 Background 2.1 Diagnostic Accuracy Measures Diagnostic accuracy refers to the ability of a test to distinguish between diseased and non-diseased patients, where the disease status has to be clearly defined [34]. The objective of a diagnostic accuracy study is to evaluate the performance of a new test compared to the gold standard test. The gold standard gives the true disease status of a patient. Here, we make the assumption that it has a perfect diagnostic accuracy. However, the gold standard test can be expensive, complex and invasive, which motivate the development of a screening test. Therefore, to evaluate the diagnostic accuracy of this test, the screening test result is compared to the true disease status as obtained by the gold standard. In clinical practice, health care providers are interested in the predictive values of a test. The positive predictive value (PPV) refers to the probability that a patient with a positive test result will be diseased whereas the negative predictive value (NPV) refers to the probability that a patient with a negative test will not have the disease [17]. However, PPV and NPV depend on disease prevalence in the studied population [17]. Sensitivity and specificity are closely related to PPV and NPV. Sensitivity is the probability that a test result will be positive given that a patient is truly diseased whereas specificity is the probability that a test result will be negative 4

15 given that a patient is truly not diseased. Even if they are less clinically relevant, sensitivity and specificity are usually reported to assess the diagnostic accuracy of a test since both measures are independent of disease prevalence [17]. From estimates of sensitivity and specificity, one can derive PPV and NPV to obtain clinically relevant measures. Of interest here are binary classifier tests such that the test result is positive (T+) or negative (T-), whereas the true disease status of a patient is diseased (D+) or non-diseased (D-). For tests that are not binary by definition (ordinal or continuous tests), a threshold has to be defined to dichotomize it. Define x k as the test result for patient k and j as a threshold. Then, for tests where higher scores are associated with the presence of the disease, x k < j = T- x k j = T+ For continuous tests, there exist an infinite number of thresholds ranging through all possible values taken by the test. For ordinal tests with J ordered categories, there exist J-1 thresholds [24]. In this work, we will focus on tests with ordered categories. For every threshold investigated, a diagnostic accuracy study reports a 2 2 contingency table showing the distribution of truly diseased and truly non-diseased patients with respect to their test results as in Table

16 Table 2 1: Reporting results of a diagnostic accuracy study for a given threshold, 2 2 table Disease status Test Result D+ D- Total T+ TP FP POS T- FN TN NEG Total n 1 n 0 n Here, n 1 and n 0 are the total number of truly diseased and truly non-diseased patients in the study, respectively, where the sample size n is n = n 1 +n 0. POS and NEG are the number of patients with a positive and negative test result, respectively. TP, FP, FN and TN respectively refer to the number of true positive, false positive, false negative and true negative patients in a study. From this table, estimates of sensitivity and specificity when considering threshold j are derived as Sensitivity j = TP n 1 and Specificity j = TN n 0 Sensitivity and specificity of a test are closely related to the chosen threshold: for a test with an increasing scale, as the threshold value increases, specificity increases while sensitivity decreases [58]. This relation between sensitivity and specificity is clearly depicted in the receiver operating characteristic (ROC) curve. The ROC curve plots pairs of sensitivity and (1-specificity) over the range of all possible thresholds of a test. The ROC curve visually describes the overall diagnostic accuracy of a test. From the ROC curve, one can compute the area under the curve (AUC) to quantify diagnostic accuracy [28]. The AUC is equivalent to the probability that a randomly chosen individual from the truly diseased sample has a greater test value 6

17 than a randomly chosen individual from the truly non-diseased sample [28]. AUC scores range from 0.5 to 1. A non-informative test has an AUC of 0.5 with its corresponding ROC curve lying on the diagonal line of the graph. A perfect test has an AUC of 1.0 with a ROC curve following the left and upper boundary of the graph. The AUC summarizes the whole ROC curve by giving the same weighting to the full range of threshold values [23, 79]. In practice, one might want to focus on a range of sensitivity (specificity) of clinical importance by giving larger weights to these values. Interpretation and comparison of AUCs must consider this weighting scheme [23]. The disadvantage of summarizing the diagnostic accuracy of a test in a single measure is that it does not distinguish between the ability of correctly detecting diseased patients (sensitivity) and non-diseased patients (specificity) [58]. Other diagnostic accuracy measures exist (positive and negative likelihood ratios, diagnostic odds ratio) [17], but will not be considered here. 2.2 Traditional and Individual Patient Data Meta-Analysis Traditional meta-analysis A meta-analysis is a statistical tool to combine results from independent studies addressing a specific clinical question [51]. Traditional meta-analyses pool published data to produce summary estimates of the effect of interest along with its associated uncertainty [51, 63]. Here, effect can refer to any measure of association between an exposure and an outcome (e.g. risk difference, odds ratio, risk ratio, proportions difference, means difference, etc). By pooling results across studies, a meta-analysis essentially computes a weighted average [51]. 7

18 A meta-analysis is usually the last step of a systematic review, but all systematic reviews do not necessarily include a meta-analysis [38, 51]. A systematic review is a method motivated by the need for healthcare decisions and policy making to rely on strong evidence from high-quality studies. It includes a comprehensive and exhaustive search of primary studies reporting results on a specific research question. It further assesses the quality of primary studies based on pre-established criteria of reproducibility and clearness. Results from selected studies are then synthesized using explicit methods, such as a meta-analysis, to produce strong evidence on the question of interest. Key requirements to conduct a meta-analysis include that selected studies report quantitative summary estimates of the effect of interest [51]. However, differences between studies are common [22, 51] and occur because of clinical heterogeneity (differences in population, outcome definition, intervention definition) and methodological heterogeneity (differences in study design or quality) [58]. These variations in study characteristics are likely to influence study-specific effect estimates. Heterogeneity should be investigated in order to increase the clinical relevance and the quality of conclusions drawn from a systematic review [9, 22, 51]. Visual assessment of heterogeneity (e.g. forrest plot, l Abbé plot) or formal statistical tests for heterogeneity (e.g. I 2 ) [38, 51] can be used for such investigation. Meta-analytic statistical methods are in fact based on the interpretation of the observed differences between results from primary studies i.e. heterogeneity. Three schools of thought may characterize this interpretation. First, one could believe that 8

19 the same effect is present in all included studies, and that observed differences between study estimates are only due to random variation. Such a belief motivates the fixed-effect model [51]. This first interpretation of observed differences across studies is highly unrealistic. Second, one could believe that primary studies estimate slightly different effects: each study estimates an underlying true study-specific effect and all true study-specific effects vary around an overall average effect. The observed differences between study-specific estimates are then a mix of two sources of variations: between-study variability given that each study is not thought to estimate the same effect and random variability. This interpretation of observed variations yields to the random-effects model [51]. This model is more plausible since it makes no assumption on the true underlying study-specific effects. Still, both the fixed- and random-effects models make several distributional assumptions [30]. Finally, one could use predefined guidelines imposed by formal statistical test for heterogeneity, and rely on these test results to decide whether to use a fixed- or random-effects model. Notwithstanding, the random-effects model remains the most realistic and plausible and should be preferred most of the time [6, 51, 68]. Traditional meta-analyses have several limitations. Publication bias occurs when statistically significant positive results are more likely to be published than studies finding no statistically significant results [13, 51]. A meta-analysis prone to publication bias might find exaggerated overall estimates. Graphical methods (e.g. funnel plot) and statistical tests exist to check the presence of publication bias [51]. Also, traditional meta-analyses highly depend on the quality of included studies or the quality of reporting [7]. Finally, subgroup analyses are very limited in conventional 9

20 meta-analyses. Most primary studies simply report overall estimates of the effect of interest and might not even be powerful enough to estimate subgroup-specific effects [7] Individual Patient Data meta-analysis An alternative way to conduct a meta-analysis is to obtain and use individual patient data (IPD) instead of published estimates of effect [63]. This alternative meta-analytic method is referred to as individual patient or participant data metaanalysis. IPD are the raw individual-level data recorded for each participant in a study whereas published data or aggregated data consists of study-level summary measures derived from these raw data [13, 63]. From IPD, one can obtain published results but published results usually do not allow recovering raw individual patient data. The objective of an IPD meta-analysis is still to summarize results from independent studies. The statistical method used to meta-analyze IPD must account for clustering of patients within studies, and never analyze the data as if all patients were coming from one single study [63]. As with traditional meta-analyses, differences between studies can occur due to clinical or methodological heterogeneity and these differences should be accounted by the statistical analysis. IPD meta-analyses are performed using two statistical analyses, both retaining study clustering and allowing fixed-effects or random-effects modelling: two-stage and one-stage approaches [13, 63]. In the two-stage approach, IPD are first summarized in each study separately and independently using an adequate statistical 10

21 method (e.g. logistic regression). Aggregated data, i.e. study-level estimates of the effect of interest along with a measure of their variability, are then available for each study. These aggregated data are combined using an appropriate traditional meta-analysis method to produce a pooled summary effect. This first approach is laborious in the sense that it implies re-analyzing data for each study and conducting a traditional meta-analysis [63, 69]. In the one-stage approach, IPD from all studies are simultaneously analyzed in a single step while accounting for clustering of patients by study. In contrast with the two-stage approach, study-level effects are not explicitly estimated for each study [63]. However, model specification and clear distinction between the sources of variability make the one-stage approach more complex for non-statisticians [63, 69]. The two-stage approach is the most popular method to conduct IPD meta-analyses [69, 72]. However, the one-stage approach is a more flexible method and allows accounting for both within- and between-study effects [63]. IPD meta-analyses have many advantages over traditional aggregated data metaanalyses [13, 63]. Aggregated data are often poorly reported and statistical methods more or less clearly detailed [69]. Aggregated data can be presented differently from one study to the other, say relative risk versus odds ratio. The availability of IPD counterbalances some of these disadvantages. IPD meta-analyses ease the standardization of the statistical analysis performed within each study [63, 69] and allow directly deriving the study-level information of interest [63]. Standardization includes consistency in dealing with missing data, consistency in modelling techniques used across studies, ability to control for inclusion and exclusion criteria, consistency 11

22 in outcome and exposure definitions and ability to adjust for the same set of covariates [63, 69], among others. In other words, the use of IPD controls for a part of the factors affecting between-study heterogeneity. Furthermore, IPD permit using information unavailable from publications [13]. In the context of diagnostic accuracy studies, IPD meta-analyses have the advantage to control for selective reporting of thresholds [43, 61]. Selective reporting of thresholds is a form of selection bias. It occurs because it is more likely that authors will report results from thresholds associated with higher sensitivity and/or specificity. A traditional meta-analysis that relies on results from these reported thresholds may produce biased (too high) pooled sensitivity and specificity due to this selective reporting [43, 61]. By using raw data, IPD meta-analyses rely on results from all thresholds and are not affected by selective reporting of thresholds. IPD meta-analyses also come with disadvantages [63]. Conducting an IPD meta-analysis can be much more time consuming and expensive than traditional meta-analyses. IPD meta-analyses require contacting study authors, obtaining their data, standardizing all acquired datasets and going through laborious ethic procedures. However, not all study authors are willing or able to share their raw data. This partial availability might introduce a bias in the IPD meta-analysis if it is associated with the study results [63, 69, 70]. Methods exist to meta-analyze aggregated data with IPD data [62, 65, 71] such that the potential partial availability bias can be counterbalanced. Also, the quality of an IPD meta-analysis still depends on the quality of included studies. An IPD meta-analysis based on poor quality primary studies may be as defective as a traditional meta-analysis based on these studies 12

23 [63]. Finally, IPD meta-analyses are still prone to publication bias since systematic reviews based on IPD are not meant to seek unpublished sources more than traditional meta-analyses [63, 70] Meta-analysis of diagnostic accuracy studies In the context of diagnostic accuracy studies, traditional and IPD meta-analyses present particular statistical challenges. Traditional meta-analytic methods of diagnostic accuracy studies depend on the type of published data available [24]. The most commonly reported measure is a pair of sensitivity and specificity of a test for a single threshold [24] and we will focus on this situation. In this context, traditional meta-analyses summarize pairs of sensitivity and specificity for that specific threshold across studies to produce a bivariate summary estimate. Such a meta-analysis should account for the correlation between sensitivity and specificity across studies. This correlation can be explained by a threshold effect [58, 60], which is closely related to clinical and methodological heterogeneity. The threshold effect can be explicit: if studies report results for different thresholds, everything else being kept equal, sensitivity and specificity are expected to be different [49]. However, if studies report results for a common threshold, implicit variations still arise due to differences in measurements, observers, equipments, laboratories or other methodological factors [49, 58]. Implicit variations are also a consequence of differences in study population say clinical heterogeneity [14, 49]. Unlike other sources of variation, the threshold effect has a specific effect on 13

24 the relationship between sensitivity and specificity across studies. Assume that an increasing score on a test is associated with a more severe disease. Lowering the threshold allows more patients with a positive result, and therefore catches more true positive patients but also more false positive patients. Lowering the threshold then results in increasing sensitivity at the expense of decreasing specificity [49, 58]. Similarly, making the threshold stricter allows fewer patients with a positive result and leads to increasing specificity at the expense of decreasing sensitivity. This tradeoff between sensitivity and specificity is fully captured in the ROC curve [49, 58]. Therefore, sensitivity and specificity will negatively correlated across studies since they mutually depend on a threshold [60]. This correlation between sensitivity and specificity across studies when considering a specific threshold is referred to as between-study correlation. Meta-analytic methods for such data should also consider the possible association between sensitivity and specificity within a study, say within-study correlation [59]. In the context of diagnostic accuracy when we focus on one specific threshold, it is reasonable to assume that within-study correlation is zero given that sensitivity and specificity estimates are derived from two independent populations, diseased and non-diseased patients, respectively [45, 59]. Simulation studies and empirical comparisons of traditional meta-analyses of diagnostic accuracy studies have shown that ignoring the within- and between-study correlation of sensitivity and specificity for a given threshold by analyzing them via two separate univariate meta-analyses is inappropriate [29, 49, 59]. The popular bivariate random-effects model is recommended in the current literature of traditional 14

25 meta-analyses methods of diagnostic accuracy studies [14, 24, 29, 50, 61]. This model will be described in detail in section 2.3. When IPD are available, there are two ways to conduct a meta-analysis of diagnostic accuracy [62]. First, one could focus on a particular threshold of interest. From IPD, it is possible to estimate sensitivity and specificity for that specific threshold for all studies and conduct a meta-analysis to obtain a pair of pooled sensitivity and specificity. This method is useful when there already exists a well-established threshold for this diagnostic test [62]. Second, IPD allow varying the threshold across all possible values of a test. For a continuous scale test, one could then estimate a ROC curve for each study and conduct a meta-analysis to obtain a pooled ROC curve [37]. For an ordinal scale test, varying the threshold allows estimating a pair of sensitivity and specificity for all possible thresholds [24]. A meta-analysis then produces pooled pairs of sensitivity and specificity for all possible thresholds. Methods to meta-analyze results from all thresholds should account for correlation between sensitivities and specificities across studies as well as their within study correlation. Between-study correlation of sensitivity and specificity still arise due to the effect of the threshold, as in traditional meta-analyses. This correlation should be taken into account by modelling sensitivities and specificities for all thresholds simultaneously. Given how sensitivities (specificities) are estimated, we also expect sensitivities (specificities) to be correlated across thresholds within a study [56]. In traditional meta-analyses, the within-study correlation only concerned the possible correlation 15

26 between sensitivity and specificity for a given threshold and this correlation was estimated by zero. However, when considering results from all thresholds, sensitivities (specificities) for different thresholds within a study are derived from the same observed results of the same diseased (healthy) population using different rules. Therefore, the within-study correlation between sensitivities for different thresholds is not zero anymore. Similarly, specificities for different thresholds within a study will be correlated. However, it is still reasonable to think that sensitivities are not correlated with specificities within a study because their estimates are derived from two independent groups of patients. Methods have been recently developed to meta-analyze results from multiple thresholds simultaneously. The ordinal multivariate random-effects model from Hamza et al. [24] is an extension of the bivariate random-effects model when results are available for the same set of thresholds across studies. This model will be described in Section 2.3. Another approach introduced by Riley et al. [66] considers the analysis of continuous scale tests when results are not reported for the same set of thresholds. This method is a missing data approach where results from missing thresholds are assumed to be missing at random i.e. the method is not designed to deal with selective reporting of thresholds. Other methods have been proposed [5, 15, 37, 55, 56]. We do not consider these methods for various reasons. Two of these methods [5, 55] do not directly concern diagnostic accuracy studies and are designed for diseases with more than two stages. Dukic et al. [15] develop a model in a bayesian framework and we do not consider this approach. Kester et al. s model [37] is similar to the ordinal multivariate random-effects model so we decide to consider the most recent 16

27 version of it. Putter et al. [56] propose a very interesting model using the framework of survival analysis. 2.3 Models IPD meta-analyses of diagnostic accuracy studies can be conducted via three methods. We will now thoroughly describe these methods Bivariate random-effects model The bivariate random-effects model(brem) was first introduced by Van Houwelingen et al. [75] in the context of meta-analyses of two-dimensional treatment groups in clinical trials. In the context of diagnostic accuracy studies, the BREM applies when each primary study reports a pair of sensitivity and specificity for a specific threshold J. Let the subscript i identify primary studies included in a meta-analysis, i = 1,...,I. The bivariate random-effects model was first described in the framework of linear mixed model (LMM) [58, 74]. The logit transformed sensitivities from primary studies are assumed to be approximately normally distributed around a pooled mean value with a certain amount of variability around this mean [58]. This is a randomeffects approach: differences in study population and any other implicit variations are incorporated in the model via the variability assumed around the pooled logit sensitivity. A similar reasoning is applied for modelling the specificities. The two normally distributed outcomes are modelled simultaneously to account for possible between-study and within-study correlations, leading to the bivariate normal 17

28 random-effects model. Define the parameters η ij as the true logit sensitivity and ξ ij as the true logit (1-specificity) for threshold J in study i. We assume that the true logit sensitivities from the primary studies follow a normal distribution centered around the pooled logit sensitivity η J with between-study variance ση. 2 Similarly, we assume that the true logit (1-specificities) from the primary studies follow a normal distribution centered around the pooled logit (1-specifity) ξ J with between-study variance σξ 2. Given the random effects u iηj and u iξj, the observed logit sensitivity ˆη ij and the observed logit (1-specificity) ˆξ ij in study i follow a bivariate normal distribution ˆξ ijξ ˆη ij N η ij,c i with C i = ξ ij s2 η J 0 0 s 2 ξ J where C i is the within-study covariance matrix. As mentioned before, it is assumed that the correlation between sensitivity and (1-specificity) within a study is zero given that they are estimated from two independent populations, which explains the off-diagonal terms in C i. Using the identity link, E(ˆη ij E(ˆξ ijξ u iηj ) = η ij = η J +u iηj u iξj ) = ξ ij = ξ J +u iξj where the random effects u iηj and u iξj are allowed to be correlated with bivariate normal distribution as 18

29 u iηj N 0, σ2 ηj u iξj 0 ρ J σ ηj σ ξj ρ J σ ηj σ ξj (2.1) σ 2 ξj We thus have σηj 2,σ2 ξj and ρ J denoting the between-study variability of logit sensitivity, the between-study variability of logit (1-specificity) and the between-study correlation of logit sensitivity and logit (1-specificity), respectively. Of interest is to estimate the pooled logit sensitivity η J and the pooled logit (1-specificity) ξ J along with their between-study variability and correlation. Chu and Cole [12] proposed to extend the normal bivariate model to accommodate non-continuous outcomes. Instead of assuming a bivariate normal distribution for the logit sensitivity and logit (1-specificity) within a study, they used the exact binomial distribution to model the number of true positive and true negatives patients in each study. Define TP ij and FP ij as the number of true positive and false positive patients, respectively, in study i when using threshold J. Define n i1 and n i0 as the total number of truly diseased and truly non-diseased patients in study i, respectively. Given the random effects u iη and u iξ, TP ij and FP ij follow two independent binomial distributions TP ij Binomial(n i1,expit(η ij )) FP ij Binomial(n i0,expit(ξ ij )) 19

30 and ( logit E ( logit E ( TPiJ n ( i1 FPiJ n i0 u iηj u iξj )) = η ij = η J +u iηj (2.2) )) = ξ ij = ξ J +u iξj where the random effects u iηj and u iξj come from a bivariate normal distribution with non-zero covariance as (2.1). Study-level covariates (e.g. type of population, mean age, type of diagnostic test, publication year) can be incorporated to the BREM model [12, 26, 58] by replacing η J and ξ J in (2.2) by η J = α 0 +β 1 X β p X p ξ J = α 0 +β 1 X β p X p where X 1,...,X p are covariates measured at the study-level. The linear combination can be adapted to the situation where a covariate is only available for the diseased or the non-diseased patients across studies or where a covariate is not available for a specific study [26]. The covariance matrix of the random effects (2.1) can also depend on the covariates [26]. Simulation studies [12, 27] and empirical comparisons [25, 27] have shown that the generalized bivariate random-effects model is preferable to the normal bivariate model. The normal bivariate random-effects model is biased because the binomial within-study exact distribution is approximated by the normal distribution [12, 27]. 20

31 This approximation is more problematic when sensitivity or specificity are close to one or zero [27]. Further references to the BREM will be for the model based on the binomial distribution. The BREM is traditionally used to meta-analyze pairs of sensitivity and specificity for a specific threshold. Methods have been developed to derive a summary ROC curve (sroc) from the pair of pooled estimates [3, 36, 49, 67]. These methods assume different parametrizations for the sroc curve and it is not clear which parametrization is preferable [3, 24]. For example, one could choose to regress ξ (logit(1-specificity)) on η (logit sensitivity). Since ξ and η are interchangeable, the regression line could as well be η on ξ [3, 24]. When the BREM is only applied to one threshold, the sroc curve make an assumption on the existence and on the shape of study specific ROC curves, which is untestable with published data only [24]. When IPD are available, the BREM may be applied threshold by threshold to produce a pooled pair of sensitivity and specificity for all possible thresholds. From these pooled estimates, a sroc curve can as well be derived but the choice of parametrization for the regression line remains a problem. However, the resulting sroc curve has the advantage of being identifiable since it relies on results from multiple thresholds [24]. Study- or patient-level covariates may be incorporated in thebremandthesemayleadtosroccurvesthatvarybylevelofthecovariate [26] Ordinal multivariate random-effects model The ordinal multivariate random-effects model (ordinal MREM) was introduced by Hamza et al. [24] precisely in the context of meta-analysis of diagnostic accuracy 21

32 studies. Their model is very similar to a method developed to meta-analyze ROC curves [37]. The ordinal MREM applies when all primary studies report pairs of sensitivity and specificity for the same set of multiple thresholds. This method is designed to support ordinal scale tests. Again, let the subscript i identify primary studies included in a meta-analysis, i = 1,...,I. The ordinal MREM is a direct extension of the exact binomial BREM when results are reported in more than one category [24]. Roughly, the model is based on the following lines of reasoning. Consider an ordinal test with J categories, corresponding to J-1 (relevant) thresholds. Within each study, the distribution of truly diseased patients across the test s categories follows a multinomial distribution. The corresponding probability parameters are functions of the J-1 true study-specific sensitivities. The random-effects approach assumes that these true sensitivities are drawn from a multivariate normal distribution centered around the J-1 pooled sensitivities. Reasoning is similar for the distribution of truly non-diseased patients (specificities) for the J-1 thresholds. The multinomial within-study modelling accounts for correlation across thresholds within study. Furthermore, sensitivities and specificities are modelled simultaneously, and thus the model correctly accounts for the correlation between sensitivity and specificity across studies. Let the subscript j identify the threshold, j = 1,...,J 1, for a test with J categories. Define η ij as the logit sensitivity and ξ ij as the logit (1-specificity) in study i with threshold j. Again, define n i1 and n i0 as the total number of truly diseased and truly non-diseased patients in study i, respectively. Finally, define x ij1 and x ij0 as the number of truly diseased and truly non-diseased patients in study 22

33 i with test results falling in category j, respectively. For a given threshold j, the observed sensitivity and specificity are respectively given by j k=1 x ik1 n i1 j k=1 and 1 x ik0 n i0 Given the random effects, the within-study model assumes that the number of diseasedsubjects(x i11,...,x ij1 )andthenumberofnon-diseasedsubjects(x i10,...,x ij0 ) follow two independent multinomial distributions with parameters (n i1 ;π i11,...,π ij1 ) and (n i0 ;π i10,...,π ij0 ), respectively. The probability parameters π ij1 are functions of the logit sensitivities η ij as ξ ij as π ij1 = exp(η ij ) 1+exp(η ij ), for j=1 exp(η ij ) 1+exp(η ij ) exp(η i,j 1) 1+exp(η i,j 1 ) 1 exp(η i,j 1) 1+exp(η i,j 1 ), for j=2,...,j-1, for j=j Similarly,theprobabilityparametersπ ij0 arefunctionsofthelogit(1-specificities) π ij0 = exp(ξ ij ) 1+exp(ξ ij ), for j=1 exp(ξ ij ) 1+exp(ξ ij ) exp(ξ i,j 1) 1+exp(ξ i,j 1 ) 1 exp(ξ i,j 1) 1+exp(ξ i,j 1 ), for j=2,...,j-1, for j=j The between-study model is specified in two steps. First, a random intercept/fixed slope linear relationship is assumed between ξ ij and η ij : η ij = α+u iα +βξ ij (2.3) 23

34 where the random effects u iα follow a normal distribution centered around zero with variance σα 2 and β is a fixed effect. It is also possible to introduce a random-effect term u iβ for the slope [24]. This linear relationship assumption is motivated by the necessity of producing a pooled ROC curve as well as simplifying the model [24]. Second, since test categories have a natural order, the proportional odds logit model [2, 5] is used to link ξ ij to the linear predictor as log ( j k=1 π ik1 J k=j+1 π ik1 i,δ ij ) = ξ ij = ξ j + i +δ ij (2.4) where i denotes the random effects associated to study clustering and δ ij denotes the random effects associated with the threshold value j, j = 1,...,J 1, within study [54]. It is assumed that the i follow a normal distribution as i N(0,σ 2). At this point, several assumptions are made on the distributions of the random effects u iα, i and δ ij in order to simplify the model. The i are allowed to be correlated with u iα, and their covariance is given by σ α. The δ ij s are assumed to be independent and identically distributed (i.i.d) with a normal distribution given by δ ij N(0,σδ 2). It is further assumed that the δ ijs are independent from i and u iα. The resulting multivariate normal distribution of the random effects is given by 24

35 u iα i +δ i1 i +δ i2. σα 2 σ α σ α σ α σ 2 +σ2 δ σ 2 σ2 N 0,. σ σ 2 i +δ i,j 1 σ α σ 2 σ 2 σ2 +σ2 δ Previous assumptions made on the random effects define a compound symmetric covariance structure. The δ ij being i.i.d., the variance of i + δ ij is constant for j = 1,...,J 1. Furthermore, since i and δ ij are assumed to be independent, we have Cov( i +δ ik, i +δ il ) = Var( i )+Cov( i,δ ik )+Cov( i,δ il )+Cov(δ il,δ ik ) = Var( i ) = σ 2 for l k, l,k = 1,2,...,J 1. As a practical interpretation, the compound symmetric structure assumes that the correlation between any pair of logit (1-specificities) ξ ik and ξ il is equal, l k, l,k = 1,2,...,J 1. This assumption might not be very appropriate since it is expected that the correlation between consecutive thresholds may be larger compared to non-consecutive thresholds [24]. Ofinterestistoestimatethefixedeffectsparametersα, β and ξ 1,..., ξ J 1 aswell as the variance and covariance of the random-effects terms σα, 2 σ 2, σ2 δ and σ α. From these estimates, one can derive estimates of the pooled logit sensitivities η j = α+β ξ j 25

36 along with their standard errors. A sroc curve is directly derived from the model estimates α and β. The ordinal MREM also easily accommodates study-level covariates [24]. Studyorpatient-levelcovariatescanbeincorporatedbyreplacingη ij in(2.3)andξ ij in(2.4) by η ij = α i +u iα +βξ ij +β 1 X 1 + +β p X p ξ ij = β 1 X 1 + +β p X p + i +δ ij where X 1,...,X p are covariates measured at the study- or patient-level. One could also decide to only adjust η ij and test the interaction between the covariates and ξ ij [24]. Again, different levels of a covariate will produce different sroc curves [24] Normal multivariate random-effects model The normal multivariate random-effects model (normal MREM) was introduced by Riley et al. [66] in the context of meta-analysis of diagnostic accuracy studies. The model applies when primary studies report pairs of sensitivity and specificity for multiple thresholds, but not necessarily for the same sets of thresholds. It can be viewed as a missing data approach. The model will be described in the context of IPD, that is when results from all thresholds are available in all primary studies. Ordinal tests with a finite number of thresholds will be considered. Imputation of missing data is out of the scope of this project and is discussed elsewhere 26

37 [61, 66]. Again, let the subscript i identify primary studies included in a metaanalysis, i = 1,...,I, and j identify the threshold value, j = 1,...,J 1. The model is a direct extension of the normal bivariate random-effects model [58, 75] just as the ordinal MREM was a direct extension of the binomial bivariate random-effects model. The model is based on the following lines of reasoning. Within each primary study, the logit sensitivities and the logit (1-specificities) for all J-1 thresholds are assumed to be normally distributed around the true study-specific logit sensitivities η ij and logit (1-specificity) ξ ij. Following the framework of linear mixed model, these true logit sensitivities and logit (1-specificities) are assumed to be drawn from a normal distribution centered around the pooled logit sensitivities η j and pooled logit (1-specificities) ξ j, for j = 1,...,J 1. Define x ij1 and x ij0 as the number of truly diseased and truly non-diseased patients in study i with test results falling in category j, respectively. Within each study, the number of diseased subjects (x i11,...,x ij1 ) and the number of non-diseased subjects(x i10,...,x ij0 )follow twoindependentmultinomialdistributionswithparameters (n i1 ;π i11,...,π ij1 ) and (n i0 ;π i10,...,π ij0 ), respectively. At this point, instead of using the framework of generalized linear mixed model (GLMM), the authors suggest to estimate the probability parameters of the two distributions by maximum likelihood. The multinomial likelihoods are given by 27

38 ( ) J 1 xij1 l i (1) (π i11 ) x i11 (π i21 ) x i21 (π i(j 1)1 ) x i(j 1)1 1 π ik1 ( ) J 1 xij0 l i (0) (π i10 ) x i10 (π i20 ) x i20 (π i(j 1)0 ) x i(j 1)0 1 π ik0 Maximization of the log-likelihood can easily be done with any statistical software [66]. From the estimates of the π ij1 s and π ij0 s, we can subsequently derive estimates of the logit sensitivity ˆη ij at threshold j in study i by ( ) j ˆη ij = logit 1 ˆπ ik1 and, similarly, estimates of the logit (1-specificity) at threshold j in study i by ( j ) ˆξ ij = logit ˆπ ik0. Estimates of the variance of the logit sensitivities s 2 ij1 and of their covariance s i(l,k)1, l k,l,k = 1,...,J 1, can be derived from estimates of the variance and covariance of the π ij1 s, for example using the delta method [8, 66]. Similarly, estimates of the variance of the logit (1-specificities) s 2 ij0 and of their covariance s i(l,k)0, l k,l,k = 1,...,J 1, can be derived from estimates of the variance and covariance of the π ij0 s. It is impossible to obtain an estimate of π ij1 (or π ij0 ) if there are no truly diseased (or truly non-diseased) patients with test result in category j [66]. The logit sensitivity (or logit (1-specificity)) estimated with that threshold j will then be equal to the estimated logit sensitivity (or 1-specificity) from the previous threshold k=1 k=1 k=1 k=1 28

An Empirical Assessment of Bivariate Methods for Meta-analysis of Test Accuracy

An Empirical Assessment of Bivariate Methods for Meta-analysis of Test Accuracy Number XX An Empirical Assessment of Bivariate Methods for Meta-analysis of Test Accuracy Prepared for: Agency for Healthcare Research and Quality U.S. Department of Health and Human Services 54 Gaither

More information

Meta-analysis of diagnostic test accuracy studies with multiple & missing thresholds

Meta-analysis of diagnostic test accuracy studies with multiple & missing thresholds Meta-analysis of diagnostic test accuracy studies with multiple & missing thresholds Richard D. Riley School of Health and Population Sciences, & School of Mathematics, University of Birmingham Collaborators:

More information

ON THE NUMBER OF PERCEIVERS IN A TRIANGLE TEST WITH REPLICATIONS

ON THE NUMBER OF PERCEIVERS IN A TRIANGLE TEST WITH REPLICATIONS ON THE NUMBER OF PERCEIVERS IN A TRIANGLE TEST WITH REPLICATIONS Michael Meyners Fachbereich Statistik, Universität Dortmund, D-4422 Dortmund, Germany E-mail: michael.meyners@udo.edu Fax: +49 (0)2 755

More information

Introduction to diagnostic accuracy meta-analysis. Yemisi Takwoingi October 2015

Introduction to diagnostic accuracy meta-analysis. Yemisi Takwoingi October 2015 Introduction to diagnostic accuracy meta-analysis Yemisi Takwoingi October 2015 Learning objectives To appreciate the concept underlying DTA meta-analytic approaches To know the Moses-Littenberg SROC method

More information

Methods Research Report. An Empirical Assessment of Bivariate Methods for Meta-Analysis of Test Accuracy

Methods Research Report. An Empirical Assessment of Bivariate Methods for Meta-Analysis of Test Accuracy Methods Research Report An Empirical Assessment of Bivariate Methods for Meta-Analysis of Test Accuracy Methods Research Report An Empirical Assessment of Bivariate Methods for Meta-Analysis of Test Accuracy

More information

Empirical assessment of univariate and bivariate meta-analyses for comparing the accuracy of diagnostic tests

Empirical assessment of univariate and bivariate meta-analyses for comparing the accuracy of diagnostic tests Empirical assessment of univariate and bivariate meta-analyses for comparing the accuracy of diagnostic tests Yemisi Takwoingi, Richard Riley and Jon Deeks Outline Rationale Methods Findings Summary Motivating

More information

Statistical methods for the meta-analysis of full ROC curves

Statistical methods for the meta-analysis of full ROC curves Statistical methods for the meta-analysis of full ROC curves Oliver Kuss (joint work with Stefan Hirt and Annika Hoyer) German Diabetes Center, Leibniz Institute for Diabetes Research at Heinrich Heine

More information

Multivariate Mixed-Effects Meta-Analysis of Paired-Comparison Studies of Diagnostic Test Accuracy

Multivariate Mixed-Effects Meta-Analysis of Paired-Comparison Studies of Diagnostic Test Accuracy Multivariate Mixed-Effects Meta-Analysis of Paired-Comparison Studies of Diagnostic Test Accuracy Ben A. Dwamena, MD The University of Michigan & VA Medical Centers, Ann Arbor SNASUG - July 24, 2008 Diagnostic

More information

Unit 1 Exploring and Understanding Data

Unit 1 Exploring and Understanding Data Unit 1 Exploring and Understanding Data Area Principle Bar Chart Boxplot Conditional Distribution Dotplot Empirical Rule Five Number Summary Frequency Distribution Frequency Polygon Histogram Interquartile

More information

Introduction to Meta-analysis of Accuracy Data

Introduction to Meta-analysis of Accuracy Data Introduction to Meta-analysis of Accuracy Data Hans Reitsma MD, PhD Dept. of Clinical Epidemiology, Biostatistics & Bioinformatics Academic Medical Center - Amsterdam Continental European Support Unit

More information

Meta-analysis of Diagnostic Test Accuracy Studies

Meta-analysis of Diagnostic Test Accuracy Studies GUIDELINE Meta-analysis of Diagnostic Test Accuracy Studies November 2014 Copyright EUnetHTA 2013. All Rights Reserved. No part of this document may be reproduced without an explicit acknowledgement of

More information

GUIDELINE COMPARATORS & COMPARISONS:

GUIDELINE COMPARATORS & COMPARISONS: GUIDELINE COMPARATORS & COMPARISONS: Direct and indirect comparisons Adapted version (2015) based on COMPARATORS & COMPARISONS: Direct and indirect comparisons - February 2013 The primary objective of

More information

Biometrical Journal 52 (2010) 61, zzz zzz / DOI: /bimj

Biometrical Journal 52 (2010) 61, zzz zzz / DOI: /bimj Biometrical Journal 52 (2010) 61, zzz zzz / DOI: 10.1002/bimj.200100000 A comparison of bivariate, multivariate random-effects, and Poisson correlated gamma-frailty models to meta-analyze individual patient

More information

STATISTICAL METHODS FOR DIAGNOSTIC TESTING: AN ILLUSTRATION USING A NEW METHOD FOR CANCER DETECTION XIN SUN. PhD, Kansas State University, 2012

STATISTICAL METHODS FOR DIAGNOSTIC TESTING: AN ILLUSTRATION USING A NEW METHOD FOR CANCER DETECTION XIN SUN. PhD, Kansas State University, 2012 STATISTICAL METHODS FOR DIAGNOSTIC TESTING: AN ILLUSTRATION USING A NEW METHOD FOR CANCER DETECTION by XIN SUN PhD, Kansas State University, 2012 A THESIS Submitted in partial fulfillment of the requirements

More information

Bayesian and Frequentist Approaches

Bayesian and Frequentist Approaches Bayesian and Frequentist Approaches G. Jogesh Babu Penn State University http://sites.stat.psu.edu/ babu http://astrostatistics.psu.edu All models are wrong But some are useful George E. P. Box (son-in-law

More information

Systematic Reviews and meta-analyses of Diagnostic Test Accuracy. Mariska Leeflang

Systematic Reviews and meta-analyses of Diagnostic Test Accuracy. Mariska Leeflang Systematic Reviews and meta-analyses of Diagnostic Test Accuracy Mariska Leeflang m.m.leeflang@amc.uva.nl This presentation 1. Introduction: accuracy? 2. QUADAS-2 exercise 3. Meta-analysis of diagnostic

More information

Clinical Epidemiology for the uninitiated

Clinical Epidemiology for the uninitiated Clinical epidemiologist have one foot in clinical care and the other in clinical practice research. As clinical epidemiologists we apply a wide array of scientific principles, strategies and tactics to

More information

Statistical methods for the meta-analysis of full ROC curves

Statistical methods for the meta-analysis of full ROC curves Statistical methods for the meta-analysis of full ROC curves Oliver Kuss (joint work with Stefan Hirt and Annika Hoyer) German Diabetes Center, Leibniz Institute for Diabetes Research at Heinrich Heine

More information

Biostatistics II

Biostatistics II Biostatistics II 514-5509 Course Description: Modern multivariable statistical analysis based on the concept of generalized linear models. Includes linear, logistic, and Poisson regression, survival analysis,

More information

SYSTEMATIC REVIEWS OF TEST ACCURACY STUDIES

SYSTEMATIC REVIEWS OF TEST ACCURACY STUDIES Biomarker & Test Evaluation Program SYSTEMATIC REVIEWS OF TEST ACCURACY STUDIES Patrick MM Bossuyt Structure 1. Clinical Scenarios 2. Test Accuracy Studies 3. Systematic Reviews 4. Meta-Analysis 5.

More information

Meta-analysis using individual participant data: one-stage and two-stage approaches, and why they may differ

Meta-analysis using individual participant data: one-stage and two-stage approaches, and why they may differ Tutorial in Biostatistics Received: 11 March 2016, Accepted: 13 September 2016 Published online 16 October 2016 in Wiley Online Library (wileyonlinelibrary.com) DOI: 10.1002/sim.7141 Meta-analysis using

More information

SISCR Module 7 Part I: Introduction Basic Concepts for Binary Biomarkers (Classifiers) and Continuous Biomarkers

SISCR Module 7 Part I: Introduction Basic Concepts for Binary Biomarkers (Classifiers) and Continuous Biomarkers SISCR Module 7 Part I: Introduction Basic Concepts for Binary Biomarkers (Classifiers) and Continuous Biomarkers Kathleen Kerr, Ph.D. Associate Professor Department of Biostatistics University of Washington

More information

Bayesian meta-analysis of Papanicolaou smear accuracy

Bayesian meta-analysis of Papanicolaou smear accuracy Gynecologic Oncology 107 (2007) S133 S137 www.elsevier.com/locate/ygyno Bayesian meta-analysis of Papanicolaou smear accuracy Xiuyu Cong a, Dennis D. Cox b, Scott B. Cantor c, a Biometrics and Data Management,

More information

Bias in regression coefficient estimates when assumptions for handling missing data are violated: a simulation study

Bias in regression coefficient estimates when assumptions for handling missing data are violated: a simulation study STATISTICAL METHODS Epidemiology Biostatistics and Public Health - 2016, Volume 13, Number 1 Bias in regression coefficient estimates when assumptions for handling missing data are violated: a simulation

More information

Lecture Outline. Biost 517 Applied Biostatistics I. Purpose of Descriptive Statistics. Purpose of Descriptive Statistics

Lecture Outline. Biost 517 Applied Biostatistics I. Purpose of Descriptive Statistics. Purpose of Descriptive Statistics Biost 517 Applied Biostatistics I Scott S. Emerson, M.D., Ph.D. Professor of Biostatistics University of Washington Lecture 3: Overview of Descriptive Statistics October 3, 2005 Lecture Outline Purpose

More information

Lecture Outline. Biost 590: Statistical Consulting. Stages of Scientific Studies. Scientific Method

Lecture Outline. Biost 590: Statistical Consulting. Stages of Scientific Studies. Scientific Method Biost 590: Statistical Consulting Statistical Classification of Scientific Studies; Approach to Consulting Lecture Outline Statistical Classification of Scientific Studies Statistical Tasks Approach to

More information

METHODS FOR DETECTING CERVICAL CANCER

METHODS FOR DETECTING CERVICAL CANCER Chapter III METHODS FOR DETECTING CERVICAL CANCER 3.1 INTRODUCTION The successful detection of cervical cancer in a variety of tissues has been reported by many researchers and baseline figures for the

More information

Module Overview. What is a Marker? Part 1 Overview

Module Overview. What is a Marker? Part 1 Overview SISCR Module 7 Part I: Introduction Basic Concepts for Binary Classification Tools and Continuous Biomarkers Kathleen Kerr, Ph.D. Associate Professor Department of Biostatistics University of Washington

More information

Cochrane Handbook for Systematic Reviews of Diagnostic Test Accuracy

Cochrane Handbook for Systematic Reviews of Diagnostic Test Accuracy Cochrane Handbook for Systematic Reviews of Diagnostic Test Accuracy Chapter 10 Analysing and Presenting Results Petra Macaskill, Constantine Gatsonis, Jonathan Deeks, Roger Harbord, Yemisi Takwoingi.

More information

Multivariate Mixed-Effects Meta-Analysis of Paired-Comparison Studies of Diagnostic Test Accuracy

Multivariate Mixed-Effects Meta-Analysis of Paired-Comparison Studies of Diagnostic Test Accuracy Multivariate Mixed-Effects Meta-Analysis of Paired-Comparison Studies of Diagnostic Test Accuracy Ben A. Dwamena, MD The University of Michigan & VA Medical Centers, Ann Arbor SNASUG - July 24, 2008 B.A.

More information

How to analyze correlated and longitudinal data?

How to analyze correlated and longitudinal data? How to analyze correlated and longitudinal data? Niloofar Ramezani, University of Northern Colorado, Greeley, Colorado ABSTRACT Longitudinal and correlated data are extensively used across disciplines

More information

Package xmeta. August 16, 2017

Package xmeta. August 16, 2017 Type Package Title A Toolbox for Multivariate Meta-Analysis Version 1.1-4 Date 2017-5-12 Package xmeta August 16, 2017 Imports aod, glmmml, numderiv, metafor, mvmeta, stats A toolbox for meta-analysis.

More information

MEA DISCUSSION PAPERS

MEA DISCUSSION PAPERS Inference Problems under a Special Form of Heteroskedasticity Helmut Farbmacher, Heinrich Kögel 03-2015 MEA DISCUSSION PAPERS mea Amalienstr. 33_D-80799 Munich_Phone+49 89 38602-355_Fax +49 89 38602-390_www.mea.mpisoc.mpg.de

More information

Introduction to ROC analysis

Introduction to ROC analysis Introduction to ROC analysis Andriy I. Bandos Department of Biostatistics University of Pittsburgh Acknowledgements Many thanks to Sam Wieand, Nancy Obuchowski, Brenda Kurland, and Todd Alonzo for previous

More information

Fixed Effect Combining

Fixed Effect Combining Meta-Analysis Workshop (part 2) Michael LaValley December 12 th 2014 Villanova University Fixed Effect Combining Each study i provides an effect size estimate d i of the population value For the inverse

More information

Psychology, 2010, 1: doi: /psych Published Online August 2010 (

Psychology, 2010, 1: doi: /psych Published Online August 2010 ( Psychology, 2010, 1: 194-198 doi:10.4236/psych.2010.13026 Published Online August 2010 (http://www.scirp.org/journal/psych) Using Generalizability Theory to Evaluate the Applicability of a Serial Bayes

More information

Bayesian graphical models for combining multiple data sources, with applications in environmental epidemiology

Bayesian graphical models for combining multiple data sources, with applications in environmental epidemiology Bayesian graphical models for combining multiple data sources, with applications in environmental epidemiology Sylvia Richardson 1 sylvia.richardson@imperial.co.uk Joint work with: Alexina Mason 1, Lawrence

More information

Treatment effect estimates adjusted for small-study effects via a limit meta-analysis

Treatment effect estimates adjusted for small-study effects via a limit meta-analysis Treatment effect estimates adjusted for small-study effects via a limit meta-analysis Gerta Rücker 1, James Carpenter 12, Guido Schwarzer 1 1 Institute of Medical Biometry and Medical Informatics, University

More information

Adjusting for mode of administration effect in surveys using mailed questionnaire and telephone interview data

Adjusting for mode of administration effect in surveys using mailed questionnaire and telephone interview data Adjusting for mode of administration effect in surveys using mailed questionnaire and telephone interview data Karl Bang Christensen National Institute of Occupational Health, Denmark Helene Feveille National

More information

MS&E 226: Small Data

MS&E 226: Small Data MS&E 226: Small Data Lecture 10: Introduction to inference (v2) Ramesh Johari ramesh.johari@stanford.edu 1 / 17 What is inference? 2 / 17 Where did our data come from? Recall our sample is: Y, the vector

More information

A Multistate Frailty Model for Sequential Events from a Family-based Study

A Multistate Frailty Model for Sequential Events from a Family-based Study A Multistate Frailty Model for Sequential Events from a Family-based Study Yun-Hee Choi 1 & Balakumar Swaminathan 2 1 Department of Epidemiology and Biostatistics, Western University, London, Ontario,

More information

Analyzing diastolic and systolic blood pressure individually or jointly?

Analyzing diastolic and systolic blood pressure individually or jointly? Analyzing diastolic and systolic blood pressure individually or jointly? Chenglin Ye a, Gary Foster a, Lisa Dolovich b, Lehana Thabane a,c a. Department of Clinical Epidemiology and Biostatistics, McMaster

More information

1 Introduction. st0020. The Stata Journal (2002) 2, Number 3, pp

1 Introduction. st0020. The Stata Journal (2002) 2, Number 3, pp The Stata Journal (22) 2, Number 3, pp. 28 289 Comparative assessment of three common algorithms for estimating the variance of the area under the nonparametric receiver operating characteristic curve

More information

Design and Analysis Plan Quantitative Synthesis of Federally-Funded Teen Pregnancy Prevention Programs HHS Contract #HHSP I 5/2/2016

Design and Analysis Plan Quantitative Synthesis of Federally-Funded Teen Pregnancy Prevention Programs HHS Contract #HHSP I 5/2/2016 Design and Analysis Plan Quantitative Synthesis of Federally-Funded Teen Pregnancy Prevention Programs HHS Contract #HHSP233201500069I 5/2/2016 Overview The goal of the meta-analysis is to assess the effects

More information

Technical Specifications

Technical Specifications Technical Specifications In order to provide summary information across a set of exercises, all tests must employ some form of scoring models. The most familiar of these scoring models is the one typically

More information

EXERCISE: HOW TO DO POWER CALCULATIONS IN OPTIMAL DESIGN SOFTWARE

EXERCISE: HOW TO DO POWER CALCULATIONS IN OPTIMAL DESIGN SOFTWARE ...... EXERCISE: HOW TO DO POWER CALCULATIONS IN OPTIMAL DESIGN SOFTWARE TABLE OF CONTENTS 73TKey Vocabulary37T... 1 73TIntroduction37T... 73TUsing the Optimal Design Software37T... 73TEstimating Sample

More information

For general queries, contact

For general queries, contact Much of the work in Bayesian econometrics has focused on showing the value of Bayesian methods for parametric models (see, for example, Geweke (2005), Koop (2003), Li and Tobias (2011), and Rossi, Allenby,

More information

Fundamental Clinical Trial Design

Fundamental Clinical Trial Design Design, Monitoring, and Analysis of Clinical Trials Session 1 Overview and Introduction Overview Scott S. Emerson, M.D., Ph.D. Professor of Biostatistics, University of Washington February 17-19, 2003

More information

CHAPTER VI RESEARCH METHODOLOGY

CHAPTER VI RESEARCH METHODOLOGY CHAPTER VI RESEARCH METHODOLOGY 6.1 Research Design Research is an organized, systematic, data based, critical, objective, scientific inquiry or investigation into a specific problem, undertaken with the

More information

Various performance measures in Binary classification An Overview of ROC study

Various performance measures in Binary classification An Overview of ROC study Various performance measures in Binary classification An Overview of ROC study Suresh Babu. Nellore Department of Statistics, S.V. University, Tirupati, India E-mail: sureshbabu.nellore@gmail.com Abstract

More information

Introduction. We can make a prediction about Y i based on X i by setting a threshold value T, and predicting Y i = 1 when X i > T.

Introduction. We can make a prediction about Y i based on X i by setting a threshold value T, and predicting Y i = 1 when X i > T. Diagnostic Tests 1 Introduction Suppose we have a quantitative measurement X i on experimental or observed units i = 1,..., n, and a characteristic Y i = 0 or Y i = 1 (e.g. case/control status). The measurement

More information

Advanced IPD meta-analysis methods for observational studies

Advanced IPD meta-analysis methods for observational studies Advanced IPD meta-analysis methods for observational studies Simon Thompson University of Cambridge, UK Part 4 IBC Victoria, July 2016 1 Outline of talk Usual measures of association (e.g. hazard ratios)

More information

Meta-Analysis. Zifei Liu. Biological and Agricultural Engineering

Meta-Analysis. Zifei Liu. Biological and Agricultural Engineering Meta-Analysis Zifei Liu What is a meta-analysis; why perform a metaanalysis? How a meta-analysis work some basic concepts and principles Steps of Meta-analysis Cautions on meta-analysis 2 What is Meta-analysis

More information

Methods for meta-analysis of individual participant data from Mendelian randomization studies with binary outcomes

Methods for meta-analysis of individual participant data from Mendelian randomization studies with binary outcomes Methods for meta-analysis of individual participant data from Mendelian randomization studies with binary outcomes Stephen Burgess Simon G. Thompson CRP CHD Genetics Collaboration May 24, 2012 Abstract

More information

Practical Bayesian Design and Analysis for Drug and Device Clinical Trials

Practical Bayesian Design and Analysis for Drug and Device Clinical Trials Practical Bayesian Design and Analysis for Drug and Device Clinical Trials p. 1/2 Practical Bayesian Design and Analysis for Drug and Device Clinical Trials Brian P. Hobbs Plan B Advisor: Bradley P. Carlin

More information

Selection and Combination of Markers for Prediction

Selection and Combination of Markers for Prediction Selection and Combination of Markers for Prediction NACC Data and Methods Meeting September, 2010 Baojiang Chen, PhD Sarah Monsell, MS Xiao-Hua Andrew Zhou, PhD Overview 1. Research motivation 2. Describe

More information

Analysis of Hearing Loss Data using Correlated Data Analysis Techniques

Analysis of Hearing Loss Data using Correlated Data Analysis Techniques Analysis of Hearing Loss Data using Correlated Data Analysis Techniques Ruth Penman and Gillian Heller, Department of Statistics, Macquarie University, Sydney, Australia. Correspondence: Ruth Penman, Department

More information

ISPOR Task Force Report: ITC & NMA Study Questionnaire

ISPOR Task Force Report: ITC & NMA Study Questionnaire INDIRECT TREATMENT COMPARISON / NETWORK META-ANALYSIS STUDY QUESTIONNAIRE TO ASSESS RELEVANCE AND CREDIBILITY TO INFORM HEALTHCARE DECISION-MAKING: AN ISPOR-AMCP-NPC GOOD PRACTICE TASK FORCE REPORT DRAFT

More information

An Introduction to Bayesian Statistics

An Introduction to Bayesian Statistics An Introduction to Bayesian Statistics Robert Weiss Department of Biostatistics UCLA Fielding School of Public Health robweiss@ucla.edu Sept 2015 Robert Weiss (UCLA) An Introduction to Bayesian Statistics

More information

Introduction to Bayesian Analysis 1

Introduction to Bayesian Analysis 1 Biostats VHM 801/802 Courses Fall 2005, Atlantic Veterinary College, PEI Henrik Stryhn Introduction to Bayesian Analysis 1 Little known outside the statistical science, there exist two different approaches

More information

Meta-analyses evaluating diagnostic test accuracy

Meta-analyses evaluating diagnostic test accuracy THE STATISTICIAN S PAGE Summary Receiver Operating Characteristic Curve Analysis Techniques in the Evaluation of Diagnostic Tests Catherine M. Jones, MBBS, BSc(Stat), and Thanos Athanasiou, MD, PhD, FETCS

More information

Using Standard and Asymmetric Confidence Intervals. Utilisation d intervalles de confiance standards et asymétriques. Christopher J.

Using Standard and Asymmetric Confidence Intervals. Utilisation d intervalles de confiance standards et asymétriques. Christopher J. Using Standard and Asymmetric Confidence Intervals Utilisation d intervalles de confiance standards et asymétriques KEY WORDS CONFIDENCE INTERVAL MEASUREMENT ERROR CONSISTENCY COEFFICIENT AGREEMENT COEFFICIENT

More information

Live WebEx meeting agenda

Live WebEx meeting agenda 10:00am 10:30am Using OpenMeta[Analyst] to extract quantitative data from published literature Live WebEx meeting agenda August 25, 10:00am-12:00pm ET 10:30am 11:20am Lecture (this will be recorded) 11:20am

More information

Meta-analysis of diagnostic accuracy studies. Mariska Leeflang (with thanks to Yemisi Takwoingi, Jon Deeks and Hans Reitsma)

Meta-analysis of diagnostic accuracy studies. Mariska Leeflang (with thanks to Yemisi Takwoingi, Jon Deeks and Hans Reitsma) Meta-analysis of diagnostic accuracy studies Mariska Leeflang (with thanks to Yemisi Takwoingi, Jon Deeks and Hans Reitsma) 1 Diagnostic Test Accuracy Reviews 1. Framing the question 2. Identification

More information

Chapter 02. Basic Research Methodology

Chapter 02. Basic Research Methodology Chapter 02 Basic Research Methodology Definition RESEARCH Research is a quest for knowledge through diligent search or investigation or experimentation aimed at the discovery and interpretation of new

More information

A review of statistical methods in the analysis of data arising from observer reliability studies (Part 11) *

A review of statistical methods in the analysis of data arising from observer reliability studies (Part 11) * A review of statistical methods in the analysis of data arising from observer reliability studies (Part 11) * by J. RICHARD LANDIS** and GARY G. KOCH** 4 Methods proposed for nominal and ordinal data Many

More information

Received: 14 April 2016, Accepted: 28 October 2016 Published online 1 December 2016 in Wiley Online Library

Received: 14 April 2016, Accepted: 28 October 2016 Published online 1 December 2016 in Wiley Online Library Research Article Received: 14 April 2016, Accepted: 28 October 2016 Published online 1 December 2016 in Wiley Online Library (wileyonlinelibrary.com) DOI: 10.1002/sim.7171 One-stage individual participant

More information

Chapter 17 Sensitivity Analysis and Model Validation

Chapter 17 Sensitivity Analysis and Model Validation Chapter 17 Sensitivity Analysis and Model Validation Justin D. Salciccioli, Yves Crutain, Matthieu Komorowski and Dominic C. Marshall Learning Objectives Appreciate that all models possess inherent limitations

More information

Structural Approach to Bias in Meta-analyses

Structural Approach to Bias in Meta-analyses Original Article Received 26 July 2011, Revised 22 November 2011, Accepted 12 December 2011 Published online 2 February 2012 in Wiley Online Library (wileyonlinelibrary.com) DOI: 10.1002/jrsm.52 Structural

More information

The Effect of Extremes in Small Sample Size on Simple Mixed Models: A Comparison of Level-1 and Level-2 Size

The Effect of Extremes in Small Sample Size on Simple Mixed Models: A Comparison of Level-1 and Level-2 Size INSTITUTE FOR DEFENSE ANALYSES The Effect of Extremes in Small Sample Size on Simple Mixed Models: A Comparison of Level-1 and Level-2 Size Jane Pinelis, Project Leader February 26, 2018 Approved for public

More information

Review. Imagine the following table being obtained as a random. Decision Test Diseased Not Diseased Positive TP FP Negative FN TN

Review. Imagine the following table being obtained as a random. Decision Test Diseased Not Diseased Positive TP FP Negative FN TN Outline 1. Review sensitivity and specificity 2. Define an ROC curve 3. Define AUC 4. Non-parametric tests for whether or not the test is informative 5. Introduce the binormal ROC model 6. Discuss non-parametric

More information

Lec 02: Estimation & Hypothesis Testing in Animal Ecology

Lec 02: Estimation & Hypothesis Testing in Animal Ecology Lec 02: Estimation & Hypothesis Testing in Animal Ecology Parameter Estimation from Samples Samples We typically observe systems incompletely, i.e., we sample according to a designed protocol. We then

More information

Meta-analysis of clinical prediction models

Meta-analysis of clinical prediction models Meta-analysis of clinical prediction models Thomas P. A. Debray Julius Center for Health Sciences and Primary Care Utrecht University Meta-analysis of clinical prediction models Julius Center for Health

More information

Generalized Estimating Equations for Depression Dose Regimes

Generalized Estimating Equations for Depression Dose Regimes Generalized Estimating Equations for Depression Dose Regimes Karen Walker, Walker Consulting LLC, Menifee CA Generalized Estimating Equations on the average produce consistent estimates of the regression

More information

An Introduction to Multilevel Regression Models

An Introduction to Multilevel Regression Models A B S T R A C T Data in health research are frequently structured hierarchically. For example, data may consist of patients nested within physicians, who in turn may be nested in hospitals or geographic

More information

Discussion Meeting for MCP-Mod Qualification Opinion Request. Novartis 10 July 2013 EMA, London, UK

Discussion Meeting for MCP-Mod Qualification Opinion Request. Novartis 10 July 2013 EMA, London, UK Discussion Meeting for MCP-Mod Qualification Opinion Request Novartis 10 July 2013 EMA, London, UK Attendees Face to face: Dr. Frank Bretz Global Statistical Methodology Head, Novartis Dr. Björn Bornkamp

More information

Regression Discontinuity Analysis

Regression Discontinuity Analysis Regression Discontinuity Analysis A researcher wants to determine whether tutoring underachieving middle school students improves their math grades. Another wonders whether providing financial aid to low-income

More information

EPSE 594: Meta-Analysis: Quantitative Research Synthesis

EPSE 594: Meta-Analysis: Quantitative Research Synthesis EPSE 594: Meta-Analysis: Quantitative Research Synthesis Ed Kroc University of British Columbia ed.kroc@ubc.ca March 28, 2019 Ed Kroc (UBC) EPSE 594 March 28, 2019 1 / 32 Last Time Publication bias Funnel

More information

Catherine A. Welch 1*, Séverine Sabia 1,2, Eric Brunner 1, Mika Kivimäki 1 and Martin J. Shipley 1

Catherine A. Welch 1*, Séverine Sabia 1,2, Eric Brunner 1, Mika Kivimäki 1 and Martin J. Shipley 1 Welch et al. BMC Medical Research Methodology (2018) 18:89 https://doi.org/10.1186/s12874-018-0548-0 RESEARCH ARTICLE Open Access Does pattern mixture modelling reduce bias due to informative attrition

More information

Multilevel analysis quantifies variation in the experimental effect while optimizing power and preventing false positives

Multilevel analysis quantifies variation in the experimental effect while optimizing power and preventing false positives DOI 10.1186/s12868-015-0228-5 BMC Neuroscience RESEARCH ARTICLE Open Access Multilevel analysis quantifies variation in the experimental effect while optimizing power and preventing false positives Emmeke

More information

Analysis of Rheumatoid Arthritis Data using Logistic Regression and Penalized Approach

Analysis of Rheumatoid Arthritis Data using Logistic Regression and Penalized Approach University of South Florida Scholar Commons Graduate Theses and Dissertations Graduate School November 2015 Analysis of Rheumatoid Arthritis Data using Logistic Regression and Penalized Approach Wei Chen

More information

A Brief Introduction to Bayesian Statistics

A Brief Introduction to Bayesian Statistics A Brief Introduction to Statistics David Kaplan Department of Educational Psychology Methods for Social Policy Research and, Washington, DC 2017 1 / 37 The Reverend Thomas Bayes, 1701 1761 2 / 37 Pierre-Simon

More information

Propensity Score Methods for Estimating Causality in the Absence of Random Assignment: Applications for Child Care Policy Research

Propensity Score Methods for Estimating Causality in the Absence of Random Assignment: Applications for Child Care Policy Research 2012 CCPRC Meeting Methodology Presession Workshop October 23, 2012, 2:00-5:00 p.m. Propensity Score Methods for Estimating Causality in the Absence of Random Assignment: Applications for Child Care Policy

More information

Analysis of Vaccine Effects on Post-Infection Endpoints Biostat 578A Lecture 3

Analysis of Vaccine Effects on Post-Infection Endpoints Biostat 578A Lecture 3 Analysis of Vaccine Effects on Post-Infection Endpoints Biostat 578A Lecture 3 Analysis of Vaccine Effects on Post-Infection Endpoints p.1/40 Data Collected in Phase IIb/III Vaccine Trial Longitudinal

More information

Title: A new statistical test for trends: establishing the properties of a test for repeated binomial observations on a set of items

Title: A new statistical test for trends: establishing the properties of a test for repeated binomial observations on a set of items Title: A new statistical test for trends: establishing the properties of a test for repeated binomial observations on a set of items Introduction Many studies of therapies with single subjects involve

More information

BMI 541/699 Lecture 16

BMI 541/699 Lecture 16 BMI 541/699 Lecture 16 Where we are: 1. Introduction and Experimental Design 2. Exploratory Data Analysis 3. Probability 4. T-based methods for continous variables 5. Proportions & contingency tables -

More information

Meta-analysis of diagnostic research. Karen R Steingart, MD, MPH Chennai, 15 December Overview

Meta-analysis of diagnostic research. Karen R Steingart, MD, MPH Chennai, 15 December Overview Meta-analysis of diagnostic research Karen R Steingart, MD, MPH karenst@uw.edu Chennai, 15 December 2010 Overview Describe key steps in a systematic review/ meta-analysis of diagnostic test accuracy studies

More information

Meta-analysis of external validation studies

Meta-analysis of external validation studies Meta-analysis of external validation studies Thomas Debray, PhD Julius Center for Health Sciences and Primary Care, University Medical Center Utrecht, The Netherlands Cochrane Netherlands, Utrecht, The

More information

Analysis A step in the research process that involves describing and then making inferences based on a set of data.

Analysis A step in the research process that involves describing and then making inferences based on a set of data. 1 Appendix 1:. Definitions of important terms. Additionality The difference between the value of an outcome after the implementation of a policy, and its value in a counterfactual scenario in which the

More information

S Imputation of Categorical Missing Data: A comparison of Multivariate Normal and. Multinomial Methods. Holmes Finch.

S Imputation of Categorical Missing Data: A comparison of Multivariate Normal and. Multinomial Methods. Holmes Finch. S05-2008 Imputation of Categorical Missing Data: A comparison of Multivariate Normal and Abstract Multinomial Methods Holmes Finch Matt Margraf Ball State University Procedures for the imputation of missing

More information

A re-randomisation design for clinical trials

A re-randomisation design for clinical trials Kahan et al. BMC Medical Research Methodology (2015) 15:96 DOI 10.1186/s12874-015-0082-2 RESEARCH ARTICLE Open Access A re-randomisation design for clinical trials Brennan C Kahan 1*, Andrew B Forbes 2,

More information

Review: Logistic regression, Gaussian naïve Bayes, linear regression, and their connections

Review: Logistic regression, Gaussian naïve Bayes, linear regression, and their connections Review: Logistic regression, Gaussian naïve Bayes, linear regression, and their connections New: Bias-variance decomposition, biasvariance tradeoff, overfitting, regularization, and feature selection Yi

More information

Lecture Slides. Elementary Statistics Eleventh Edition. by Mario F. Triola. and the Triola Statistics Series 1.1-1

Lecture Slides. Elementary Statistics Eleventh Edition. by Mario F. Triola. and the Triola Statistics Series 1.1-1 Lecture Slides Elementary Statistics Eleventh Edition and the Triola Statistics Series by Mario F. Triola 1.1-1 Chapter 1 Introduction to Statistics 1-1 Review and Preview 1-2 Statistical Thinking 1-3

More information

Models for potentially biased evidence in meta-analysis using empirically based priors

Models for potentially biased evidence in meta-analysis using empirically based priors Models for potentially biased evidence in meta-analysis using empirically based priors Nicky Welton Thanks to: Tony Ades, John Carlin, Doug Altman, Jonathan Sterne, Ross Harris RSS Avon Local Group Meeting,

More information

Individual Participant Data (IPD) Meta-analysis of prediction modelling studies

Individual Participant Data (IPD) Meta-analysis of prediction modelling studies Individual Participant Data (IPD) Meta-analysis of prediction modelling studies Thomas Debray, PhD Julius Center for Health Sciences and Primary Care Utrecht, The Netherlands March 7, 2016 Prediction

More information

Multiple Bivariate Gaussian Plotting and Checking

Multiple Bivariate Gaussian Plotting and Checking Multiple Bivariate Gaussian Plotting and Checking Jared L. Deutsch and Clayton V. Deutsch The geostatistical modeling of continuous variables relies heavily on the multivariate Gaussian distribution. It

More information

MIDAS RETOUCH REGARDING DIAGNOSTIC ACCURACY META-ANALYSIS

MIDAS RETOUCH REGARDING DIAGNOSTIC ACCURACY META-ANALYSIS MIDAS RETOUCH REGARDING DIAGNOSTIC ACCURACY META-ANALYSIS Ben A. Dwamena, MD University of Michigan Radiology and VA Nuclear Medicine, Ann Arbor 2014 Stata Conference, Boston, MA - July 31, 2014 Dwamena

More information

Derivative-Free Optimization for Hyper-Parameter Tuning in Machine Learning Problems

Derivative-Free Optimization for Hyper-Parameter Tuning in Machine Learning Problems Derivative-Free Optimization for Hyper-Parameter Tuning in Machine Learning Problems Hiva Ghanbari Jointed work with Prof. Katya Scheinberg Industrial and Systems Engineering Department Lehigh University

More information