DRAFT REPORT FOR REVIEW VERSION AUGUST 22, 2013

Size: px
Start display at page:

Download "DRAFT REPORT FOR REVIEW VERSION AUGUST 22, 2013"

Transcription

1 RETROSPECTIVE OBSERVATIONAL STUDY QUESTIONNAIRE TO ASSESS RELEVANCE AND CREDIBILITY TO INFORM HEALTH CARE DECISION MAKING: AN ISPOR-AMCP-NPC GOOD PRACTICE TASK FORCE REPORT DRAFT REPORT FOR REVIEW VERSION AUGUST 22,

2 ABSTRACT (To be added) Keywords: questionnaire, checklist, decision-making, validity, credibility, relevance, bias, confounding 2

3 BACKGROUND TO THE TASK FORCE (To be added) INTRODUCTION Four Good Practices Task Forces developed a consensus-based set of questionnaires to help decision makers evaluate 1) prospective and 2) retrospective observational studies, 3) network meta-analysis (Indirect Treatment Comparison), and 4) decision analytic modeling studies with greater uniformity and transparency.[1-3] The primary audiences of these questionnaires are assessors and reviewers of health care research studies for health technology assessment, drug formulary, and health care services decisions requiring varying level of knowledge and expertise. This report focuses on the questionnaire to assess the relevance and credibility of prospective observational studies Although randomized controlled trials (RCTS) are often sought to inform health system decisions, there is increasing recognition of the limitations of relying on RCTS alone. These studies may be absent, due to financial, ethical or time limitations or they may lack sufficient information for decision-making in a realworld setting and recognizing a diverse range of populations and practice settings. Other types of studies, such as observational, modeling, and network meta-analysis, are increasingly sought to fill this gap. [4] However, there may be barriers to the use of these studies due to the limited number of accepted principles for their evaluation and interpretation. There is a need for transparent and uniform ways to assess their quality. [5] A structured approach reduces the potential for subjectivity to influence the interpretation of evidence and can promote consistency in decision-making. [6] Previous tools, including grading systems, scorecards, and checklists, have been developed to facilitate structured approaches to critically appraising clinical research.[7, 8] Some are elaborate, requiring software and deliberation among a broad range of experts [9], others are very simple, using scoring systems and best suited for randomized clinical trials.[10] It was believed a questionnaire that would not be time-consuming and easy to use with a basic understanding of epidemiology, would create awareness of issues related to alternative study designs, and could be widely promoted. Development of this questionnaire was informed by prior efforts and with several guiding principles derived from the focus group payors comments described above. First, questionnaires had to be easy-to-use and not time consuming for individuals without a broad range of expertise and without in-depth study design and statistical knowledge. Second, questionnaires had to be sufficiently comprehensive to promote awareness of the appropriate application of different study designs to decision-making; we also sought to produce a 3

4 questionnaire that would prompt users to obtain additional education on the underlying methodologies. Lastly, the use of questionnaire would need to be facilitated by comprehensive educational programs. The Task Force defines prospective observational studies as ones in which participants are not randomized or otherwise assigned to an exposure for which the consequential outcomes of interest occur after study commencement (including creation of a study protocol and analysis plan, and study initiation). They are often longitudinal in nature. Exposure to any of the interventions being studied may have been recorded before the study initiation such as when a prospective observational study uses an existing registry cohort. Exposure may include a pharmaceutical intervention, surgery, medical device, prescription, or decision to treat. This definition contrasts with retrospective observational studies that use existing data sources in which both exposure and outcomes have already occurred. [11] QUESTIONNAIRE DEVELOPMENT The first issue was whether the questionnaires should be linked to checklists, scorecards, or annotated scorecards. Concerns were raised that a scoring system may be misleading if it did not have adequate measurement properties. Scoring systems have been shown to be problematic in the interpretation of randomized trials.[12] An alternative to a scorecard is a checklist. However, the Task Force members believed that checklists might also mislead users because a study may satisfy all of the elements of a checklist and still harbor fatal flaws. Moreover, users might have the tendency to count up the number of elements present converting it into a score, and then apply the score to their overall assessment of the evidence. In addition, the strength of a study finding may depend on other evidence that addresses the specific issue or the decision being made. A questionnaire without an accompanying score or checklist was felt to be the best way to allow analysts to be aware of the strengths and weaknesses of each piece of evidence and apply his/her own reasoning. Questions were developed based on a review of items in previous questionnaires, guidance documents, and previous ISPOR Task Force recommendations [11] as well as methods and reporting guidance s.[13-16] Through user testing and consensus, items from previous efforts were grouped into conceptual domains The questionnaire is divided into two main sections, Relevance and Credibility based on the key elements essential to evaluating comparative effectiveness evidence. Four similar questions were developed for the relevance section across all questionnaires. Credibility is further divided into several key domains. For this questionnaire to obtain broad acceptance, and based on the focus group 4

5 recommendations, it was limited to ~30 questions. Whenever possible, efforts were made to avoid the use of jargon and to employ similar wording across all four questionnaires. Also, as shown in Figure 1, there is substantial overlap in the design and flow of this questionnaire and that developed for retrospective observational studies. FIGURE 1 Upon completion of questions in the relevance section, users are asked to rate whether the study is sufficient or insufficient for inclusion. If a study was not considered sufficiently relevant, a user could then opt to truncate the review of its credibility. In the credibility category, users rate each domain as a strength, a weakness, or neutral. Based upon these evaluations, the user then similarly rates the credibility of the research study as either sufficient or insufficient to inform decision making. For some questions in the credibility section, a user would be notified that they had detected a fatal flaw. The presence of a fatal flaw suggests significant opportunities for the findings to be misleading. Consequently, the decision maker should use extreme caution in applying the findings to inform decisions. However, the occurrence of a fatal flaw does not prevent a user from completing the questionnaire nor does it require the user to judge the evidence as insufficient for use in decision making. The presence of a fatal flaw is intended to raise a strong caution and should be carefully considered when the overall body of evidence is reviewed. The questionnaire includes a summary in a structured paragraph as follows: In evaluating this study, I made the following judgments: I found the study (relevant/not relevant) for decision making because I considered that the population/interventions/outcomes/setting (applied/did not apply) to the decision I am informing. I found the study (credible/not credible) for decision making because: o There (were/were not any) fatal flaws i.e. critical elements that call into question the validity of the findings. The presence of a fatal flaw suggests significant opportunities for the findings to be misleading and misinterpreted; extreme caution should be used in applying the findings to inform decisions. The following domains contained fatal flaws: o There are strengths and weakness in the study: The following domains were evaluated as strengths: The following domains were evaluated as weaknesses 5

6 QUESTIONNAIRE ITEMS Questions that fall under the main categories of relevance and credibility appear in Table 1. Explanations of each question along with specific definitions are provided in the following section, to facilitate understanding of the appropriate use of the questionnaire. 162 TABLE RELEVANCE Relevance addresses whether the results of the study/apply to the setting of interest to the decision-maker. It addresses issues of external validity (population, comparators, endpoints, timeframe) and the direction and magnitude of difference meaningful to the decision maker. There is no correct answer for relevance. Relevance is determined by each decision-maker and the relevance assessment determined by one decision-maker will not necessarily apply to other decision-makers Is the population relevant? This question addresses whether the population analyzed in the study sufficiently matches the population of interest to the decision-maker. Population characteristics to consider include demographics such as age and gender, nationality, ethnicity; risk factors such as average blood pressure, cholesterol levels, and body mass index levels; behaviors such as smoking; stage/severity of the condition; past and current treatments for the condition; and clinical issues such as co-morbidities Are any relevant interventions missing? This question addresses whether the interventions analyzed in the study match the ones of interest to the decision-maker and whether all relevant comparators have been considered. Intervention characteristics to consider include: technology characteristics (e.g., if the issue is screening for osteoporosis, is the technology dual-energy X-ray absorptiometry (DEXA) of the spine? DEXA of the wrist? Another scanning method?); technique (e.g., if the issue is endovascular aneurysm repair, does it involve an iliac branch device?); dose of a drug or biologic; duration of treatment; mode of administration (e.g., oral, intravenous); skill level of provider; post-treatment monitoring and care; and duration of follow-up. In addition, analysts should consider to what extent alternative modes of care (e.g., another intervention or standard care) match the modes of care in the decision setting Are the outcomes relevant? This question asks what outcomes are assessed in the study and whether the outcomes are meaningful to the decision maker. Outcomes such as cardiovascular events (e.g., rates of myocardial infarction or stroke), patient functioning, or health-related quality of life (e.g., scores from short form-36 health survey 6

7 or EuroQol-5D instruments) may be more relevant than surrogate or intermediate endpoints (e.g., cholesterol levels). 4. Is the context (settings & practice patterns) applicable? The context of the study refers to factors that may influence the generalizability of the study findings to other settings. Factors that should be considered may include the study time frame, the payer setting, provider characteristics, or the geographic area. Some or all of these factors may be different than the population to which the user wants to apply the study results; however, if it is suspected that differences in these factors may influence the treatment response, it should influence the user s judgment of the extent that these findings could be applied to another setting. CREDIBILITY Credibility addresses the extent to which the study accurately answers the question it is designed or intended to answer and is determined by the design and conduct of the study. It addresses issues of internal validity, error, and confounding. For example, the observed effect of a new treatment may be due to the degree to which patients were followed and their outcomes reliably measured and not due to differences in treatment effectiveness. Appropriate study design and analytic approaches can better separate the contribution of the intervention to observed outcomes versus other factors. The credibility section of the questionnaire was divided into the following domains: design, data, analysis, reporting, interpretation, and conflicts of interest. DESIGN 5. Were the study hypotheses or goals pre-specified a priori? 'A priori' in this context refers to the development of the objectives or study plan prior to the execution of the study. This is often a difficult point to assess in published reports since many authors do not explicitly report which if any objectives were pre-specified. The best evidence that the study developed 'a priori' objectives would be to compare a study report with the trial registry where the study was registered, however, only small fraction of observational studies are registered in these trial registries. Alternatively, users could rely on terms commonly used in reporting observational studies that indicate pre-specified 'a priori.' Phrases and terms such as 'pre-specified' or 'planned analyses,' when describing the objectives, methods, or even the results can indicate pre-specified objectives. Other indicants that the objectives may have been developed prior to the study execution would be reporting of IRB review or grant awards 7

8 specific to the research study. 'A priori' objectives cannot be assumed when a study does not offer any indicant that the objectives were developed beforehand. 6. If one or more comparison groups were used were they concurrent comparators or did they justify the use of historical comparisons group(s)? Concurrent comparators are derived from the same sample of subjects and are followed over the same time frame. Historical controls are derived from a sample of subjects derived from a time period prior to the treatment that is compared. Concurrent comparators should be used where possible, however, the use of historical comparators can be justified when a treatment becomes a standard of care and nearly all subjects that could be treated either receive the new treatment, or it becomes unethical to withhold the new treatment. 7. Was there evidence that a formal study protocol including an analysis plan was prespecified and guided the execution of the study? Altering study definitions, subject inclusion exclusion criteria, model specification, and other study procedures can have dramatic effects on study findings and permit reported results to be influenced by investigator biases.[17-19] Ideally, study reports would be fully transparent about pre-planned study procedures. They should report findings based on the original plan and be fully transparent regarding post-hoc changes to the analysis plan to justify those post-hoc alterations. Unfortunately, the conventional reporting practices of observational studies are rarely detailed enough to allow readers to adequately assess what was pre-planned and what was not. [20] The best evidence that the study had an analytic plan would be to compare a study report with the trial registry where the study was registered; however, only a small fraction of observational studies are registered in these trial registries. Alternatively, users could rely on terms commonly used, such as 'pre-specified' and 'planned analyses' when describing the methods. The results can also indicate pre-specified objectives and analyses by declaring results that were pre-planned versus those that were post-hoc. If the study reports 'post-hoc' analyses, the user may cautiously infer that the other analyses were pre-planned. Other indicants that an analysis plan was developed prior to the study execution would be the reporting of an IRB review or grant awards specific to the research study. A pre-specified analysis plan cannot be assumed when a study does not offer any indication that there was such a plan developed beforehand. 8. Was sample size and statistical power addressed? 8

9 An observational study attempts to approximate a randomized study and therefore a sample size or power calculation should be performed in advance of the study. [11] In the retrospective framework where subjects are not prospectively recruited, sample size estimates do not dictate the number of subjects to be included in the study as the investigator will typically include all available subjects recorded in the data source and follow them for as long as subject data are recorded. However, a sample size estimate / power calculation will enable readers to interpret 'null' findings. A study that results in a null finding of no difference between comparators that has low power (e.g. < 80%) is difficult to interpret since the findings may be due to inadequate power or due to a true null finding. 9. Was a study design employed to minimize or account for confounding such as inception cohorts, new-user designs, multiple comparator groups, matching designs, and assessment of outcomes not thought to be impacted by the intervention compared? Some study designs can provide stronger methods to deal with potential confounding that may occur due to lack of randomization. A confounder is a factor that distorts the true relationship of the study variables of central interest by virtue of being related to the outcome of interest, but extraneous to the study question and unequally distributed among the groups being compared. A variety of considerations will factor into the choice of study design including cost, feasibility, and ethical considerations. Box 1: Study Designs Employed to Minimize the Effect of Confounding Variables Inception cohorts are designated groups of persons assembled at a common time early in the development of a specific clinical disorder (e.g., at first exposure to the putative cause or at initial diagnosis), who are followed thereafter. A new-user design begins by identifying all of the patients in a defined population (both in terms of people and time) who start a course of treatment with the study medication. Study follow-up for endpoints begins at precisely the same time as initiation of therapy, (t0). The study is further restricted to patients with a minimum period of non-use (washout) prior to t0. Newuser designs mitigate channeling bias by excluding subjects that tolerate existing treatments. Matching designs include a deliberate process of making a study group and a comparison group comparable with respect to factors that are extraneous to the purpose of the investigation, but which might interfere with the interpretation of the study's findings. For example, in case-control 9

10 studies, individual cases might be matched or paired with a specific control on the basis of comparable age, sex, clinical features, or a combination of these. Assessment of outcomes thought not to be impacted by the intervention compared or 'falsification tests' may permit an assessment of residual confounding [11] 10. Were the sources, criteria, and methods for selecting participants appropriate to address the study hypotheses? The sources, criteria, and methods for selecting percipients to study should be similar for the different groups of patients being assessed, whether they are defined based on outcome (case-control studies) or based on exposure (cohort studies). Bias can be introduced if patient selection varies by comparator group or the data source or methods for assessing or selecting patient groups varies.[21] Also the data source should provide some level of assurance that key measures are reliably recorded. Some key questions in assessing the sources, criteria, and methods for selecting patients include; 1. Was there an adequate rationale provided for key inclusion and exclusion criteria? 2. Was there an assurance that subject encounters or data were adequately recorded over the entire study time frame for each subject? Depending on the data source, the assurance that relevant health information was captured will vary. Administrative data sources should be checked to ensure subjects had continuous eligibility for health benefits and are eligible to receive all relevant sources of care. t all persons eligible for insurance will be eligible for pharmacy or mental benefits for example. Studies that rely on provider-supplied data such as EMRs should attempt to ensure that persons are receiving care in the provider network and not from other providers. 11. Was the study sample restricted so that comparison groups would be sufficiently similar to each other (e.g., indications for treatment)? One of the most common starting points to enhance the comparability of treatment groups is to restrict subjects to a common set of conditions or patient characteristics. For example, if one were to compare beta blockers and diuretics as antihypertensive therapy, it would be important to restrict both treated groups to those without any evidence of previous cardiovascular disease including angina since beta blockers have indications for angina (a strong risk factor for subsequent MI) whereas diuretics do not and an unrestricted analysis would suffer by confounding by indication. DATA 10

11 Was exposure defined and measured in a valid way? Exposure to treatment is ideally documented by evidence that patients actually took the medication or received some form of treatment, though this is rarely available. [22] Exposure may be documented by evidence of a prescription being written, being fulfilled, a claim being filed, or expressed in continuous terms such as the medication possession ratio. The most basic level of exposure measurement defines the presence or absence that a subject received (or did not receive) a treatment, but exposure may also be defined in terms of the intensity of exposure, including the dose and/or duration of treatment(s). A valid exposure measure will have low levels of exposure misclassification. Exposure misclassification may occur when the data used to define exposure are errantly recorded. For example, subjects may be misclassified when they experience a treatment but are defined as unexposed in a study and vice versa. Specific exposure misclassification, common in retrospective analyses of claims data, may occur when drug treatments are not covered by a health plan and persons obtain treatments without filing claims. Drugs that can be acquired over the counter, such as lifestyle drugs that are not commonly reimbursed, treatments that may have some stigma associated with them, or treatments that are inexpensive where claims may not be filed, are more prone to exposure misclassification. Some key questions one should consider when assessing the validity of the studies exposure measure include; 1. Does the study describe the data source(s) used in the study of the ascertainment of exposure (e.g., pharmacy dispensing, general practice prescribing, claims data, self-report, face-to-face interview)? 2. Does the study describe how exposure is defined and measured (e.g., operational details for defining and categorizing exposure)? 3. Does the study discuss the validity of exposure measurement (e.g., precision, accuracy, prospective ascertainment, exposure information recorded before outcome occurred, use of validation sub study)? The latter question regarding the accuracy of the study exposure measure can be obtained by a sub-study comparing alternative data sources or by citing previous validation studies of the study exposure measures. 13. Have the primary outcomes been defined and measured in a valid way? Ideally, some evidence of the validity of the outcome measure definitions (which often include measures of sensitivity, specificity, and positive predictive value) should be described. Researchers may conduct a sub study to verify the accuracy of their outcome definitions using a more detailed source such as a review of electronic medical records, chart reviews, or patient/provider surveys, and then report the performance of their outcome definitions. More often, investigators may cite previous studies that validated outcome measures/definitions with an alternative source. Some judgment may be exercised in assessing the validity of an outcome measure. Outcome measure definitions that have been used in 11

12 previous analyses permit comparison between studies but do not ensure validity. Some outcome measures that are measured objectively such as inpatient episodes derived from an administrative claims database, may be presumed to be valid since there are strong financial incentives for these services to be recorded, and they are not subject to personal judgment. The validity of other measures, and in particular those that are subject to clinical judgment such as patient functioning or symptoms such as pain or severity of disease, will be context-specific and more prone to errors in measurement. Additional steps such as blinded assessment or the use of multiple clinician assessments should be considered. 14. Were the data sources sufficient to support the study? The data source should contain valid measures of treatment, outcome, and covariates for confounding variables, possess a large enough sample, and be of sufficient duration to detect differences. Often a single data source will not contain all the information necessary to conduct the study, and linkages to other data sources may be necessary; those data linkages should be described well enough to assess their validity. This is a broad concept to assess as the quality of the study data and its use in executing the study are influenced by multiple factors. Questions to consider in assessing these criteria include; 1. Were all outcomes, exposures, predictors, potential confounders, and effect modifiers clearly defined? (Give diagnostic criteria, if applicable.) 2. Were sources of data and details of methods of assessment for each variable of interest described? 3. Were the assessment methods and/or definitions the same across treatment groups? 4. Have the reliability and validity of the data been described, including any data quality checks and data cleaning procedures? 5. Were reliable and valid data available on important confounders or effect modifiers? 15. Was the follow-up time similar among comparisons groups OR were differences in follow-up accounted for in the analysis? Persons that are followed for longer durations are more likely to experience outcome events and the duration of follow-up must be accounted for in the analysis. Restricting the sample to only those subjects that have the same follow up (e.g., one year) is one approach to account for person time in the analysis. However, this approach can lead to immortal time bias (also referred to as survivorship bias), where subjects that die or otherwise lose enrollment in the study are excluded, which may be differential across treatment groups. The more likely that the duration of follow-up time can be influenced by the treatment group selections, the greater the prospects that immortal time bias can influence the associations. An alternative approach is to allow subjects to be followed for varying durations and to use a survival analytic approach. However, this does not eliminate the possibility of immortal time bias occurring if being selected to a cohort is dependent upon some person time which may not be overt criteria, but rather 12

13 one that may be de-facto to be exposed to a treatment. A good study will explain as best as possible these reasons and use appropriate statistical techniques (e.g., time to event such as cox proportional hazards) to mitigate the impact. ANALYSES 16. Were sensitivity analyses performed to assess key assumptions or definitions? A variety of decisions must be made in designing a study. How are the populations, interventions, and outcomes defined? How are missing data dealt with? How are outliers dealt with? To what extent may unmeasured confounders influence the results? A good study will indicate which of these decisions had an important impact on the results of the analyses and will report the effect of using a reasonable range of alternative choices on the results. Key issues to consider include; 1. Were sensitivity analyses reported using different statistical approaches? 2. Were sensitivity analyses reported to test the impact of key definitions? 3. Did the analysis account for outliers and examine their impact in a sensitivity analysis? 4. Did the authors discuss or conduct a sensitivity analysis to estimate the effect of unmeasured confounders on the observed difference in effect? 17. Was there a thorough assessment of potential measured and unmeasured confounders? Carefully identifying all potential confounders is a critical step in any observational comparative effectiveness research study. Confounders are variables or factors that are correlated with treatment and outcome variables that lie outside the causal pathway between treatment and outcome. Confounders may be an artifact of common epidemiologic biases such as channeling or indication bias. For example, persons with more severe disease may be more likely to receive one treatment over another. Treatment inferences from observational studies are all potentially biased from imbalances across treatment groups on confounding variables whether the variables are observed or not. Carefully identifying all potential confounders is a critical step in any observational comparative effectiveness research study. Confounding can be controlled statistically using a wide range of multivariate approaches, however, if a statistical model excludes a confounding variable, the estimates of treatment effects suffer from omitted variable bias in just about all analyses except when a valid instrumental variables approach is undertaken. When assessing a research study, one should look to see that the authors have considered all potential confounding factors and conducted a literature review to identify variables that are known to influence the outcome variable. The literature search strategy for identifying confounders is not typically reported in manuscripts; however, a table of potential confounders with citations to previous research describing 13

14 these associations is sometimes available in a manuscript and would be suggestive of an explicit search to identify known confounding variables. In addition to a literature search, researchers should use clinical judgment to identify confounders. Often the data will not contain information for some confounding variables (e.g., race, income, exercise) and will be omitted variables from the analysis. When the analysis does not include key confounders, the potential impact of these should be discussed including the direction and magnitude of the potential bias. 18. Were analyses of subgroups or interactions of effects reported for comparison groups? Exploring and identifying heterogeneous treatment effects, or effect modification, are one of the potential advantages of large observational studies. Interaction occurs when the association of one exposure differs in the presence of another exposure or factor. The most common and basic approaches for identifying heterogeneous treatment effects are to conduct subgroup analyses or to incorporate interaction terms with the treatment variable. [23] Interactions may be explored using additive or multiplicative approaches where the differences in effect depart from either the addition of the effects of the two factors or exposures, or the multiplicative effect of those factors. Caution should be warranted when the main treatment effect is not significantly associated with the outcome, but significant subgroup results are reported (which could have been the result of a series of post-hoc sub-group analyses). REPORTING 19. Were the methods described in sufficient detail? Replication is one of the hallmarks of the scientific process, and given the increasing availability of retrospective data source studies, it is increasingly possible to replicate many retrospective studies whether utilizing the same data source or a similar data source. Ideally, a published manuscript should describe the study in sufficient detail to enable a knowledgeable person to replicate the study with the same sample criteria, study measures, and analytic approach. However, the word limits imposed by many journals in their printed versions often do not permit enough space to sufficiently describe a study to enable replication. When a study is not described in sufficient detail, an online appendix or a statement indicating that a technical appendix will be made available upon request would be sufficient. 20. Were the numbers of individuals at each stage of the subject selection process reported after applying inclusion and exclusion criteria? 14

15 Reporting the number of individuals screened at each stage of the selection process is important to assess the potential selection bias of participants. The final analyzable sample should be described in text or displayed as a flow diagram describing the initial pool of potential subjects and the sample after each inclusion and exclusion is applied. This will allow the reader to assess the extent to which the analyzable sample may differ from the target population and which criteria materially influenced the final sample. 21. Were the descriptive statistics of study participants adequately reported? Descriptive statistics include demographics, comorbidity, and other potential confounders reported by treatment groups that will enable the reader to assess the potential for confounding. Vast differences on key confounding measures may suggest a higher likelihood of residual confounding even after adjusting for the observable characteristics. 22. Did the authors describe the uncertainty of their findings through reporting confidence intervals and or p-values? There will always be uncertainty when evaluating outcomes in observational research, and that uncertainty should be presented in the form of either a p-value (chance that the relationship being assessed is observed by chance), or a confidence interval (range of values that contain the true estimate with a pre-defined probability - e.g. 95%). 23. Did the authors describe and report the key components of their statistical approaches? The authors should fully describe their statistical approach and provide citations for any nuanced econometric or statistical methods that fully describe the approach. Some factors to consider in assessing the adequacy of the statistical reporting includes; 1. If the authors utilize multivariate statistical techniques, do they discuss how well the models predict what it is intended to predict (e.g. R-square, pseudo r-square, c-statistics, and c-indices)? 2. Were any modeling assumptions addressed such as multiplicity adjustment or the Bonferroni correction? 3. Were unadjusted estimates of treatment effects (outcomes) reported? 4. Was the full regression model (not just the adjusted treatment effects) available in the publication or at least an appendix? 5. If propensity score methods were used, were the methods appropriately described? This includes the method of developing the propensity score, use of the propensity score (matching, weighting, regression, etc.), and evaluation of the propensity score (e.g., standardized differences before and after matching). 6. If instrumental variable methods were used, were 15

16 the methods appropriately described (rational for the instrumental variable, evaluation of the strength of the instrument)? 24. Were confounder-adjusted estimates of treatment effects reported? Confounder-adjusted estimates of treatment effects can be obtained in a variety of ways. Most commonly, treatment effects are estimated from a coefficient of a treatment variable(s) in a multivariate regression, or systems of equations that include a set of control covariates that represent potential confounders. Treatment effects may also be estimated by taking differences from propensity matched treated and comparison subjects. Any non-randomized study must report confounder-adjusted estimates if they are attempting to make any inference regarding a treatment comparison. Unadjusted estimates should also be reported to allow for comparison with the adjusted results. 25. Was the extent of missing data and how they were handled reported? Unlike clinical trials or prospective registries, it is often difficult to assess when and for whom data are missing because subjects are rarely enrolled in a study with pre-determined data collection time points. When data are clearly missing, for example, a prescription claim is identified in a claims database but the days supply is missing from the claim, the frequency this occurs should be reported. Additionally, the methods for handling the missing data should be described including whether subjects were excluded, the missing values were imputed, or analyses were stratified. 26. Were absolute and relative measures of treatment effects reported? Reporting the effect of treatment(s) in both absolute and relative terms provides the decision maker the greatest understanding of the magnitude of the effect. [24] Absolute measures of effect included differences in proportions, means, rates, NNH, and NNT and should be reported for a meaningful time period. Relative measures of effect are defined as a ratio of rates, proportions, or other measures and include Odds Ratios (ORs), incidence rate ratios, relative risks, HRs. INTERPRETATION 16

17 Were the results consistent with prior known information? If not, was an adequate explanation provided for the inconsistent results? The authors should undertake a thorough review of the literature to identify and present all known previous findings exploring the same or similar objectives and then contrast their findings. When studies differ in the direction of the findings, the study should motivate plausible explanations for the disparate findings and identify methodological differences and/or advance a theoretical or biologic rationale for the differences. Key questions to consider include; Is the direction of the effect in this study comparable to similar studies? Is the magnitude of effect in this study comparable to similar studies? If the direction or size of effect is different, have the authors provided an adequate explanation? Is the explanation plausible? Are there other more important factors more likely to explain the difference? 28. Are the results (differences demonstrated) considered clinically meaningful? In analyses of large databases, sometimes relatively minor differences between treatment groups can attain levels of statistical significance due to the large sample sizes. The results should be interpreted not only in terms of their statistical association, but also by the magnitude of effect in terms of clinical or economic importance. Additionally, the larger the treatment effect that is observed, the smaller the chances that residual confounding can change a significant finding to a null finding. Key questions to consider include; was the magnitude of differences in results between groups large? Have the statistical findings been interpreted in terms of their clinical or economic relevance? 29. Are the conclusions fair and balanced? Overstating the implications of the study results is commonly encountered in the literature. The study should be fully transparent describing the study limitations and importantly how study limitations could influence the direction and magnitude of the findings and ultimately the study conclusions. Key considerations include; was a cautious/appropriate overall interpretation of results considering objectives, limitations, multiplicity of analyses, results from similar studies, and other relevant evidence presented? Were the limitations of the study, including the direction and magnitude of any potential bias, reported? 30. Was the impact of unmeasured confounding factors discussed? 17

18 Unmeasured confounding is always a potential issue in any observational research framework. Unmeasured confounding is more likely to bias the results when patients might be channeled to one treatment over another with studies that investigate known or likely outcomes. Outcome events that are lesser known or unsuspected are less likely to be considered in the clinician s or patient s decision process, and in turn, are less likely to suffer from confounding. A good discussion will identify factors thought to be confounders that were not recorded or were un-measurable, and identifies the potential direction of the bias. Ideally, residual confounding could be accessed through simulations to explore how strongly a confounder would have to be correlated with treatment and outcome to move the results to a null finding. CONFLICTS OF INTEREST 31. Were there any potential conflicts of interest? Potential conflicts of interest care may be present for many reasons including financial interests in the study results, desire for professional recognition, or the desire to do favors for others. 32. If there were potential conflicts of interest, were steps taken to address these? Steps to address potential conflicts of interest include disclosure of any potential conflicts of interest, involving third parties in the design, conduct, and analysis of studies, and agreements that provide independence of researchers (including freedom to publicly disseminate results) from funding entities. [25] DISCUSSION User Testing: Following testing by members of the Assessing Prospective and Retrospective Observational, Indirect Treatment and Modeling Studies Task Forces and subsequent modification of the questionnaire based on these tests, a revised questionnaire was made available to volunteers from the payers community as well as the pharmaceutical industry. Each volunteer was asked to test one questionnaire using three studies and rate them accordingly. Studies were previously rated by the Task Force as either good quality, medium quality, or poor quality. Ninety-three volunteers were solicited to participate, with 65 participating and 24 individuals assigned to the retrospective observational study testing. The retrospective observational study questionnaire response rate was 70%. Although there were not enough 18

19 users to perform a formal psychometric evaluation, the good quality study was generally rated as sufficient with respect to relevance and credibility, while the poor study was generally rated not sufficient. Based upon the answer to the question: is the study Sufficiently or Insufficiently credible to include in the body of evidence? -- There was 54% agreement overall among the ratings provided. Ratings were not provided 36% of the time. Multi-rater agreement exceeded 80% for most of the credibility domains. Few users completed supplementary questions. Educational Needs Internationally, the resources and expertise available to inform health care decision makers varies widely. While there is broad experience in evaluating evidence from RCTs, there is less experience and greater skepticism regarding the value of other types of evidence. (11, 18) However, the volume and variety of real-world evidence is increasingly rapidly with the increasing adoption of electronic medical records (EMRs) and the linkage of claims data with laboratory, imaging and EMR data. Volume, variety and velocity (the speed at which data is generated) are three of the hallmarks of the era of big data in health care. The amount of information from these sources could easily eclipse that from RCTs in coming years. This implies it is an ideal time for health care decision makers and those that support evidence evaluation to enhance their ability to evaluate this information. Although there is skepticism about the value of evidence from observational, network metaanalysis/indirect treatment comparison and modeling studies, they continue to fill important gaps in a knowledge base important to payers, providers, and patients. ISPOR has provided Good Clinical Research recommendations on what rigorous design, conduct, and analysis looks like for these sources of evidence. [11] These questionnaires are an extension of those recommendations and serve as a platform to assist the decision maker in understanding what a systematic evaluation of this research requires. By understanding what a systematic structured approach to the appraisal of this research entails, it is hoped that this will lead to a general increase in sophistication by decision makers in the use of this evidence. To that end, we anticipate additional educational efforts and promotion of these questionnaires will be developed and made available to members. In addition, an interactive (i.e., web-based) questionnaire would facilitate uptake and support the educational goal that the questionnaire provides 19

20 REFERENCES 1. Jansen, J.P. Et Al. An Indirect Treatment Comparison And Network Meta-Analysis Study Questionnaire To Assess Study Relevance And Credibility Of To Inform Healthcare Decision- Making: An ISPOR-AMCP- NPC Good Practice Task Force Report, Value in Health, 17:1 2. Martin, et al. A Retrospective Observational Study Questionnaire to Assess Relevance and Credibility to Inform Healthcare Decision-Making: An ISPOR-AMCP-NPC Good Practice Task Force Report, Value in Health, 17:1. 3. Caro JJ, et al. A Modeling Study Questionnaire to Assess Study Relevance and Credibility to Inform Healthcare Decision-Making: An ISPOR-AMCP-NPC Good Practice Task Force Report. Value in Health,; 17:1 4. Garrison Jr. LP, Neumann PJ, Erickson P, et al. Using real-world data for coverage and payment decisions: The ISPOR real-world data task force report. Value Health 2007;10: Brixner DI, Holtorf AP, Neumann PJ, Malone DC, Watkins JB. Standardizing Quality Assessment of Observational Studies for Decision Making in Health Care. JMCP. 2009;15(3); Balshem H, Helfand M, Schünemann HJ, Oxman AD, Kunz R, Brozek J, et al. GRADE guidelines: 3. Rating the quality of evidence. Journal of Clinical Epidemiology Apr;64(4): Atkins D, Eccles M, Flottorp S, Guyatt GH, Henry D, et al. (2004) Systems for grading the quality of evidence and the strength of recommendations I: critical appraisal of existing approaches The GRADE Working Group. BMC Health Serv Res 4: 38. doi: / Glasziou P (2004) Assessing the quality of research. BMJ 328: doi: /bmj Guyatt GH, Oxman AD, Vist GE, Kunz R, Falck-Ytter Y, et al. (2008) GRADE: an emerging consensus on rating quality of evidence and strength of recommendations. BMJ 336: doi: /bmj ad. 10. Moher D, Jadad AR, Tugwell P (1996) Assessing the quality of randomized controlled trials. Current issues and future directions. Int J Technol Assess Health Care 12: Berger ML, Dreyer N, Anderson Fred, et al. Prospective Observational Studies to Assess Comparative Effectiveness: The ISPOR Good Research Practices Task Force Report. Value Health 2012;15: Jüni P, Witschi A, Bloch R, Egger M (1999). The Hazards of Scoring the Quality of Clinical Trials for meta- analysis. Jama J Am Med Assoc 282: GRACE Principles. Available from: Accessed on: August 9, STROBE checklist. Available from: g/fileadmin/strobe/uploads/checklists/strobe_checklist_v4_cohort.pdf Accessed on: August 9, EnCePP. Available from: rdsinpe.pdf Accessed on: August 9,

21 AHRQ User Guide for Developing a Protocol for Observation Comparative Effectiveness Research. Accessed on: August 9, Thomas L, Peterson ED. The value of statistical analysis plans in observational research: defining high-quality research from the start. JAMA Aug 22;308(8): Cox E, Martin BC, Van Staa T, Garbe E, Siebert U, Johnson ML. Good research practices for comparative effectiveness research: approaches to mitigate bias and confounding in the design of nonrandomized studies of treatment effects using secondary data sources: the International Society for Pharmacoeconomics and Outcomes Research Good Research Practices for Retrospective Database Analysis Task Force Report--Part II. Value Health Dec;12(8): Berger ML, Mamdani M, Atkins D, Johnson ML. Good research practices for comparative effectiveness research: defining, reporting and interpreting nonrandomized studies of treatment effects using secondary data sources: the ISPOR Good Research Practices for Retrospective Database Analysis Task Force Report--Part I. Value Health Dec;12(8): Altman DG (2000) Statistics in medical journals: some recent trends. Stat Med 19: Ellenberg JH. Selection bias in observational and experimental studies. Statistics in Medicine. 1994;13(5-7): Gordis L. (1979) Conceptual and methodologic problems in measuring patient compliance. In: Haynes B, Taylor DW, Sackett DL, eds. Compliance in Health Care. Baltimore: The John Hopkins University Press, Rosenbaum PR. Heterogeneity and causality: Unit heterogeneity and design sensitivity in observational studies. Am Stat May;59(2): Forrow L, Taylor WC, Arnold RM. Absolutely relative: how research results are summarized can affect treatment decisions. Am J Med1992;92: Husereau D, Drummond M, Petrou S, et al. Consolidated health economic evaluation reporting standards (CHEERS) Explanation and elaboration: A report of the ISPOR health economic evaluations publication guidelines good reporting practices task force. Value Health 2013;16:

22 Figure 1 Retrospective observational study assessment questionnaire flowchart TABLE 1 Questionnaire to assess the relevance and credibility of a Retrospective observational study The questionnaire consists of 31 questions related to the relevance and credibility of a Retrospective observational study. Relevance questions relate to usefulness of the prospective observational study to inform health care decision making. Each question will be scored with / / Can t Answer. Based on the scoring of the individual questions, the overall relevance of the prospective observational study needs to be judged as Sufficient or Insufficient. If the Retrospective observational study is considered sufficiently relevant, the credibility is going to be assessed. The credibility is captured with questions in the following 6 domains, Design, Data, Analysis, 22

Prospective Observational Study Assessment Questionnaire

Prospective Observational Study Assessment Questionnaire 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 PROSPECTIVE OBSERVATIONAL STUDY QUESTIONNAIRE TO ASSESS RELEVANCE AND CREDIBILITY TO INFORM HEALTHCARE DECISION-MAKING: AN ISPOR-AMCP-NPC GOOD PRACTICE

More information

ISPOR Task Force Report: ITC & NMA Study Questionnaire

ISPOR Task Force Report: ITC & NMA Study Questionnaire INDIRECT TREATMENT COMPARISON / NETWORK META-ANALYSIS STUDY QUESTIONNAIRE TO ASSESS RELEVANCE AND CREDIBILITY TO INFORM HEALTHCARE DECISION-MAKING: AN ISPOR-AMCP-NPC GOOD PRACTICE TASK FORCE REPORT DRAFT

More information

Interpreting Prospective Studies

Interpreting Prospective Studies Comparative Effectiveness Research Collaborative Initiative (CER CI) PART 1: INTERPRETING OUTCOMES RESEARCH STUDIES FOR HEALTH CARE DECISION MAKERS ASSESSING PROSPECTIVE DATABASE STUDIES: A PROPOSED MEASUREMENT

More information

Online Supplementary Material

Online Supplementary Material Section 1. Adapted Newcastle-Ottawa Scale The adaptation consisted of allowing case-control studies to earn a star when the case definition is based on record linkage, to liken the evaluation of case-control

More information

CONSORT 2010 checklist of information to include when reporting a randomised trial*

CONSORT 2010 checklist of information to include when reporting a randomised trial* CONSORT 2010 checklist of information to include when reporting a randomised trial* Section/Topic Title and abstract Introduction Background and objectives Item No Checklist item 1a Identification as a

More information

Comparative Effectiveness Research Collaborative Initiative (CER-CI) PART 1: INTERPRETING OUTCOMES RESEARCH STUDIES FOR HEALTH CARE DECISION MAKERS

Comparative Effectiveness Research Collaborative Initiative (CER-CI) PART 1: INTERPRETING OUTCOMES RESEARCH STUDIES FOR HEALTH CARE DECISION MAKERS Comparative Effectiveness Research Collaborative Initiative (CER-CI) PART 1: INTERPRETING OUTCOMES RESEARCH STUDIES FOR HEALTH CARE DECISION MAKERS HEALTH EVIDENCE FOR DECISION MAKING: ASSESSMENT TOOL

More information

Recent developments for combining evidence within evidence streams: bias-adjusted meta-analysis

Recent developments for combining evidence within evidence streams: bias-adjusted meta-analysis EFSA/EBTC Colloquium, 25 October 2017 Recent developments for combining evidence within evidence streams: bias-adjusted meta-analysis Julian Higgins University of Bristol 1 Introduction to concepts Standard

More information

Systematic reviews and meta-analyses of observational studies (MOOSE): Checklist.

Systematic reviews and meta-analyses of observational studies (MOOSE): Checklist. Systematic reviews and meta-analyses of observational studies (MOOSE): Checklist. MOOSE Checklist Infliximab reduces hospitalizations and surgery interventions in patients with inflammatory bowel disease:

More information

Comparative Effectiveness Research Collaborative Initiative (CER-CI) PART 1: INTERPRETING OUTCOMES RESEARCH STUDIES FOR HEALTH CARE DECISION MAKERS

Comparative Effectiveness Research Collaborative Initiative (CER-CI) PART 1: INTERPRETING OUTCOMES RESEARCH STUDIES FOR HEALTH CARE DECISION MAKERS Comparative Effectiveness Research Collaborative Initiative (CER-CI) PART 1: INTERPRETING OUTCOMES RESEARCH STUDIES FOR HEALTH CARE DECISION MAKERS ASSESSING NETWORK META ANALYSIS STUDIES: A PROPOSED MEASUREMENT

More information

Clinical research in AKI Timing of initiation of dialysis in AKI

Clinical research in AKI Timing of initiation of dialysis in AKI Clinical research in AKI Timing of initiation of dialysis in AKI Josée Bouchard, MD Krescent Workshop December 10 th, 2011 1 Acute kidney injury in ICU 15 25% of critically ill patients experience AKI

More information

Observational Study Designs. Review. Today. Measures of disease occurrence. Cohort Studies

Observational Study Designs. Review. Today. Measures of disease occurrence. Cohort Studies Observational Study Designs Denise Boudreau, PhD Center for Health Studies Group Health Cooperative Today Review cohort studies Case-control studies Design Identifying cases and controls Measuring exposure

More information

Biases in clinical research. Seungho Ryu, MD, PhD Kanguk Samsung Hospital, Sungkyunkwan University

Biases in clinical research. Seungho Ryu, MD, PhD Kanguk Samsung Hospital, Sungkyunkwan University Biases in clinical research Seungho Ryu, MD, PhD Kanguk Samsung Hospital, Sungkyunkwan University Learning objectives Describe the threats to causal inferences in clinical studies Understand the role of

More information

Evidence Informed Practice Online Learning Module Glossary

Evidence Informed Practice Online Learning Module Glossary Term Abstract Associations Attrition Bias Background and Significance Baseline Basic Science Bias Blinding Definition An abstract is a summary of a research article. It usually includes the purpose, methods,

More information

CRITICAL APPRAISAL OF CLINICAL PRACTICE GUIDELINE (CPG)

CRITICAL APPRAISAL OF CLINICAL PRACTICE GUIDELINE (CPG) CRITICAL APPRAISAL OF CLINICAL PRACTICE GUIDELINE (CPG) Atiporn Ingsathit MD, PhD. 1 Rationale for Clinical Practice Guidelines Worldwide concerns about: Unexplained variations in clinical practice Rising

More information

May 16, Division of Dockets Management (HFA-305) Food and Drug Administration 5630 Fishers Lane, Room 1061 Rockville, MD 20852

May 16, Division of Dockets Management (HFA-305) Food and Drug Administration 5630 Fishers Lane, Room 1061 Rockville, MD 20852 701 Pennsylvania Avenue, NW Suite 800 Washington, D.C. 20004 2654 Tel: 202 783 8700 Fax: 202 783 8750 www.advamed.org May 16, 2014 Division of Dockets Management (HFA-305) Food and Drug Administration

More information

Standards for the reporting of new Cochrane Intervention Reviews

Standards for the reporting of new Cochrane Intervention Reviews Methodological Expectations of Cochrane Intervention Reviews (MECIR) Standards for the reporting of new Cochrane Intervention Reviews 24 September 2012 Preface The standards below summarize proposed attributes

More information

Glossary of Practical Epidemiology Concepts

Glossary of Practical Epidemiology Concepts Glossary of Practical Epidemiology Concepts - 2009 Adapted from the McMaster EBCP Workshop 2003, McMaster University, Hamilton, Ont. Note that open access to the much of the materials used in the Epi-546

More information

CHAMP: CHecklist for the Appraisal of Moderators and Predictors

CHAMP: CHecklist for the Appraisal of Moderators and Predictors CHAMP - Page 1 of 13 CHAMP: CHecklist for the Appraisal of Moderators and Predictors About the checklist In this document, a CHecklist for the Appraisal of Moderators and Predictors (CHAMP) is presented.

More information

PubH 7405: REGRESSION ANALYSIS. Propensity Score

PubH 7405: REGRESSION ANALYSIS. Propensity Score PubH 7405: REGRESSION ANALYSIS Propensity Score INTRODUCTION: There is a growing interest in using observational (or nonrandomized) studies to estimate the effects of treatments on outcomes. In observational

More information

Systematic Reviews and Meta- Analysis in Kidney Transplantation

Systematic Reviews and Meta- Analysis in Kidney Transplantation Systematic Reviews and Meta- Analysis in Kidney Transplantation Greg Knoll MD MSc Associate Professor of Medicine Medical Director, Kidney Transplantation University of Ottawa and The Ottawa Hospital KRESCENT

More information

ISPOR Good Research Practices for Retrospective Database Analysis Task Force

ISPOR Good Research Practices for Retrospective Database Analysis Task Force ISPOR Good Research Practices for Retrospective Database Analysis Task Force II. GOOD RESEARCH PRACTICES FOR COMPARATIVE EFFECTIVENESS RESEARCH: APPROACHES TO MITIGATE BIAS AND CONFOUNDING IN THE DESIGN

More information

PHO MetaQAT Guide. Critical appraisal in public health. PHO Meta-tool for quality appraisal

PHO MetaQAT Guide. Critical appraisal in public health. PHO Meta-tool for quality appraisal PHO MetaQAT Guide Critical appraisal in public health Critical appraisal is a necessary part of evidence-based practice and decision-making, allowing us to understand the strengths and weaknesses of evidence,

More information

CLINICAL REGISTRIES Use and Emerging Best Practices

CLINICAL REGISTRIES Use and Emerging Best Practices CLINICAL REGISTRIES Use and Emerging Best Practices Tim Friede Department of Medical Statistics University Medical Center Göttingen DZHK (German Center for Cardiovascular Research) OUTLINE Background:

More information

Guideline development in TB diagnostics. Karen R Steingart, MD, MPH McGill University, Montreal, July 2011

Guideline development in TB diagnostics. Karen R Steingart, MD, MPH McGill University, Montreal, July 2011 Guideline development in TB diagnostics Karen R Steingart, MD, MPH McGill University, Montreal, July 2011 karenst@uw.edu Overview What are guidelines? Quality of guidelines The Grade approach IOM standards

More information

ISA 540, Auditing Accounting Estimates, Including Fair Value Accounting Estimates, and Related Disclosures Issues and Task Force Recommendations

ISA 540, Auditing Accounting Estimates, Including Fair Value Accounting Estimates, and Related Disclosures Issues and Task Force Recommendations Agenda Item 1-A ISA 540, Auditing Accounting Estimates, Including Fair Value Accounting Estimates, and Related Disclosures Issues and Task Force Recommendations Introduction 1. Since the September 2016

More information

CHECK-LISTS AND Tools DR F. R E Z A E I DR E. G H A D E R I K U R D I S TA N U N I V E R S I T Y O F M E D I C A L S C I E N C E S

CHECK-LISTS AND Tools DR F. R E Z A E I DR E. G H A D E R I K U R D I S TA N U N I V E R S I T Y O F M E D I C A L S C I E N C E S CHECK-LISTS AND Tools DR F. R E Z A E I DR E. G H A D E R I K U R D I S TA N U N I V E R S I T Y O F M E D I C A L S C I E N C E S What is critical appraisal? Critical appraisal is the assessment of evidence

More information

SUPPLEMENTARY DATA. Supplementary Figure S1. Search terms*

SUPPLEMENTARY DATA. Supplementary Figure S1. Search terms* Supplementary Figure S1. Search terms* *mh = exploded MeSH: Medical subject heading (Medline medical index term); tw = text word; pt = publication type; the asterisk (*) stands for any character(s) #1:

More information

Controlled Trials. Spyros Kitsiou, PhD

Controlled Trials. Spyros Kitsiou, PhD Assessing Risk of Bias in Randomized Controlled Trials Spyros Kitsiou, PhD Assistant Professor Department of Biomedical and Health Information Sciences College of Applied Health Sciences University of

More information

Washington, DC, November 9, 2009 Institute of Medicine

Washington, DC, November 9, 2009 Institute of Medicine Holger Schünemann, MD, PhD Chair, Department of Clinical Epidemiology & Biostatistics Michael Gent Chair in Healthcare Research McMaster University, Hamilton, Canada Washington, DC, November 9, 2009 Institute

More information

The Joanna Briggs Institute Reviewers Manual 2014

The Joanna Briggs Institute Reviewers Manual 2014 The Joanna Briggs Institute Reviewers Manual 2014 Summary of Findings Tables for Joanna Briggs Institute Systematic Reviews www.joannabriggs.org Joanna Briggs Institute Reviewers Manual: 2014 edition/supplement

More information

INTERNAL VALIDITY, BIAS AND CONFOUNDING

INTERNAL VALIDITY, BIAS AND CONFOUNDING OCW Epidemiology and Biostatistics, 2010 J. Forrester, PhD Tufts University School of Medicine October 6, 2010 INTERNAL VALIDITY, BIAS AND CONFOUNDING Learning objectives for this session: 1) Understand

More information

The QUOROM Statement: revised recommendations for improving the quality of reports of systematic reviews

The QUOROM Statement: revised recommendations for improving the quality of reports of systematic reviews The QUOROM Statement: revised recommendations for improving the quality of reports of systematic reviews David Moher 1, Alessandro Liberati 2, Douglas G Altman 3, Jennifer Tetzlaff 1 for the QUOROM Group

More information

Randomized Controlled Trial

Randomized Controlled Trial Randomized Controlled Trial Training Course in Sexual and Reproductive Health Research Geneva 2016 Dr Khalifa Elmusharaf MBBS, PgDip, FRSPH, PHD Senior Lecturer in Public Health Graduate Entry Medical

More information

EQUATOR Network: promises and results of reporting guidelines

EQUATOR Network: promises and results of reporting guidelines EQUATOR Network: promises and results of reporting guidelines Doug Altman The EQUATOR Network Centre for Statistics in Medicine, Oxford, UK Key principles of research publications A published research

More information

Research Questions and Survey Development

Research Questions and Survey Development Research Questions and Survey Development R. Eric Heidel, PhD Associate Professor of Biostatistics Department of Surgery University of Tennessee Graduate School of Medicine Research Questions 1 Research

More information

Challenges of Observational and Retrospective Studies

Challenges of Observational and Retrospective Studies Challenges of Observational and Retrospective Studies Kyoungmi Kim, Ph.D. March 8, 2017 This seminar is jointly supported by the following NIH-funded centers: Background There are several methods in which

More information

Downloaded from:

Downloaded from: Arnup, SJ; Forbes, AB; Kahan, BC; Morgan, KE; McKenzie, JE (2016) The quality of reporting in cluster randomised crossover trials: proposal for reporting items and an assessment of reporting quality. Trials,

More information

Epidemiologic Methods and Counting Infections: The Basics of Surveillance

Epidemiologic Methods and Counting Infections: The Basics of Surveillance Epidemiologic Methods and Counting Infections: The Basics of Surveillance Ebbing Lautenbach, MD, MPH, MSCE University of Pennsylvania School of Medicine Nothing to disclose PENN Outline Definitions / Historical

More information

Quantitative Research Methods and Tools

Quantitative Research Methods and Tools Quantitative Research Methods and Tools Fraser Health Authority, 2011 The Fraser Health Authority ( FH ) authorizes the use, reproduction and/or modification of this publication for purposes other than

More information

Title:Continuity of GP care is associated with lower use of complementary and alternative medical providers A population-based cross-sectional survey

Title:Continuity of GP care is associated with lower use of complementary and alternative medical providers A population-based cross-sectional survey Author's response to reviews Title:Continuity of GP care is associated with lower use of complementary and alternative medical providers A population-based cross-sectional survey Authors: Anne Helen Hansen

More information

Biases in clinical research. Seungho Ryu, MD, PhD Kanguk Samsung Hospital, Sungkyunkwan University

Biases in clinical research. Seungho Ryu, MD, PhD Kanguk Samsung Hospital, Sungkyunkwan University Biases in clinical research Seungho Ryu, MD, PhD Kanguk Samsung Hospital, Sungkyunkwan University Learning objectives Understand goal of measurement and definition of accuracy Describe the threats to causal

More information

1. Draft checklist for judging on quality of animal studies (Van der Worp et al., 2010)

1. Draft checklist for judging on quality of animal studies (Van der Worp et al., 2010) Appendix C Quality appraisal tools (QATs) 1. Draft checklist for judging on quality of animal studies (Van der Worp et al., 2010) 2. NICE Checklist for qualitative studies (2012) 3. The Critical Appraisal

More information

Protocol Development: The Guiding Light of Any Clinical Study

Protocol Development: The Guiding Light of Any Clinical Study Protocol Development: The Guiding Light of Any Clinical Study Susan G. Fisher, Ph.D. Chair, Department of Clinical Sciences 1 Introduction Importance/ relevance/ gaps in knowledge Specific purpose of the

More information

Critical Review Form Clinical Decision Analysis

Critical Review Form Clinical Decision Analysis Critical Review Form Clinical Decision Analysis An Interdisciplinary Initiative to Reduce Radiation Exposure: Evaluation of Appendicitis in a Pediatric Emergency Department with Clinical Assessment Supported

More information

Finland and Sweden and UK GP-HOSP datasets

Finland and Sweden and UK GP-HOSP datasets Web appendix: Supplementary material Table 1 Specific diagnosis codes used to identify bladder cancer cases in each dataset Finland and Sweden and UK GP-HOSP datasets Netherlands hospital and cancer registry

More information

Meta-Analysis. Zifei Liu. Biological and Agricultural Engineering

Meta-Analysis. Zifei Liu. Biological and Agricultural Engineering Meta-Analysis Zifei Liu What is a meta-analysis; why perform a metaanalysis? How a meta-analysis work some basic concepts and principles Steps of Meta-analysis Cautions on meta-analysis 2 What is Meta-analysis

More information

GLOSSARY OF GENERAL TERMS

GLOSSARY OF GENERAL TERMS GLOSSARY OF GENERAL TERMS Absolute risk reduction Absolute risk reduction (ARR) is the difference between the event rate in the control group (CER) and the event rate in the treated group (EER). ARR =

More information

Fitting the Method to the Question

Fitting the Method to the Question Quantitative Research Design: Fitting the Method to the Question Deborah Eldredge, PhD, RN Director, Nursing Quality, Research & Magnet Oregon Health & Science University Healthcare Margo A. Halm, RN,

More information

Appendix 2 Quality assessment tools. Cochrane risk of bias tool for RCTs. Support for judgment

Appendix 2 Quality assessment tools. Cochrane risk of bias tool for RCTs. Support for judgment Appendix 2 Quality assessment tools Cochrane risk of bias tool for RCTs Item Judgment (H/L/Unclear) Random sequence generation (selection bias) Allocation concealment (selection bias) Blinding of participants

More information

About Reading Scientific Studies

About Reading Scientific Studies About Reading Scientific Studies TABLE OF CONTENTS About Reading Scientific Studies... 1 Why are these skills important?... 1 Create a Checklist... 1 Introduction... 1 Abstract... 1 Background... 2 Methods...

More information

OBSERVATIONAL MEDICAL OUTCOMES PARTNERSHIP

OBSERVATIONAL MEDICAL OUTCOMES PARTNERSHIP OBSERVATIONAL Patient-centered observational analytics: New directions toward studying the effects of medical products Patrick Ryan on behalf of OMOP Research Team May 22, 2012 Observational Medical Outcomes

More information

Trials and Tribulations of Systematic Reviews and Meta-Analyses

Trials and Tribulations of Systematic Reviews and Meta-Analyses Trials and Tribulations of Systematic Reviews and Meta-Analyses Mark A. Crowther and Deborah J. Cook St. Joseph s Hospital, Hamilton, Ontario, Canada; McMaster University, Hamilton, Ontario, Canada Systematic

More information

Improved Transparency in Key Operational Decisions in Real World Evidence

Improved Transparency in Key Operational Decisions in Real World Evidence PharmaSUG 2018 - Paper RW-06 Improved Transparency in Key Operational Decisions in Real World Evidence Rebecca Levin, Irene Cosmatos, Jamie Reifsnyder United BioSource Corp. ABSTRACT The joint International

More information

Live WebEx meeting agenda

Live WebEx meeting agenda 10:00am 10:30am Using OpenMeta[Analyst] to extract quantitative data from published literature Live WebEx meeting agenda August 25, 10:00am-12:00pm ET 10:30am 11:20am Lecture (this will be recorded) 11:20am

More information

CHECKLIST FOR EVALUATING A RESEARCH REPORT Provided by Dr. Blevins

CHECKLIST FOR EVALUATING A RESEARCH REPORT Provided by Dr. Blevins CHECKLIST FOR EVALUATING A RESEARCH REPORT Provided by Dr. Blevins 1. The Title a. Is it clear and concise? b. Does it promise no more than the study can provide? INTRODUCTION 2. The Problem a. It is clearly

More information

Rapid appraisal of the literature: Identifying study biases

Rapid appraisal of the literature: Identifying study biases Rapid appraisal of the literature: Identifying study biases Rita Popat, PhD Clinical Assistant Professor Division of Epidemiology Stanford University School of Medicine August 7, 2007 What is critical

More information

Guidelines for Reporting Non-Randomised Studies

Guidelines for Reporting Non-Randomised Studies Revised and edited by Renatus Ziegler B.C. Reeves a W. Gaus b a Department of Public Health and Policy, London School of Hygiene and Tropical Medicine, Great Britain b Biometrie und Medizinische Dokumentation,

More information

Learning objectives. Examining the reliability of published research findings

Learning objectives. Examining the reliability of published research findings Examining the reliability of published research findings Roger Chou, MD Associate Professor of Medicine Department of Medicine and Department of Medical Informatics and Clinical Epidemiology Scientific

More information

Checklist for Case Control Studies. The Joanna Briggs Institute Critical Appraisal tools for use in JBI Systematic Reviews

Checklist for Case Control Studies. The Joanna Briggs Institute Critical Appraisal tools for use in JBI Systematic Reviews The Joanna Briggs Institute Critical Appraisal tools for use in JBI Systematic Reviews Checklist for http://joannabriggs.org/research/critical-appraisal-tools.html www.joannabriggs.org The Joanna Briggs

More information

Fitting the Method to the Question

Fitting the Method to the Question Quantitative Research Design: Fitting the Method to the Question Deborah Eldredge, PhD, RN Director, Nursing Quality, Research & Magnet Oregon Health & Science University Healthcare Margo A. Halm, RN,

More information

Evidence-Based Medicine and Publication Bias Desmond Thompson Merck & Co.

Evidence-Based Medicine and Publication Bias Desmond Thompson Merck & Co. Evidence-Based Medicine and Publication Bias Desmond Thompson Merck & Co. Meta-Analysis Defined A meta-analysis is: the statistical combination of two or more separate studies In other words: overview,

More information

Cochrane Pregnancy and Childbirth Group Methodological Guidelines

Cochrane Pregnancy and Childbirth Group Methodological Guidelines Cochrane Pregnancy and Childbirth Group Methodological Guidelines [Prepared by Simon Gates: July 2009, updated July 2012] These guidelines are intended to aid quality and consistency across the reviews

More information

Clinical Research Scientific Writing. K. A. Koram NMIMR

Clinical Research Scientific Writing. K. A. Koram NMIMR Clinical Research Scientific Writing K. A. Koram NMIMR Clinical Research Branch of medical science that determines the safety and effectiveness of medications, devices, diagnostic products and treatment

More information

CHAPTER VI RESEARCH METHODOLOGY

CHAPTER VI RESEARCH METHODOLOGY CHAPTER VI RESEARCH METHODOLOGY 6.1 Research Design Research is an organized, systematic, data based, critical, objective, scientific inquiry or investigation into a specific problem, undertaken with the

More information

Results. NeuRA Worldwide incidence April 2016

Results. NeuRA Worldwide incidence April 2016 Introduction The incidence of schizophrenia refers to how many new cases there are per population in a specified time period. It is different from prevalence, which refers to how many existing cases there

More information

Declaration of interests. Register-based research on safety and effectiveness opportunities and challenges 08/04/2018

Declaration of interests. Register-based research on safety and effectiveness opportunities and challenges 08/04/2018 Register-based research on safety and effectiveness opportunities and challenges Morten Andersen Department of Drug Design And Pharmacology Declaration of interests Participate(d) in research projects

More information

GRADE. Grading of Recommendations Assessment, Development and Evaluation. British Association of Dermatologists April 2018

GRADE. Grading of Recommendations Assessment, Development and Evaluation. British Association of Dermatologists April 2018 GRADE Grading of Recommendations Assessment, Development and Evaluation British Association of Dermatologists April 2018 Previous grading system Level of evidence Strength of recommendation Level of evidence

More information

Confounding and Bias

Confounding and Bias 28 th International Conference on Pharmacoepidemiology and Therapeutic Risk Management Barcelona, Spain August 22, 2012 Confounding and Bias Tobias Gerhard, PhD Assistant Professor, Ernest Mario School

More information

Results. NeuRA Treatments for internalised stigma December 2017

Results. NeuRA Treatments for internalised stigma December 2017 Introduction Internalised stigma occurs within an individual, such that a person s attitude may reinforce a negative self-perception of mental disorders, resulting in reduced sense of selfworth, anticipation

More information

Critical appraisal: Systematic Review & Meta-analysis

Critical appraisal: Systematic Review & Meta-analysis Critical appraisal: Systematic Review & Meta-analysis Atiporn Ingsathit MD.PhD. Section for Clinical Epidemiology and biostatistics Faculty of Medicine Ramathibodi Hospital Mahidol University What is a

More information

BIOSTATISTICAL METHODS

BIOSTATISTICAL METHODS BIOSTATISTICAL METHODS FOR TRANSLATIONAL & CLINICAL RESEARCH PROPENSITY SCORE Confounding Definition: A situation in which the effect or association between an exposure (a predictor or risk factor) and

More information

The comparison or control group may be allocated a placebo intervention, an alternative real intervention or no intervention at all.

The comparison or control group may be allocated a placebo intervention, an alternative real intervention or no intervention at all. 1. RANDOMISED CONTROLLED TRIALS (Treatment studies) (Relevant JAMA User s Guides, Numbers IIA & B: references (3,4) Introduction: The most valid study design for assessing the effectiveness (both the benefits

More information

Improving reporting for observational studies: STROBE statement

Improving reporting for observational studies: STROBE statement Improving reporting for observational studies: STROBE statement Technical meeting on the reporting of human studies submitted for the scientific substantiation of health claims EFSA Parma 20 November 2013

More information

Checklist for appraisal of study relevance (child sex offenses)

Checklist for appraisal of study relevance (child sex offenses) Appendix 3 Evaluation protocols. [posted as supplied by author] Checklist for appraisal of study relevance (child sex offenses) First author, year, reference number Relevance Yes No Cannot answer Not applicable

More information

Biases in clinical research. Seungho Ryu, MD, PhD Kanguk Samsung Hospital, Sungkyunkwan University

Biases in clinical research. Seungho Ryu, MD, PhD Kanguk Samsung Hospital, Sungkyunkwan University Biases in clinical research Seungho Ryu, MD, PhD Kanguk Samsung Hospital, Sungkyunkwan University Learning objectives Describe the threats to causal inferences in clinical studies Understand the role of

More information

Blood Pressure and Complications in Individuals with Type 2 Diabetes and No Previous Cardiovascular Disease. ID BMJ

Blood Pressure and Complications in Individuals with Type 2 Diabetes and No Previous Cardiovascular Disease. ID BMJ 1 Blood Pressure and Complications in Individuals with Type 2 Diabetes and No Previous Cardiovascular Disease. ID BMJ 2016.033440 Dear Editor, Editorial Committee and Reviewers Thank you for your appreciation

More information

Complications of Proton Pump Inhibitor Therapy. Gastroenterology 2017; 153:35-48 발표자 ; F1 김선화

Complications of Proton Pump Inhibitor Therapy. Gastroenterology 2017; 153:35-48 발표자 ; F1 김선화 Complications of Proton Pump Inhibitor Therapy Gastroenterology 2017; 153:35-48 발표자 ; F1 김선화 Background Proton pump inhibitors (PPIs) are among the most commonly prescribed medicines for gastroesophageal

More information

Nature and significance of the local problem

Nature and significance of the local problem Revised Standards for Quality Improvement Reporting Excellence (SQUIRE 2.0) September 15, 2015 Text Section and Item Section or Item Description Name The SQUIRE guidelines provide a framework for reporting

More information

Meta-analysis of safety thoughts from CIOMS X

Meta-analysis of safety thoughts from CIOMS X CIOMS Working Group X Meta-analysis of safety thoughts from CIOMS X Stephen.Evans@Lshtm.ac.uk Improving health worldwide www.lshtm.ac.uk Evans: ENCePP CIOMS Meta Analysis 1 Acknowledgements, conflicts,

More information

NATIONAL INSTITUTE FOR HEALTH AND CLINICAL EXCELLENCE

NATIONAL INSTITUTE FOR HEALTH AND CLINICAL EXCELLENCE NATIONAL INSTITUTE FOR HEALTH AND CLINICAL EXCELLENCE QUALITY AND OUTCOMES FRAMEWORK (QOF) INDICATOR DEVELOPMENT PROGRAMME Briefing paper QOF indicator area: Peripheral arterial disease Potential output:

More information

Fixed Effect Combining

Fixed Effect Combining Meta-Analysis Workshop (part 2) Michael LaValley December 12 th 2014 Villanova University Fixed Effect Combining Each study i provides an effect size estimate d i of the population value For the inverse

More information

Traumatic brain injury

Traumatic brain injury Introduction It is well established that traumatic brain injury increases the risk for a wide range of neuropsychiatric disturbances, however there is little consensus on whether it is a risk factor for

More information

Epidemiologic Methods I & II Epidem 201AB Winter & Spring 2002

Epidemiologic Methods I & II Epidem 201AB Winter & Spring 2002 DETAILED COURSE OUTLINE Epidemiologic Methods I & II Epidem 201AB Winter & Spring 2002 Hal Morgenstern, Ph.D. Department of Epidemiology UCLA School of Public Health Page 1 I. THE NATURE OF EPIDEMIOLOGIC

More information

University of Wollongong. Research Online. Australian Health Services Research Institute

University of Wollongong. Research Online. Australian Health Services Research Institute University of Wollongong Research Online Australian Health Services Research Institute Faculty of Business 2011 Measurement of error Janet E. Sansoni University of Wollongong, jans@uow.edu.au Publication

More information

Evidence-Based Medicine Journal Club. A Primer in Statistics, Study Design, and Epidemiology. August, 2013

Evidence-Based Medicine Journal Club. A Primer in Statistics, Study Design, and Epidemiology. August, 2013 Evidence-Based Medicine Journal Club A Primer in Statistics, Study Design, and Epidemiology August, 2013 Rationale for EBM Conscientious, explicit, and judicious use Beyond clinical experience and physiologic

More information

Lecture Slides. Elementary Statistics Eleventh Edition. by Mario F. Triola. and the Triola Statistics Series 1.1-1

Lecture Slides. Elementary Statistics Eleventh Edition. by Mario F. Triola. and the Triola Statistics Series 1.1-1 Lecture Slides Elementary Statistics Eleventh Edition and the Triola Statistics Series by Mario F. Triola 1.1-1 Chapter 1 Introduction to Statistics 1-1 Review and Preview 1-2 Statistical Thinking 1-3

More information

Uses and misuses of the STROBE statement: bibliographic study

Uses and misuses of the STROBE statement: bibliographic study Uses and misuses of the STROBE statement: bibliographic study Bruno R. da Costa 1, Myriam Cevallos 1, 2, Douglas G. Altman 3, Anne W.S. Rutjes 1, Matthias Egger 1 1. Institute of Social & Preventive Medicine

More information

Incorporating Clinical Information into the Label

Incorporating Clinical Information into the Label ECC Population Health Group LLC expertise-driven consulting in global maternal-child health & pharmacoepidemiology Incorporating Clinical Information into the Label Labels without Categories: A Workshop

More information

Results. NeuRA Hypnosis June 2016

Results. NeuRA Hypnosis June 2016 Introduction may be experienced as an altered state of consciousness or as a state of relaxation. There is no agreed framework for administering hypnosis, but the procedure often involves induction (such

More information

Propensity Score Methods for Estimating Causality in the Absence of Random Assignment: Applications for Child Care Policy Research

Propensity Score Methods for Estimating Causality in the Absence of Random Assignment: Applications for Child Care Policy Research 2012 CCPRC Meeting Methodology Presession Workshop October 23, 2012, 2:00-5:00 p.m. Propensity Score Methods for Estimating Causality in the Absence of Random Assignment: Applications for Child Care Policy

More information

Real-world data in pragmatic trials

Real-world data in pragmatic trials Real-world data in pragmatic trials Harold C. Sox,MD The Patient-Centered Outcomes Research Institute Washington, DC Presenter Disclosure Information In compliance with the accrediting board policies,

More information

Essential Skills for Evidence-based Practice Understanding and Using Systematic Reviews

Essential Skills for Evidence-based Practice Understanding and Using Systematic Reviews J Nurs Sci Vol.28 No.4 Oct - Dec 2010 Essential Skills for Evidence-based Practice Understanding and Using Systematic Reviews Jeanne Grace Corresponding author: J Grace E-mail: Jeanne_Grace@urmc.rochester.edu

More information

Amyotrophic Lateral Sclerosis: Developing Drugs for Treatment Guidance for Industry

Amyotrophic Lateral Sclerosis: Developing Drugs for Treatment Guidance for Industry Amyotrophic Lateral Sclerosis: Developing Drugs for Treatment Guidance for Industry DRAFT GUIDANCE This guidance document is being distributed for comment purposes only. Comments and suggestions regarding

More information

Results. NeuRA Mindfulness and acceptance therapies August 2018

Results. NeuRA Mindfulness and acceptance therapies August 2018 Introduction involve intentional and non-judgmental focus of one's attention on emotions, thoughts and sensations that are occurring in the present moment. The aim is to open awareness to present experiences,

More information

JAMA. 2011;305(24): Nora A. Kalagi, MSc

JAMA. 2011;305(24): Nora A. Kalagi, MSc JAMA. 2011;305(24):2556-2564 By Nora A. Kalagi, MSc Cardiovascular disease (CVD) is the number one cause of mortality and morbidity world wide Reducing high blood cholesterol which is a risk factor for

More information

THE DEPARTMENT OF VETERANS AFFAIRS HEALTH SERVICES RESEARCH & DEVELOPMENT SERVICE 810 Vermont Avenue Washington, DC 20420

THE DEPARTMENT OF VETERANS AFFAIRS HEALTH SERVICES RESEARCH & DEVELOPMENT SERVICE 810 Vermont Avenue Washington, DC 20420 HSR&D Service Document #39-32-010 THE DEPARTMENT OF VETERANS AFFAIRS HEALTH SERVICES RESEARCH & DEVELOPMENT SERVICE 810 Vermont Avenue Washington, DC 20420 June 1989 Common Methodological Problems in Investigator

More information

Evaluating Scientific Journal Articles. Tufts CTSI s Mission & Purpose. Tufts Clinical and Translational Science Institute

Evaluating Scientific Journal Articles. Tufts CTSI s Mission & Purpose. Tufts Clinical and Translational Science Institute Tufts Clinical and Translational Science Institute Lori Lyn Price, MAS Biostatistics, Epidemiology, and Research Design (BERD) Tufts Clinical and Translational Science Institute (CTSI) Tufts CTSI s Mission

More information

Appraising the Literature Overview of Study Designs

Appraising the Literature Overview of Study Designs Chapter 5 Appraising the Literature Overview of Study Designs Barbara M. Sullivan, PhD Department of Research, NUHS Jerrilyn A. Cambron, PhD, DC Department of Researach, NUHS EBP@NUHS Ch 5 - Overview of

More information

Analysis A step in the research process that involves describing and then making inferences based on a set of data.

Analysis A step in the research process that involves describing and then making inferences based on a set of data. 1 Appendix 1:. Definitions of important terms. Additionality The difference between the value of an outcome after the implementation of a policy, and its value in a counterfactual scenario in which the

More information

The ROBINS-I tool is reproduced from riskofbias.info with the permission of the authors. The tool should not be modified for use.

The ROBINS-I tool is reproduced from riskofbias.info with the permission of the authors. The tool should not be modified for use. Table A. The Risk Of Bias In Non-romized Studies of Interventions (ROBINS-I) I) assessment tool The ROBINS-I tool is reproduced from riskofbias.info with the permission of the auths. The tool should not

More information