GLOSSARY OF GENERAL TERMS

Similar documents
the standard deviation (SD) is a measure of how much dispersion exists from the mean SD = square root (variance)

Critical Appraisal Series

Evidence-based practice tutorial Critical Appraisal Skills

Glossary of Practical Epidemiology Concepts

GRADE. Grading of Recommendations Assessment, Development and Evaluation. British Association of Dermatologists April 2018

Evidence Based Medicine

Vocabulary. Bias. Blinding. Block. Cluster sample

GATE: Graphic Appraisal Tool for Epidemiology picture, 2 formulas & 3 acronyms

Critical Appraisal of Evidence

Clinical Epidemiology for the uninitiated

GRADE. Grading of Recommendations Assessment, Development and Evaluation. British Association of Dermatologists April 2014

Evidence Informed Practice Online Learning Module Glossary

The comparison or control group may be allocated a placebo intervention, an alternative real intervention or no intervention at all.

11 questions to help you make sense of a case control study

Lecture 5 Conducting Interviews and Focus Groups

RATING OF A RESEARCH PAPER. By: Neti Juniarti, S.Kp., M.Kes., MNurs

GATE CAT Intervention RCT/Cohort Studies

CRITICAL APPRAISAL WORKSHEET

Outline. What is Evidence-Based Practice? EVIDENCE-BASED PRACTICE. What EBP is Not:

Clinical problems and choice of study designs

Critical Appraisal of Evidence A Focus on Intervention/Treatment Studies

GATE: Graphic Appraisal Tool for Epidemiology picture, 2 formulas & 3 acronyms

About Reading Scientific Studies

Essential Skills for Evidence-based Practice: Statistics for Therapy Questions

Chapter 13 Summary Experiments and Observational Studies

Journal Club Critical Appraisal Worksheets. Dr David Walbridge

Module 5. The Epidemiological Basis of Randomised Controlled Trials. Landon Myer School of Public Health & Family Medicine, University of Cape Town

Chapter 13. Experiments and Observational Studies. Copyright 2012, 2008, 2005 Pearson Education, Inc.

Critical Appraisal. Dave Abbott Senior Medicines Information Pharmacist

CRITICAL APPRAISAL WORKSHEET 1

MCQ Course in Pediatrics Al Yamamah Hospital June Dr M A Maleque Molla, FRCP, FRCPCH

Guidelines for Writing and Reviewing an Informed Consent Manuscript From the Editors of Clinical Research in Practice: The Journal of Team Hippocrates

Consider the following hypothetical

Journal Club: Fairfax Internal Medicine. Stephanie Shin PGY-2 Bianca Ummat PGY-2 July 27, 2012

ACR OA Guideline Development Process Knee and Hip

Copyright GRADE ING THE QUALITY OF EVIDENCE AND STRENGTH OF RECOMMENDATIONS NANCY SANTESSO, RD, PHD

Overview of Study Designs in Clinical Research

Journal Club Handbook

CHECK-LISTS AND Tools DR F. R E Z A E I DR E. G H A D E R I K U R D I S TA N U N I V E R S I T Y O F M E D I C A L S C I E N C E S

1. Draft checklist for judging on quality of animal studies (Van der Worp et al., 2010)

Cochrane Pregnancy and Childbirth Group Methodological Guidelines

Are the likely benefits worth the potential harms and costs? From McMaster EBCP Workshop/Duke University Medical Center

Funnelling Used to describe a process of narrowing down of focus within a literature review. So, the writer begins with a broad discussion providing b

THERAPY WORKSHEET: page 1 of 2 adapted from Sackett 1996

Chapter 11: Experiments and Observational Studies p 318

Systematic Reviews and Meta- Analysis in Kidney Transplantation

Checklist for appraisal of study relevance (child sex offenses)

The Royal College of Pathologists Journal article evaluation questions

CONSORT 2010 checklist of information to include when reporting a randomised trial*

Critical Appraisal of a Meta-Analysis: Rosiglitazone and CV Death. Debra Moy Faculty of Pharmacy University of Toronto

Chapter Three Research Methodology

Evidence Based Medicine

Unit 1 Exploring and Understanding Data

4/10/2018. Choosing a study design to answer a specific research question. Importance of study design. Types of study design. Types of study design

Meta Analysis. David R Urbach MD MSc Outcomes Research Course December 4, 2014

Common Statistical Issues in Biomedical Research

Thursday, April 25, 13. Intervention Studies

Overview. Goals of Interpretation. Methodology. Reasons to Read and Evaluate

ISC- GRADE XI HUMANITIES ( ) PSYCHOLOGY. Chapter 2- Methods of Psychology

Randomized Controlled Trial

Purpose. Study Designs. Objectives. Observational Studies. Analytic Studies

Learning Objectives 9/9/2013. Hypothesis Testing. Conflicts of Interest. Descriptive statistics: Numerical methods Measures of Central Tendency

Role of evidence from observational studies in the process of health care decision making

lab exam lab exam Experimental Design Experimental Design when: Nov 27 - Dec 1 format: length = 1 hour each lab section divided in two

Thursday, April 25, 13. Intervention Studies

Appraising the Literature Overview of Study Designs

Types of Data. Systematic Reviews: Data Synthesis Professor Jodie Dodd 4/12/2014. Acknowledgements: Emily Bain Australasian Cochrane Centre

New Math: Taking on Overdiagnosis Waste and Harm Flipped Classroom: Pre-work

Chapter 13. Experiments and Observational Studies

Recent developments for combining evidence within evidence streams: bias-adjusted meta-analysis

A to Z OF RESEARCH METHODS AND TERMS APPLICABLE WITHIN SOCIAL SCIENCE RESEARCH

Systematic reviews: From evidence to recommendation. Marcel Dijkers, PhD, FACRM Icahn School of Medicine at Mount Sinai

The role of Randomized Controlled Trials

CRITICAL EVALUATION OF BIOMEDICAL LITERATURE

STUDY DESIGNS WHICH ONE IS BEST?

Journal Club September 29, Vanessa AKIKI PGYlII Internal Medicine

Glossary. Ó 2010 John Wiley & Sons, Ltd

Delfini Evidence Tool Kit

observational studies Descriptive studies

Design strategies in quantitative research an introduction

Epidemiological study design. Paul Pharoah Department of Public Health and Primary Care

The Practice of Statistics 1 Week 2: Relationships and Data Collection

Problem solving therapy

Dr. Engle has received an unrestricted educational grant for the MD Anderson Pain Medicine Fellowship.

GLOSSARY of research terms

Higher Psychology RESEARCH REVISION

9/4/2013. Decision Errors. Hypothesis Testing. Conflicts of Interest. Descriptive statistics: Numerical methods Measures of Central Tendency

17/10/2012. Could a persistent cough be whooping cough? Epidemiology and Statistics Module Lecture 3. Sandra Eldridge

Appendix A: Literature search strategy

SOCQ121/BIOQ121. Session 8. Methods and Methodology. Department of Social Science. endeavour.edu.au

The RoB 2.0 tool (individually randomized, cross-over trials)

\ jar gon \ BUSTER. For research terms A B C D E F G H I J K L M N O P Q R S T U V W X Y Z 1

The Cochrane Collaboration

CRITICAL APPRAISAL SKILLS PROGRAMME Making sense of evidence about clinical effectiveness. 11 questions to help you make sense of case control study

Research Methods. It is actually way more exciting than it sounds!!!!

Critical Appraisal Istanbul 2011

Checklist for Randomized Controlled Trials. The Joanna Briggs Institute Critical Appraisal tools for use in JBI Systematic Reviews

How to use this appraisal tool: Three broad issues need to be considered when appraising a case control study:

Overview of Study Designs

Theory. = an explanation using an integrated set of principles that organizes observations and predicts behaviors or events.

Transcription:

GLOSSARY OF GENERAL TERMS Absolute risk reduction Absolute risk reduction (ARR) is the difference between the event rate in the control group (CER) and the event rate in the treated group (EER). ARR = CER EER. Blinding The term blinding (also masking ) is used in randomised controlled trials and includes single blinding and double blinding. Single blinding occurs when the participants are unaware of which intervention they are receiving. Double blinding occurs when both the participants and the health professionals are unaware of who is receiving the assessed intervention and the control interventions. Case control study Increasingly used to investigate causes of diseases, especially rare diseases. A case control study involves the comparison of those with a given condition (cases) and those without the condition (controls). Levels of exposure to a factor are compared between the two groups. Case control studies have been called retrospective studies. Case study In-depth analysis and systematic description of one patient or group of similar patients to promote a detailed understanding of their circumstance. Chi-squared A test statistic used to assess the statistical significance of a finding, based on the difference between the observed frequency of an event and that which would be expected if the null hypothesis were true. Clinical audit The systematic and critical analysis of the quality of clinical care, including the procedures used for the diagnosis, treatment and care, the associated use of resources and the resulting outcome and quality of life for the patient or client. Clinical effectiveness The extent to which specific clinical interventions, when deployed in the field for a particular patient or population, do what they are intended to do, i.e. maintain and improve health and secure the greatest possible health gain from the available resources. To be reasonably certain that an intervention has produced health benefits, it needs to be shown to be capable of producing worthwhile benefit (efficacy and cost effectiveness) and to have produced that benefit in practice. Clinical governance A framework through which NHS organisations are accountable for continuously improving the quality of their services. Page 1 of 6

Clinical and health outcomes These refer to the extent to which the expected health benefit (see clinical effectiveness) is achieved and can be attributed to the relevant clinical and health interventions. The NHS Executive uses the following working definition of health outcome: the attributable effect of intervention or its lack on a previous health state. Clinical significance Whether a treatment is relevant to the patient in practice this is different to statistical significance. For example, a trial could show a statistically significant difference between a treatment and a control, reducing mean length of symptoms from 7 to 6.5 days. Everyone in the treatment group gets better 0.5 days more quickly, which may be statistically significant, but makes little difference in real life. It is therefore unlikely that the new treatment would be widely adopted; it is of little clinical significance. Cohort study A cohort study (also known as a follow-up study or a longitudinal study) involves the identification of two groups (cohorts), one of which is exposed and the other not exposed to a potential factor. These cohorts are followed up in time, and the incidence of the outcome in one group compared with the incidence in the other. Confidence interval (CI) A confidence interval quantifies the uncertainty in measurement. It is usually reported as 95% CI, which is the range of values within which we can be 95% sure that the true value for the whole population lies. Controls Controls in an RCT are people in a comparison group. They are allocated not to receive the new intervention which is allocated to people in the experimental group. They could instead receive a placebo, conventional therapy or no treatment at all. Control event rate (CER) This is the event rate or risk in the control group. See also Event rate. Cost-effectiveness The cost-effectiveness of a particular form of health care depends upon the ratio of the costs of health care to its health outcomes. Critical appraisal The process of assessing and interpreting evidence, by systematically considering its validity, results and relevance to practice. Crossover study A type of trial comparing experimental and control interventions in which the participants are divided into two groups. One group starts with the control intervention and switches halfway through the trial to the experimental intervention. The other group does the opposite. Cross-sectional study A study in which information on the putative risk factors and the putative outcomes are measured simultaneously at one point in time. A survey is an example, or in epidemiology, a measure of prevalence. Page 2 of 6

Event rate Event rate (sometimes called risk) is the proportion of patients in a group in whom the event is observed. Thus, if out of 100 patients, the event is observed in 27, the event rate is 0.27. Control event rate (CER) and experimental event rate (EER) are terms used to refer to event rates in control and experimental groups of patients respectively. Experiment An experiment sets out to test causal relationships by contrasting one or more variables and predictions or hypothesised outcomes. Experimental event rate (EER) This is the event rate or risk in the experimental group. See also Event rate. Experimental group The experimental group are the group of people in an RCT who are subject to an intervention which can be a treatment, a change in practice or procedure or the introduction of a screening programme. Follow-up Results obtained from the participants in a trial all the way through the study period. Losing patients in the follow-up phase can distort the results. Hawthorne effect Psychological response in which subjects change their behaviour simply because they are participants in a study, not because of the treatment being researched. Heterogeneity A statistical test often used in meta-analysis. It determines if there are any differences between studies, in terms of the population, intervention or outcome, which could influence the intervention effect and therefore make it inappropriate to combine the studies statistically. Homogeneity Homogeneity means similarity. Studies are said to be homogeneous if their results vary no more than might be expected due to chance. The opposite of homogeneity is heterogeneity. Incidence The rate at which new cases occur in a population during a specified period. Inclusion criteria The criteria used by authors of a review to decide whether to include studies. Intervention An intervention can be a treatment, a change in practice or procedure or the introduction of a screening programme. Mean The average value, which is calculated by adding all the measurements and dividing by the number of measurements. Page 3 of 6

Median Is the value on the scale that divides the distribution into two equal parts. Half of the observations have a value less than or equal to the median, and half have a value greater than or equal to the median. Meta-analysis A statistical technique which summarises the results of several studies into a single estimate, giving more weight to results from larger studies. Methodology Methodology is the approach or design, including the methods used in a piece of work. Mode The most frequently occurring value in a set of observations. Null hypothesis The proposition that there is no difference between, for example, two treatments other than that due to chance. Number needed to harm (NNH) NNH is the number of patients who need to be treated to cause one additional bad outcome (eg side effects). For example, in a trial where side effects are one of the outcomes, if NNH = 10, for every 10 people treated one extra person will suffer side effects. Number needed to treat (NNT) NNT is the number of patients who need to be treated to benefit one additional person. For example, if the NNT in a trial looking at giving up smoking by using nicotine replacement therapy = 14, you need to treat 14 people for one additional person to stop smoking. Odds The quantified probability of an event happening. Odds ratio (OR) The odds in the treatment group divided by the odds in the control group. It is one measure of a treatment s clinical effectiveness. Outcome Result of an intervention. Outcomes can be desirable, such as improvement in the patient s condition or quality of life, or undesirable, such as side effects. Peer-reviewed Of an article published in a journal, checked by other experts to ensure that the authors have used sound methods and described their methods in sufficient detail to allow others to try and reproduce their results. Placebo therapy Placebo therapy is a biologically inert treatment, often given to controls in trials. It can help blinding if patients in the control group are given dummy inactive tablets which look and taste like the real intervention, so that patients do not know which treatments group they are in. Page 4 of 6

Population The people that you or the researchers are interested in. Information about them might include their age, gender and state of health. Prevalence The proportion of a population that are cases at a point in time. Prognosis The expected outcome, predicted on the basis of the normal course of a disease. P-value The probability that an observed difference in the results of a study could have occurred by chance. The convention is to believe that if p is greater than 5% (0.05), then the difference may have occurred by chance and not be due entirely to the intervention. Randomisation In an RCT, subjects in a population are randomly allocated to groups, usually called treatment and control groups. Allocation of participants to groups is therefore determined by chance. Robust randomisation procedures means that all individuals have the same probability of being allocated either to the experimental or to the control groups. Randomised controlled trial (RCT) A RCT is a trial in which subjects are randomly assigned to two groups: one (the experimental group) receiving the intervention that is being tested, and the other (the comparison group or control group) receiving an alternative treatment. The results of the trial are assessed by comparing the outcomes in the different groups. The RCT is the gold standard for assessing the effectiveness of an intervention because this study design has the aim of reducing bias and confounding. Random error Random error or random variation refers to the differences in results that are due to chance rather than to one of the other variables being studied. Differences caused by random error cause results to be scattered randomly about the mean or best estimate. Reliability In naturalistic research, reliability is the process of establishing that data analysis and coding remain constant when reviewed at different times by the same researcher (stability) or another researcher (reproducibility). Reproducible Capable of being reproduced. Authors of research reports should describe their methods thoroughly so that others can try to reproduce their results. Review Any summary of a particular topic. Risk The chances of something happening. Researchers often use the work risk to denote the proportion of patients in a group in whom an event is observed. Another phrase used for risk in this sense is event rate. Page 5 of 6

Risk ratio The ratio of risk in the treated group to the risk in the control group: Risk ratio is the risk in the experimental or treatment group (EER) divided by that in the control group (CER). Sometimes called relative risk (RR). RR = EER divided by CER. Sampling The process of selecting participants for research on the basis that they can provide detailed information relevant to the enquiry. Screening A diagnostic test (used on a person or group) for the presence or absence of a disease, or for risk factors for an increased probability of disease. Secondary publication A publication that selects and publishes evidence from other publications, usually with expert commentary. An example could be a book or journals, such as Evidence- Based Medicine, which use primary publications as resources. Sensitivity analysis Sensitivity analysis is a statistical technique used to see how the results of a trial or review might be changed by doubt about the data, participants who have dropped out during the course of the research or changes in research methods. Sets Groups of related items. Standard deviation The mean, median and mode are measures of central tendency and are useful for summarising a frequency distribution, but they do not indicate the spread of values. The standard deviation measures the amount of scatter in results. Approximately two-thirds of the values will fall within one standard deviation of the mean and 95% fall within two standard deviations of the mean. Statistically significant A result that is very unlikely to have happened by chance is often described as statistically significant. Researchers often use statistical tests such as chi squared and the probability value to check whether their results are statistically significant. Systematic review A review in which evidence on a topic has been systematically identified, appraised and summarised according to predetermined criteria. Transferability The extent to which findings from the data can be applied to other settings or groups. Validity Refers to the soundness or rigour of a study. A study is valid if the way it is designed and carried out means that the results are unbiased that is, it gives you a true estimate of clinical effectiveness. Page 6 of 6