Sensitivity Analysis in Observational Research: Introducing the E-value

Similar documents
Recent advances in non-experimental comparison group designs

Assessing the impact of unmeasured confounding: confounding functions for causal inference

Making comparisons. Previous sessions looked at how to describe a single group of subjects However, we are often interested in comparing two groups

Objective: To describe a new approach to neighborhood effects studies based on residential mobility and demonstrate this approach in the context of

16:35 17:20 Alexander Luedtke (Fred Hutchinson Cancer Research Center)

Dylan Small Department of Statistics, Wharton School, University of Pennsylvania. Based on joint work with Paul Rosenbaum

Two-sample Categorical data: Measuring association

Introduction to Observational Studies. Jane Pinelis

Brief introduction to instrumental variables. IV Workshop, Bristol, Miguel A. Hernán Department of Epidemiology Harvard School of Public Health

Estimating Direct Effects of New HIV Prevention Methods. Focus: the MIRA Trial

Challenges of Observational and Retrospective Studies

Controlling Bias & Confounding

Complier Average Causal Effect (CACE)

How should the propensity score be estimated when some confounders are partially observed?

Evidence-Based Medicine Journal Club. A Primer in Statistics, Study Design, and Epidemiology. August, 2013

Propensity scores: what, why and why not?

INTERNAL VALIDITY, BIAS AND CONFOUNDING

Matched Cohort designs.

Measures of Association

George B. Ploubidis. The role of sensitivity analysis in the estimation of causal pathways from observational data. Improving health worldwide

Propensity Score Methods for Estimating Causality in the Absence of Random Assignment: Applications for Child Care Policy Research

Measuring association in contingency tables

Role of respondents education as a mediator and moderator in the association between childhood socio-economic status and later health and wellbeing

A Brief Introduction to Bayesian Statistics

Welcome to this third module in a three-part series focused on epidemiologic measures of association and impact.

Causal Mediation Analysis with the CAUSALMED Procedure

Measuring association in contingency tables

Fixed-Effect Versus Random-Effects Models

Supplementary Online Content

Simple Sensitivity Analyses for Matched Samples Thomas E. Love, Ph.D. ASA Course Atlanta Georgia

University of California, Berkeley

Midterm Exam ANSWERS Categorical Data Analysis, CHL5407H

Causal Association : Cause To Effect. Dr. Akhilesh Bhargava MD, DHA, PGDHRM Prof. Community Medicine & Director-SIHFW, Jaipur

PARTIAL IDENTIFICATION OF PROBABILITY DISTRIBUTIONS. Charles F. Manski. Springer-Verlag, 2003

A Primer on Causality in Data Science

BIOSTATISTICAL METHODS

The ACCE method: an approach for obtaining quantitative or qualitative estimates of residual confounding that includes unmeasured confounding

Propensity Score Analysis Shenyang Guo, Ph.D.

observational studies Descriptive studies

CAUSAL INFERENCE IN HIV/AIDS RESEARCH: GENERALIZABILITY AND APPLICATIONS

Beyond the intention-to treat effect: Per-protocol effects in randomized trials

Confounding and Bias

Confounding by indication developments in matching, and instrumental variable methods. Richard Grieve London School of Hygiene and Tropical Medicine

Functionalist theories of content

Challenges in design and analysis of large register-based epidemiological studies

Identifying Peer Influence Effects in Observational Social Network Data: An Evaluation of Propensity Score Methods

Propensity Score Methods for Causal Inference with the PSMATCH Procedure

MS&E 226: Small Data

Quantitative Methods. Lonnie Berger. Research Training Policy Practice

PubH 7405: REGRESSION ANALYSIS. Propensity Score

Checking the counterarguments confirms that publication bias contaminated studies relating social class and unethical behavior

Pros. University of Chicago and NORC at the University of Chicago, USA, and IZA, Germany

University of Wollongong. Research Online. Australian Health Services Research Institute

A Case Study: Two-sample categorical data

Author's response to reviews

Instrumental Variables Estimation: An Introduction

PharmaSUG Paper HA-04 Two Roads Diverged in a Narrow Dataset...When Coarsened Exact Matching is More Appropriate than Propensity Score Matching

SAMPLING AND SAMPLE SIZE

Ec331: Research in Applied Economics Spring term, Panel Data: brief outlines

Addendum to the 12th Report on Carcinogens

Imputation approaches for potential outcomes in causal inference

Overview of Perspectives on Causal Inference: Campbell and Rubin. Stephen G. West Arizona State University Freie Universität Berlin, Germany

Comparative Effectiveness Research Using Observational Data: Active Comparators to Emulate Target Trials with Inactive Comparators

Methods for Addressing Selection Bias in Observational Studies

Book review of Herbert I. Weisberg: Bias and Causation, Models and Judgment for Valid Comparisons Reviewed by Judea Pearl

The Logic of Causal Inference. Uwe Siebert, MD, MPH, MSc, ScD

Results. NeuRA Mindfulness and acceptance therapies August 2018

Mendelian randomization with a binary exposure variable: interpretation and presentation of causal estimates

Recent developments for combining evidence within evidence streams: bias-adjusted meta-analysis

Propensity score methods to adjust for confounding in assessing treatment effects: bias and precision

University of California, Berkeley

Supplement 2. Use of Directed Acyclic Graphs (DAGs)

Analysis methods for improved external validity

Rapid appraisal of the literature: Identifying study biases

Empirical Tools of Public Finance. 131 Undergraduate Public Economics Emmanuel Saez UC Berkeley

Lecture II: Difference in Difference. Causality is difficult to Show from cross

Meta-Analysis of Correlation Coefficients: A Monte Carlo Comparison of Fixed- and Random-Effects Methods

W e have previously described the disease impact

Methods to control for confounding - Introduction & Overview - Nicolle M Gatto 18 February 2015

Chapter 11. Experimental Design: One-Way Independent Samples Design

Traumatic brain injury

Missing data. Patrick Breheny. April 23. Introduction Missing response data Missing covariate data

Using Ensemble-Based Methods for Directly Estimating Causal Effects: An Investigation of Tree-Based G-Computation

THE INTERPRETATION OF EFFECT SIZE IN PUBLISHED ARTICLES. Rink Hoekstra University of Groningen, The Netherlands

- Decide on an estimator for the parameter. - Calculate distribution of estimator; usually involves unknown parameter

Results. NeuRA Family relationships May 2017

Results. NeuRA Forensic settings April 2016

Underlying Theory & Basic Issues

Generalizing the right question, which is?

An Introduction to Propensity Score Methods for Reducing the Effects of Confounding in Observational Studies

Conclusions of the BioInitiative Report. Michael Kundi Medical University of Vienna BioInitiative Organizing Committee

Estimands, Missing Data and Sensitivity Analysis: some overview remarks. Roderick Little

Stratified Tables. Example: Effect of seat belt use on accident fatality

NIH Public Access Author Manuscript Epidemiology. Author manuscript; available in PMC 2015 July 01.

Problem solving therapy

Two-Way Independent ANOVA

Distraction techniques

Regression Discontinuity Designs: An Approach to Causal Inference Using Observational Data

NeuRA Sleep disturbance April 2016

EC352 Econometric Methods: Week 07

Transcription:

Sensitivity Analysis in Observational Research: Introducing the E-value Tyler J. VanderWeele Harvard T.H. Chan School of Public Health Departments of Epidemiology and Biostatistics 1

Plan of Presentation (1) Introduction to Sensitivity Analysis (2) Cornfield Conditions (3) E-values for Sensitivity Analysis (4) Empirical Examples (5) Other Scales - Hazard Ratios, Odds Ratios, Continuous Outcomes (6) Other Sensitivity Analysis Results (7) Concluding Remarks 2

Unmeasured Confounding Unmeasured/uncontrolled confounding is a common problem in observational epidemiology Attempts are made to collect data on and control for as many covariates as possible that are related to the exposure and the outcome of interest Often, however, one or more important covariates will remain unmeasured U that might confound the relationship between the exposure and the outcome C A Y U Such unmeasured confounding variables can bias estimate of the effect of the exposure A on outcome Y 3

Sensitivity Analysis Bias analysis or Sensitivity analysis can help assess the extent to which an unmeasured variable (or variables) U would have to affect both exposure A and outcome Y in order for the observed associations between A and Y to be attributable solely to confounding rather than a causal effect of A on Y C A Y U Sensitivity analysis can also be useful in assessing a plausible range of values for the causal effect of A on Y corresponding to a plausible range of assumptions concerning the relationship between the unmeasured confounder U and the exposure A and outcome Y Example: Breastfeeding is associated with better infant and maternal outcomes might this be explained by the fact that breastfeeding mothers 4 have higher SES, rather than breastfeeding itself?

Sensitivity Analysis The term sensitivity analysis is used outside the context of unmeasured confounding bias to address other types of bias such as selection bias and measurement error In other fields sensitivity analysis is also referred to as bias analysis (e.g. Rothman et al., 2008; Lash et al., 2009) An early application of this approach was the work of Cornfield et al. (1959) who showed that the association between smoking and lungcancer association was unlikely to be entirely due to unmeasured confounding by an unknown genetic factor 5

Cornfield Conditions Let RR denote the estimated risk ratio relating exposure A and outcome Y, possibly conditional on covariates C Let U denote a binary unmeasured confounding variable Cornfield et al. (1959) considered how strong the associations between U and A (and between U and Y, Schlesselman, 1978) would have to be to completely explain away the observed RR relating A and Y Let U A denote the risk ratio relating A and U Let U Y denote the risk relating U and Y The Cornfield Conditions are that U could not be entirely responsible for the observed RR between A and Y unless: U A >RR and U Y >RR i.e. if RR was greater than either U A or U Y then U could not be entirely responsible

Cornfield Conditions Fisher thought the smoking-lung cancer relationship could be explained by a genetic variant U (Fisher, 1958, Nature) But RR = 10.73 (95% CI: 8.02, 14.36; Hammond and Horn, 1958) So we would need (Cornfield et al., 1959): RR UY >10.73 and RR AU >10.73 This does not seem plausible Almost certainly there is an effect Cornfield conditions are useful for assessing if an effect could be completely explained away, but are less useful for assessing plausible effects sizes; sensitivity analysis is more useful for this

Sensitivity Analysis for Total Effects The estimate adjusted for only measured covariates C is The estimate adjusted for C and U is Causal Effects: Let Y a be the potential outcome or counterfactual outcome for each individual if, possibly contrary to fact, A were set to a If the effect of A on Y is unconfounded conditional on C (i.e. Y a the first expression is P(Y 1 =1 c) / P(Y 0 =1 c) A C) then If the effect of A on Y is unconfoudned conditional on (U,C) but not on C alone then the second expression is = P(Y 1 =1 c) / P(Y 0 =1 c) but the first is not We let the bias factor B be the ratio between the estimate and the truth 8 We provide bounds for this bias based on sensitivity analysis parameters

Sensitivity Analysis for Total Effects Numerous sensitivity analysis techniques exist for risk ratios (relating B to sensitivity analysis parameters for U-Y and U-A associations) Many techniques also are available for differences in average outcomes However many of these techniques make numerous assumptions e.g. Binary Confounder: Cornfield et al. 1959; Bross 1966, 1967; Schlesselman 1978; Rosenbaum and Rubin 1983 No Interaction b/w Exposure and Confounder: Cornfield et al. 1959; Schlesselman 1978; Imbens, 2003 Only One Confounder: Almost all techniques Techniques are sometimes criticized because they make assumptions to assess assumptions Some techniques do not impose these assumptions but require specifying many parameters (Flanders and Khoury, 1990; VanderWeele and Arah, 2011) or parameters that make no reference to an unmeasured confounder (Robins et al, 2000; Diaz and 9 van der Laan, 2013)

Sensitivity Analysis w/o Assumptions Here we give an approach that in some sense makes no assumptions (Ding and VanderWeele, 2016) Assume we have one or more unmeasured confounders U We define the following parameters These are essentially the maximum effect of U on Y across exposure groups and the maximum risk ratio relating A and any category of U If U is binary these are just ordinary risk ratios, but apply more generally Note that the parameters condition on the measured covariate C They capture the confounding after having already controlled for C 10

Sensitivity Analysis w/o Assumptions Let RR UY be the maximum risk ratio relating any two categories of U to Y conditional on covariates C and across exposure A Let RR AU be the maximum risk ratio relating A to any category of U The largest factor by which such a U could reduce an observed risk ratio estimate is given by (Ding and VanderWeele (2016, Epidemiology): B = RR UY *RR AU (RR UY +RR AU - 1) We can divide the estimate and CI by B to obtain a corrected estimate and CI (i.e. the maximum a confounder could shift the estimate) The result holds without making any assumptions about the structure of U The bound is sharp in the sense that some U can always achieve them If the association is protective we multiply rather than divide 11

Smoking Lung Cancer Fisher thought the smoking-lung cancer relationship could be explained by a genetic variant U But RR = 10.73 (95% CI: 8.02, 14.36) Suppose we had: RR UY =2 RR AU =3 B = RR UY *RR AU = 2 * 3 = 1.5 (RR UY +RR AU - 1) 2+3-1 This would shift the estimate to 10.73/1.5 = 7.1 (95% CI: 5.3, 9.5) Still a large effect

Smoking Lung Cancer Fisher thought the smoking-lung cancer relationship could be explained by a genetic variant U But RR = 10.73 (95% CI: 8.02, 14.36; Hammond and Horn, 1958) Suppose we had: RR UY =10.73 RR AU =10.73 B = RR UY *RR AU = 10.7 * 10.7 = 5.6 (RR UY +RR AU - 1) 10.7+10.7-1 Not even this amount of confounding would explain the association We would still have an adjusted RR of RR = 1.91 (95% CI: 1.43, 2.56)

Sensitivity Analysis w/o Assumptions The Cornfield conditions are generally conservative The largest factor by which such a U could reduce an observed risk ratio estimate is given by (take inverses for protective exposures): B = RR UY *RR AU (RR UY +RR AU - 1) If we let RR UY go to infinity we get If we let RR UA go to infinity we get B = RR AU B = RR UY i.e. we recover the Cornfield Conditions (but without assuming a binary confounder); the Cornfield conditions essentially assume the other parameter is infinite; it is conservative in the other parameter 14

Table of Bias Factors We can consider many different parameters and calculate the bias factor (i.e. maximum bias that can be generated) 15

Approaches to Sensitivity Analysis Sensitivity analysis does not give one right answer but a range It is sometimes objected that there is too much subjectivity in sensitivity analysis Possible Approaches: (1) Report how large the effects of the confounder would have to be to completely explain away (i) the effect and (ii) the CI (2) Create a table with many values of all sensitivity analysis parameters; include those one thinks are too extreme; let the reader decide (3) Find the most important measured confounder variable; examine if an unmeasured confounder of similar strength would change conclusions (4) Use external data or expert opinion to inform sensitivity analysis parameters We will focus on the first approach of explaining away 16

E-Values The E-value (VanderWeele and Ding, 2017): the minimum strength of association on the risk ratio scale that an unmeasured confounder would need to have with both the exposure and the outcome, conditional on the measured covariates, to fully explain away a specific exposure-outcome association. The E-value is thus a measure pertaining to evidence for causality With an observed risk ratio of RR, we have that if RR UY and RR AU are greater than (VanderWeele, and Ding, 2017): E-value = RR + sqrt[ RR*(RR-1) ] then this could suffice, but weaker confounding could not 17

E-Values We can also ask how much confounding would explain away an estimate and its confidence interval; we apply the formula: E-value = RR + sqrt[ RR*(RR-1) ] We can apply this in a routine manner to both the estimate and the confidence interval limit closest to the null If RR<1, take inverses first and apply the formula We might call this the E-value (evidence for causality) RR = 10.73 (95% CI: 8.02, 14.36) E-Value for Estimate: 20.9; E-value for CI: 15.5 With an observed risk ratio of RR=10.7, an unmeasured confounder that was associated with both the outcome and the exposure by a risk ratio of 20.9-fold each, above and beyond the 18 measured confounders, could explain away the estimate, but weaker confounding could not

Sensitivity Analysis w/o Assumptions Smoking and Lung Cancer (Hammond and Horn, 1958): RR = 10.73 (95% CI: 8.02, 14.36) 19

Breast-feeding Example Consider adjusted associations between breastfeeding and health We often worried these are confounded by SES Examples (all highly statistically significant ) (AHRQ, 2007; Moorman, 2008) Lower Respiratory Tract Infection: RR = 0.28 (95% CI: 0.14, 0.54) To calculate E-value for estimate for LRTI: 1/0.28 = 3.57 E-value = RR + sqrt[ RR*(RR-1) ] = 3.57+sqrt{3.57 (3.57-1)} = 6.6 To calculate E-value for confidence interval 1/0.54 = 1.85 E-value = RR + sqrt[ RR*(RR-1) ] = 1.85+sqrt{1.85 (1.85-1)}=3.1 Substantial confounding, above and beyond the measured covariates, would be required 20

Breast-feeding Example Examples (all highly statistically significant ) (AHRQ, 2007; Moorman, 2008) Lower Respiratory Tract Infection: RR = 0.28 (95% CI: 0.14, 0.54) Maternal Ovarian Cancer: RR = 0.5 (95% CI: 0.3, 0.8) Childhood Leukemia: RR = 0.81 (95% CI: 0.71, 0.91) E-values: Estimate Upper CI Lower Respiratory Tract Infection: 6.6 3.1 Maternal Ovarian Cancer: 3.4 1.8 Childhood Leukemia: 1.7 1.4 Strong evidence for LRTI; moderate for ovarian cancer; modest for leukemia Just because an unmeasured confounder can explain an effect away does not mean that it does (effect might still be present, but evidence isn t definitive) 21

E-Values E-value Formula: E-value = RR + sqrt[ RR*(RR-1) ] The E-value could always be reported for the estimate and CI This would allow for better assessment of causality P-value gives evidence for association E-value gives evidence that the association is causal The two can diverge: E-values: Estimate Upper CI Maternal Ovarian Cancer: 3.4 1.8 Childhood Leukemia: 1.7 1.4 P-value is more extreme for leukemia (p<.001) than ovarian cancer (P=.006) 22 E-value is more extreme for ovarian cancer than leukemia

E-Values Properties of the E-Value: ØP-value can always be made more extreme by larger sample size ØE-value cannot ØThe E-value for the CI will generally get larger as sample size increases, but converge to the E-value for the observed association ØIn a different context Rosenbaum (2009) refers to this limit as design sensitivity ; but what is most relevant is arguably the data available We can interpret the E-Value as e.g. With an observed risk ratio of RR=0.28, an unmeasured confounder that was associated with both the outcome and the exposure by a risk ratio of 6.6-fold each, above and beyond the measured confounders, could explain away the estimate, but weaker confounding could not; to move the confidence interval to include the null, an unmeasured confounder that was associated with the outcome and the exposure by a risk ratio of 3.1-fold each could do so, but weaker confounding could not. 23 We could include statements like this in all empirical papers

E-Values Further Important Points for Interpretation: ØWhat is a large E-value depends on the measure covariates; the parameters are conditional on the measured covariates; if control is made for many measured covariates, it is harder to explain away a given estimate ØWhat is a large E-value is relative to the outcome and exposure ØE-value is conservative; it makes no assumptions about the unmeasured confounder; if the unmeasured confounder is very rare it will not generate the amount of bias suggested by the E-value or bounding factor ØIf, after control for many covariates, we still have evidence for an effect after calculating the E-value, this can be very persuasive ØBut we cannot really use the E-value in the same way to provide evidence for no effect because the bounding factor and E-value considers the maximum bias a confounder could induce ØAbsence of evidence is not evidence of absence ØOne can also calculate non-null E-values to see how much confounding 24 is needed to move an estimate to any other value

Non-Null E-Values If we want to see how much confounding would shift the estimate to a value other than 1, e.g. to RR true AY, we can do this as well We just take the ratio of the ratio RR = RR true ED / RR and then again apply the E-value formula: E-value = RR + sqrt[ RR *(RR -1) ] As before, if the ratio RR true ED / RR less is than 1, then take inverses before applying the formula 25

Other Scales The same approach also works with: ² Odds Ratios with a rare outcome ² Hazard Ratios with a rare outcome at end of follow-up ² Ratios of count outcomes ² Continuous positive outcomes (using a ratio of expectations) We just replace the RR s above with HR s, OR s, etc. For odds ratios with a common outcome, the OR is often much larger than the RR and these will not be comparable One can get an approximate E-value by using RR sqrt(or) and then applying the E-value formulas (VanderWeele, 2017) Sensitivity analyses that present the sensitivity analysis parameters on an OR scale when using a common outcome are likely to exaggerate the robustness 26 Other formulas for HR s with a common outcome

Difference Scales For a difference in continuous outcomes, an approximate E-value can be obtained as follows: Consider a standardized effect sizes d (the outcome variable should be divided by its standard deviation) and a standard error for this effect size s d Use approximation RR»exp(0.91 d) Use approximate CI: (exp{0.91 d 1.78 s d }, exp{0.91 d + 1.78 s d }). Then apply the E-value formula: RR + sqrt[ RR*(RR-1) ] This approach relies on approximations from meta-analysis (Borenstein et al., 2009; Hasselblad and Hedges, 1995) to convert standardized effect sizes into OR s, and then approximations between OR and RR Other sensitivity analysis approaches can be used for continuous outcomes 27 but also make additional assumptions

Advantages of E-Values Advantages of E-values: (1) Easy to compute Ø By hand, or with an online calculator, or R package: https://evalue.hmdc.harvard.edu/app/ (2) Easy to report and interpret (3) Standardized metric across different scales (4) Can be calculated without assumptions (5) Removal of subjectivity in sensitivity analysis (6) Moving away from OR parameters avoids exaggeration When to avoid? Possibly when the unmeasured confounder is known and rare since the E-value approach will be too conservative 28

Prevalence Specification If we are willing to specify the prevalence (and make further assumptions) we can sometimes obtain smaller bias, and also get the exact bias Result (Schlesselman,1978) : If there is no unmeasured confounding given (C,U) i.e. Y a A (C,U) and U is a binary unmeasured confounder with the same risk ratio, γ, on Y for both exposed and unexposed so that Then we have that: We can use the bias formula by specifying: (i) γ = the effect of U and (ii) the prevalence of U amongst the exposed and unexposed 29

Prevalence Specification Once we have calculated the bias term B mult (c), we can simply estimate our risk ratio controlling only for C (e.g. if the outcome is rare we just fit a logistic regression) and we divide our estimate by B mult (c) to get the corrected estimate for the causal risk ratio i.e. what we would have obtained if we had adjusted for U also This is now exact, not conservative, but we must make assumptions, and also specify prevalences We can obtain corrected confidence intervals by dividing both limits of the confidence interval by B mult (c) The result above is given by Schlesselman (1978) building on the work of Bross (1966) 30

Prevalence Specification The estimates for breastfeeding for 6-12 months on ovarian cancer was OR of 0.5 (95% CI: 0.3, 0.8) Suppose we thought that low SES increased the risk of ovarian cancer by 2 fold and that 30% of the 6-12 month breastfeeding group was low SES but 60% of the reference group (no breast-feeding) was low SES The bias factor is then: [1 + (2-1)*(0.3)] / [1 + (2-1)*(0.6)] = 0.81 The corrected estimate would be 0.61 (95% CI: 0.37, 0.99) Note the E-value for the CI here was 1.8; but with the prevalences specified and additional assumptions the risk ratio of 2-fold each do not explain away the CI 31

Prevalence Specification Suppose we thought that low SES increased the risk of ovarian cancer by 2 fold and that 10% of the 6-12 month breastfeeding group was low SES and 20% of the reference group (no breast-feeding) was low SES The bias factor is then: [1 + (2-1)*(0.1)] / [1 + (2-1)*(0.2)] = 0.91 The corrected estimate would be 0.55 (95% CI: 0.33, 0.88) With lower prevalence the confounding generates less bias However, in specifying the prevalences there can be temptation to specify low prevalence to make the estimates look good The formula also makes homogeneity assumptions The E-Value gets around both of these issues 32 But if we know the prevalence is indeed low, specifying it may be preferable

Difference Scale and Continuous The bias for the causal effect conditional on C on the difference scale: B add (c) = Outcomes Result. Suppose that we have a binary unmeasured confounder U with the effect of U on Y on the difference scale the same for the exposed and unexposed with γ= then the conditional bias is: B add (c) = γδ where δ = E(U A=1,c)-E(U A=0,c) is the difference in expectations of U between the exposed and unexposed conditional on C=c (prevalence difference if U is binary) Note these parameters are conditional on / control for measured C 33

Difference Scale and Continuous Outcomes Once we have calculated the bias term B add (c) we can simply estimate our causal effect conditional on C and then subtract the bias term to get the corrected estimate To get corrected confidence intervals on can simply subtract γδ from both limits of the estimated confidence intervals for either conditional effects or for average causal effects if B add (c) is constant over c The result was given in a regression context by Cochran (1935) and re-derived numerous times subsequently, though with increasingly weaker assumptions (Lin et al., 1998; VanderWeele and Arah, 2011) It can be applied regardless of the estimation method and requires only that the effect of U on Y is the same in both exposure strata (no U*A interaction on the difference scale) But this approach still does make this homogeneity assumption 34

Covariate Conditioning Ø All of the results (and all sensitivity analysis parameters) are conditional on measured covariates Ø Unmeasured confounding, to alter the estimate, must affect the exposure and outcome through pathways independent of the measured covariates Ø Often, with extensive confounding control, an additional covariate (which may make a big difference when adjusting compared to the crude estimate) doesn t matter much under control for other covariate Ø E.g. Income may not matter as much if we have also controlled for education, wealth, home ownership, and occupation Ø It can be instructive examining this with one s data Ø Sensitivity methods (e.g. Altonji et al., 2005) that assume unobserved confounding is comparable in magnitude to observed confounding are problematic b/c it looks like confounding gets worse with additional control Ø Somewhat more sensible is to compare what proportion of the effect of measured confounding the unobserved confounding would have to have to explain away the effect (Altonji et al., 2005) Ø Interpretation here as well is also with respect to the measured covariates 35

Conclusions 1) Sensitivity analysis is a powerful tool to assess robustness; and techniques are now available which effectively make no assumptions 2) We can use these to calculate the minimum amount of confounding that would suffice to explain away the effect (i.e. the E-value) 3) If the E-value suggests that the evidence is robust, this can be a very strong argument 4) It is more difficult to provide evidence that an effect is absent 5) Too many of our interpretative practices for observational studies are inherited from RCTs 6) The E-value approach corrects this; it is easy to use and is applicable with most effect measures 7) It could be used very broadly; if this were done our inferences about causation and science would likely be considerably improved; whether this is done depends in part on you it is easy to employ 36

Additional Slides Tyler J. VanderWeele Harvard T.H. Chan School of Public Health Departments of Epidemiology and Biostatistics 37

Appendix: Parameter Interpretation Let RR UY be the maximum risk ratio relating any two categories of U to Y conditional on covariates C and across exposure A Let RR AU be the maximum risk ratio relating A to any category of U The largest factor which such a U could reduce a RR is given by: B = RR UY *RR AU (RR UY +RR AU - 1) This holds true for any U such that (C,U) suffice to control for confounding We could take the minimum over all U such that (C,U) suffice for confounding If one only requires approximate unconfoundedness (e.g. RR true /RR obs < 1.03) then RR UY and RR AU could be considered over all U and bias factor B could be taken as the minimum over all U satisfying approximate unconfoundedness The E-value could then be interpreted the minimum strength of association on the risk ratio scale that any and every unmeasured confounder, or coarsening thereof, that suffices with the observed covariates to bring the observed association within 3% of the true causal effect, would have to have with both the exposure and the outcome, above and beyond the measured covariates, to explain away the 38 observed exposure-outcome association

Transformation of the Effect Size The E-value is just a transformation of the effect size: E-value = RR + sqrt[ RR*(RR-1) ] We recommend reporting the E-value for the estimate and the limit of the CI closest to the null There is nothing in this information that is not also in the estimate and CI itself But the transformation provides a different scale for which it is easier to assess robustness to unmeasured confounding In many contexts, there is likewise nothing in the p-value that is not also in the estimate and CI, but we cannot generally do the transformation in our head The scale of the p-value is helpful for weighing evidence; likewise the E-value Consider two estimates: RR=1.18 (95% CI: 1.04, 1.33) and the other as RR=1.18 (95% CI: 1.12, 1.24) We might ordinarily interpret these quite similarly But the E-value for the CI are 1.24 and 1.49 respectively We quite frequently find RR of 1.25, but much less frequently 1.5 The strength of the evidence here is arguably considerably different The E-value allows us to see this more easily 39

Relation to Rosenbaum s Design Sensitivity The E-value bears some relation to Rosenbaum s design sensitivity (Rosenbaum, 2004) but (1) Design sensitivity focuses on much how the odds of treatment of two units with identical covariates may differ, rather than relations with the outcome; the sensitivity analysis amplification (Rosenbaum and Silber, 2009) considers the relation with the outcome but under parametric models rather than non-parametrically (2) Design sensitivity provides sensitivity to unmeasured confounding for infinite samples, rather than the actual sample (3) Design sensitivity uses odds ratio scales, which can vastly exaggerate risk ratios and hence the intuitive understanding of robustness of results 40

Inference With and Using the E-value One could calculate the confidence interval for the E-value for the confounded association but this would give statements like: Across repeated samples, at least 95% of the time, the minimum strength of association on the risk ratio scale that an unmeasured confounder would have to have with both the exposure and the outcome, conditional on the measured covariates, to explain away the actual confounded exposure-outcome association will lie in the confidence interval provided. If instead we calculate the E-value for the lower limit of the confidence interval we can make statements of the form: Across repeated samples, at least 95% of the time, if the actual confounding parameters RR UY and RR EU are both less than the E-value for the confidence interval that was calculated, then there will be a true effect in the same direction as the observed association The object of inference is arguably the causal effect itself, not the E-value; the E-value is just a tool; statements of the latter type are arguably generally 41 of greater interest

Sensitivity Analysis w/o Assumptions The sensitivity analysis is without assumptions in the sense that without making assumptions we can make statements of the form: For the unmeasured confounding to shift the observed estimate to 1 the unmeasured confounder(s) would have to have parameters that satisfy: RR UY *RR AU (RR UY +RR AU - 1) > RR Or statements employing the E-value: To explain away the observed risk ratio of RR=0.28, an unmeasured confounder that was associated with both the outcome and the treatment by a risk ratio of 6.6-fold each could suffice, but weaker confounding could not where the strength of confounding is measured by the bounding factor 42

Sensitivity Analysis w/o Assumptions Smoking and Lung Cancer (Hammond and Horn, 1958): RR = 10.73 (95% CI: 8.02, 14.36) 43