DRAFT REPORT HIERARCHY OF METHODS TO CHARACTERIZE UNCERTAINTY: STATE OF SCIENCE OF METHODS FOR DESCRIBING AND QUANTIFYING UNCERTAINTY

Size: px
Start display at page:

Download "DRAFT REPORT HIERARCHY OF METHODS TO CHARACTERIZE UNCERTAINTY: STATE OF SCIENCE OF METHODS FOR DESCRIBING AND QUANTIFYING UNCERTAINTY"

Transcription

1 DRAFT REPORT HIERARCHY OF METHODS TO CHARACTERIZE UNCERTAINTY: STATE OF SCIENCE OF METHODS FOR DESCRIBING AND QUANTIFYING UNCERTAINTY EPA Contract No. 68-D Work Assignment No Prepared by: Chris Frey Doug Crawford-Brown Zheng Junyu Dan Loughlin For: Jim Wilson, Project Manager E.H. Pechan & Associates, Inc B Hempstead Way Springfield, VA For submittal to: Lisa Conner, Work Assignment Manager U.S. Environmental Protection Agency Office of Air Quality Planning and Standards Innovative Strategies and Economics Group (MD-C339-01) Research Triangle Park, NC /12/2003

2 Table of Contents 1 Introduction Objective of this Document Motivations National Academy of Science Report State of Practice Risk, Cost, Uncertainty and Decisions Sources of Variability and Uncertainty Sources of Variability Sources of Uncertainty General Classification of Methods Statistical Methods Based Upon Empirical Data Statistical Methods Based upon Judgment Other Quantitative Methods Qualitative Methods Sensitivity Analysis Statistical Methods Based Upon Empirical Data Introduction Characterizing Distributions Based Upon Data Analysis Empirical Distributions Parametric Distributions Mixture Distributions Censored Data Other Issues Quantification of Uncertainty Based Upon Variability Propagating Uncertainty Analytical propagation techniques Approximation methods Numerical propagation techniques Statistical Methods Based Upon Judgment Heuristics and Biases to Avoid Protocols for Expert Elicitation Expert Groups vs. Individuals Resource Requirements for Expert Elicitation Bayesian Methods Introduction Bayesian Framework for Uncertainty Analysis i

3 3.5.3 Prior Distribution Likelihood Functions Bayesian Updating: Posterior Distribution Propagation of Uncertainty through Model Research Needs for Bayesian Methods Other Quantitative Methods, including Approximation or Interval Methods Interval Methods Other Quantitative Methods Fuzzy Methods Meta-Analysis Methods for Scenario and Model Uncertainty Scenario Uncertainty Model Uncertainty Qualitative Methods Qualitative Assessment and Rationality Principles of Rationality Lines of Reasoning Direct empirical evidence Semi-empirical evidence Empirical correlations Theory-based inference Existential insight Combining the five categories of evidence Structuring the Discourse Interpreting the Results Weight-of-Evidence Sensitivity Analysis Local Sensitivity Analysis Methodologies Global Sensitivity Analysis Methodologies Summary and Conclusions Key Criteria for Classification of Input Uncertainty Characterization Methods Key Criteria for Classification of Uncertainty Propagation Methods Identification of Key Criteria for Classification of Sensitivity Analysis Methods References ii

4 1 Introduction 1.1 Objective of this Document The analysis of uncertainty is playing an increasing role in regulatory decisions, driven by several factors: (i) the tools of uncertainty analysis have been significantly improved over the past decade; (ii) the rise of decision processes involving stakeholders (e.g. regulatory negotiation) has brought competing estimates of risk to the forefront; (iii) the success of the EPA and other environmental managers in removing some of the more obvious threats to health has caused the regulatory arena to focus on sources of risk that are not as obvious and are less well studied; and (iv) legal challenges to regulatory decisions have necessitated a focus on the quality of the risk estimates, which must withstand these challenges in court. While the importance of uncertainty analysis in decisions has increased, so has the awareness that such analyses often require significant resources of time, money, and expertise. Resources spent on uncertainty analysis may draw resources away from analyses of multiple policy options, since it may be possible to conduct full uncertainty analyses on only a few policy options in considering a regulatory decision. It also is the case that some decisions require a full uncertainty analysis, while others may be based on less detailed analysis of uncertainties, or the use of decision tools that do not involve quantitative uncertainty analysis. Finally, an issue can be raised as to whether some aspects of uncertainty (model quality comes to mind) are susceptible to fully quantitative analysis, or may be better approached through a more qualitative process. All of these issues suggest that uncertainty analyses may need to be tailored to the specific decision problem, with the level of detail required of the analysis being related to the kind of decision problem posed. One way to view this situation is to imagine a hierarchy of tools for uncertainty analysis. This hierarchy proceeds from methods that are relatively quick and require few resources, but also provide the least detailed understanding of uncertainty, to methods that require significantly greater resources but produce highly detailed quantitative analyses of uncertainty that can be used in formal tools of decision analysis. 1.2 Motivations There are three motivations for this report. The first is that the EPA, in essentially all of its 1

5 offices, is moving rapidly in the use of uncertainty analyses. This suggests it is timely to consider some common understanding of the available tools, in part to ensure uniformity of application in regulatory and other decisions. An example is the publication of Agency-wide guidelines on the use of Monte Carlo analysis (EPA, Proposed Policy for Use of Monte Carlo Analysis in Agency Risk Assessments, Washington, DC., 1997). The second is that the field of uncertainty analysis has progressed rapidly in the past decade. Many new tools have been developed, and computational power has increased to the point where highly detailed, probabilistic assessments of uncertainty, unwieldy in the past, can now be performed routinely in regulatory assessments. The third is a recent review of a specific use of uncertainty analysis by the Agency in estimating costs and benefits of reductions in air pollution. This review by the National Academy of Sciences Committee on Estimating the Health-Risk-Reduction Benefits of Proposed Air Pollution Regulations (Estimating Public Health Benefits of Air Pollution Regulations, National Academy Press, 2003) recommended several ways in which the Agency might improve the use of uncertainty analysis. These three motivations are discussed in subsections through below, in reverse order from the listing above National Academy of Science Report The National Academy of Sciences was asked by Congress to review recent EPA estimates indicating that thousands of premature deaths and numerous cases of illness, such as chronic bronchitis and asthma attacks, could be prevented by reducing exposure to air pollution (NAS, Estimating Public Health Benefits of Air Pollution Regulations, National Academy Press, 2003). The report further comments that the estimates are often controversial, and the methods used to prepare them have been questioned. The NAS was asked to conduct a study of this issue and recommend to the agency a common methodology to be followed in all future analyses. Notice that the charge itself suggests a belief that there is a single methodology, rather than a suite of methodologies, any one of which might be appropriate in some decision problems but not in others. The committee reviewed a large number of aspects of the agency assessment, much of it focused on the risk and cost estimation processes themselves. Those parts of the NAS recommendations are not considered here. Instead, we focus on those parts of the NAS recommendations related to uncertainty analysis. There is, however, one aspect of the NAS report related to the conduct of an 2

6 assessment, but with clear implications for uncertainty analysis. The committee notes that In three of the four EPA analyses reviewed by the committee, EPA focused on evaluating a single regulatory option. They go on to state that a realistic range of options guided by expert opinion and technical feasibility should be represented in EPA s benefit analyses. Two points are evident here. First, there is a call for defining decision problems as ones requiring choice between several policy options. This means risk, cost, and benefit assessments would need to be performed for each of these options. This in turn raises issues related to the amount of resources that would be needed to conduct full uncertainty analyses for each of these assessments. As a larger number of policy options is considered, it may be necessary to balance the resource requirements of these assessments against the resources required to conduct fully quantitative analyses of uncertainty. Second, note the use of the term expert opinion in the NAS text. This stems from the sense on the committee that many aspects of scientific study of policy options, both in the risk and cost calculations, will not be determined entirely by the available data. These data will need to be abstracted, interpreted through theories, weighed one against the other when they conflict, etc. In these cases, expert judgment will be needed. While the committee does not specify whether this need will arise in the primary assessments of risk and cost, or in the secondary analysis of uncertainty, it is likely that it will arise in both. The committee goes on to consider two ways in which the EPA might improve uncertainty analyses. The first issue focuses on the conduct of the analysis, and the second on the use of uncertainty analyses in decisions. With respect to the conduct of the analysis, they describe the EPA approach as being two-tiered. The first tier produces a probability distribution for each health outcome evaluated. They express concern that the agency has considered only random sampling error in generating this distribution (and that being restricted to the concentrationresponse function, rather than all steps of the assessment), and that the focus was on uncertainty in mean values of risk, rather than on the uncertainty in other characteristics of the inter-subject variability distribution. This leads, they claim, to a view of the results as being more certain than they are, presumably because the uncertainties in the risks associated with the tails of intersubject variability distributions would be larger than uncertainties in means. 3

7 The second tier consists of ancillary uncertainty analyses, which include alternative and supplementary calculations for some uncertainties and sensitivity analyses for others. Ancillary analyses might, for example, consist of re-generating the uncertainty analyses from the first tier under alternative models of atmospheric dispersion, or alternative assumptions about the presence of a dose-response threshold. The critique by the committee is that the ancillary analyses usually examine one source of uncertainty at a time and therefore do not adequately convey the aggregate uncertainty from other sources, nor do they discern the relative degrees of uncertainty in the various components of the health benefit analysis. The committee is calling for a more comprehensive method of uncertainty analysis in which all sources of uncertainty are considered simultaneously (including source characterization; dispersion and fate; exposure as this relates to human activity patterns; and exposure-response relationships). This same methodology would form the groundwork for a detailed sensitivity analysis, which then could be used to direct research resources most effectively towards those aspects of the assessments contributing most heavily to uncertainty in the final risk, cost and benefits picture and, hence, to any decision. The recommended methodology is not fully described, but is mentioned as being a probabilistic, multiple-source uncertainty analysis, which seems to suggest some form of nested variability and uncertainty analysis accounting simultaneously for all significant sources of uncertainty. The committee goes on to make several other recommendations that are germane here. They note that the uncertainty distributions, however they are generated, will need to be based on available data and expert judgment. This calls for a methodology capable of incorporating both kinds of information. They also note that experts whose judgments are used should be identified, and the rationales and empirical bases for their judgments described. This clearly calls not only for encoding of expert judgments, but some way of describing how those judgments were formed, and the degree of rationality behind the judgments. Finally, the committee notes that there may always be some sources of uncertainty that are not reflected in the primary analysis (i.e. the analysis producing the quantitative probability density function describing uncertainty), and that the agency should clearly identify these. No guidance is provided on how to use this identification, but it is reasonable to expect that the significance of this non-quantified source of uncertainty on decisions should be described. 4

8 The NAS report (Estimating Public Health Benefits of Air Pollution Regulations, National Academy Press, 2003), therefore, presents several challenges: To formalize the analysis of uncertainty as much as possible; To make the results of the analysis as detailed as feasible given resources and the need to simulate a number of policy options; To make the results as quantitative as possible, summarized in a probability density function (PDF); To reflect as many sources of uncertainty as possible in this PDF, including components from source characterization to risk; To perform uncertainty analysis not only on mean values of risk, but on other aspects of the inter-subject variability distribution; To use not only expectation values of risk in decisions, but other characteristics of the uncertainty PDF; To find a way to characterize expert judgment, and to fold this into the analysis of uncertainty as appropriate; and To find a way to reflect sources of uncertainty which have not been included in developing the PDF, either because they weren t considered or (more interestingly) because they are not necessarily amenable to mathematical treatment State of Practice We turn next to the issue of how the conduct of uncertainty analyses has improved recently, and the ways in which these improvements have been incorporated into assessment and decisionmaking processes in a variety of settings. A goal of this report is to present a picture of the state of development of uncertainty analysis in decision-making, so a decision-maker can determine what is feasible at present and what will meet current standards of practice in the risk assessment and decision community. 5

9 The use of probabilistic analysis methods for dealing with variability and uncertainty is becoming more widely recognized and recommended for environmental modeling and assessment applications. The National Research Council (NRC) and others have recommend that EPA use quantitative probabilistic analysis methods that distinguish between variability and uncertainty (NRC, 1994). One of the recommendations of the Emission Inventory Improvement Program (EIIP), which is jointly sponsored by EPA and other organizations, is to encourage the use of quantitative methods to characterize variability and uncertainties in emission inventories (Radian, 1996). In 2002, NRC (2002) recommended that EPA should begin to move the assessment of uncertainties from its ancillary analyses to its primary analyses in estimating public health benefits of air pollution regulations. EPA has been responsive to these recommendations. For example, EPA has sponsored workshops regarding Monte Carlo simulation methods, has developed a guidance document on Monte Carlo methods, and has included guidance regarding probabilistic analysis in Risk Assessment Guidance for Superfund (EPA, 1996; EPA, 1997; EPA, 1999a; EPA, 199b; EPA, 2001a); has developed computational methods for sensitivity and uncertainty for environmental and biological models (EPA, 2001b) and methods for characterizing uncertainty in exposure data (EPA, 2000). For example, EPA has applied two-stage Monte Carlo simulation to characterize both variability and uncertainty in human cumulative exposure and dose to pollutants, such as pesticides and particulate matter (PM), in the development of SHEDS (Stochastic Human Exposure Dose Simulation) series models (Zartarian, 2000). Uncertainty analysis is now part of the planning process for major assessments performed by EPA, such as the National Air Toxics Assessment and exposure assessment. Recently, the National Research Council released a report on mobile source emissions estimation that calls for new efforts to quantify uncertainty in such emissions (NRC, 2000). The Intergovernmental Panel on Climate Change (IPCC) recently issues a good practice document regarding uncertainty analysis for greenhouse gas emission inventories (IPCC, 2000). Thus, the quantification of variability and uncertainty has become widely accepted not only in human health risk assessment but also in supporting or related areas, such as emissions estimation. In addition, there is a growing track record of the demonstrated use of quantitative methods for characterizing variability and uncertainty applied to emission factors, emission inventories, air 6

10 quality modeling, exposure assessment, risk assessment, policy analysis, decision-making under uncertainty, and economic assessment. Some examples of these are briefly mentioned here. There have been a number of projects aimed at quantifying variability and uncertainty in highway vehicle emissions, including uncertainty estimates associated with the Mobile5a emission factor model and with the EMFAC emission factor model used in California (Kini and Frey, 1997; Frey, 1997; Frey, Bharvirkar and Zheng, 1999; Pollack et al.,.1999; Frey and Zheng, 2002). Frey and Eichenberger (1997) and Frey et al. (2001) have quantified uncertainty in highway vehicle emission factors estimated based upon measured data collected using remote sensing and on-board instrumentation, respectively. Frey et al. (2002) have recommended modeling methods for the New Generation Model (NGM) that will succeed the Mobile6 emission factor model. These methods include quantification of unexplained inter-vehicle variability and fleet average uncertainty. There have been a number of efforts aimed at probabilistic analysis of various other emission sources, including power plants, non-road mobile sources, natural gas-fired engines, and specific area sources (Frey and Rhodes, 1996; Frey, Bharvirkar and Zheng, 1999; Frey and Zheng, 2000; Frey and Bammi, 2002a&b; Frey, and Zheng, 2002; Frey and Bharvirkar, 2002; Li and Frey, 2002, Abdel-Aziz and Frey, 2002). Probabilistic anlayses have also been applied to air quality models; Hanna et al. (1998) used Monte Carlo simulation to quantify uncertainty in ozone predictions for New York City by using the Urban Air Shed Model (UAM-IV). Hanna et al. (2001) employed the same approach to the ozone transport assessment group (OTAG) domain. In the area of exposure and risk assessment, there have been a number of analyses in which uncertainty analysis was performed. These include, for example, Bogen and Spear (1987; 1990), Frey (1992), Hoffman and Hammonds (1996), Cohen et al. (1996), Evans et al. (1994), Greenland et al. (1999), Hattis and Burmaster (1994), Krewski et al.(1999), Whitfield et al.(1991), and others. For policy and decision-making, there have been a number of analyses where quantitative uncertainty analysis was used. These include, for example, Morgan et al. (1984), Morgan and Henrion (1990), Harrison (2002), Chao and Hobbs, 1997; Venkatesh and Hobbs, 1999, and 7

11 others. Quantitative uncertainty analysis has also been used in economic analysis field. These include, for example, Clarke and Low (1993), Belli et al s Economic Analysis of Investment Operations (World Bank Institute 2001), Holland et al. (1999), and others Risk, Cost, Uncertainty and Decisions This document is concerned not only with the development of tools for uncertainty analysis, but also with determining which tools are appropriate for which applications in decision-making. There are several distinct ways in which uncertainty enters a decision problem: Sound Science. As the economic impacts (costs) of environmental decisions have increased, there has been greater scrutiny of the scientific basis for risk estimates. The concern has been that some risk estimates may be so uncertain as to be an unreliable guide for decisions as to whether a risk exists in the first place, and the magnitude of that risk if it does exist. This has led to a call for the use of sound science as a basis for risk estimates. While the term is contentious, and can often be related to attempts to simply impede regulatory action, it does have a footing in the philosophical concept of minimal epistemic status (Crawford-Brown, Chapter 1, Risk-Based Environmental Decisions: Methods and Culture, Kluwer Academic Publishers, 1999). Epistemic status refers to the rational basis for belief in a statement, such as a statement about risk (or cost). As the evidence underlying a belief builds up, the degree of support for that belief increases, and it becomes more rational to adopt that belief in reaching decisions. This rationality might be because the estimate of risk is more true, in the sense that it corresponds better to the real risk (e.g. the real probability of effect), because it works better in some sense (leading to decisions that produce the results we desired), or some combination of these two criteria. Regardless, increasing evidence, and better use of that evidence in reasoning, increases the epistemic status of a belief. At some point as the evidence increases, the belief becomes a rational basis for decisions. It should be clear that perfect evidence and reasoning produces a rational basis for belief, and that the existence of zero evidence makes the belief an irrational basis for decisions. As the strength 8

12 of the evidence and reasoning increases, the epistemic status rises from zero and moves towards perfection. At some threshold level, called the level of minimal epistemic status, the belief is justified as a basis for decisions. Where that level lies is open to debate, as is the issue of how the epistemic status itself is to be characterized, but the concept of epistemic status and its role in decisions is clear. This issue is discussed in greater detail in Sections and 4.3, and in Chapter 5. A key characteristic of this aspect of uncertainty is that it is not necessarily quantitative. Rational analysis of epistemic status focuses on the evidence and reasoning process underlying a belief, asking whether that combination is strong, how it compares to other combinations of evidence and reasoning, etc. A judgment that the results of a risk assessment represent sound science, therefore, often involves a judgment of the scientific process, and often rests on some combination of objective and subjective statements. A full assessment of the epistemic status of a statement about risk (or cost) cannot avoid the quantitative aspects of uncertainty analysis, but it also cannot be fully reduced to those quantitative aspects. A full assessment would produce probability density functions (PDFs) describing confidence in risk estimates, but it would set these PDFs inside a larger, and often qualitative, assessment of the quality of the science that went into producing these PDFs (which are necessarily conditional on the state of existing science). Similar thoughts seem to be expressed by the NAS in their recommendations mentioned earlier, where they call for both a quantitative assessment of uncertainty and a discussion of other causes of uncertainty that are not included quantitatively. Finally, following the NAS, we note that significant uncertainties do not necessarily paralyze decisions. A quantitative PDF, however broad it might be, can still be used in decision theoretic approaches to produce rational decisions. And even a judgment of low epistemic status might allow a decision to proceed, although the kind of decision allowed might be tailored to the degree of epistemic status (e.g. a low epistemic status might require adoption of no regrets policies, while higher epistemic status might allow policies requiring greater costs). That is why it is so important to determine how a statement of risk (or cost) will be used in decisions before the criterion of minimal epistemic status is established. Precaution. The earliest rulings on EPA decisions, such as the cases of benzene and vinyl 9

13 chloride, established the basic framework within which agency decisions on protection of public health are interpreted. This framework is based on three key principles: (1) that the agency need not protect individuals against ALL risk; there is some level of risk that is unacceptable, but this level is not equal to zero (even if it might be contentious where it does lie); (2) in the face of inter-subject variability of risk, it is not necessary to protect ALL individuals through environmental policies (some individuals might need to be protected by other forms of mitigation), but rather to protect some `reasonable fraction of individuals ; and (3) in the face of uncertainty, it is not necessary that a decision be based on complete confidence in the results of an assessment, but rather that there be some standard of reasonable confidence. These latter two principles are related to the general idea of precaution, and have been formalized in various concepts such as adequate protection against unreasonable risk, margin of safety, etc. Common to these concepts is recognition that risk varies throughout a population and that there is uncertainty in any estimate of risk, as well as less than perfect epistemic status in the science underlying the risk estimates. To deal with this, procedures have been put in place to ensure that decisions are based on risk estimates that are likely to reflect individuals at the upper end of any distribution of risk in a population, and (in the face of uncertainty) are more likely to overestimate risk to these individuals than to underestimate that risk. By using these conservative assumptions, the resulting risk estimates are thought of as having a margin of safety or analogous concept built into them. The margin is not necessarily known in any detailed, quantitative or probabilistic sense, but it provides risk estimates the decision-maker can reasonably claim are protective of health (and, hence, a rational basis for protection of health). Such an approach does not require a fully quantitative uncertainty analysis. It requires, instead, identification of exposure scenarios, models, parameter values, etc, that produce confidence in the claim that the resulting risk estimate is significantly more likely to be an overestimate of risk than to be an underestimate. Uncertainty analysis then is treated as a process of discourse in which the basis for this claim is judged. An advantage of the approach is that it is not very resource intensive, and has formed the basis of many decisions that have withstood court challenges. The major disadvantage is that it usually is not possible to relate the resulting risk estimate to any specific point in the inter-subject variability distribution or the uncertainty distribution, although work by Dourson and others (Dourson, Felter and Robinson, Evolution of 10

14 Science-Based Uncertainty Factors in Noncancer Risk Assessment, Regulatory Toxicology and Pharmacology, 24, , 1996) has helped move at least the concept of uncertainty factors applied in developing RfDs closer to a formal, probabilistic interpretation. Since one cannot specify the point in the variability and uncertainty distributions represented by the risk estimate, it is not possible to specify the magnitude of the margin of safety or analogous concept. This leaves the assessment open to the charge that the actual margin is not reasonable, since there is the possibility that it is either too small or too large (bearing in mind that a margin that is too small may threaten public health, and a margin that is too large carries economic consequences). Probabilistic arguments. The third realm of uncertainty analysis is in the development of decision processes rooted in probabilistic arguments. Following the discussion of margins of safety above, imagine a decision problem stated as follows: locate a policy that will protect at least X% of the exposed population against unacceptable risk Y, and that does this with Z% confidence. For example, this might be to locate an air pollution policy for which one can state with at least 90% confidence that it will protect at least 99% of the population against an excess risk of cancer of 1E-5. Such a probabilistic decision criterion requires development of a fully nested variability and uncertainty analysis. The strength of such an approach is that it allows the decision-maker to know precisely where a given estimate falls in the variability and uncertainty distributions. This removes one of the key arguments against approaches rooted in concepts like margin of safety, which is that those concepts do not allow the decision-maker to know the magnitude of any degree of conservatism. The weakness of the fully probabilistic approach is that it is significantly more resource intensive, and requires that all sources of uncertainty be described in probabilistic terms. Part of this weakness can be reduced by noting the context within which probabilistic assessments of variability and uncertainty are set. If the scenario being simulated is well specified, and if it is stated clearly that the assessment is contingent on adopting a specific set of models (e.g. a specific set of dispersion models or models of exposure-response), then one can perform a contingent uncertainty analysis. The results do not imply that the absolute uncertainty, as it would be described in an analysis of epistemic status, has been assessed. Instead they imply that given the sets of assumptions considered in the assessment, the relative confidence in any estimate of risk is reflected in the resulting PDF describing uncertainty. This 11

15 then breaks the decision process into several steps: a step bounding the decision problem by specifying the scenarios, models, data, etc to be considered; a step performing a probabilistic analysis of uncertainty conditional on these specified components; and a final judgment as to whether the original bounding of the problem was justifiable through use of a concept such as bounded rationality, which states that decision-makers cannot possibly take into account all factors that affect the outcome of a decision, but must instead focus on an approximation of the decision problem that is reasonably complete and can allow the assessment to proceed with the available resources. Questions asked from different policy motivations How do these three motivations for consideration of uncertainty in decisions differ in the questions they address? The specific questions that must be answered in the assessment (qualitative or quantitative) of uncertainty are: Sound science: Is the science sufficient to support any estimate of risk? If yes, which estimate has the highest epistemic quality, and what range includes all risk estimates that pass some standard of being reasonably sound (i.e. are scientifically plausible)? Precaution: What is the scientifically plausible upper value the risk might have in the exposed population, taking into account both inter-subject variability and risk? Probabilistic: What is the probability density function (or cumulative distribution function) describing (i) the best estimate of the variability of risk in an exposed population and (ii) the uncertainty in any specific characteristic (mean, 95 th percentile, etc) of this variability distribution? 1.3 Sources of Variability and Uncertainty As indicated previously, variability and uncertainty are two distinct concepts within a decision problem, even though they often have been lumped together in environmental analyses. Variability results from natural stochastic behavior, periodicity, or variance in a trait across a population. Uncertainty, in contrast, is related to lack of knowledge and originates from sources 12

16 such as measurement error, sampling bias, lack of data, spatial and temporal averaging, and model formulation.(bogan & Spear, 1987; Morgan & Henrion, 1990; Frey, 1992; NRC, 1994; Cullen & Frey, 1999) If unaccounted for, both variability and uncertainty have the potential to affect the quality of an environmental assessment. For example, in a human health risk assessment context, individuals in a population will have differing levels of exposure because of variability in physiology, activity patterns, and temporal and spatial variability in pollutant concentrations. Failure to account for this variability provides results that do not address the wide range of exposure values possible, and may lead to an overstatement or understatement of risk. Accounting for variability helps analysts and decision makers to determine the portion of the population that is subject to high levels of exposure and, hence, to identify priorities for mitigation. Similarly, failure to account for uncertainty may lead to assumptions of precision that do not convey the true state of knowledge. For example, an air quality modeling prediction of an 8-hour average ground-level ozone concentration of 80 parts per billion (ppb) suggests that a region may be in compliance with the 85 ppb standard. Accounting for uncertainty may provide a confidence bound of plus-or-minus 20 ppb, however, suggesting that there is a considerable likelihood for exceeding the standard. Thus, both variability and uncertainty may have ramifications on policy assessments and the ultimate success of the resulting policies. Disassociating and determining the relative effects of different sources of variability and uncertainty may be advantageous in determining how to address these factors most effectively within a decision-making context. Further, understanding which sources of variability and uncertainty are reducible provides insight to the process of determining how to most appropriately allocate resources to improve the certainty in the results of an analysis. These are both common goals associated with an uncertainty analysis. Understanding the sources of variability and uncertainty thus is critical in identifying their existence and characterizing their impact Sources of Variability Variability is present in any data set that is not static in space, time, or across members of a population. Common sources of variability include: 13

17 Stochasticity. Stochasticity is random, non-predictable behavior that is common in many physical phenomena. Stochasticity is often caused by the random motions of particles at the molecular and submolecular levels. This random motion affects processes such as dispersion, which, in turn, produce variability in pollutant concentrations across space. Stochasticity is irreducible, although it can typically be represented over time or space with a frequency distribution. Use of averaging periods can also reduce the effect of stochasticity. For example, stochastic fluctuations have limited influence on annual average particulate matter (PM) concentrations. Periodicity. Periodicity implies cyclical behavior. For example, ground-level temperature is cyclical, tending to rise during the day and decrease at night. Daily mean temperatures are also cyclical, tending to be higher in the summer and cooler in the winter. Other examples of cyclical phenomena include tides, weather patterns, global climate patterns, and emissions. Periodicity is sometimes addressed using averaging periods or assessing maximum values. Time series approaches can also be used to represent periodicity more explicitly. This is often accomplished using a sinusoidal curve or representing the value of a parameter as a function of prior values of the same parameter. Population variance. Many traits vary across a population of individuals. Examples include human physiology (e.g., height), attitudes (e.g., risk-avoidance), and activities (e.g., commute distance to work). These traits often can be represented with frequency distributions. Existence of subpopulations. In some instances, a population of data may include distinct (or not-so distinct) subpopulations that have significantly different distributions for a given trait. If the subpopulations are aggregated together, the variance of the trait in the total population may be greater than that in the individual subpopulations. If distinct subpopulations can be identified, it can be advantageous to perform an assessment of each subpopulation separately. A simple example is that of height in an adult human population. The variability in height among men or among women typically would be less than the variability in height among the entire adult population. Similarly, in a risk exposure setting, an analysis directed toward at-risk subpopulations (e.g., the elderly or children) may provide more precise information that an analysis on the entire population. 14

18 1.3.2 Sources of Uncertainty Sources of uncertainty include the following: Random error. Random error, like stochasticity, represents random, non-predictable behavior. For example, in a time series analysis, the behavior of a time-dependent trait is characterized with a function. This function typically does not capture all of the variability, however. An error term, often having a normal probability distribution, is used to capture or represent the unexplainable behavior. Lack of precision. The equipment used to measure a trait in a data set may lack precision. This, in turn, may add uncertainty to measured values. In some instances, this uncertainty may be represented using a probability distribution. To compensate for limited precision, data is sometimes censored. For example, some meteorological monitoring sites have a threshold wind speed under which only 0.0 meters per second is reported. Censoring may introduce biases that should be corrected to more adequately reflect variability. Systematic Error. Systematic error involves the introduction of bias into a data set. Bias may originate from sources such as the imperfect calibration of equipment, simplifying or incorrect assumptions, and any other errors introduced in the selection and implementation of methodologies for collecting and utilizing data. Lack of Data. Characterizing uncertainty in an input to an assessment can often be accomplished by analyzing sampled data. A consequence of the Central Limit Theorem is that the estimates of any statistics (e.g., mean, median, percentiles) of a distribution converge as sample size increases. If the sample size is not sufficient, however, sampling error is introduced into the estimates. This is a factor in many environmental and human health assessments where data are limited. For example, emissions factors used to generate emissions estimates are often calculated using a limited number of emissions tests. Thus, there is uncertainty in the true value of the statistics that describe the data. Parametric bootstrapping approaches are useful in characterizing uncertainty in the distribution parameters. Lack of representativeness. In a health or environmental assessment, data used as input to the assessment may not be completely representative of the study objectives. For example, use of 15

19 mortality data for the general population for a study involving seniors is a source of uncertainty. Similarly, emissions factor data used to estimate emissions may have been generated using different kinds of equipment or on equipment that was used under different operating or maintenance conditions. Use of surrogate data is a related source of uncertainty. When data describing an input to an assessment is unavailable, limited, or not practical to collect, surrogate data is often used. For example, toxicity studies in rodents are often used to predict toxicity in humans, even though the exact relationship between the effects in rodents and those in humans is unknown. Similarly, county-level population data are sometimes used as a surrogate to estimate area source emissions, neglecting the differences in emissions not accounted for in population data. Disagreement of experts. Expert opinion is often used to select appropriate values or distributions for inputs into an assessment. For example, experts may suggest the most appropriate reaction rate, or, in a Bayesian analysis, experts may supply a subjective prior distribution. Often different experts opinions on these data and distributions may differ, however. This disagreement results in uncertainty regarding the most appropriate values or distributions to use. Problem and scenario specification. Specification of the goals and scope of an environmental assessment includes the determination of assessment boundaries (e.g., the time period and region to be studied), the pollutant transport media (e.g., water or air), the pathways by which the pollutants reach their end points (e.g., inhalation or ingestion), and various other decision about which components to include. In some instances, problem and scenario specification decisions are made such that important factors may be omitted. For example, the computational intensity of air quality modeling often limits the number of meteorological episodes that can be evaluated. Thus, episodes that would have an impact on the assessment may be omitted. Similarly, a study involving the health effects associated with air pollutants might neglect the impacts associated with water pollution resulting from air pollution control equipment. Modeling. The computer models used in assessments represent only caricatures of reality, and thus do not capture all aspects of the problem. To meet various practical constraints (e.g., understanding of the problem, availability of data for model calibration, and time, financial, and 16

20 computational limitations), tradeoffs involving model formulation and scope must be made. Specific sources of uncertainty in modeling include: (Frey, 1992; Cullen & Frey, 1999; Isukapalli, 1999) conceptual model The analyst s conceptual model of the problem may be different from reality, with the analyst inadvertently omitting important factors, disproportionately emphasizing specific factors, omitting or misrepresenting dependence, etc. model structure There may be competing model formulations that may have advantages or disadvantages over each other. The alternative formulations may also represent different schools of thought or assumptions. mathematical implementation Model implementation and parameterization decisions, particularly in finite difference and finite element models, may result in the introduction of uncertainty through numerical instability.(hornbeck, 1975) There is also the possibility that coding errors introduce additional uncertainty. detail Model detail refers to the degree to which physical and chemical processes are described within the model. There is often the assumption is that the greater the level of detail, the greater less uncertainty in the model predictions. In some cases this is true. However, more detailed models may require the introduction of new parameters, the values of which may be uncertain. Thus, increasing detail may lead to greater uncertainty. resolution Model resolution refers to factors such as the time steps and grid sizes used. Decreasing these parameters can yield more detailed results. For example, a model with a grid size of 5 km would be expected to replicate observed values more closely than a model with a 20 km grid size. The tradeoff would be the additional computational cost. boundary conditions The selection of boundary conditions can have a significant impact upon model results. Boundary conditions, such as background concentrations and initial conditions may be based on expert judgment, assumed behavior, or the outputs of other models. Each source potentially contributes to uncertainty. aggregation Spatial and temporal averaging in modeling (i.e, determining the average 17

21 hourly concentration within a grid cell) do not facilitate the estimation of sub-grid cell or instantaneous conditions. While techniques such as kriging exist to obtain estimates of subgrid cell conditions by considering concentrations in surrounding cells, these techniques themselves introduce some uncertainty. Further models are not capable of capturing withincell variability due to random effects. calibration When applying a model for a specific application, an initial step is often to calibrate the model by comparing its performance to observed data and adjusting various parameters to improve the model performance. Calibration is a complex process. It most often requires a large number of model runs. Further, multiple calibrations could potentially produce similar results, raising the question of whether the selected calibration is the most appropriate. extrapolation Extrapolation involves the use of a model for conditions different from how it was calibrated or developed. For example, in cases where data are limited, it may be necessary to use an uncalibrated model or a model that was calibrated for a different location or meteorological episode. Another example would be the use of a statistical model (e.g., a regression model) to extrapolate behavior outside of the range in which the model was generated. 1.4 General Classification of Methods This section provides the basic rationale for a general classification of methods that are discussed further in later sections of this report. The major classes of methods include statistical methods based upon empirical data, statistical methods based upon judgment, other quantitative methods that either are approximation methods or that are non-statistical, qualitative methods, and methods for sensitivity analysis. Each of these are briefly discussed Statistical Methods Based Upon Empirical Data Statistical methods that are based upon analysis of empirical data are typically termed as frequentist methods, although sometimes the term classical is used (e.g., Warren-Hicks and Butler, 1996; Morgan and Henrion, 1990; Cullen and Frey, 1999). However, the term classical is sometimes connotated with thought experiments (e.g., what happens with a roll of a die) as opposed to inference from empirical data (DeGroot, 1986). Therefore, we use the term 18

22 frequentist. Frequentist methods are fundamentally predicated upon statistical inference based upon long-run frequencies. For example, suppose that one wishes to estimate the mean emission factor for a specific pollutant emitted from a specific source category under specific conditions. Because of the cost of collecting measurements, it is not practically to measure each and every such emission source, which would result in a census of the actual population distribution of emissions. With limited resources, one instead would prefer to randomly select a representative sample of such sources. Suppose 10 sources were selected. The mean emission rate is calculated based upon these ten sources and a probability distribution model could be fit to the random sample of data. If this process is repeated many times, with a different set of ten random samples each time, the results will vary. The variation in results for estimates of a given statistic, such as the mean, based upon random sampling is quantified using a sampling distribution. From sampling distributions, confidence intervals are obtained. Thus, the commonly used 95 percent confidence interval for the mean is a frequentist inference based upon how the estimates of the mean vary because of random sampling for a finite sample of data. Statistical inference can be used to develop compact representations of data sets (Box and Tiao, 1973). For example, a probability distribution model, such as a normal, lognormal, or other, can be fit to a random sample of empirical data. The use of probability distribution models is a convenient way to summarize information. The parameters of the distributions are subject to random sampling error and statistical methods can be applied to evaluate goodness-of-fit of the model to the data (Hahn and Shapiro, 1967). Goodness-of-fit methods are typically based upon comparison of a test statistic with a critical value, taking into account the sample size and desired level of significance (Cullen and Frey, 1999). Frequentist statistical methods are powerful tools for working with empirical data. Although there appears to be a common misperception that one must have a lot of data in order to use frequentist statistics, in fact the fundamental starting point for a frequentist analysis is to have a random representative sample. As long as this assumption is valid, it is possible to make statistical inferences even for very small data sets. The trade-off with regard to sample size is that the sampling distributions for estimates of statistics, such as the mean, distribution parameters, and others, become narrower as the sample size increases. Thus, inferences based upon data with small sample sizes will typically have wider confidence intervals than those 19

A Brief Introduction to Bayesian Statistics

A Brief Introduction to Bayesian Statistics A Brief Introduction to Statistics David Kaplan Department of Educational Psychology Methods for Social Policy Research and, Washington, DC 2017 1 / 37 The Reverend Thomas Bayes, 1701 1761 2 / 37 Pierre-Simon

More information

Model calibration and Bayesian methods for probabilistic projections

Model calibration and Bayesian methods for probabilistic projections ETH Zurich Reto Knutti Model calibration and Bayesian methods for probabilistic projections Reto Knutti, IAC ETH Toy model Model: obs = linear trend + noise(variance, spectrum) 1) Short term predictability,

More information

Expert judgements in risk and reliability analysis

Expert judgements in risk and reliability analysis Expert judgements in risk and reliability analysis Jan-Erik Holmberg M2-E2117 Risk analysis, April 17, 2016 Aalto university Contents Definitions Applications Bayesian approach Examples 2 Background Objective,

More information

Climate-Related Decision Making Under Uncertainty. Chris Weaver February 2016

Climate-Related Decision Making Under Uncertainty. Chris Weaver February 2016 Climate-Related Decision Making Under Uncertainty Chris Weaver February 2016 Why is climate change decision making hard? Climate change is a uniquely tricky problem But effective, science-based decision

More information

6.3.5 Uncertainty Assessment

6.3.5 Uncertainty Assessment 6.3.5 Uncertainty Assessment Because risk characterization is a bridge between risk assessment and risk management, it is important that the major assumptions, professional judgments, and estimates of

More information

Evaluation Models STUDIES OF DIAGNOSTIC EFFICIENCY

Evaluation Models STUDIES OF DIAGNOSTIC EFFICIENCY 2. Evaluation Model 2 Evaluation Models To understand the strengths and weaknesses of evaluation, one must keep in mind its fundamental purpose: to inform those who make decisions. The inferences drawn

More information

PAH - Sources and Measurement Mike Woodfield - AEAT

PAH - Sources and Measurement Mike Woodfield - AEAT PAH - Sources and Measurement Mike Woodfield - AEAT The Background (1) PAH in Ambient include substances known to produce cancer; World Health Organisation considers the risk to human health from ambient

More information

Journal of Political Economy, Vol. 93, No. 2 (Apr., 1985)

Journal of Political Economy, Vol. 93, No. 2 (Apr., 1985) Confirmations and Contradictions Journal of Political Economy, Vol. 93, No. 2 (Apr., 1985) Estimates of the Deterrent Effect of Capital Punishment: The Importance of the Researcher's Prior Beliefs Walter

More information

Bayesian and Frequentist Approaches

Bayesian and Frequentist Approaches Bayesian and Frequentist Approaches G. Jogesh Babu Penn State University http://sites.stat.psu.edu/ babu http://astrostatistics.psu.edu All models are wrong But some are useful George E. P. Box (son-in-law

More information

Quantifying uncertainty under a predictive, epistemic approach to risk analysis

Quantifying uncertainty under a predictive, epistemic approach to risk analysis Quantifying uncertainty under a predictive, epistemic approach to risk analysis S.Apeland Stavanger University College, Norway T.Aven Stavanger University College, Norway T.Nilsen RF Rogaland Research,

More information

Analysis of Environmental Data Conceptual Foundations: En viro n m e n tal Data

Analysis of Environmental Data Conceptual Foundations: En viro n m e n tal Data Analysis of Environmental Data Conceptual Foundations: En viro n m e n tal Data 1. Purpose of data collection...................................................... 2 2. Samples and populations.......................................................

More information

Risk Characterization

Risk Characterization Risk Characterization 1 Learning Objectives By the end of this module, participants should have an understanding of: The purpose of risk characterization Integrating the results of hazard identification,

More information

John Quigley, Tim Bedford, Lesley Walls Department of Management Science, University of Strathclyde, Glasgow

John Quigley, Tim Bedford, Lesley Walls Department of Management Science, University of Strathclyde, Glasgow Empirical Bayes Estimates of Development Reliability for One Shot Devices John Quigley, Tim Bedford, Lesley Walls Department of Management Science, University of Strathclyde, Glasgow This article describes

More information

Fundamental Concepts for Using Diagnostic Classification Models. Section #2 NCME 2016 Training Session. NCME 2016 Training Session: Section 2

Fundamental Concepts for Using Diagnostic Classification Models. Section #2 NCME 2016 Training Session. NCME 2016 Training Session: Section 2 Fundamental Concepts for Using Diagnostic Classification Models Section #2 NCME 2016 Training Session NCME 2016 Training Session: Section 2 Lecture Overview Nature of attributes What s in a name? Grain

More information

The Context of Detection and Attribution

The Context of Detection and Attribution 10.2.1 The Context of Detection and Attribution In IPCC Assessments, detection and attribution involve quantifying the evidence for a causal link between external drivers of climate change and observed

More information

You must answer question 1.

You must answer question 1. Research Methods and Statistics Specialty Area Exam October 28, 2015 Part I: Statistics Committee: Richard Williams (Chair), Elizabeth McClintock, Sarah Mustillo You must answer question 1. 1. Suppose

More information

BACKGROUND + GENERAL COMMENTS

BACKGROUND + GENERAL COMMENTS Response on behalf of Sobi (Swedish Orphan Biovitrum AB) to the European Commission s Public Consultation on a Commission Notice on the Application of Articles 3, 5 and 7 of Regulation (EC) No. 141/2000

More information

Modeling Terrorist Beliefs and Motivations

Modeling Terrorist Beliefs and Motivations CREATE Research Archive Research Project Summaries 2010 Modeling Terrorist Beliefs and Motivations Richard John University of Southern California, richardj@usc.edu Heather Rosoff University of Southern

More information

Technical Specifications

Technical Specifications Technical Specifications In order to provide summary information across a set of exercises, all tests must employ some form of scoring models. The most familiar of these scoring models is the one typically

More information

1, 2, 3-Trichloropropane (TCP): Assessment of Risks from Drinking Water

1, 2, 3-Trichloropropane (TCP): Assessment of Risks from Drinking Water 1, 2, 3-Trichloropropane (TCP): Assessment of Risks from Drinking Water 1, 2, 3-Trichloropropane (TCP): Assessment of Risks from Drinking Water Prepared for Alaimo Group (Engineers and Architects) 200

More information

Correlation Neglect in Belief Formation

Correlation Neglect in Belief Formation Correlation Neglect in Belief Formation Benjamin Enke Florian Zimmermann Bonn Graduate School of Economics University of Zurich NYU Bounded Rationality in Choice Conference May 31, 2015 Benjamin Enke (Bonn)

More information

CHAPTER 8 RISK CHARACTERIZATION

CHAPTER 8 RISK CHARACTERIZATION CHAPTER 8 RISK CHARACTERIZATION This chapter describes the final step of the baseline health risk assessment process, risk characterization. In this step, the toxicity and exposure assessments are summarized

More information

Combining Risks from Several Tumors Using Markov Chain Monte Carlo

Combining Risks from Several Tumors Using Markov Chain Monte Carlo University of Nebraska - Lincoln DigitalCommons@University of Nebraska - Lincoln U.S. Environmental Protection Agency Papers U.S. Environmental Protection Agency 2009 Combining Risks from Several Tumors

More information

Answers to end of chapter questions

Answers to end of chapter questions Answers to end of chapter questions Chapter 1 What are the three most important characteristics of QCA as a method of data analysis? QCA is (1) systematic, (2) flexible, and (3) it reduces data. What are

More information

Excess BTM Production. Issue Paper

Excess BTM Production. Issue Paper Excess BTM Production Issue Paper Date paper published: June 28, 2018 Table of Contents 1. Executive Summary... 3 2. Plan for Stakeholder Engagement... 4 3. Background... 4 4. Scope... 5 5. Clarification

More information

Russian Journal of Agricultural and Socio-Economic Sciences, 3(15)

Russian Journal of Agricultural and Socio-Economic Sciences, 3(15) ON THE COMPARISON OF BAYESIAN INFORMATION CRITERION AND DRAPER S INFORMATION CRITERION IN SELECTION OF AN ASYMMETRIC PRICE RELATIONSHIP: BOOTSTRAP SIMULATION RESULTS Henry de-graft Acquah, Senior Lecturer

More information

NEW METHODS FOR SENSITIVITY TESTS OF EXPLOSIVE DEVICES

NEW METHODS FOR SENSITIVITY TESTS OF EXPLOSIVE DEVICES NEW METHODS FOR SENSITIVITY TESTS OF EXPLOSIVE DEVICES Amit Teller 1, David M. Steinberg 2, Lina Teper 1, Rotem Rozenblum 2, Liran Mendel 2, and Mordechai Jaeger 2 1 RAFAEL, POB 2250, Haifa, 3102102, Israel

More information

CSC2130: Empirical Research Methods for Software Engineering

CSC2130: Empirical Research Methods for Software Engineering CSC2130: Empirical Research Methods for Software Engineering Steve Easterbrook sme@cs.toronto.edu www.cs.toronto.edu/~sme/csc2130/ 2004-5 Steve Easterbrook. This presentation is available free for non-commercial

More information

Rethinking Approaches To Low Dose Extrapolation for Environmental Health Risk Assessment

Rethinking Approaches To Low Dose Extrapolation for Environmental Health Risk Assessment Rethinking Approaches To Low Dose Extrapolation for Environmental Health Risk Assessment Introduction Over the past half-century, methods for quantitatively assessing the human health risks from exposures

More information

DRAFT (Final) Concept Paper On choosing appropriate estimands and defining sensitivity analyses in confirmatory clinical trials

DRAFT (Final) Concept Paper On choosing appropriate estimands and defining sensitivity analyses in confirmatory clinical trials DRAFT (Final) Concept Paper On choosing appropriate estimands and defining sensitivity analyses in confirmatory clinical trials EFSPI Comments Page General Priority (H/M/L) Comment The concept to develop

More information

ANIMAL HEALTH SURVEILLANCE

ANIMAL HEALTH SURVEILLANCE 291 Annex 33a CHAPTER 1.4. ANIMAL HEALTH SURVEILLANCE Article 1.4.1. Introduction and objectives 1) In general, surveillance is aimed at demonstrating the absence of infection or infestation, determining

More information

CARISMA-LMS Workshop on Statistics for Risk Analysis

CARISMA-LMS Workshop on Statistics for Risk Analysis Department of Mathematics CARISMA-LMS Workshop on Statistics for Risk Analysis Thursday 28 th May 2015 Location: Department of Mathematics, John Crank Building, Room JNCK128 (Campus map can be found at

More information

Lec 02: Estimation & Hypothesis Testing in Animal Ecology

Lec 02: Estimation & Hypothesis Testing in Animal Ecology Lec 02: Estimation & Hypothesis Testing in Animal Ecology Parameter Estimation from Samples Samples We typically observe systems incompletely, i.e., we sample according to a designed protocol. We then

More information

Unit 1 Exploring and Understanding Data

Unit 1 Exploring and Understanding Data Unit 1 Exploring and Understanding Data Area Principle Bar Chart Boxplot Conditional Distribution Dotplot Empirical Rule Five Number Summary Frequency Distribution Frequency Polygon Histogram Interquartile

More information

[1] provides a philosophical introduction to the subject. Simon [21] discusses numerous topics in economics; see [2] for a broad economic survey.

[1] provides a philosophical introduction to the subject. Simon [21] discusses numerous topics in economics; see [2] for a broad economic survey. Draft of an article to appear in The MIT Encyclopedia of the Cognitive Sciences (Rob Wilson and Frank Kiel, editors), Cambridge, Massachusetts: MIT Press, 1997. Copyright c 1997 Jon Doyle. All rights reserved

More information

Discussion Meeting for MCP-Mod Qualification Opinion Request. Novartis 10 July 2013 EMA, London, UK

Discussion Meeting for MCP-Mod Qualification Opinion Request. Novartis 10 July 2013 EMA, London, UK Discussion Meeting for MCP-Mod Qualification Opinion Request Novartis 10 July 2013 EMA, London, UK Attendees Face to face: Dr. Frank Bretz Global Statistical Methodology Head, Novartis Dr. Björn Bornkamp

More information

Part 2. Chemical and physical aspects

Part 2. Chemical and physical aspects Part 2. Chemical and physical aspects 12. Chemical and physical aspects: introduction 12.1 Background information used The assessment of the toxicity of drinking-water contaminants has been made on the

More information

Improving the U.S. Public s Understanding of Climate Change

Improving the U.S. Public s Understanding of Climate Change Improving the U.S. Public s Understanding of Climate Change Paul C. Stern, Director, Committee on the Human Dimensions of Global Change National Research Council, USA Presentation to Symposium on Communicating

More information

FINAL. Recommendations for Update to Arsenic Soil CTL Computation. Methodology Focus Group. Contaminated Soils Forum. Prepared by:

FINAL. Recommendations for Update to Arsenic Soil CTL Computation. Methodology Focus Group. Contaminated Soils Forum. Prepared by: A stakeholder body advising the Florida Department of Environmental Protection FINAL Recommendations for Update to Arsenic Soil CTL Computation Prepared by: Methodology Focus Group Contaminated Soils Forum

More information

US power plant carbon standards and clean air and health co-benefits

US power plant carbon standards and clean air and health co-benefits US power plant carbon standards and clean air and health co-benefits Emissions Modeling For each scenario, we used emissions output from ICF International s Integrated Planning Model (IPM) conducted by

More information

JunkScience.com. June 16, Mr. Ray Wassel National Academy of Sciences 500 Fifth St., N.W. Washington, D.C

JunkScience.com. June 16, Mr. Ray Wassel National Academy of Sciences 500 Fifth St., N.W. Washington, D.C JunkScience.com June 16, 2016 Mr. Ray Wassel National Academy of Sciences 500 Fifth St., N.W. Washington, D.C. 20001 Re: Comment on Assessing Toxicologic Risks to Human Subjects Used in Controlled Exposure

More information

Actor (and network) models

Actor (and network) models Actor (and network) models EPA2142 Policy and Decision Models Research Seminar Leon Hermans 29-06-10 Delft University of Technology The original lecture slides by Leon Hermans have been adapted for OpenCourseWare

More information

Background EVM. FAO/WHO technical workshop on nutrient risk assessment, Geneva, May 2005, published 2006.

Background EVM. FAO/WHO technical workshop on nutrient risk assessment, Geneva, May 2005, published 2006. UK GOVERNMENT RESPONSE TO THE EUROPEAN COMMISSION S DISCUSSION PAPER ON THE SETTING OF MAXIMUM AND MINIMUM AMOUNTS FOR VITAMINS AND MINERALS IN FOODSTUFFS. Background The United Kingdom (UK) Government

More information

Practitioner s Guide To Stratified Random Sampling: Part 1

Practitioner s Guide To Stratified Random Sampling: Part 1 Practitioner s Guide To Stratified Random Sampling: Part 1 By Brian Kriegler November 30, 2018, 3:53 PM EST This is the first of two articles on stratified random sampling. In the first article, I discuss

More information

PLANNING THE RESEARCH PROJECT

PLANNING THE RESEARCH PROJECT Van Der Velde / Guide to Business Research Methods First Proof 6.11.2003 4:53pm page 1 Part I PLANNING THE RESEARCH PROJECT Van Der Velde / Guide to Business Research Methods First Proof 6.11.2003 4:53pm

More information

Agreement Coefficients and Statistical Inference

Agreement Coefficients and Statistical Inference CHAPTER Agreement Coefficients and Statistical Inference OBJECTIVE This chapter describes several approaches for evaluating the precision associated with the inter-rater reliability coefficients of the

More information

PHO MetaQAT Guide. Critical appraisal in public health. PHO Meta-tool for quality appraisal

PHO MetaQAT Guide. Critical appraisal in public health. PHO Meta-tool for quality appraisal PHO MetaQAT Guide Critical appraisal in public health Critical appraisal is a necessary part of evidence-based practice and decision-making, allowing us to understand the strengths and weaknesses of evidence,

More information

Chapter 1 Social Science and Its Methods

Chapter 1 Social Science and Its Methods Chapter 1 Social Science and Its Methods MULTIPLE CHOICE 1) Scientific knowledge is knowledge that has been: A) systematically gathered, classified, related, and interpreted. B) rediscovered and cherished

More information

FMEA AND RPN NUMBERS. Failure Mode Severity Occurrence Detection RPN A B

FMEA AND RPN NUMBERS. Failure Mode Severity Occurrence Detection RPN A B FMEA AND RPN NUMBERS An important part of risk is to remember that risk is a vector: one aspect of risk is the severity of the effect of the event and the other aspect is the probability or frequency of

More information

6. A theory that has been substantially verified is sometimes called a a. law. b. model.

6. A theory that has been substantially verified is sometimes called a a. law. b. model. Chapter 2 Multiple Choice Questions 1. A theory is a(n) a. a plausible or scientifically acceptable, well-substantiated explanation of some aspect of the natural world. b. a well-substantiated explanation

More information

Chapter 02 Developing and Evaluating Theories of Behavior

Chapter 02 Developing and Evaluating Theories of Behavior Chapter 02 Developing and Evaluating Theories of Behavior Multiple Choice Questions 1. A theory is a(n): A. plausible or scientifically acceptable, well-substantiated explanation of some aspect of the

More information

2012 Course : The Statistician Brain: the Bayesian Revolution in Cognitive Science

2012 Course : The Statistician Brain: the Bayesian Revolution in Cognitive Science 2012 Course : The Statistician Brain: the Bayesian Revolution in Cognitive Science Stanislas Dehaene Chair in Experimental Cognitive Psychology Lecture No. 4 Constraints combination and selection of a

More information

Guidance Note for Lead Authors of the IPCC Fifth Assessment Report on Consistent Treatment of Uncertainties

Guidance Note for Lead Authors of the IPCC Fifth Assessment Report on Consistent Treatment of Uncertainties Guidance Note for Lead Authors of the IPCC Fifth Assessment Report on Consistent Treatment of Uncertainties IPCC Cross-Working Group Meeting on Consistent Treatment of Uncertainties Jasper Ridge, CA, USA

More information

Bayesian Tolerance Intervals for Sparse Data Margin Assessment

Bayesian Tolerance Intervals for Sparse Data Margin Assessment Bayesian Tolerance Intervals for Sparse Data Margin Assessment Benjamin Schroeder and Lauren Hund ASME V&V Symposium May 3, 2017 - Las Vegas, NV SAND2017-4590 C - (UUR) Sandia National Laboratories is

More information

International Conference on Harmonisation of Technical Requirements for Registration of Pharmaceuticals for Human Use

International Conference on Harmonisation of Technical Requirements for Registration of Pharmaceuticals for Human Use Final Concept Paper E9(R1): Addendum to Statistical Principles for Clinical Trials on Choosing Appropriate Estimands and Defining Sensitivity Analyses in Clinical Trials dated 22 October 2014 Endorsed

More information

The Regression-Discontinuity Design

The Regression-Discontinuity Design Page 1 of 10 Home» Design» Quasi-Experimental Design» The Regression-Discontinuity Design The regression-discontinuity design. What a terrible name! In everyday language both parts of the term have connotations

More information

Chapter 02. Basic Research Methodology

Chapter 02. Basic Research Methodology Chapter 02 Basic Research Methodology Definition RESEARCH Research is a quest for knowledge through diligent search or investigation or experimentation aimed at the discovery and interpretation of new

More information

A Proposal for the Validation of Control Banding Using Bayesian Decision Analysis Techniques

A Proposal for the Validation of Control Banding Using Bayesian Decision Analysis Techniques A Proposal for the Validation of Control Banding Using Bayesian Decision Analysis Techniques Paul Hewett Exposure Assessment Solutions, Inc. Morgantown, WV John Mulhausen 3M Company St. Paul, MN Perry

More information

EXECUTIVE SUMMARY 9. Executive Summary

EXECUTIVE SUMMARY 9. Executive Summary EXECUTIVE SUMMARY 9 Executive Summary Education affects people s lives in ways that go far beyond what can be measured by labour market earnings and economic growth. Important as they are, these social

More information

Confirmation Bias. this entry appeared in pp of in M. Kattan (Ed.), The Encyclopedia of Medical Decision Making.

Confirmation Bias. this entry appeared in pp of in M. Kattan (Ed.), The Encyclopedia of Medical Decision Making. Confirmation Bias Jonathan D Nelson^ and Craig R M McKenzie + this entry appeared in pp. 167-171 of in M. Kattan (Ed.), The Encyclopedia of Medical Decision Making. London, UK: Sage the full Encyclopedia

More information

Bayesian Confidence Intervals for Means and Variances of Lognormal and Bivariate Lognormal Distributions

Bayesian Confidence Intervals for Means and Variances of Lognormal and Bivariate Lognormal Distributions Bayesian Confidence Intervals for Means and Variances of Lognormal and Bivariate Lognormal Distributions J. Harvey a,b, & A.J. van der Merwe b a Centre for Statistical Consultation Department of Statistics

More information

CHAPTER 2: RISK ANALYSIS

CHAPTER 2: RISK ANALYSIS Update Project Chapter : Risk Analysis Draft May 00 0 0 0 0 0 PRINCIPLES AND METHODS FOR THE RISK ASSESSMENT OF CHEMICALS IN FOOD CHAPTER : RISK ANALYSIS Contents CHAPTER : RISK ANALYSIS.... INTRODUCTION....

More information

ISC- GRADE XI HUMANITIES ( ) PSYCHOLOGY. Chapter 2- Methods of Psychology

ISC- GRADE XI HUMANITIES ( ) PSYCHOLOGY. Chapter 2- Methods of Psychology ISC- GRADE XI HUMANITIES (2018-19) PSYCHOLOGY Chapter 2- Methods of Psychology OUTLINE OF THE CHAPTER (i) Scientific Methods in Psychology -observation, case study, surveys, psychological tests, experimentation

More information

January 2, Overview

January 2, Overview American Statistical Association Position on Statistical Statements for Forensic Evidence Presented under the guidance of the ASA Forensic Science Advisory Committee * January 2, 2019 Overview The American

More information

A Decision-Theoretic Approach to Evaluating Posterior Probabilities of Mental Models

A Decision-Theoretic Approach to Evaluating Posterior Probabilities of Mental Models A Decision-Theoretic Approach to Evaluating Posterior Probabilities of Mental Models Jonathan Y. Ito and David V. Pynadath and Stacy C. Marsella Information Sciences Institute, University of Southern California

More information

Handling Partial Preferences in the Belief AHP Method: Application to Life Cycle Assessment

Handling Partial Preferences in the Belief AHP Method: Application to Life Cycle Assessment Handling Partial Preferences in the Belief AHP Method: Application to Life Cycle Assessment Amel Ennaceur 1, Zied Elouedi 1, and Eric Lefevre 2 1 University of Tunis, Institut Supérieur de Gestion de Tunis,

More information

Analysis A step in the research process that involves describing and then making inferences based on a set of data.

Analysis A step in the research process that involves describing and then making inferences based on a set of data. 1 Appendix 1:. Definitions of important terms. Additionality The difference between the value of an outcome after the implementation of a policy, and its value in a counterfactual scenario in which the

More information

Kahneman, Daniel. Thinking Fast and Slow. New York: Farrar, Straus & Giroux, 2011.

Kahneman, Daniel. Thinking Fast and Slow. New York: Farrar, Straus & Giroux, 2011. The accumulating research indicates that individuals cognitive and behavioral orientations to objects (their thoughts and actions) are frequently based on rapid shortcuts or heuristics. The past few decades

More information

Pooling Subjective Confidence Intervals

Pooling Subjective Confidence Intervals Spring, 1999 1 Administrative Things Pooling Subjective Confidence Intervals Assignment 7 due Friday You should consider only two indices, the S&P and the Nikkei. Sorry for causing the confusion. Reading

More information

Running head: INDIVIDUAL DIFFERENCES 1. Why to treat subjects as fixed effects. James S. Adelman. University of Warwick.

Running head: INDIVIDUAL DIFFERENCES 1. Why to treat subjects as fixed effects. James S. Adelman. University of Warwick. Running head: INDIVIDUAL DIFFERENCES 1 Why to treat subjects as fixed effects James S. Adelman University of Warwick Zachary Estes Bocconi University Corresponding Author: James S. Adelman Department of

More information

Recent developments for combining evidence within evidence streams: bias-adjusted meta-analysis

Recent developments for combining evidence within evidence streams: bias-adjusted meta-analysis EFSA/EBTC Colloquium, 25 October 2017 Recent developments for combining evidence within evidence streams: bias-adjusted meta-analysis Julian Higgins University of Bristol 1 Introduction to concepts Standard

More information

Scientific Working Group on Digital Evidence

Scientific Working Group on Digital Evidence Disclaimer: As a condition to the use of this document and the information contained therein, the SWGDE requests notification by e-mail before or contemporaneous to the introduction of this document, or

More information

Competency Rubric Bank for the Sciences (CRBS)

Competency Rubric Bank for the Sciences (CRBS) Competency Rubric Bank for the Sciences (CRBS) Content Knowledge 1 Content Knowledge: Accuracy of scientific understanding Higher Order Cognitive Skills (HOCS) 3 Analysis: Clarity of Research Question

More information

Addendum to the 12th Report on Carcinogens

Addendum to the 12th Report on Carcinogens Addendum to the 12th Report on Carcinogens Published by the U.S. Department of Health and Human Services, National Toxicology Program The twelfth edition of the National Toxicology Program (NTP) Report

More information

Putting it together: The potential role of modeling to explore the impact of FMD

Putting it together: The potential role of modeling to explore the impact of FMD Putting it together: The potential role of modeling to explore the impact of FMD Mo Salman Animal Population Health Institute Colorado State University m.d.salman@colostate.edu And Melissa McLaws European

More information

AUG Expert responses to issues raised by British Gas during the query period for the first draft 2017/18 AUG Statement, 14 March 2017.

AUG Expert responses to issues raised by British Gas during the query period for the first draft 2017/18 AUG Statement, 14 March 2017. AUG Expert responses to issues raised by British Gas during the query period for the first draft 2017/18 AUG Statement, 14 March 2017. Below are the issues raised by British Gas. Following each issue is

More information

Cochrane Pregnancy and Childbirth Group Methodological Guidelines

Cochrane Pregnancy and Childbirth Group Methodological Guidelines Cochrane Pregnancy and Childbirth Group Methodological Guidelines [Prepared by Simon Gates: July 2009, updated July 2012] These guidelines are intended to aid quality and consistency across the reviews

More information

By the end of the activities described in Section 5, consensus has

By the end of the activities described in Section 5, consensus has 6 Selecting Indicators 81 By the end of the activities described in Section 5, consensus has been reached on the goals and objectives of the project, the information which needs to be collected and analyzed

More information

Sources of uncertainty in intuitive physics

Sources of uncertainty in intuitive physics Sources of uncertainty in intuitive physics Kevin A Smith (k2smith@ucsd.edu) and Edward Vul (evul@ucsd.edu) University of California, San Diego Department of Psychology, 9500 Gilman Dr. La Jolla, CA 92093

More information

The Logotherapy Evidence Base: A Practitioner s Review. Marshall H. Lewis

The Logotherapy Evidence Base: A Practitioner s Review. Marshall H. Lewis Why A Practitioner s Review? The Logotherapy Evidence Base: A Practitioner s Review Marshall H. Lewis Why do we need a practitioner s review of the logotherapy evidence base? Isn t this something we should

More information

An Emerging New Risk Analysis Science: Foundations and Implications

An Emerging New Risk Analysis Science: Foundations and Implications Risk Analysis, Vol. 38, No. 5, 2018 DOI: 10.1111/risa.12899 An Emerging New Risk Analysis Science: Foundations and Implications Terje Aven To solve real-life problems such as those related to technology,

More information

Welcome to next lecture in the class. During this week we will introduce the concepts of risk and hazard analysis and go over the processes that

Welcome to next lecture in the class. During this week we will introduce the concepts of risk and hazard analysis and go over the processes that Welcome to next lecture in the class. During this week we will introduce the concepts of risk and hazard analysis and go over the processes that these analysts use and how this can relate to fire and fuels

More information

International Framework for Assurance Engagements

International Framework for Assurance Engagements FRAMEWORK March 2015 Framework International Framework for Assurance Engagements Explanatory Foreword The Council of the Malaysian Institute of Accountants has approved this International Framework for

More information

High-level Vision. Bernd Neumann Slides for the course in WS 2004/05. Faculty of Informatics Hamburg University Germany

High-level Vision. Bernd Neumann Slides for the course in WS 2004/05. Faculty of Informatics Hamburg University Germany High-level Vision Bernd Neumann Slides for the course in WS 2004/05 Faculty of Informatics Hamburg University Germany neumann@informatik.uni-hamburg.de http://kogs-www.informatik.uni-hamburg.de 1 Contents

More information

Macroeconometric Analysis. Chapter 1. Introduction

Macroeconometric Analysis. Chapter 1. Introduction Macroeconometric Analysis Chapter 1. Introduction Chetan Dave David N. DeJong 1 Background The seminal contribution of Kydland and Prescott (1982) marked the crest of a sea change in the way macroeconomists

More information

A Case Study: Two-sample categorical data

A Case Study: Two-sample categorical data A Case Study: Two-sample categorical data Patrick Breheny January 31 Patrick Breheny BST 701: Bayesian Modeling in Biostatistics 1/43 Introduction Model specification Continuous vs. mixture priors Choice

More information

E11(R1) Addendum to E11: Clinical Investigation of Medicinal Products in the Pediatric Population Step4

E11(R1) Addendum to E11: Clinical Investigation of Medicinal Products in the Pediatric Population Step4 E11(R1) Addendum to E11: Clinical Investigation of Medicinal Products in the Step4 October 2017 International Council for Harmonisation of Technical Requirements for Pharmaceuticals for Human Use 1 Legal

More information

APPENDIX IV.I.1. Delivery Truck HRA

APPENDIX IV.I.1. Delivery Truck HRA APPENDIX IV.I.1 Delivery Truck HRA Diesel Particulate Matter (DPM) Health Risk Assessment (HRA) of Delivery Trucks from the Project at Pico and Sepulveda, Los Angeles, California Introduction This health

More information

2012 Course: The Statistician Brain: the Bayesian Revolution in Cognitive Sciences

2012 Course: The Statistician Brain: the Bayesian Revolution in Cognitive Sciences 2012 Course: The Statistician Brain: the Bayesian Revolution in Cognitive Sciences Stanislas Dehaene Chair of Experimental Cognitive Psychology Lecture n 5 Bayesian Decision-Making Lecture material translated

More information

European Federation of Statisticians in the Pharmaceutical Industry (EFSPI)

European Federation of Statisticians in the Pharmaceutical Industry (EFSPI) Page 1 of 14 European Federation of Statisticians in the Pharmaceutical Industry (EFSPI) COMMENTS ON DRAFT FDA Guidance for Industry - Non-Inferiority Clinical Trials Rapporteur: Bernhard Huitfeldt (bernhard.huitfeldt@astrazeneca.com)

More information

UNIT 4 ALGEBRA II TEMPLATE CREATED BY REGION 1 ESA UNIT 4

UNIT 4 ALGEBRA II TEMPLATE CREATED BY REGION 1 ESA UNIT 4 UNIT 4 ALGEBRA II TEMPLATE CREATED BY REGION 1 ESA UNIT 4 Algebra II Unit 4 Overview: Inferences and Conclusions from Data In this unit, students see how the visual displays and summary statistics they

More information

Approaches to Integrating Evidence From Animal and Human Studies in Chemical Assessments

Approaches to Integrating Evidence From Animal and Human Studies in Chemical Assessments Approaches to Integrating Evidence From Animal and Human Studies in Chemical Assessments Kris Thayer, National Center for Environmental Assessment (NCEA) Integrated Risk Information System (IRIS) Division

More information

Political Science 15, Winter 2014 Final Review

Political Science 15, Winter 2014 Final Review Political Science 15, Winter 2014 Final Review The major topics covered in class are listed below. You should also take a look at the readings listed on the class website. Studying Politics Scientifically

More information

Methods of eliciting time preferences for health A pilot study

Methods of eliciting time preferences for health A pilot study Methods of eliciting time preferences for health A pilot study Dorte Gyrd-Hansen Health Economics Papers 2000:1 Methods of eliciting time preferences for health A pilot study Dorte Gyrd-Hansen Contents

More information

Inconsistent Inference in Qualitative Risk Assessment

Inconsistent Inference in Qualitative Risk Assessment Inconsistent Inference in Qualitative Risk Assessment November 10, 2013 Prepared by Kailan Shang 1 1 Kailan Shang, FSA, CFA, PRM, SCJP, of Manulife Financial, can be reached at klshang81@gmail.com. Page

More information

Should Cochrane apply error-adjustment methods when conducting repeated meta-analyses?

Should Cochrane apply error-adjustment methods when conducting repeated meta-analyses? Cochrane Scientific Committee Should Cochrane apply error-adjustment methods when conducting repeated meta-analyses? Document initially prepared by Christopher Schmid and Jackie Chandler Edits by Panel

More information

EXPERIMENTAL RESEARCH DESIGNS

EXPERIMENTAL RESEARCH DESIGNS ARTHUR PSYC 204 (EXPERIMENTAL PSYCHOLOGY) 14A LECTURE NOTES [02/28/14] EXPERIMENTAL RESEARCH DESIGNS PAGE 1 Topic #5 EXPERIMENTAL RESEARCH DESIGNS As a strict technical definition, an experiment is a study

More information

Continuous/Discrete Non Parametric Bayesian Belief Nets with UNICORN and UNINET

Continuous/Discrete Non Parametric Bayesian Belief Nets with UNICORN and UNINET Continuous/Discrete Non Parametric Bayesian Belief Nets with UNICORN and UNINET R.M. Cooke 1, D. Kurowicka, A. M. Hanea, O. Morales, D. A. Ababei Department of Mathematics Delft University of Technology

More information