Bayesian and Frequentist Approaches

Similar documents
A Brief Introduction to Bayesian Statistics

BAYESIAN HYPOTHESIS TESTING WITH SPSS AMOS

Introduction to Bayesian Analysis 1

Hierarchy of Statistical Goals

A Case Study: Two-sample categorical data

STATISTICAL INFERENCE 1 Richard A. Johnson Professor Emeritus Department of Statistics University of Wisconsin

Introduction. Patrick Breheny. January 10. The meaning of probability The Bayesian approach Preview of MCMC methods

What is a probability? Two schools in statistics: frequentists and Bayesians.

Bayesian Statistics Estimation of a Single Mean and Variance MCMC Diagnostics and Missing Data

Chapter 23. Inference About Means. Copyright 2010 Pearson Education, Inc.

Bayesian Inference Bayes Laplace

Statistical Decision Theory and Bayesian Analysis

Lec 02: Estimation & Hypothesis Testing in Animal Ecology

Fundamental Clinical Trial Design

ST440/550: Applied Bayesian Statistics. (10) Frequentist Properties of Bayesian Methods

MS&E 226: Small Data

Combining Risks from Several Tumors Using Markov Chain Monte Carlo

Response to the ASA s statement on p-values: context, process, and purpose

How Does Analysis of Competing Hypotheses (ACH) Improve Intelligence Analysis?

BOOTSTRAPPING CONFIDENCE LEVELS FOR HYPOTHESES ABOUT QUADRATIC (U-SHAPED) REGRESSION MODELS

Practical Bayesian Design and Analysis for Drug and Device Clinical Trials

Att vara eller inte vara (en Bayesian)?... Sherlock-conundrum

Lecture Outline Biost 517 Applied Biostatistics I. Statistical Goals of Studies Role of Statistical Inference

Using historical data for Bayesian sample size determination

Lecture Outline. Biost 590: Statistical Consulting. Stages of Scientific Studies. Scientific Method

Journal of Clinical and Translational Research special issue on negative results /jctres S2.007

Missing data. Patrick Breheny. April 23. Introduction Missing response data Missing covariate data

An Introduction to Bayesian Statistics

1 The conceptual underpinnings of statistical power

Inference About Magnitudes of Effects

Introductory Statistical Inference with the Likelihood Function

Bayesian Confidence Intervals for Means and Variances of Lognormal and Bivariate Lognormal Distributions

Bayesian and Classical Hypothesis Testing: Practical Differences for a Controversial Area of Research

Lecture Outline Biost 517 Applied Biostatistics I

Statistics and Full Belief

Handout on Perfect Bayesian Equilibrium

11/18/2013. Correlational Research. Correlational Designs. Why Use a Correlational Design? CORRELATIONAL RESEARCH STUDIES

Advanced Bayesian Models for the Social Sciences. TA: Elizabeth Menninga (University of North Carolina, Chapel Hill)

Full title: A likelihood-based approach to early stopping in single arm phase II cancer clinical trials

Signal Detection Theory and Bayesian Modeling

Power & Sample Size. Dr. Andrea Benedetti

Advanced Bayesian Models for the Social Sciences

For general queries, contact

The Century of Bayes

Introduction & Basics

Lecture 2: Learning and Equilibrium Extensive-Form Games

Chapter 02. Basic Research Methodology

Introduction On Assessing Agreement With Continuous Measurement

Model calibration and Bayesian methods for probabilistic projections

Coherence and calibration: comments on subjectivity and objectivity in Bayesian analysis (Comment on Articles by Berger and by Goldstein)

Beyond Subjective and Objective in Statistics

What is probability. A way of quantifying uncertainty. Mathematical theory originally developed to model outcomes in games of chance.

Bayesian Methods in Regulatory Science

Bayesian Hierarchical Models for Fitting Dose-Response Relationships

Hierarchical Bayesian Modeling of Individual Differences in Texture Discrimination

Recognizing Ambiguity

arxiv: v1 [math.st] 8 Jun 2012

Chapter 1 Introduction to Educational Research

MTAT Bayesian Networks. Introductory Lecture. Sven Laur University of Tartu

UW Biostatistics Working Paper Series

UNIT 4 ALGEBRA II TEMPLATE CREATED BY REGION 1 ESA UNIT 4

You must answer question 1.

A Bayesian Nonparametric Model Fit statistic of Item Response Models

Bayes Factors for t tests and one way Analysis of Variance; in R

Ordinal Data Modeling

UNLOCKING VALUE WITH DATA SCIENCE BAYES APPROACH: MAKING DATA WORK HARDER

Data and Statistics 101: Key Concepts in the Collection, Analysis, and Application of Child Welfare Data

Professor Deborah G. Mayo

HOW IS HAIR GEL QUANTIFIED?

Making Null Effects Informative: Statistical Techniques and Inferential Frameworks

Lessons in biostatistics

Response to Comment on Cognitive Science in the field: Does exercising core mathematical concepts improve school readiness?

Unit 1 Exploring and Understanding Data

How do People Really Think about Climate Change?

Computational Capacity and Statistical Inference: A Never Ending Interaction. Finbarr Sloane EHR/DRL

Individual Differences in Attention During Category Learning

Estimating Bayes Factors for Linear Models with Random Slopes. on Continuous Predictors

Checking the counterarguments confirms that publication bias contaminated studies relating social class and unethical behavior

Chapter 12: Introduction to Analysis of Variance

INTRODUC TION TO BAYESIAN STATISTICS W ILLI A M M. BOLSTA D JA MES M. CUR R A N

INTRODUCTION TO BAYESIAN REASONING

DECISION ANALYSIS WITH BAYESIAN NETWORKS

How many speakers? How many tokens?:

Generalization and Theory-Building in Software Engineering Research

INHERENT DIFFICULTIES WITH ACTIVE CONTROL EQUIVALENCE STUDIES*

ROLE OF RANDOMIZATION IN BAYESIAN ANALYSIS AN EXPOSITORY OVERVIEW by Jayanta K. Ghosh Purdue University and I.S.I. Technical Report #05-04

Statistical Hocus Pocus? Assessing the Accuracy of a Diagnostic Screening Test When You Don t Even Know Who Has the Disease

Where does "analysis" enter the experimental process?

An Exercise in Bayesian Econometric Analysis Probit and Linear Probability Models

The Fallacy of Placing Confidence in Confidence Intervals

Evaluation Models STUDIES OF DIAGNOSTIC EFFICIENCY

Russian Journal of Agricultural and Socio-Economic Sciences, 3(15)

Ecological Statistics

Chapter 12. The One- Sample

The Case for Objective Bayesian Analysis

Statistical inference provides methods for drawing conclusions about a population from sample data.

Survival Skills for Researchers. Study Design

Bayesian Tolerance Intervals for Sparse Data Margin Assessment

T-Statistic-based Up&Down Design for Dose-Finding Competes Favorably with Bayesian 4-parameter Logistic Design

Transcription:

Bayesian and Frequentist Approaches G. Jogesh Babu Penn State University http://sites.stat.psu.edu/ babu http://astrostatistics.psu.edu

All models are wrong But some are useful George E. P. Box (son-in-law of Sir Ronald Fisher) Box is well known for his contributions to time series analysis, design of experiments and Bayesian inference

Introduction The Bayesian vs. Frequentist debate is largely muted in statistics community. The debate often misses some salient features on both sides. Both the Bayesian and frequentist ideas have a lot to offer to practioners. It is always good statistical practice to analyze the data by several methods and compare. Inherently joint Bayesian-frequentist situations. Hypothesis testing and Estimation

Bayesian methods are based on Decision theoretic principle. Actions are dictated by risk management. Minimizing the expected loss. Both frequentist and Bayesian methods often lead to the same solutions, when no external information (other than the data and model itself) is to be introduced into the analysis. Frequentist methods are computationally easy. No need for numerical integration in most cases. Interpretations are different. Parts of the presentation is based on the paper by Bayarri and Berger in Statistical Science, entitled The Interplay of Bayesian and Frequentist Analysis.

Frequentist Principle In repeated practical use of a statistical procedure, the long-run average actual accuracy should not be less than (and ideally should equal) the long-run average reported accuracy. This is actually a joint frequentist-bayesian principle. Example (95% classical confidence interval for a normal mean) The text book statement focuses on fixing the value of the normal mean and imagining repeatedly drawing data from the given model and utilizing the confidence procedure repeatedly on this data. However, in practice the confidence procedure is used on a series of different problems. Hence, in evaluating the procedure, we should simultaneously be averaging over the differing means and data. Conceptually, it is the combined frequentist-bayesian average that is practically relevant.

Hypothesis testing Type I error This is the error made when a test rejects a true null hypothesis. A type I error is false positive, in signal detection. These are philosophically a focus of skepticism and Occam s razor. Type II error Error made when a test fails to reject a false null hypothesis. A type II error may be compared with a so-called false negative. Such an error occurs if we believe there is no signal when in actuality there is a signal.

Null H 0 : θ Θ 0 verses alternative H 1 : θ Θ 1 In classical testing H 0 and H 1 are not treated symmetrically while they are in Bayesian analysis. Bayes factor B in favor of Θ 0 is given by B = Posterior odds ratio Prior odds ratio = α 0π 1 α 1 π 0 In the simple vs. simple hypothesis testing it is the likelihood ratio. α i = P (Θ i x), posterior and π i are prior probabilities, i = 0, 1. Bayesian chooses H 0 if B is greater than 1. Treats both hypotheses symmetrically. The procedure does not provide error probabilities.

Rejection region of Bayesian test by decision theoretic approach under binary loss is { } K 1 C = x : P (Θ 1 x) > K 0 + K 1 Typically similar to classical test. K i are cost factors in the loss function.

P-value against H 0 P-value is the probability, when θ = θ 0 of observing an X more extreme than the actual data x. p-value is not a posterior probability of a hypothesis. p-value of 0.05 typically means that one can be pretty sure that H 0 is wrong. Any reasonably fair Bayesian analysis will show that there is at best very weak evidence against H 0. Criticism of p-values ignores the power of the test, which has quite a bit of hidden information. Bayesian analysis can be questioned because of the choice of the prior. A Bayesian has no recourse but to attempt subjective specification of the feature.

Neyman-Pearson testing theory reports the same error probability regardless of the size of the test statistic Problematic in the view of many statisticians. This led Fisher to push p-values. p-values do not have a frequentist justification. Berger, Brown and Wolpert (1994) proposed a solution for testing by defining the Type I and Type II errors conditioned on the observed values of a statistic measuring the strength of evidence in the data.

Conditional Frequentist Testing Example Suppose the data X arises from the simple hypotheses H 0 : f = f 0 vs. H 1 : f = f 1 Conditional error probabilities based on a selected statistic S = S(X) are computed as α(s) = P (Type I error S = s) P 0 (reject H 0 S(X) = s) β(s) = P (Type II error S = s) P 1 (accept H 0 S(X) = s) S and associated test utilize p-values to measure the strength of the evidence in the data. No connection with Bayesianism till this point.

Conditioning is completely allowed (and encouraged) within the frequentist paradigm. The Bayesian connection arises as it is known that α(s) = B(x) 1 + B(x) and β(s) = 1 1 + B(x) B(x) is the Bayes factor (or likelihood ratio) These expressions are precisely the Bayesian posterior probabilities of H 0 and H 1, assuming the hypotheses have equal prior probabilities of 1/2. A conditional frequentist can simply compute the objective Bayesian posterior probabilities of the hypotheses Declare that they are the conditional frequentist error probabilities. No need to formally derive the conditioning statistic or perform the conditional frequentist computations.

Bayesian Computation Gibbs sampling and other Markov chain Monte Carlo methods have become relatively standard to deal with Hierarchical, Multilevel or Mixed Models. These methods are commonly known as Bayesian computation methods. The popularity of the methods is not necessarily due to their intrinsic virtues, but rather because the Bayesian computation is now much easier than computation via more classical routes. On the other hand, any MCMC method relies fundamentally on frequentist reasoning to do the computation. Diagnostics for MCMC convergence are almost universally based on frequentist tools.

Use of frequentist methods in Prior development Example A subjective Bayesian need not worry about frequentist ideas, if a prior distribution accurately reflects prior beliefs. However, it is rare to have a (mathematical) prior distribution that accurately reflects all prior beliefs. Suppose that the only unknown model parameter is a normal mean µ. Complete assessment of the prior distribution for µ involves an infinite number of judgments. Such as specification of the probability that µ belongs to the interval (, r] for any rational number r. In practice, only a few assessments are ever made (e.g., choose prior as a Cauchy density, with median and first quartile specified). Clearly one should worry about the effect of features of the prior that were not elicited.

Consistency The simplest frequentist estimation tool that a Bayesian can usefully employ is consistency. Bayes estimates are virtually always consistent if the parameter space is finite-dimensional. This need not be true if the parameter space is not finite-dimensional or in irregular cases. Example (Neyman-Scott Problem) X ij are normally distributed with mean µ i and variance σ 2, i = 1,..., n; j = 1, 2 Until relatively recently, the most commonly used objective prior was the Jeffreys-rule prior. The resulting Bayesian estimate of σ 2 approximates to half the value, leading to an inconsistent estimate. The Jeffreys-rule prior is often inappropriate in highdimensional settings, yet it can be difficult or impossible to assess this problem within the Bayesian paradigm itself.

Estimation In the standard parametric estimation problems, objective Bayesian and frequentist methods often give similar or even identical answers. For the standard normal linear model, frequentist estimates and confidence intervals coincide exactly with the standard objective Bayesian estimates and credible intervals. More generally, this occurs in the presence of invariance structure It is still possible to achieve near-agreement between frequentist and Bayesian estimation procedure, in more complicated problems.

Conclusions Both Bayesian and frequentist methodology are here to stay. There are many areas of frequentist methodology that should be replaced by (existing) Bayesian methodology that provides superior answers. The verdict is still out on those Bayesian methodologies that have been exposed as having potentially serious frequentist problems. Philosophical unification of the Bayesian and frequentist positions is not likely, nor desirable, since each illuminates a different aspect of statistical inference. These two do different things. Analyze using both and decide. Understand the assumptions behind the theory and the mathematical conclusions. Properly interpret the results.