Chapter 11: Advanced Remedial Measures. Weighted Least Squares (WLS)

Similar documents
Analysis of Rheumatoid Arthritis Data using Logistic Regression and Penalized Approach

Applying Machine Learning Methods in Medical Research Studies

2.75: 84% 2.5: 80% 2.25: 78% 2: 74% 1.75: 70% 1.5: 66% 1.25: 64% 1.0: 60% 0.5: 50% 0.25: 25% 0: 0%

Comparison of Adaptive and M Estimation in Linear Regression

Performance of Median and Least Squares Regression for Slightly Skewed Data

Anale. Seria Informatică. Vol. XVI fasc Annals. Computer Science Series. 16 th Tome 1 st Fasc. 2018

Citation for published version (APA): Ebbes, P. (2004). Latent instrumental variables: a new approach to solve for endogeneity s.n.

MODEL SELECTION STRATEGIES. Tony Panzarella

Chapter 2 Organizing and Summarizing Data. Chapter 3 Numerically Summarizing Data. Chapter 4 Describing the Relation between Two Variables

WELCOME! Lecture 11 Thommy Perlinger

Introduction to Econometrics

A Comparative Study of Some Estimation Methods for Multicollinear Data

Study Guide #2: MULTIPLE REGRESSION in education

VARIABLE SELECTION WHEN CONFRONTED WITH MISSING DATA

Chapter 1: Exploring Data

Statistics as a Tool. A set of tools for collecting, organizing, presenting and analyzing numerical facts or observations.

What is Regularization? Example by Sean Owen

CHILD HEALTH AND DEVELOPMENT STUDY

Modern Regression Methods

UNIVERSITY of PENNSYLVANIA CIS 520: Machine Learning Final, Fall 2014

CLASSICAL AND. MODERN REGRESSION WITH APPLICATIONS

A Comparison of Robust and Nonparametric Estimators Under the Simple Linear Regression Model

WDHS Curriculum Map Probability and Statistics. What is Statistics and how does it relate to you?

Bootstrapping Residuals to Estimate the Standard Error of Simple Linear Regression Coefficients

Section on Survey Research Methods JSM 2009

Multiple Regression Analysis

Study of cigarette sales in the United States Ge Cheng1, a,

STATISTICS & PROBABILITY

Results & Statistics: Description and Correlation. I. Scales of Measurement A Review

Computer Age Statistical Inference. Algorithms, Evidence, and Data Science. BRADLEY EFRON Stanford University, California

Limited dependent variable regression models

Correlation and regression

Identifying Susceptibility in Epidemiology Studies: Implications for Risk Assessment. Joel Schwartz Harvard TH Chan School of Public Health

11/18/2013. Correlational Research. Correlational Designs. Why Use a Correlational Design? CORRELATIONAL RESEARCH STUDIES

Unit 1 Exploring and Understanding Data

Linear Regression Analysis

Understandable Statistics

Staff Papers Series. Department of Agricultural and Applied Economics

Identification of Tissue Independent Cancer Driver Genes

Statistical Techniques. Meta-Stat provides a wealth of statistical tools to help you examine your data. Overview

Simple Linear Regression the model, estimation and testing

IAPT: Regression. Regression analyses

Examining Relationships Least-squares regression. Sections 2.3

MULTIPLE REGRESSION OF CPS DATA

This exam consists of three parts. Provide answers to ALL THREE sections.

M15_BERE8380_12_SE_C15.6.qxd 2/21/11 8:21 PM Page Influence Analysis 1

6. Unusual and Influential Data

Research Methods in Forest Sciences: Learning Diary. Yoko Lu December Research process

Selection and Combination of Markers for Prediction

BIOINFORMATICS ORIGINAL PAPER

Machine Learning to Inform Breast Cancer Post-Recovery Surveillance

Simple Sensitivity Analyses for Matched Samples Thomas E. Love, Ph.D. ASA Course Atlanta Georgia

CRITERIA FOR USE. A GRAPHICAL EXPLANATION OF BI-VARIATE (2 VARIABLE) REGRESSION ANALYSISSys

Validation of consistency of Mendelian sampling variance in national evaluation models

Lecture 12: more Chapter 5, Section 3 Relationships between Two Quantitative Variables; Regression

SW 9300 Applied Regression Analysis and Generalized Linear Models 3 Credits. Master Syllabus

Chapter 2--Norms and Basic Statistics for Testing

Article from. Forecasting and Futurism. Month Year July 2015 Issue Number 11

UNIVERSITY of PENNSYLVANIA CIS 520: Machine Learning Midterm, 2016

Still important ideas

Pitfalls in Linear Regression Analysis

Chapter 11 Multiple Regression

Section 3.2 Least-Squares Regression

Econometric Game 2012: infants birthweight?

Chapter 3 CORRELATION AND REGRESSION

(CORRELATIONAL DESIGN AND COMPARATIVE DESIGN)

Multivariable Systems. Lawrence Hubert. July 31, 2011

Final Exam - section 2. Thursday, December hours, 30 minutes

Sawtooth Software. MaxDiff Analysis: Simple Counting, Individual-Level Logit, and HB RESEARCH PAPER SERIES. Bryan Orme, Sawtooth Software, Inc.

The impact of pre-selected variance inflation factor thresholds on the stability and predictive power of logistic regression models in credit scoring

Score Tests of Normality in Bivariate Probit Models

Describe what is meant by a placebo Contrast the double-blind procedure with the single-blind procedure Review the structure for organizing a memo

Business Statistics Probability

THE EFFECT OF WEIGHT TRIMMING ON NONLINEAR SURVEY ESTIMATES

Catherine A. Welch 1*, Séverine Sabia 1,2, Eric Brunner 1, Mika Kivimäki 1 and Martin J. Shipley 1

Psychology Research Process

Meta-Analysis and Publication Bias: How Well Does the FAT-PET-PEESE Procedure Work?

Chapter 3: Examining Relationships

Making Inferences from Experiments

Stepwise method Modern Model Selection Methods Quantile-Quantile plot and tests for normality

Lecture Outline. Biost 517 Applied Biostatistics I. Purpose of Descriptive Statistics. Purpose of Descriptive Statistics

The Non-Response in the Population Ratio of Mean for Current Occasion in Sampling on Two Occasions. Amelia Victoria Garcia Luengo 1

Bringing machine learning to the point of care to inform suicide prevention

RISK PREDICTION MODEL: PENALIZED REGRESSIONS

Ridge Regression and Multicollinearity: An In-Depth Review

Lecture 6B: more Chapter 5, Section 3 Relationships between Two Quantitative Variables; Regression

Write your identification number on each paper and cover sheet (the number stated in the upper right hand corner on your exam cover).

10.1 Estimating with Confidence. Chapter 10 Introduction to Inference

11/24/2017. Do not imply a cause-and-effect relationship

Regression Discontinuity Analysis

Readings: Textbook readings: OpenStax - Chapters 1 13 (emphasis on Chapter 12) Online readings: Appendix D, E & F

In this module I provide a few illustrations of options within lavaan for handling various situations.

Combining machine learning and matching techniques to improve causal inference in program evaluation

Chapter 2. Behavioral Variability and Research

FORM C Dr. Sanocki, PSY 3204 EXAM 1 NAME

Russian Journal of Agricultural and Socio-Economic Sciences, 3(15)

CHAPTER TWO REGRESSION

Ordinary Least Squares Regression

Treatment effect estimates adjusted for small-study effects via a limit meta-analysis

Transcription:

Chapter : Advanced Remedial Measures Weighted Least Squares (WLS) When the error variance appears nonconstant, a transformation (of Y and/or X) is a quick remedy. But it may not solve the problem, or it may create an inappropriate regression relationship. A more advanced approach is WLS regression. 2 If var( Yi ) i, then give observations with higher variance weight in the regression fitting. For example, let and use But are typically unknown. 2 2,, n Note 2 i Thus = To estimate how i varies a function of Xi (or of Yˆ i ), regress

Procedure for Determining Weights wi: () Regress Y against predictor variable(s) as usual (OLS). (2) Regress absolute residuals ei against predictor Xj (if error variance is nonconstant as a function of Xj) or against fitted (if error variance is nonconstant as a function of Yˆ ). valuesyˆ (3) (4) (5) SAS or R will do WLS once we find the weights wi. SAS Example (Blood pressure data): Note: R 2 does not have a standard interpretation with WLS. Note: Standard error of b decreases somewhat in the WLS regression Note: If the WLS estimates differ greatly from the OLS estimates, we may iterate this algorithm one or two times. Note: In WLS, standard inferences about coefficients may not be valid for small sample sizes, when the weights are estimated from data. Note: If the MSE of the WLS regression is near, then our estimation of the error standard deviation function is trustworthy.

Ridge Regression and LASSO Regression Ridge regression is an advanced remedy for multicollinearity. Idea: Instead of using our unbiased ordinary estimate, use a biased estimate, denoted Although br is biased, it may have less variance (so that multicollinearity is reduced). Procedure [typically these calculations are done on standardized (centered and scaled) regression coefficients]: Add a biasing constant c into the normal equations (see pg. 433). If c = 0, then As c increases: R provides automated choices for c, and it will perform ridge regression. R example (body fat data):

Disadvantage to ridge regression: We cannot use ordinary inference procedures (bootstrapping can be used for inference) Ridge regression is an example of shrinkage estimation: The process will typically shrink the least-squares estimates toward zero because of the biasing constant. This shrinkage increases bias, but reduces variance. Ridge regression estimates may be obtained by minimizing the penalized least-squares criterion: The solution to this is the vector br that minimizes The LASSO is a similar method which chooses b to minimize An advantage of LASSO regression (and ridge regression, to some degree) is that this constraint leads to some bj s being set very close to zero, so LASSO can be viewed as a method of variable selection as well as coefficient estimation.

Traditionally, ridge regression estimates have been easier to obtain computationally than the LASSO estimates. In 2000, an efficient algorithm was developed to solve for the LASSO estimates, making LASSO regression very popular. R example with LASSO regression (body fat data): Robust Regression If we have highly influential observations, we can reduce their impact on the regression equation (without discarding them entirely) using robust regression methods. Similarly, robust regression is effective when the error distribution is not normal, but rather heavy-tailed. M-estimation is a general class of estimation methods. We choose

Note: () If p(u) = u 2, then this is (2) If p(u) = u, then the criterion is This method is called Least Absolute Residuals (LAR) regression, also called L regression. It uses absolute residuals rather than squared residuals, so the effect of outliers is not as great. Note: Residuals from LAR regression might not sum to zero. (3) Huber s method uses a p(. ) function that is a compromise between least-squares and LAR regression: R example (math proficiency data): Note: Inference on the regression coefficients is more complex for robust regression. For large samples, the robust estimators are approximately normal, so we can perform approximate CIs and tests about the coefficients.

Inference Using the Bootstrap Method We have seen how to perform inference (CIs, tests) with the general linear model with normal errors. Bootstrapping is a general method of inference that can often be used in nonstandard situations when our usual inferential methods are not valid. Examples: We can evaluate the precision of estimates such as estimated coefficients and fitted values in: Weighted least squares Ridge and LASSO regression Robust regression General Procedure: () We select a random sample (of size n), with replacement, from the observations in the original sample. This is called a bootstrap sample. This bootstrap sample will likely contain some duplicate values from the original data, and some original data will be omitted in the bootstrap sample. (2) We perform the original regression procedure on the [] b bootstrap sample, and obtain the estimate of interest, say,. (3) We repeat the sampling with replacement a large number (say, B) of times, and for each new bootstrap sample, we obtain the estimate of interest, so that we have a collection of [] [ B] bootstrap estimates, e.g., b,, b. (4) The estimated standard deviation of these bootstrap [] [ B] estimates b,, b is an estimate of the standard error of the original estimator b itself.

Two Types of Bootstrap Sampling in Regression Fixed X resampling which is used when: () (2) AND (3) With fixed X resampling, we fit the original regression and sample the residuals e,, en, with replacement, to obtain the bootstrap sample of n residuals e,, en. Then the bootstrap sample of response values is Then we regress the Y i obtain the bootstrap estimate, say, values against the original Xi values to b This is done B times, so that we obtain. b, b. [], Random X resampling which is used when: () (2) OR (3) [ B] With random X resampling, we sample the data pairs (Xi, Yi) with replacement, so that we obtain a bootstrap sample of n data pairs (Xi, Yi). Then we regress the Yi values against the X i values to obtain the bootstrap estimate, say, b. [] [ B] This is done B times, so that we obtain b, b.,

Bootstrap Confidence Intervals Bootstrap CIs are based on the empirical distribution of b. The percentile method to obtain a 00( )% bootstrap CI for, say, is to use the interval (L, U), where The reflection method to obtain a 00( )% bootstrap CI for, say, is The above methods tend to produce similar results. It is recommended to let B be at least 500 when constructing bootstrap CIs (often 000 resamples are used). Examples in R (Toluca data and blood pressure data):