#22 Uncertainty of Derived Parameters

Slides:



Advertisements
Similar presentations
Welcome to PHYS 225a Lab Introduction, class rules, error analysis Julia Velkovska.
Advertisements

3. Binary Choice – Inference. Hypothesis Testing in Binary Choice Models.
Bayesian Estimation in MARK
Sampling Distributions (§ )
Markov-Chain Monte Carlo
Computing the Posterior Probability The posterior probability distribution contains the complete information concerning the parameters, but need often.
The University of Texas at Austin, CS 395T, Spring 2008, Prof. William H. Press IMPRS Summer School 2009, Prof. William H. Press 1 4th IMPRS Astronomy.
1 Analysis of Variance This technique is designed to test the null hypothesis that three or more group means are equal.
The University of Texas at Austin, CS 395T, Spring 2008, Prof. William H. Press IMPRS Summer School 2009, Prof. William H. Press 1 4th IMPRS Astronomy.
Copyright © 2005 Department of Computer Science CPSC 641 Winter Data Analysis and Presentation There are many “tricks of the trade” used in data.
Lesson #17 Sampling Distributions. The mean of a sampling distribution is called the expected value of the statistic. The standard deviation of a sampling.
G. Cowan RHUL Physics Bayesian Higgs combination page 1 Bayesian Higgs combination using shapes ATLAS Statistics Meeting CERN, 19 December, 2007 Glen Cowan.
Physics 114: Lecture 15 Probability Tests & Linear Fitting Dale E. Gary NJIT Physics Department.
Using car4ams, the Bayesian AMS data-analysis code V. Palonen, P. Tikkanen, and J. Keinonen Department of Physics, Division of Materials Physics.
Combined Uncertainty P M V Subbarao Professor Mechanical Engineering Department A Model for Propagation of Uncertainty ….
R. Kass/W03P416/Lecture 7 1 Lecture 7 Some Advanced Topics using Propagation of Errors and Least Squares Fitting Error on the mean (review from Lecture.
Chanyoung Park Raphael T. Haftka Paper Helicopter Project.
Variance Harry R. Erwin, PhD School of Computing and Technology University of Sunderland.
Part 2: Model and Inference 2-1/49 Regression Models Professor William Greene Stern School of Business IOMS Department Department of Economics.
University of Colorado Boulder ASEN 5070: Statistical Orbit Determination I Fall 2014 Professor Brandon A. Jones Lecture 14: Probability Wrap-Up and Statistical.
Data Modeling Patrice Koehl Department of Biological Sciences National University of Singapore
The University of Texas at Austin, CS 395T, Spring 2008, Prof. William H. Press 1 Computational Statistics with Application to Bioinformatics Prof. William.
The University of Texas at Austin, CS 395T, Spring 2008, Prof. William H. Press 1 Computational Statistics with Application to Bioinformatics Prof. William.
NON-LINEAR REGRESSION Introduction Section 0 Lecture 1 Slide 1 Lecture 6 Slide 1 INTRODUCTION TO Modern Physics PHYX 2710 Fall 2004 Intermediate 3870 Fall.
G. Cowan Lectures on Statistical Data Analysis Lecture 4 page 1 Statistical Data Analysis: Lecture 4 1Probability, Bayes’ theorem 2Random variables and.
University of Colorado Boulder ASEN 5070: Statistical Orbit Determination I Fall 2015 Professor Brandon A. Jones Lecture 14: Probability and Statistics.
Professor William H. Press, Department of Computer Science, the University of Texas at Austin1 Opinionated in Statistics by Bill Press Lessons #31 A Tale.
Review of statistical modeling and probability theory Alan Moses ML4bio.
Professor William H. Press, Department of Computer Science, the University of Texas at Austin1 Opinionated in Statistics by Bill Press Lessons #46 Interpolation.
Density Estimation in R Ha Le and Nikolaos Sarafianos COSC 7362 – Advanced Machine Learning Professor: Dr. Christoph F. Eick 1.
Estimating standard error using bootstrap
Data Modeling Patrice Koehl Department of Biological Sciences
Applied statistics Usman Roshan.
The simple linear regression model and parameter estimation
Statistical Methods Michael J. Watts
Physics 114: Lecture 13 Probability Tests & Linear Fitting
Statistical Methods Michael J. Watts
REGRESSION (CONTINUED)
Ch3: Model Building through Regression
Introduction, class rules, error analysis Julia Velkovska
CH 5: Multivariate Methods
Analyzing Redistribution Matrix with Wavelet
Maximum Likelihood Estimation
Statistical Methods For Engineers
Predictive distributions
Modern Spectral Estimation
CPSC 531: System Modeling and Simulation
Data Presentation Carey Williamson Department of Computer Science
Sampling Distribution
Sampling Distribution
Multidimensional Integration Part I
Filtering and State Estimation: Basic Concepts
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Today (2/23/16) Learning objectives:
Nonlinear regression.
Regression in the 21st Century
PSY 626: Bayesian Statistics for Psychological Science
Principles of the Global Positioning System Lecture 11
Carey Williamson Department of Computer Science University of Calgary
#21 Marginalize vs. Condition Uninteresting Fitted Parameters
Decomposition of Stat/Sys Errors
Opinionated Lessons #39 MCMC and Gibbs Sampling in Statistics
#23 Bootstrap Estimation of Uncertainty
#10 The Central Limit Theorem
Sampling Distributions (§ )
#8 Some Standard Distributions
Opinionated Lessons #13 The Yeast Genome in Statistics by Bill Press
Hairong Qi, Gonzalez Family Professor
Uncertainty Propagation
Classical regression review
Presentation transcript:

#22 Uncertainty of Derived Parameters Opinionated Lessons in Statistics by Bill Press #22 Uncertainty of Derived Parameters Professor William H. Press, Department of Computer Science, the University of Texas at Austin

b ´ ¡ r f b f ´ ( b ) = + r ¢ h f i ¼ ( b ) + r = ­ f ® ¡ h i ¼ ( b ) What is the uncertainty in quantities other than the fitted coefficients: Method 1: Linearized propagation of errors nerdy math note: is technically a row (not column) vector, because it’s a one-form r f is the MLE parameters estimate b is the RV as the parameters fluctuate b 1 ´ ¡ f ´ ( b ) = + r 1 ¢ h f i ¼ ( b ) + r 1 = ­ f 2 ® ¡ h i ¼ ( b ) r 1 + = T § Professor William H. Press, Department of Computer Science, the University of Texas at Austin

f = b b = : 9 8 § 1 r f = ( ; b ) r f § = b + : 6 p : 3 6 = 1 8 In our example, if we are interested in the area of the “hump”, bfit = 1.1235 1.5210 0.6582 3.2654 1.4832 covar = 0.1349 0.2224 0.0068 -0.0309 0.0135 0.2224 0.6918 0.0052 -0.1598 0.1585 0.0068 0.0052 0.0049 0.0016 -0.0094 -0.0309 -0.1598 0.0016 0.0746 -0.0444 0.0135 0.1585 -0.0094 -0.0444 0.0948 f = b 3 5 r f = ( ; b 5 3 ) r f § T = b 2 5 3 + : 6 p : 3 6 = 1 8 the one standard deviation (1-s) error bar So b 3 5 = : 9 8 § 1 Is it normally distributed? Absolutely not! A function of normals is not normal (although, if they are all narrow, it might be close). Professor William H. Press, Department of Computer Science, the University of Texas at Austin

Method 2: Sample from the posterior distribution What is the uncertainty in quantities other than the fitted coefficients: Method 2: Sample from the posterior distribution 1. Generate a large number of (vector) b’s 2. Compute your separately for each b 3. Histogram Note again that b is typically (close to) m.v. normal because of the CLT, but your (nonlinear) f may not, in general, be anything even close to normal! Professor William H. Press, Department of Computer Science, the University of Texas at Austin

b Our example: bees = mvnrnd(bfit,covar,10000); humps = bees(:,3).*bees(:,5); hist(humps,30); std(humps) std = 0.1833 b 3 5 Did I really need to use the full covar, not just the 2x2 piece for parameters 3 and 5? Professor William H. Press, Department of Computer Science, the University of Texas at Austin

Compare linear propagation of errors to sampling the posterior Note that even with lots of data, so that the distribution of the b’s really  multivariate normal, a derived quantity might be very non-Normal. In this case, sampling the posterior is a good idea! For example, the ratio of two normals of zero mean is Cauchy which is very non-Normal!. So, sampling the posterior is a more powerful method than linear propagation of errors. even when optimistically (or in ignorance) assuming multivariate Gaussian for the fitted parameters In fact, sampling the posterior distribution of large Bayesian models whose parameters are not at all Gaussian is, under the name MCMC, the most powerful technique in modern computational statistics. we’ll come back to this! Professor William H. Press, Department of Computer Science, the University of Texas at Austin