Uncertainty and confidence intervals Statistical estimation methods, Finse Friday 10.9.2010, 12.45–14.05 Andreas Lindén.

Slides:



Advertisements
Similar presentations
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 9 Inferences Based on Two Samples.
Advertisements

Chapter 12 Inference for Linear Regression
Lecture (11,12) Parameter Estimation of PDF and Fitting a Distribution Function.
Hypothesis Testing Steps in Hypothesis Testing:
CmpE 104 SOFTWARE STATISTICAL TOOLS & METHODS MEASURING & ESTIMATING SOFTWARE SIZE AND RESOURCE & SCHEDULE ESTIMATING.
Statistical Estimation and Sampling Distributions
Sampling Distributions (§ )
Objectives (BPS chapter 24)
Confidence Intervals Underlying model: Unknown parameter We know how to calculate point estimates E.g. regression analysis But different data would change.
Multiple regression analysis
Resampling techniques Why resampling? Jacknife Cross-validation Bootstrap Examples of application of bootstrap.
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem, random variables, pdfs 2Functions.
Maximum likelihood (ML)
Final Review Session.
2008 Chingchun 1 Bootstrap Chingchun Huang ( 黃敬群 ) Vision Lab, NCTU.
Lecture 16 – Thurs, Oct. 30 Inference for Regression (Sections ): –Hypothesis Tests and Confidence Intervals for Intercept and Slope –Confidence.
Maximum likelihood (ML)
Chapter 12 Section 1 Inference for Linear Regression.
Standard error of estimate & Confidence interval.
Bootstrapping applied to t-tests
Confidence Interval A confidence interval (or interval estimate) is a range (or an interval) of values used to estimate the true value of a population.
Review of normal distribution. Exercise Solution.
Estimation Goal: Use sample data to make predictions regarding unknown population parameters Point Estimate - Single value that is best guess of true parameter.
Inference for regression - Simple linear regression
Copyright © Cengage Learning. All rights reserved. 13 Linear Correlation and Regression Analysis.
Copyright © 2013, 2010 and 2007 Pearson Education, Inc. Chapter Inference on the Least-Squares Regression Model and Multiple Regression 14.
The Triangle of Statistical Inference: Likelihoood
1 SAMPLE MEAN and its distribution. 2 CENTRAL LIMIT THEOREM: If sufficiently large sample is taken from population with any distribution with mean  and.
Population All members of a set which have a given characteristic. Population Data Data associated with a certain population. Population Parameter A measure.
Inferences in Regression and Correlation Analysis Ayona Chatterjee Spring 2008 Math 4803/5803.
G. Cowan Lectures on Statistical Data Analysis Lecture 3 page 1 Lecture 3 1 Probability (90 min.) Definition, Bayes’ theorem, probability densities and.
1 G Lect 10a G Lecture 10a Revisited Example: Okazaki’s inferences from a survey Inferences on correlation Correlation: Power and effect.
University of Ottawa - Bio 4118 – Applied Biostatistics © Antoine Morin and Scott Findlay 08/10/ :23 PM 1 Some basic statistical concepts, statistics.
+ Chapter 12: Inference for Regression Inference for Linear Regression.
1 Advances in the Construction of Efficient Stated Choice Experimental Designs John Rose 1 Michiel Bliemer 1,2 1 The University of Sydney, Australia 2.
Statistical estimation, confidence intervals
Lecture 8 Simple Linear Regression (cont.). Section Objectives: Statistical model for linear regression Data for simple linear regression Estimation.
Inference for Regression Simple Linear Regression IPS Chapter 10.1 © 2009 W.H. Freeman and Company.
Chapter 7 Sampling and Sampling Distributions ©. Simple Random Sample simple random sample Suppose that we want to select a sample of n objects from a.
ELEC 303 – Random Signals Lecture 18 – Classical Statistical Inference, Dr. Farinaz Koushanfar ECE Dept., Rice University Nov 4, 2010.
Stat 112: Notes 2 Today’s class: Section 3.3. –Full description of simple linear regression model. –Checking the assumptions of the simple linear regression.
Limits to Statistical Theory Bootstrap analysis ESM April 2006.
Introduction to logistic regression and Generalized Linear Models July 14, 2011 Introduction to Statistical Measurement and Modeling Karen Bandeen-Roche,
Robust Estimators.
Review of Statistical Inference Prepared by Vera Tabakova, East Carolina University ECON 4550 Econometrics Memorial University of Newfoundland.
© Copyright McGraw-Hill 2004
Chapter 5 Sampling Distributions. The Concept of Sampling Distributions Parameter – numerical descriptive measure of a population. It is usually unknown.
Section 6.4 Inferences for Variances. Chi-square probability densities.
Review of statistical modeling and probability theory Alan Moses ML4bio.
Review of Statistical Inference Prepared by Vera Tabakova, East Carolina University.
Ex St 801 Statistical Methods Inference about a Single Population Mean (CI)
Parameter Estimation. Statistics Probability specified inferred Steam engine pump “prediction” “estimation”
Hypothesis Testing. Statistical Inference – dealing with parameter and model uncertainty  Confidence Intervals (credible intervals)  Hypothesis Tests.
1 STAT 500 – Statistics for Managers STAT 500 Statistics for Managers.
Learning Theory Reza Shadmehr Distribution of the ML estimates of model parameters Signal dependent noise models.
Bootstrapping James G. Anderson, Ph.D. Purdue University.
Confidence Intervals. Point Estimate u A specific numerical value estimate of a parameter. u The best point estimate for the population mean is the sample.
Statistical principles: the normal distribution and methods of testing Or, “Explaining the arrangement of things”
Statistics and probability Dr. Khaled Ismael Almghari Phone No:
Canadian Bioinformatics Workshops
Estimating standard error using bootstrap
Inference for Regression (Chapter 14) A.P. Stats Review Topic #3
STAT 312 Chapter 7 - Statistical Intervals Based on a Single Sample
CHAPTER 12 More About Regression
Estimation Goal: Use sample data to make predictions regarding unknown population parameters Point Estimate - Single value that is best guess of true parameter.
CHAPTER 12 More About Regression
CHAPTER 12 More About Regression
Sampling Distributions (§ )
Chapter 6 Confidence Intervals.
Introductory Statistics
Presentation transcript:

Uncertainty and confidence intervals Statistical estimation methods, Finse Friday , 12.45–14.05 Andreas Lindén

Outline Point estimates and uncertainty Sampling distribution –Standard error –Covariation between parameters Finding the VC-matrix for the parameter estimates –Analytical formulas –From the Hessian matrix –Bootstrapping The idea behind confidence intervals General methods for constructing confidence intervals of parameters –CI based on the central limit theorem –Profile likelihood CI –CI by bootstrapping

3 Point estimates and uncertainty The main output in any statistical model fitting are the parameter estimates –Point estimates — one value for each parameter –The effect sizes –Answers the question “how much” Point estimates are of little use without any assessment of uncertainty –Standard error –Confidence intervals –p-values –Estimated sampling distribution –Bayesian credible intervals –Plotting Bayesian posterior distribution

4 Sampling distribution The probability distribution of a parameter estimate –Calculated from a sample –Variability due to sampling effects Typically depends on sample size or the number of degrees of freedom (df) Examples of common sampling distributions –Student’s t-distribution –F-distribution –χ²-distribution

5 Degrees of freedom Y X In a linear regression df = n – 2

6 Properties of the sampling distribution The standard error (SE) of a parameter, is the estimated standard deviation of the sampling distribution –Square root of parameter variance Parameters are not necessarily unrelated –The sampling distribution of several parameters is multivariate –Example: regression slope and intercept

7 Linear regression – simulated data Param.abσ² True value Estim Estim Estim Estim Estim Estim Estim Estim Estim Estim ……… … Estim

8 Properties of the sampling distribution The standard error (SE) of a parameter, is the estimated standard deviation of the sampling distribution –Square root of parameter variance Parameters are not necessarily unrelated –The sampling distribution of several parameters is multivariate –Example: regression slope and intercept COV = CORR =

9 Properties of the sampling distribution The standard error (SE) of a parameter, is the estimated standard deviation of the sampling distribution –Square root of parameter variance Parameters are not necessarily unrelated –The sampling distribution of several parameters is multivariate –Example: regression slope and intercept Methods to obtain the VC-matrix (or standard errors) for a set of parameters –Analytical formulas –Bootstrap –The inverse of the Hessian matrix

10 Parameter variances analytically For many common situations the SE and VC-matrix of a set of parameters can be calculated with analytical formulas Standard error of the sample mean Standard error of the estimated binomial probability

11 Bootstrap The bootstrap is a general and common resampling method Used to simulate the sampling distribution Information in the sample itself is used to mimic the original sampling procedure –Non-parametric bootstrap — sampling with replacement –Parametric bootstrap — simulation based on parameter estimates The procedure is repeated B times (e.g. B = 1000) To make inference from the bootstrapped estimates –Sample standard deviation = bootstrap estimate of SE –Sample VC-matrix = bootstrap estimate of VC-matrix –Mean = difference between bootstrap mean and original estimate is an estimate of bias

12 VC-matrix from the Hessian The Hessian matrix (H) –2 nd derivative of the (multivariate) negative log-likelihood at the ML-estimate –Typically given as an output by software for numerical optimization The inverse of the Hessian is an estimate of the parameters’ variance-covariance matrix

13 Confidence interval (CI) An frequentistic interval estimate of one or several parameters A fraction α of all correctly produced CI:s will fail to include the true parameter value –Trust your 95% CI and take the risk α = 0.05 NB! Should not be confused with Bayesian credible intervals –CI:s should not be thought to contain the parameter with 95% probability –The CI is based on the sampling distribution, not on an estimated probability distribution for the parameter of interest

14

15 CI based on central limit theorem The sum/mean of many random values are approximately normally distributed –Actually t-distributed with df depending on sample size and model complexity –Might matter with small sample size As a rule of thumb, an arbitrary parameter estimate ± 2*SE produce an approximate 95% confidence interval –With infinitely many observations ± 1.96*SE

16 CI from profile likelihood The profile deviance –The change in −2*log-likelihood, in comparison to the ML- estimate –Asymptotically χ²-distributed (assuming infinite sample size) Confidence intervals can be obtained as the range around the ML-estimate, for which the profile deviance is under a critical level –The 1 – α quantile from χ²-distribution –One-parameter -> df = 1 (e.g for α = 0.05) –k-dimensional profile deviance -> df = k

17 95% CI from profile deviance –2*LL Parameter value F min F min

18 2-D confidence regions Parameter a Parameter b 99% confidence region, deviance χ² df2 = % confidence region, deviance χ² df2 = 5.992

19 CI by bootstrapping A 100*(1 – α)% CI for a parameter can be calculated from the sampling distribution –The α / 2 and 1 – α /2 quantiles (e.g and with α = 0.05) In bootstrapping, simply use the sample quantiles of simulated values

Exercises Data: The prevalence of an infectious disease in a human population is investigated. The infection is recorded with 100% detection efficiency. In a sample of N = 80 humans X = 18 infections were found. Model: Assume that infection (x = 0 or 1) of a host individual is an independent Bernoulli trial with probability p i, such that the probability of infection is constant over all hosts. (This equals a logistic regression with an intercept only. Host specific explanatory variables, such as age, condition, etc. could be used to improve the model of p i closer.)

Do the following in R: a)Calculate and plot the profile (log) likelihood of infection probability p b)What is the maximum likelihood estimate of p (called p̂ )? c)Construct 95% and 99% confidence intervals for p̂ based on the profile likelihood d)Calculate the analytic SE for p̂ e)Construct symmetric 95% confidence interval for p̂ based on the central limit theorem and the SE obtained in previous exercise f)Simulate and plot the sampling distribution of p̂ by parametric bootstrapping (B = 10000) g)Calculate the bootstrap SE of p̂ h)Construct 95% confidence interval for p̂ based on the bootstrap