Uncertainty in fall time surrogate Prediction variance vs. data sensitivity – Non-uniform noise – Example 3.2.1 Uncertainty in fall time data Bootstrapping.

Slides:



Advertisements
Similar presentations
Kin 304 Regression Linear Regression Least Sum of Squares
Advertisements

Kriging.
Cost of surrogates In linear regression, the process of fitting involves solving a set of linear equations once. For moving least squares, we need to.
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
Simple Linear Regression. G. Baker, Department of Statistics University of South Carolina; Slide 2 Relationship Between Two Quantitative Variables If.
Ch11 Curve Fitting Dr. Deshi Ye
Simple Linear Regression and Correlation
Simple Linear Regression
Chapter 10: Sampling and Sampling Distributions
© 2010 Pearson Prentice Hall. All rights reserved Least Squares Regression Models.
Resampling techniques Why resampling? Jacknife Cross-validation Bootstrap Examples of application of bootstrap.
Chapter 10 Simple Regression.
Accuracy of Prediction How accurate are predictions based on a correlation?
Chapter 11 Multiple Regression.
Lecture 16 – Thurs, Oct. 30 Inference for Regression (Sections ): –Hypothesis Tests and Confidence Intervals for Intercept and Slope –Confidence.
Simple Linear Regression Analysis
Quantitative Business Analysis for Decision Making Simple Linear Regression.
REGRESSION Predict future scores on Y based on measured scores on X Predictions are based on a correlation from a sample where both X and Y were measured.
1 BA 555 Practical Business Analysis Review of Statistics Confidence Interval Estimation Hypothesis Testing Linear Regression Analysis Introduction Case.
Business Statistics - QBM117 Statistical inference for regression.
Correlation and Regression Analysis
Simple Linear Regression and Correlation
Simple Linear Regression Analysis
Normal and Sampling Distributions A normal distribution is uniquely determined by its mean, , and variance,  2 The random variable Z = (X-  /  is.
Statistical Analysis. Purpose of Statistical Analysis Determines whether the results found in an experiment are meaningful. Answers the question: –Does.
Lecture 5 Correlation and Regression
Copyright © 2013, 2010 and 2007 Pearson Education, Inc. Chapter Inference on the Least-Squares Regression Model and Multiple Regression 14.
(a.k.a: The statistical bare minimum I should take along from STAT 101)
BPS - 3rd Ed. Chapter 211 Inference for Regression.
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Chapter 15 Multiple Regression n Multiple Regression Model n Least Squares Method n Multiple.
© 2011 Cengage Learning. All Rights Reserved. May not be copied, scanned, or duplicated, in whole or in part, except for use as permitted in a license.
Correlation and Regression Chapter 9. § 9.3 Measures of Regression and Prediction Intervals.
Statistics 1: tests and linear models. How to get started? Exploring data graphically: Scatterplot HistogramBoxplot.
Multiple Regression and Model Building Chapter 15 Copyright © 2014 by The McGraw-Hill Companies, Inc. All rights reserved.McGraw-Hill/Irwin.
Review of Building Multiple Regression Models Generalization of univariate linear regression models. One unit of data with a value of dependent variable.
Lesson Multiple Regression Models. Objectives Obtain the correlation matrix Use technology to find a multiple regression equation Interpret the.
Section 5.4 Sampling Distributions and the Central Limit Theorem Larson/Farber 4th ed.
Essential Question:  How do scientists use statistical analyses to draw meaningful conclusions from experimental results?
Robust Estimators.
Review - Confidence Interval Most variables used in social science research (e.g., age, officer cynicism) are normally distributed, meaning that their.
Linear Prediction Correlation can be used to make predictions – Values on X can be used to predict values on Y – Stronger relationships between X and Y.
Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Simple Linear Regression Analysis Chapter 13.
Review of Statistical Terms Population Sample Parameter Statistic.
1 Simple Linear Regression and Correlation Least Squares Method The Model Estimating the Coefficients EXAMPLE 1: USED CAR SALES.
Tutorial I: Missing Value Analysis
Intro to Statistics for the Behavioral Sciences PSYC 1900 Lecture 7: Regression.
Global predictors of regression fidelity A single number to characterize the overall quality of the surrogate. Equivalence measures –Coefficient of multiple.
Bootstrapping James G. Anderson, Ph.D. Purdue University.
BPS - 5th Ed. Chapter 231 Inference for Regression.
Bias-Variance Analysis in Regression  True function is y = f(x) +  where  is normally distributed with zero mean and standard deviation .  Given a.
Global predictors of regression fidelity A single number to characterize the overall quality of the surrogate. Equivalence measures –Coefficient of multiple.
Global predictors of regression fidelity A single number to characterize the overall quality of the surrogate. Equivalence measures –Coefficient of multiple.
Estimating standard error using bootstrap
Questions from lectures
Chapter 14 Inference on the Least-Squares Regression Model and Multiple Regression.
Multiple Imputation using SOLAS for Missing Data Analysis
Statistical Quality Control, 7th Edition by Douglas C. Montgomery.
B&A ; and REGRESSION - ANCOVA B&A ; and
Kin 304 Regression Linear Regression Least Sum of Squares
BPK 304W Regression Linear Regression Least Sum of Squares
Quantitative Methods Simple Regression.
BPK 304W Correlation.
S519: Evaluation of Information Systems
Lecture Slides Elementary Statistics Thirteenth Edition
REGRESSION.
Multiple Regression Models
Bootstrapping Jackknifing
Simple Linear Regression
Sample vs Population (true mean) (sample mean) (sample variance)
Bootstrapping and Bootstrapping Regression Models
Presentation transcript:

Uncertainty in fall time surrogate Prediction variance vs. data sensitivity – Non-uniform noise – Example Uncertainty in fall time data Bootstrapping – Estimating accuracy of statistics

Linear Regression Functional form For linear approximation Define then Regression coefficients Altogether Differentiate with respect to ith component of y.

Example Given data Linear fit X-2012 Y

Prediction variance with variable noise Prediction variance based on assumptions on noise Variance of surrogate prediction Allows different variances. Note that with different variances better to use weighted least squares.

Comparison for example at x=3 Prediction variance (surprisingly small, why?) Variance of prediction If all data variances are the same, check you get the same If not, variance of y 5 is most important

Bootstrapping When we calculate statistics from random data bootstrapping can provide error estimates. If we had multiple samples we can use them to estimate the error in the computation. With bootstrapping we perform the amazing feat of getting the error from a single sample. This is done by resampling with replacement the same data. We draw a samples from the original data without removing it so that the new sample may have repetitions. We repeat for many bootstrap samples to get a distribution of the statistic of interest.

Example with sample mean x=randn(1,10) x = bootsam(:,1:5) ans = Each column contains the indices of one boot strap sample. For example, the last column indicates that we drew x(2)= four times, x(6) twice, along with x(3), x(5), x(9), and x(10). What is the probability of getting no repetitions?

Matlab bootstrp routine bootstat = bootstrp(nboot,bootfun,d1,...) draws nboot bootstrap data samples, computes statistics on each sample using bootfun, and returns the results in the matrix bootstat. bootfun is a function handle specified Each row of bootstat contains the results of applying bootfun to one bootstrap sample. [bootstat,bootsam] = bootstrp(...) returns an n-by- nboot matrix of bootstrap indices, bootsam. Each column in bootsam contains indices of the values that were drawn from the original data sets to constitute the corresponding bootstrap sample

Statistics for sample mean mean(x) = mean(bootstat)= std(x) = std(bootstat)= In this case we know that the standard deviation of the mean is the native standard deviation divided by the square root of the sample size, or about 0.56 In other cases we may not have a formula. May use bootstrapping to estimate accuracy of probability

Sample standard deviation mean(bootstat)= std(bootstat)= Check ratio a=randn(10,10000); s=std(a); mean(s) = std(s)= Bootstrap ratio is 0.208, actual ratio 0.237

Exercise The variables x and y are normally distributed with N(0,1) marginal distributions and a correlation coefficient of Generate a sample of 10 pairs and use bootstrap to estimate the accuracy of the correlation coefficient you obtain from the sample. 2.Compare to the accuracy you can get from a formula or by repeating step 1 many times.