- 1 - Preliminaries Multivariate normal model (section 3.6, Gelman) –For a multi-parameter vector y, multivariate normal distribution is where  is covariance.

Slides:



Advertisements
Similar presentations
A. The Basic Principle We consider the multivariate extension of multiple linear regression – modeling the relationship between m responses Y 1,…,Y m and.
Advertisements

Regression and correlation methods
Kin 304 Regression Linear Regression Least Sum of Squares
Copula Regression By Rahul A. Parsa Drake University &
Chapter 12 Simple Linear Regression
Bayesian inference of normal distribution
Kriging.
Cost of surrogates In linear regression, the process of fitting involves solving a set of linear equations once. For moving least squares, we need to form.
Irwin/McGraw-Hill © Andrew F. Siegel, 1997 and l Chapter 12 l Multiple Regression: Predicting One Factor from Several Others.
Cost of surrogates In linear regression, the process of fitting involves solving a set of linear equations once. For moving least squares, we need to.
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
Ch11 Curve Fitting Dr. Deshi Ye
1 Chapter 2 Simple Linear Regression Ray-Bing Chen Institute of Statistics National University of Kaohsiung.
The Multiple Regression Model Prepared by Vera Tabakova, East Carolina University.
The Simple Linear Regression Model: Specification and Estimation
1-1 Regression Models  Population Deterministic Regression Model Y i =  0 +  1 X i u Y i only depends on the value of X i and no other factor can affect.
Chapter 10 Simple Regression.
CHAPTER 3 ECONOMETRICS x x x x x Chapter 2: Estimating the parameters of a linear regression model. Y i = b 1 + b 2 X i + e i Using OLS Chapter 3: Testing.
1 Chapter 3 Multiple Linear Regression Ray-Bing Chen Institute of Statistics National University of Kaohsiung.
Probability theory 2011 The multivariate normal distribution  Characterizing properties of the univariate normal distribution  Different definitions.
Petter Mostad Linear regression Petter Mostad
Using ranking and DCE data to value health states on the QALY scale using conventional and Bayesian methods Theresa Cain.
Quantitative Business Analysis for Decision Making Simple Linear Regression.
MACHINE LEARNING 6. Multivariate Methods 1. Based on E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 2 Motivating Example  Loan.
Probability theory 2008 Outline of lecture 5 The multivariate normal distribution  Characterizing properties of the univariate normal distribution  Different.
Simple Linear Regression Analysis
Ordinary Least Squares
Lecture 5 Correlation and Regression
Objectives of Multiple Regression
Regression and Correlation Methods Judy Zhong Ph.D.
Copyright © 2013, 2010 and 2007 Pearson Education, Inc. Chapter Inference on the Least-Squares Regression Model and Multiple Regression 14.
Chapter 4-5: Analytical Solutions to OLS
Random Regressors and Moment Based Estimation Prepared by Vera Tabakova, East Carolina University.
Repeated Measurements Analysis. Repeated Measures Analysis of Variance Situations in which biologists would make repeated measurements on same individual.
SUPA Advanced Data Analysis Course, Jan 6th – 7th 2009 Advanced Data Analysis for the Physical Sciences Dr Martin Hendry Dept of Physics and Astronomy.
Predicting Output from Computer Experiments Design and Analysis of Computer Experiments Chapter 3 Kevin Leyton-Brown.
ELEC 303 – Random Signals Lecture 18 – Classical Statistical Inference, Dr. Farinaz Koushanfar ECE Dept., Rice University Nov 4, 2010.
Regression Analysis Part C Confidence Intervals and Hypothesis Testing
Example: Bioassay experiment Problem statement –Observations: At each level of dose, 5 animals are tested, and number of death are observed.
The Simple Linear Regression Model: Specification and Estimation ECON 4550 Econometrics Memorial University of Newfoundland Adapted from Vera Tabakova’s.
Chapter 7 Point Estimation of Parameters. Learning Objectives Explain the general concepts of estimating Explain important properties of point estimators.
–The shortest distance is the one that crosses at 90° the vector u Statistical Inference on correlation and regression.
Additional Topics in Prediction Methodology. Introduction Predictive distribution for random variable Y 0 is meant to capture all the information about.
Simulation Study for Longitudinal Data with Nonignorable Missing Data Rong Liu, PhD Candidate Dr. Ramakrishnan, Advisor Department of Biostatistics Virginia.
- 1 - Calibration with discrepancy Major references –Calibration lecture is not in the book. –Kennedy, Marc C., and Anthony O'Hagan. "Bayesian calibration.
Lecture 1: Basic Statistical Tools. A random variable (RV) = outcome (realization) not a set value, but rather drawn from some probability distribution.
ESTIMATION METHODS We know how to calculate confidence intervals for estimates of  and  2 Now, we need procedures to calculate  and  2, themselves.
MathematicalMarketing Slide 5.1 OLS Chapter 5: Ordinary Least Square Regression We will be discussing  The Linear Regression Model  Estimation of the.
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Slides by JOHN LOUCKS St. Edward’s University.
Kriging - Introduction Method invented in the 1950s by South African geologist Daniel Krige (1919-) for predicting distribution of minerals. Became very.
Inference about the slope parameter and correlation
The simple linear regression model and parameter estimation
Probability Theory and Parameter Estimation I
Multiple Imputation using SOLAS for Missing Data Analysis
Ch3: Model Building through Regression
CH 5: Multivariate Methods
Set-up of the Bayesian Regression Model
Evgeniya Anatolievna Kolomak, Professor
Inference for Geostatistical Data: Kriging for Spatial Interpolation
Propagating Uncertainty In POMDP Value Iteration with Gaussian Process
Basic Econometrics Chapter 4: THE NORMALITY ASSUMPTION:
The regression model in matrix form
Simple Linear Regression
Regression Lecture-5 Additional chapters of mathematics
Simple Linear Regression
OVERVIEW OF LINEAR MODELS
ESTIMATION METHODS We know how to calculate confidence intervals for estimates of  and 2 Now, we need procedures to calculate  and 2 , themselves.
Multivariate Methods Berlin Chen, 2005 References:
Probabilistic Surrogate Models
Classical regression review
Presentation transcript:

- 1 - Preliminaries Multivariate normal model (section 3.6, Gelman) –For a multi-parameter vector y, multivariate normal distribution is where  is covariance matrix with size d, includes correlation betw i & j. –Compare this with single parameter (univariate) distribution Corresponds to  . –Compare this with multi-parameter but independent distribution. Corresponds to  = diag(  1 2,  2 2, …,  d 2 ). No correlation with each other. –Multivariate normal distribution with two parameters (  : correlation coeff)

- 2 - Preliminaries Multivariate normal model example –Two random variables x=[x1, x2] following multivariate normal distribution with zero means, unit variance and correlation . Simulate x by drawing samples with N=1000. When there is no correlation, i.e.,  =0, it is the same as sampling x1 and x2 independently, and just pair them. As the correlation is stronger, we get the following trend. If x1 is high, x2 is also likely to be high. Independent or  =0.01  =0.5  =0.9  =1.0

- 3 - Correlated regression Departures from ordinary linear regression –Previously, there were a number of assumptions. Linearity of response w.r.t. the explanatory variables Normality of the error terms Independent (uncorrelated) observations with equal variance –Non of which is true in practice. Correlations –Consider error between data and regression. Sometimes, error can’t be -1 at a point, +1 at the next. The change should be gradual. In this sense, errors at two adjacent points can be correlated. –If we ignore this, the posterior inference may lead to wrong result. Furthermore, prediction for future will also be wrong.

- 4 - Classical regression review Important equations –Functional form of the regression –Regression coefficients –Standard error –Coefficient of determination –Variance of coefficients –Variance of regression –Variance of prediction

- 5 - Classical regression review: Bayesian version Analytical procedure –Joint posterior pdf of  2 –Factorization –Marginal pdf of  2 –Conditional pdf of  –Posterior prediction version 1 Posterior prediction version 2

- 6 - Correlated regression Likelihood of correlated regression model –Consider n number of observed data y at x, where the errors are correlated. Then the likelihood of y is given by where  is covariance matrix with the size n. –Compare this with classical regression where   I, i.e., diagonal matrix with variance.

- 7 - Bayesian correlated regression Covariance matrix –We often consider variance matrix  in the following form, where Q represents correlation structure, while  2 accounts for global variance. In the matrix Q, d is a factor that controls degree of correlation. –Let’s see the behavior of function Q = exp{ -(x/d) 2 } –Let’s see the behavior of correlation matrix Q (11x11) over x=0:0.1:1. If d is small, function is narrow, correlation is small, data are independent. If d is large, function is wide, correlation is strong. d=0.01 d=0.2 d=5

- 8 - Bayesian correlated regression Correlated y example –y ~ multivariate normal dist. with zero mean & covariance Q over x=(0,1). Simulate y(x) over x = 0:0.1:1 with N=4, which follows multivariate normal distribution. When there is no correlation, y(x) at adjacent x is independent each other. As the correlation is stronger, we get the following trend. Independent or d=0.01 d=0.2 d=0.5d=5 d=100

- 9 - Bayesian correlated regression

Bayesian correlated regression

Bayesian correlated regression Sampling of posterior distribution by factorization –Data –Samples of 3 parameters (  1,  2 &  2 ) are drawn with N=5000 using the factorization approach. y=[0.95, 1.08, 1.28, 1.23, 1.42, 1.45]'; x=[ ]'; posterior distribution of b1 & b2posterior distribution of s2

Bayesian correlated regression

Bayesian correlated regression d=0.01 d=0.1 d=0.5Classical regression

Discussions on the posterior prediction Interpolation: –At every observation point, any realization of predicted y always interpolates the data y. The variance at the data point also vanishes. –The reason is explained in the following.

Discussions on the posterior prediction Choice of correlation factor –The distribution shape largely depends on the value of d. –For small d, which is small correlation, y(x) at adjacent x independent each other, which leads to the result closer to classical regression. –For large d, strong correlation between adjacent points leads to the unique knotted smoother shape like the figure, because of still passing thru the data. d=0.01 d=0.1 d=0.5

Discussions on the posterior prediction Choice of correlation factor –In the Bayesian inference, this d can be added as the unknowns as well. From the joint posterior pdf, not only the samples of  2 but also d are obtained. –In this case, we have no other way but to use MCMC. d=0.01 d=0.1 d=0.5

Practice example Sampling by MCMC –This will be implemented and discussed depending on the availability. –Only the results are suggested here. d unknown d=0.1 d=0.2

Kriging surrogate model

Kriging surrogate model