Bayesian inference of normal distribution

Slides:



Advertisements
Similar presentations
Introduction to Monte Carlo Markov chain (MCMC) methods
Advertisements

Statistical Inference Statistical Inference is the process of making judgments about a population based on properties of the sample Statistical Inference.
CHAPTER 8 More About Estimation. 8.1 Bayesian Estimation In this chapter we introduce the concepts related to estimation and begin this by considering.
Bayesian Estimation in MARK
ELEC 303 – Random Signals Lecture 18 – Statistics, Confidence Intervals Dr. Farinaz Koushanfar ECE Dept., Rice University Nov 10, 2009.
Probability & Statistical Inference Lecture 7 MSc in Computing (Data Analytics)
Descriptive statistics Experiment  Data  Sample Statistics Sample mean Sample variance Normalize sample variance by N-1 Standard deviation goes as square-root.
Statistical Inference Chapter 12/13. COMP 5340/6340 Statistical Inference2 Statistical Inference Given a sample of observations from a population, the.
Part 2b Parameter Estimation CSE717, FALL 2008 CUBS, Univ at Buffalo.
Descriptive statistics Experiment  Data  Sample Statistics Experiment  Data  Sample Statistics Sample mean Sample mean Sample variance Sample variance.
Overview of STAT 270 Ch 1-9 of Devore + Various Applications.
Using ranking and DCE data to value health states on the QALY scale using conventional and Bayesian methods Theresa Cain.
- 1 - Summary of P-box Probability bound analysis (PBA) PBA can be implemented by nested Monte Carlo simulation. –Generate CDF for different instances.
Computer vision: models, learning and inference
July 3, A36 Theory of Statistics Course within the Master’s program in Statistics and Data mining Fall semester 2011.
Lecture 7 1 Statistics Statistics: 1. Model 2. Estimation 3. Hypothesis test.
1 Inference About a Population Variance Sometimes we are interested in making inference about the variability of processes. Examples: –Investors use variance.
Review What you have learned in QA 128 Business Statistics I.
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
8-1 Introduction In the previous chapter we illustrated how a parameter can be estimated from sample data. However, it is important to understand how.
1 Confidence Intervals for Means. 2 When the sample size n< 30 case1-1. the underlying distribution is normal with known variance case1-2. the underlying.
(1) A probability model respecting those covariance observations: Gaussian Maximum entropy probability distribution for a given covariance observation.
Lecture II-2: Probability Review
1 Bayesian methods for parameter estimation and data assimilation with crop models Part 2: Likelihood function and prior distribution David Makowski and.
1 Introduction to Estimation Chapter Concepts of Estimation The objective of estimation is to determine the value of a population parameter on the.
Statistics for Data Miners: Part I (continued) S.T. Balke.
G. Cowan Lectures on Statistical Data Analysis Lecture 3 page 1 Lecture 3 1 Probability (90 min.) Definition, Bayes’ theorem, probability densities and.
Bayesian inference review Objective –estimate unknown parameter  based on observations y. Result is given by probability distribution. Bayesian inference.
Exam I review Understanding the meaning of the terminology we use. Quick calculations that indicate understanding of the basis of methods. Many of the.
12.1 Heteroskedasticity: Remedies Normality Assumption.
7.4 – Sampling Distribution Statistic: a numerical descriptive measure of a sample Parameter: a numerical descriptive measure of a population.
MEGN 537 – Probabilistic Biomechanics Ch.5 – Determining Distributions and Parameters from Observed Data Anthony J Petrella, PhD.
Inference for Regression Simple Linear Regression IPS Chapter 10.1 © 2009 W.H. Freeman and Company.
1 Two Functions of Two Random Variables In the spirit of the previous lecture, let us look at an immediate generalization: Suppose X and Y are two random.
Statistical Inference Statistical Inference is the process of making judgments about a population based on properties of the sample Statistical Inference.
Determination of Sample Size: A Review of Statistical Theory
ELEC 303 – Random Signals Lecture 18 – Classical Statistical Inference, Dr. Farinaz Koushanfar ECE Dept., Rice University Nov 4, 2010.
Simulation techniques Summary of the methods we used so far Other methods –Rejection sampling –Importance sampling Very good slides from Dr. Joo-Ho Choi.
- 1 - Bayesian inference of binomial problem Estimating a probability from binomial data –Objective is to estimate unknown proportion (or probability of.
Example: Bioassay experiment Problem statement –Observations: At each level of dose, 5 animals are tested, and number of death are observed.
Confidence Interval & Unbiased Estimator Review and Foreword.
© 2001 Prentice-Hall, Inc.Chap 7-1 BA 201 Lecture 11 Sampling Distributions.
1 Bayesian Essentials Slides by Peter Rossi and David Madigan.
- 1 - Overall procedure of validation Calibration Validation Figure 12.4 Validation, calibration, and prediction (Oberkampf and Barone, 2004 ). Model accuracy.
Chapter 20 Classification and Estimation Classification – Feature selection Good feature have four characteristics: –Discrimination. Features.
- 1 - Matlab statistics fundamentals Normal distribution % basic functions mew=100; sig=10; x=90; normpdf(x,mew,sig) 1/sig/sqrt(2*pi)*exp(-(x-mew)^2/sig^2/2)
- 1 - Computer model under uncertainty In previous lecture on accuracy assessment –We considered mostly deterministic models. – We did not distinguish.
Lecture 4 Confidence Intervals. Lecture Summary Last lecture, we talked about summary statistics and how “good” they were in estimating the parameters.
CS Statistical Machine learning Lecture 25 Yuan (Alan) Qi Purdue CS Nov
- 1 - Preliminaries Multivariate normal model (section 3.6, Gelman) –For a multi-parameter vector y, multivariate normal distribution is where  is covariance.
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
MEGN 537 – Probabilistic Biomechanics Ch.5 – Determining Distributions and Parameters from Observed Data Anthony J Petrella, PhD.
WARM UP: Penny Sampling 1.) Take a look at the graphs that you made yesterday. What are some intuitive takeaways just from looking at the graphs?
Outline Historical note about Bayes’ rule Bayesian updating for probability density functions –Salary offer estimate Coin trials example Reading material:
Bayesian Estimation and Confidence Intervals Lecture XXII.
Bayesian Inference: Multiple Parameters
Bayesian Estimation and Confidence Intervals
Statistics in Applied Science and Technology
CI for μ When σ is Unknown
Predictive distributions
CONCEPTS OF ESTIMATION
OVERVIEW OF BAYESIAN INFERENCE: PART 1
STAT 5372: Experimental Statistics
More about Posterior Distributions
STATISTICS INTERVAL ESTIMATION
CS639: Data Management for Data Science
Homework We wish to classify two species of fish (salmon and bass) based on their luminosity as measured by a camera. We have the following likelihoods:
Sampling Distributions
Introduction to Probability: Solutions for Quizzes 4 and 5
Classical regression review
Presentation transcript:

Bayesian inference of normal distribution Problem statement Objective is to estimate unknown two parameters q={m,s2} of normal distribution based on observations y = {y1, y2, …}. Prior Joint pdf of non-informative prior Joint posterior distribution This is function of two variables, which can be plotted as a surface or contour. In the todays lecture, the same thing is addressed but with different problem. In this case, it is normal distribution. And the obejctive is … remember that the normal … In this case, we have two unknwon parameters. So at first we estimate mean only while the other is assumed as given a priori. Then we move to the other case, the two parameters are switched. I mean, estiamte variance while the mean is assumed as known. Finally, which is the most practical case, estimate both parameters as unknowns.

Joint posterior distribution There is no inherent pdf function by matlab in this case. This is function of two variables, which can be plotted as a surface or contour. Let’s consider a case with n=20; ȳ =2.9; s=0.2; Remark Analysis of posterior pdf: mean, median & confidence bounds. Marginal distribution Once we have the marginal pdf, we can evaluate its mean and confidence bounds. Posterior prediction: predictive distribution of new y based on observed y. We need some basic understanding of this function within the matlab environment. So lets start the matlab program. Consider the parameters being 100 & 10. First, we can draw the shape of the function. We can compute a pdf value at a certain x. like 90. This can also be obtained using the original expression the same value obtained. The probability less than an x which is the definition of the cdf is also obtained at x=90. Or we can draw the cdf over a range of x too. See here the value at x=90 represents the cdf value which is the probability that the value is less than 90. We can also get the inverse of the cdf. Which means that when a probability is given, then the inverse is where the probability that x is less than that value. We can further generate virtual data using random sampling. In case of 50 #, we can make data using ***. Histogram can be drawn at the interval of 5 from 50 to 150. Mean, mode and median of the data are obtained. Repeat the random sampling, then you will find out the different results are obtained, which is no doubt at all. During the explanation, I would advise the students may follow the same thign that I do. So that you may understand at your hand. Because we have made virtual data frm the true parameters mew and sig, theoretically we expect the mean and stdev may be close to the true values. If we try smaller #, however, the results are more distant from the true values. This is likelihood of new y This is posterior pdf of m & s2

Marginal distributions Analytical solutions Marginal mean Marginal variance How to evaluate marginal distributions Before doing this, note that in normal dist., variable y can be transformed to standard normal variable z. Marginal mean & variance can be evaluated in the same manner. One can evaluate various characteristics using the matlab functions. Unfortunately, unlike the normal case, matlab does not provide t dist of original variable.

Marginal distributions Supplementary information for normal pdf The original pdf fY(y) and normalized pdf fZ(z) have following relation. Therefore, If we want use f(z) instead of f(y), use this relation. Supplementary information for chi2 pdf

Posterior prediction Predictive distribution of new y based on observed y Analytical solution One can evaluate characteristics of this using the matlab function. Compare with the marginal mean. Marginal mean: t-distribution with location ȳ and variance s^2/n with n-1 dof. Predictive distribution: t-distribution with location ȳ and variance s^2*(1+1/n) with n-1 dof. Likelihood of new y posterior pdf of m & s2

Simulation of joint posterior distribution Why ? Even when analytic solutions available, some are not easy to evaluate. Once exercised, may find it more convenient and more general. Practice simulation & validate with analytic solution. Factorization approach Review: joint probability of A & B Likewise, joint posterior pdf of m & s2 The two at the right are marginal pdfs. Simulation by drawing random samples We can evaluate characteristics of joint posterior pdf using simulation techniques.

Simulation of joint posterior distribution Approach 1: use marginal variance. Conditional pdf of m on s2 is already derived, which is the pdf of mean with known variance. (page 10 of Lec. #4) Marginal pdf of s2 is given in this lecture. In order to sample the posterior pdf of p(m,s2|y) Draw s2 from the marginal pdf Draw m from the conditional pdf Conditional pdf of m on s2 Marginal pdf of s2

Simulation of joint posterior distribution Approach 1: use marginal variance. Once you have obtained samples of joint pdf, compare (validate) the results with the analytic solution. Compare the samples of (M,S2) with the analytic joint pdf. In terms of scattered pot & contour. In terms of 3-D histogram & surface (mesh). Compare the samples of M with the marginal pdf, which is t distribution. Compare the samples of S2 with the marginal pdf which is inv-chi2 distribution. Extract features of the samples M and compare with analytic solution. Extract features of the samples S2 and compare with analytic solution.

Simulation of joint posterior distribution Approach 2: use marginal mean. Conditional pdf of s2 on m is already derived, which is the pdf of variance with known mean. (page 13 of Lec. #4) Marginal pdf of m is given in this lecture. In order to sample the posterior pdf of p(m,s2|y) Draw m from the marginal pdf Draw s2 from the conditional pdf Compare (validate) the resulting samples with the analytic solution. Conditional pdf of s2 on m Marginal pdf of m

Simulation of joint posterior distribution Approach 2: use marginal mean. Once you have obtained samples of joint pdf, compare (validate) the results with the analytic solution. Details are omitted for limited time.

Posterior prediction by simulation Analytic approach Result of analytic solution by double integral over infinity to get Simulation In practice, posterior predictive distribution is obtained by random draws. Once we have posterior distribution for m & s2 in the form of samples, the predictive new y are easily obtained by drawing each one from conditional on each individual m & s2. Mean & conf. intervals of posterior prediction can be obtained easily.

Homework

Homework

Homework Problems 1 Use the approach 1 to obtain samples of joint pdf (M,S2). Compare scattered dots with contour of analytic solution. Compare 3-D histogram with mesh shape of analytic solution. Compare the samples of M with the marginal pdf, which is t distribution. Compare the samples of S2 with the marginal pdf which is inv-chi2 dist. Obtain mean, 95% conf. interval of M, & compare with analytic solution. Obtain mean, 95% conf. interval of S2, & compare with analytic solution. Obtain samples of posterior prediction, and obtain mean, 95% conf. interval of ynew & compare with analytic solution.