Outline Historical note about Bayes’ rule Bayesian updating for probability density functions –Salary offer estimate Coin trials example Reading material:

Slides:



Advertisements
Similar presentations
Bayes rule, priors and maximum a posteriori
Advertisements

Lecture 4A: Probability Theory Review Advanced Artificial Intelligence.
Bayesian inference of normal distribution
Estimation, Variation and Uncertainty Simon French
Psychology 290 Special Topics Study Course: Advanced Meta-analysis April 7, 2014.
A Brief Introduction to Bayesian Inference Robert Van Dine 1.
Flipping A Biased Coin Suppose you have a coin with an unknown bias, θ ≡ P(head). You flip the coin multiple times and observe the outcome. From observations,
1 Methods of Experimental Particle Physics Alexei Safonov Lecture #21.
Sampling Distributions (§ )
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 10: The Bayesian way to fit models Geoffrey Hinton.
Ai in game programming it university of copenhagen Statistical Learning Methods Marco Loog.
Bayesian inference Gil McVean, Department of Statistics Monday 17 th November 2008.
Descriptive statistics Experiment  Data  Sample Statistics Sample mean Sample variance Normalize sample variance by N-1 Standard deviation goes as square-root.
A/Prof Geraint Lewis A/Prof Peter Tuthill
BHS Methods in Behavioral Sciences I
Using ranking and DCE data to value health states on the QALY scale using conventional and Bayesian methods Theresa Cain.
1.  Why understanding probability is important?  What is normal curve  How to compute and interpret z scores. 2.
QMS 6351 Statistics and Research Methods Probability and Probability distributions Chapter 4, page 161 Chapter 5 (5.1) Chapter 6 (6.2) Prof. Vera Adamchik.
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 7 1Probability, Bayes’ theorem, random variables, pdfs 2Functions of.
R. Kass/S07 P416 Lec 3 1 Lecture 3 The Gaussian Probability Distribution Function Plot of Gaussian pdf x p(x)p(x) Introduction l The Gaussian probability.
Probability, Bayes’ Theorem and the Monty Hall Problem
Chapter 5 Sampling and Statistics Math 6203 Fall 2009 Instructor: Ayona Chatterjee.
Additional Slides on Bayesian Statistics for STA 101 Prof. Jerry Reiter Fall 2008.
A quick intro to Bayesian thinking 104 Frequentist Approach 10/14 Probability of 1 head next: = X Probability of 2 heads next: = 0.51.
Statistical Decision Theory
Statistics for Data Miners: Part I (continued) S.T. Balke.
Introduction to Management Science
11-1 Copyright © 2010 Pearson Education, Inc. Publishing as Prentice Hall Probability and Statistics Chapter 11.
Exam I review Understanding the meaning of the terminology we use. Quick calculations that indicate understanding of the basis of methods. Many of the.
Theory of Probability Statistics for Business and Economics.
Bayesian Networks for Data Mining David Heckerman Microsoft Research (Data Mining and Knowledge Discovery 1, (1997))
ECE 8443 – Pattern Recognition LECTURE 07: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Class-Conditional Density The Multivariate Case General.
G. Cowan Lectures on Statistical Data Analysis Lecture 1 page 1 Lectures on Statistical Data Analysis London Postgraduate Lectures on Particle Physics;
Bayesian vs. frequentist inference frequentist: 1) Deductive hypothesis testing of Popper--ruling out alternative explanations Falsification: can prove.
Simulation techniques Summary of the methods we used so far Other methods –Rejection sampling –Importance sampling Very good slides from Dr. Joo-Ho Choi.
- 1 - Bayesian inference of binomial problem Estimating a probability from binomial data –Objective is to estimate unknown proportion (or probability of.
Uncertainty in Expert Systems
Statistical Decision Theory Bayes’ theorem: For discrete events For probability density functions.
Data Collection & Sampling Dr. Guerette. Gathering Data Three ways a researcher collects data: Three ways a researcher collects data: By asking questions.
1 1 Slide © 2007 Thomson South-Western. All Rights Reserved Chapter 8 Interval Estimation Population Mean:  Known Population Mean:  Known Population.
Ch15: Decision Theory & Bayesian Inference 15.1: INTRO: We are back to some theoretical statistics: 1.Decision Theory –Make decisions in the presence of.
- 1 - Overall procedure of validation Calibration Validation Figure 12.4 Validation, calibration, and prediction (Oberkampf and Barone, 2004 ). Model accuracy.
Chapter 20 Classification and Estimation Classification – Feature selection Good feature have four characteristics: –Discrimination. Features.
Bayes Theorem. Prior Probabilities On way to party, you ask “Has Karl already had too many beers?” Your prior probabilities are 20% yes, 80% no.
G. Cowan Lectures on Statistical Data Analysis Lecture 8 page 1 Statistical Data Analysis: Lecture 8 1Probability, Bayes’ theorem 2Random variables and.
MPS/MSc in StatisticsAdaptive & Bayesian - Lect 71 Lecture 7 Bayesian methods: a refresher 7.1 Principles of the Bayesian approach 7.2 The beta distribution.
Course on Bayesian Methods in Environmental Valuation
CS Statistical Machine learning Lecture 25 Yuan (Alan) Qi Purdue CS Nov
- 1 - Outline Introduction to the Bayesian theory –Bayesian Probability –Bayes’ Rule –Bayesian Inference –Historical Note Coin trials example Bayes rule.
Statistical NLP: Lecture 4 Mathematical Foundations I: Probability Theory (Ch2)
FIXETH LIKELIHOODS this is correct. Bayesian methods I: theory.
CSC321: Lecture 8: The Bayesian way to fit models Geoffrey Hinton.
Theoretical distributions: the Normal distribution.
Bayesian Estimation and Confidence Intervals Lecture XXII.
Confidence Intervals.
Statistics 200 Objectives:
Advanced Quantitative Techniques
Bayesian Estimation and Confidence Intervals
Bayesian data analysis
Bayes Net Learning: Bayesian Approaches
Econometric Models The most basic econometric model consists of a relationship between two variables which is disturbed by a random error. We need to use.
Statistical NLP: Lecture 4
Wellcome Trust Centre for Neuroimaging
Bayes for Beginners Luca Chech and Jolanda Malamud
Sampling Distributions (§ )
CS639: Data Management for Data Science
Mathematical Foundations of BME Reza Shadmehr
Classical regression review
Some Key Ingredients for Inferential Statistics
Presentation transcript:

Outline Historical note about Bayes’ rule Bayesian updating for probability density functions –Salary offer estimate Coin trials example Reading material: –Gelman, Andrew, et al. Bayesian data analysis. CRC press, 2003, Chapter 1. Slides based in part on lecture by Prof. Joo-Ho Choi of Korea Aerospace University

Historical Note Birth of Bayesian –Rev. Thomas Bayes proposed Bayes’ theory (1763):  of Binomial dist. is estimated using observed data. Laplace discovered, put his name (1812), generalized to many prob’s. –For more 100 years, Bayesian “degree of belief” was rejected as vague and subjective. Objective “frequency” was accepted in statistics. –Jeffreys (1939) rediscovered, made modern theory (1961). Until 80s, still limited due to requirement for computation. Flourishing of Bayesian –From 1990, rapid advance of HW & SW, made it practical. –Bayesian technique applied to areas of science (economics, medical) & engineering., 1999

Bayesian Probability What is Bayesian probability ? –Classical: relative frequency of an event, given many repeated trials (e.g., probability of throwing 10 with pair of dice) –Bayesian: degree of belief that it is true based on evidence at hand Saturn mass estimation –Classical: mass is fixed but unknown. –Bayesian: mass described probabilistically based on observations (e.g, uniformly in interval (a,b).

Bayes rule for pdf’s θ is a probability density to estimate based on data y. Conditional probability density functions Leading to Bayes’ rule: Often written as L used because p(y| θ ) is called the likelihood function. Instead of dividing by p(y) can divide by area under curve.

Bayesian updating –The process schematically Updated prior PDF Observed data added Prior distribution Likelihood function Observed data Posterior distribution

Salary estimate example You are considering an engineering position for which salary offers θ (in thousand dollars)have recently followed the triangular distribution Your friend received a $93K offer for a similar position, and you know that their range of offers for such positions is no more than $5K. Before your friend’s data, what was your chance of an offer <$93K ? Estimate the distribution of the expected offer and the likeliest value.

Self evaluation question What value of salary offer to your friend would leave you with the least uncertainty about your own expected offer?

Coin Trials Example Problem –For a weighted (uneven) coin, probability of heads is to be determined based on the experiments. –Assume the true θ is 0.78, obtained after ∞ trials. But we don’t know this. Only infer based on experiments. Bayesian parameter estimation This is the parameter  to be estimated. Experiment data: x times out of n trials. 4 out of 5 trials 78 out of 100 trials Experiment data: x times out of n trials. 4 out of 5 trials 78 out of 100 trials Prior knowledge on p 0 (  ) 1. No prior information 2. Normal dist centered at 0.5 with  = Uniform distribution [0.5, 0.7] Prior knowledge on p 0 (  ) 1. No prior information 2. Normal dist centered at 0.5 with  = Uniform distribution [0.5, 0.7]

Probability of heads posterior distributions Prior 1.No prior (uniform) 2.N(0.5,0.05), poor prior slows convergence. 3.U(0.5,0.7) cannot exceed barrier due to incorrect prior figure Red: prior Wide: 4 out of 5 Narrow; 78 out of 100

Probability of 5 consecutive heads Prediction using posterior (no prior case) Exact value is binom(5,5,0.78) = = Posterior PDF of  Draw random samples of  from PDF Posterior prediction process Compute p based on each  binom(5,5,  ) Compute p based on each  binom(5,5,  ) median5% CI95% CI Estimation process 10,000 samples of  10,000 samples of predicted p

practice problems 1.For the salary estimate problem, what is the probability of getting a better offer than your friend? 2.For the salary problem, calculate the 95% confidence bounds on your salary around the mean and median of your expected salary distribution. 3. Slide 9 shows the risks associated with using a prior. When is it important to use a prior? Source: Smithsonian Institution Number: