A Brief Introduction to Bayesian Inference Robert Van Dine 1.

Slides:



Advertisements
Similar presentations
Bayes rule, priors and maximum a posteriori
Advertisements

Week 11 Review: Statistical Model A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution.
Bayesian inference “Very much lies in the posterior distribution” Bayesian definition of sufficiency: A statistic T (x 1, …, x n ) is sufficient for 
Psychology 290 Special Topics Study Course: Advanced Meta-analysis April 7, 2014.
LECTURE 11: BAYESIAN PARAMETER ESTIMATION
Copyright ©2010 Pearson Education, Inc. publishing as Prentice Hall 9- 1 Basic Marketing Research: Using Microsoft Excel Data Analysis, 3 rd edition Alvin.
A/Prof Geraint Lewis A/Prof Peter Tuthill
Lecture Inference for a population mean when the stdev is unknown; one more example 12.3 Testing a population variance 12.4 Testing a population.
I The meaning of chance Axiomatization. E Plurbus Unum.
Using ranking and DCE data to value health states on the QALY scale using conventional and Bayesian methods Theresa Cain.
Applied Bayesian Analysis for the Social Sciences Philip Pendergast Computing and Research Services Department of Sociology
7-2 Estimating a Population Proportion
Lecture 9: p-value functions and intro to Bayesian thinking Matthew Fox Advanced Epidemiology.
Statistical Analysis of Systematic Errors and Small Signals Reinhard Schwienhorst University of Minnesota 10/26/99.
1 BA 275 Quantitative Business Methods Hypothesis Testing Elements of a Test Concept behind a Test Examples Agenda.
1 Bayesian methods for parameter estimation and data assimilation with crop models Part 2: Likelihood function and prior distribution David Makowski and.
Additional Slides on Bayesian Statistics for STA 101 Prof. Jerry Reiter Fall 2008.
Bayesian Inference, Basics Professor Wei Zhu 1. Bayes Theorem Bayesian statistics named after Thomas Bayes ( ) -- an English statistician, philosopher.
Topic 5 Statistical inference: point and interval estimate
Chap 20-1 Statistics for Business and Economics, 6e © 2007 Pearson Education, Inc. Chapter 20 Sampling: Additional Topics in Sampling Statistics for Business.
Prof. Dr. S. K. Bhattacharjee Department of Statistics University of Rajshahi.
Random Sampling, Point Estimation and Maximum Likelihood.
Estimating a Population Proportion
Education Research 250:205 Writing Chapter 3. Objectives Subjects Instrumentation Procedures Experimental Design Statistical Analysis  Displaying data.
Introduction Osborn. Daubert is a benchmark!!!: Daubert (1993)- Judges are the “gatekeepers” of scientific evidence. Must determine if the science is.
Applied Bayesian Inference, KSU, April 29, 2012 § ❷ / §❷ An Introduction to Bayesian inference Robert J. Tempelman 1.
ECE 8443 – Pattern Recognition LECTURE 07: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Class-Conditional Density The Multivariate Case General.
HYPOTHESIS TESTING. Statistical Methods Estimation Hypothesis Testing Inferential Statistics Descriptive Statistics Statistical Methods.
Maximum Likelihood - "Frequentist" inference x 1,x 2,....,x n ~ iid N( ,  2 ) Joint pdf for the whole random sample Maximum likelihood estimates.
Bayesian vs. frequentist inference frequentist: 1) Deductive hypothesis testing of Popper--ruling out alternative explanations Falsification: can prove.
Bayesian statistics Probabilities for everything.
Copyright © 2010, 2007, 2004 Pearson Education, Inc. All Rights Reserved. Section 7-1 Review and Preview.
Making sense of randomness
Statistical Decision Theory Bayes’ theorem: For discrete events For probability density functions.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 07: BAYESIAN ESTIMATION (Cont.) Objectives:
Confidence Interval & Unbiased Estimator Review and Foreword.
Ch15: Decision Theory & Bayesian Inference 15.1: INTRO: We are back to some theoretical statistics: 1.Decision Theory –Make decisions in the presence of.
MATH 643 Bayesian Statistics. 2 Discrete Case n There are 3 suspects in a murder case –Based on available information, the police think the following.
Chapter 20 Classification and Estimation Classification – Feature selection Good feature have four characteristics: –Discrimination. Features.
CHAPTER 9 Inference: Estimation The essential nature of inferential statistics, as verses descriptive statistics is one of knowledge. In descriptive statistics,
Statistical Inference Statistical inference is concerned with the use of sample data to make inferences about unknown population parameters. For example,
Computer vision: models, learning and inference Chapter 2 Introduction to probability.
Statistics Sampling Distributions and Point Estimation of Parameters Contents, figures, and exercises come from the textbook: Applied Statistics and Probability.
- 1 - Outline Introduction to the Bayesian theory –Bayesian Probability –Bayes’ Rule –Bayesian Inference –Historical Note Coin trials example Bayes rule.
Statistical Methods. 2 Concepts and Notations Sample unit – the basic landscape unit at which we wish to establish the presence/absence of the species.
CHAPTER 3: BAYESIAN DECISION THEORY. Making Decision Under Uncertainty Based on E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
Parameter Estimation. Statistics Probability specified inferred Steam engine pump “prediction” “estimation”
Bayesian Brain Probabilistic Approaches to Neural Coding 1.1 A Probability Primer Bayesian Brain Probabilistic Approaches to Neural Coding 1.1 A Probability.
Outline Historical note about Bayes’ rule Bayesian updating for probability density functions –Salary offer estimate Coin trials example Reading material:
Exploring the connections…. NeurosciencePhilosophy.
Canadian Bioinformatics Workshops
Class Six Turn In: Chapter 15: 30, 32, 38, 44, 48, 50 Chapter 17: 28, 38, 44 For Class Seven: Chapter 18: 32, 34, 36 Chapter 19: 26, 34, 44 Quiz 3 Read.
Unit 5: Hypothesis Testing
Presented by: Karen Miller
Heisenberg’s Uncertainty principle
Bayes for Beginners Stephanie Azzopardi & Hrvoje Stojic
CHAPTER 9 Testing a Claim
(Very Brief) Introduction to Bayesian Statistics
Bayesian Inference, Basics
Section 7.7 Introduction to Inference
Significance Tests: The Basics
Significance Tests: The Basics
Basic Practice of Statistics - 3rd Edition Introduction to Inference
Bayes for Beginners Luca Chech and Jolanda Malamud
CS 594: Empirical Methods in HCC Introduction to Bayesian Analysis
Parametric Methods Berlin Chen, 2005 References:
Inference on the Mean of a Population -Variance Known
CS639: Data Management for Data Science
Mathematical Foundations of BME Reza Shadmehr
Applied Statistics and Probability for Engineers
Presentation transcript:

A Brief Introduction to Bayesian Inference Robert Van Dine 1

"We may regard the present state of the universe as the effect of its past and the cause of its future. An intellect which at a certain moment would know all forces that set nature in motion, and all positions of all items of which nature is composed, if this intellect were also vast enough to submit these data to analysis, it would embrace in a single formula the movements of the greatest bodies of the universe and those of the tiniest atom; for such an intellect nothing would be uncertain and the future just like the past would be present before its eyes." (Laplace, 1814). 2

A Brief History Bayes’ theorem was developed by the Revered Thomas Bayes ( ), however it was not published until after his death The idea developed by Thomas Bayes was not well known until it was independently published by Pierre-Simone Laplace in 1774, who was one of the first mathematicians to apply probabilistic ideas to scientific inquiry In 1814 Laplace published A Philosophical Essay on Probabilities which developed the Bayesian interpretation of probability more thoroughly – some refer to Bayesian inference as Laplacian inference because of this 3

In 1939 Theory of Probability, written by Harold Jeffreys, revived interest in Bayesian inference During World War II, Alan Turing used an early form of Bayesian networks in his work to decode the Nazi Enigma machine In 1946 Richard Cox showed that the rules of Bayesian inference have a well-formulated axiomatic basis and that it is the only inferential approach that is logically consistent 4 A Brief History Continued

Some Definitions Prior : the distribution of the parameter that is assumed before any data is observed  The prior captures our knowledge and belief about the parameter Evidence: the marginal likelihood of the data Posterior : the distribution of the parameter after taking into account the observed data 5

Overview of Bayesian Inference In Bayesian inference, the parameter of interest is considered a random variable rather than a fixed value The rules of probability are used to make direct inferences about the parameter Probability statements about the parameter are interpreted as “degree of belief” Bayes’ theorem is used to revise our beliefs about the parameter after getting the data – it expresses how a subjective degree of belief should rationally change to account for new evidence 6

Derivation of Bayes’ Theorem 7

Interpretation of Bayes’ Theorem 8

Example: Estimating a Binomial Parameter 9

Review : The Beta Distribution 10

The Beta Prior 11 Simply add the number of successes to a and the number of failures to b to obtain the posterior

How to Choose a Prior 12

Flavors of the Beta Distribution 13

The BernBeta Function 14 A snippet of the R code: -> The BernBeta function takes a Beta prior and data vector as arguments and produces plots of the prior, likelihood and posterior distributions

The BernBeta Function 15

Output of BernBeta 16

Example 2 – Ned and Fred Suppose two NBA scouts, Ned and Fred, are scouting player A and player B respectively for the NBA draft They would like to know the players’ shooting percentages on jump shots off the pick and roll – this is an ability that is very important in the NBA, but it is not an officially kept statistic Suppose further that the true (unmeasurable) abilities of player A and player B are identical – both are 45% jump shooters off the pick and roll, but because of some unknown biases Ned and Fred have very different prior beliefs 17

Ned and Fred – NBA Scouts Ned believes player A is a 25% shooter and Fred believes player B is a 60% shooter In reality both players are 40% shooters and they average 5 such shots per game Ned observes player A for 10 games and Fred observes player B for 10 games – using Bayesian reasoning to update their beliefs, how much will their prior beliefs affect their conclusions about the players? 18

The post.Demo function 19

Ned Fred 20

Inference The posterior distribution summarizes our belief about the parameter after looking at the data According to the Bayesian point of view, inferences about the parameter are drawn from the posterior distribution – they are conditional on the sample that actually occurred Frequentist inferences about the parameter involve probabilities calculated from the sampling distribution based on all possible sample that could have occurred, probabilities that are not conditional on the sample that did occur 21

Point Estimation 22

MSE Comparison 23

Bayesian vs. Frequentist MSE 24 Frequentist Bayesian

Interval Estimation The Bayesian credible interval is calculated directly from the posterior distribution It has a straightforward “degree of belief” probability interpretation – it summarizes the parameter values that could be credibly believed given the observed data Contrast with frequentist confidence intervals 25

Credible Intervals 26

Hypothesis Testing: One Sided 27

Hypothesis Testing: Two Sided 28

References Introduction to Bayesian statistics By: Bolstad, William M. Wiley-Interscience 2004 Doing Bayesian data analysis: a tutorial with R and BUGS By: Kruschke, John K. Academic Press