Bayesian Brain Probabilistic Approaches to Neural Coding 1.1 A Probability Primer Bayesian Brain Probabilistic Approaches to Neural Coding 1.1 A Probability.

Slides:



Advertisements
Similar presentations
Modeling of Data. Basic Bayes theorem Bayes theorem relates the conditional probabilities of two events A, and B: A might be a hypothesis and B might.
Advertisements

Naïve Bayes. Bayesian Reasoning Bayesian reasoning provides a probabilistic approach to inference. It is based on the assumption that the quantities of.
Probabilistic models Haixu Tang School of Informatics.
Pattern Recognition and Machine Learning
Rutgers CS440, Fall 2003 Review session. Rutgers CS440, Fall 2003 Topics Final will cover the following topics (after midterm): 1.Uncertainty & introduction.
ECE 8443 – Pattern Recognition LECTURE 05: MAXIMUM LIKELIHOOD ESTIMATION Objectives: Discrete Features Maximum Likelihood Resources: D.H.S: Chapter 3 (Part.
Biointelligence Laboratory, Seoul National University
Simulation with ArenaAppendix C – A Refresher on Probability and StatisticsSlide 1 of 33 A Refresher on Probability and Statistics Appendix C.
Integration of sensory modalities
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 10: The Bayesian way to fit models Geoffrey Hinton.
Introduction of Probabilistic Reasoning and Bayesian Networks
Chapter 4: Linear Models for Classification
Laboratory for Social & Neural Systems Research (SNS) PATTERN RECOGNITION AND MACHINE LEARNING Institute of Empirical Research in Economics (IEW)
Visual Recognition Tutorial
1 Def: Let and be random variables of the discrete type with the joint p.m.f. on the space S. (1) is called the mean of (2) is called the variance of (3)
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem, random variables, pdfs 2Functions.
G. Cowan Lectures on Statistical Data Analysis Lecture 2 page 1 Statistical Data Analysis: Lecture 2 1Probability, Bayes’ theorem 2Random variables and.
Machine Learning CMPT 726 Simon Fraser University CHAPTER 1: INTRODUCTION.
CSE 221: Probabilistic Analysis of Computer Systems Topics covered: Statistical inference (Sec. )
Descriptive statistics Experiment  Data  Sample Statistics Experiment  Data  Sample Statistics Sample mean Sample mean Sample variance Sample variance.
Machine Learning CMPT 726 Simon Fraser University
Artificial Intelligence Term Project #3 Kyu-Baek Hwang Biointelligence Lab School of Computer Science and Engineering Seoul National University
Computer vision: models, learning and inference
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
Probability and Statistics Review Thursday Sep 11.
(1) A probability model respecting those covariance observations: Gaussian Maximum entropy probability distribution for a given covariance observation.
1 Bayesian methods for parameter estimation and data assimilation with crop models Part 2: Likelihood function and prior distribution David Makowski and.
Jointly Distributed Random Variables
Estimating parameters in a statistical model Likelihood and Maximum likelihood estimation Bayesian point estimates Maximum a posteriori point.
1 Foundations of Statistical Natural Language Processing By Christopher Manning & Hinrich Schutze Course Book.
7-1 Introduction The field of statistical inference consists of those methods used to make decisions or to draw conclusions about a population. These.
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 11: Bayesian learning continued Geoffrey Hinton.
Computer Science, Software Engineering & Robotics Workshop, FGCU, April 27-28, 2012 Fault Prediction with Particle Filters by David Hatfield mentors: Dr.
ECE 8443 – Pattern Recognition LECTURE 07: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Class-Conditional Density The Multivariate Case General.
Ch 2. Probability Distributions (1/2) Pattern Recognition and Machine Learning, C. M. Bishop, Summarized by Yung-Kyun Noh and Joo-kyung Kim Biointelligence.
Ch 8. Graphical Models Pattern Recognition and Machine Learning, C. M. Bishop, Revised by M.-O. Heo Summarized by J.W. Nam Biointelligence Laboratory,
Correlation Assume you have two measurements, x and y, on a set of objects, and would like to know if x and y are related. If they are directly related,
BCS547 Neural Decoding. Population Code Tuning CurvesPattern of activity (r) Direction (deg) Activity
Mathematical Foundations Elementary Probability Theory Essential Information Theory Updated 11/11/2005.
INTRODUCTION TO Machine Learning 3rd Edition
BCS547 Neural Decoding.
Ch15: Decision Theory & Bayesian Inference 15.1: INTRO: We are back to some theoretical statistics: 1.Decision Theory –Make decisions in the presence of.
Lecture 2: Statistical learning primer for biologists
Lecture 1: Basic Statistical Tools. A random variable (RV) = outcome (realization) not a set value, but rather drawn from some probability distribution.
6. Population Codes Presented by Rhee, Je-Keun © 2008, SNU Biointelligence Lab,
1 Chapter 8: Model Inference and Averaging Presented by Hui Fang.
Review of statistical modeling and probability theory Alan Moses ML4bio.
Ch 2. Probability Distributions (1/2) Pattern Recognition and Machine Learning, C. M. Bishop, Summarized by Joo-kyung Kim Biointelligence Laboratory,
Parameter Estimation. Statistics Probability specified inferred Steam engine pump “prediction” “estimation”
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
Essential Probability & Statistics (Lecture for CS397-CXZ Algorithms in Bioinformatics) Jan. 23, 2004 ChengXiang Zhai Department of Computer Science University.
Ch 1. Introduction (Latter) Pattern Recognition and Machine Learning, C. M. Bishop, Summarized by J.W. Ha Biointelligence Laboratory, Seoul National.
Learning Machine Learning Further Introduction from Eleftherios inspired mainly from Ghahramani’s and Bishop’s lectures.
Ch 1. Introduction Pattern Recognition and Machine Learning, C. M. Bishop, Updated by J.-H. Eom (2 nd round revision) Summarized by K.-I.
CSC321: Lecture 8: The Bayesian way to fit models Geoffrey Hinton.
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 1: INTRODUCTION.
Probability Theory and Parameter Estimation I
Probability for Machine Learning
ICS 280 Learning in Graphical Models
Ch3: Model Building through Regression
Course: Autonomous Machine Learning
Special Topics In Scientific Computing
Statistical NLP: Lecture 4
Integration of sensory modalities
Biointelligence Laboratory, Seoul National University
LECTURE 23: INFORMATION THEORY REVIEW
Biointelligence Laboratory, Seoul National University
Ch 3. Linear Models for Regression (2/2) Pattern Recognition and Machine Learning, C. M. Bishop, Previously summarized by Yung-Kyun Noh Updated.
Mathematical Foundations of BME Reza Shadmehr
CS639: Data Management for Data Science
Presentation transcript:

Bayesian Brain Probabilistic Approaches to Neural Coding 1.1 A Probability Primer Bayesian Brain Probabilistic Approaches to Neural Coding 1.1 A Probability Primer Jinsan Yang Biointelligence Lab School of Computer Sci. & Eng. Seoul National University

(c) SNU CSE Biointelligence Lab2 I A Probability Primer 1.1 What is Probability 1.2 Bayes Theorem 1.3 Measuring Information 1.4 Making an Inferance 1.5 Learning from Data 1.6 Graphical Models and Other Bayesian Algorithms

(c) SNU CSE Biointelligence Lab3 1.1 What is Probability Two views of interpretation for probability  Frequentist view  Bayesian view Probability distribution and density  Random variable, vector from sample space  Probability mass function (pmf), probability density function (pdf)  Cumulative distribution function (cdf), distribution function Expectation and Statistics  mean, variance, covariance, correlation

(c) SNU CSE Biointelligence Lab4 1.1 What is Probability Joint, conditional and marginal probability Independence and correlation  Independent rv’s:  Uncorrelated rv’s: If two variables are independent, they are uncorrelated but the reverse is not always true

(c) SNU CSE Biointelligence Lab5 1.2 Bayes Theorem How to update the belief of a hypothesis based on how well the acquired data were predicted from the hypothesis  Prior, posterior probability  Generative model  Marginal data likelihood

(c) SNU CSE Biointelligence Lab6 1.3 Measuring information Smaller P(X) means more informative Entropy  Average information  Measure of randomness Mutual information  Decrease of uncertainty about the world X by observing Y

(c) SNU CSE Biointelligence Lab7 1.3 Measuring information Kullback-Leibler Divergence  Measure the difference in two probability distributions by information  Difference in information between P(X) and Q(X) when X follow the distribution P(X)  Does not satisfy the symmetry condition

(c) SNU CSE Biointelligence Lab8 1.4 Making an inference Maximum likelihood estimate  Finding the state of the world X by maximizing the likelihood P(Y|X) of sensory input y to the brain  Point estimate may not be enough Maximum a posteriori estimate  Combining sensory information and prior probability of world states Bayesian estimate  Full probability distribution instead of a point estimate

(c) SNU CSE Biointelligence Lab9 1.4 Making an inference Bayes filtering  Using the posterior probability as the prior probability in the next step  When the state changes by a state transition probability  Examples : Kalman filter, particle filter

(c) SNU CSE Biointelligence Lab Learning from data Learning sensory transformation or state transition from experience by parameter estimation Fisher information: measure of steepness of the likelihood (How good is the estimation) Bayesian Learning: principled way of adding regularization terms

(c) SNU CSE Biointelligence Lab Learning from data Bayesian Learning: principled way of adding regularization terms Maximizing wrt. w is the same as minimizing the least mean squared error with a penalty term

(c) SNU CSE Biointelligence Lab Learning from data Marginal likelihood is a good criterion to see whether the prior is consistent with the observed data (called evidence) Represent dependency relations by graphical models  Bayesian network: directed acyclic graph (DAG) 1.6 Graphical models and other Bayesian algorithms