Neural Computation Chapter 3. Neural Computation Outline Comparison of behavioral and neural response on a discrimination task –Bayes rule –ROC curves.

Slides:



Advertisements
Similar presentations
Bayes rule, priors and maximum a posteriori
Advertisements

Lecture XXIII.  In general there are two kinds of hypotheses: one concerns the form of the probability distribution (i.e. is the random variable normally.
ECE 8443 – Pattern Recognition LECTURE 05: MAXIMUM LIKELIHOOD ESTIMATION Objectives: Discrete Features Maximum Likelihood Resources: D.H.S: Chapter 3 (Part.
Spike Train Statistics Sabri IPM. Review of spike train  Extracting information from spike trains  Noisy environment:  in vitro  in vivo  measurement.
Reading population codes: a neural implementation of ideal observers Sophie Deneve, Peter Latham, and Alexandre Pouget.
How well can we learn what the stimulus is by looking at the neural responses? We will discuss two approaches: devise and evaluate explicit algorithms.
For stimulus s, have estimated s est Bias: Cramer-Rao bound: Mean square error: Variance: Fisher information How good is our estimate? (ML is unbiased:
Bayesian Decision Theory Chapter 2 (Duda et al.) – Sections
Parameter Estimation: Maximum Likelihood Estimation Chapter 3 (Duda et al.) – Sections CS479/679 Pattern Recognition Dr. George Bebis.
Classification and risk prediction
0 Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
AGC DSP AGC DSP Professor A G Constantinides© Estimation Theory We seek to determine from a set of data, a set of parameters such that their values would.
Pattern Classification, Chapter 3 Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P.
Statistical Decision Theory, Bayes Classifier
Spike Train decoding Summary Decoding of stimulus from response –Two choice case Discrimination ROC curves –Population decoding MAP and ML estimators.
Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 1 Computational Architectures in.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Data Selection In Ad-Hoc Wireless Sensor Networks Olawoye Oyeyele 11/24/2003.
EE513 Audio Signals and Systems Statistical Pattern Classification Kevin D. Donohue Electrical and Computer Engineering University of Kentucky.
Principles of Pattern Recognition
ECE 8443 – Pattern Recognition LECTURE 06: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Bias in ML Estimates Bayesian Estimation Example Resources:
Population Coding Alexandre Pouget Okinawa Computational Neuroscience Course Okinawa, Japan November 2004.
1 Information Geometry on Classification Logistic, AdaBoost, Area under ROC curve Shinto Eguchi – – ISM seminor on 17/1/2001 This talk is based on one.
1 / 41 Inference and Computation with Population Codes 13 November 2012 Inference and Computation with Population Codes Alexandre Pouget, Peter Dayan,
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 03: GAUSSIAN CLASSIFIERS Objectives: Whitening.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum.
Synergy, redundancy, and independence in population codes, revisited. or Are correlations important? Peter Latham* and Sheila Nirenberg † *Gatsby Computational.
Jun-Won Suh Intelligent Electronic Systems Human and Systems Engineering Department of Electrical and Computer Engineering Speaker Verification System.
Lecture notes for Stat 231: Pattern Recognition and Machine Learning 3. Bayes Decision Theory: Part II. Prof. A.L. Yuille Stat 231. Fall 2004.
1 E. Fatemizadeh Statistical Pattern Recognition.
Statistics In HEP Helge VossHadron Collider Physics Summer School June 8-17, 2011― Statistics in HEP 1 How do we understand/interpret our measurements.
Bayesian Decision Theory (Classification) 主講人:虞台文.
: Chapter 3: Maximum-Likelihood and Baysian Parameter Estimation 1 Montri Karnjanadecha ac.th/~montri.
Population coding Population code formulation Methods for decoding: population vector Bayesian inference maximum a posteriori maximum likelihood Fisher.
1 Chapter 12 Probabilistic Reasoning and Bayesian Belief Networks.
1 Pattern Recognition: Statistical and Neural Lonnie C. Ludeman Lecture 12 Sept 30, 2005 Nanjing University of Science & Technology.
Chapter 3: Maximum-Likelihood Parameter Estimation l Introduction l Maximum-Likelihood Estimation l Multivariate Case: unknown , known  l Univariate.
BCS547 Neural Decoding. Population Code Tuning CurvesPattern of activity (r) Direction (deg) Activity
1 Pattern Recognition: Statistical and Neural Lonnie C. Ludeman Lecture 8 Sept 23, 2005 Nanjing University of Science & Technology.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
ECE 471/571 – Lecture 2 Bayesian Decision Theory 08/25/15.
BCS547 Neural Decoding.
Bayesian Decision Theory Basic Concepts Discriminant Functions The Normal Density ROC Curves.
Chapter 20 Classification and Estimation Classification – Feature selection Good feature have four characteristics: –Discrimination. Features.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
6. Population Codes Presented by Rhee, Je-Keun © 2008, SNU Biointelligence Lab,
EE 551/451, Fall, 2006 Communication Systems Zhu Han Department of Electrical and Computer Engineering Class 15 Oct. 10 th, 2006.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Lecture 1.31 Criteria for optimal reception of radio signals.
Decoding How well can we learn what the stimulus is by looking
CS479/679 Pattern Recognition Dr. George Bebis
Chapter 3: Maximum-Likelihood Parameter Estimation
LECTURE 03: DECISION SURFACES
Parameter Estimation 主講人:虞台文.
Synapses Signal is carried chemically across the synaptic cleft.
LECTURE 05: THRESHOLD DECODING
9 Tests of Hypotheses for a Single Sample CHAPTER OUTLINE
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Parametric Estimation
spike-triggering stimulus features
EE513 Audio Signals and Systems
Probabilistic Population Codes for Bayesian Decision Making
Parametric Methods Berlin Chen, 2005 References:
Mathematical Foundations of BME
LECTURE 05: THRESHOLD DECODING
ECE – Pattern Recognition Lecture 4 – Parametric Estimation
Presentation transcript:

Neural Computation Chapter 3

Neural Computation Outline Comparison of behavioral and neural response on a discrimination task –Bayes rule –ROC curves –Neyman Pearson Lemma Population decoding –Cricket cercal system –Monkey M1 motoneurons Optimal decoding –MAP and Bayesian estimates –Relation to population vector –Fisher information

Neural Computation Bayes’ rule Let s denote a stimulus and r=(r 1,…,r N ) denote the response of one or more neurons. We define –The stimulus probability p(s) –The response probability p(r) –The joint probability p(r,s) and conditional probabilities p(r|s) and p(s|r) Bayes rule:

Neural Computation Discrimination of movements Stimulus is moving dot pattern with variable % of coherently moving dots. Monkey behavioral forced choice task to report direction of motion (+ or -) as a function of coherence in the stimulus (filled circles) Monkey decides on basis of neural response. Open circles are optimal discrimination performance given the neural response data

Neural Computation Two alternative forced choice

Neural Computation

Optimal decision: ROC curve The classification requires the definition of a threshold. The threshold z affects the classification performance: –Define the false alarm rate (size)  =p(r>z|s=-) and hit rate (power)  =p(r>z|s=+) –ROC (‘receiver operating characteristic’) plot  (z) vs.  (z) Area under curve s d   / classification performance: –½ is random guessing –1 is perfect classification

Neural Computation Two alternative forced choice Given a stimulus s=+ and the response in two neurons p+ and p-, That give rates r+ and r- respectively. Stimulus is classified by the highest rate. What is the probability of correct classification?

Neural Computation Two alternative forced choice White circles are p(correct) as a function of stimulus coherence Monkeys response is as if based on two alternative forced choice

Neural Computation Discriminability

Neural Computation Discriminability (details) Discuss ex. 3.1

Neural Computation Likelihood ratio

Neural Computation Neyman-Pearson Lemma

Neural Computation Neyman-Pearson Lemma

Neural Computation Neyman-Pearson Lemma

Neural Computation Cricket cercal system Consider response of a population of neurons p(r 1,…,r N |s) Consider a stimulus that is parametrized by a continuous value Cricket cercal system –Hair cells send spike when deflected by wind –4 inter-neurons receive input from thousands of hair cells

Neural Computation Cricket cercal system

Neural Computation Cricket cercal system

Neural Computation Monkey primary motor cortex (M1)

Neural Computation Monkey primary motor cortex (M1)

Neural Computation Optimal decoding

Neural Computation Optimal decoding

Optimal decoding details log p(r|s) \propto \sum_{a=1}^4 (r_a-f_a(s))^2 Assume f_a(s)=c_{ai} v_i Then log p(r|s) as a function of v has maximum at Sum_{ja} c_{ia} c_{ja} v_j = sum_a c_{ia} r_a This is the MAP estimate Bayesian estimate requires p(s|r) which is a normalized version of p(r|s) as a function of s. Neural Computation

Bayesian vs. Population vector decoding

Neural Computation Bayesian vs. Population vector decoding

Neural Computation Bayesian vs. Population vector decoding

Neural Computation Bias and variance

Neural Computation Bias and variance

Neural Computation Fisher information

Neural Computation

Fisher information

Neural Computation Fisher information

Neural Computation Fisher information

Neural Computation

S_est=sum_i a_i r_i sum_i a_i=1 r_i is unbiased estimator of s = sum_i a_i s= s Sigma^2_est=sum_i a_i^2 sigma^2 A_i =1/n -> Sigma^2_est = sigma^2/n optimal A_i different is suboptimal.

Neural Computation Fisher information

Neural Computation Fisher information

Neural Computation

Spike Train decoding

Neural Computation Spike Train decoding

Neural Computation Spike Train decoding

Neural Computation Summary Decoding of stimulus from response –Two choice case Discrimination ROC curves –Population decoding MAP and ML estimators Bias and variance Fisher information, Cramer-Rao bound –Spike train decoding

Neural Computation