BCS547 Neural Decoding.

Slides:



Advertisements
Similar presentations
Pattern Recognition and Machine Learning
Advertisements

Pattern Recognition and Machine Learning
What is the neural code? Puchalla et al., What is the neural code? Encoding: how does a stimulus cause the pattern of responses? what are the responses.
Pattern Recognition and Machine Learning
ECE 8443 – Pattern Recognition LECTURE 05: MAXIMUM LIKELIHOOD ESTIMATION Objectives: Discrete Features Maximum Likelihood Resources: D.H.S: Chapter 3 (Part.
Pattern Recognition and Machine Learning
Probability and Statistics Basic concepts II (from a physicist point of view) Benoit CLEMENT – Université J. Fourier / LPSC
Neural Computation Chapter 3. Neural Computation Outline Comparison of behavioral and neural response on a discrimination task –Bayes rule –ROC curves.
1 12. Principles of Parameter Estimation The purpose of this lecture is to illustrate the usefulness of the various concepts introduced and studied in.
Chapter 7. Statistical Estimation and Sampling Distributions
Chapter 7 Title and Outline 1 7 Sampling Distributions and Point Estimation of Parameters 7-1 Point Estimation 7-2 Sampling Distributions and the Central.
Pattern Classification. Chapter 2 (Part 1): Bayesian Decision Theory (Sections ) Introduction Bayesian Decision Theory–Continuous Features.
Chapter 4: Linear Models for Classification
Visual Recognition Tutorial
Reading population codes: a neural implementation of ideal observers Sophie Deneve, Peter Latham, and Alexandre Pouget.
How well can we learn what the stimulus is by looking at the neural responses? We will discuss two approaches: devise and evaluate explicit algorithms.
For stimulus s, have estimated s est Bias: Cramer-Rao bound: Mean square error: Variance: Fisher information How good is our estimate? (ML is unbiased:
Parameter Estimation: Maximum Likelihood Estimation Chapter 3 (Duda et al.) – Sections CS479/679 Pattern Recognition Dr. George Bebis.
Maximum likelihood (ML) and likelihood ratio (LR) test
0 Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Maximum likelihood (ML)
AGC DSP AGC DSP Professor A G Constantinides© Estimation Theory We seek to determine from a set of data, a set of parameters such that their values would.
Maximum likelihood (ML) and likelihood ratio (LR) test
Spike Train decoding Summary Decoding of stimulus from response –Two choice case Discrimination ROC curves –Population decoding MAP and ML estimators.
Visual Recognition Tutorial
Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 1 Computational Architectures in.
Maximum likelihood (ML)
Binary Variables (1) Coin flipping: heads=1, tails=0 Bernoulli Distribution.
PATTERN RECOGNITION AND MACHINE LEARNING
Population Coding Alexandre Pouget Okinawa Computational Neuroscience Course Okinawa, Japan November 2004.
Speech Recognition Pattern Classification. 22 September 2015Veton Këpuska2 Pattern Classification  Introduction  Parametric classifiers  Semi-parametric.
Model Inference and Averaging
Prof. Dr. S. K. Bhattacharjee Department of Statistics University of Rajshahi.
Random Sampling, Point Estimation and Maximum Likelihood.
1 / 41 Inference and Computation with Population Codes 13 November 2012 Inference and Computation with Population Codes Alexandre Pouget, Peter Dayan,
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum.
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 3: LINEAR MODELS FOR REGRESSION.
Projects: 1.Predictive coding in balanced spiking networks (Erwan Ledoux). 2.Using Canonical Correlation Analysis (CCA) to analyse neural data (David Schulz).
ECE 8443 – Pattern Recognition LECTURE 07: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Class-Conditional Density The Multivariate Case General.
CS 782 – Machine Learning Lecture 4 Linear Models for Classification  Probabilistic generative models  Probabilistic discriminative models.
Ch 4. Linear Models for Classification (1/2) Pattern Recognition and Machine Learning, C. M. Bishop, Summarized and revised by Hee-Woong Lim.
PROBABILITY AND STATISTICS FOR ENGINEERING Hossein Sameti Department of Computer Engineering Sharif University of Technology Principles of Parameter Estimation.
Population coding Population code formulation Methods for decoding: population vector Bayesian inference maximum a posteriori maximum likelihood Fisher.
BCS547 Neural Decoding. Population Code Tuning CurvesPattern of activity (r) Direction (deg) Activity
Chapter 7 Point Estimation of Parameters. Learning Objectives Explain the general concepts of estimating Explain important properties of point estimators.
Lecture 2: Statistical learning primer for biologists
Chapter 20 Classification and Estimation Classification – Feature selection Good feature have four characteristics: –Discrimination. Features.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Occam’s Razor No Free Lunch Theorem Minimum.
Machine Learning 5. Parametric Methods.
6. Population Codes Presented by Rhee, Je-Keun © 2008, SNU Biointelligence Lab,
Review of statistical modeling and probability theory Alan Moses ML4bio.
Chapter 8 Estimation ©. Estimator and Estimate estimator estimate An estimator of a population parameter is a random variable that depends on the sample.
Parameter Estimation. Statistics Probability specified inferred Steam engine pump “prediction” “estimation”
Learning Theory Reza Shadmehr Distribution of the ML estimates of model parameters Signal dependent noise models.
Bayesian Brain Probabilistic Approaches to Neural Coding 1.1 A Probability Primer Bayesian Brain Probabilistic Approaches to Neural Coding 1.1 A Probability.
Week 21 Statistical Model A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution that produced.
Bayesian Perception.
Presentation : “ Maximum Likelihood Estimation” Presented By : Jesu Kiran Spurgen Date :
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
CS479/679 Pattern Recognition Dr. George Bebis
12. Principles of Parameter Estimation
Probability Theory and Parameter Estimation I
Ch3: Model Building through Regression
Maximum Likelihood Estimation
Synapses Signal is carried chemically across the synaptic cleft.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Confidence as Bayesian Probability: From Neural Origins to Behavior
Parametric Methods Berlin Chen, 2005 References:
12. Principles of Parameter Estimation
Presentation transcript:

BCS547 Neural Decoding

Nature of the problem In response to a stimulus with unknown orientation q, you observe a pattern of activity A. What can you say about q given A? Estimation theory: come up with a single value estimate from A Bayesian approach: recover P(q|A) (the posterior distribution)

Population Code Tuning Curves Pattern of activity (A)

Estimation theory

` A1 A2 A200 Decoder Trial 1 Encoder Decoder Trial 2 Encoder Decoder -100 100 20 40 60 80 Preferred retinal location A1 Decoder Trial 1 Encoder ` -100 100 20 40 60 80 Preferred retinal location A2 Decoder Trial 2 Encoder Decoder Trial 200 Encoder -100 100 20 40 60 80 Preferred retinal location A200

Estimation theory is a random variable. To determine the quality of this estimate we can compute its mean, E[ ], and its variance, s 2. If E[ ] =q, the estimate is said to be unbiased If s 2 is as small as possible, the estimate is said to be efficient

Estimation theory A common measure of decoding performance is the mean square error between the estimate and the true value This error can be decomposed as:

Efficient Estimators The smallest achievable variance for an unbiased estimator is known as the Cramer-Rao bound, sCR2. An efficient estimator is such that In general :

Fisher Information Fisher information is defined as: and it is equal to: where P(A| q) is the distribution of the neuronal noise.

Fisher Information

Fisher Information For one neuron with Poisson noise For n independent neurons : Large slope is good! The more neurons, the better! Small variance is good!

Fisher Information and Tuning Curves Fisher information is maximum where the slope is maximum This is consistent with adaptation experiments

Fisher Information In 1D, Fisher information decreases with the width of the tuning curves In 2D, Fisher information does not depend on the width of the tuning curve In 3D and above, Fisher information increases with the width of the tuning curves. ATTENTION: this is true for independent gaussian noise.

Ideal observer The discrimination threshold of an ideal observer, dq, is proportional to the variance of the Cramer-Rao Bound. In other words, an efficient estimator is an ideal observer.

An ideal observer is an observer that can recover all the Fisher information in the activity (easy link between Fisher information and behavioral performance) If all distributions are gaussian, Fisher information is the same as Shannon information.

Estimation theory Examples of decoders

Voting Methods Optimal Linear Estimator

Voting Methods Optimal Linear Estimator

Voting Methods Optimal Linear Estimator Center of Mass

Center of Mass/Population Vector The center of mass is optimal (unbiased and efficient) iff: The tuning curves are gaussian with a zero baseline, uniformly distributed and the noise follows a Poisson distribution In general, the center of mass has a large bias and a large variance

Voting Methods Optimal Linear Estimator Center of Mass Population Vector

Population Vector P aiPi q

Population Vector Typically, Population vector is not the optimal linear estimator.

Population Vector Population vector is optimal iff: The tuning curves are cosine, uniformly distributed and the noise follows a normal distribution with fixed variance In most cases, the population vector is biased and has a large variance The variance of the population vector estimate does not reflect Fisher information

Population Vector Population vector should NEVER be used to estimate Fisher Information Population vector should NEVER be used to estimate information content!!!!

Population Vector

Maximum Likelihood

Maximum Likelihood The estimate is the value of q that maximizes the likelihood P(A|q). Therefore, we seek such that:

Maximum Likelihood If the noise is gaussian and independent Therefore and the estimate is given by: Distance measure: Template matching

Maximum Likelihood

Gradient descent for ML To minimize the likelihood function with respect to q, one can use a gradient descent technique in which q is updated according to :

Gaussian noise with variance proportional to the mean If the noise is gaussian with variance proportional to the mean, the distance being minimized changes to: Data point with small variance are weighted more heavily

Poisson noise If the noise is Poisson then And :

ML and template matching Maximum likelihood is a template matching procedure BUT the metric used is not always the Euclidean distance, it depends on the noise distribution.

Bayesian approach We want to recover P(q|A). Using Bayes theorem, we have: prior distribution over q likelihood of q posterior distribution over q prior distribution over A

Bayesian approach What is the likelihood of q, P(A| q)? It is the distribution of the noise… It is the same distribution we used for maximum likelihood.

Bayesian approach The prior P(q) correspond to any knowledge we may have about q before we get to see any activity. Ex: Zhang et al.

Bayesian approach Once we have P(q|A), we can proceed in two different ways. We can keep this distribution for Bayesian inferences (as we would do in a Bayesian network) or we can make a decision about q. For instance, we can estimate q as being the value that maximizes P(q|A). This is known as the maximum a posteriori estimate (MAP). For flat prior, ML and MAP are equivalent.

Using the prior: Zhang et al For a time varying variable, one can use the distribution over the previous estimate as a prior for the next one. Prior Nasty but independent of Xt+1

Bayesian approach Limitations: the Bayesian approach and ML require a lot of data… Alternative: estimate P(q|A) directly using a nonlinear estimate.

Bayesian approach: logistic regression Example: Decoding finger movements in M1. On each trial, we observe 100 cells and we want to know which one of the 5 fingers is being moved. P(F5|A) 1 5 categories 1 2 3 4 5 … 100 input units 1 2 3 100 A

Bayesian approach: multinomial distributions Example: Decoding finger movements in M1. Each finger can take 3 mutually exclusive states: no movement, flexion, extension. Probability of no movement Probability of flexion Probability of extension Activity of the N M1 neurons W Digit 1 Wrist Softmax Digit 2 Digit 3 Digit 4 Digit 5

Decoding time varying signals s(t) r(t)

Decoding time varying signals

Decoding time varying signals

Decoding time varying signals s(t) r(t)