Center for Computational Biology Department of Mathematical Sciences Montana State University Collaborators: Alexander Dimitrov John P. Miller Zane Aldworth.

Slides:



Advertisements
Similar presentations
Feature Selection as Relevant Information Encoding Naftali Tishby School of Computer Science and Engineering The Hebrew University, Jerusalem, Israel NIPS.
Advertisements

2806 Neural Computation Self-Organizing Maps Lecture Ari Visa.
Deep Learning and Neural Nets Spring 2015
Chapter 6 The Structural Risk Minimization Principle Junping Zhang Intelligent Information Processing Laboratory, Fudan University.
Center for Computational Biology Department of Mathematical Sciences Montana State University Collaborators: Alexander Dimitrov Tomas Gedeon John P. Miller.
Support Vector Machines
Continuation and Symmetry Breaking Bifurcation of the Information Distortion Function September 19, 2002 Albert E. Parker Complex Biological Systems Department.
1 Testing the Efficiency of Sensory Coding with Optimal Stimulus Ensembles C. K. Machens, T. Gollisch, O. Kolesnikova, and A.V.M. Herz Presented by Tomoki.
BCS547 Neural Encoding.
The University of Manchester Introducción al análisis del código neuronal con métodos de la teoría de la información Dr Marcelo A Montemurro
Chapter 6 Information Theory
On Systems with Limited Communication PhD Thesis Defense Jian Zou May 6, 2004.
For stimulus s, have estimated s est Bias: Cramer-Rao bound: Mean square error: Variance: Fisher information How good is our estimate? (ML is unbiased:
Mutual Information Mathematical Biology Seminar
Modelling and Control Issues Arising in the Quest for a Neural Decoder Computation, Control, and Biological Systems Conference VIII, July 30, 2003 Albert.
Fundamental limits in Information Theory Chapter 10 :
Center for Computational Biology Department of Mathematical Sciences Montana State University Collaborators: Alexander Dimitrov John P. Miller Zane Aldworth.
Efficient Statistical Pruning for Maximum Likelihood Decoding Radhika Gowaikar Babak Hassibi California Institute of Technology July 3, 2003.
Symmetry Breaking Bifurcation of the Distortion Problem Albert E. Parker Complex Biological Systems Department of Mathematical Sciences Center for Computational.
ENGR 351 Numerical Methods Instructor: Dr. L.R. Chevalier
We use Numerical continuation Bifurcation theory with symmetries to analyze a class of optimization problems of the form max F(q,  )=max (G(q)+  D(q)).
1 A hybrid particle swarm optimization algorithm for optimal task assignment in distributed system Peng-Yeng Yin and Pei-Pei Wang Department of Information.
Symmetry breaking clusters when deciphering the neural code September 12, 2005 Albert E. Parker Department of Mathematical Sciences Center for Computational.
Center for Computational Biology Department of Mathematical Sciences Montana State University Collaborators: Alexander Dimitrov John P. Miller Zane Aldworth.
Expectation Maximization Algorithm
A Bifurcation Theoretical Approach to the Solving the Neural Coding Problem June 28 Albert E. Parker Complex Biological Systems Department of Mathematical.
Collaborators: Tomas Gedeon Alexander Dimitrov John P. Miller Zane Aldworth Information Theory and Neural Coding PhD Oral Examination November 29, 2001.
We use Numerical continuation Bifurcation theory with symmetries to analyze a class of optimization problems of the form max F(q,  )=max (G(q)+  D(q)).
Neural Coding Through The Ages February 1, 2002 Albert E. Parker Complex Biological Systems Department of Mathematical Sciences Center for Computational.
Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 1 Computational Architectures in.
Linear Codes for Distributed Source Coding: Reconstruction of a Function of the Sources -D. Krithivasan and S. Sandeep Pradhan -University of Michigan,
Phase Transitions in the Information Distortion NIPS 2003 workshop on Information Theory and Learning: The Bottleneck and Distortion Approach December.
Noise, Information Theory, and Entropy
Application of Methods of Queuing Theory to Scheduling in GRID A Queuing Theory-based mathematical model is presented, and an explicit form of the optimal.
Noise, Information Theory, and Entropy
Basic Concepts in Information Theory
INFORMATION THEORY BYK.SWARAJA ASSOCIATE PROFESSOR MREC.
Tahereh Toosi IPM. Recap 2 [Churchland and Abbott, 2012]
On comparison of different approaches to the stability radius calculation Olga Karelkina Department of Mathematics University of Turku MCDM 2011.
1 Hybrid methods for solving large-scale parameter estimation problems Carlos A. Quintero 1 Miguel Argáez 1 Hector Klie 2 Leticia Velázquez 1 Mary Wheeler.
FDA- A scalable evolutionary algorithm for the optimization of ADFs By Hossein Momeni.
1 / 41 Inference and Computation with Population Codes 13 November 2012 Inference and Computation with Population Codes Alexandre Pouget, Peter Dayan,
Generative Topographic Mapping by Deterministic Annealing Jong Youl Choi, Judy Qiu, Marlon Pierce, and Geoffrey Fox School of Informatics and Computing.
Extremal Problems of Information Combining Alexei Ashikhmin  Information Combining: formulation of the problem  Mutual Information Function for the Single.
Projects: 1.Predictive coding in balanced spiking networks (Erwan Ledoux). 2.Using Canonical Correlation Analysis (CCA) to analyse neural data (David Schulz).
Derivative Free Optimization G.Anuradha. Contents Genetic Algorithm Simulated Annealing Random search method Downhill simplex method.
Exact and heuristics algorithms
The Information Bottleneck Method clusters the response space, Y, into a much smaller space, T. In order to informatively cluster the response space, the.
Coding Theory Efficient and Reliable Transfer of Information
Vaida Bartkutė, Leonidas Sakalauskas
July 11, 2006Bayesian Inference and Maximum Entropy Probing the covariance matrix Kenneth M. Hanson T-16, Nuclear Physics; Theoretical Division Los.
1 Source Coding and Compression Dr.-Ing. Khaled Shawky Hassan Room: C3-222, ext: 1204, Lecture 10 Rate-Distortion.
1 Lecture 7 System Models Attributes of a man-made system. Concerns in the design of a distributed system Communication channels Entropy and mutual information.
Final Exam Review CS479/679 Pattern Recognition Dr. George Bebis 1.
6. Population Codes Presented by Rhee, Je-Keun © 2008, SNU Biointelligence Lab,
1 On the Channel Capacity of Wireless Fading Channels C. D. Charalambous and S. Z. Denic School of Information Technology and Engineering, University of.
Basic Concepts of Information Theory A measure of uncertainty. Entropy. 1.
Rate Distortion Theory. Introduction The description of an arbitrary real number requires an infinite number of bits, so a finite representation of a.
Chance Constrained Robust Energy Efficiency in Cognitive Radio Networks with Channel Uncertainty Yongjun Xu and Xiaohui Zhao College of Communication Engineering,
UNIT I. Entropy and Uncertainty Entropy is the irreducible complexity below which a signal cannot be compressed. Entropy is the irreducible complexity.
Introduction to genetic algorithm
Semi-Supervised Clustering
Introduction to Information theory
The Doctoral Dissertation
Random Noise in Seismic Data: Types, Origins, Estimation, and Removal
Akio Utsugi National Institute of Bioscience and Human-technology,
SMEM Algorithm for Mixture Models
What are optimization methods?
Watermarking with Side Information
Stochastic Methods.
Presentation transcript:

Center for Computational Biology Department of Mathematical Sciences Montana State University Collaborators: Alexander Dimitrov John P. Miller Zane Aldworth Tomas Gedeon Brendan Mumey Neural Coding and Decoding Albert E. Parker

Problem: How does neural ensemble activity represent information about sensory stimuli? Our Approach: Build a model using Information Theory My Research: Probability Theory Use the model to determine a coding scheme My Research: Numerical Optimization Techniques

Neural Coding and Decoding. Goal: Determine a coding scheme: How does neural ensemble activity represent information about sensory stimuli? Demands: An animal needs to recognize the same object on repeated exposures. Coding has to be deterministic at this level. The code must deal with uncertainties introduced by the environment and neural architecture. Coding is by necessity stochastic at this finer scale. Major Problem: The search for a coding scheme requires large amounts of data

How to determine a coding scheme? Idea: Model a part of a neural system as a communication channel using Information Theory. This model enables us to: Meet the demands of a coding scheme: oDefine a coding scheme as a relation between stimulus and neural response classes. oConstruct a coding scheme that is stochastic on the finer scale yet almost deterministic on the classes. Deal with the major problem: oUse whatever quantity of data is available to construct coarse but optimally informative approximations of the coding scheme. oRefine the coding scheme as more data becomes available. Investigate the cricket cercal sensory system.

Y X stimulus sequences response sequences stimulus/response sequence pairs distinguishable classes of stimulus/response pairs Stimulus and Response Classes

Information Theoretic Quantities A quantizer or encoder, Q, relates the environmental stimulus, X, to the neural response, Y, through a process called quantization. In general, Q is a stochastic map The Reproduction space Y is a quantization of X. This can be repeated: Let Y f be a reproduction of Y. So there is a quantizer Use Mutual information to measure the degree of dependence between X and Y f. Use Conditional Entropy to measure the self-information of Y f given Y

The Model Problem: To determine a coding scheme between X and Y requires large amounts of data Idea: Determine the coding scheme between X and Y f, a squashing (reproduction) of Y, such that: Y f preserves as much information (mutual information) with X as possible and the self-information (entropy) of Y f |Y is maximized. That is, we are searching for an optimal mapping (quantizer): that satisfies these conditions. Justification: Jayne's maximum entropy principle, which states that of all the quantizers that satisfy a given set of constraints, choose the one that maximizes the entropy.

Equivalent Optimization Problems  Maximum entropy: maximize F ( q(y f |y) ) = H(Y f |Y) constrained by I(X;Y f )  I o I o determines the informativeness of the reproduction.  Deterministic annealing (Rose, ’98): maximize F ( q(y f |y) ) = H(Y f |Y) -  D I (Y,Y f ). Small  favor maximum entropy, large  - minimum D I.  Simplex Algorithm: maximize I(X,Y f ) over vertices of constraint space  Implicit solution:

? ?

Signal Nervous system Communication channel Modeling the cricket cercal sensory system as a communication channel

Wind Stimulus and Neural Response in the cricket cercal system Neural Responses (over a 30 minute recording) caused by white noise wind stimulus. T, ms Neural Responses (these are all doublets) for a 12 ms window Some of the air current stimuli preceding one of the neural responses Time in ms. A t T=0, the first spike occurs X Y

YfYf Y Quantization: A quantizer is any map f: Y -> Y f from Y to a reproduction space Y f with finitely many elements. Quantizers can be deterministic or refined Y probabilistic

Applying the algorithm to cricket sensory data. Y YfYf YfYf

Conclusions We model a part of the neural system as a communication channel. define a coding scheme through relations between classes of stimulus/response pairs. -Coding is probabilistic on the individual elements of X and Y. -Coding is almost deterministic on the stimulus/response classes. To recover such a coding scheme, we propose a new method to quantify neural spike trains. -Quantize the response patterns to a small finite space (Y f ). -Use information theoretic measures to determine optimal quantizer for a fixed reproduction size. -Refine the coding scheme by increasing the reproduction size. present preliminary results with cricket sensory data.

Random clusters Linear relation Application to synthetic data.

The cricket case: dealing with complex stimuli. D I =I(X,Y)-I(X,Y N ) cannot be estimated directly for rich stimulus sets. Use an upper bound. Start by D I =H(X)-H(X|Y)- ( H(X) - H(X|Y N ) ). Only H(X|Y N ) depends on the quantizer so use D eff = H(X|Y N ). Use H(X|y N ) is bounder by a Gaussian: where is the conditional covariance of the stimulus. This gives an upper bound of D eff. Minimize the upper bound. Better model = tighter bound. explicitly as a function of the quantizer.

Directions Properties of the information distortion function. –Bifurcation structure (conjecture: there are only pitchfork bifurcations). –Classes of problems with exact solution. –Structure of the set of optima (conjecture: global minima can be found at vertices of the domain). Algorithmic development –Improved numeric optimization schemes. –Convergence properties. Neural coding model. –Better models for rich stimulus sets. Better upper bounds. –Explicit dependence on the size of the dataset (possibly for each model). Applications to non-neural problems.