Presentation is loading. Please wait.

Presentation is loading. Please wait.

SHANNON LECTURE LIVING INFORMATION THEORY Toby Berger Cornell University IEEE ISIT – LAUSANNE 4 July 2002.

Similar presentations


Presentation on theme: "SHANNON LECTURE LIVING INFORMATION THEORY Toby Berger Cornell University IEEE ISIT – LAUSANNE 4 July 2002."— Presentation transcript:

1 SHANNON LECTURE LIVING INFORMATION THEORY Toby Berger Cornell University IEEE ISIT – LAUSANNE 4 July 2002

2 PART III: NEURAL COALITIONS AND MAIN THEOREM

3 DEFINITION OF A COALITION OF SENSORY NEURONS THE AXONS IN A COALITION OF SENSORY NEURONS FORM MANY OF THEIR SYNAPSES WITH OTHER NEURONS IN THE COALITION (INTER-COALITION FEEDBACK). SOMETIMES THE LOCAL CONNECTIVITY IS CLOSE TO 50%, AS OPPOSED TO ONLY BRAINWIDE. THE REMAINDER OF THE SYNAPSES IN WHICH THE AXONS OF A COALITION ARE INVOLVED ARE SPLIT ON AVERAGE ABOUT EQUALLY BETWEEN “LOWER” NEURONS AND “HIGHER” NEURONS. HERE, “LOWER” MEANS CLOSER TO THE SENSORY ORGAN AND “HIGHER” MEANS CLOSER TO THE ‘TOP BRAIN’.

4 BOTTOM-UP v. TOP-DOWN INFORMATION TRAVELING ALONG AXONS THAT GO FROM LOWER TO HIGHER IS CALLED BOTTOM-UP, OR FEEDFORWARD, INFORMATION. THAT WHICH TRAVELS ALONG AXONS THAT GO FROM HIGHER TO LOWER IS CALLED TOP-DOWN, OR FEEDBACK, INFORMATION.

5 Real neural spikes occur in continuous time. We introduce a time-discrete model with a time step of circa 2 ms (typical spike width). This captures the essence, provided the exact time of occurrence of a neural spike is not significant to an accuracy much greater than 1 ms.

6 “FIGURE 1” OF NEURAL INFORMATION THEORY SCHEMATIC OF A COALITION

7 PSPs

8 THE COALITION CHANNEL NEXT SLIDE SHOWS THAT A NEURAL COALITION IS A LINEAR (BUT THRESHOLDED) RANDOMIZED DYNAMICAL SYSTEM

9 COALITION EQUATION where =Signed weight of synapse (i,m) Random size of quantal contribution of synapse (i,m) to PSP of coalition member m at time k = 0 if synapse (i,m) fails at time k, 1 if not. (The fascinating phenomenon of “quantal failure” deserves a full lecture of its own – see Levy and Baxter, JCN 2002 and Proc. ISIT ’02, p. 18.) analogously defined Spiking threshold of neuron m at time k =

10 MAXIMUM INFORMATION RATE HYPOTHESIS The input (afferent) process {X(k))} that drives a coalition will have the property that it maximizes the mutual information rate between itself and the output (efferent) process {Y(k)} that it generates over all processes that lead to the same or smaller energy expenditure in the Y-neurons. Remark: Energy is expended both in the synapses in receiving and responding to afferent excitation and in the axons both to restore chemical concentrations during refractory periods following action potential generation and, to a lesser extent, to drive spikes down the axonal ‘transmission lines’.

11 JUSTIFICATIONS FOR THE MAX INFO RATE HYPOTHESIS Brain consumes 25% to 50% of metabolic energy under low physical exertion, so its computations should be energy-aware*. Inefficient to build and maintain a coalition and then expend more energy than necessary to transfer whatever information about its afferent process it conveys to its efferent process. The max info rate hypothesis leads to Markovianness of {X(k),Y(k)} which reduces latency by encapsulating memory. *L. Sokoloff (1989), “Circulation and energy metabolism of the brain,” in Basic Neurochemistry: Molecular, Cellular and Medical Aspects, 4 th ed., G. Siegel et al., Eds.

12 The Brain as a Markov Chain MAIN THEOREM: IF MAXIMUM INFORMATION RATE HYPTOTHESIS IS TRUE, THEN: {(X k,Y k } is a first-order (non-homog) Markov chain {Y k } is a first-order (non-homog) Markov chain {X k } is not necessarily Markovian PROOF: IF TIME PERMITS ( This is joint work with Yuzheng Ying with thanks to Claudio Weidmann.)

13 REMARKS: Max info rate hypothesis says the source {X(k)} is robustly “matched” to the channel’s transition matrix, P(y|x). If double matching prevails, as we suspect it does, then the failure rate parameterizes the rate-distortion function, and distortion is measured by a Weber-Fechner fidelity criterion of the form The Markovianness of the Main Theorem is essential to the brain’s low-latency processing of sensory information. Without it, bottom-up delay would accumulate too fast to allow for the number of coalitions needed to achieve the sophisticated distinctions of which the brain is capable.

14 From T. S. Lee and M. Nguyen, Dynamics of subjective contour formation in the early visual cortex. PNAS 98(4): , Temporal Dynamics of a V1 Neuron’s Response to Real and Illusory Contours

15 All problems in bio-IT are open, including those I may have implied are solved. What is memory? How is it physically stored and accessed? Can the max information rate hypothesis be proved by appealing to a least action principal in chemical statistical mechanics? (Perhaps this can be approached via the fact that the solution of multiphase chemical equilibrium problems is obtained by solving for the minimum of the Gibbs/Helmholtz Free Energy which, in turn, is mathematically fully analogous to calculation of a point on a rate-distortion function.) OPEN PROBLEMS

16 CLOSING CONJECTURE The most remarkable forms of information processing – those we call consciousness, intelligence and creativity – all are deeply reflexive. They are products of our self- awareness. It seems reasonable to conjecture that these forms of thinking emerge from the brain’s pervasive and intricate use of feedback.

17


Download ppt "SHANNON LECTURE LIVING INFORMATION THEORY Toby Berger Cornell University IEEE ISIT – LAUSANNE 4 July 2002."

Similar presentations


Ads by Google