ENERGY-EFFICIENT RECURSIVE ESTIMATION BY VARIABLE THRESHOLD NEURONS Toby Berger, Cornell & Uva Chip Levy, Uva Medical Galen Reeves, Cornell & UCB CoSyNe.

Slides:



Advertisements
Similar presentations
Detection Chia-Hsin Cheng. Wireless Access Tech. Lab. CCU Wireless Access Tech. Lab. 2 Outlines Detection Theory Simple Binary Hypothesis Tests Bayes.
Advertisements

Outline Transmitters (Chapters 3 and 4, Source Coding and Modulation) (week 1 and 2) Receivers (Chapter 5) (week 3 and 4) Received Signal Synchronization.
Design of the fast-pick area Based on Bartholdi & Hackman, Chpt. 7.
1. INTRODUCTION In order to transmit digital information over * bandpass channels, we have to transfer the information to a carrier wave of.appropriate.
ECE 8443 – Pattern Recognition LECTURE 05: MAXIMUM LIKELIHOOD ESTIMATION Objectives: Discrete Features Maximum Likelihood Resources: D.H.S: Chapter 3 (Part.
Information Theory EE322 Al-Sanie.
1 12. Principles of Parameter Estimation The purpose of this lecture is to illustrate the usefulness of the various concepts introduced and studied in.
Capacity of Wireless Channels
Quantization Prof. Siripong Potisuk.
Chapter 6 Information Theory
Copyright © Cengage Learning. All rights reserved. 9 Inferences Based on Two Samples.
Digital Data Transmission ECE 457 Spring Information Representation Communication systems convert information into a form suitable for transmission.
Continuous Random Variables and Probability Distributions
Inferences About Process Quality
Page 1 of 37 Density Evolution, Capacity Limits, and the "5k" Code Result (L. Schirber 11/22/11) The Density Evolution (DE) algorithm calculates a "threshold.
EE513 Audio Signals and Systems Statistical Pattern Classification Kevin D. Donohue Electrical and Computer Engineering University of Kentucky.
INFORMATION THEORY BYK.SWARAJA ASSOCIATE PROFESSOR MREC.
Lecture II Introduction to Digital Communications Following Lecture III next week: 4. … Matched Filtering ( … continued from L2) (ch. 2 – part 0 “ Notes.
Multiple Input Multiple Output
Dept. of EE, NDHU 1 Chapter Three Baseband Demodulation/Detection.
1 As we have seen in section 4 conditional probability density functions are useful to update the information about an event based on the knowledge about.
TELECOMMUNICATIONS Dr. Hugh Blanton ENTC 4307/ENTC 5307.
Estimating parameters in a statistical model Likelihood and Maximum likelihood estimation Bayesian point estimates Maximum a posteriori point.
Biological Modeling of Neural Networks Week 8 – Noisy input models: Barrage of spike arrivals Wulfram Gerstner EPFL, Lausanne, Switzerland 8.1 Variation.
1 7. Two Random Variables In many experiments, the observations are expressible not as a single quantity, but as a family of quantities. For example to.
Channel Capacity.
1 E. Fatemizadeh Statistical Pattern Recognition.
Biological Modeling of Neural Networks Week 8 – Noisy output models: Escape rate and soft threshold Wulfram Gerstner EPFL, Lausanne, Switzerland 8.1 Variation.
PROBABILITY AND STATISTICS FOR ENGINEERING Hossein Sameti Department of Computer Engineering Sharif University of Technology Principles of Parameter Estimation.
Chapter 4: Baseband Pulse Transmission Digital Communication Systems 2012 R.Sokullu1/46 CHAPTER 4 BASEBAND PULSE TRANSMISSION.
CHAPTER 5 SIGNAL SPACE ANALYSIS
Name Iterative Source- and Channel Decoding Speaker: Inga Trusova Advisor: Joachim Hagenauer.
BCS547 Neural Decoding. Population Code Tuning CurvesPattern of activity (r) Direction (deg) Activity
1 CONTEXT DEPENDENT CLASSIFICATION  Remember: Bayes rule  Here: The class to which a feature vector belongs depends on:  Its own value  The values.
BCS547 Neural Decoding.
CY1B2 Statistics1 (ii) Poisson distribution The Poisson distribution resembles the binomial distribution if the probability of an accident is very small.
1 Central Limit Theorem The theorem states that the sum of a large number of independent observations from the same distribution has, under certain general.
Timo O. Korhonen, HUT Communication Laboratory 1 Convolutional encoding u Convolutional codes are applied in applications that require good performance.
Expected values of discrete Random Variables. The function that maps S into S X in R and which is denoted by X(.) is called a random variable. The name.
1 EE571 PART 3 Random Processes Huseyin Bilgekul Eeng571 Probability and astochastic Processes Department of Electrical and Electronic Engineering Eastern.
Outline Transmitters (Chapters 3 and 4, Source Coding and Modulation) (week 1 and 2) Receivers (Chapter 5) (week 3 and 4) Received Signal Synchronization.
Baseband Receiver Receiver Design: Demodulation Matched Filter Correlator Receiver Detection Max. Likelihood Detector Probability of Error.
1 On the Channel Capacity of Wireless Fading Channels C. D. Charalambous and S. Z. Denic School of Information Technology and Engineering, University of.
Paging Area Optimization Based on Interval Estimation in Wireless Personal Communication Networks By Z. Lei, C. U. Saraydar and N. B. Mandayam.
Introduction to Estimation Theory: A Tutorial
Source Encoder Channel Encoder Noisy channel Source Decoder Channel Decoder Figure 1.1. A communication system: source and channel coding.
Uncertainty and Reliability Analysis D Nagesh Kumar, IISc Water Resources Planning and Management: M6L2 Stochastic Optimization.
1 Information Content Tristan L’Ecuyer. 2 Degrees of Freedom Using the expression for the state vector that minimizes the cost function it is relatively.
CPH Dr. Charnigo Chap. 11 Notes Figure 11.2 provides a diagram which shows, at a glance, what a neural network does. Inputs X 1, X 2,.., X P are.
Biological Modeling of Neural Networks Week 11 – Variability and Noise: Autocorrelation Wulfram Gerstner EPFL, Lausanne, Switzerland 11.1 Variation of.
Lecture 1.31 Criteria for optimal reception of radio signals.
12. Principles of Parameter Estimation
Advanced Wireless Networks
Cumulative distribution functions and expected values
Outline Introduction Signal, random variable, random process and spectra Analog modulation Analog to digital conversion Digital transmission through baseband.
General Properties of Radiation
7. Two Random Variables In many experiments, the observations are expressible not as a single quantity, but as a family of quantities. For example to record.
Statistical NLP: Lecture 4
Lav R. Varshney, Per Jesper Sjöström, Dmitri B. Chklovskii  Neuron 
CONTEXT DEPENDENT CLASSIFICATION
EE513 Audio Signals and Systems
Lav R. Varshney, Per Jesper Sjöström, Dmitri B. Chklovskii  Neuron 
11. Conditional Density Functions and Conditional Expected Values
7. Two Random Variables In many experiments, the observations are expressible not as a single quantity, but as a family of quantities. For example to record.
Improved Spread Spectrum: A New Modulation Technique for Robust Watermarking IEEE Trans. On Signal Processing, April 2003 Multimedia Security.
Information Theoretical Analysis of Digital Watermarking
11. Conditional Density Functions and Conditional Expected Values
5 pair of RVs.
12. Principles of Parameter Estimation
7. Two Random Variables In many experiments, the observations are expressible not as a single quantity, but as a family of quantities. For example to record.
Presentation transcript:

ENERGY-EFFICIENT RECURSIVE ESTIMATION BY VARIABLE THRESHOLD NEURONS Toby Berger, Cornell & Uva Chip Levy, Uva Medical Galen Reeves, Cornell & UCB CoSyNe Workshop on Info-Neuro, Park City, UT 2/25-26,2007

In time slot k, let X k denote the mean contribution to energy-efficient neuron N’s PSP in response to action potentials afferent to N’s synapses during the slot. The value of X k is assumed to be in 1-1 correspondence with a random variable,, that is of interest to the neurons in N’s efferent cohort and that remains effectively constant over any two successive slots. This implies that X also remains effectively constant over the two slots we shall consider in this paper. Moreover, for any r.v. V, I(X;V) = I( ;V), where I( ; ) is Shannon mutual information.

The process of PSP formation is random because of Poisson noise with variance proportional to the mean input signal strength, among other noises. In this preliminary investigation we simplistically assume that the overall noise in slot k, denoted N k, is additive, causing the net PSP in slot k due to both signal and noise to be of the form: Moreover, in our “toy problem” study, we assume that the noise also is (i) independent from slot to slot, (ii) independent of the signal strength, and (iii) uniformly distributed over [-1,1]. Although assumptions (ii) and (iii) are unrealistic, they’ve shamelessly been made purely for analytical convenience.

What’s more, we assume that X is uniformly distributed over the interval [-A,A], so that A may be thought of as the square root of the signal-to- noise ratio. Information Theory Remark: It is easy to show that a time-discrete,amplitude-continuous, memoryless channel that adds white uniform [-1,1] noise to its inputs to produce its outputs, and whose input amplitude magnitude is constrained not to exceed A, has Shannon capacity C = log(A+1). This capacity is achieved with an input distribution that concentrates its probability uniformly on the A+1 discrete points {A, A-2, A-4, …,-A}. However, distributing the input uniformly over the continuous input space [-A,A] achieves almost the full capacity when A >> 1.

Case 1: Only one slot. N encodes by generating an action potential in the slot if and only if Y exceeds a judiciously preset threshold, T. N does not set T at the median of Y even though that would maximize the information that the spike v. nonspike decision provides about. Rather, N uses a higher T than that in order to reduce the expected energy expenditure, E, down to where I/E is the quantity that gets maximized (cf., Levy and Baxter).

Fig. 2:with A = 5 and T 2 chosen a priori. Case 2: Two slots: This case divides into three subcases as follows, wherein Q i is a binary r.v. that equals 1 if Y i > T i and equals 0 otherwise, I = 1,2 : Case 2A: Maximize I(X;Q 1,Q 2 ) with T2 chosen before Q 1 is observed.

Fig. 3: I/E as a function of T 1 (Case 1)

Case 2B 1 : T 2 = T 2 (Q 1 ), i.e., second threshold set a posteriori We pick the initial threshold T 1, resulting in the binary output Q 1. Then we pick T 2 as a function of Q 1, say T 2 (Q 1 ), so as to achieve Case 2B 2 : Same as Case 2B 1 except we maximize bits/joule, where the maximum is over T 1, T 2 (Q 1 ), R and A. THIS IS AN IMPORTANT CASE because it addresses the task of maximizing bits per unit of energy via a slot-by-slot recursive strategy. Some graphical results follow, after which we introduce Case 2C

Fig 4: Maximum I/E as a function of T 1 (Case 2B)

Fig. 5: Optimal T 2 (0), T 2 (1) as a function of T 1 (Case 2B)

Case 2C: Refractoriness: This is similar to Case 2B except we apply a refractoriness constraint that prohibits N from firing in the second slot if it has just fired in the first slot by requiring that T 2 (Q 1 =1) be infinite. The refractoriness requirement changes the T1 that maximizes overall I/E, as shown in Figure 6. Figure 7 shows the optimum T 2 (0) as a function of T 1, with T 2 (1) = infinity.

Fig. 6: Maximum I/E as a function of T 1 (Case 2C)

Fig. 7: Optimal T 2 (0) as a function of T 1 (Case 2C)