Low Level Visual Processing. Information Maximization in the Retina Hypothesis: ganglion cells try to transmit as much information as possible about the.

Slides:



Advertisements
Similar presentations
What do we know about Primary Visual Cortex (V1)
Advertisements

Shapelets Correlated with Surface Normals Produce Surfaces Peter Kovesi School of Computer Science & Software Engineering The University of Western Australia.
Computer Vision Lecture 7: The Fourier Transform
Chapter 2.
Gabor Filter: A model of visual processing in primary visual cortex (V1) Presented by: CHEN Wei (Rosary) Supervisor: Dr. Richard So.
A Neural Model for Detecting and Labeling Motion Patterns in Image Sequences Marc Pomplun 1 Julio Martinez-Trujillo 2 Yueju Liu 2 Evgueni Simine 2 John.
E ffi cient Coding: From Retina Ganglion Cells To V2 Cells Honghao Shan Garrison W. Cottrell The Temporal Dynamics of Learning Center Gary's Unbelievable.
DFT/FFT and Wavelets ● Additive Synthesis demonstration (wave addition) ● Standard Definitions ● Computing the DFT and FFT ● Sine and cosine wave multiplication.
黃文中 Preview 2 3 The Saliency Map is a topographically arranged map that represents visual saliency of a corresponding visual scene. 4.
Noise. Noise is like a weed. Just as a weed is a plant where you do not wish it to be, noise is a signal where you do not wish it to be. The noise signal.
Application of Statistical Techniques to Neural Data Analysis Aniket Kaloti 03/07/2006.
CSCE 641 Computer Graphics: Image Sampling and Reconstruction Jinxiang Chai.
Spike-triggering stimulus features stimulus X(t) multidimensional decision function spike output Y(t) x1x1 x2x2 x3x3 f1f1 f2f2 f3f3 Functional models of.
MSU CSE 803 Stockman Fall 2009 Vectors [and more on masks] Vector space theory applies directly to several image processing/representation problems.
Levels in Computational Neuroscience Reasonably good understanding (for our purposes!) Poor understanding Poorer understanding Very poorer understanding.
Last week…. Fourier Analysis Re-writing a signal as a sum of sines and cosines.
1 MSU CSE 803 Fall 2014 Vectors [and more on masks] Vector space theory applies directly to several image processing/representation problems.
Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 1 Computational Architectures in.
ELEC 303 – Random Signals Lecture 21 – Random processes
October 8, 2013Computer Vision Lecture 11: The Hough Transform 1 Fitting Curve Models to Edges Most contours can be well described by combining several.
Spatial-based Enhancements Lecture 3 prepared by R. Lathrop 10/99 updated 10/03 ERDAS Field Guide 6th Ed. Ch 5: ;
Another viewpoint: V1 cells are spatial frequency filters
Low Level Visual Processing. Information Maximization in the Retina Hypothesis: ganglion cells try to transmit as much information as possible about the.
1 Let g(t) be periodic; period = T o. Fundamental frequency = f o = 1/ T o Hz or  o = 2  / T o rad/sec. Harmonics =n f o, n =2,3 4,... Trigonometric.
1 Computational Vision CSCI 363, Fall 2012 Lecture 26 Review for Exam 2.
Active Vision Key points: Acting to obtain information Eye movements Depth from motion parallax Extracting motion information from a spatio-temporal pattern.
Module 2 SPECTRAL ANALYSIS OF COMMUNICATION SIGNAL.
Seeram Chapter 7: Image Reconstruction
1 Computational Vision CSCI 363, Fall 2012 Lecture 31 Heading Models.
1 Computational Vision CSCI 363, Fall 2012 Lecture 20 Stereo, Motion.
CSC321: Neural Networks Lecture 13: Learning without a teacher: Autoencoders and Principal Components Analysis Geoffrey Hinton.
Projects: 1.Predictive coding in balanced spiking networks (Erwan Ledoux). 2.Using Canonical Correlation Analysis (CCA) to analyse neural data (David Schulz).
Chapter 3: Neural Processing and Perception. Neural Processing and Perception Neural processing is the interaction of signals in many neurons.
1 Computational Vision CSCI 363, Fall 2012 Lecture 21 Motion II.
Chapter 2. Signals and Linear Systems
1 Computational Vision CSCI 363, Fall 2012 Lecture 5 The Retina.
Human vision Jitendra Malik U.C. Berkeley. Visual Areas.
Wireless Multiple Access Schemes in a Class of Frequency Selective Channels with Uncertain Channel State Information Christopher Steger February 2, 2004.
BCS547 Neural Decoding. Population Code Tuning CurvesPattern of activity (r) Direction (deg) Activity
BCS547 Neural Decoding.
Chapter 11 Filter Design 11.1 Introduction 11.2 Lowpass Filters
7- 1 Chapter 7: Fourier Analysis Fourier analysis = Series + Transform ◎ Fourier Series -- A periodic (T) function f(x) can be written as the sum of sines.
1 Computational Vision CSCI 363, Fall 2012 Lecture 24 Computing Motion.
Autonomous Robots Vision © Manfred Huber 2014.
CSC2515: Lecture 7 (post) Independent Components Analysis, and Autoencoders Geoffrey Hinton.
Visual Computation I. Physiological Foundations
October 16, 2014Computer Vision Lecture 12: Image Segmentation II 1 Hough Transform The Hough transform is a very general technique for feature detection.
CSC321 Lecture 5 Applying backpropagation to shape recognition Geoffrey Hinton.
Understanding early visual coding from information theory By Li Zhaoping Lecture at EU advanced course in computational neuroscience, Arcachon, France,
Computational Vision CSCI 363, Fall 2012 Lecture 22 Motion III
Baseband Receiver Receiver Design: Demodulation Matched Filter Correlator Receiver Detection Max. Likelihood Detector Probability of Error.
6. Population Codes Presented by Rhee, Je-Keun © 2008, SNU Biointelligence Lab,
1 Computational Vision CSCI 363, Fall 2012 Lecture 16 Stereopsis.
Sensation & Perception. Motion Vision I: Basic Motion Vision.
Signal Analyzers. Introduction In the first 14 chapters we discussed measurement techniques in the time domain, that is, measurement of parameters that.
Understanding Psychophysics: Spatial Frequency & Contrast
The Frequency Domain Digital Image Processing – Chapter 8.
CSC2535: Computation in Neural Networks Lecture 11 Extracting coherent properties by maximizing mutual information across space or time Geoffrey Hinton.
Lecture 1.30 Structure of the optimal receiver deterministic signals.
Brodmann’s Areas. Brodmann’s Areas The Primary Visual Cortex Hubel and Weisel discovered simple, complex and hypercomplex cells in the striate.
Fitting Curve Models to Edges
Early Processing in Biological Vision
Computer Vision Lecture 9: Edge Detection II
Spatial Vision (continued)
Volume 21, Issue 19, Pages (October 2011)
Orientation tuning: strongest response to one orientation
Volume 55, Issue 3, Pages (August 2007)
Motion-Based Analysis of Spatial Patterns by the Human Visual System
Binocular Disparity and the Perception of Depth
8.6 Autocorrelation instrument, mathematical definition, and properties autocorrelation and Fourier transforms cosine and sine waves sum of cosines Johnson.
Presentation transcript:

Low Level Visual Processing

Information Maximization in the Retina Hypothesis: ganglion cells try to transmit as much information as possible about the image. What kind of receptive field maximizes information transfer?

Information Maximization in the Retina In this particular context, information is maximized for a factorial code: For a factorial code, the mutual information is 0 (there are no redundancies):

Information Maximization in the Retina Independence is hard to achieve. Instead, we can look for a code which decorrelates the activity of the ganglion cells. This is a lot easier because decorrelation can be achieved with a simple linear transformation.

Information Maximization in the Retina We assume that ganglion cells are linear: The goal is to find a receptive field profile, D s (x), for which the ganglion cells are decorrelated (i.e., a whitening filter).

Information Maximization in the Retina Correlations are captured by the crosscorrelogram (all signals are assumed to be zero mean): The crosscorrelogram is also a convolution

Fourier Transform The Fourier transform of a convolution is equal to the product of the individual spectra: The spectrum of a Dirac function is flat.

Information Maximization in the Retina To decorrelate, we need to ensure that the crosscorrelogram is a Dirac function, i.e., its Fourier transform should be as flat as possible.

Information Maximization in the Retina

If we assume that the retina adds noise on top of the signal to be transmitted to the brain, the previous filter is a bad idea because it amplifies the noise. Solution: use a noise filter first, :

Information Maximization in the Retina

The shape of the whitening filter depends on the noise level. For high contrast/low noise: bandpass filter. Center-surround RF. For low contrast/high noise: low pass filter. Gaussian RF

Information Maximization in the Retina j +

temporal frequency (Hz) temporal frequency (H

Information Maximization beyond the Retina The bottleneck argument can only work once… The whitening filter only decorrelates. To find independent components, use ICA: predicts oriented filter Use other constrained beside infomax, such as sparseness.

Center Surround Receptive Fields The center surround receptive fields are decent edge detectors +

Center Surround Receptive Fields

Feature extraction: Energy Filters

2D Fourier Transform Frequency Orientation

2D Fourier Transform

Motion Energy Filters Space Time

Motion Energy Filters In a space time diagram 1 st order motion shows up as diagonal lines. The slope of the line indicates the velocity A space-time Fourier transform can therefore recover the speed of motion

Motion Energy Filters

1 st order motion Time Space

Motion Energy Filters 2 nd order motion

Motion Energy Filters In a space time diagram, 2 nd order motion does not show up as a diagonal line… Methods based on linear filtering of the image followed by a nonlinearity cannot work You need to apply a nonlinearity to the image first

Motion Energy Filters A Fourier transform returns a set of complex coefficients:

Motion Energy Filters The power spectrum is given by

Motion Energy Filters

cos sin + ajaj bjbj cjcj x2x2 x2x2 I

Motion Energy Filters cos sin + ajaj bjbj cjcj x2x2 x2x2 I

tt xx

Motion Energy Filters Therefore, taking a Fourier transform in space time is sufficient to compute motion. To compute velocity, just look at where the power is and compute the angle. Better still, use the Fourier spectrum as your observation and design an optimal estimator of velocity (tricky because the noise is poorly defined)…

Motion Energy Filters How do you compute a Fourier transform with neurons? Use neurons with spatio- temporal filters looking like oriented sine and cosine functions. Problem: the receptive fields are non local and would have a hard time dealing with multiple objects in space and multiple events in time…

Motion Energy Filters Solution: use oriented Gabor-like filters or causal version of Gabor-like filters. To recover the spectrum, take quadrature pairs, square them and add them: this is what is called an Energy Filter.

Motion Energy Filters x2x2 x2x2 +

From V1 to MT V1 cells are tuned to velocity but they are also tuned to spatial and temporal frequencies tt xx

From V1 to MT MT cells are tuned to velocity across a wide range of spatial and temporal frequencies tt xx

MT Cells

Pooling across Filters Motion opponency: it is not possible to perceive transparent motion within the same spatial bandwidth. This suggests that the neural read out mechanism for speed computes the difference between filters tuned to different spatial frequencies within the same temporal bandwidth.

Pooling across Filters + Flicker

Energy Filters For second order motion, apply a nonlinearity to the image and then run a motion energy filter.

Motion Processing: Bayesian Approach The energy filter approach is not the only game in town… Bayesian integration provides a better account of psychophysical results

Energy Filters: Generalization The same technique can be used to compute orientation, disparity, … etc.

Energy Filters: Generalization The case of stereopsis: constant disparity correspond to oriented line in righ/left RF diagram.

Energy Filters: Generalization

Orientation Selectivity At first sight: a simple, if not downright stupid, problem. Use an orientation energy filter The only challenge: finding out exactly how the brain does it… Two classes of models: feedforward and lateral connection models.

Orientation Selectivity: Feedforward (Hubel and Wiesel)  LGN CTX

Orientation Selectivity: The Lateral Connection Model  LGN CTX - +

Orientation Selectivity: Feedforward (Hubel and Wiesel) -- +

Orientation Selectivity Take a quadrature pair, rectifiy and square their outputs, sum and you get a complex cell tuned to orientation.

Orientation Selectivity Most people think this model is wrong, yet the evidence in its favor are overwhelming. –No further tuning over time –Aspect ratio consistent with this model –LGN input to layer 4 cells as tuned as output –LGN/Cortex connectivity as predicted by feedforward model Why are there lateral connections?