Presented by Hanan Shteingart Physiology A – 2009, ICNC May 2009

Slides:



Advertisements
Similar presentations
Presentation of Illusory Haptic Shape via Contact Location Trajectory on the Fingerpad Hanifa Dostmohamed and Vincent Hayward Haptics Laboratory, Center.
Advertisements

Chapter 2: Marr’s theory of vision. Cognitive Science  José Luis Bermúdez / Cambridge University Press 2010 Overview Introduce Marr’s distinction between.
Bayesian Robust Principal Component Analysis Presenter: Raghu Ranganathan ECE / CMR Tennessee Technological University January 21, 2011 Reading Group (Xinghao.
CS292 Computational Vision and Language Visual Features - Colour and Texture.
Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 1 Computational Architectures in.
Statistical Color Models (SCM) Kyungnam Kim. Contents Introduction Trivariate Gaussian model Chromaticity models –Fixed planar chromaticity models –Zhu.
Careers for Psychology and Neuroscience Majors Oct. 19th5-7pm in SU 300 Ballroom B.
Computer vision.
Studying Visual Attention with the Visual Search Paradigm Marc Pomplun Department of Computer Science University of Massachusetts at Boston
Cristian Urs and Ben Riveira. Introduction The article we chose focuses on improving the performance of Genetic Algorithms by: Use of predictive models.
Perception Introduction Pattern Recognition Image Formation
Lecture 2b Readings: Kandell Schwartz et al Ch 27 Wolfe et al Chs 3 and 4.
Yaomin Jin Design of Experiments Morris Method.
How natural scenes might shape neural machinery for computing shape from texture? Qiaochu Li (Blaine) Advisor: Tai Sing Lee.
December 9, 2014Computer Vision Lecture 23: Motion Analysis 1 Now we will talk about… Motion Analysis.
Spatial Smoothing and Multiple Comparisons Correction for Dummies Alexa Morcom, Matthew Brett Acknowledgements.
From cortical anisotropy to failures of 3-D shape constancy Qasim Zaidi Elias H. Cohen State University of New York College of Optometry.
Image features and properties. Image content representation The simplest representation of an image pattern is to list image pixels, one after the other.
Introduction to emulators Tony O’Hagan University of Sheffield.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology Advisor : Dr. Hsu Graduate : Yu Cheng Chen Author: Michael.
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 1: INTRODUCTION.
1 Nonlinear models for Natural Image Statistics Urs Köster & Aapo Hyvärinen University of Helsinki.
Module 9: Choosing the Sampling Strategy
CSC2535: Computation in Neural Networks Lecture 11 Extracting coherent properties by maximizing mutual information across space or time Geoffrey Hinton.
Vision.
Lecture 07 13/12/2011 Shai Avidan הבהרה: החומר המחייב הוא החומר הנלמד בכיתה ולא זה המופיע / לא מופיע במצגת.
Digital Image Processing Lecture 10: Image Restoration
Recognizing Deformable Shapes
Common Classification Tasks
Neural Correlates of Shape from Shading
Attention Narrows Position Tuning of Population Responses in V1
Volume 60, Issue 4, Pages (November 2008)
MODELING MST OPTIC FLOW RESPONSES
Computer Vision Lecture 16: Texture II
Nonlinear processing in LGN neurons
Xaq Pitkow, Dora E. Angelaki  Neuron 
A Sparse Object Coding Scheme in Area V4
Satoru Suzuki, Marcia Grabowecky  Neuron 
Christopher C. Pack, Richard T. Born, Margaret S. Livingstone  Neuron 
Volume 54, Issue 6, Pages (June 2007)
MODELING MST OPTIC FLOW RESPONSES
Volume 66, Issue 6, Pages (June 2010)
Perception.
Probabilistic Map Based Localization
Monte Carlo I Previous lecture Analytical illumination formula
Hindsight Bias Tendency to believe, after learning an outcome, that one would have foreseen it. “I knew.
Chapter 5: Producing Data
A Channel for 3D Environmental Shape in Anterior Inferotemporal Cortex
Siavash Vaziri, Charles E. Connor  Current Biology 
Attentional Modulations Related to Spatial Gating but Not to Allocation of Limited Resources in Primate V1  Yuzhi Chen, Eyal Seidemann  Neuron  Volume.
Satoru Suzuki, Marcia Grabowecky  Neuron 
A Map for Horizontal Disparity in Monkey V2
Jack Grinband, Joy Hirsch, Vincent P. Ferrera  Neuron 
Nonparametric Hypothesis Tests for Dependency Structures
Object recognition in the inferior temporal cortex
Parametric Methods Berlin Chen, 2005 References:
Medial Axis Shape Coding in Macaque Inferotemporal Cortex
CSE 185 Introduction to Computer Vision
Volume 89, Issue 6, Pages (March 2016)
Perceptual learning Nisheeth 15th February 2019.
Stephen V. David, Benjamin Y. Hayden, James A. Mazer, Jack L. Gallant 
Receptive Fields of Disparity-Tuned Simple Cells in Macaque V1
Encoding of Stimulus Probability in Macaque Inferior Temporal Cortex
by Khaled Nasr, Pooja Viswanathan, and Andreas Nieder
Dynamic Shape Synthesis in Posterior Inferotemporal Cortex
Gregor Rainer, Earl K Miller  Neuron 
Higher-Order Figure Discrimination in Fly and Human Vision
Memory-Based Learning Instance-Based Learning K-Nearest Neighbor
Perception.
Conserved Sequence Processing in Primate Frontal Cortex
Presentation transcript:

Presented by Hanan Shteingart Physiology A – 2009, ICNC May 2009 Macaca Mulatta Presented by Hanan Shteingart Physiology A – 2009, ICNC May 2009

Something to Wake You Up Go to icnc.wordpress.com to make your own dragon

How does the brain (IT) codes 3D objects? Main Question How does the brain (IT) codes 3D objects? ? macaque monkey inferotemporal cortex (IT)

Abstract Complex shape codes - focused on 2D. 3D requires : higher-D coding (more info) computation (harder to decipher) The Good News: evidence for an explicit neural code. The Bad News: very complicated paper! (fancy object creation, evolutionary stimulus strategy, linear/nonlinear models with and crazy statistical tests)

Introduction object boundary fragments (features) integration into explicit signal component level shape (or holistic via learning) V2 IT V4 ~isomorphic (106 pixels) unwieldy and unstable  not useful for object perception 2D studies

The Hypothesis H0: IT neurons encode three-dimensional spatial configurations of surface fragments. H1: Complex shape perception is based primarily on two-dimensional image processing. Previous studies differential responses across a small number of 3D shapes or tuning along a single depth-related dimension Question: how does H0 and H1 different? The paper says that “However, these results and the multiple-views hypothesis itself are also compatible with three-dimensional representation29 (see Discussion).”

The 3D Decoding Problem Which 3D shape factors are associated with neural responses? 3D  complex, multi tuning properties  Large stimulus set wide range of 3D elements combined in many different ways A conventional random or systematic (grid-based) approach can never produce sufficiently dense combinatorial sampling. This has not been attempted before.

Decoding Solution Evolutionary stimulus strategy Two advantages: Focused Variant This evolutionary stimulus strategy made it possible for the first time to test the three-dimensional configural coding hypothesis at the neural level. The second advantage is not clear (page 4 left) Focused around the response range of the neuron (far less time) shape characteristics encoded by the neuron are repeatedly varied and recombined with other shape features

Evolutionary Algorithms Population Initialization Fitness Scoring Selection Variation Termination Initial generation of 50 random 3D shapes Averaged response = probability to reproduce Descended morphed, either locally or globally

Evolutionary Stimulus Strategy

Response Model Feature Space: Stimuli were characterized by 7 surface fragments: X,Y,Z XY, YZ angles, max/min curvature Two Gaussian function amplitude at the stimulus point closest to the Gaussian peak + nonlinear interactions.

Results Overview Strong cross-prediction of responses Model order = 2 1 Strong cross-prediction of responses Model order = 2 Most fragments are outside plane (3D) 3D shape tuning independent of lighting, position, size and depth. wide range of tuned configurations. 3D representation in IT is not holistic Convexity bias 2 geometric similarities between runs are imposed by the neuron itself and thus serve to confirm the generality of the tuning model The neuron remained responsive when either disparity or shading cues for depth were present, but failed to respond when both were eliminated

Tuning Functions Consistency: Non consistence: Lighting Depth, Position, Size Non consistence: orientation Light angle depth position rotation size xy plane

3D Verification The modulation index is the response difference between the high- and low response stimuli, normalized by maximum response across all conditions. three-dimensional shape tuning was largely independent of lighting direction, stimulus position, stimulus size and stimulus depth. Separability is represented here by the fraction of response variance (r2) explainable by a matrix product between separate tuning functions for shape and shading, depth, position or size ??????????????

Surface-Fragment Configurations Tuning models spanned a wide range of surface-fragment configurations 3D shape representation in IT is not generally holistic (model covers partially the whole shape)

Convexity Bias stimuli model Tuning was markedly biased in the curvature domain toward high values, especially on the convex end of the scale. spatial curvature

Discussion Convexity bias may reflect functional importance (emphasize pointy parts) Substantial fraction of IT neurons followed H0: they were simultaneously tuned for 3D shape Neurons tuned for multiple regions in this domain consistent with classic theories of configural shape representation (‘geons’) with 2 differs: No rotating of reference frame Multi-part configuration for single neuron H0 = neurons encode the three-dimensional shape, three-dimensional orientation and relative three-dimensional position of object parts

The Need for 3D Encoding? Why would the brain explicitly represent complex 3D object shape, considering the computational expense of inferring 3D structure from the two-dimensional retinal image and the higher neural tuning dimensionality required? Speculation: Representation of 3D object structure supports other aspects of object vision beyond identification (e.g. usage).

Configural coding of three-dimensional object structure Henry Moore’s ‘‘Sheep Piece’’ (1971–1972)

Question to Eli “The average neural response to each stimulus determined the probability with which it produced morphed descendants in subsequent stimulus generations” (pg 4) but in methods it says: a typical second generation would contain ten stimuli generated de novo, four descendants of stimuli in the highest response bin, four descendants from the second highest bin, etc. [equal probabilities] Fig 3b. Response consistency was measured by separability of tuning for shape? Why is this consistency? It’s more independence of variables or something.