Standard Brain Model for Vision

Slides:



Advertisements
Similar presentations
Object Recognition from Local Scale-Invariant Features David G. Lowe Presented by Ashley L. Kapron.
Advertisements

Human Detection Phanindra Varma. Detection -- Overview  Human detection in static images is based on the HOG (Histogram of Oriented Gradients) encoding.
CSCE 643 Computer Vision: Template Matching, Image Pyramids and Denoising Jinxiang Chai.
Group Meeting Presented by Wyman 10/14/2006
Classification Classification Examples
Image classification Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing them?
Object recognition and scene “understanding”
Object Recognition with Features Inspired by Visual Cortex T. Serre, L. Wolf, T. Poggio Presented by Andrew C. Gallagher Jan. 25, 2007.
CSE 473/573 Computer Vision and Image Processing (CVIP)
Data Mining Classification: Alternative Techniques
Support Vector Machines
Image Analysis Phases Image pre-processing –Noise suppression, linear and non-linear filters, deconvolution, etc. Image segmentation –Detection of objects.
HMAX Models Architecture Jim Mutch March 31, 2010.
Computer Vision for Human-Computer InteractionResearch Group, Universität Karlsruhe (TH) cv:hci Dr. Edgar Seemann 1 Computer Vision: Histograms of Oriented.
Semi-Supervised Hierarchical Models for 3D Human Pose Reconstruction Atul Kanaujia, CBIM, Rutgers Cristian Sminchisescu, TTI-C Dimitris Metaxas,CBIM, Rutgers.
Face detection Many slides adapted from P. Viola.
Ghunhui Gu, Joseph J. Lim, Pablo Arbeláez, Jitendra Malik University of California at Berkeley Berkeley, CA
Robust Moving Object Detection & Categorization using self- improving classifiers Omar Javed, Saad Ali & Mubarak Shah.
Patch Descriptors CSE P 576 Larry Zitnick
1 Image Recognition - I. Global appearance patterns Slides by K. Grauman, B. Leibe.
1 Learning to Detect Objects in Images via a Sparse, Part-Based Representation S. Agarwal, A. Awan and D. Roth IEEE Transactions on Pattern Analysis and.
Graz University of Technology, AUSTRIA Institute for Computer Graphics and Vision Fast Visual Object Identification and Categorization Michael Grabner,
A Study of Approaches for Object Recognition
Segmentation Divide the image into segments. Each segment:
Un Supervised Learning & Self Organizing Maps Learning From Examples
Robust Object Recognition with Cortex-Like Mechanisms Thomas Serre, Lior Wolf, Stanley Bileshi, Maximilian Riesenhuber, and Tomaso Poggio, Member, IEEE.
Fitting a Model to Data Reading: 15.1,
Associative Learning in Hierarchical Self Organizing Learning Arrays Janusz A. Starzyk, Zhen Zhu, and Yue Li School of Electrical Engineering and Computer.
5/30/2006EE 148, Spring Visual Categorization with Bags of Keypoints Gabriella Csurka Christopher R. Dance Lixin Fan Jutta Willamowski Cedric Bray.
Smart Traveller with Visual Translator for OCR and Face Recognition LYU0203 FYP.
Lecture 6: Feature matching and alignment CS4670: Computer Vision Noah Snavely.
Radial-Basis Function Networks
FACE DETECTION AND RECOGNITION By: Paranjith Singh Lohiya Ravi Babu Lavu.
Face Recognition Using Neural Networks Presented By: Hadis Mohseni Leila Taghavi Atefeh Mirsafian.
Multiclass object recognition
Distinctive Image Features from Scale-Invariant Keypoints By David G. Lowe, University of British Columbia Presented by: Tim Havinga, Joël van Neerbos.
Computer vision.
Prakash Chockalingam Clemson University Non-Rigid Multi-Modal Object Tracking Using Gaussian Mixture Models Committee Members Dr Stan Birchfield (chair)
2 2  Background  Vision in Human Brain  Efficient Coding Theory  Motivation  Natural Pictures  Methodology  Statistical Characteristics  Models.
Window-based models for generic object detection Mei-Chen Yeh 04/24/2012.
Classifying Images with Visual/Textual Cues By Steven Kappes and Yan Cao.
Representations for object class recognition David Lowe Department of Computer Science University of British Columbia Vancouver, Canada Sept. 21, 2006.
Classifiers Given a feature representation for images, how do we learn a model for distinguishing features from different classes? Zebra Non-zebra Decision.
Object Recognition in Images Slides originally created by Bernd Heisele.
An Introduction to Support Vector Machines (M. Law)
Recognition II Ali Farhadi. We have talked about Nearest Neighbor Naïve Bayes Logistic Regression Boosting.
Face Detection Ying Wu Electrical and Computer Engineering Northwestern University, Evanston, IL
Handwritten digit recognition
Face Detection Using Large Margin Classifiers Ming-Hsuan Yang Dan Roth Narendra Ahuja Presented by Kiang “Sean” Zhou Beckman Institute University of Illinois.
Visual Categorization With Bags of Keypoints Original Authors: G. Csurka, C.R. Dance, L. Fan, J. Willamowski, C. Bray ECCV Workshop on Statistical Learning.
Class 21, 1999 CBCl/AI MIT Neuroscience II T. Poggio.
Robust Real Time Face Detection
Local features and image matching October 1 st 2015 Devi Parikh Virginia Tech Disclaimer: Many slides have been borrowed from Kristen Grauman, who may.
CSC321 Lecture 5 Applying backpropagation to shape recognition Geoffrey Hinton.
Convolutional Restricted Boltzmann Machines for Feature Learning Mohammad Norouzi Advisor: Dr. Greg Mori Simon Fraser University 27 Nov
Face detection Many slides adapted from P. Viola.
Statistical Modeling and Learning in Vision --- cortex-like generative models Ying Nian Wu UCLA Department of Statistics JSM, August 2010.
Activity Recognition Journal Club “Neural Mechanisms for the Recognition of Biological Movements” Martin Giese, Tomaso Poggio (Nature Neuroscience Review,
Another Example: Circle Detection
9.012 Presentation by Alex Rakhlin March 16, 2001
Data Mining, Neural Network and Genetic Programming
MIRA, SVM, k-NN Lirong Xia. MIRA, SVM, k-NN Lirong Xia.
Article Review Todd Hricik.
Recognizing Deformable Shapes
Recognition using Nearest Neighbor (or kNN)
Paper Presentation: Shape and Matching
R-CNN region By Ilia Iofedov 11/11/2018 BGU, DNN course 2016.
Local features and image matching
Fourier Transform of Boundaries
MIRA, SVM, k-NN Lirong Xia. MIRA, SVM, k-NN Lirong Xia.
Presentation transcript:

Standard Brain Model for Vision The talk is given by Tomer Livne and Maria Zeldin

Overview Introduction to biological basis of vision Computer analogy to biology Implementation Discussion

Overview of biological vision Hierarchical structure From simple features to complex ones (Hubel & Weisel) Increased invariance

The basic idea Hubel and Weisel (1962, 1965) following experimental results proposed a model in which neighbouring simple cells are combined into complex cell. The result is complex cells with phase independence.

Max vs. sum pooling Electrophysiological results indicate that pooling may not be linear, the response of a complex cell can be best described by the activity of its maximal afferent.

From simple to complex cells:

A straightforward extension of this is to start with simple cells and end up with “higher-order-hyper-complex cells”. This is the basis for all the hierarchy idea!

The hierarchy based on the brain model: Hierarchical models of object recognition in cortex. Reisenhuber and Poggio. Nature, november 1999.

Clearer explanation of the hierarchy orientations - | \ / 1 0.7 Simple cells Max pooling - | \ / 1 0.7 Complex cells

Computer vision Usual approach – image patching Biological motivated approach - hierarchy

Representing objects by invariant complex features The IT area in the brain is dealing with object recognition. In this area there are cells that respond best to a specific object Hierarchical models of object recognition in cortex. Reisenhuber and Poggio. Nature, november 1999.

Recognize the same faces

In the previous task our brains did a very good job in recognizing same face even thou the scale, impression, illumination were different. And did not classified different faces as same even thou they have similar physical conditions

Motivation The presented approach is trying to implement into a computer system the hierarchical idea that was presented. In order to achieve similar robustness.

The models that we present deal with more general problem which is object classification. We can say that the problem of recognition of different transformations of an object is similar to the problem of classification.

Can computers reach similar properties to biology? Reisenhuber & Poggio (1999) demonstrate that it can. Comparing electrophysiological results from cells in the monkey brain with implemented hierarchical model.

Training stage: The monkey was trained to recognize restricted set of views of unfamiliar target stimuli resembling paperclips. They check which IT cell responds best to all views. After finding the cell that responded the most was picked for the study.

Test stage: The best reaction of the cell was to the trained data. The second best was to new transformations of the trained object. And very little response to new objects (distractors)

Learning the results: Hierarchical models of object recognition in cortex. Reisenhuber and Poggio. Nature America Inc, november 1999.

The hierarchy based on the brain model: We saw this part Now lets compare it to the model Hierarchical models of object recognition in cortex. Reisenhuber and Poggio. Nature, november 1999.

Hierarchical models of object recognition in cortex Hierarchical models of object recognition in cortex. Reisenhuber and Poggio. Nature America Inc, november 1999.

Results of scrambling Hierarchical models of object recognition in cortex. Reisenhuber and Poggio. Nature America Inc, november 1999.

Summary Goal- brain based object classification Biology view of the problem implementation of hierarchical structure comparing true results to model results

What’s next? Models based on the hierarchical idea we already discussed Riesenhuber & Poggio (1999) Serre & Riesenhuber (2004) Serre, Wolf, Bileschi, Riesenhuber, & Poggio (2007) Mutch & Lowe (2006) Modifications of the basic ideas limitations and shortcomings

Method #1 Riesenhuber & Poggio , ”Hierarchical models of objects recognition in cortex”, Nature 1999 Later it was modified by Serre, Wolf, Bileschi, Riesenhuber, & Poggio, “Robust object recognition with cortex-like mechanisms”, 2007.

16 different sizes (7X7, 9X9,…,37X37) 4 orientations S1 – Gabor filters 16 different sizes (7X7, 9X9,…,37X37) 4 orientations A total of 64 S1 type detectors Robust object recognition with cortex-like mechanisms. Serre, Wolf, Bileschi, Reisenhuber and Poggio. IEEE, march 2007.

A serial implementation of filtering

8 different sizes (8X8, 10X10,…,22X22) 4 orientations C1 – MAX pooling 8 different sizes (8X8, 10X10,…,22X22) 4 orientations A total of 32 C1 type detectors Used to define features during the learning stage Robust object recognition with cortex-like mechanisms. Serre, Wolf, Bileschi, Reisenhuber and Poggio. IEEE, march 2007.

S2 – learned features Holds N learned features 4 patch sizes (4X4, 8X8, 12X12, 16X16) indicating how many C1 neighboring cells are considered (this is done separately for each C1 scale) For each image patch X, a Gaussian radial basis function that depends on an Euclidean distance, is calculated from each of the stored features Pi (i=1:N) r=exp(-β ||X – Pi||²)

Robust object recognition with cortex-like mechanisms Robust object recognition with cortex-like mechanisms. Serre, Wolf, Bileschi, Reisenhuber and Poggio. IEEE, march 2007.

For each stored feature the best match (closest) Classifier C2 – max pooling For each stored feature the best match (closest) Classifier Classification is based on both C1 and C2 Robust object recognition with cortex-like mechanisms. Serre, Wolf, Bileschi, Reisenhuber and Poggio. IEEE, march 2007.

Summery 4 Layers of processing 2 types of operations (Max, Sum) Output – N dimensional vector

Model’s performance Testing the model Defining features Flexibility of the design

Robustness to background Ignoring presented unrelated data Training and test images contains both targets and distractors Performed best with C2 type detectors Simple detection – present/absent (no location information) Approaches maximal performance with 1000-5000 features Performance improve with increased training (more examples) Robust object recognition with cortex-like mechanisms. Serre, Wolf, Bileschi, Reisenhuber and Poggio. IEEE, march 2007.

Object specific features or a universal dictionary A Universal dictionary based system is good for small training sets (10,000 features) An object specific based system is better when using large training sets (improves with practice – increased number of features [200 an image])

Robust object recognition with cortex-like mechanisms Robust object recognition with cortex-like mechanisms. Serre, Wolf, Bileschi, Reisenhuber and Poggio. IEEE, march 2007.

Object recognition without a clutter Scene understanding using a windowing strategy Large inter-category variability Training sets of only either positive (target) or negative (no target) 2 classification systems: C1 and C2 based C1 based system performs better (able to efficiently represent objects’ boundaries)

Robust object recognition with cortex-like mechanisms Robust object recognition with cortex-like mechanisms. Serre, Wolf, Bileschi, Reisenhuber and Poggio. IEEE, march 2007.

Texture based objects Again C1 and C2 based classifiers C2 features are now evaluated only locally, not over all image locations C2 based classification is better (the features are more invariant and complex) Evaluated by correct labeling of pixels in the image Robust object recognition with cortex-like mechanisms. Serre, Wolf, Bileschi, Reisenhuber and Poggio. IEEE, march 2007.

A unified system – looking at multiple processing levels The hierarchical nature of the described system enables the use of multiple levels of feature Recognizing both shape and texture based objects in the same image Two processing pathways

Robust object recognition with cortex-like mechanisms Robust object recognition with cortex-like mechanisms. Serre, Wolf, Bileschi, Reisenhuber and Poggio. IEEE, march 2007.

Scene understanding task Complex scene understanding requires more than just detection of objects, location information of the detected objects is also required Shape-based objects C1 based classification, using a windowing approach, for both identification and localization Local neighborhood suppression by the maximal detected result Texture-based objects C2 based classification texture boundaries posses a problem (solved by additionally segmenting the image and averaging the responses within each segment)

Model summery Hierarchical design Efficiency Multiple processing pathways Universality Vs. specificity Limitations

Method #2 Mutch & Lowe Multiclass Object Recognition with Sparse, Localized Features. 2006.

Image scaling – 10 scales S1 – Gabor filters Single scale (11X11) 4 orientations applied to every location Evaluated at all possible locations Multiclass Object Recognition with Sparse, Localized Features. By Mutch & Lowe. IEEE 2006

C1 – local invariance Max pooling using a 10X10(size)X2(scale) filter Each orientation is tested separately used to define features during the learning stage Larger skips Multiclass Object Recognition with Sparse, Localized Features. By Mutch & Lowe. IEEE 2006

S2 – intermediate features 4 filter sizes (4X4, 8X8, 12X12, 16X16) defined by the stored features A Universal feature set Response to each filter (feature) is calculated as R(X,P) = exp[-(||X – P||²)/2σ²α] Multiclass Object Recognition with Sparse, Localized Features. By Mutch & Lowe. IEEE 2006

C2 – Global invariance SVM classifier A vector of size d of the maximal response (anywhere in the image) to each feature. SVM classifier Majority-voting based decision Multiclass Object Recognition with Sparse, Localized Features. By Mutch & Lowe. IEEE 2006

The overall look on all the stages: Multiclass Object Recognition with Sparse, Localized Features. By Mutch & Lowe. IEEE 2006

Summary Similar assumptions Differences in construction

Model performance and improvements Testing classification More biologically motivated improvements

Tests classification 101 categories (from Caltech101) Trained sets of 15 (or 30) images of each category Learn random features (in both size and location), an equal number for each category Construct C2 vectors Train the SVM (on the improved model also perform feature selection) Test stage

Results of the test: Multiclass Object Recognition with Sparse, Localized Features. By Mutch & Lowe. IEEE 2006

To get better results, some improvements were added to the model: S2 – encodes only the dominant orientation at each location. Increased number of tested orientations (from 4 to 12) Lateral inhibition – suppressing below threshold filter outputs in S1 & C1 layers Limited S2 invariance – in order to allow for preserving a certain amount of geometrical relations, S2 feature are limited to certain places in the image (relative to the center of the object) Select only good features for classification

Running the previous test on the improved model lead to the following results: Multiclass Object Recognition with Sparse, Localized Features. By Mutch & Lowe. IEEE 2006

Refining the model Multiclass Object Recognition with Sparse, Localized Features. By Mutch & Lowe. IEEE 2006

Tests detection/localization Sliding window Merging overlapping detections Single/multiple scale test images Multiclass Object Recognition with Sparse, Localized Features. By Mutch & Lowe. IEEE 2006

Summery Efficiency Improvements Limitations

THE END Thank you for listening!

Simple cell is an early visual neuron meaning it responds best to a line of a specific size, orientation, and phase. This cell responds best to 180 deg. phase. This cell responds best to 90 deg. phase.

back

back Image Simple cell (phase sensitive) Complex cell (phase insensitive) back