A Model of Saliency-Based Visual Attention for Rapid Scene Analysis Laurent Itti, Christof Koch, and Ernst Niebur IEEE PAMI, 1998.

Slides:



Advertisements
Similar presentations
Spike Based Visual Encoding Activity level (a m ) Visual encoder implemented in the NEF as network of 1024 laterally inhibiting neural columns Network.
Advertisements

A Neural Model for Detecting and Labeling Motion Patterns in Image Sequences Marc Pomplun 1 Julio Martinez-Trujillo 2 Yueju Liu 2 Evgueni Simine 2 John.
Object recognition and scene “understanding”
Change Detection C. Stauffer and W.E.L. Grimson, “Learning patterns of activity using real time tracking,” IEEE Trans. On PAMI, 22(8): , Aug 2000.
Image Understanding Roxanne Canosa, Ph.D.. Introduction Computer vision  Give machines the ability to see  The goal is to duplicate the effect of human.
Leveraging Stereopsis for Saliency Analysis
Attention - Overview Definition Theories of Attention Neural Correlates of Attention Human neurophysiology and neuroimaging Change Blindness Deficits of.
黃文中 Preview 2 3 The Saliency Map is a topographically arranged map that represents visual saliency of a corresponding visual scene. 4.
A saliency map model explains the effects of random variations along irrelevant dimensions in texture segmentation and visual search Li Zhaoping, University.
Control of Attention and Gaze in the Natural World.
Attention, Awareness, and the Computational Theory of Surprise Research Qualifying Exam August 30 th, 2006.
A Novel Method for Generation of Motion Saliency Yang Xia, Ruimin Hu, Zhenkun Huang, and Yin Su ICIP 2010.
 Introduction  Principles of context – aware saliency  Detection of context – aware saliency  Result  Application  Conclusion.
IEEE TCSVT 2011 Wonjun Kim Chanho Jung Changick Kim
Learning to Detect A Salient Object Reporter: 鄭綱 (3/2)
Stas Goferman Lihi Zelnik-Manor Ayellet Tal. …
Tracking Objects with Dynamics Computer Vision CS 543 / ECE 549 University of Illinois Derek Hoiem 04/21/15 some slides from Amin Sadeghi, Lana Lazebnik,
Michigan State University1 Visual Attention and Recognition Through Neuromorphic Modeling of “Where” and “What” Pathways Zhengping Ji Embodied Intelligence.
Visual Attention More information in visual field than we can process at a given moment Solutions Shifts of Visual Attention related to eye movements Some.
Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 6: Low-level features 1 Computational Architectures in Biological.
Signal Processing Institute Swiss Federal Institute of Technology, Lausanne 1 “OBJECTIVE AND SUBJECTIVE IDENTIFICATION OF INTERESTING AREAS IN VIDEO SEQUENCES”
CS 664, Session 19 1 General architecture. CS 664, Session 19 2 Minimal Subscene Working definition: The smallest set of objects, actors and actions in.
Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC. Lecture 12: Visual Attention 1 Computational Architectures in Biological Vision,
1 Segmentation of Salient Regions in Outdoor Scenes Using Imagery and 3-D Data Gunhee Kim Daniel Huber Martial Hebert Carnegie Mellon University Robotics.
Visual Attention Jeremy Wyatt. Where to look? Many visual processes are expensive Humans don’t process the whole visual field How do we decide what to.
Christian Siagian Laurent Itti Univ. Southern California, CA, USA
Michigan State University 1 “Saliency-Based Visual Attention” “Computational Modeling of Visual Attention”, Itti, Koch, (Nature Reviews – Neuroscience.
Attention in Computer Vision Mica Arie-Nachimson and Michal Kiwkowitz May 22, 2005 Advanced Topics in Computer Vision Weizmann Institute of Science.
Overview 1.The Structure of the Visual Cortex 2.Using Selective Tuning to Model Visual Attention 3.The Motion Hierarchy Model 4.Simulation Results 5.Conclusions.
Michael Arbib & Laurent Itti: CS664 – USC, spring Lecture 6: Object Recognition 1 CS664, USC, Spring 2002 Lecture 6. Object Recognition Reading Assignments:
Michael Arbib & Laurent Itti: CS664 – Spring Lecture 5: Visual Attention (bottom-up) 1 CS 664, USC Spring 2002 Lecture 5. Visual Attention (bottom-up)
BATTLECAM™: A Dynamic Camera System for Real-Time Strategy Games Yangli Hector Yee Graphics Programmer, Petroglyph Elie Arabian.
Introduction of Saliency Map
Computer Science Department, Duke UniversityPhD Defense TalkMay 4, 2005 Fast Extraction of Feature Salience Maps for Rapid Video Data Analysis Nikos P.
Studying Visual Attention with the Visual Search Paradigm Marc Pomplun Department of Computer Science University of Massachusetts at Boston
Text Lecture 2 Schwartz.  The problem for the designer is to ensure all visual queries can be effectively and rapidly served.  Semantically meaningful.
Title of On the Implementation of a Information Hiding Design based on Saliency Map A.Basu, T. S. Das and S. K. Sarkar/ Jadavpur University/ Kolkata/ India/
REU Presentation Week 3 Nicholas Baker.  What features “pop out” in a scene?  No prior information/goal  Identify areas of large feature contrasts.
Visual Attention Accelerated Vehicle Detection in Low-Altitude Airborne Video of Urban Environment Xianbin Cao, Senior Member, IEEE, Renjun Lin, Pingkun.
Visual Attention Derek Hoiem March 14, 2007 Misc Reading Group.
黃文中 Introduction The Model Results Conclusion 2.
N n Debanga Raj Neog, Anurag Ranjan, João L. Cardoso, Dinesh K. Pai Sensorimotor Systems Lab, Department of Computer Science The University of British.
National Taiwan A Road Sign Recognition System Based on a Dynamic Visual Model C. Y. Fang Department of Information and.
Geodesic Saliency Using Background Priors
Department of Psychology & The Human Computer Interaction Program Vision Sciences Society’s Annual Meeting, Sarasota, FL May 13, 2007 Jeremiah D. Still,
Region-Based Saliency Detection and Its Application in Object Recognition IEEE TRANSACTIONS ON CIRCUITS AND SYSTEM FOR VIDEO TECHNOLOGY, VOL. 24 NO. 5,
Street Smarts: Visual Attention on the Go Alexander Patrikalakis May 13, XXX.
Efficient Color Boundary Detection with Color-opponent Mechanisms CVPR2013 Posters.
Computer Science Readings: Reinforcement Learning Presentation by: Arif OZGELEN.
Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 12: Visual Attention 1 Computational Architectures in Biological.
Neural Models of Visual Attention John K. Tsotsos Center for Vision Research York University, Toronto, Canada Marc Pomplun Department of Computer Science.
Spatio-temporal saliency model to predict eye movements in video free viewing Gipsa-lab, Grenoble Département Images et Signal CNRS, UMR 5216 S. Marat,
An MPEG-7 Based Semantic Album for Home Entertainment Presented by Chen-hsiu Huang 2003/08/12 Presented by Chen-hsiu Huang 2003/08/12.
CSC321: Neural Networks Lecture 18: Distributed Representations
A Model of Saliency-Based Visual Attention for Rapid Scene Analysis
Perception and VR MONT 104S, Fall 2008 Lecture 8 Seeing Depth
Biological Modeling of Neural Networks: Week 10 – Neuronal Populations Wulfram Gerstner EPFL, Lausanne, Switzerland 10.1 Cortical Populations - columns.
ICCV 2009 Tilke Judd, Krista Ehinger, Fr´edo Durand, Antonio Torralba.
Biologically Inspired Vision-based Indoor Localization Zhihao Li, Ming Yang
National Taiwan Normal A System to Detect Complex Motion of Nearby Vehicles on Freeways C. Y. Fang Department of Information.
A Neurodynamical Cortical Model of Visual Attention and Invariant Object Recognition Gustavo Deco Edmund T. Rolls Vision Research, 2004.
Face recognition using Histograms of Oriented Gradients
Introduction to Skin and Face Detection
Perception in 3D Computer Graphics
Tracking Objects with Dynamics
BATTLECAM™: A Dynamic Camera System for Real-Time Strategy Games
Implementation of a Visual Attention Model
Attentional Modulations Related to Spatial Gating but Not to Allocation of Limited Resources in Primate V1  Yuzhi Chen, Eyal Seidemann  Neuron  Volume.
The functional architecture of attention
Presentation transcript:

A Model of Saliency-Based Visual Attention for Rapid Scene Analysis Laurent Itti, Christof Koch, and Ernst Niebur IEEE PAMI, 1998

What is Saliency? ● Something is said to be salient if it stands out ● E.g. road signs should have high saliency

Introduction ● Trying to model visual attention ● Find locations of Focus of Attention in an image ● Use the idea of saliency as a basis for their model ● For primates focus of attention directed from: ● Bottom-up: rapid, saliency driven, task- independent ● Top-down: slower, task dependent

Results of the Model Only considering “Bottom-up”  task-independent

Model diagram

Model ● Input: static images (640x480) ● Each image at 8 different scales (640x480, 320x240, 160x120, …) ● Use different scales for computing “centre- surround” differences (similar to assignment) + - Fine scale Course scale

Feature Maps 1. Intensity contrast (6 maps) ● Using “centre-surround” ● Similar to neurons sensitive to dark centre, bright surround, and opposite 2. Color (12 maps) ● Similar to intensity map, but using different color channels ● E.g. high response to centre red, surround green

Feature Maps 3. Orientation maps (24 maps) ● Gabor filters at 0º, 45º, 90º, and 135º ● Also at different scales  Total of 42 feature maps are combined into the saliency map

Saliency Map ● Purpose: represent saliency at all locations with a scalar quantity ● Feature maps combined into three “conspicuity maps” ● Intensity (I) ● Color (C) ● Orientation (O) ● Before they are combined they need to be normalized

Normalization Operator

Example of operation Leaky integrate-and-fire neurons “Inhibition of return”

Model diagram

Example of operation Using 2D “winner- take-all” neural network at scale 4 FOA shifts every ms Inhibition lasts ms

Results Image Saliency Map High saliency Locations (yellow circles)

Results ● Tested on both synthetic and natural images ● Typically finds objects of interest, e.g. traffic signs, faces, flags, buildings… ● Generally robust to noise (less to multicoloured noise)

Uses ● Real-time systems ● Could be implemented in hardware ● Great reduction of data volume ● Video compression (Parkhurst & Niebur) ● Compress less important parts of images

Summary ● Basic idea: ● Find multiple saliency measures in parallel ● Normalize ● Combine them to a single map ● Use 2D integrate-and-fire layer of neurons to determine position of FOA ● Model appears to work accurately and robustly (but difficult to evaluate) ● Can be extended with other feature maps

References ● Itti, Koch, and Niebur: “A Model of Saliency- Based Visual Attention for Rapid Scene Analysis” IEEE PAMI Vol. 20, No. 11, November (1998) ● Itti, Koch: “Computational Modeling of Visual Attention”, Nature Reviews – Neuroscience Vol. 2 (2001) ● Parkhurst, Law, Niebur: “Modeling the role of salience in the allocation of overt visual attention”, Vision Research 42 (2002)