Computer Vision Lecture 2: Vision, Attention, and Eye Movements

Slides:



Advertisements
Similar presentations
Read this article for Friday next week [1]Chelazzi L, Miller EK, Duncan J, Desimone R. A neural basis for visual search in inferior temporal cortex. Nature.
Advertisements

Introduction to Eye Tracking
A Neural Model for Detecting and Labeling Motion Patterns in Image Sequences Marc Pomplun 1 Julio Martinez-Trujillo 2 Yueju Liu 2 Evgueni Simine 2 John.
Perception Chapter 9: Event Perception Event Perception: an event is defined as a change in both time and space. Thus far we have discussed how our visual.
Visual Attention Attention is the ability to select objects of interest from the surrounding environment A reliable measure of attention is eye movement.
Psych 216: Movement Attention. What is attention? There is too much information available in the world to process it all. Demonstration: change-detection.
Chapter 6: Visual Attention. Scanning a Scene Visual scanning – looking from place to place –Fixation –Saccadic eye movement Overt attention involves.
Use a pen on the test. The distinct modes of vision offered by feedforward and recurrent processing Victor A.F. Lamme and Pieter R. Roelfsema.
Next Tuesday Read article by Anne Treisman. Moving from Perception to Cognition You will now find chapters in the Cognition textbook on reserve to be.
Orienting Attention. Control of Attention Major Distinctions: VoluntaryReflexive.
We are on track for an exam on NOVEMBER 2 nd To cover everything since last exam up to Friday the 28th.
1 Biological Neural Networks Example: The Visual System.
Cognitive Processes PSY 334 Chapter 3 – Attention July 8, 2003.
Read article by Anne Treisman. Stages of Selection.
Disorders of Orienting Lesions to parietal cortex can produce some strange behavioural consequences –patients fail to notice events on the contralesional.
Electro-Oculography (EOG) Measurement System The goal : To measure eye movement with maximum accuracy using skin electrodes around the eyes that detect.
Jeff B. Pelz, Roxanne Canosa, Jason Babcock, & Eric Knappenberger Visual Perception Laboratory Carlson Center for Imaging Science Rochester Institute of.
Tracking multiple independent targets: Evidence for a parallel tracking mechanism Zenon Pylyshyn and Ron Storm presented by Nick Howe.
September 14, 2010Neural Networks Lecture 3: Models of Neurons and Neural Networks 1 Visual Illusions demonstrate how we perceive an “interpreted version”
November 5, 2009Introduction to Cognitive Science Lecture 16: Symbolic vs. Connectionist AI 1 Symbolism vs. Connectionism There is another major division.
December 1, 2009Introduction to Cognitive Science Lecture 22: Neural Models of Mental Processes 1 Some YouTube movies: The Neocognitron Part I:
Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC. Lecture 12: Visual Attention 1 Computational Architectures in Biological Vision,
Color vision Different cone photo- receptors have opsin molecules which are differentially sensitive to certain wavelengths of light – these are the physical.
Jeff B. Pelz, Roxanne Canosa, Jason Babcock, & Eric Knappenberger Visual Perception Laboratory Carlson Center for Imaging Science Rochester Institute of.
Overview 1.The Structure of the Visual Cortex 2.Using Selective Tuning to Model Visual Attention 3.The Motion Hierarchy Model 4.Simulation Results 5.Conclusions.
Michael Arbib & Laurent Itti: CS664 – Spring Lecture 5: Visual Attention (bottom-up) 1 CS 664, USC Spring 2002 Lecture 5. Visual Attention (bottom-up)
1B50 – Percepts and Concepts Daniel J Hulme. Outline Cognitive Vision –Why do we want computers to see? –Why can’t computers see? –Introducing percepts.
Eye Movements and Visual Attention
1 Perception, Illusion and VR HNRS 299, Spring 2008 Lecture 9 Visual Development Thanks to website at University of Calgary:
Active Vision Key points: Acting to obtain information Eye movements Depth from motion parallax Extracting motion information from a spatio-temporal pattern.
 The most intelligent device - “Human Brain”.  The machine that revolutionized the whole world – “computer”.  Inefficiencies of the computer has lead.
1 Computational Vision CSCI 363, Fall 2012 Lecture 3 Neurons Central Visual Pathways See Reading Assignment on "Assignments page"
THE VISUAL SYSTEM: EYE TO CORTEX Outline 1. The Eyes a. Structure b. Accommodation c. Binocular Disparity 2. The Retina a. Structure b. Completion c. Cone.
The primate visual systemHelmuth Radrich, The primate visual system 1.Structure of the eye 2.Neural responses to light 3.Brightness perception.
Chapter 8: Perceiving Motion
Visually guided attention during flying OR Pilots “do not like” fovea because they cannot pay attention to more than 1% of space at any one time.
Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 12: Visual Attention 1 Computational Architectures in Biological.
Perception and VR MONT 104S, Fall 2008 Lecture 8 Seeing Depth
1 Perception and VR MONT 104S, Spring 2008 Lecture 3 Central Visual Pathways.
March 31, 2016Introduction to Artificial Intelligence Lecture 16: Neural Network Paradigms I 1 … let us move on to… Artificial Neural Networks.
General Principles: The senses as physical instruments
Blindsight Patients with scotomas could move eyes to the location of a light flash (Poppel et al., 1973). Case D.B. (Larry Weizkrantz) hemianopic with.
Assist. Prof. Dr. Ilmiye Seçer Fall
Vision.
Computational Vision --- a window to our brain
Perception Unit How can we so easily perceive the world? Objects may be moving, partially hidden, varying in orientation, and projected as a 2D image.
How do we realize design? What should we consider? Technical Visual Interaction Search Context of use Information Interacting/ transacting.
CORTICAL MECHANISMS OF VISION
Volume 60, Issue 4, Pages (November 2008)
Optic Nerve Projections
Space Perception and Binocular Vision
Cognitive Processes PSY 334
Mind, Brain & Behavior Wednesday February 12, 2003.
Sensorimotor Learning and the Development of Position Invariance
Creating Data Representations
Orientation tuning: strongest response to one orientation
Independent Aftereffects of Attention and Motion
Artificial Neural Networks
Attentional Modulations Related to Spatial Gating but Not to Allocation of Limited Resources in Primate V1  Yuzhi Chen, Eyal Seidemann  Neuron  Volume.
Cognitive Processes PSY 334
Introduction to Artificial Intelligence Lecture 24: Computer Vision IV
Neural Mechanisms of Visual Motion Perception in Primates
Chapter 7 - Visual Attention
Attention & Perception
Voluntary Attention Modulates fMRI Activity in Human MT–MST
Introduction to Perception: Visual Perception
Learning Sensorimotor Contingencies
The functional architecture of attention
Attention and Scene Perception
Unit 3 Biological Bases of Behavior
Presentation transcript:

Computer Vision Lecture 2: Vision, Attention, and Eye Movements January 25, 2018 Computer Vision Lecture 2: Vision, Attention, and Eye Movements

Computer Vision Lecture 2: Vision, Attention, and Eye Movements Stimuli in receptive field of neuron January 25, 2018 Computer Vision Lecture 2: Vision, Attention, and Eye Movements

Computer Vision Lecture 2: Vision, Attention, and Eye Movements Cat V1 (striate cortex) Orientation preference map Ocular dominance map January 25, 2018 Computer Vision Lecture 2: Vision, Attention, and Eye Movements

Computer Vision Lecture 2: Vision, Attention, and Eye Movements January 25, 2018 Computer Vision Lecture 2: Vision, Attention, and Eye Movements

Structure of NNs (and some ANNs) In biological systems, neurons of similar functionality are usually organized in separate areas (or layers). Often, there is a hierarchy of interconnected layers with the lowest layer receiving sensory input and neurons in higher layers computing more complex functions. For example, neurons in macaque visual cortex have been identified that are activated only when there is a face (monkey, human, or drawing) in the macaque’s visual field. January 25, 2018 Computer Vision Lecture 2: Vision, Attention, and Eye Movements

Computer Vision Lecture 2: Vision, Attention, and Eye Movements “Data Flow Diagram” of Visual Areas in Macaque Brain Blue: motion perception pathway Green: object recognition pathway January 25, 2018 Computer Vision Lecture 2: Vision, Attention, and Eye Movements

Receptive Fields in Hierarchical Neural Networks neuron A Complexity tells us to use pyramids - but new problem... Routing is one of those problems that was little considered before Anderson & Van Essen pointed it out (Koch & Ullman did say selected features routed to high level, but not how). They used shifts in space (a giant multiplexor) controlled by outside signals to route a signal from anywhere int he input field to the appropriate portions of the output layer. Claimed that stereo,motion can also be handled analogously. Never implemented or tested the idea; in Olshausen’s thesis. connectivity was prohibitive. ‘Patchy’ connectivity solved this. Burt used a Laplacian pyramid: each level is determined as the difference of the previous two levels of the Guassian pyramid. Fovea imposed on this with 1/theta acuity solves routing. Only attended signal reaches output (rest is discarded at each level) Key to my routing is observation that stimulus feeds a forward cone and that any output ayer unit is root of a sub-pyramid. Potential for cross-talk is large Single attentional fixation at a time - can’t route two without interference! receptive field of A January 25, 2018 Computer Vision Lecture 2: Vision, Attention, and Eye Movements

Receptive Fields in Hierarchical Neural Networks neuron A in top layer Complexity tells us to use pyramids - but new problem... Routing is one of those problems that was little considered before Anderson & Van Essen pointed it out (Koch & Ullman did say selected features routed to high level, but not how). They used shifts in space (a giant multiplexor) controlled by outside signals to route a signal from anywhere int he input field to the appropriate portions of the output layer. Claimed that stereo,motion can also be handled analogously. Never implemented or tested the idea; in Olshausen’s thesis. connectivity was prohibitive. ‘Patchy’ connectivity solved this. Burt used a Laplacian pyramid: each level is determined as the difference of the previous two levels of the Guassian pyramid. Fovea imposed on this with 1/theta acuity solves routing. Only attended signal reaches output (rest is discarded at each level) Key to my routing is observation that stimulus feeds a forward cone and that any output ayer unit is root of a sub-pyramid. Potential for cross-talk is large Single attentional fixation at a time - can’t route two without interference! receptive field of A in input layer January 25, 2018 Computer Vision Lecture 2: Vision, Attention, and Eye Movements

Computer Vision Lecture 2: Vision, Attention, and Eye Movements January 25, 2018 Computer Vision Lecture 2: Vision, Attention, and Eye Movements

Face aftereffect – Thanks to Arash Afraz for the slides! January 25, 2018 Computer Vision Lecture 2: Vision, Attention, and Eye Movements

Computer Vision Lecture 2: Vision, Attention, and Eye Movements Visual Illusions Visual Illusions demonstrate how we perceive an “interpreted version” of the incoming light pattern rather that the exact pattern itself. January 25, 2018 Computer Vision Lecture 2: Vision, Attention, and Eye Movements

Computer Vision Lecture 2: Vision, Attention, and Eye Movements Visual Illusions He we see that the squares A and B from the previous image actually have the same luminance (but in their visual context are interpreted differently). January 25, 2018 Computer Vision Lecture 2: Vision, Attention, and Eye Movements

Computer Vision Lecture 2: Vision, Attention, and Eye Movements Visual Attention Visual attention is the selective allocation of visual processing resources. For example, we can focus our attention on a particular object of interest in the visual field. Visual processing of that object is enhanced while being rather shallow for other objects. Also, we can respond more quickly and accurately to changes in an attended region. This prioritization is necessary due to our limited processing resources. January 25, 2018 Computer Vision Lecture 2: Vision, Attention, and Eye Movements

Computer Vision Lecture 2: Vision, Attention, and Eye Movements Visual Attention The attentional cueing task introduced by Michael Posner gives insight into the dynamics of visual attention. Subjects are instructed to fixate on the central cross. One of two boxes (left or right) flashes to capture the subject’s attention (an automatic, involuntary response). After some a short delay (stimulus onset asynchrony - SOA) an asterisk appears in one of the boxes. The subject has to report as quickly as possible in which box the asterisk appeared. January 25, 2018 Computer Vision Lecture 2: Vision, Attention, and Eye Movements

The Posner Attention Task x January 25, 2018 Computer Vision Lecture 2: Vision, Attention, and Eye Movements

The Posner Attention Task x January 25, 2018 Computer Vision Lecture 2: Vision, Attention, and Eye Movements

The Posner Attention Task x January 25, 2018 Computer Vision Lecture 2: Vision, Attention, and Eye Movements

The Posner Attention Task * x January 25, 2018 Computer Vision Lecture 2: Vision, Attention, and Eye Movements

The Posner Attention Task x January 25, 2018 Computer Vision Lecture 2: Vision, Attention, and Eye Movements

The Posner Attention Task For short SOAs (< 200 ms), subjects respond faster if flash and asterisk appear on the same side. Cueing of attention to relevant location allows faster response. For longer SOAs, subjects respond more slowly if flash and asterisk appear on the same side. Inhibition-of-Return mechanism makes attention less likely to remain on the side of the flash until the asterisk appears. January 25, 2018 Computer Vision Lecture 2: Vision, Attention, and Eye Movements

Computer Vision Lecture 2: Vision, Attention, and Eye Movements Eye Muscles January 25, 2018 Computer Vision Lecture 2: Vision, Attention, and Eye Movements

Computer Vision Lecture 2: Vision, Attention, and Eye Movements Types of Eye Movement Fixations: The eye is almost motionless, for example, while reading a single, short word. The information from the scene is almost entirely acquired during fixation. Duration varies from 100-1000 ms, typically between 200-600 ms. Typical fixation frequency is about 3 Hz. Fixations are interspersed with saccades. January 25, 2018 Computer Vision Lecture 2: Vision, Attention, and Eye Movements

Computer Vision Lecture 2: Vision, Attention, and Eye Movements Types of Eye Movement Saccades: Quick “jumps” that connect fixations Duration is typically between 30 and 120 ms Very fast (up to 700 degrees/second) Saccades are ballistic, i.e., the target of a saccade cannot be changed during the movement. Vision is suppressed during saccades to allow stable perception of surroundings. Saccades are used to move the fovea to the next object/region of interest. January 25, 2018 Computer Vision Lecture 2: Vision, Attention, and Eye Movements

Computer Vision Lecture 2: Vision, Attention, and Eye Movements Types of Eye Movement Smooth Pursuit Eye Movements: Smooth movement of the eyes for visually tracking a moving object Cannot be performed in static scenes (fixation/saccade behavior instead) January 25, 2018 Computer Vision Lecture 2: Vision, Attention, and Eye Movements

Why Eye-Movement Research? About eye movements and visual attention: Usually, saccades follow shifts of attention to provide high acuity at the attended position. It is possible to look at an object without paying attention to it (staring). It is possible to shift attention without eye movement (covert shifts of attention). It is impossible to perform a saccade while not shifting attention. During specific, natural tasks it is reasonable to assume that saccades follow shifts of attention. January 25, 2018 Computer Vision Lecture 2: Vision, Attention, and Eye Movements

Why Eye-Movement Research? The investigation of visual attention, in turn, is at the core of cognitive science. Studying visual attention yields insight into general attentional mechanisms. It can provide information on a person’s stream of conscious and unconscious processing while solving a task. Attention is closely linked to the concept of consciousness. Attentional mechanisms could improve artificial vision systems. January 25, 2018 Computer Vision Lecture 2: Vision, Attention, and Eye Movements

Computer Vision Lecture 2: Vision, Attention, and Eye Movements Eye-Movement Studies Eye movements while watching a girl’s face (early study by Yarbus, 1967) January 25, 2018 Computer Vision Lecture 2: Vision, Attention, and Eye Movements

Computer Vision Lecture 2: Vision, Attention, and Eye Movements Eye-Movement Studies Eye movements as indicators of cognitive processes (Yarbus): trace 1: examine at will trace 2: estimate wealth trace 3: estimate ages trace 4: guess previous activity trace 5: remember clothing trace 6: remember position trace 7: time since last visit January 25, 2018 Computer Vision Lecture 2: Vision, Attention, and Eye Movements

Computer Vision Lecture 2: Vision, Attention, and Eye Movements Eye-Movement Studies Visual scan paths on instruments/dashboards – studies for the improvement of human-computer interfaces January 25, 2018 Computer Vision Lecture 2: Vision, Attention, and Eye Movements

Eye-Movement Studies Gaze trajectory measurement for the optimization of web page layout January 25, 2018 Computer Vision Lecture 2: Vision, Attention, and Eye Movements

Eye-Movement Studies Improving advertisements with eye-movement studies January 25, 2018 Computer Vision Lecture 2: Vision, Attention, and Eye Movements

Selectivity in Complex Scenes January 25, 2018 Computer Vision Lecture 2: Vision, Attention, and Eye Movements

Computer Vision Lecture 2: Vision, Attention, and Eye Movements Face Recognition Gaze-contingent window deteriorates face recognition, allows to identify relevant visual information. January 25, 2018 Computer Vision Lecture 2: Vision, Attention, and Eye Movements