Active Vision Key points: Acting to obtain information Eye movements Depth from motion parallax Extracting motion information from a spatio-temporal pattern.

Slides:



Advertisements
Similar presentations
Ter Haar Romeny, EMBS Berder 2004 How can we find a dense optic flow field from a motion sequence in 2D and 3D? Many approaches are taken: - gradient based.
Advertisements

Perception Chapter 9: Event Perception Event Perception: an event is defined as a change in both time and space. Thus far we have discussed how our visual.
What is vision Aristotle - vision is knowing what is where by looking.
PERCEPTION Chapter 4.5. Gestalt Principles  Gestalt principles are based on the idea that the whole is greater than the sum of the parts.  These principles.
Perception.
To represent the world, we must detect physical energy (a stimulus) from the environment and convert it into neural signals. This is a process called__________________.
What happens when no correspondence is possible? Highly mismatched stereo-pairs lead to ‘binocular rivalry’ TANGENT ALERT! Open question: Can rivalry and.
December 5, 2013Computer Vision Lecture 20: Hidden Markov Models/Depth 1 Stereo Vision Due to the limited resolution of images, increasing the baseline.
Motion Tracking. Image Processing and Computer Vision: 82 Introduction Finding how objects have moved in an image sequence Movement in space Movement.
Vision Computing An Introduction. Visual Perception Sight is our most impressive sense. It gives us, without conscious effort, detailed information about.
Computing motion between images
CSSE463: Image Recognition Day 30 Due Friday – Project plan Due Friday – Project plan Evidence that you’ve tried something and what specifically you hope.
Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 8: Stereoscopic Vision 1 Computational Architectures in Biological.
Visual Surveillance Chip. 1. Visual Change Detection 2. Photoreceptors and change detection 3. superior colliculus / optic tectum Visual Surveillance.
Ch 31 Sensation & Perception Ch. 3: Vision © Takashi Yamauchi (Dept. of Psychology, Texas A&M University) Main topics –convergence –Inhibition, lateral.
CS292 Computational Vision and Language Visual Features - Colour and Texture.
Visual motion Many slides adapted from S. Seitz, R. Szeliski, M. Pollefeys.
1 Motion in 2D image sequences Definitely used in human vision Object detection and tracking Navigation and obstacle avoidance Analysis of actions or.
Computational Architectures in Biological Vision, USC, Spring 2001
Motion detection with movement detectors. It is a non-linear device: response to velocity a and velocity b is not equal to velocity a+b movement detection.
1 Video Surveillance systems for Traffic Monitoring Simeon Indupalli.
Optical flow (motion vector) computation Course: Computer Graphics and Image Processing Semester:Fall 2002 Presenter:Nilesh Ghubade
1B50 – Percepts and Concepts Daniel J Hulme. Outline Cognitive Vision –Why do we want computers to see? –Why can’t computers see? –Introducing percepts.
MASKS © 2004 Invitation to 3D vision Lecture 3 Image Primitives andCorrespondence.
Studying Visual Attention with the Visual Search Paradigm Marc Pomplun Department of Computer Science University of Massachusetts at Boston
Sensing self motion Key points: Why robots need self-sensing Sensors for proprioception in biological systems in robot systems Position sensing Velocity.
1 Computational Vision CSCI 363, Fall 2012 Lecture 26 Review for Exam 2.
CSSE463: Image Recognition Day 30 This week This week Today: motion vectors and tracking Today: motion vectors and tracking Friday: Project workday. First.
Open Problems in Computer Vision(oProCV) Topic 4. Motion MSc. Luis A. Mateos.
Ch 81 Sensation & Perception Ch. 8: Perceiving Movement © Takashi Yamauchi (Dept. of Psychology, Texas A&M University) Main topics The functions of motion.
1 Perception and VR MONT 104S, Fall 2008 Session 13 Visual Attention.
1 Computational Vision CSCI 363, Fall 2012 Lecture 31 Heading Models.
Computer Vision, Robert Pless Lecture 11 our goal is to understand the process of multi-camera vision. Last time, we studies the “Essential” and “Fundamental”
THE VISUAL SYSTEM: EYE TO CORTEX Outline 1. The Eyes a. Structure b. Accommodation c. Binocular Disparity 2. The Retina a. Structure b. Completion c. Cone.
Lecture 2b Readings: Kandell Schwartz et al Ch 27 Wolfe et al Chs 3 and 4.
1 Computational Vision CSCI 363, Fall 2012 Lecture 20 Stereo, Motion.
Subject wearing a VR helmet immersed in the virtual environment on the right, with obstacles and targets. Subjects follow the path, avoid the obstacels,
CHAPTER 4 – SENSATION AND PERCEPTION SECTION 1 – SENSATION AND PERCEPTION: THE BASICS Objective: DISTINGUISH BETWEEN SENSATION AND PERCEPTION, AND EXPLAIN.
Visual motion Many slides adapted from S. Seitz, R. Szeliski, M. Pollefeys.
Recognizing Action at a Distance Alexei A. Efros, Alexander C. Berg, Greg Mori, Jitendra Malik Computer Science Division, UC Berkeley Presented by Pundik.
Chapter 8: Perceiving Motion
1 Computational Vision CSCI 363, Fall 2012 Lecture 21 Motion II.
1 Artificial Intelligence: Vision Stages of analysis Low level vision Surfaces and distance Object Matching.
CSE 185 Introduction to Computer Vision Stereo. Taken at the same time or sequential in time stereo vision structure from motion optical flow Multiple.
3D Imaging Motion.
Jack Pinches INFO410 & INFO350 S INFORMATION SCIENCE Computer Vision I.
Segmentation of Vehicles in Traffic Video Tun-Yu Chiang Wilson Lau.
Features, Feature descriptors, Matching Jana Kosecka George Mason University.
1 Perception and VR MONT 104S, Fall 2008 Lecture 6 Seeing Motion.
Colour and Texture. Extract 3-D information Using Vision Extract 3-D information for performing certain tasks such as manipulation, navigation, and recognition.
Course14 Dynamic Vision. Biological vision can cope with changing world Moving and changing objects Change illumination Change View-point.
Motion Estimation using Markov Random Fields Hrvoje Bogunović Image Processing Group Faculty of Electrical Engineering and Computing University of Zagreb.
Give examples of the way that virtual reality can be used in Psychology.
How is vision used to catch a ball?
Perception and VR MONT 104S, Fall 2008 Lecture 8 Seeing Depth
Correspondence and Stereopsis Original notes by W. Correa. Figures from [Forsyth & Ponce] and [Trucco & Verri]
Sensation & Perception. Motion Vision I: Basic Motion Vision.
Eye Movements and Working Memory Marc Pomplun Department of Computer Science University of Massachusetts at Boston Homepage:
Optical flow and keypoint tracking Many slides adapted from S. Seitz, R. Szeliski, M. Pollefeys.
1 Computational Vision CSCI 363, Fall 2012 Lecture 29 Structure from motion, Heading.
MASKS © 2004 Invitation to 3D vision Lecture 3 Image Primitives andCorrespondence.
What visual image information is needed for the things we do? How is vision used to acquire information from the world?
Correspondence and Stereopsis. Introduction Disparity – Informally: difference between two pictures – Allows us to gain a strong sense of depth Stereopsis.
Vision.
Image Primitives and Correspondence
Early Processing in Biological Vision
Presented by: Cindy Yan EE6358 Computer Vision
Experiencing the World
Optical flow and keypoint tracking
PERCEPTION Def: the mental process of organizing sensory input into meaningful patterns.
Presentation transcript:

Active Vision Key points: Acting to obtain information Eye movements Depth from motion parallax Extracting motion information from a spatio-temporal pattern Obtaining structure from motion

The importance of activity Computer vision is often approached as a problem of passive extraction of information from single images But most natural vision systems try to actively sample the visual scene Can solve some problems with active system that are hard with passive system

Segmenting (Kruger, 1998)

Improving resolution Making best use of fixed number of receptors e.g. 100x100 pixels (Ballard, 1991)

Eye movements Increase the effective resolution by saccade movements of high resolution area (fovea) Creates impression that see complete detailed scene, but this is illusory

Yarbus (1967) eye movements vary to extract different information Eye movement patterns indicate attention and task 1. Describe room. 2. What was happening before? 3. People’s ages.

Eye movements Increase the effective resolution by saccade movements of high resolution area (fovea) Creates impression that see detailed scene, but this is illusory Task dependent, indicates attention

Mid-lecture Question The human eye/brain system? 1 - sees the whole scene 2 - has higher resolution in the fovea 3 - saccades 4 - is directed by an attention process

Eye movements and localisation Knowing where the eye/camera is pointing tells us the direction of objects of interest (requires proprioception to know relative angles) Can also extract depth through motion parallax

Motion Parallax B B B A A A Closer objects change retinal position more when we move Geometrically equivalent to binocular stereopsis

Motion perception Like parallax, much important information about the world comes from sensing visual motion E.g. breaking camouflage, sensing self motion, seeing what is happening… ‘Active vision’ sometimes taken to mean vision based on sequences of images Aim is to extract the flow-field:

Method 1: feature tracking Assumes have first extracted features Assumes gradual change, and may need to deal with false matches or disappearing features In biological systems, motion perception is more fundamental than seeing objects change position Image at time 1Image at time 2 feature search window neighbourhood best match neighbourhood Motion vector

Motion as a perceptual primitive Can detect motion at faster times and smaller distances than can detect motionless features Patients with motion-blindness ( Zihl et al, 1983) Retinal motion detector neurons Visual cortex motion features detectors

Method 1: correlation DELAY A(t) B(t-1)A(t-1) B(t) If image is moving from A to B, then A(t-1) will correlate with B(t) If image is moving from A to B, then A(t) will not correlate with B(t-1) Sign indicates direction of movement Observed neural circuits in brain region V1 Image Motion A(t):B(t-1) - B(t):A(t-1)

Method 2: gradients I x t I Spatial gradient =dI/dx Temporal gradient =dI/dt Optical Flow Constraint Equation:

Method 4: energy models Consider space-time ‘image’ I(x,y,t) (pictures from Bruce et al 1996) Summing and differencing their outputs can produce equivalent output to gradient or correlation models Create filters that detect intensity edges (or similar features) oriented in space-time

Structure from motion Motion field contains information about the 3-d structure of objects (e.g. strong depth effect) If rigid body, and can track points, can geometrically recover structure of scene and movement of camera - active field in Computer Vision.

Structure from motion 2 What is this?

Structure from motion 3 3D structure emerges from pattern of motion

Structure from motion Source images Reconstructed from tracked feature points:

Attention from motion Can use flow-field to determine where to redirect the eyes – moving stimuli are salient Mechanism to determine new eye position: –Calculate the flow field –Enhance changes to detect new stimuli –Smooth to offset noise –Implement ‘winner-take-all’ connection to choose most salient movement, and inhibit return to same location Note that then have to solve problem of mapping visual target onto correct motion of camera

Vijayakumar et al. 2001

Eye movements Increase the effective resolution by saccade movements of high resolution area (fovea) Creates impression that see detailed scene, but this is illusory Task dependent, indicates attention Why/how do we interpret the world as static when the image is constantly moving?

Explaining movement in retinal image? YES:NO: See motionDon’t see motion IF eyes not moving IF eyes are moving Perceive stationary Perceive moving world world during saccade (e.g. tracking, stabilised image) Visual system uses info to construct perception – but which info?

How does the visual system take eye movements into account? Inflow theory Outflow theory Brain Perceived motion Motion on retina Stretch in eye muscles Difference Brain Perceived motion Signal for eye to move Motion on retina copy Difference

If immobolise the eyes and attempt to move them: Inflow theory predicts see no motion Outflow theory predicts see motion No motion Motion on retina = 0 Stretch in eye muscles = 0 diff. Brain See motion Signal for eye to move = + Motion on retina = 0 copy diff.

If eye is moved passively: Inflow theory predicts see no motion Outflow theory predicts see motion No motion Motion on retina = + Stretch in eye muscles = + diff. Brain See motion Signal for eye to move = 0 Motion on retina = + copy diff.

Active Vision Summary Changing Visual Field gives info about: Visual details Groupings Object depth and 3D structure Interesting Objects Issues to consider: Self or scene movement Depth info from: features, area correlation, optical flow, spatio-temporal patterns