National Research Council Canada Conseil national de recherches Canada National Research Council Canada Conseil national de recherches Canada Canada Dmitry.

Slides:



Advertisements
Similar presentations
Detection and tracking of pianist hands and fingers Dmitry O. Gorodnichy 1 and Arjun Yogeswaran 23 1 Institute for Information Technology, National Research.
Advertisements

8.1si31_2001 SI31 Advanced Computer Graphics AGR Lecture 8 Polygon Rendering.
Miroslav Hlaváč Martin Kozák Fish position determination in 3D space by stereo vision.
Chayatat Ratanasawanya Min He May 13, Background information The goal Tasks involved in implementation Depth estimation Pitch & yaw correction angle.
Matthias Wimmer, Bernd Radig, Michael Beetz Chair for Image Understanding Computer Science TU München, Germany A Person and Context.
Visual Servo Control Tutorial Part 1: Basic Approaches Chayatat Ratanasawanya December 2, 2009 Ref: Article by Francois Chaumette & Seth Hutchinson.
The fundamental matrix F
Lecture 11: Two-view geometry
Face Alignment with Part-Based Modeling
MASKS © 2004 Invitation to 3D vision Lecture 7 Step-by-Step Model Buidling.
Facial feature localization Presented by: Harvest Jang Spring 2002.
Reference Guide: Visual feedback provided by Nousor for different states NouseBoard + NousePad Touch corner for next letter selection Go to block containing.
Instructor: Mircea Nicolescu Lecture 13 CS 485 / 685 Computer Vision.
December 5, 2013Computer Vision Lecture 20: Hidden Markov Models/Depth 1 Stereo Vision Due to the limited resolution of images, increasing the baseline.
A Versatile Depalletizer of Boxes Based on Range Imagery Dimitrios Katsoulas*, Lothar Bergen*, Lambis Tassakos** *University of Freiburg **Inos Automation-software.
1 of 25 1 of 22 Blind-Spot Experiment Draw an image similar to that below on a piece of paper (the dot and cross are about 6 inches apart) Close your right.
Camera calibration and epipolar geometry
Structure from motion.
CH24 in Robotics Handbook Presented by Wen Li Ph.D. student Texas A&M University.
Geometry of Images Pinhole camera, projection A taste of projective geometry Two view geometry:  Homography  Epipolar geometry, the essential matrix.
Object Recognition with Invariant Features n Definition: Identify objects or scenes and determine their pose and model parameters n Applications l Industrial.
Structure from motion. Multiple-view geometry questions Scene geometry (structure): Given 2D point matches in two or more images, where are the corresponding.
Head Tracking and Virtual Reality by Benjamin Nielsen.
Lecture 21: Multiple-view geometry and structure from motion
Chamfer Matching & Hausdorff Distance Presented by Ankur Datta Slides Courtesy Mark Bouts Arasanathan Thayananthan.
Lecture 11: Structure from motion CS6670: Computer Vision Noah Snavely.
Multiple View Geometry Marc Pollefeys University of North Carolina at Chapel Hill Modified by Philippos Mordohai.
MSU CSE 803 Fall 2008 Stockman1 CV: 3D sensing and calibration Coordinate system changes; perspective transformation; Stereo and structured light.
May 2004Stereo1 Introduction to Computer Vision CS / ECE 181B Tuesday, May 11, 2004  Multiple view geometry and stereo  Handout #6 available (check with.
Camera parameters Extrinisic parameters define location and orientation of camera reference frame with respect to world frame Intrinsic parameters define.
Lecture 6: Feature matching and alignment CS4670: Computer Vision Noah Snavely.
Lecture 12: Structure from motion CS6670: Computer Vision Noah Snavely.
כמה מהתעשייה? מבנה הקורס השתנה Computer vision.
Automatic Camera Calibration
Computer vision: models, learning and inference
Electronic Visualization Laboratory University of Illinois at Chicago Interaction between Real and Virtual Humans: Playing Checkers R. Torre, S. Balcisoy.
The Camera Mouse: Visual Tracking of Body Features to Provide Computer Access for People With Severe Disabilities.
Computer vision.
05 - Feature Detection Overview Feature Detection –Intensity Extrema –Blob Detection –Corner Detection Feature Descriptors Feature Matching Conclusion.
A HIGH RESOLUTION 3D TIRE AND FOOTPRINT IMPRESSION ACQUISITION DEVICE FOR FORENSICS APPLICATIONS RUWAN EGODA GAMAGE, ABHISHEK JOSHI, JIANG YU ZHENG, MIHRAN.
Final Exam Review CS485/685 Computer Vision Prof. Bebis.
Hand Tracking for Virtual Object Manipulation
Reconstructing 3D mesh from video image sequences supervisor : Mgr. Martin Samuelčik by Martin Bujňák specifications Master thesis
视觉的三维运动理解 刘允才 上海交通大学 2002 年 11 月 16 日 Understanding 3D Motion from Images Yuncai Liu Shanghai Jiao Tong University November 16, 2002.
December 4, 2014Computer Vision Lecture 22: Depth 1 Stereo Vision Comparing the similar triangles PMC l and p l LC l, we get: Similarly, for PNC r and.
An Information Fusion Approach for Multiview Feature Tracking Esra Ataer-Cansizoglu and Margrit Betke ) Image and.
National Research Council Canada Conseil national de recherches Canada National Research Council Canada Conseil national de recherches Canada Institute.
Generalized Hough Transform
Acquiring 3D models of objects via a robotic stereo head David Virasinghe Department of Computer Science University of Adelaide Supervisors: Mike Brooks.
1 Reconstructing head models from photograph for individualized 3D-audio processing Matteo Dellepiane, Nico Pietroni, Nicolas Tsingos, Manuel Asselot,
Lec 22: Stereo CS4670 / 5670: Computer Vision Kavita Bala.
Lecture 7: Features Part 2 CS4670/5670: Computer Vision Noah Snavely.
Chapter 5 Multi-Cue 3D Model- Based Object Tracking Geoffrey Taylor Lindsay Kleeman Intelligent Robotics Research Centre (IRRC) Department of Electrical.
A Flexible New Technique for Camera Calibration Zhengyou Zhang Sung Huh CSPS 643 Individual Presentation 1 February 25,
Two-view geometry. Epipolar Plane – plane containing baseline (1D family) Epipoles = intersections of baseline with image planes = projections of the.
2016/1/141 A novel method for detecting lips, eyes and faces in real time Real-Time Imaging (2003) 277–287 Cheng-Chin Chiang*,Wen-Kai Tai,Mau-Tsuen Yang,
Computer vision: models, learning and inference M Ahad Multiple Cameras
Lecture 9 Feature Extraction and Motion Estimation Slides by: Michael Black Clark F. Olson Jean Ponce.
Visual Odometry for Ground Vehicle Applications David Nistér, Oleg Naroditsky, and James Bergen Sarnoff Corporation CN5300 Princeton, New Jersey
Augmented Reality and 3D modelling Done by Stafford Joemat Supervised by Mr James Connan.
3D Reconstruction Using Image Sequence
Face Recognition Technology By Catherine jenni christy.M.sc.
Correspondence and Stereopsis. Introduction Disparity – Informally: difference between two pictures – Allows us to gain a strong sense of depth Stereopsis.
Lecture 07 13/12/2011 Shai Avidan הבהרה: החומר המחייב הוא החומר הנלמד בכיתה ולא זה המופיע / לא מופיע במצגת.
CS4670 / 5670: Computer Vision Kavita Bala Lec 27: Stereo.
René Vidal and Xiaodong Fan Center for Imaging Science
Common Classification Tasks
Reconstruction.
Presented by Xu Miao April 20, 2005
Towards building user seeing computers
Presentation transcript:

National Research Council Canada Conseil national de recherches Canada National Research Council Canada Conseil national de recherches Canada Canada Dmitry Gorodnichy, Gerhard Roth, Shahzad Mallik Institute for Information Technology Institut de technologie de l'information Institute for Information Technology Institut de technologie de l'information Computational Video Group Groupe Vidéo Informatique Computational Video Group Groupe Vidéo Informatique Nouse “Use your N ose as M ouse ” – New Technology for Hands-free Games and Interfaces Nouse “Use your N ose as M ouse ” – New Technology for Hands-free Games and Interfaces

(2) Face-Tracking Based User Interfaces 1.Replacing cumbersome track-ball (track-stick) on laptops. 2.Extra degree of control (e.g. to switch the focus of attention). 3.Hands free control (e.g. for handicap users). 4.Interactive games: more physical, entertaining, 3D control, multiple-user Fig.1. A user plays an aim-n-shoot Bubble-Frenzy game aiming the turret by pointing with her nose. (slight rotation of head allows to aim precisely in 180 o range)

(3) Fig. 2. Two users play a virtual ping-pong game, bouncing the ball with their heads. Image-based tracking allows one to track heads, however it doesn’t allow one to pin-point with head. Key Issues and Approaches 1.Speed (in real time). 2. Affordability (with cheap easy-to-install, but low-quality USB cams) 3. Robustness Precision 3. Robustness (to normal head motion). 4. Precision (with pixel precision) Image-based Face Tracking: - Uses global facial cues: skin colour, head shape, head motion not precise - Doesn’t require high-quility images, robust, but not precise

(4) Should be used for precise tracking. However, it’s not robust. [ Bradsky, Toyama, Gee, Cipolla, Zelinsky, Matsumoto, Yang, Baluja, Newman, …] …”still not ready for practical implementation” - Feature f is associated with vector V f (obtained by centering a mask on the feature) - Features are tracked by template matching with V f in the local area of interest (calculated with image- based cues) - The pixel u=(i,j) which has the largest score s(V u, V f ) is returned Feature-based Face Tracking Fig. 3. Tracking eyes (from [Gorodnichy97]). Question: What features to use? Proposition 1: Robust and precise tracking can be achieved by designing an invariant to head motion feature template.

(5) Features are conventionally thought of as visually distinctive (ie with large  I(f) ). Hence, the commonly used features are edge-based, such as corners of brows, eyes, lips, nostrils etc. They however are not robust not always visible Desired feature properties: 1. Uniqueness: s(V f, V u )  min 2. Robustness: s(V f t=0, V f t )  max 3. Continuity (for sub-pixel accuracy): the closer a pixel u in an image is to the pixel corresponding to f, the larger the score between V u and V f (Then evidence-based convolution can be applied to refine feature position u ) Edge-based Features Edge-based Features – not good

(6) Convex-shape features Convex-shape features – much better Definition 1: Convex-shape feature is defined as an extremum of a convex-shape surface Shape-from-Shading theory shows, that these features exhibit the desired properties (for the fixed camera-user-light configuration) Nose feature Definition 2: Nose feature is the extremum of the 3D nose surface curvature defined as z=f(x,y) in camera centered coordinate system. Thus defined, Nose feature is Very robust Can be detected with sub-pixel precision PLUS, It is always visible!

(7) Nouse TM Face Tracking Technology Based on tracking the convex-shape nose feature. Enables precise hand-free 2D control in a) joystick or b) mouse modes. Allows aiming and drawing with the nose. Just think of your nose as a chalk or a joystick handle! NB: Left/Right head motion is very natural and can be easily applied for control, provided it can tracked precisely. Affordable and downloadable. Uses a generic USB camera! Zero intialization of NouseUsing Nouse for Painting

(8) Performance: Robustness & Precision The range of head motion tracked ‘No’ motion ‘Yes’ motion Robustness to rotation Robustness to scale Test: The user rotates his head only! (the shoulders do not move)

(9) Demo: Range of Tracked Motion

(10) On Importance of Two Cameras For humans: it is much easier to track with two eyes than with one eye. Not only extends tracking from 2D to 3D, but also makes tracking more precise and robust! For computers however: … 1. The relationship between “eyes” is not known. 2. Tracking of features is not robust (to rotation and scale) StereoTracker from CVG NRC: Tracks face in 3D with two USB cams to control a virtual man, by using 1) Projective Vision Theory and 2) robust Nose Feature Tracking

(11) StereoTracking with USB webcams Stage 1: Self-calibration The relationship between the cameras is represented using the Fundamental Matrix F : (u left, F u right )=0 F can be found automatically for any two cameras by observing the same scene using : find corners  matching  filtering  robust solution with 7 selected corners ( RANSAC )  F Stage 2: Feature selection and calibration verification Select features in one image Verify that the epipolar line passes thru each feature in the second image Use nose tip feature and two other common features (eg brow corners) More at

(12) Using Nose for StereoTracking Proposition 2 : With F known, the tracked 3D feature is the one that minimizes the epipolar error defined by Proposition 3 : First detect convex-shape nose feature. Then use rigidity constraint to find other features.

(13) Demo: Stereotracking at Work Robustness to Scale and Rotation around all (!) axis of rotation

(14)Conclusions Acknowledgements Nouse TM is trademark of Computational Video Group IIT NRC BubbleFrenzy game is provided by Nose is a very unique feature. Humans are lucky to have it! Nose allows us to track a face very robustly and precisely. Pointing with Nose is natural. This makes 2D perceptual user interfaces a reality! Nose helps recovering other facial features. Two cameras (even bad webcams) make tracking more robust. This makes 3D face tracking affordable, precise and robust. Use your Nose as MouseUse NouseUse your Nose as Mouse! – Use Nouse! Nouse TM is open for public evaluation at