Over the recent years, computer vision has started to play a significant role in the Human Computer Interaction (HCI). With efficient object tracking.

Slides:



Advertisements
Similar presentations
Applications of one-class classification
Advertisements

Mouse Training Instruction by: Connie Hutchison & Christopher McCoy.
Road-Sign Detection and Recognition Based on Support Vector Machines Saturnino, Sergio et al. Yunjia Man ECG 782 Dr. Brendan.
電腦視覺 Computer and Robot Vision I
No soda cans were harmed in the production of this project RAT Soda can detection system What is it RAT is an attempt to create a system to detect any.
Computer Training for Elders.  We make some assumptions about who you are: ◦ You are interested in using a computer for a few basic tasks (to begin with)
VisHap: Guangqi Ye, Jason J. Corso, Gregory D. Hager, Allison M. Okamura Presented By: Adelle C. Knight Augmented Reality Combining Haptics and Vision.
Multi-scenario Gesture Recognition Using Kinect Student : Sin- Jhu YE Student Id : MA Computer Engineering & Computer Science University of Louisville.
1 Video Processing Lecture on the image part (8+9) Automatic Perception Volker Krüger Aalborg Media Lab Aalborg University Copenhagen
Lecture 07 Segmentation Lecture 07 Segmentation Mata kuliah: T Computer Vision Tahun: 2010.
Image Maps and Graphics Internet Basics and Far Beyond! Mrs. Wilson.
 INTRODUCTION  STEPS OF GESTURE RECOGNITION  TRACKING TECHNOLOGIES  SPEECH WITH GESTURE  APPLICATIONS.
Image Segmentation Region growing & Contour following Hyeun-gu Choi Advisor: Dr. Harvey Rhody Center for Imaging Science.
1 Formation et Analyse d’Images Session 3 Daniela Hall 14 October 2004.
Edge detection. Edge Detection in Images Finding the contour of objects in a scene.
Ensemble Tracking Shai Avidan IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE February 2007.
Multiple Human Objects Tracking in Crowded Scenes Yao-Te Tsai, Huang-Chia Shih, and Chung-Lin Huang Dept. of EE, NTHU International Conference on Pattern.
Real-time Hand Pose Recognition Using Low- Resolution Depth Images
Cindy Song Sharena Paripatyadar. Use vision for HCI Determine steps necessary to incorporate vision in HCI applications Examine concerns & implications.
Objective of Computer Vision
Face Processing System Presented by: Harvest Jang Group meeting Fall 2002.
Visual Screen: Transforming an Ordinary Screen into a Touch Screen Zhengyou Zhang & Ying Shan Vision Technology Group Microsoft Research
Real-Time Face Detection and Tracking Using Multiple Cameras RIT Computer Engineering Senior Design Project John RuppertJustin HnatowJared Holsopple This.
ICBV Course Final Project Arik Krol Aviad Pinkovezky.
Sana Naghipour, Saba Naghipour Mentor: Phani Chavali Advisers: Ed Richter, Prof. Arye Nehorai.
Knowledge Systems Lab JN 8/24/2015 A Method for Temporal Hand Gesture Recognition Joshua R. New Knowledge Systems Laboratory Jacksonville State University.
Gesture Recognition Using Laser-Based Tracking System Stéphane Perrin, Alvaro Cassinelli and Masatoshi Ishikawa Ishikawa Namiki Laboratory UNIVERSITY OF.
WP3 - 3D reprojection Goal: reproject 2D ball positions from both cameras into 3D space Inputs: – 2D ball positions estimated by WP2 – 2D table positions.
Abstract Some Examples The Eye tracker project is a research initiative to enable people, who are suffering from Amyotrophic Lateral Sclerosis (ALS), to.
Information Extraction from Cricket Videos Syed Ahsan Ishtiaque Kumar Srijan.
Canyon Adventure Technology David Maung, Tristan Reichardt, Dan Bibyk, Juan Roman Department of Computer Science and Engineering The Ohio State University.
Knowledge Systems Lab JN 9/10/2002 Computer Vision: Gesture Recognition from Images Joshua R. New Knowledge Systems Laboratory Jacksonville State University.
1.Obtain laser pointer coordinates 2.Send laser pointer coordinates to the laptop through the standard PS/2 mouse port 3.Pixel Mapping 4.Attach a button.
Automated Face Detection Peter Brende David Black-Schaffer Veni Bourakov.
September 23, 2014Computer Vision Lecture 5: Binary Image Processing 1 Binary Images Binary images are grayscale images with only two possible levels of.
Digital Image Processing CCS331 Relationships of Pixel 1.
1 Research Question  Can a vision-based mobile robot  with limited computation and memory,  and rapidly varying camera positions,  operate autonomously.
Design Studies 20 ‘Show Off’ Project How to make a computer monitor In Google Sketchup By: Liam Jack.
Interactive Sand Art Drawing Using RGB-D Sensor
Edge Detection and Geometric Primitive Extraction Jinxiang Chai.
3d Pose Detection Used by Kinect
1 Formation et Analyse d’Images Session 4 Daniela Hall 10 October 2005.
Team Members Ming-Chun Chang Lungisa Matshoba Steven Preston Supervisors Dr James Gain Dr Patrick Marais.
1 Arrays of Arrays An array can represent a collection of any type of object - including other arrays! The world is filled with examples Monthly magazine:
Nottingham Image Analysis School, 23 – 25 June NITS Image Segmentation Guoping Qiu School of Computer Science, University of Nottingham
Robotics Chapter 6 – Machine Vision Dr. Amit Goradia.
Instructor: Mircea Nicolescu Lecture 5 CS 485 / 685 Computer Vision.
Vision Based hand tracking for Interaction The 7th International Conference on Applications and Principles of Information Science (APIS2008) Dept. of Visual.
An intelligent strategy for checking the annual inspection status of motorcycles based on license plate recognition Yo-Ping Huang a, Chien-Hung Chen b,
Face Detection – EE368 Group 10 May 30, Face Detection EE 368 Group 10 Waqar Mohsin Noman Ahmed Chung-Tse Mar.
SPACE MOUSE. INTRODUCTION  It is a human computer interaction technology  Helps in movement of manipulator in 6 degree of freedom * 3 translation degree.
Content Based Coding of Face Images
CMSC5711 Image processing and computer vision
Course : T Computer Vision
Hand Gestures Based Applications
1.3 Mouse Training.
Paper – Stephen Se, David Lowe, Jim Little
LECTURE Course Name: Computer Application
CMSC5711 Revision (1) (v.7.b) revised
Tremor Detection Using Motion Filtering and SVM Bilge Soran, Jenq-Neng Hwang, Linda Shapiro, ICPR, /16/2018.
CMSC5711 Image processing and computer vision
Higher School of Economics , Moscow, 2016
Presented by :- Vishal Vijayshankar Mishra
1.3 Mouse Training Mouse Training.
Computer and Robot Vision I
Computer and Robot Vision I
Higher School of Economics , Moscow, 2016
Higher School of Economics , Moscow, 2016
Presentation transcript:

Over the recent years, computer vision has started to play a significant role in the Human Computer Interaction (HCI). With efficient object tracking algorithms, it is possible to track the motion of a human hand in real time using a simple web camera. This presentation discusses the design of a system that tracks the fingertip of the index finger for the purpose of controlling mouse pointer on the screen.

 A single camera (web camera) is used to track the motion of the fingertip in real time.  The camera is mounted on the top of the computer monitor or hooked on the laptop screen.

 To move the mouse pointer, the user moves his index finger on a 2D plane (for example on a table). All other fingers are folded to form a fist.  The underlying tracking algorithm works by segmenting the skin colour from the background, it is required for the efficient tracking of the finger tip that user's other hand is not in the scene.

 To do left click, user unfolds his thumb and then folds it back.  Please note that if the hand pose is not as described above (all fingers in a folded form to form a fist except the index finger), then we simply track the hand but do not move the mouse pointer. Similarly the clicks are recognized only if index finger was spotted in the scene.

 The system consists of three parts: Online training Fingertip tracking and gesture recognition Computing mouse coordinates and smoothing mouse pointer path One TimeContinuous

 It allows the system to learn variations of skin colours due to change in illumination and hand poses.  A 2D look up table of Hue and Saturation values ( probability density values) from HSV color space is computed

 A CAMSHIFT algorithm is used to detect hand region, that is the detection of hand region is confined to Region of Interest (ROI).  Initially, the ROI is set to the entire captured frame.  The hand region is detected using the H and S values from the 2D look-up table computed during the training phase.  A threshold operation gives us the Binary image B(x,y) of the hand region, where B(x,y) = 1 for pixels in the skin region. 

 The centre of the hand region (xc, yc) and its orientation (θ) is calculated using the 0th, 1st, 2nd order image moments.  The ROI of the next frame is then computed from this binary image.  The centre of ROI for the next region is set to the centre of hand region in the current frame.

 The horizontal and vertical lengths (Rx, Ry) of the ROI of next frame are computed as follows: Rx = sx Moo Ry = sy M00 Where sx = cos|θ| + 1, sy = sin|θ| + 1 And M00 = 0th order image moment

To detect finger tip, first it is found out if the pose of the hand is our mouse pointer pose (only index finger in the unfolded position, rest all folded to form a fist). This is done by first cropping the image around the hand region, smoothing it by using Gaussian kernel so as to reduce noise and then analyzing the shape of the hand pose. A simple method is implemented to analyze the shape of the hand.

The image is converted into Binary image and the system scans the image from top to bottom row by row. The system counts the pixels that it encounters in each row and try to match it with the width of the finger. If enough rows are found with number of pixels greater than equal to the width of the finger, the system proceeds to check if there is a fist following the finger.

If enough number of rows with pixel count greater than fist width are found, then system shows that the pointing pose is detected. Basically an finite state machine has been implemented to detect the pointing pose in the image by analyzing the number of pixels in the rows. The dimensions of the finger, fist are determined during the on-line training process.

The x coordinate of finger tip is set to the first row having number of pixels greater than finger width. The y coordinate is set to the centre of the finger.

First the system proceeds to find out if thumb is present in the image. For this a similar Finite state machine is implemented. But this time the system scans the image column by column to find out thumb is present in the image or not. Depending upon if the user is left handed or right handed, the scanning will begin from left side or right side of the image.

The system tries to find out if enough number of columns with pixel count greater than thumb width are present in the image or not. If sufficient number of such columns detected, the system checks if there is a fist present in the system or not. Once the fist is also detected, the system declares that a thumb is detected in the image which is equivalent to mouse's left button down event.

When the user folds his thumb back, the system generates a mouse button up event. Also note that if pointing pose was not detected in the image, then mouse clicks are not detected at all. The system maintains the status of the mouse button (left button only).

 Once the fingertip is detected, we need to map the coordinates of the fingertip to the coordinates of the mouse on the monitor.  But, we can not use the fingertip locations directly due to the following problems:

Noise from sources such as segmentation error make it difficult to position the mouse pointer accurately Due to the limitation on the tracking rate, we might get discontinuous coordinates of fingertip locations Difference of resolution between the input image from the camera and monitor makes it difficult to position the mouse pointer accurately

 The displacement of detected finger tip is averaged over few frames and this average displacement is used to displace the mouse cursor on the screen.  Also if this displacement is found to be less than a threshold value, the mouse cursor is not moved.

 This application gives user an easy way to move mouse cursor on the screen.  Since user is using his hand, the homing time (time to place hand on the mouse) is reduced a lot.  The click is implemented with a very easy gesture.  With more robust finger tip detection, this application can replace the use of mouse.

 If background contains colors similar to skin, then the algorithm will loose track of the hand or falsely report its location.  When the camera's height is changed, the system has reported false pose detection. A better way to detect pointing pose is to use Machine learning algorithm (for example Neural network ).  The mouse cursor movement on the screen required more smoothing. Also user is not able to cover the entire screen.