Stanford hci group / cs376 u Scott Klemmer · 28 November 2006 Vision- Based Interacti on.

Slides:



Advertisements
Similar presentations
A Natural Interactive Game By Zak Wilson. Background This project was my second year group project at University and I have chosen it to present as it.
Advertisements

OpenCV Introduction Hang Xiao Oct 26, History  1999 Jan : lanched by Intel, real time machine vision library for UI, optimized code for intel 
Hand Gesture for Taking Self Portrait Shaowei Chu and Jiro Tanaka University of Tsukuba Japan 12th July 15 minutes talk.
Robust Part-Based Hand Gesture Recognition Using Kinect Sensor
Move With Me S.W Graduation Project An Najah National University Engineering Faculty Computer Engineering Department Supervisor : Dr. Raed Al-Qadi Ghada.
Sensor-based Situated, Individualized, and Personalized Interaction in Smart Environments Simone Hämmerle, Matthias Wimmer, Bernd Radig, Michael Beetz.
Human-Computer Interaction in the New Millenium Dr. Cheryl Seals.
Stanford hci group / cs376 research topics in human-computer interaction Vision-based Interaction Scott Klemmer 17 November 2005.
1 Studying Development and Debugging To Help Create a Better Programming Environment Brad A. Myers and Andrew Ko Human-Computer Interaction Institute School.
OpenCV Stacy O’Malley CS-590 Summer, What is OpenCV? Open source library of functions relating to computer vision. Cross-platform (Linux, OS X,
Supervised by Prof. LYU, Rung Tsong Michael Department of Computer Science & Engineering The Chinese University of Hong Kong Prepared by: Chan Pik Wah,
Stanford hci group / cs376 research topics in human-computer interaction Multimodal Interfaces Scott Klemmer 15 November 2005.
Computer Vision for Interactive Computer Graphics Mrudang Rawal.
Tracking Gökhan Tekkaya Gürkan Vural Can Eroğul. Outline Tracking –Overview –Head Tracking –Eye Tracking –Finger/Hand Tracking Demos.
Stockman MSU Fall Computing Motion from Images Chapter 9 of S&S plus otherwork.
[cvPONG] A 3-D Pong Game Controlled Using Computer Vision Techniques Quan Yu and Chris Wagner.
Image Formation Mohan Sridharan Based on slides created by Edward Angel CS4395: Computer Graphics 1.
Ch 1 Intro to Graphics page 1CS 367 First Day Agenda Best course you have ever had (survey) Info Cards Name, , Nickname C / C++ experience, EOS experience.
Stanford hci group / cs376 research topics in human-computer interaction I/O Toolkits Scott Klemmer 29 November 2005.
Application Programming Interface For Tracking Face & Eye Motion Team Members Tharaka Roshan Pathberiya Nimesh Saveendra Chamara Susantha Gayan Gunarathne.
I mage and M edia U nderstanding L aboratory for Performance Evaluation of Vision-based Real-time Motion Capture Naoto Date, Hiromasa Yoshimoto, Daisaku.
Electronic Visualization Laboratory University of Illinois at Chicago Interaction between Real and Virtual Humans: Playing Checkers R. Torre, S. Balcisoy.
Introduction to ECE432 Instructor: Ying Wu Dept. Electrical & Computer Engr. Northwestern University Evanston, IL 60208
1Computer Graphics Lecture 3 - Image Formation John Shearer Culture Lab – space 2
Knowledge Systems Lab JN 9/10/2002 Computer Vision: Gesture Recognition from Images Joshua R. New Knowledge Systems Laboratory Jacksonville State University.
Multimedia Specification Design and Production 2013 / Semester 2 / week 8 Lecturer: Dr. Nikos Gazepidis
IMAGE Daniel Harmon Michael Ryan Stu SPLITTERS Rabess Keener Dack Kao Haas.
PortableVision-based HCI A Hand Mouse System on Portable Devices 連矩鋒 (Burt C.F. Lien) Computer Science and Information Engineering Department National.
Real-time object tracking using Kalman filter Siddharth Verma P.hD. Candidate Mechanical Engineering.
A Method for Hand Gesture Recognition Jaya Shukla Department of Computer Science Shiv Nadar University Gautam Budh Nagar, India Ashutosh Dwivedi.
SUPERVISOR: MR. J. CONNAN.  The main purpose of system is to track an object across multiple frames using fixed input source.
Learning about Programming Languages By: Mike and Sean.
Portable Vision-Based HCI A Real-Time Hand Mouse System on Portable Devices 連矩鋒 (Burt C.F. Lien) Department of Computer Science and Information Engineering.
Stanford hci group / cs376 u Scott Klemmer · 16 November 2006 Speech & Multimod al.
出處: Signal Processing and Communications Applications, 2006 IEEE 作者: Asanterabi Malima, Erol Ozgur, and Miijdat Cetin 2015/10/251 指導教授:張財榮 學生:陳建宏 學號: M97G0209.
Research Interests of Dr. Dennis J Bouvier Fall 2007.
Enabling User Interactions with Video Contents Khalad Hasan, Yang Wang, Wing Kwong and Pourang Irani.
Human Computer Interaction © 2014 Project Lead The Way, Inc.Computer Science and Software Engineering.
1 The Rendering Pipeline. CS788 Topic of HCI 2 Outline  Introduction  The Graphics Rendering Pipeline  Three functional stages  Example  Bottleneck.
Developing An Educational Rigid Body Dynamics Physics Engine By Neal Milstein.
12/7/10 Looking Back, Moving Forward Computational Photography Derek Hoiem, University of Illinois Photo Credit Lee Cullivan.
Object Recognition in ROS Using Feature Extractors and Feature Matchers By Dolev Shapira.
Universitas Gunadarma Human-Computer Interaction Parno, SKom., MMSI Teknik Informatika, UG, 2008 staffsite.gunadarma.ac.id/parno.
Student: Ibraheem Frieslaar Supervisor: Mehrdad Ghaziasgar.
Delivering Business Value through IT Face feature detection using Java and OpenCV 1.
Developing An Educational Rigid Body Dynamics Physics Engine By Neal Milstein.
Spring 2007 COMP TUI 1 Computer Vision for Tangible User Interfaces.
Video Surveillance Under The Guidance of Smt. D.Neelima M.Tech., Asst. Professor Submitted by G. Subrahmanyam Roll No: 10021F0013 M.C.A.
Final Year Project. Project Title Kalman Tracking For Image Processing Applications.
Augmented Reality and 3D modelling Done by Stafford Joemat Supervised by Mr James Connan.
Vision-Guided Robot Position Control SKYNET Tony BaumgartnerBrock Shepard Jeff Clements Norm Pond Nicholas Vidovich Advisors: Dr. Juliet Hurtig & Dr. J.D.
Tools for the Tool Makers CS 376, Fall 2005 Kevin Collins.
Toolkits and Languages CSE 490JL Section Dec 1 st & 3 rd 2004 Richard C. Davis & Kate Everitt.
MIT Artificial Intelligence Laboratory — Research Directions Intelligent Perceptual Interfaces Trevor Darrell Eric Grimson.
Papier-Mache: Toolkit Support for Tangible Input HCI Group University of California Scott R. Klemmer Jack Li James Lin DUB Group University of Washington.
Visual Information Processing. Human Perception V.S. Machine Perception  Human perception: pictorial information improvement for human interpretation.
Automatic License Plate Recognition for Electronic Payment system Chiu Wing Cheung d.
Hand Gestures Based Applications
AHED Automatic Human Emotion Detection
FISH IDENTIFICATION SYSTEM
Multimodal Interfaces
Video-based human motion recognition using 3D mocap data
Vision-based Interaction
Speech & Multimodal Scott Klemmer · 16 November 2006.
I/O Toolkits Scott Klemmer · 16 November 2006.
Project #2 Multimodal Caricatural Mirror Intermediate report
Coupled Horn-Schunck and Lukas-Kanade for image processing
AHED Automatic Human Emotion Detection
CSE (c) S. Tanimoto, 2001 Image Understanding
Presentation transcript:

stanford hci group / cs376 u Scott Klemmer · 28 November 2006 Vision- Based Interacti on

2 cs547: Blake Ross and Asa Dotzler Mozilla: Creating simple software in a geek-driven culture

3 The first vision-based interface  Myron Krueger used computer vision to create Responsive Environments (1970s)  “Reaction is the Medium”  timeline/videoplace_video.html

4 How it works  Video and background are separated in analog using chroma key techniques (think broadcast news)  The first and last points of each raster are stored in the computer, and represent the person’s outline

5 Vision-based UIs: “Verbs”  Detecting and Tracking elements of a certain type in a scene  Capturing contents of detected objects  Recognizing individual members in an object class

6 Vision-based UIs: “Verbs”  Detecting and Tracking elements of a certain type in a scene

7 Vision-based UIs: “Verbs”  Capturing contents of detected objects

8 Vision-based UIs: “Verbs”  Recognizing individual members in a class

9 Vision-based UIs: “Nouns”  People (one or multiple)  Bodies  Faces  Hands  Documents  Objects

10 Vision-based UIs: “Nouns”  People (one or multiple)  Bodies  Faces  Hands  Documents  Objects

11 Vision-based UIs: “Nouns”  People (one or multiple)  Bodies  Faces  Hands  Documents  Objects

12 Background Subtraction I N F R A S T R U C T U R E

13 Image Moments (of Inertia)  0 th moment is mass (total number of pixels)

14 Image Moments (of Inertia)  1 st moment is center

15 Image Moments (of Inertia)  2 nd moment is orientation

16 Tools for Vision apps  Intel’s OpenCV  C API to highly optimized image processing functions (threshold, dilate, optical flow, …)  ncv  Fast to run! Slow to develop  Great for vision folks; too low-level for app folks  Papier-Mâché  Java API (and to some extent visual UI) for vision (and other physical input)   Fast to develop! Slow to run  Great for app folks; ~5 fps can sometimes be too slow

17 Good Vision Books  Computer Vision: A Modern Approach  David Forsyth and Jean Ponce (2003)  Fantastic book; but goal is more theoretical understanding than practical application  Robot Vision  Berthold Horn (1987)  More focused on apps and interactive algorithms  Somewhat out of date

18 Next Time… Software Tools Past, Present, and Future of User Interface Software Tools, Brad Myers, Scott E. Hudson, Randy Pausch Natural Programming Languages and Environments, Brad A. Myers, John F. Pane, Andy Ko