San Diego May 22, 2013 Giovanni Saponaro Giampiero Salvi

Slides:



Advertisements
Similar presentations
V-Detector: A Negative Selection Algorithm Zhou Ji, advised by Prof. Dasgupta Computer Science Research Day The University of Memphis March 25, 2005.
Advertisements

Evidential modeling for pose estimation Fabio Cuzzolin, Ruggero Frezza Computer Science Department UCLA.
Breakout session B questions. Research directions/areas Multi-modal perception cognition and interaction Learning, adaptation and imitation Design and.
Skill Presentation Chapter 7.
Perception and Perspective in Robotics Paul Fitzpatrick MIT Computer Science and Artificial Intelligence Laboratory Humanoid Robotics Group Goal To build.
A vision-based system for grasping novel objects in cluttered environments Ashutosh Saxena, Lawson Wong, Morgan Quigley, Andrew Y. Ng 2007 Learning to.
Database-Based Hand Pose Estimation CSE 6367 – Computer Vision Vassilis Athitsos University of Texas at Arlington.
Vision Based Control Motion Matt Baker Kevin VanDyke.
Robot Localization Using Bayesian Methods
Real-Time Human Pose Recognition in Parts from Single Depth Images Presented by: Mohammad A. Gowayyed.
 INTRODUCTION  STEPS OF GESTURE RECOGNITION  TRACKING TECHNOLOGIES  SPEECH WITH GESTURE  APPLICATIONS.
L ABORATORY FOR P ERCEPTUAL R OBOTICS U NIVERSITY OF M ASSACHUSETTS A MHERST D EPARTMENT OF C OMPUTER S CIENCE A Relational Representation for Procedural.
Overview of Computer Vision CS491E/791E. What is Computer Vision? Deals with the development of the theoretical and algorithmic basis by which useful.
1 Unsupervised Learning With Non-ignorable Missing Data Machine Learning Group Talk University of Toronto Monday Oct 4, 2004 Ben Marlin Sam Roweis Rich.
Vision-Based Interactive Systems Martin Jagersand c610.
Human-robot interaction Michal de Vries. Humanoid robots as cooperative partners for people Breazeal, Brooks, Gray, Hoffman, Kidd, Lee, Lieberman, Lockerd.
L ABORATORY FOR P ERCEPTUAL R OBOTICS U NIVERSITY OF M ASSACHUSETTS A MHERST D EPARTMENT OF C OMPUTER S CIENCE Intent Recognition as a Basis for Imitation.
Computational Vision Jitendra Malik University of California at Berkeley Jitendra Malik University of California at Berkeley.
Information Retrieval in Practice
Knowledge Systems Lab JN 8/24/2015 A Method for Temporal Hand Gesture Recognition Joshua R. New Knowledge Systems Laboratory Jacksonville State University.
Feature Extraction Spring Semester, Accelerometer Based Gestural Control of Browser Applications M. Kauppila et al., In Proc. of Int. Workshop on.
Learning to classify the visual dynamics of a scene Nicoletta Noceti Università degli Studi di Genova Corso di Dottorato.
A Review of Children, Humanoid Robots and Caregivers (Arsenio, 2004) COM3240 – Week 3 Presented by Gizdem Akdur.
Extracting Places and Activities from GPS Traces Using Hierarchical Conditional Random Fields Yong-Joong Kim Dept. of Computer Science Yonsei.
Ink and Gesture recognition techniques. Definitions Gesture – some type of body movement –a hand movement –Head movement, lips, eyes Depending on the.
Juhana Leiwo – Marco Torti.  Position and movement  Direction of acceleration (gravity) ‏  Proximity and collision sensing  3-dimensional spatial.
Juhana Leiwo – Marco Torti.  Position and movement  Direction of acceleration (gravity) ‏  Proximity and collision sensing  3-dimensional spatial.
Multimedia Specification Design and Production 2013 / Semester 2 / week 8 Lecturer: Dr. Nikos Gazepidis
Zhengyou Zhang Microsoft Research Digital Object Identifier: /MMUL Publication Year: 2012, Page(s): Professor: Yih-Ran Sheu Student.
SPIE'01CIRL-JHU1 Dynamic Composition of Tracking Primitives for Interactive Vision-Guided Navigation D. Burschka and G. Hager Computational Interaction.
EE 492 ENGINEERING PROJECT LIP TRACKING Yusuf Ziya Işık & Ashat Turlibayev Yusuf Ziya Işık & Ashat Turlibayev Advisor: Prof. Dr. Bülent Sankur Advisor:
Towards Cognitive Robotics Biointelligence Laboratory School of Computer Science and Engineering Seoul National University Christian.
CSCE 5013 Computer Vision Fall 2011 Prof. John Gauch
K. J. O’Hara AMRS: Behavior Recognition and Opponent Modeling Oct Behavior Recognition and Opponent Modeling in Autonomous Multi-Robot Systems.
CVPR Workshop on RTV4HCI 7/2/2004, Washington D.C. Gesture Recognition Using 3D Appearance and Motion Features Guangqi Ye, Jason J. Corso, Gregory D. Hager.
NUS CS5247 Deadlock-Free and Collision-Free Coordination of Two Robot Manipulators By Patrick A. O’Donnell and Tomás Lozano-Pérez MIT Artificial Intelligence.
Tracking with CACTuS on Jetson Running a Bayesian multi object tracker on a low power, embedded system School of Information Technology & Mathematical.
Computer Vision Michael Isard and Dimitris Metaxas.
Vision-based human motion analysis: An overview Computer Vision and Image Understanding(2007)
Model of the Human  Name Stan  Emotion Happy  Command Watch me  Face Location (x,y,z) = (122, 34, 205)  Hand Locations (x,y,z) = (85, -10, 175) (x,y,z)
Natural Tasking of Robots Based on Human Interaction Cues Brian Scassellati, Bryan Adams, Aaron Edsinger, Matthew Marjanovic MIT Artificial Intelligence.
Interactive Learning of the Acoustic Properties of Objects by a Robot
Robotic Chapter 8. Artificial IntelligenceChapter 72 Robotic 1) Robotics is the intelligent connection of perception action. 2) A robotic is anything.
Chapter 5 Multi-Cue 3D Model- Based Object Tracking Geoffrey Taylor Lindsay Kleeman Intelligent Robotics Research Centre (IRRC) Department of Electrical.
Confidence Based Autonomy: Policy Learning by Demonstration Manuela M. Veloso Thanks to Sonia Chernova Computer Science Department Carnegie Mellon University.
Object Lesson: Discovering and Learning to Recognize Objects Object Lesson: Discovering and Learning to Recognize Objects – Paul Fitzpatrick – MIT CSAIL.
Evaluating Perceptual Cue Reliabilities Robert Jacobs Department of Brain and Cognitive Sciences University of Rochester.
Inferring High-Level Behavior from Low-Level Sensors Donald J. Patterson, Lin Liao, Dieter Fox, and Henry Kautz.
Chapter 8. Learning of Gestures by Imitation in a Humanoid Robot in Imitation and Social Learning in Robots, Calinon and Billard. Course: Robots Learning.
Ghislain Fouodji Tasse Supervisor: Dr. Karen Bradshaw Computer Science Department Rhodes University 24 March 2009.
1 Robotic Chapter AI & ESChapter 7 Robotic 2 Robotic 1) Robotics is the intelligent connection of perception action. 2) A robotic is anything.
Chapter 21 Robotic Perception and action Chapter 21 Robotic Perception and action Artificial Intelligence ดร. วิภาดา เวทย์ประสิทธิ์ ภาควิชาวิทยาการคอมพิวเตอร์
Shape2Pose: Human Centric Shape Analysis CMPT888 Vladimir G. Kim Siddhartha Chaudhuri Leonidas Guibas Thomas Funkhouser Stanford University Princeton University.
Statistical environment representation to support navigation of mobile robots in unstructured environments Sumare workshop Stefan Rolfes Maria.
A Multi-Touch Display for Robotic Team Control
COGNITIVE APPROACH TO ROBOT SPATIAL MAPPING
IPAB Research Areas and Strengths
Gait Recognition Gökhan ŞENGÜL.
Supervised Time Series Pattern Discovery through Local Importance
Thrust IC: Action Selection in Joint-Human-Robot Teams
Dynamical Statistical Shape Priors for Level Set Based Tracking
Identifying Human-Object Interaction in Range and Video Data
Making Statistical Inferences
Learning about Objects
Statistical environment representation to support navigation of mobile robots in unstructured environments Stefan Rolfes Maria Joao Rendas
An Improved Neural Network Algorithm for Classifying the Transmission Line Faults Slavko Vasilic Dr Mladen Kezunovic Texas A&M University.
Department of Electrical Engineering
Computer Graphics Lecture 15.
EE 492 ENGINEERING PROJECT
Chapter 4 . Trajectory planning and Inverse kinematics
Presentation transcript:

Robot Anticipation of Human Intentions through Continuous Gesture Recognition San Diego May 22, 2013 Giovanni Saponaro Giampiero Salvi Alexandre Bernardino 1

Motivation Recognize human task actions in real time, enable robots to provide appropriate collaborative support. Approach based on computer vision, statistical models and cognitive robotics. example human-robot collaboration scenario assume Push-Tap-Grasp sequence as correct strategy (intention) human user has to move object avoiding collisions robot recognizes human gesture, provides appropriate support

Outline Overview of affordances Action (Gesture) recognition Conclusions

Affordances

Affordances All action possibilities latent in the environment, in relation to the actor capabilities (Gibson 77). Action 1 Environment items Action 2 Action n

Object Affordances Effects are related to properties of objects: Graphical Model 7

Learning Object Affordances Effect Actions Actions: Grasp, Tap and Touch Object features: shape, size, color Effects: object velocity, object-hand distance, hand velocity, contact 8

Learned object affordances Learn structure, then learn parameters A – Action C – Color Sh – Shape S – Size Di – Distance Ct – Contact V – Velocity Object Effect Actions 10

Using Affordances Generative model: Allows inferring any set of variables, given any others: Object, Action -> Effect (prediction: self or others) Object, Effect -> Action (recognition, planning) Effect, Action -> Object (selection, recognition)

Which action gives the same effect? Imitation games Objective: select action and object to obtain the same effect Demonstration (grasp on small box) Which action gives the same effect? 13

Action Recognition

Action Recognition

Action Recognition Goal: Trainable statistical model able to recognize individual gestures in a continuous sequence. Motivation: extend affordance model with body gestures equip robots with action recognition capabilities use body gestures as a cue to recognize actions and intention anticipate effects while partner's action is still taking place robot learning by demonstration, instruct robot with gestures Assumptions: sensor, actions repertoire and model

Action Recognition Issues finding gesture boundaries (temporal segmentation) accurate detection of joints trade-off between precision and invasiveness of sensors sub-gesture (substring) problem other common assumptions: small, known vocabulary; spatial restrictions; availability of whole data (offline processing) Examples library of human actions for human-robot interactive scenario: push, tap, grasp human user can speak and move at the same time, the two modalities are linked previously: robot actions described by external narrator future: human actions described by the human herself

Action Recognition Feature space: 3D position coordinates of hand joint in time

Action Recognition Vocabulary of simple manipulation gestures

Statistical models (Hidden Markov Models) Action Recognition Statistical models (Hidden Markov Models)

Action Recognition Inference algorithms:

Action Recognition Example 1: correct local action recognition (lexicon), correct global strategy (syntax)

Action Recognition Example 2: correct local action recognition (lexicon), incorrect global strategy (syntax)

Action Recognition Example 3: incorrect recognition

Action Recognition Conclusions: capture predictive power of actions build a model that understands gestures in order to improve interactions with robots, extend affordance model Future work: relax assumptions on sensor, make statistical models more robust (features, training) estimate attributes of actions and objects