Video-based human motion recognition using 3D mocap data

Slides:



Advertisements
Similar presentations
Gestures Recognition. Image acquisition Image acquisition at BBC R&D studios in London using eight different viewpoints. Sequence frame-by-frame segmentation.
Advertisements

1 Gesture recognition Using HMMs and size functions.
Evidential modeling for pose estimation Fabio Cuzzolin, Ruggero Frezza Computer Science Department UCLA.
FEATURE PERFORMANCE COMPARISON FEATURE PERFORMANCE COMPARISON y SC is a training set of k-dimensional observations with labels S and C b C is a parameter.
Pseudo-Relevance Feedback For Multimedia Retrieval By Rong Yan, Alexander G. and Rong Jin Mwangi S. Kariuki
Ziming Zhang, Yucheng Zhao and Yiwen Wan.  Introduction&Motivation  Problem Statement  Paper Summeries  Discussion and Conclusions.
Hand Gesture for Taking Self Portrait Shaowei Chu and Jiro Tanaka University of Tsukuba Japan 12th July 15 minutes talk.
By: Ryan Wendel.  It is an ongoing analysis in which videos are analyzed frame by frame  Most of the video recognition is pulled from 3-D graphic engines.
Víctor Ponce Miguel Reyes Xavier Baró Mario Gorga Sergio Escalera Two-level GMM Clustering of Human Poses for Automatic Human Behavior Analysis Departament.
SmartPlayer: User-Centric Video Fast-Forwarding K.-Y. Cheng, S.-J. Luo, B.-Y. Chen, and H.-H. Chu ACM CHI 2009 (international conference on Human factors.
Actions in video Monday, April 25 Kristen Grauman UT-Austin.
M. Emre Sargın, Ferda Ofli, Yelena Yasinnik, Oya Aran, Alexey Karpov, Stephen Wilson,Engin Erzin, Yücel Yemez, A. Murat Tekalp Combined Gesture- Speech.
 INTRODUCTION  STEPS OF GESTURE RECOGNITION  TRACKING TECHNOLOGIES  SPEECH WITH GESTURE  APPLICATIONS.
SOMM: Self Organizing Markov Map for Gesture Recognition Pattern Recognition 2010 Spring Seung-Hyun Lee G. Caridakis et al., Pattern Recognition, Vol.
Move With Me S.W Graduation Project An Najah National University Engineering Faculty Computer Engineering Department Supervisor : Dr. Raed Al-Qadi Ghada.
Computer and Robot Vision I
Haojie Li Jinhui Tang Si Wu Yongdong Zhang Shouxun Lin Automatic Detection and Analysis of Player Action in Moving Background Sports Video Sequences IEEE.
Stanford hci group / cs376 research topics in human-computer interaction Vision-based Interaction Scott Klemmer 17 November 2005.
MVL (Machine Vision Lab) UIC HUMAN MOTION VIDEO DATABASE Jezekiel Ben-Arie ECE Department University Of Illinois at Chicago Scripts, Queries, Recognition.
Instructor : Dr. K. R. Rao Presented by: Rajesh Radhakrishnan.
Viewpoint Tracking for 3D Display Systems A look at the system proposed by Yusuf Bediz, Gözde Bozdağı Akar.
Learning and Recognizing Activities in Streams of Video Dinesh Govindaraju.
Machine Learning Damon Waring 22 April of 15 Agenda Problem, Solution, Benefits Problem, Solution, Benefits Machine Learning Overview/Basics Machine.
Knowledge Systems Lab JN 8/24/2015 A Method for Temporal Hand Gesture Recognition Joshua R. New Knowledge Systems Laboratory Jacksonville State University.
Action and Gait Recognition From Recovered 3-D Human Joints IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS— PART B: CYBERNETICS, VOL. 40, NO. 4, AUGUST.
Topic regards: ◆ Browsing of Search Results ◆ Video Retrieval using Spatio-Temporal ◆ Object Tracking ◆ Face tracking Yuan-Hao Lai.
Feature and object tracking algorithms for video tracking Student: Oren Shevach Instructor: Arie nakhmani.
Multimedia Databases (MMDB)
Knowledge Systems Lab JN 9/10/2002 Computer Vision: Gesture Recognition from Images Joshua R. New Knowledge Systems Laboratory Jacksonville State University.
Multimedia Specification Design and Production 2013 / Semester 2 / week 8 Lecturer: Dr. Nikos Gazepidis
Cut-And-Stitch: Efficient Parallel Learning of Linear Dynamical Systems on SMPs Lei Li Computer Science Department School of Computer Science Carnegie.
3D Motion Capture Assisted Video human motion recognition based on the Layered HMM Myunghoon Suk & Ashok Ramadass Advisor : Dr. B. Prabhakaran Multimedia.
7-Speech Recognition Speech Recognition Concepts
A Method for Hand Gesture Recognition Jaya Shukla Department of Computer Science Shiv Nadar University Gautam Budh Nagar, India Ashutosh Dwivedi.
Introduction EE 520: Image Analysis & Computer Vision.
An Information Fusion Approach for Multiview Feature Tracking Esra Ataer-Cansizoglu and Margrit Betke ) Image and.
Recognition, Analysis and Synthesis of Gesture Expressivity George Caridakis IVML-ICCS.
Research Projects 6v81 Multimedia Database Yohan Jin, T.A.
Online Kinect Handwritten Digit Recognition Based on Dynamic Time Warping and Support Vector Machine Journal of Information & Computational Science, 2015.
Using Inactivity to Detect Unusual behavior Presenter : Siang Wang Advisor : Dr. Yen - Ting Chen Date : Motion and video Computing, WMVC.
Relative Hidden Markov Models Qiang Zhang, Baoxin Li Arizona State University.
Vehicle Segmentation and Tracking From a Low-Angle Off-Axis Camera Neeraj K. Kanhere Committee members Dr. Stanley Birchfield Dr. Robert Schalkoff Dr.
Action and Gait Recognition From Recovered 3-D Human Joints IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS— PART B: CYBERNETICS, VOL. 40, NO. 4, AUGUST.
A NOVEL METHOD FOR COLOR FACE RECOGNITION USING KNN CLASSIFIER
Spatiotemporal Saliency Map of a Video Sequence in FPGA hardware David Boland Acknowledgements: Professor Peter Cheung Mr Yang Liu.
Digital Video Library Network Supervisor: Prof. Michael Lyu Student: Ma Chak Kei, Jacky.
Team Members Ming-Chun Chang Lungisa Matshoba Steven Preston Supervisors Dr James Gain Dr Patrick Marais.
Chapter 8. Learning of Gestures by Imitation in a Humanoid Robot in Imitation and Social Learning in Robots, Calinon and Billard. Course: Robots Learning.
IEEE International Conference on Multimedia and Expo.
Learning video saliency from human gaze using candidate selection CVPR2013 Poster.
3D Motion Classification Partial Image Retrieval and Download Multimedia Project Multimedia and Network Lab, Department of Computer Science.
Portable Camera-Based Assistive Text and Product Label Reading From Hand-Held Objects for Blind Persons.
Motion tracking TEAM D, Project 11: Laura Gui - Timisoara Calin Garboni - Timisoara Peter Horvath - Szeged Peter Kovacs - Debrecen.
Visual Information Processing. Human Perception V.S. Machine Perception  Human perception: pictorial information improvement for human interpretation.
MAV Optical Navigation Software System April 30, 2012 Tom Fritz, Pamela Warman, Richard Woodham Jr, Justin Clark, Andre DeRoux Sponsor: Dr. Adrian Lauf.
Hand Gestures Based Applications
A. M. R. R. Bandara & L. Ranathunga
Visual Information Retrieval
Horn-Schunk Optical Flow
AHED Automatic Human Emotion Detection
Gait Recognition Gökhan ŞENGÜL.
Saliency-guided Video Classification via Adaptively weighted learning
Computational NeuroEngineering Lab
Vehicle Segmentation and Tracking in the Presence of Occlusions
Image Segmentation Techniques
AHED Automatic Human Emotion Detection
AHED Automatic Human Emotion Detection
Visual Recognition of American Sign Language Using Hidden Markov Models 문현구 문현구.
Week 3 Volodymyr Bobyr.
Presentation transcript:

Video-based human motion recognition using 3D mocap data Myunghoon Suk Advisor : Dr. B. Prabhakaran Multimedia and Networking Lab The University of Texas at Dallas

The University of Texas at Dallas Contents Background KBH method Extended Optical Flow algorithm Compare with Baum-Welch method Conclusion Project examples November 9, 2018 The University of Texas at Dallas

The University of Texas at Dallas Background Video-based human motion recognition is a prerequisite for many of useful applications Gaming, Automatic visual surveillance, Human-computer interaction (HCI), and so on. November 9, 2018 The University of Texas at Dallas

The University of Texas at Dallas Background Hidden Markov Models(HMMs) Popularly been employed in video human motion recognition 2D feature data Easy to take it, but very Noisy 3D Motion Capture(MOCAP) data Expensive to get it, but very Clean and Semantic November 9, 2018 The University of Texas at Dallas

The University of Texas at Dallas System Concept Training Section 2D FEATURE EXTRACTOR Body Position, Direction, … 2D Feature Vector 3D MOCAP Data Head position Testing Section Classifier HMM New Actions HMM Fall down HMM HMM … Non-Fall down November 9, 2018 The University of Texas at Dallas

Optical Flow Algorithm Optical flow or optic flow is the pattern of apparent motion of objects, surfaces, and edges in a visual scene caused by the relative motion between an observer (an eye or a camera) and the scene. Given a set of points in an image, finding those same points in another image is the basis of optical flow algorithms Set of points – Good features to track algorithm Track the given points – Pyramidal Lucas Kanade Algorithm.

Extended Optical Flow Algorithm Objective Eliminating Noisy Features by determining useful features. Determining low level feature data including speed and direction of the human motion. Definitions Noisy Features –> All the Features that are tracked outside the person and certain features that are tracking on the person. Useful Features - >Features that show considerable movement in their position spatially with respect to a sequence of images Techniques Frame Jump -> To determine the useful features Threshold -> To remove the noisy features Feature extraction November 9, 2018

Extended Optical Flow November 9, 2018

Feature Extraction Six types of low level feature data: Average x and y feature coordinate position of the useful features. Speed of the moving body in terms of x and y coordinates Direction of the moving body Euclidean distance of separation of the useful features between two frames. November 9, 2018

The University of Texas at Dallas What is KBH? KBH(Knowledge-Based Hybrid) Method The proposed method to use 3D MOCAP data as Knowledge Base Hidden state comes from mixing 2D feature vector and 3D MOCAP data, when creating HMMs for actions Provide higher accuracy without learning which is doing iteratively November 9, 2018 The University of Texas at Dallas

The University of Texas at Dallas 3D Motion Capture data (1 column) How does it work? (1) 2D feature data (6 columns) Frame # Hybrid data (7 columns) Hidden state and Observation seq. Creating HMMs November 9, 2018 The University of Texas at Dallas

Baum-Welch (EM) method Learning ( Parameter Estimation ) Given an observation sequence O, find the model that is most likely to produce the sequences November 9, 2018 The University of Texas at Dallas

Comparison between KBH and BW Baum-Welch Given data Observation Sequence State Sequence 2D feature data 3D mocap data Iteration No Yes Compute directly a model parameter λ Expectation & Maximization Repeat Learning Time About 10 seconds About 4 hrs Accuracy* 91.2% 81.5% November 9, 2018 The University of Texas at Dallas

The University of Texas at Dallas Experiment Results Data Set # of Video motion clips : 647 # of Motion capture clips : 8 # of subjects : 5 # of motion classes : 8 crawl, fall down, jump, pick up, run, sit down, stand up, and walk 2D feature vector (6 dimensions) x, y coordinates, differential information, speed, direction 3D motion capture data (1 dimension) Head position data: x,y,z coordinates November 9, 2018 The University of Texas at Dallas

The University of Texas at Dallas Conclusion KBH method has advantages such as reducing learning time and providing high accuracy compared to traditional Baum-Welch method for HMMs in video-based human motion recognition. November 9, 2018 The University of Texas at Dallas

Project 1: Object Tracking DESCRIPTION: Object tracking is one of fundamental techniques in computer vision. The feature data obtained from object tracking can be used for action recognition later. REQUIREMENT: Basic knowledge of OpenCV library Developing an application (e.g. simple game) by implementing the existing object tracking algorithm. REFERENCES: Yilmaz, A., Javed, O., and Shah, M. 2006. “Object tracking: A survey.” ACM Comput. Surv. 38, 4 (Dec. 2006), 13. (http://portal.acm.org/beta/citation.cfm?id=1177355) November 9, 2018 The University of Texas at Dallas

Project 2: Gesture Recognition DESCRIPTION: Video-based human gesture recognition is a very interesting topic in computer vision and Machine Learning. It can be applied to many applications such as game, surveillance, or user interface of computer. REQUIREMENT: Implementing a machine learning algorithm (e.g. Finite State Machine (FSM)) as a recognizer. REFERENCES: Pengyu Hong, Thomas S. Huang, Matthew Turk, "Gesture Modeling and Recognition Using Finite State Machines," fg, pp.410, Fourth IEEE International Conference on Automatic Face and Gesture Recognition (FG'00), 2000 November 9, 2018 The University of Texas at Dallas

Thank you Myunghoon Suk Email: mhsuk@utallas.edu Lab: ECSS 4.620