Action and Gait Recognition From Recovered 3-D Human Joints IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS— PART B: CYBERNETICS, VOL. 40, NO. 4, AUGUST.

Slides:



Advertisements
Similar presentations
We consider situations in which the object is unknown the only way of doing pose estimation is then building a map between image measurements (features)
Advertisements

Evidential modeling for pose estimation Fabio Cuzzolin, Ruggero Frezza Computer Science Department UCLA.
Recognizing Human Actions by Attributes CVPR2011 Jingen Liu, Benjamin Kuipers, Silvio Savarese Dept. of Electrical Engineering and Computer Science University.
By: Ryan Wendel.  It is an ongoing analysis in which videos are analyzed frame by frame  Most of the video recognition is pulled from 3-D graphic engines.
Reducing Drift in Parametric Motion Tracking
3D Human Body Pose Estimation from Monocular Video Moin Nabi Computer Vision Group Institute for Research in Fundamental Sciences (IPM)
International Conference on Automatic Face and Gesture Recognition, 2006 A Layered Deformable Model for Gait Analysis Haiping Lu, K.N. Plataniotis and.
Intelligent Systems Lab. Recognizing Human actions from Still Images with Latent Poses Authors: Weilong Yang, Yang Wang, and Greg Mori Simon Fraser University,
Real-Time Human Pose Recognition in Parts from Single Depth Images Presented by: Mohammad A. Gowayyed.
Silhouette Lookup for Automatic Pose Tracking N ICK H OWE.
Abandoned Object Detection for Public Surveillance Video Student: Wei-Hao Tung Advisor: Jia-Shung Wang Dept. of Computer Science National Tsing Hua University.
Adviser:Ming-Yuan Shieh Student:shun-te chuang SN:M
Adviser : Ming-Yuan Shieh Student ID : M Student : Chung-Chieh Lien VIDEO OBJECT SEGMENTATION AND ITS SALIENT MOTION DETECTION USING ADAPTIVE BACKGROUND.
Shape and Dynamics in Human Movement Analysis Ashok Veeraraghavan.
Haojie Li Jinhui Tang Si Wu Yongdong Zhang Shouxun Lin Automatic Detection and Analysis of Player Action in Moving Background Sports Video Sequences IEEE.
Broadcast Court-Net Sports Video Analysis Using Fast 3-D Camera Modeling Jungong Han Dirk Farin Peter H. N. IEEE CSVT 2008.
Modeling Pixel Process with Scale Invariant Local Patterns for Background Subtraction in Complex Scenes (CVPR’10) Shengcai Liao, Guoying Zhao, Vili Kellokumpu,
Exchanging Faces in Images SIGGRAPH ’04 Blanz V., Scherbaum K., Vetter T., Seidel HP. Speaker: Alvin Date: 21 July 2004.
Shape and Dynamics in Human Movement Analysis Ashok Veeraraghavan.
Motion Detection And Analysis Michael Knowles Tuesday 13 th January 2004.
Multiple Human Objects Tracking in Crowded Scenes Yao-Te Tsai, Huang-Chia Shih, and Chung-Lin Huang Dept. of EE, NTHU International Conference on Pattern.
3D Hand Pose Estimation by Finding Appearance-Based Matches in a Large Database of Training Views
Object Detection and Tracking Mike Knowles 11 th January 2005
Student: Hsu-Yung Cheng Advisor: Jenq-Neng Hwang, Professor
Effective Gaussian mixture learning for video background subtraction Dar-Shyang Lee, Member, IEEE.
Gait Recognition Simon Smith Jamie Hutton Thomas Moore David Newman.
Hand Signals Recognition from Video Using 3D Motion Capture Archive Tai-Peng Tian Stan Sclaroff Computer Science Department B OSTON U NIVERSITY I. Introduction.
Real-Time Decentralized Articulated Motion Analysis and Object Tracking From Videos Wei Qu, Member, IEEE, and Dan Schonfeld, Senior Member, IEEE.
Face Recognition and Retrieval in Video Basic concept of Face Recog. & retrieval And their basic methods. C.S.E. Kwon Min Hyuk.
Extracting Places and Activities from GPS Traces Using Hierarchical Conditional Random Fields Yong-Joong Kim Dept. of Computer Science Yonsei.
Tracking Pedestrians Using Local Spatio- Temporal Motion Patterns in Extremely Crowded Scenes Louis Kratz and Ko Nishino IEEE TRANSACTIONS ON PATTERN ANALYSIS.
Shape Recognition and Pose Estimation for Mobile Augmented Reality Author : N. Hagbi, J. El-Sana, O. Bergig, and M. Billinghurst Date : Speaker.
3D Fingertip and Palm Tracking in Depth Image Sequences
Prakash Chockalingam Clemson University Non-Rigid Multi-Modal Object Tracking Using Gaussian Mixture Models Committee Members Dr Stan Birchfield (chair)
Shape-Based Human Detection and Segmentation via Hierarchical Part- Template Matching Zhe Lin, Member, IEEE Larry S. Davis, Fellow, IEEE IEEE TRANSACTIONS.
Motion Object Segmentation, Recognition and Tracking Huiqiong Chen; Yun Zhang; Derek Rivait Faculty of Computer Science Dalhousie University.
Learning and Recognizing Human Dynamics in Video Sequences Christoph Bregler Alvina Goh Reading group: 07/06/06.
A General Framework for Tracking Multiple People from a Moving Camera
International Conference on Intelligent and Advanced Systems 2007 Chee-Ming Ting Sh-Hussain Salleh Tian-Swee Tan A. K. Ariff. Jain-De,Lee.
Person detection, tracking and human body analysis in multi-camera scenarios Montse Pardàs (UPC) ACV, Bilkent University, MTA-SZTAKI, Technion-ML, University.
Scientific Writing Abstract Writing. Why ? Most important part of the paper Number of Readers ! Make people read your work. Sell your work. Make your.
An Information Fusion Approach for Multiview Feature Tracking Esra Ataer-Cansizoglu and Margrit Betke ) Image and.
Jun-Won Suh Intelligent Electronic Systems Human and Systems Engineering Department of Electrical and Computer Engineering Speaker Verification System.
Vision-based human motion analysis: An overview Computer Vision and Image Understanding(2007)
Introduction to Video Background Subtraction 1. Motivation In video action analysis, there are many popular applications like surveillance for security,
Human pose recognition from depth image MS Research Cambridge.
Action and Gait Recognition From Recovered 3-D Human Joints IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS— PART B: CYBERNETICS, VOL. 40, NO. 4, AUGUST.
Efficient Visual Object Tracking with Online Nearest Neighbor Classifier Many slides adapt from Steve Gu.
Chapter 5 Multi-Cue 3D Model- Based Object Tracking Geoffrey Taylor Lindsay Kleeman Intelligent Robotics Research Centre (IRRC) Department of Electrical.
P RW GEI: Poisson Random Walk based Gait Recognition Intelligent Systems Research Centre School of Computing and Intelligent Systems,
Rick Parent - CIS681 Motion Analysis – Human Figure Processing video to extract information of objects Motion tracking Pose reconstruction Motion and subject.
Face Image-Based Gender Recognition Using Complex-Valued Neural Network Instructor :Dr. Dong-Chul Kim Indrani Gorripati.
 Present by 陳群元.  Introduction  Previous work  Predicting motion patterns  Spatio-temporal transition distribution  Discerning pedestrians  Experimental.
Skeleton Based Action Recognition with Convolutional Neural Network
Hand Gesture Recognition Using Haar-Like Features and a Stochastic Context-Free Grammar IEEE 高裕凱 陳思安.
Big traffic data processing framework for intelligent monitoring and recording systems 學生 : 賴弘偉 教授 : 許毅然 作者 : Yingjie Xia a, JinlongChen a,b,n, XindaiLu.
Journal of Visual Communication and Image Representation
CSCI 631 – Foundations of Computer Vision March 15, 2016 Ashwini Imran Image Stitching.
Motion tracking TEAM D, Project 11: Laura Gui - Timisoara Calin Garboni - Timisoara Peter Horvath - Szeged Peter Kovacs - Debrecen.
CSCI 631 – Foundations of Computer Vision March 15, 2016 Ashwini Imran Image Stitching Link: singhashwini.mesinghashwini.me.
Bayesian Decision Theory Case Studies CS479/679 Pattern Recognition Dr. George Bebis.
Gait Recognition Gökhan ŞENGÜL.
Bag-of-Visual-Words Based Feature Extraction
Saliency-guided Video Classification via Adaptively weighted learning
Motion Detection And Analysis
Real-Time Human Pose Recognition in Parts from Single Depth Image
Dynamical Statistical Shape Priors for Level Set Based Tracking
Video-based human motion recognition using 3D mocap data
PRAKASH CHOCKALINGAM, NALIN PRADEEP, AND STAN BIRCHFIELD
Oral presentation for ACM International Conference on Multimedia, 2014
Presentation transcript:

Action and Gait Recognition From Recovered 3-D Human Joints IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS— PART B: CYBERNETICS, VOL. 40, NO. 4, AUGUST 2010 Junxia Gu, Member, IEEE, Xiaoqing Ding, Senior Member, IEEE, Shengjin Wang, Member, IEEE, and Youshou Wu Adviser : Ming-Yuan Shieh Student : shun-te chuang ppt 製作: 100%

Outline ABSTRACT INTRODUCTION PREVIOUS WORK FULL-BODY TRACKING METHOD FOR POSE RECOVERY CLASSIFICATION EXPERIMENTAL RESULTS AND ANALYSIS CONCLUSION

Abstract A common viewpoint-free framework that fuses pose recovery and classification for action and gait recognition is presented in this paper. First, a markerless pose recovery method is adopted to automatically capture the 3-D human joint and pose parameter sequences from volume data. Second, multiple configuration features (combination of joints) and movement features (position, orientation, and height of the body) are extracted from the recovered 3-D human joint and pose parameter sequences.

Abstract A hidden Markov model (HMM) and an exemplar- based HMM are then used to model the movement features and configuration features, respectively. Finally, actions are classified by a hierarchical classifier that fuses the movement features and the configuration features, and persons are recognized from their gait sequences with the configuration features.

INTRODUCTION VIDEO-BASED study of human motion has been receiving increased attention in the past decades. This has been motivated by the desire for application of intelligent video surveillance and human– computer interaction. With increased awareness in security issues, motion analysis is becoming increasingly important in surveillance systems.

INTRODUCTION Action recognition is a new requirement for understanding what the person is doing. Current intelligent surveillance systems are in urgent need of noninvasive and viewpoint-free research on motion analysis. This paper focuses on the movement of main body segments (arms, legs, and torso). A human gait is extracted from a “walk” action.

INTRODUCTION In this paper, a vision-based markerless pose recovery approach is proposed to extract 3-D human joints. Human joint sequence is one of the most effective and discriminative representations of human motion. It contains much information, including position, orientation, and joint position. The information is categorized into two types: movement features and configuration features.

INTRODUCTION The changes of position, orientation, and height of the body, which describe the global movement of the subject, are defined as movement features. The sequences of human joint positions, which describe the change of relative configuration of body segments, are defined as configuration features. A hidden Markov model (HMM) and an exemplar-based HMM (EHMM) are employed to characterize the movement and configuration features, respectively. Both the HMM and the EHMM have been used to recognize an action and a gait.

INTRODUCTION Fig. 1. Flowchart of the video-based motion recognition.

PREVIOUS WORK 1. Appearance-Based Methods : Appearance-based approaches are widely used in action and gait representation. They directly represent human motion using image information, such as a silhouette, an edge, and an optical flow. 2. Human Model-Based Methods : Human model-based approaches represent an action or a gait with body segments, joint positions, or pose parameters.

PREVIOUS WORK This paper combined stochastic search with a gradient descent for local pose refinement to recover complex whole body motion. The initialization of the model was automatic, with an initialization pose standing upright with his/her arms and legs spread in the “Da Vinci” pose. The tracking speed was below 1 s per frame. In this paper, an adaptive particle filter method is proposed for pose recovery.

PREVIOUS WORK First, the whole body of a subject in each frame is segmented into several body segments. A particle filter with an adaptive particle number is then used to track each body segment. This method decomposes the search space and reduces the computational complexity.

FULL-BODY TRACKING METHOD FOR POSE RECOVERY Human Model Human Model-Based Full-Body Tracking

Human Model

Human Model-Based Full-Body Tracking

CLASSIFICATION A. HMM and EHMM Learning B. Classifier for Gait Recognition C. Classifier for Action Recognition

A. HMM and EHMM Learning The EHMM is different from the HMM in the definition of observation densities. For the HMM, the general representation of observation densities is the Gaussian mixture model (GMM) of the following form: In the EHMM, the definition of the observation probability is as follows:

B. Classifier for Gait Recognition For each person c ∈ {1,..., Cp} in the database, we learn an EHMM gait model λ (c) Sgait with features Sgait and an EHMM gait model λ( c) Lgait with features Lgait. A testing gait sequence Y = {y0,..., yK} is classified with the following MAP estimation:

C. Classifier for Action Recognition Testing action sequence Y = {y0,..., yK}, with features SY, RY, PY, OY, and HY, is classified with a two-layer classifier fused multiple features. The first layer is a weighted-MAP classifier that fuses three movement features and the configuration feature of the whole body as If the decision c1 of the first layer belongs to single-arm actions, sequence Y will be recognized by the second MAP classifier with arm features as

EXPERIMENTAL RESULTS AND ANALYSIS There are 11 actions, and each subject plays each action three times. All samples contain approximately frames in total. These actions include “check watch,” “cross arm,” “scratch head,” “sit down,” “get up,” “turn around,” “walk in a circle,” “wave hand,” “punch,” “kick,” and “pick up,”.To demonstrate view invariance, subjects freely change their orientations. The acquisition is achieved using five standard firewire cameras. The image resolution is 390 × 291 pixels, and the volume of interest is divided into 64 × 64 × 64 voxels.

EXPERIMENTAL RESULTS AND ANALYSIS Fig. 8. Samples of images and 3-D volume data.

EXPERIMENTAL RESULTS AND ANALYSIS Fig. 9. Annotation of the human joint. (a) Annotation of the knee joint. (b) Results of annotation.

EXPERIMENTAL RESULTS AND ANALYSIS Fig. 11. Average error of individual joint positions. Fig. 10. Average error of the joint positions.

EXPERIMENTAL RESULTS AND ANALYSIS Fig. 12. Results of the pose recovery.

EXPERIMENTAL RESULTS AND ANALYSIS Fig. 13. Selected exemplars and recognition rate versus the number of exemplars.

EXPERIMENTAL RESULTS AND ANALYSIS Fig. 14. Comparison of convergence performance between the EHMM and the HMM.

EXPERIMENTAL RESULTS AND ANALYSIS Fig. 15. Average recognition rates of actions.

CONCLUSION The main contribution of this paper is the fusion of pose recovery and motion recognition. Future work plans include automatically segmenting temporal sequence, reducing computational complexity, analyzing actions that are more complex, and recognizing 2-D actions based on the 3-D EHMM. The free-viewpoint 3-D human joint sequence contains a significant amount of information for motion analysis. In addition to representing the single actions used in this paper, it can be used for more applications, such as analysis of complex actions.

CONCLUSION High DOF and huge 3-D points make the human model-based pose recovery method very time consuming. To solve this problem, parallel computing, code optimization, and a GPU can be used to lessen time cost. It is difficult to obtain robust volume data of subjects in surveillance and content analysis scenarios at present. Actions/gaits are affected by different factors, including clothing, age, and gender. In the future, performance with these factors will be analyzed in larger databases.