A Bayesian Formulation For 3d Articulated Upper Body Segmentation And Tracking From Dense Disparity Maps Navin Goel Dr Ara V Nefian Dr George Bebis.

Slides:



Advertisements
Similar presentations
Evidential modeling for pose estimation Fabio Cuzzolin, Ruggero Frezza Computer Science Department UCLA.
Advertisements

Probabilistic Inverse Dynamics for Nonlinear Blood Pattern Reconstruction Benjamin Cecchetto, Wolfgang Heidrich University of British Columbia.
Pose Estimation and Segmentation of People in 3D Movies Karteek Alahari, Guillaume Seguin, Josef Sivic, Ivan Laptev Inria, Ecole Normale Superieure ICCV.
For Internal Use Only. © CT T IN EM. All rights reserved. 3D Reconstruction Using Aerial Images A Dense Structure from Motion pipeline Ramakrishna Vedantam.
Change Detection C. Stauffer and W.E.L. Grimson, “Learning patterns of activity using real time tracking,” IEEE Trans. On PAMI, 22(8): , Aug 2000.
Hilal Tayara ADVANCED INTELLIGENT ROBOTICS 1 Depth Camera Based Indoor Mobile Robot Localization and Navigation.
Multiple Frame Motion Inference Using Belief Propagation Jiang Gao Jianbo Shi Presented By: Gilad Kapelushnik Visual Recognition, Spring 2005, Technion.
MASKS © 2004 Invitation to 3D vision Lecture 7 Step-by-Step Model Buidling.
A Modified EM Algorithm for Hand Gesture Segmentation in RGB-D Data 2014 IEEE International Conference on Fuzzy Systems (FUZZ-IEEE) July 6-11, 2014, Beijing,
Segmentation and Fitting Using Probabilistic Methods
Foreground Background detection from video Foreground Background detection from video מאת : אבישג אנגרמן.
Instructor: Dr. G. Bebis Reza Amayeh Fall 2005
Motion Detection And Analysis Michael Knowles Tuesday 13 th January 2004.
Tracking with Online Appearance Model Bohyung Han
Segmentation and Tracking of Multiple Humans in Crowded Environments Tao Zhao, Ram Nevatia, Bo Wu IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE,
A Bayesian algorithm for tracking multiple moving objects in outdoor surveillance video Department of Electrical Engineering and Computer Science The University.
Prénom Nom Document Analysis: Data Analysis and Clustering Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Object Recognition by Parts Object recognition started with line segments. - Roberts recognized objects from line segments and junctions. - This led to.
Multiple Human Objects Tracking in Crowded Scenes Yao-Te Tsai, Huang-Chia Shih, and Chung-Lin Huang Dept. of EE, NTHU International Conference on Pattern.
Object Detection and Tracking Mike Knowles 11 th January 2005
1 Integration of Background Modeling and Object Tracking Yu-Ting Chen, Chu-Song Chen, Yi-Ping Hung IEEE ICME, 2006.
FACE RECOGNITION, EXPERIMENTS WITH RANDOM PROJECTION
Effective Gaussian mixture learning for video background subtraction Dar-Shyang Lee, Member, IEEE.
A Probabilistic Framework for Video Representation Arnaldo Mayer, Hayit Greenspan Dept. of Biomedical Engineering Faculty of Engineering Tel-Aviv University,
Pattern Recognition. Introduction. Definitions.. Recognition process. Recognition process relates input signal to the stored concepts about the object.
Tracking with Linear Dynamic Models. Introduction Tracking is the problem of generating an inference about the motion of an object given a sequence of.
Algorithm Evaluation and Error Analysis class 7 Multiple View Geometry Comp Marc Pollefeys.
Face Processing System Presented by: Harvest Jang Group meeting Fall 2002.
Segmentation and tracking of the upper body model from range data with applications in hand gesture recognition Navin Goel Intel Corporation Department.
Object Recognition by Parts Object recognition started with line segments. - Roberts recognized objects from line segments and junctions. - This led to.
I mage and M edia U nderstanding L aboratory for Performance Evaluation of Vision-based Real-time Motion Capture Naoto Date, Hiromasa Yoshimoto, Daisaku.
Statistical Color Models (SCM) Kyungnam Kim. Contents Introduction Trivariate Gaussian model Chromaticity models –Fixed planar chromaticity models –Zhu.
Feature and object tracking algorithms for video tracking Student: Oren Shevach Instructor: Arie nakhmani.
3D Fingertip and Palm Tracking in Depth Image Sequences
Multimodal Interaction Dr. Mike Spann
Prakash Chockalingam Clemson University Non-Rigid Multi-Modal Object Tracking Using Gaussian Mixture Models Committee Members Dr Stan Birchfield (chair)
BraMBLe: The Bayesian Multiple-BLob Tracker By Michael Isard and John MacCormick Presented by Kristin Branson CSE 252C, Fall 2003.
1 Physical Fluctuomatics 5th and 6th Probabilistic information processing by Gaussian graphical model Kazuyuki Tanaka Graduate School of Information Sciences,
Object Stereo- Joint Stereo Matching and Object Segmentation Computer Vision and Pattern Recognition (CVPR), 2011 IEEE Conference on Michael Bleyer Vienna.
Kevin Cherry Robert Firth Manohar Karki. Accurate detection of moving objects within scenes with dynamic background, in scenarios where the camera is.
Vision-based human motion analysis: An overview Computer Vision and Image Understanding(2007)
A Region Based Stereo Matching Algorithm Using Cooperative Optimization Zeng-Fu Wang, Zhi-Gang Zheng University of Science and Technology of China Computer.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: ML and Simple Regression Bias of the ML Estimate Variance of the ML Estimate.
Non-Photorealistic Rendering and Content- Based Image Retrieval Yuan-Hao Lai Pacific Graphics (2003)
Expectation-Maximization (EM) Case Studies
CVPR2013 Poster Detecting and Naming Actors in Movies using Generative Appearance Models.
CS Statistical Machine learning Lecture 24
Computer Vision Lecture 6. Probabilistic Methods in Segmentation.
Lecture 2: Statistical learning primer for biologists
Chapter 20 Classification and Estimation Classification – Feature selection Good feature have four characteristics: –Discrimination. Features.
Presented by: Idan Aharoni
Visual Tracking by Cluster Analysis Arthur Pece Department of Computer Science University of Copenhagen
Stereo Vision Local Map Alignment for Robot Environment Mapping Computer Vision Center Dept. Ciències de la Computació UAB Ricardo Toledo Morales (CVC)
Discriminative Training and Machine Learning Approaches Machine Learning Lab, Dept. of CSIE, NCKU Chih-Pin Liao.
Hidden Markov Models. A Hidden Markov Model consists of 1.A sequence of states {X t |t  T } = {X 1, X 2,..., X T }, and 2.A sequence of observations.
Hand Detection with a Cascade of Boosted Classifiers Using Haar-like Features Qing Chen Discover Lab, SITE, University of Ottawa May 2, 2006.
May 2003 SUT Color image segmentation – an innovative approach Amin Fazel May 2003 Sharif University of Technology Course Presentation base on a paper.
Visual Recognition Tutorial1 Markov models Hidden Markov models Forward/Backward algorithm Viterbi algorithm Baum-Welch estimation algorithm Hidden.
Zhaoxia Fu, Yan Han Measurement Volume 45, Issue 4, May 2012, Pages 650–655 Reporter: Jing-Siang, Chen.
Gaussian Mixture Model classification of Multi-Color Fluorescence In Situ Hybridization (M-FISH) Images Amin Fazel 2006 Department of Computer Science.
11/25/03 3D Model Acquisition by Tracking 2D Wireframes Presenter: Jing Han Shiau M. Brown, T. Drummond and R. Cipolla Department of Engineering University.
Urban Scene Analysis James Elder & Patrick Denis York University.
Motion Detection And Analysis
A segmentation and tracking algorithm
Object Recognition by Parts
PRAKASH CHOCKALINGAM, NALIN PRADEEP, AND STAN BIRCHFIELD
“grabcut”- Interactive Foreground Extraction using Iterated Graph Cuts
Computer and Robot Vision I
EM Algorithm and its Applications
Object Recognition by Parts
Presentation transcript:

A Bayesian Formulation For 3d Articulated Upper Body Segmentation And Tracking From Dense Disparity Maps Navin Goel Dr Ara V Nefian Dr George Bebis

Overview Introduction Gesture Recognition System 3D Upper Body Model Initialization Segmentation Tracking Results Future work

Introduction Applications: Human-Computer Interaction Systems, American Sign Language Recognition, Security-Monitoring Systems, Gesture Recognition. Requirements: Background and scene independent, Robust hand free initialization, Robust tracking.

Introduction Related Work: A. V. Nefian, R. Grzeszczuk, and V. Eruhimov. A statistical upper body model for 3D static and dynamic gesture recognition from stereo sequences. In IEEE International Conference on Image Processing, volume 3, pages , Success: 96% Accuracy.

Gesture Recognition System:

Assumptions: One user in the camera view, The torso is the largest visible component, The torso plane is approximately perpendicular to the camera, The user’s head is approximately in vertical position.

Observation Vectors: Observation Vectors = depth + color Upper Body Model Pixel OdOd OcOc Threshold is chosen in such a way that the foreground pixels consist mostly on the upper body pixels.

Head: Where, N is a Gaussian density function, µ H and C H are mean and covariance of gaussian density function. U is a uniform distribution function, H is Head & S H is the size of head. Upper Body Model

Torso: Where, Upper Body Model T is Torso & W T is Width of Torso

Arms – Linear PDF Where, R F is radius of Fore Arm, F l is Left fore arm, ρ ij & σ is mean and covariance of Gaussian. Upper Body Model

EM Algorithm The E Step Compute: P(O ij | J,C) The M step finds: J,C = arg max(P(O ij | J,C)) Where, J are Joints & C are Components of upper body model

Head & Torso Starting with an initial position of the neck, the parameters of the torso plane are estimated below the neck plane using the EM approach. Using the torso center and anthropological measures of the head, parameters of the head are estimated using the EM approach. Estimate the position of Neck: N(x) = μ He (x) N(y) = μ He (y) + S H /2 N(z) = aN(x) + bN(y) + c Initialization

Head & Torso Estimate the position of Shoulders: S l (x) = µ T (x) + W T /2 S l (y) = N(y) S l (z) = aS l (x) + bS l (y) + c & S r (x) = µ T (x) - W T /2 S r (y) = N(y) S l (z) = aS l (x) + bS l (y) + c Repeat the above steps for updated J B until convergence error falls under a threshold. Initialization

Arms For each possible arm joint configuration we estimate the mean of the linear pdfs corresponding to the upper and fore arms, and the mean of the normal pdf for hands. For each joint configuration of the arms, we determine the best state assignment of the observation vectors. Find the max likelihood over all joint configuration and determine the best set of joints and the corresponding best state assignment. Initialization

Head & Torso: Estimate head and torso parameters and position of joints using the algorithm in Initialization – 1. The value of neck is obtained from previous frame. TRACKING

Arms: Step 1: Find 4 points in sphere that are equidistant from original elbow position based on theta and phi values passed. These points constitute the new set of possible elbow joints to search in the current iteration. FOR(I = 0; I < 5; I++) Step 2: Find log likelihood of Probability Density Function (PDF) found by EM given 3d planar distance. Step 3: For each new elbow joint, find 4 new positions for wrist + original based on theta and phi values passed. FOR(J = 0; J < 5; J++) Step 4: Find log likelihood of Probability Density Function (PDF) found by EM given 3d planar distance. PDF is defined as start (elbow) and end (wrist) points of joints. TRACKING

Arms: Step 5: For each new wrist joint, find 4 new positions for hand center + original based on theta and phi values passed. FOR(K = 0; K < 5; K++) Step 6: Find log likelihood of Probability Density Function (PDF) found by EM given 3d planar distance. PDF is defined as start (wrist) and end (hand center) points of joints. Step 7: Calculate Max Tracking Log Likelihood of Single Arm. Calculate argmaxElWl{ P(Oij|{Ul, Fl, Hl},{El,Wl})} Convergence Error = Best Probability – Previous Probability Step 8: Repeat the above steps until convergence error. TRACKING

Success: Results

Success: Results

Failure: Results

Future Work: The system collapses when hand is adjacent to head. Integration with other systems. Learning anthropological measures of the person.

Acknowledgements: National Science Foundation University Of Nevada, Reno Intel Corporation, Santa Clara