Motion Features for Action Recognition YeHao 3/11/2014.

Slides:



Advertisements
Similar presentations
Feature Detection. Description Localization More Points Robust to occlusion Works with less texture More Repeatable Robust detection Precise localization.
Advertisements

Simultaneous surveillance camera calibration and foot-head homology estimation from human detection 1 Author : Micusic & Pajdla Presenter : Shiu, Jia-Hau.
Outline Feature Extraction and Matching (for Larger Motion)
MASKS © 2004 Invitation to 3D vision Lecture 7 Step-by-Step Model Buidling.
Activity Recognition Aneeq Zia. Agenda What is activity recognition Typical methods used for action recognition “Evaluation of local spatio-temporal features.
Matching with Invariant Features
Computer Vision Optical Flow
A New Block Based Motion Estimation with True Region Motion Field Jozef Huska & Peter Kulla EUROCON 2007 The International Conference on “Computer as a.
3D Computer Vision and Video Computing 3D Vision Topic 4 of Part II Visual Motion CSc I6716 Fall 2011 Cover Image/video credits: Rick Szeliski, MSR Zhigang.
Personal Driving Diary: Constructing a Video Archive of Everyday Driving Events IEEE workshop on Motion and Video Computing ( WMVC) 2011 IEEE Workshop.
LUCAS KANADE FEATURE TRACKER a pyramidal implementation
Announcements Quiz Thursday Quiz Review Tomorrow: AV Williams 4424, 4pm. Practice Quiz handout.
Direct Methods for Visual Scene Reconstruction Paper by Richard Szeliski & Sing Bing Kang Presented by Kristin Branson November 7, 2002.
Feature matching and tracking Class 5 Read Section 4.1 of course notes Read Shi and Tomasi’s paper on.
Visual motion Many slides adapted from S. Seitz, R. Szeliski, M. Pollefeys.
Visual motion Many slides adapted from S. Seitz, R. Szeliski, M. Pollefeys.
Motion Estimation Today’s Readings Trucco & Verri, 8.3 – 8.4 (skip 8.3.3, read only top half of p. 199) Numerical Recipes (Newton-Raphson), 9.4 (first.
COMP 290 Computer Vision - Spring Motion II - Estimation of Motion field / 3-D construction from motion Yongjik Kim.
3D Rigid/Nonrigid RegistrationRegistration 1)Known features, correspondences, transformation model – feature basedfeature based 2)Specific motion type,
Matching Compare region of image to region of image. –We talked about this for stereo. –Important for motion. Epipolar constraint unknown. But motion small.
KLT tracker & triangulation Class 6 Read Shi and Tomasi’s paper on good features to track
Optical Flow Digital Photography CSE558, Spring 2003 Richard Szeliski (notes cribbed from P. Anandan)
3D Computer Vision and Video Computing 3D Vision Topic 8 of Part 2 Visual Motion (II) CSC I6716 Spring 2004 Zhigang Zhu, NAC 8/203A
Summary of Previous Lecture A homography transforms one 3d plane to another 3d plane, under perspective projections. Those planes can be camera imaging.
Automatic Camera Calibration
Feature Tracking and Optical Flow
CSE 473/573 Computer Vision and Image Processing (CVIP) Ifeoma Nwogu Lecture 35 – Review for midterm.
Action recognition with improved trajectories
TP15 - Tracking Computer Vision, FCUP, 2013 Miguel Coimbra Slides by Prof. Kristen Grauman.
Exploiting video information for Meeting Structuring ….
The Brightness Constraint
Tzu ming Su Advisor : S.J.Wang MOTION DETAIL PRESERVING OPTICAL FLOW ESTIMATION 2013/1/28 L. Xu, J. Jia, and Y. Matsushita. Motion detail preserving optical.
Flow Separation for Fast and Robust Stereo Odometry [ICRA 2009]
Video-Vigilance and Biometrics
Introduction EE 520: Image Analysis & Computer Vision.
CSCE 643 Computer Vision: Structure from Motion
Mentor: Salman Khokhar Action Recognition in Crowds Week 7.
CSE 185 Introduction to Computer Vision Feature Tracking and Optical Flow.
Visual motion Many slides adapted from S. Seitz, R. Szeliski, M. Pollefeys.
Recognizing Action at a Distance Alexei A. Efros, Alexander C. Berg, Greg Mori, Jitendra Malik Computer Science Division, UC Berkeley Presented by Pundik.
December 9, 2014Computer Vision Lecture 23: Motion Analysis 1 Now we will talk about… Motion Analysis.
Feature Reconstruction Using Lucas-Kanade Feature Tracking and Tomasi-Kanade Factorization EE7740 Project I Dr. Gunturk.
Sean M. Ficht.  Problem Definition  Previous Work  Methods & Theory  Results.
Jack Pinches INFO410 & INFO350 S INFORMATION SCIENCE Computer Vision I.
Virtual Image Peephole By Kyle Patience Supervisor: Reg Dodds Co Supervisor: Mehrdad Ghaziasgar.
IEEE International Conference on Multimedia and Expo.
Linearizing (assuming small (u,v)): Brightness Constancy Equation: The Brightness Constraint Where:),(),(yxJyxII t  Each pixel provides 1 equation in.
Line Matching Jonghee Park GIST CV-Lab..  Lines –Fundamental feature in many computer vision fields 3D reconstruction, SLAM, motion estimation –Useful.
776 Computer Vision Jan-Michael Frahm Spring 2012.
Motion / Optical Flow II Estimation of Motion Field Avneesh Sud.
Person Following with a Mobile Robot Using Binocular Feature-Based Tracking Zhichao Chen and Stanley T. Birchfield Dept. of Electrical and Computer Engineering.
2D to 3D Conversion Using 3D Database For Football Scenes Kiana Calagari Final Project of CMPT880 July 2013.
Optical flow and keypoint tracking Many slides adapted from S. Seitz, R. Szeliski, M. Pollefeys.
Linearizing (assuming small (u,v)): Brightness Constancy Equation: The Brightness Constraint Where:),(),(yxJyxII t  Each pixel provides 1 equation in.
Recognizing specific objects Matching with SIFT Original suggestion Lowe, 1999,2004.
SIFT.
Action Recognition in Video
776 Computer Vision Jan-Michael Frahm Spring 2012.
SIFT Scale-Invariant Feature Transform David Lowe
Feature Tracking and Optical Flow
CSE 577 Image and Video Analysis
The Brightness Constraint
The Brightness Constraint
Two-Stream Convolutional Networks for Action Recognition in Videos
Effective and Efficient Detection of Moving Targets From a UAV’s Camera
The SIFT (Scale Invariant Feature Transform) Detector and Descriptor
Coupled Horn-Schunck and Lukas-Kanade for image processing
AHED Automatic Human Emotion Detection
Optical flow Computer Vision Spring 2019, Lecture 21
Optical flow and keypoint tracking
Presentation transcript:

Motion Features for Action Recognition YeHao 3/11/2014

Motion Information DNN Dense Trajectory

Trajectory Tracking Interest Points – Tracking Harris3D Interest point KLT Tracker: Sparse Interest Points. – Matching SIFT descriptor Computationally Expensive

KLT Tracker 1. Three Assumptions: – Intensity – Velocity – Space

Derivation of KLT Tracker

Derivation of KLT Tracker (II)

Derivation of KLT Tracker (III)

Good Features to Track Intuitively, a good feature needs at least: – Texture – Corner But what does this mean formally? Shi/Tomasi. Intuitive result really part of motion equation. – High eigenvalues imply reliable solvability. Good Feature has big eigenvalues, imply: Texture Corner

Dense Trajectory

Optical Flow: OpenCV Drifting: Limit to L frames Sudden large displacement: Remove

Dense Trajectory

Trajectory-aligned Descriptors Histograms of Oriented Gradients (HOG) – Appearance Information Histograms of Optical Flow (HOF) – Local motion information Motion Boundary Histogram (MBH) – Relative motion between pixels

Camera Motion

Improved Trajectory

Camera Motion Estimation – Two consecutive frames are related by a homography Match SURF feature Match optical flow vector Estimate homography by RANSAC

Homography Conditions Both images are viewing the same plane from a different angle Both images are taken from the same camera but from a different angle Homography relationship is independent of the scene structure – It does not depend on what the cameras are looking at – Relationship holds regardless of what is seen in the images

Homography The homography relates the pixel co- ordinates in two images if x’ = M x When applied to every pixel the new image is a warped version of the original image

Homography Consider a point x = (u,v,1) in one image and x’=(u’,v’,1) in another image A homography is a 3 by 3 matrix M

Removing inconsistent matches due to humans

Video Classification with Convolutional Neural Networks

Two-Stream Convolutional Networks

Optical Flow

Stacking Optical FlowTrajectory

Accuracy

Reference [1] H. Wang, “Evaluation of local spatio-temporal features for action recognition,” presented at the CRV '12: Proceedings of the 2012 Ninth Conference on Computer and Robot Vision, 2012, pp. 468–475. [2] H. Wang, A. Klaser, C. Schmid, and C.-L. Liu, “Action recognition by dense trajectories,” presented at the Computer Vision and Pattern Recognition (CVPR), 2011 IEEE Conference on, 2011, pp. 3169–3176. [3] A. Karpathy, G. Toderici, and S. Shetty, “Large-scale video classification with convolutional neural networks,” … on Computer Vision …, [4] K. Simonyan and A. Zisserman, “Two-Stream Convolutional Networks for Action Recognition in Videos,” arXiv.org, vol. cs.CV. 09-Jun-2014.