Download presentation

Presentation is loading. Please wait.

Published byLeila Songer Modified over 2 years ago

1
A CTION R ECOGNITION FROM V IDEO U SING F EATURE C OVARIANCE M ATRICES Kai Guo, Prakash Ishwar, Senior Member, IEEE, and Janusz Konrad, Fellow, IEEE

2
O UTLINE Introduction Framework Action Feature Experiments Conclusion

3
I NTRODUCTION A new approach to action representation—one based on the empirical covariance matrix of a bag of local action features. We apply the covariance matrix representation to two types of local feature collections: 1. A sequence of silhouettes of an object (the so – called silhouette tunnel) 2. The optical flow.

4
I NTRODUCTION We focus on two distinct types of classifiers: 1. The nearest-neighbor (NN) classifier. 2. the sparse-linear approximation (SLA) classifier. Transformation of the supervised classification problem in the closed convex cone of covariance matrices into an equivalent problem in the vector space of symmetric matrices via the matrix logarithm.

5
F RAMEWORK Feature Covariance Matrices We adopt a “bag of dense local feature vectors” modeling approach. Inspired by Tuzel et al. ’s work, the feature- covariance matrix can provide a very discriminative representation for action recognition.

6
F RAMEWORK Let F = { f n } denote a “bag of feature vectors” extracted from a video sample, the size of the feature set | F | be N. The empirical estimate of the covariance estimate of the covariance matrix of F is given by: Where is the empirical mean feature vector.

7
F RAMEWORK Log-Covariance Matrices A key idea is to map the convex cone of covariance matrices to the vector space of symmetric matrices1 by using the matrix logarithm proposed by Arsigny et al.. The eigen-decomposition of C is given by C = Then log (C) :=, where is a diagonal matrix obtained from D by replacing D ’s diagonal entries by their logarithms.

8
F RAMEWORK Classification Using Log-Covariance Matrices Nearest-Neighbor (NN) Classification: Given a query sample, find the most similar sample in the annotated training set, where similarity is measured with respect to some distance measure, and assign its label to the query sample.

9
F RAMEWORK Sparse Linear Approximation (SLA) Classification: We approximate the log-covariance matrix of a query sample by a sparse linear combination of log- covariance matrices of all training samples p 1,..., p N.

10
F RAMEWORK Given a query sample, one may attempt to express it as a linear combination of training samples by solving the matrix-vector equation given by By solving the following NP-hard optimization problem: If the optimal solution α ∗ is sufficiently sparse: This difficulty can be overcome by introducing a noise term as follows: where z is an additive noise term whose length is assumed to be bounded by ε, This leads to the following -minimization problem:

11
F RAMEWORK Use a reconstruction residual error (RRE) measure to decide the query class. Let α ∗ denote the coefficients associated with class i (having label li ), corresponding to columns of training matrix Pi. The RRE measure of class i is defined as : To annotate the sample we assign the class label that leads to the minimum RRE

12
A CTION F EATURE Silhouette Tunnel Shape Features Our goal is to reliably discriminate between shapes; not to accurately reconstruct them. Hence a coarse, low- dimensional representation of shape would suffice. We capture the shape of the 3D silhouette tunnel by the empirical covariance matrix of a bag of thirteen- dimensional local shape features.

13
A CTION F EATURE We associate the following 13-dimensional feature vector f ( s ) that captures certain shape characteristics of the tunnel:

14
A CTION F EATURE After obtaining 13-dimensional silhouette shape feature vectors, we can compute their 13 × 13 covariance matrix, denoted by C, using (1) (with N = | S |): Where is the mean feature vector. Thus, C is an empirical covariance matrix of the collection of vectors F.

15
A CTION F EATURE Optical Flow Features Here we use a variant of the Horn and Schunck method, which optimizes a functional based on residuals from the intensity constraints and a smoothness regularization term. Let I (x, y, t) denote the luminance of the raw video sequence at pixel position (x, y, t) and let u (x, y, t) represent the corresponding optical flow vector. Based on I (x, y, t) and u (x, y, t), we use the following feature vector f (x, y, t) :

16
E XPERIMENTS

23
C ONCLUSION The action recognition framework that we have developed in this paper is conceptually simple, easy to implement, has good run-time performance. The TRECVID [63] and VIRAT [64] video datasets exemplify these types of realworld challenges and much work remains to be done to address them.

24
C ONCLUSION Our method’s relative simplicity, as compared to some of the top methods in the literature, enables almost tuning-free rapid deployment and real- time operation. This opens new application areas outside the traditional surveillance/security arena, for example in sports video annotation and customizable human-computer interaction.

25
T HE E ND

Similar presentations

OK

Over-fitting and Regularization Chapter 4 textbook Lectures 11 and 12 on amlbook.com.

Over-fitting and Regularization Chapter 4 textbook Lectures 11 and 12 on amlbook.com.

© 2017 SlidePlayer.com Inc.

All rights reserved.

Ads by Google

Ppt on acute coronary syndrome icd-9 Ppt on chapter 3 drainage cap Ppt on views in dbms functions Ppt on ideal gas law density Ppt on 2 dimensional figures and 3 dimensional slides Ppt on lost city of atlantis mystery Edit ppt on iphone Ppt on un security council Ppt on project life cycle phases Ppt on suspension type insulators home