A Convergent Solution to Tensor Subspace Learning.

Slides:



Advertisements
Similar presentations
Active Appearance Models
Advertisements

Eigen Decomposition and Singular Value Decomposition
Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
Tensors and Component Analysis Musawir Ali. Tensor: Generalization of an n-dimensional array Vector: order-1 tensor Matrix: order-2 tensor Order-3 tensor.
Graph Embedding and Extensions: A General Framework for Dimensionality Reduction Keywords: Dimensionality reduction, manifold learning, subspace learning,
Dimension reduction (1)
A Constraint Generation Approach to Learning Stable Linear Dynamical Systems Sajid M. Siddiqi Byron Boots Geoffrey J. Gordon Carnegie Mellon University.
1cs542g-term High Dimensional Data  So far we’ve considered scalar data values f i (or interpolated/approximated each component of vector values.
Principal Component Analysis CMPUT 466/551 Nilanjan Ray.
Principal Component Analysis
Linear Transformations
Symmetric Matrices and Quadratic Forms
Principal Component Analysis
Dimension reduction : PCA and Clustering Agnieszka S. Juncker Slides: Christopher Workman and Agnieszka S. Juncker Center for Biological Sequence Analysis.
Face Recognition using PCA (Eigenfaces) and LDA (Fisherfaces)
Real-time Combined 2D+3D Active Appearance Models Jing Xiao, Simon Baker,Iain Matthew, and Takeo Kanade CVPR 2004 Presented by Pat Chan 23/11/2004.
Chapter 3 Determinants and Matrices
Introduction Given a Matrix of distances D, (which contains zeros in the main diagonal and is squared and symmetric), find variables which could be able,
Dimension reduction : PCA and Clustering Christopher Workman Center for Biological Sequence Analysis DTU.
3D Geometry for Computer Graphics
6 1 Linear Transformations. 6 2 Hopfield Network Questions.
1cs542g-term Notes  Extra class next week (Oct 12, not this Friday)  To submit your assignment: me the URL of a page containing (links to)
Stats & Linear Models.
Linear Algebra Review By Tim K. Marks UCSD Borrows heavily from: Jana Kosecka Virginia de Sa (UCSD) Cogsci 108F Linear.
Nonlinear Dimensionality Reduction Approaches. Dimensionality Reduction The goal: The meaningful low-dimensional structures hidden in their high-dimensional.
Introduction The central problems of Linear Algebra are to study the properties of matrices and to investigate the solutions of systems of linear equations.
Manifold learning: Locally Linear Embedding Jieping Ye Department of Computer Science and Engineering Arizona State University
Summarized by Soo-Jin Kim
Enhancing Tensor Subspace Learning by Element Rearrangement
Presented By Wanchen Lu 2/25/2013
Linear Algebra Review 1 CS479/679 Pattern Recognition Dr. George Bebis.
1 Graph Embedding (GE) & Marginal Fisher Analysis (MFA) 吳沛勳 劉冠成 韓仁智
Feature extraction 1.Introduction 2.T-test 3.Signal Noise Ratio (SNR) 4.Linear Correlation Coefficient (LCC) 5.Principle component analysis (PCA) 6.Linear.
Graph Embedding: A General Framework for Dimensionality Reduction Dong XU School of Computer Engineering Nanyang Technological University
1 Recognition by Appearance Appearance-based recognition is a competing paradigm to features and alignment. No features are extracted! Images are represented.
IEEE TRANSSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE
Quantum One: Lecture Representation Independent Properties of Linear Operators 3.
A Tensorial Approach to Access Cognitive Workload related to Mental Arithmetic from EEG Functional Connectivity Estimates S.I. Dimitriadis, Yu Sun, K.
Exploring Intrinsic Structures from Samples: Supervised, Unsupervised, and Semisupervised Frameworks Supervised by Prof. Xiaoou Tang & Prof. Jianzhuang.
Chapter 5 Eigenvalues and Eigenvectors 大葉大學 資訊工程系 黃鈴玲 Linear Algebra.
Computer Vision Lab. SNU Young Ki Baik Nonlinear Dimensionality Reduction Approach (ISOMAP, LLE)
SINGULAR VALUE DECOMPOSITION (SVD)
Transductive Regression Piloted by Inter-Manifold Relations.
Diagonalization and Similar Matrices In Section 4.2 we showed how to compute eigenpairs (,p) of a matrix A by determining the roots of the characteristic.
Manifold learning: MDS and Isomap
Spectral Sequencing Based on Graph Distance Rong Liu, Hao Zhang, Oliver van Kaick {lrong, haoz, cs.sfu.ca {lrong, haoz, cs.sfu.ca.
EE4-62 MLCV Lecture Face Recognition – Subspace/Manifold Learning Tae-Kyun Kim 1 EE4-62 MLCV.
Review of Matrix Operations Vector: a sequence of elements (the order is important) e.g., x = (2, 1) denotes a vector length = sqrt(2*2+1*1) orientation.
Feature Extraction 主講人:虞台文. Content Principal Component Analysis (PCA) PCA Calculation — for Fewer-Sample Case Factor Analysis Fisher’s Linear Discriminant.
2D-LDA: A statistical linear discriminant analysis for image matrix
Ultra-high dimensional feature selection Yun Li
4.8 Rank Rank enables one to relate matrices to vectors, and vice versa. Definition Let A be an m  n matrix. The rows of A may be viewed as row vectors.
Chapter 61 Chapter 7 Review of Matrix Methods Including: Eigen Vectors, Eigen Values, Principle Components, Singular Value Decomposition.
Principal Components Analysis ( PCA)
Dimension reduction (2) EDR space Sliced inverse regression Multi-dimensional LDA Partial Least Squares Network Component analysis.
Spectral Methods for Dimensionality
Principal Component Analysis (PCA)
Introduction The central problems of Linear Algebra are to study the properties of matrices and to investigate the solutions of systems of linear equations.
Intrinsic Data Geometry from a Training Set
Review of Matrix Operations
Lecture 8:Eigenfaces and Shared Features
Lecture: Face Recognition and Feature Reduction
Lecture 21 SVD and Latent Semantic Indexing and Dimensional Reduction
Principal Component Analysis
Spectral Clustering Eric Xing Lecture 8, August 13, 2010
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Feature space tansformation methods
Symmetric Matrices and Quadratic Forms
Symmetric Matrices and Quadratic Forms
Marios Mattheakis and Pavlos Protopapas
Presentation transcript:

A Convergent Solution to Tensor Subspace Learning

Concept Tensor Subspace Learning. Concept Tensor: multi-dimensional (or multi-way) arrays of components

Application Tensor Subspace Learning. Application real-world data are affected by multifarious factors for the person identification, we may have facial images of different ► views and poses ► lightening conditions ► expressions the observed data evolve differently along the variation of different factors ► image columns and rows

Application Tensor Subspace Learning. Application it is desirable to dig through the intrinsic connections among different affection factors of the data. Tensor provides a concise and effective representation. Illumination pose expression Image columns Image rows Images

Tensor Subspace Learning algorithms Traditional Tensor Discriminant algorithms Tensor Subspace Analysis He et.al Two-dimensional Linear Discriminant Analysis Discriminant Analysis with Tensor Representation Ye et.al Yan et.al project the tensor along different dimensions or ways projection matrices for different dimensions are derived iteratively solve an trace ratio optimization problem DO NOT CONVERGE !

Tensor Subspace Learning algorithms Graph Embedding – a general framework An undirected intrinsic graph G={X,W} is constructed to represent the pairwise similarities over sample data. A penalty graph or a scale normalization item is constructed to impose extra constraints on the transform. intrinsic graph penalty graph

Discriminant Analysis Objective Solve the projection matrices iteratively: leave one projection matrix as variable while keeping others as constant. No closed form solution Mode-k unfolding of the tensor

Objective Deduction Discriminant Analysis Objective Trace Ratio: General Formulation for the objectives of the Discriminant Analysis based Algorithms. DATER: TSA: Within Class Scatter of the unfolded data Between Class Scatter of the unfolded data Diagonal Matrix with weights Constructed from Image Manifold

Disagreement between the Objective and the Optimization Process Why do previous algorithms not converge? GEVD The conversion from Trace Ratio to Ratio Trace induces an inconsistency among the objectives of different dimensions!

from Trace Ratio to Trace Difference What will we do? from Trace Ratio to Trace Difference Objective: Define Then Trace Ratio Trace Difference Find So that

from Trace Ratio to Trace Difference What will we do? from Trace Ratio to Trace Difference Constraint Let We have Thus The Objective rises monotonously! Projection matrices of different dimensions share the same objective Whereare the leading eigen vectors of.

Main Algorithm Process Main Algorithm 1: Initialization. Initialize as arbitrary column orthogonal matrices. 2: Iterative optimization. For t=1, 2,..., Tmax, Do For k=1, 2,..., n, Do 1. Set. 2. Compute and. 3. Conduct Eigenvalue Decomposition: 4. Reshape the projection directions 5. 3: Output the projection matrices

Strict Monotony and Convergency If Then Thus orthogonal matrix Meanwhile, both and have been calculated from the leading eigenvectors of the projected samples. span the same space So The algorithm is strictly monotonous! Theorem[Meyer,1976]: Assume that the algorithm Ω is strictly monotonic with respect to J and it generates a sequence which lies in a compact set. If is normed, then. The algorithm is guaranteed to converge!

Hightlights of the Trace Ratio based algorithm Highlights of our algorithm The objective value is guaranteed to monotonously increase; and the multiple projection matrices are proved to converge. Only eigenvalue decomposition method is applied for iterative optimization, which makes the algorithm extremely efficient. Enhanced potential classification capability of the derived low- dimensional representation from the subspace learning algorithms. The algorithm does not suffer from the singularity problem that is often encountered by the traditional generalized eigenvalue decomposition method used to solve the ratio trace optimization problem.

Monotony of the Objective Experimental Results The traditional ratio trace based procedure does not converge, while our new solution procedure guarantees the monotonous increase of the objective function value and commonly our new procedure will converge after about 4-10 iterations. Moreover, the final converged value of the objective function from our new procedure is much larger than the value of the objective function for any iteration of the ratio trace based procedure.

Convergency of the Projection Matrices Experimental Results The projection matrices converge after 4-10 iterations for our new solution procedure; while for the traditional procedure, heavy oscillations exist and the solution does not converge.

Face Recognition Results Experimental Results 1. TMFA TR mostly outperforms all the other methods concerned in this work, with only one exception for the case G5P5 on the CMU PIE database. 2. For vector-based algorithms, the trace ratio based formulation is consistently superior to the ratio trace based one for subspace learning. 3. Tensor representation has the potential to improve the classification performance for both trace ratio and ratio trace formulations of subspace learning.

Summary A novel iterative procedure was proposed to directly optimize the objective function of general subspace learning based on tensor representation. The convergence of the projection matrices and the monotony property of the objective function value were proven. The first work to give a convergent solution for the general tensor-based subspace learning.

Thank You!