Graph Embedding and Extensions: A General Framework for Dimensionality Reduction Keywords: Dimensionality reduction, manifold learning, subspace learning,

Slides:



Advertisements
Similar presentations
Text mining Gergely Kótyuk Laboratory of Cryptography and System Security (CrySyS) Budapest University of Technology and Economics
Advertisements

M. Belkin and P. Niyogi, Neural Computation, pp. 1373–1396, 2003.
1 Manifold Alignment for Multitemporal Hyperspectral Image Classification H. Lexie Yang 1, Melba M. Crawford 2 School of Civil Engineering, Purdue University.
Aggregating local image descriptors into compact codes
SVM - Support Vector Machines A new classification method for both linear and nonlinear data It uses a nonlinear mapping to transform the original training.
AGE ESTIMATION: A CLASSIFICATION PROBLEM HANDE ALEMDAR, BERNA ALTINEL, NEŞE ALYÜZ, SERHAN DANİŞ.
One-Shot Multi-Set Non-rigid Feature-Spatial Matching
Pattern Recognition and Machine Learning
Support Vector Machines (SVMs) Chapter 5 (Duda et al.)
Principal Component Analysis
Dimensional reduction, PCA
Three Algorithms for Nonlinear Dimensionality Reduction Haixuan Yang Group Meeting Jan. 011, 2005.
A Global Geometric Framework for Nonlinear Dimensionality Reduction Joshua B. Tenenbaum, Vin de Silva, John C. Langford Presented by Napat Triroj.
Atul Singh Junior Undergraduate CSE, IIT Kanpur.  Dimension reduction is a technique which is used to represent a high dimensional data in a more compact.
NonLinear Dimensionality Reduction or Unfolding Manifolds Tennenbaum|Silva|Langford [Isomap] Roweis|Saul [Locally Linear Embedding] Presented by Vikas.
Lightseminar: Learned Representation in AI An Introduction to Locally Linear Embedding Lawrence K. Saul Sam T. Roweis presented by Chan-Su Lee.
Nonlinear Dimensionality Reduction Approaches. Dimensionality Reduction The goal: The meaningful low-dimensional structures hidden in their high-dimensional.
Representative Previous Work
Manifold learning: Locally Linear Embedding Jieping Ye Department of Computer Science and Engineering Arizona State University
Enhancing Tensor Subspace Learning by Element Rearrangement
1 Graph Embedding (GE) & Marginal Fisher Analysis (MFA) 吳沛勳 劉冠成 韓仁智
General Tensor Discriminant Analysis and Gabor Features for Gait Recognition by D. Tao, X. Li, and J. Maybank, TPAMI 2007 Presented by Iulian Pruteanu.
Graph Embedding: A General Framework for Dimensionality Reduction Dong XU School of Computer Engineering Nanyang Technological University
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology Adaptive nonlinear manifolds and their applications to pattern.
计算机学院 计算感知 Support Vector Machines. 2 University of Texas at Austin Machine Learning Group 计算感知 计算机学院 Perceptron Revisited: Linear Separators Binary classification.
IEEE TRANSSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE
Learning a Kernel Matrix for Nonlinear Dimensionality Reduction By K. Weinberger, F. Sha, and L. Saul Presented by Michael Barnathan.
Computer Vision Lab. SNU Young Ki Baik Nonlinear Dimensionality Reduction Approach (ISOMAP, LLE)
A Two-level Pose Estimation Framework Using Majority Voting of Gabor Wavelets and Bunch Graph Analysis J. Wu, J. M. Pedersen, D. Putthividhya, D. Norgaard,
ECE 8443 – Pattern Recognition LECTURE 10: HETEROSCEDASTIC LINEAR DISCRIMINANT ANALYSIS AND INDEPENDENT COMPONENT ANALYSIS Objectives: Generalization of.
Transductive Regression Piloted by Inter-Manifold Relations.
Local Fisher Discriminant Analysis for Supervised Dimensionality Reduction Presented by Xianwang Wang Masashi Sugiyama.
Computational Intelligence: Methods and Applications Lecture 23 Logistic discrimination and support vectors Włodzisław Duch Dept. of Informatics, UMK Google:
GRASP Learning a Kernel Matrix for Nonlinear Dimensionality Reduction Kilian Q. Weinberger, Fei Sha and Lawrence K. Saul ICML’04 Department of Computer.
Spoken Language Group Chinese Information Processing Lab. Institute of Information Science Academia Sinica, Taipei, Taiwan
Manifold learning: MDS and Isomap
SemiBoost : Boosting for Semi-supervised Learning Pavan Kumar Mallapragada, Student Member, IEEE, Rong Jin, Member, IEEE, Anil K. Jain, Fellow, IEEE, and.
Nonlinear Dimensionality Reduction Approach (ISOMAP)
H. Lexie Yang1, Dr. Melba M. Crawford2
EE4-62 MLCV Lecture Face Recognition – Subspace/Manifold Learning Tae-Kyun Kim 1 EE4-62 MLCV.
Optimal Dimensionality of Metric Space for kNN Classification Wei Zhang, Xiangyang Xue, Zichen Sun Yuefei Guo, and Hong Lu Dept. of Computer Science &
Tony Jebara, Columbia University Advanced Machine Learning & Perception Instructor: Tony Jebara.
A Convergent Solution to Tensor Subspace Learning.
CZ5225: Modeling and Simulation in Biology Lecture 7, Microarray Class Classification by Machine learning Methods Prof. Chen Yu Zong Tel:
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 12: Advanced Discriminant Analysis Objectives:
Dimensionality reduction
Math 285 Project Diffusion Maps Xiaoyan Chong Department of Mathematics and Statistics San Jose State University.
June 25-29, 2006ICML2006, Pittsburgh, USA Local Fisher Discriminant Analysis for Supervised Dimensionality Reduction Masashi Sugiyama Tokyo Institute of.
2D-LDA: A statistical linear discriminant analysis for image matrix
A Kernel Approach for Learning From Almost Orthogonal Pattern * CIS 525 Class Presentation Professor: Slobodan Vucetic Presenter: Yilian Qin * B. Scholkopf.
Nonlinear Dimension Reduction: Semi-Definite Embedding vs. Local Linear Embedding Li Zhang and Lin Liao.
Out of sample extension of PCA, Kernel PCA, and MDS WILSON A. FLORERO-SALINAS DAN LI MATH 285, FALL
Support Vector Machines (SVMs) Chapter 5 (Duda et al.) CS479/679 Pattern Recognition Dr. George Bebis.
Spectral Methods for Dimensionality
Principal Component Analysis (PCA)
Deeply learned face representations are sparse, selective, and robust
LECTURE 09: BAYESIAN ESTIMATION (Cont.)
Unsupervised Riemannian Clustering of Probability Density Functions
کاربرد نگاشت با حفظ تنکی در شناسایی چهره
CS 2750: Machine Learning Dimensionality Reduction
Machine Learning Basics
Outline Nonlinear Dimension Reduction Brief introduction Isomap LLE
Image Retrieval Longin Jan Latecki.
Learning with information of features
CS 2750: Machine Learning Support Vector Machines
Feature space tansformation methods
Generally Discriminant Analysis
Globally Maximizing Locally Minimizing unsupervised discriminant projection with applications to face and palm biometrics PAMI 2007 Bo Yang 2/25/2019.
Nonlinear Dimension Reduction:
NonLinear Dimensionality Reduction or Unfolding Manifolds
Presentation transcript:

Graph Embedding and Extensions: A General Framework for Dimensionality Reduction Keywords: Dimensionality reduction, manifold learning, subspace learning, graph embedding framework.

1.Introduction Techniques for dimensionality reduction Linear: PCA/LDA/LPP... Nonlinear: ISOMAP/Laplacian Eigenmap/LLE... Linear  Nonlinear: kernel trick Graph embedding framework A unified view for understanding and explaining many popular algorithms such as the ones mentioned above. A platform for developing new dimension reduction algorithms.

2.Graph embedding 2.1Graph embedding Let m is often very large so we need to find Intrinsic graph: --similarity matrix Penalty graph: --the similarity to be suppressed in the dimension-reduced feature space Y

Our graph-preserving criterion is: L is called Laplacian matrix B typically is diagonal for scale normalization or L- matrix of the penalty graph

Linearization: Kernelization: Both can be obtained by solving:

Tensorization: 2.2General Framework for Dimensionality Reduction

The adjacency graphs for PCA and LDA. (a) Constraint and intrinsic graph in PCA. (b) Penalty and intrinsic graphs in LDA.

2.3 Related Works and Discussions Kernel Interpretation and Out-of-Sample Extension Ham et al. [13] proposed a kernel interpretation of KPCA,ISOMAP, LLE, and Laplacian Eigenmap Bengio et al. [4] presented a method for computing the low dimensional representation of out-of-sample data. Comparison: Kernel Interpretation Graph embeding normalized similarity matrix laplacian matrix unsupervised learning both supervised&unsupervised

2.3.2 Brand’s Work [5] Brand’s Work can be viewed as a special case of the graph embedding framework

2.3.3 Laplacian Eigenmap [3] and LPP [10] Single graph B=D Nonnegative similarity matrix Although [10] attempts to use LPP to explain PCA and LDA, this explanation is incomplete. The constraint matrix B is fixed to D in LPP, while the constraint matrix of LDA is comes from a penalty graph that connects all samples with equal weights;hence, LPP cannot explain LPP. Also,a minimization algorithm, does not explain why PCA maximizes the objective function.

3 MARGINAL FISHER ANALYSIS 3.1 Marginal Fisher Analysis Limitation of LDA:data distribution assumption limited available projection directions MFA overcomed the limitation by characterizing intraclass compactness and interclass separability. intrinsic graph: each sample is connected to its k1 nearest neighbors of the same class (intraclass compactness) penalty graph: each sample is connected to its k2 nearest neighbors of other classes (interclass separability)

Procedure of MFA PCA projection Constructing the intraclass compactness and interclass separability graphs. Marginal Fisher Criterion Output the final linear projection direction

The available projection directions are much greater than that of LDA There is no assumption on the data distribution of each class Without prior information on data distributions Advantages of MFA

KMFA Projection direction: The distance between sample xi and xj is For a new data point x, its projection to the derived optimal direction is obtained as

TMFA:

4.Experiments 4.1face recognition MFA>Fisherface(LDA+PCA)>PCA PCA+MFA>PCA+LDA>PCA Kernel trick KDA>LDA,KMFA>MFA KMFA>PCA,Fisherface,LPP

Trainingset Adequate: LPP > Fisherface,PCA Inadequate: Fisherface > LPP>PCA anyway, MFA>=LPP Performance can be substantially improved by exploring a certain range of PCA dimensions first. PCA+MFA>MFA,Bayesian face >PCA,Fisherface,LPP Tensor representation brings encouraging improvements compared with vector-based algorithms it is critical to collect sufficient samples for all subjects!

4.2 A Non-Gaussian Case

5.CONCLUSION AND FUTURE WORK All possible extensions of the algorithms mentioned in this paper Combination of the kernel trick and tensorization The selection of parameters k1 and k2 How to utilize higher order statistics of the data set in the graph embedding framework?