Computer Vision Lab. SNU Young Ki Baik Nonlinear Dimensionality Reduction Approach (ISOMAP, LLE)

Slides:



Advertisements
Similar presentations
Coherent Laplacian 3D protrusion segmentation Oxford Brookes Vision Group Queen Mary, University of London, 11/12/2009 Fabio Cuzzolin.
Advertisements

Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
Manifold Learning Dimensionality Reduction. Outline Introduction Dim. Reduction Manifold Isomap Overall procedure Approximating geodesic dist. Dijkstra’s.
Presented by: Mingyuan Zhou Duke University, ECE April 3, 2009
1 A Survey on Distance Metric Learning (Part 1) Gerry Tesauro IBM T.J.Watson Research Center.
Graph Laplacian Regularization for Large-Scale Semidefinite Programming Kilian Weinberger et al. NIPS 2006 presented by Aggeliki Tsoli.
Non-linear Dimensionality Reduction CMPUT 466/551 Nilanjan Ray Prepared on materials from the book Non-linear dimensionality reduction By Lee and Verleysen,
University of Joensuu Dept. of Computer Science P.O. Box 111 FIN Joensuu Tel fax Isomap Algorithm.
“Random Projections on Smooth Manifolds” -A short summary
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Dimensionality R e d u c t i o n. Another unsupervised task Clustering, etc. -- all forms of data modeling Trying to identify statistically supportable.
Principal Component Analysis
LLE and ISOMAP Analysis of Robot Images Rong Xu. Background Intuition of Dimensionality Reduction Linear Approach –PCA(Principal Component Analysis) Nonlinear.
Image Manifolds : Learning-based Methods in Vision Alexei Efros, CMU, Spring 2007 © A.A. Efros With slides by Dave Thompson.
1 Numerical geometry of non-rigid shapes Spectral Methods Tutorial. Spectral Methods Tutorial 6 © Maks Ovsjanikov tosca.cs.technion.ac.il/book Numerical.
Manifold Learning: ISOMAP Alan O'Connor April 29, 2008.
Three Algorithms for Nonlinear Dimensionality Reduction Haixuan Yang Group Meeting Jan. 011, 2005.
Distance Metric Learning: A Comprehensive Survey
A Global Geometric Framework for Nonlinear Dimensionality Reduction Joshua B. Tenenbaum, Vin de Silva, John C. Langford Presented by Napat Triroj.
Atul Singh Junior Undergraduate CSE, IIT Kanpur.  Dimension reduction is a technique which is used to represent a high dimensional data in a more compact.
NonLinear Dimensionality Reduction or Unfolding Manifolds Tennenbaum|Silva|Langford [Isomap] Roweis|Saul [Locally Linear Embedding] Presented by Vikas.
Lightseminar: Learned Representation in AI An Introduction to Locally Linear Embedding Lawrence K. Saul Sam T. Roweis presented by Chan-Su Lee.
Nonlinear Dimensionality Reduction by Locally Linear Embedding Sam T. Roweis and Lawrence K. Saul Reference: "Nonlinear dimensionality reduction by locally.
CS 485/685 Computer Vision Face Recognition Using Principal Components Analysis (PCA) M. Turk, A. Pentland, "Eigenfaces for Recognition", Journal of Cognitive.
Nonlinear Dimensionality Reduction Approaches. Dimensionality Reduction The goal: The meaningful low-dimensional structures hidden in their high-dimensional.
Manifold learning: Locally Linear Embedding Jieping Ye Department of Computer Science and Engineering Arizona State University
Summarized by Soo-Jin Kim
Graph Embedding: A General Framework for Dimensionality Reduction Dong XU School of Computer Engineering Nanyang Technological University
1 Recognition by Appearance Appearance-based recognition is a competing paradigm to features and alignment. No features are extracted! Images are represented.
IEEE TRANSSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE
Data Reduction. 1.Overview 2.The Curse of Dimensionality 3.Data Sampling 4.Binning and Reduction of Cardinality.
Learning a Kernel Matrix for Nonlinear Dimensionality Reduction By K. Weinberger, F. Sha, and L. Saul Presented by Michael Barnathan.
THE MANIFOLDS OF SPATIAL HEARING Ramani Duraiswami | Vikas C. Raykar Perceptual Interfaces and Reality Lab University of Maryland, College park.
ISOMAP TRACKING WITH PARTICLE FILTER Presented by Nikhil Rane.
GRASP Learning a Kernel Matrix for Nonlinear Dimensionality Reduction Kilian Q. Weinberger, Fei Sha and Lawrence K. Saul ICML’04 Department of Computer.
CSE 185 Introduction to Computer Vision Face Recognition.
Dimensionality Reduction
Manifold learning: MDS and Isomap
CSC2535: Computation in Neural Networks Lecture 12: Non-linear dimensionality reduction Geoffrey Hinton.
Dimensionality Reduction Part 2: Nonlinear Methods
1 LING 696B: MDS and non-linear methods of dimension reduction.
Nonlinear Dimensionality Reduction Approach (ISOMAP)
Jan Kamenický.  Many features ⇒ many dimensions  Dimensionality reduction ◦ Feature extraction (useful representation) ◦ Classification ◦ Visualization.
Data Mining Course 0 Manifold learning Xin Yang. Data Mining Course 1 Outline Manifold and Manifold Learning Classical Dimensionality Reduction Semi-Supervised.
Non-Linear Dimensionality Reduction
Tony Jebara, Columbia University Advanced Machine Learning & Perception Instructor: Tony Jebara.
Data Projections & Visualization Rajmonda Caceres MIT Lincoln Laboratory.
Supervisor: Nakhmani Arie Semester: Winter 2007 Target Recognition Harmatz Isca.
Data Mining Course 2007 Eric Postma Clustering. Overview Three approaches to clustering 1.Minimization of reconstruction error PCA, nlPCA, k-means clustering.
Math 285 Project Diffusion Maps Xiaoyan Chong Department of Mathematics and Statistics San Jose State University.
CSC321: Lecture 25: Non-linear dimensionality reduction Geoffrey Hinton.
CS685 : Special Topics in Data Mining, UKY The UNIVERSITY of KENTUCKY Dimensionality Reduction CS 685: Special Topics in Data Mining Spring 2008 Jinze.
Nonlinear Dimension Reduction: Semi-Definite Embedding vs. Local Linear Embedding Li Zhang and Lin Liao.
CSC321: Extra Lecture (not on the exam) Non-linear dimensionality reduction Geoffrey Hinton.
Manifold Learning JAMES MCQUEEN – UW DEPARTMENT OF STATISTICS.
The UNIVERSITY of NORTH CAROLINA at CHAPEL HILL Dimensionality Reduction Part 1: Linear Methods Comp Spring 2007.
Multi-index Evaluation Algorithm Based on Locally Linear Embedding for the Node importance in Complex Networks Fang Hu
Machine Learning Supervised Learning Classification and Regression K-Nearest Neighbor Classification Fisher’s Criteria & Linear Discriminant Analysis Perceptron:
Eric Xing © Eric CMU, Machine Learning Data visualization and dimensionality reduction Eric Xing Lecture 7, August 13, 2010.
Spectral Methods for Dimensionality
Nonlinear Dimensionality Reduction
Unsupervised Riemannian Clustering of Probability Density Functions
CS 2750: Machine Learning Dimensionality Reduction
Dimensionality Reduction
ISOMAP TRACKING WITH PARTICLE FILTERING
Outline Nonlinear Dimension Reduction Brief introduction Isomap LLE
Dimensionality Reduction
Object Modeling with Layers
Nonlinear Dimension Reduction:
NonLinear Dimensionality Reduction or Unfolding Manifolds
Presentation transcript:

Computer Vision Lab. SNU Young Ki Baik Nonlinear Dimensionality Reduction Approach (ISOMAP, LLE)

References ISOMAP A global geometric framework for nonlinear dimensionality reduction J.B.Tenenbaum, V.De Silva, J.C.Langford (science 2000) LLE Nonlinear Dimensionality Reduction by Locally Linear Embedding Sam T. Roweis and Lawrence K. Saul (science 2000) ISOMAP and LLE LLE and Isomap Analysis of Spectra and Colour Images Dejan Kulpinski (Thesis 1999) Out-of-Sample Extensions for LLE, Isomap, MDS, Eignemaps, and Spectral Clustering Yoshua Bengio et. Al. (TR2003)

Contents Introduction PCA and MDS ISOMAP and LLE Conclusion

Dimensionality Reduction Problem Complex stimuli can be represented by points in a high-dimensional vector space. They typically have a much more compact description. The goal The meaningful low-dimensional structures hidden in their high-dimensional observations in order to compress the signals in size and discover compact representations of their variable.

Dimensionality Reduction Simple example 3-D data

Dimensionality Reduction Linear method PCA (Principle Component Analysis) Preserves the variance MDS (Multi Dimensional Scaling) Preserves inter-point distance Non-linear method ISOMAP LLE …

Linear Dimensionality Reduction PCA Find a low-dimensional embedding of the data points that best preserves their variance as measured in the high-dimensional input space. Eigenvectors are the principal directions, and eigen- values represent the variance of the data along each principal direction. is the marginal variance along the principle direction

Linear Dimensionality Reduction PCA Projecting onto e 1 captures the majority of the variance and hence it minimizes the error. Choosing subspace dimension M: Large M means lower expected error in the subspace data approximation Reduction

Linear Dimensionality Reduction MDS Find an embedding that preserves the inter- point distances, equivalent to PCA when the distances are Euclidean. PCA MDS

distances Relation Linear Dimensionality Reduction

MDS Providing dimension reduction. Relating tools Linear Dimensionality Reduction PCA MDS Method 1 Method 2 Method … Dimension Reduction

Nonlinear Dimensionality Reduction Many data sets contain essential nonlinear structures that invisible to PCA and MDS. Resort to some nonlinear dimensionality reduction approaches.

ISOMAP Example of non-linear structure(swiss roll) Only the geodesic distances reflect the true low- dimensional geometry of the manifold. ISOMAP (Isometric feature Mapping) Preserves the intrinsic geometry of the data. Uses the geodesic manifold distances between all pairs.

ISOMAP (algorithm description) Step 1 Determining neighboring points within a fixed radius based on the input space distance These neighborhood relation are represented as a weighted graph G over the data points. Step 2 Estimating the geodesic distances between all pairs of points on the manifold by computing their shortest path distances in the graph G. Step 3 Constructing an embedding of the data in d-dimensional Euclidean space Y that best preserves the manifold’s geometry.

ISOMAP (algorithm description) ε K=4 i j k Step 1 Determining neighboring points within a fixed radius based on the input space distance # ε-radius # K-nearest neighbors These neighborhood relations are represented as a weighted graph G over the data points.

ISOMAP (algorithm description) Step 2 Estimating the geodesic distances between all pairs of points on the manifold by computing their shortest path distances in the graph G. Can be done using Floyd’s algorithm or Dijkstra’s algorithm i j k

ISOMAP (algorithm description) Step 3 Constructing an embedding of the data in d-dimensional Euclidean space Y that best preserves the manifold’s geometry. Minimize the cost function Solution: take top d eigenvectors of the matrix

Manifold Recovery Guarantee of ISOMAP Isomap is guaranteed asymptotically to recover the true dimensionality and geometric structure of nonlinear manifolds. As the sample data points increases, the graph distances provide increasingly better approximations to the intrinsic geodesic distances.

Experimental Results (ISOMAP) # Face # Hand writing : face pose and illumination : bottom loop and top arch MDS : open triangles Isomap : filled circles

LLE LLE (Locally Linear Embedding) Neighborhood preserving embeddings. Mapping to global coordinate system of low dimensionality. Recovering global nonlinear structure from locally linear fits. Each data point and it’s neighbors is expected to lie on or close to a locally linear patch. Each data point is constructed by it’s neighbors: Where W ij summarize the contribution of j-th data point to the i-th data reconstruction and is what we will estimated by optimizing the error. Reconstructed from only its neighbors.

LLE (algorithm description) We want to minimize the error function With the constraints : Solution (using lagrange multipliers):

LLE (algorithm description) Choose d-dimensional coordinates, Y, to minimize: Under : Solution : compute bottom d+1 eigenvectors of M. (discard the last one) Quadratic form: where:

LLE (algorithm summary) Step 1 Compute the neighbors of each data point, X i Step 2 Compute the weight W ij that best reconstruct each data point X i from its neighbors, minimizing the cost in eq(1) by constrainted linear fits. Step 3 Compute the vectors Y i best reconstructed by the weights W ij, minimizing the quadratic form in eq(2) by its bottom nonzero eigenvectors. 1 2

Experimental Results (LLE) Lips # PCA # LLE

Conclusion ISOMAP Use the geodesic manifold distances between all pairs. LLE Recovers global nonlinear structure from locally linear fits. ISOMAP vs LLE Preserving the neighborhoods and their geometric relation. LLE requires massive input data sets and it must have same weight dimension. Merit of Isomap is fast processing time with dijkstra’s algorithm. Isomap is more practical than LLE.