A Global Geometric Framework for Nonlinear Dimensionality Reduction Joshua B. Tenenbaum, Vin de Silva, John C. Langford Presented by Napat Triroj.

Slides:



Advertisements
Similar presentations
Computer examples Tenenbaum, de Silva, Langford “A Global Geometric Framework for Nonlinear Dimensionality Reduction”
Advertisements

Self-Organizing Maps Projection of p dimensional observations to a two (or one) dimensional grid space Constraint version of K-means clustering –Prototypes.
Manifold Learning Dimensionality Reduction. Outline Introduction Dim. Reduction Manifold Isomap Overall procedure Approximating geodesic dist. Dijkstra’s.
Presented by: Mingyuan Zhou Duke University, ECE April 3, 2009
1 A Survey on Distance Metric Learning (Part 1) Gerry Tesauro IBM T.J.Watson Research Center.
Non-linear Dimensionality Reduction CMPUT 466/551 Nilanjan Ray Prepared on materials from the book Non-linear dimensionality reduction By Lee and Verleysen,
University of Joensuu Dept. of Computer Science P.O. Box 111 FIN Joensuu Tel fax Isomap Algorithm.
1 High dimensionality Evgeny Maksakov CS533C Department of Computer Science UBC.
“Random Projections on Smooth Manifolds” -A short summary
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
1 High dimensionality Evgeny Maksakov CS533C Department of Computer Science UBC.
Dimensionality Reduction and Embeddings
LLE and ISOMAP Analysis of Robot Images Rong Xu. Background Intuition of Dimensionality Reduction Linear Approach –PCA(Principal Component Analysis) Nonlinear.
Localization from Mere Connectivity Yi Shang (University of Missouri - Columbia); Wheeler Ruml (Palo Alto Research Center ); Ying Zhang; Markus Fromherz.
Correspondence & Symmetry
1 Numerical geometry of non-rigid shapes Spectral Methods Tutorial. Spectral Methods Tutorial 6 © Maks Ovsjanikov tosca.cs.technion.ac.il/book Numerical.
Manifold Learning: ISOMAP Alan O'Connor April 29, 2008.
1 NONLINEAR MAPPING: APPROACHES BASED ON OPTIMIZING AN INDEX OF CONTINUITY AND APPLYING CLASSICAL METRIC MDS TO REVISED DISTANCES By Ulas Akkucuk & J.
Three Algorithms for Nonlinear Dimensionality Reduction Haixuan Yang Group Meeting Jan. 011, 2005.
Atul Singh Junior Undergraduate CSE, IIT Kanpur.  Dimension reduction is a technique which is used to represent a high dimensional data in a more compact.
Dimensionality Reduction
1 Numerical geometry of non-rigid shapes Non-Euclidean Embedding Non-Euclidean Embedding Lecture 6 © Alexander & Michael Bronstein tosca.cs.technion.ac.il/book.
NonLinear Dimensionality Reduction or Unfolding Manifolds Tennenbaum|Silva|Langford [Isomap] Roweis|Saul [Locally Linear Embedding] Presented by Vikas.
Lightseminar: Learned Representation in AI An Introduction to Locally Linear Embedding Lawrence K. Saul Sam T. Roweis presented by Chan-Su Lee.
Dimensionality Reduction. Multimedia DBs Many multimedia applications require efficient indexing in high-dimensions (time-series, images and videos, etc)
Nonlinear Dimensionality Reduction by Locally Linear Embedding Sam T. Roweis and Lawrence K. Saul Reference: "Nonlinear dimensionality reduction by locally.
CS 485/685 Computer Vision Face Recognition Using Principal Components Analysis (PCA) M. Turk, A. Pentland, "Eigenfaces for Recognition", Journal of Cognitive.
Nonlinear Dimensionality Reduction Approaches. Dimensionality Reduction The goal: The meaningful low-dimensional structures hidden in their high-dimensional.
Manifold learning: Locally Linear Embedding Jieping Ye Department of Computer Science and Engineering Arizona State University
University of Joensuu Dept. of Computer Science P.O. Box 111 FIN Joensuu Tel fax Speech and.
Graph Embedding: A General Framework for Dimensionality Reduction Dong XU School of Computer Engineering Nanyang Technological University
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology Adaptive nonlinear manifolds and their applications to pattern.
IEEE TRANSSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE
Computer Vision Lab. SNU Young Ki Baik Nonlinear Dimensionality Reduction Approach (ISOMAP, LLE)
Computer examples Tenenbaum, de Silva, Langford “A Global Geometric Framework for Nonlinear Dimensionality Reduction”
Materials Process Design and Control Laboratory A NONLINEAR DIMENSION REDUCTION STRATEGY FOR GENERATING DATA DRIVEN STOCHASTIC INPUT MODELS Baskar Ganapathysubramanian.
ISOMAP TRACKING WITH PARTICLE FILTER Presented by Nikhil Rane.
GRASP Learning a Kernel Matrix for Nonlinear Dimensionality Reduction Kilian Q. Weinberger, Fei Sha and Lawrence K. Saul ICML’04 Department of Computer.
CSE 185 Introduction to Computer Vision Face Recognition.
Dimensionality Reduction
Manifold learning: MDS and Isomap
CSC2535: Computation in Neural Networks Lecture 12: Non-linear dimensionality reduction Geoffrey Hinton.
Geometry of Shape Manifolds
1 LING 696B: MDS and non-linear methods of dimension reduction.
Nonlinear Dimensionality Reduction Approach (ISOMAP)
Jan Kamenický.  Many features ⇒ many dimensions  Dimensionality reduction ◦ Feature extraction (useful representation) ◦ Classification ◦ Visualization.
Non-Linear Dimensionality Reduction
Spectral Sequencing Based on Graph Distance Rong Liu, Hao Zhang, Oliver van Kaick {lrong, haoz, cs.sfu.ca {lrong, haoz, cs.sfu.ca.
Tony Jebara, Columbia University Advanced Machine Learning & Perception Instructor: Tony Jebara.
EECS 274 Computer Vision Affine Structure from Motion.
Data Mining Course 2007 Eric Postma Clustering. Overview Three approaches to clustering 1.Minimization of reconstruction error PCA, nlPCA, k-means clustering.
Multidimensional Scaling By Marc Sobel. The Goal  We observe (possibly non-euclidean) proximity data. For each pair of objects number ‘i’ and ‘j’ we.
Math 285 Project Diffusion Maps Xiaoyan Chong Department of Mathematics and Statistics San Jose State University.
CSC321: Lecture 25: Non-linear dimensionality reduction Geoffrey Hinton.
Nonlinear Dimension Reduction: Semi-Definite Embedding vs. Local Linear Embedding Li Zhang and Lin Liao.
CSC321: Extra Lecture (not on the exam) Non-linear dimensionality reduction Geoffrey Hinton.
The UNIVERSITY of NORTH CAROLINA at CHAPEL HILL Dimensionality Reduction Part 1: Linear Methods Comp Spring 2007.
Eric Xing © Eric CMU, Machine Learning Data visualization and dimensionality reduction Eric Xing Lecture 7, August 13, 2010.
Spectral Methods for Dimensionality
Nonlinear Dimensionality Reduction
Ch 12. Continuous Latent Variables ~ 12
Unsupervised Riemannian Clustering of Probability Density Functions
Dimensionality Reduction
Spectral Methods Tutorial 6 1 © Maks Ovsjanikov
ISOMAP TRACKING WITH PARTICLE FILTERING
Outline Nonlinear Dimension Reduction Brief introduction Isomap LLE
Dimensionality Reduction
Outline H. Murase, and S. K. Nayar, “Visual learning and recognition of 3-D objects from appearance,” International Journal of Computer Vision, vol. 14,
Nonlinear Dimension Reduction:
NonLinear Dimensionality Reduction or Unfolding Manifolds
Presentation transcript:

A Global Geometric Framework for Nonlinear Dimensionality Reduction Joshua B. Tenenbaum, Vin de Silva, John C. Langford Presented by Napat Triroj

Dimensionality Reduction Objective: to find a small number of features that represent a large number of observed dimensions. For each image: there are 64x64 = 4096 pixels (observed dimensions)

Dimensionality Reduction Reduced to 3-dimensional manifold: two pose variables + azimuthal lighting angle

Isometric Feature Mapping (Isomap) A nonlinear method for dimensionality reduction Finds the map that preserves the global, nonlinear geometry of the data by preserving the geodesic manifold interpoint distances Geodesic: Shortest curve along the manifold connecting two points First approximates the geodesic interpoint distances, then runs MDS to find the projection that preserves these distances 2 components: - How do we measure the geodesic distances on the manifold? - How to map points on the Euclidean space in lower dimensional space?

Multidimensional Scaling (MDS) The algorithm detects meaningful underlying dimensions that explain observed similarities or dissimilarities (distances) between the investigated objects. Given: n x n matrix of dissimilarities between n objects (∆ T = ∆; δ ij ≥ 0; δ ii = 0) OR n x n matrix of similarities between n objects ( Θ T = Θ ; θ ij ≤ θ ii ) convert to dissimilarities  δ ij 2 = θ ii + θ jj - 2 θ ij Goal: Find a configuration in a low-dimensional Euclidean space R k whose interpoint distances d(x i,x j ) closely match dissimilarities.

Measure of goodness-of-fit: Stress Let the new configuration be obtained by projecting x 1,…,x n onto a k-D subspace. Minimizing “lack of fit” between the two configurations (dissimilarities and distances) Stress of the configuration: Φ(∆, D) = Σ (δ ij 2 – d ij 2 ) Then, the subspace is spanned by the k largest principal components.

Classical Solution: Metric method Given: nxn ∆ dissimilarities matrix of n points in m-D space Questions: - Is ∆ an interpoint distance matrix in Euclidean space? - If yes, what is the dimension?  What are the coordinates? First, check whether ∆ is Euclidean: define A nxn with elements a ij = -(1/2) δ ij 2 B = HAH = UΛU T “centered inner product matrix” H nxn = I – n T “centering matrix”  ∆ is Euclidean if and only if B is positive semidefinite i.e. λ(B) ≥ 0, with rank ≤ m

Classical Solution: Metric method If ∆ is Euclidean: - Construct the matrix A - Obtain the matrix B with elements b ij = a ij - a i· - a ·j + a ·· - Find the k largest eigenvalues λ 1 > … > λ k of B, with corresponding eigenvectors X = (x 1,…, x k ) which are normalized by x i T x i = λ i, i = 1,…,k (k< m, is whatever we pick) - Then, the required coordinates of the reconstruction points P i are x i = (x i1,…, x in ) T

The “Swiss roll” data set Unlike the geodesic distance, the Euclidean distance cannot reflect the geometric structure of the data points

Approximating Geodesic Distances Neighboring points: input-space distance Faraway points: a sequence of “short hops” between neighboring points Method: Finding shortest paths in a graph with edges connecting neighboring data points

Isomap Algorithm StepNameDescription 1 O(DN 2 ) Construct neighborhood graph, G Compute matrix D G ={d X (i,j)} d x (i,j) = Euclidean distance between neighbors 2 O(DN 2 ) Compute shortest paths between all pairs Compute matrix D G ={d G (i,j)} d G (i,j) = sequence of hops = approx geodesic dist. 3 O(dN 2 ) Construct k- dimensional coordinate vectors Apply MDS to D G instead of D X

Isomap Output The 2-D embedding recovered by Isomap

Summary on Isomap Algorithm Advantages: Nonlinear Non-iterative Globally optimal Parameters: k or ε (chosen fixed radius) Disadvantages: Graph discreteness overestimates the geodesic distance k must be high to avoid “linear shortcuts” near regions of high surface curvature