M. Belkin and P. Niyogi, Neural Computation, pp. 1373–1396, 2003.

Slides:



Advertisements
Similar presentations
Ho m o l ogy. Not to be confused with: Homeomorphism Homomorphism Homotopy Oh, and it is NOT the mixing of the cream and skim milk.
Advertisements

Leting Wu Xiaowei Ying, Xintao Wu Aidong Lu and Zhi-Hua Zhou PAKDD 2011 Spectral Analysis of k-balanced Signed Graphs 1.
A Geometric Perspective on Machine Learning 何晓飞 浙江大学计算机学院 1.
Graph Embedding and Extensions: A General Framework for Dimensionality Reduction Keywords: Dimensionality reduction, manifold learning, subspace learning,
AGE ESTIMATION: A CLASSIFICATION PROBLEM HANDE ALEMDAR, BERNA ALTINEL, NEŞE ALYÜZ, SERHAN DANİŞ.
Non-linear Dimensionality Reduction CMPUT 466/551 Nilanjan Ray Prepared on materials from the book Non-linear dimensionality reduction By Lee and Verleysen,
Clustering and Dimensionality Reduction Brendan and Yifang April
Watching Unlabeled Video Helps Learn New Human Actions from Very Few Labeled Snapshots Chao-Yeh Chen and Kristen Grauman University of Texas at Austin.
One-Shot Multi-Set Non-rigid Feature-Spatial Matching
“Random Projections on Smooth Manifolds” -A short summary
1 Manifold Clustering of Shapes Dragomir Yankov, Eamonn Keogh Dept. of Computer Science & Eng. University of California Riverside.
6. One-Dimensional Continuous Groups 6.1 The Rotation Group SO(2) 6.2 The Generator of SO(2) 6.3 Irreducible Representations of SO(2) 6.4 Invariant Integration.
Charts. Coordinate Chart  A chart is a neighborhood W and function . W  XW  X  : W  E d  : W  E d  is a diffeomorphism  is a diffeomorphism.
Planning Paths for Elastic Objects Under Manipulation Constraints Florent Lamiraux Lydia E. Kavraki Rice University Presented by: Michael Adams.
Manifold Learning: ISOMAP Alan O'Connor April 29, 2008.
Three Algorithms for Nonlinear Dimensionality Reduction Haixuan Yang Group Meeting Jan. 011, 2005.
1 NHDC and PHDC: Local and Global Heat Diffusion Based Classifiers Haixuan Yang Group Meeting Sep 26, 2005.
A Global Geometric Framework for Nonlinear Dimensionality Reduction Joshua B. Tenenbaum, Vin de Silva, John C. Langford Presented by Napat Triroj.
Complete Path Planning for Planar Closed Chains Among Point Obstacles Guanfeng Liu and Jeff Trinkle Rensselaer Polytechnic Institute.
Dilations.
NonLinear Dimensionality Reduction or Unfolding Manifolds Tennenbaum|Silva|Langford [Isomap] Roweis|Saul [Locally Linear Embedding] Presented by Vikas.
Nonlinear Dimensionality Reduction by Locally Linear Embedding Sam T. Roweis and Lawrence K. Saul Reference: "Nonlinear dimensionality reduction by locally.
Diffusion Maps and Spectral Clustering
Gender and 3D Facial Symmetry: What’s the Relationship ? Xia BAIQIANG (University Lille1/LIFL) Boulbaba Ben Amor (TELECOM Lille1/LIFL) Hassen Drira (TELECOM.
Manifold learning: Locally Linear Embedding Jieping Ye Department of Computer Science and Engineering Arizona State University
TEDI: Efficient Shortest Path Query Answering on Graphs Author: Fang Wei SIGMOD 2010 Presentation: Dr. Greg Speegle.
Structure Preserving Embedding Blake Shaw, Tony Jebara ICML 2009 (Best Student Paper nominee) Presented by Feng Chen.
Tessellation Data Structures
Graph Embedding: A General Framework for Dimensionality Reduction Dong XU School of Computer Engineering Nanyang Technological University
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology Adaptive nonlinear manifolds and their applications to pattern.
IEEE TRANSSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE
1 Learning from Shadows Dimensionality Reduction and its Application in Artificial Intelligence, Signal Processing and Robotics Ali Ghodsi Department of.
Computer Vision Lab. SNU Young Ki Baik Nonlinear Dimensionality Reduction Approach (ISOMAP, LLE)
Approximation Algorithms For Protein Folding Prediction Giancarlo MAURI,Antonio PICCOLBONI and Giulio PAVESI Symposium on Discrete Algorithms, pp ,
Double and Triple Integrals Using Iterated Integrals to find area Using Double Integrals to find Volume Using Triple Integrals to find Volume.
Manifold learning: MDS and Isomap
1 Heat Diffusion Classifier on a Graph Haixuan Yang, Irwin King, Michael R. Lyu The Chinese University of Hong Kong Group Meeting 2006.
1 LING 696B: MDS and non-linear methods of dimension reduction.
Spectral Clustering Jianping Fan Dept of Computer Science UNC, Charlotte.
Nonlinear Dimensionality Reduction Approach (ISOMAP)
Jan Kamenický.  Many features ⇒ many dimensions  Dimensionality reduction ◦ Feature extraction (useful representation) ◦ Classification ◦ Visualization.
Non-Linear Dimensionality Reduction
Math 285 Project Diffusion Maps Xiaoyan Chong Department of Mathematics and Statistics San Jose State University.
Dijkstra-Scholten and Shavit-Francez termination algorithms
Value Function Approximation with Diffusion Wavelets and Laplacian Eigenfunctions by S. Mahadevan & M. Maggioni Discussion led by Qi An ECE, Duke University.
Zhonglong Zheng, Jie Yang, Yitan Zhu Engineering Applications of Artificial Intelligence 20 (2007) 101–110 Presenter Chia-Cheng Chen 1.
Date of download: 6/22/2016 Copyright © 2016 SPIE. All rights reserved. Flowchart for the proposed unified unimodal biometric system. Figure Legend: From:
Spectral Methods for Dimensionality
Visual exploratory data analysis:
Unsupervised Riemannian Clustering of Probability Density Functions
Visual exploratory data analysis:
Introduction to linear Lie groups
Dipartimento di Ingegneria «Enzo Ferrari»,
Non-manifold Multiresolution Modeling (some preliminary results)
Jianping Fan Dept of CS UNC-Charlotte
Outline Nonlinear Dimension Reduction Brief introduction Isomap LLE
Topological Ordering Algorithm: Example
Goodfellow: Chapter 14 Autoencoders
Outline H. Murase, and S. K. Nayar, “Visual learning and recognition of 3-D objects from appearance,” International Journal of Computer Vision, vol. 14,
Introduction to linear Lie groups -- continued
Algorithm design (computational geometry)
Topological Ordering Algorithm: Example
Lecture 15: Least Square Regression Metric Embeddings
Topological Ordering Algorithm: Example
Using Manifold Structure for Partially Labeled Classification
Ho-Ramamoorthy 2-Phase Deadlock Detection Algorithm
NonLinear Dimensionality Reduction or Unfolding Manifolds
Topological Signatures For Fast Mobility Analysis
Topological Ordering Algorithm: Example
Goodfellow: Chapter 14 Autoencoders
Presentation transcript:

M. Belkin and P. Niyogi, Neural Computation, pp. 1373–1396, 2003

Outline IIntroduction AAlgorithm EExperimental Results AApplications CConclusions 2

Manifold  A manifold is a topological space which is locally Euclidean. In general, any object which is nearly "flat" on small scales is a manifold.  Examples of 1-D manifolds include a line, a circle, and two separate circles. 3

Embedding  An embedding is a representation of a topological object, manifold, graph, field, etc. in a certain space in such a way that its connectivity or algebraic properties are preserved.  Examples: 4

Manifold and Dimensionality Reduction (1)  Manifold: generalized “subspace” in R n  Points in a local region on a manifold can be indexed by a subset of R k (k<<n) 5 M z RnRn X1X1 R2R2 X2X2 x x : coordinate of z

Manifold and Dimensionality Reduction (2)  If there is a global indexing scheme for M… 6 M z RnRn X1X1 R2R2 X2X2 x x : coordinate of z y: data point closest  reduced dimension representation of y

Introduction (1)  We consider the problem of constructing a representation for data lying on a low dimensional manifold embedded in a high dimensional space 7

Introduction (2)  Linear methods - PCA (Principal Component Analysis) MDS (Multidimensional Scaling) 1952  Nonlinear methods - ISOMAP LLE (Locally Linear Embedding) LE (Laplacian Eigenmap)

Linear Methods (1)  What are “linear” methods? - Assume that data is a linear function of the parameters  Deficiencies of linear methods - Data may not be best summarized by linear combination of features 9

Linear Methods (2)  PCA: rotate data so that principal axes lie in direction of maximum variance  MDS: find coordinates that best preserve pairwise distances  Linear methods do nothing more than “globally transform” (rotate/translate/scale) data. 10 

ISOMAP, LLE and Laplacian Eigenmap  The graph-based algorithms have 3 basic steps. 1. Find K nearest neighbors. 2. Estimate local properties of manifold by looking at neighborhoods found in Step Find a global embedding that preserves the properties found in Step 2. 11

Geodesic Distance (1)  Geodesic: the shortest curve on a manifold that connects two points on the manifold Example: on a sphere, geodesics are great circles  Geodesic distance: length of the geodesic 12 small circle great circle

Geodesic Distance (2)  Euclidean distance needs not be a good measure between two points on a manifold Length of geodesic is more appropriate 13

ISOMAP 14  Comes from Isometric feature mapping Step1: Take a distance matrix {  ij } as input Step2: Estimate geodesic distance between any two points by “a chain of short paths”  Approximate the geodesic distance by Euclidean distance Step3: Perform MDSMDS

LLE (1)  Assumption: manifold is approximately “linear” when viewed locally 15 Xi Xj Xk Wij Wik 1. select neighbors 2. reconstruct with linear weights

LLE (2)  The geometrical property is best preserved if the error below is small i.e. choose the best W to minimize the cost function 16 Linear reconstruction of x i

Outline IIntroduction AAlgorithm EExperimental Results AApplications CConclusions 17

Some Aspects of the Algorithm  It reflects the intrinsic geometric structure of the manifold  The manifold is approximated by the adjacency graph computed from the data points  The Laplace Beltrami operator is approximated by the weighted Laplacian of the adjacency graph 18

Laplace Beltrami Operator (1)  The Laplace operator is a second order differential operator in the n-dimensional Euclidean space:  Laplace Beltrami operator: The Laplacian can be extended to functions defined on surfaces, or more generally, on Riemannian and pseudo-Riemannian manifolds. 19

Laplace Beltrami Operator (2)  We can justify that the eigenfunctions of the Laplace Beltrami operator have properties desirable for embedding… 20

Lapalcian of a Graph (1)  Let G(V,E) be a undirected graph without graph loops. The Laplacian of the graph is d ij if i=j (degree of node i) L ij = -1 if i≠j and (i,j) belongs to E 0 otherwise 21

Lapalcian of a Graph (2) D W(weight matrix)

Laplacian Eigenmap (1)  Consider that, and M is a manifold embedded in R l. Find y 1,.., y n in R m such that y i represents x i (m<< l ) 23

Laplacian Eigenmap (2)  Construct the adjacency graph to approximate the manifold L = ︰ 0︰ =D-W ︰

Laplacian Eigenmap (3)  There are two variations for W (weight matrix) - simple-minded (1 if connected, 0 o.w.) - heat kernel (t is real) 25

Laplacian Eigenmap (4)  Consider the problem of mapping the graph G to a line so that connected points stay as close together as possible  To choose a good “map”, we have to minimize the objective function W ij, (y i -y j ) y T Ly where y = [y 1 … y n ] T 26

Laplacian Eigenmap (5)  Therefore, this problem reduces to find argmin y T Ly subjects to y T Dy = 1 (removes an arbitrary scaling factor in the embedding)  The solution y is the eigenvector corresponding to the minimum eigenvalue of the generalized eigenvalue problem Ly = λDy 27

Laplacian Eigenmap (6)  Now we consider the more general problem of embedding the graph into m-dimensional Euclidean space  Let Y be such a n*m map 28

Laplacian Eigenmap (7)  To sum up: Step1: Construct adjacency graph Step2: Choosing the weights Step3: Eigenmaps Ly = λDy Ly 0 = λ 0 Dy 0, Ly 1 = λ 1 Dy 1 … 0= λ 0 ≦ λ 1 ≦ … ≦ λ n-1 x i  (y 0 (i), y 1 (i),…, y m (i)) 29 Recall that we have n data points, so L and D is n×n and y is a n×1 vector

ISOMAP, LLE and Laplacian Eigenmap  The graph-based algorithms have 3 basic steps. 1. Find K nearest neighbors. 2. Estimate local properties of manifold by looking at neighborhoods found in Step Find a global embedding that preserves the properties found in Step 2. 30

Outline IIntroduction AAlgorithm EExperimental Results AApplications CConclusions 31

 The following material is from

Swiss Roll (1) 33

Swiss Roll (2) 34 MDS is very slow, and ISOMAP is extremely slow. MDS and PCA don’t can’t unroll Swiss Roll, use no manifold information. LLE and Laplacian can’t handle this data.

Swiss Roll (3)  Isomap provides a isometric embedding that preserves global geodesic distances  It works only when the surface is flat  Laplacian eigenmap tries to preserve the geometric characteristics of the surface 35

Non-Convexity (1) 36

Non-Convexity (2) 37 Only Hessian LLE can handle non-convexity. ISOMAP, LLE, and Laplacian find the hole but the set is distorted.

Curvature & Non-uniform Sampling  Gaussian: We can randomly sample a Gaussian distribution.  We increase the curvature by decreasing the standard deviation.  Coloring on the z-axis, we should map to concentric circles 38

39 For std = 1 (low curvature), MDS and PCA can project accurately. Laplacian Eigenmap cannot handle the change in sampling.

40 For std = 0.4 (higher curvature), PCA projects from the side rather than top-down. Laplacian looks even worse.

41 For std = 0.3 (high curvature), none of the methods can project correctly.

Corner  Corner Planes: We bend a plane with a lift angle A.  We want to bend it back down to a plane. 42 A

43 For angle A=75, we see some disortions in PCA and Laplacian.

44 For A = 135, MDS, PCA, and Hessian LLE overwrite the data points. Diffusion Maps work very well for Sigma < 1. LLE handles corners surprisingly well.

Clustering  3D Clusters: Generate M non-overlapping clusters with random centers. Connect the clusters with a line. 45

46 For M = 3 clusters, MDS and PCA can project correctly. LLE compresses each cluster into a single point.

47 For M=8 clusters, MDS and PCA can still recover. LLE and ISOMAP are decent, but Hessian and Laplacian fail.

Sparse Data & Non-uniform Sampling  Punctured Sphere: the sampling is very sparse at the bottom and dense at the top. 48

49 Only LLE and Laplacian get decent results. PCA projects the sphere from the side. MDS turns it inside-out.

50 MDSPCAISOMAPLLELaplacianDiffusion Map KNN Diffusion Hessian SpeedVery slow Extremely fast Extremely slow Fast Slow Infers geometry? NO YES MAYBE YES Handles non-convex? NO MAYBE YES Handles non-uniform sampling? YES NOYES MAYBE Handles curvature? NO YES MAYBE YES Handles corners? NO YES NO Clusters?YES NOYES NO Handles noise? YES MAYBENOYES Handles sparsity? YES NO may crash Sensitive to parameters? NO YES VERY YES

Outline IIntroduction AAlgorithm EExperimental Results AApplications CConclusions 51

Applications  We can apply manifold learning to pattern recognition (face, handwriting etc)  Recently, ISOMAP and Laplacian eigenmap are used to initialize the human body model. 52

Outline IIntroduction AAlgorithm EExperimental Results AApplications CConclusions 53

Conclusions  Laplacian eigenmap provides a computationally efficient approach to non- linear dimensionality reduction that has locality preserving properties  Laplcian and LLE attempts to approximate or preserve neighborhood information, while ISOMAP attempts to faithfully approximate all geodesic distances on the manifold 54

Reference    ISOMAP Joshua B. Tenenbaum, Vin de Silva, and John C. Langford, “A Global Geometric Framework for Nonlinear Dimensionality Reduction,” Science, vol. 290, Dec.,  LLE Sam T. Roweis, and Lawrence K. Saul, “Nonlinear Dimensionality Reduction by Locally Linear Embedding,” Science, vol. 290, Dec., 2000  Laplacian eigenmap M. Belkin and P. Niyogi. “Laplacian eigenmaps for dimensionality reduction and data representation,” Neural Comput.,15(6):1373– 1396,