Atul Singh Junior Undergraduate CSE, IIT Kanpur.  Dimension reduction is a technique which is used to represent a high dimensional data in a more compact.

Slides:



Advertisements
Similar presentations
EigenFaces and EigenPatches Useful model of variation in a region –Region must be fixed shape (eg rectangle) Developed for face recognition Generalised.
Advertisements

1 Manifold Alignment for Multitemporal Hyperspectral Image Classification H. Lexie Yang 1, Melba M. Crawford 2 School of Civil Engineering, Purdue University.
Manifold Learning Dimensionality Reduction. Outline Introduction Dim. Reduction Manifold Isomap Overall procedure Approximating geodesic dist. Dijkstra’s.
Presented by: Mingyuan Zhou Duke University, ECE April 3, 2009
1 A Survey on Distance Metric Learning (Part 1) Gerry Tesauro IBM T.J.Watson Research Center.
Non-linear Dimensionality Reduction CMPUT 466/551 Nilanjan Ray Prepared on materials from the book Non-linear dimensionality reduction By Lee and Verleysen,
University of Joensuu Dept. of Computer Science P.O. Box 111 FIN Joensuu Tel fax Isomap Algorithm.
“Random Projections on Smooth Manifolds” -A short summary
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Principal Component Analysis
LLE and ISOMAP Analysis of Robot Images Rong Xu. Background Intuition of Dimensionality Reduction Linear Approach –PCA(Principal Component Analysis) Nonlinear.
Image Manifolds : Learning-based Methods in Vision Alexei Efros, CMU, Spring 2007 © A.A. Efros With slides by Dave Thompson.
1 Numerical geometry of non-rigid shapes Spectral Methods Tutorial. Spectral Methods Tutorial 6 © Maks Ovsjanikov tosca.cs.technion.ac.il/book Numerical.
Manifold Learning: ISOMAP Alan O'Connor April 29, 2008.
Face Recognition Jeremy Wyatt.
Three Algorithms for Nonlinear Dimensionality Reduction Haixuan Yang Group Meeting Jan. 011, 2005.
Distance Metric Learning: A Comprehensive Survey
A Global Geometric Framework for Nonlinear Dimensionality Reduction Joshua B. Tenenbaum, Vin de Silva, John C. Langford Presented by Napat Triroj.
NonLinear Dimensionality Reduction or Unfolding Manifolds Tennenbaum|Silva|Langford [Isomap] Roweis|Saul [Locally Linear Embedding] Presented by Vikas.
Lightseminar: Learned Representation in AI An Introduction to Locally Linear Embedding Lawrence K. Saul Sam T. Roweis presented by Chan-Su Lee.
Nonlinear Dimensionality Reduction by Locally Linear Embedding Sam T. Roweis and Lawrence K. Saul Reference: "Nonlinear dimensionality reduction by locally.
Nonlinear Dimensionality Reduction Approaches. Dimensionality Reduction The goal: The meaningful low-dimensional structures hidden in their high-dimensional.
Manifold learning: Locally Linear Embedding Jieping Ye Department of Computer Science and Engineering Arizona State University
Nonlinear Dimensionality Reduction for Hyperspectral Image Classification Tim Doster Advisors: John Benedetto & Wojciech Czaja.
Dimensionality Reduction: Principal Components Analysis Optional Reading: Smith, A Tutorial on Principal Components Analysis (linked to class webpage)
Graph Embedding: A General Framework for Dimensionality Reduction Dong XU School of Computer Engineering Nanyang Technological University
IEEE TRANSSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE
Data Reduction. 1.Overview 2.The Curse of Dimensionality 3.Data Sampling 4.Binning and Reduction of Cardinality.
Learning a Kernel Matrix for Nonlinear Dimensionality Reduction By K. Weinberger, F. Sha, and L. Saul Presented by Michael Barnathan.
1 Learning from Shadows Dimensionality Reduction and its Application in Artificial Intelligence, Signal Processing and Robotics Ali Ghodsi Department of.
THE MANIFOLDS OF SPATIAL HEARING Ramani Duraiswami | Vikas C. Raykar Perceptual Interfaces and Reality Lab University of Maryland, College park.
Computer Vision Lab. SNU Young Ki Baik Nonlinear Dimensionality Reduction Approach (ISOMAP, LLE)
ISOMAP TRACKING WITH PARTICLE FILTER Presented by Nikhil Rane.
GRASP Learning a Kernel Matrix for Nonlinear Dimensionality Reduction Kilian Q. Weinberger, Fei Sha and Lawrence K. Saul ICML’04 Department of Computer.
CSE 185 Introduction to Computer Vision Face Recognition.
Dimensionality Reduction
Manifold learning: MDS and Isomap
CSC2535: Computation in Neural Networks Lecture 12: Non-linear dimensionality reduction Geoffrey Hinton.
Dimensionality Reduction Part 2: Nonlinear Methods
1 LING 696B: MDS and non-linear methods of dimension reduction.
Nonlinear Dimensionality Reduction Approach (ISOMAP)
Jan Kamenický.  Many features ⇒ many dimensions  Dimensionality reduction ◦ Feature extraction (useful representation) ◦ Classification ◦ Visualization.
H. Lexie Yang1, Dr. Melba M. Crawford2
Non-Linear Dimensionality Reduction
Project 11: Determining the Intrinsic Dimensionality of a Distribution Okke Formsma, Nicolas Roussis and Per Løwenborg.
Project 11: Determining the Intrinsic Dimensionality of a Distribution Okke Formsma, Nicolas Roussis and Per Løwenborg.
Tony Jebara, Columbia University Advanced Machine Learning & Perception Instructor: Tony Jebara.
Data Projections & Visualization Rajmonda Caceres MIT Lincoln Laboratory.
Data Mining Course 2007 Eric Postma Clustering. Overview Three approaches to clustering 1.Minimization of reconstruction error PCA, nlPCA, k-means clustering.
Math 285 Project Diffusion Maps Xiaoyan Chong Department of Mathematics and Statistics San Jose State University.
CSC321: Lecture 25: Non-linear dimensionality reduction Geoffrey Hinton.
Nonlinear Dimension Reduction: Semi-Definite Embedding vs. Local Linear Embedding Li Zhang and Lin Liao.
Eric Xing © Eric CMU, Machine Learning Data visualization and dimensionality reduction Eric Xing Lecture 7, August 13, 2010.
Spectral Methods for Dimensionality
Nonlinear Dimensionality Reduction
Intrinsic Data Geometry from a Training Set
Unsupervised Riemannian Clustering of Probability Density Functions
کاربرد نگاشت با حفظ تنکی در شناسایی چهره
Dipartimento di Ingegneria «Enzo Ferrari»,
Principal Component Analysis (PCA)
Dimensionality Reduction
Spectral Methods Tutorial 6 1 © Maks Ovsjanikov
Machine Learning Dimensionality Reduction
ISOMAP TRACKING WITH PARTICLE FILTERING
Outline Nonlinear Dimension Reduction Brief introduction Isomap LLE
Dimensionality Reduction
Object Modeling with Layers
Learning with information of features
NonLinear Dimensionality Reduction or Unfolding Manifolds
Marios Mattheakis and Pavlos Protopapas
Presentation transcript:

Atul Singh Junior Undergraduate CSE, IIT Kanpur

 Dimension reduction is a technique which is used to represent a high dimensional data in a more compact way using some alternate representation  Many process of data generation generate a large data set which is embedded in a low dimensional manifold of a high dimensional space. Dimensional reduction can be applied to those data sets

 In statistical pattern recognition, we often encounter a data set in a high dimensional space  Many times the data is correlated in such a manner that there are very few independent dimensions  Possible to represent the data using much lower dimensions. Some benefits are - ◦ Compact representation ◦ Less processing time ◦ Visualization of high dimensional data ◦ Interpolate meaningful dimensions

 Linear Methods ◦ Principal Component Analysis (PCA) ◦ Independent Component Analysis (ICA) ◦ Multi-dimensional Scaling (MDS)  Non-linear Methods ◦ Global  Isomap and its variants ◦ Local  Locally Linear Embedding (LLE)  Laplacian Eigenmaps

 Principal Component Analysis ◦ Involves finding directions having large covariance ◦ Express the data as a linear combination of eigenvectors along those directions  Multidimensional Scaling ◦ Keep inter point distances as invariant ◦ Again a linear methodology

 Many variables can’t be represented as linear combinations of some vectors ◦ Examples – Swiss roll, faces data, etc  In general the low dimension data is embedded on some non linear manifold  Not possible to transform these manifolds to low dimension space using only translation, rotation, rescaling

 Linear methods have a pre conceived notion of dimension reduction  Goal is to – automate the estimation (infer the degrees of freedom of data manifold) and classification (embed data in low dimension space) process  So need to go beyond Linear methods  Non Linear Methods ◦ Isomap ◦ Locally Linear Embedding

 Global manifold learning algorithm  A swiss roll (fig a) embedded as a manifold in high dimension space. Goal is to reduce it to two dimensions (fig c)

 Consider a human face ◦ How is it represented/stored in brain? ◦ How is it represented/stored in a computer?  Do we need to store all the information (every pixel)?  Just need to figure out some important structures

 Basic steps involved are ◦ Construct Neighborhood Graph: Determine the neighbors of each point and assign edge weights d x (i, j) to the graph thus formed ◦ Compute Geodesic distances: Estimate the geodesic distances d M (i, j) using Dijkstra’s algorithm ◦ Use MDS to lower dimensions: Apply MDS on the computed shortest path distance matrix and thus reduce the dimensionality  So basically similar to MDS just using geodesic distances rather than Euclidean

 Another non-linear dimension reduction algorithm  Follows a local approach to reduce the dimensions  The idea is based on the assumption that a point on a manifold reside on a hyper plane determined by the point and its some nearest neighbors  Key Question – How to combine these parts of hyper planes and map them to low dimensional space?

 The basic steps involved are ◦ Assign neighbors: To each point assign neighbors using nearest neighbor approach ◦ Weight calculation: We compute weights W ij such that X i is best reconstructed from its neighbors ◦ Compute Low dimension embedding: Using the above computed weight matrix find corresponding embedding vectors Y i in lower dimensional space minimizing error function

 The weights computed for reconstruction are invariant with respect to translation, rotation and rescaling  The same weights should reconstruct the map in reduced dimensions  So we can conclude that the local geometry is preserved

 Shortcomings of Isomap ◦ Need a dense sampling of data on the manifold ◦ If k is chosen very small then residual error will be too large ◦ If k is chosen very large then short-circuiting may happen  Shortcomings of LLE ◦ Due to local nature doesn’t give a complete picture of the data ◦ Again problems with selection of k

 Short-circuiting ◦ “When the distance between the folds is very less or there is noise such that a point from a different fold is chosen to be a neighbour of the point, the distance computed does not represent geodesic distance and hence the algorithm fails”  Insight ◦ This problem arises due to selection of neighbors just on the basis of their Euclidean distance. The basic selection criteria are  Select all the points within a ball of radius ε  Select K nearest neighbors ◦ Locally Linear isomaps overcomes this problem by modifying the neighbor selection criteria  Proposed algorithm – K LL Isomaps

 Similar to Tanenbaum’s Isomap except for the selection of nearest neighbors  Previous algorithms (Isomap, LLE) consider only the Euclidean distance to judge neighborhood criteria  The proposed algorithm is ◦ Find a candidate neighborhood using K-nn approach ◦ Construct the data point using candidate neighbors (as in LLE) to minimize the reconstruction error ◦ The neighbours whose Euclidean distance is less and those lying on the locally linear patch of the manifold get higher weights and hence are selected preferably ◦ Now K LL ≤ K neighbors are chosen based on the post construction weights

 K LL Isomap has been demonstrated to perform better than Isomap under ◦ Sparsely sampled data ◦ Noisy data ◦ Dense data without noise  The metric used for Analysis are “Short circuit edges” and “Residual variance” To establish a formal proof for better performance of this algorithm

 Tenenbaum J.B., Silva V.d., Langford J.C.: A Global Geometric Framework for Nonlinear Dimensionality Reduction  Roweis S.T., Saul L.K.: Nonlinear Dimensionality Reduction by Locally Linear Embedding  Balasubramanian M., Schwartz E.L., Tenenbaum J.B., Silva V.d., Langford J.C.: The Isomap Algorithm and Topological Stability  Silva V.d., Tenenbaum J.B.: Global versus local methods in nonlinear dimensionality reduction  Roweis S.T., Saul L.K.: An Introduction to Locally Linear Embedding  Saxena A., Gupta A., Mukherjee A.: Non-linear Dimensionality Reduction by Locally Linear Isomaps