Unsupervised Riemannian Clustering of Probability Density Functions

Slides:



Advertisements
Similar presentations
Coherent Laplacian 3D protrusion segmentation Oxford Brookes Vision Group Queen Mary, University of London, 11/12/2009 Fabio Cuzzolin.
Advertisements

Text mining Gergely Kótyuk Laboratory of Cryptography and System Security (CrySyS) Budapest University of Technology and Economics
M. Belkin and P. Niyogi, Neural Computation, pp. 1373–1396, 2003.
1 Manifold Alignment for Multitemporal Hyperspectral Image Classification H. Lexie Yang 1, Melba M. Crawford 2 School of Civil Engineering, Purdue University.
Aggregating local image descriptors into compact codes
Component Analysis (Review)
Nonlinear Dimension Reduction Presenter: Xingwei Yang The powerpoint is organized from: 1.Ronald R. Coifman et al. (Yale University) 2. Jieping Ye, (Arizona.
Graph Embedding and Extensions: A General Framework for Dimensionality Reduction Keywords: Dimensionality reduction, manifold learning, subspace learning,
Computer vision: models, learning and inference Chapter 13 Image preprocessing and feature extraction.
AGE ESTIMATION: A CLASSIFICATION PROBLEM HANDE ALEMDAR, BERNA ALTINEL, NEŞE ALYÜZ, SERHAN DANİŞ.
Presented by: Mingyuan Zhou Duke University, ECE April 3, 2009
Nonlinear Unsupervised Feature Learning How Local Similarities Lead to Global Coding Amirreza Shaban.
Non-linear Dimensionality Reduction CMPUT 466/551 Nilanjan Ray Prepared on materials from the book Non-linear dimensionality reduction By Lee and Verleysen,
Watching Unlabeled Video Helps Learn New Human Actions from Very Few Labeled Snapshots Chao-Yeh Chen and Kristen Grauman University of Texas at Austin.
One-Shot Multi-Set Non-rigid Feature-Spatial Matching
“Random Projections on Smooth Manifolds” -A short summary
Manifold Learning: ISOMAP Alan O'Connor April 29, 2008.
Support Vector Machines and Kernel Methods
Three Algorithms for Nonlinear Dimensionality Reduction Haixuan Yang Group Meeting Jan. 011, 2005.
A Global Geometric Framework for Nonlinear Dimensionality Reduction Joshua B. Tenenbaum, Vin de Silva, John C. Langford Presented by Napat Triroj.
Pattern Recognition. Introduction. Definitions.. Recognition process. Recognition process relates input signal to the stored concepts about the object.
Atul Singh Junior Undergraduate CSE, IIT Kanpur.  Dimension reduction is a technique which is used to represent a high dimensional data in a more compact.
NonLinear Dimensionality Reduction or Unfolding Manifolds Tennenbaum|Silva|Langford [Isomap] Roweis|Saul [Locally Linear Embedding] Presented by Vikas.
Lightseminar: Learned Representation in AI An Introduction to Locally Linear Embedding Lawrence K. Saul Sam T. Roweis presented by Chan-Su Lee.
Nonlinear Dimensionality Reduction by Locally Linear Embedding Sam T. Roweis and Lawrence K. Saul Reference: "Nonlinear dimensionality reduction by locally.
Nonlinear Dimensionality Reduction Approaches. Dimensionality Reduction The goal: The meaningful low-dimensional structures hidden in their high-dimensional.
Linear Algebra and Image Processing
Manifold learning: Locally Linear Embedding Jieping Ye Department of Computer Science and Engineering Arizona State University
Graph Embedding: A General Framework for Dimensionality Reduction Dong XU School of Computer Engineering Nanyang Technological University
IEEE TRANSSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE
Combined Central and Subspace Clustering for Computer Vision Applications Le Lu 1 René Vidal 2 1 Computer Science Department, Johns Hopkins University,
Data Reduction. 1.Overview 2.The Curse of Dimensionality 3.Data Sampling 4.Binning and Reduction of Cardinality.
Learning a Kernel Matrix for Nonlinear Dimensionality Reduction By K. Weinberger, F. Sha, and L. Saul Presented by Michael Barnathan.
Computer Vision Lab. SNU Young Ki Baik Nonlinear Dimensionality Reduction Approach (ISOMAP, LLE)
A Two-level Pose Estimation Framework Using Majority Voting of Gabor Wavelets and Bunch Graph Analysis J. Wu, J. M. Pedersen, D. Putthividhya, D. Norgaard,
Transductive Regression Piloted by Inter-Manifold Relations.
Local Fisher Discriminant Analysis for Supervised Dimensionality Reduction Presented by Xianwang Wang Masashi Sugiyama.
Non-Euclidean Example: The Unit Sphere. Differential Geometry Formal mathematical theory Work with small ‘patches’ –the ‘patches’ look Euclidean Do calculus.
Handwritten digit recognition
ISOMAP TRACKING WITH PARTICLE FILTER Presented by Nikhil Rane.
GRASP Learning a Kernel Matrix for Nonlinear Dimensionality Reduction Kilian Q. Weinberger, Fei Sha and Lawrence K. Saul ICML’04 Department of Computer.
Spoken Language Group Chinese Information Processing Lab. Institute of Information Science Academia Sinica, Taipei, Taiwan
Manifold learning: MDS and Isomap
Nonlinear Dimensionality Reduction Approach (ISOMAP)
Jan Kamenický.  Many features ⇒ many dimensions  Dimensionality reduction ◦ Feature extraction (useful representation) ◦ Classification ◦ Visualization.
H. Lexie Yang1, Dr. Melba M. Crawford2
Data Mining Course 0 Manifold learning Xin Yang. Data Mining Course 1 Outline Manifold and Manifold Learning Classical Dimensionality Reduction Semi-Supervised.
Tony Jebara, Columbia University Advanced Machine Learning & Perception Instructor: Tony Jebara.
Math 285 Project Diffusion Maps Xiaoyan Chong Department of Mathematics and Statistics San Jose State University.
PART III: TRANSIENT INTERFERENCE SUPPRESSION USING DIFFUSION MAPS.
Spectral Methods for Dimensionality
Nonlinear Dimensionality Reduction
Shuang Hong Yang College of Computing, Georgia Tech, USA Hongyuan Zha
Nonparametric Density Estimation – k-nearest neighbor (kNN) 02/20/17
LECTURE 10: DISCRIMINANT ANALYSIS
Computer Vision, Robotics, Machine Learning and Control Lab
کاربرد نگاشت با حفظ تنکی در شناسایی چهره
René Vidal Center for Imaging Science
René Vidal Time/Place: T-Th 4.30pm-6pm, Hodson 301
T. Chernyakova, A. Aberdam, E. Bar-Ilan, Y. C. Eldar
Dipartimento di Ingegneria «Enzo Ferrari»,
Generalized Principal Component Analysis CVPR 2008
Principal Component Analysis (PCA)
Machine Learning Basics
Outline Nonlinear Dimension Reduction Brief introduction Isomap LLE
Object Modeling with Layers
Outline S. C. Zhu, X. Liu, and Y. Wu, “Exploring Texture Ensembles by Efficient Markov Chain Monte Carlo”, IEEE Transactions On Pattern Analysis And Machine.
Outline H. Murase, and S. K. Nayar, “Visual learning and recognition of 3-D objects from appearance,” International Journal of Computer Vision, vol. 14,
LECTURE 09: DISCRIMINANT ANALYSIS
NonLinear Dimensionality Reduction or Unfolding Manifolds
Presentation transcript:

Unsupervised Riemannian Clustering of Probability Density Functions Alvina Goh and René Vidal Department of Biomedical Engineering Center for Imaging Science Johns Hopkins University 1

Introduction Data clustering and dimensionality reduction is an important topic having applications in various areas. In our work, every data point is not just a point, but a probability density function (pdf). Therefore, the question is how to do clustering and dimensionality reduction of pdfs. ?

Motivation Texture is a function of the spatial variation in pixel intensities. By convoluting an image with a set of filters, it is possible to generate a histogram representing the patterns present. Modern object recognition techniques use similar ideas; categories are represented with histograms of codewords. In high angular resolution diffusion imaging, the data at each voxel is the orientation distribution function. Texture properties include uniformity, density, coarseness, roughness, regularity, linearity, directionality, direction, frequency, and phase.

Challenges How do we develop a computationally simple framework that allows us to group the pdfs into similar families? These pdfs are determined from data and they are non-parametric in nature. These are infinitely dimensional objects. In order to compare them and cluster them, we need a metric in the space. This is a clustering problem in the statistical manifold of pdfs.

Main Contributions To develop Riemannian clustering and nonlinear dimensionality reduction for pdfs, that proceeds by making use of the square-root representation and the resulting Riemannian geometry, generalized LLE, LE, HLLE with Riemannian metric, generating a mapping from different submanifolds to different central clusters.

Nonlinear dimensionality reduction Global techniques Preserve global properties of the data lying on a sub-manifold. Similar to PCA for a linear subspace. Isomap, Kernel PCA. Local techniques Preserve local properties obtained from small neighborhoods around points. Also retain the global properties of the data via spectral analysis. Locally Linear Embedding (LLE) (Roweis, Saul 03), Laplacian Eigenmaps (LE) (Belkin, Niyogi 02), Hessian LLE (Donoho, Grimes 03) 6

Local techniques

Local techniques

Local techniques

Local techniques

Riemannian Analysis of Probability Density Functions Class of constrained non-negative continuous functions The Fisher-Rao metric is the unique intrinsic metric where are tangent vectors and is the set containing the function tangent to at the point . Difficult to work with as ensuring the geodesic between two elements lie on is not easy.

Riemannian Analysis of Probability Density Functions Square-root representation The functions lie on a unit sphere. The Fisher-Rao metric is then are tangent vectors Closed form formulae for geodesic distance exponential map logarithm map

Extending NLDR to Riemannian manifolds 13 13

Extending NLDR to Riemannian manifolds 14 14

Extending NLDR to Riemannian manifolds 15 15

Extending NLDR to Riemannian manifolds 16 16

Extending NLDR to Riemannian manifolds Manifold geometry essential only in first two steps of each algorithm. How to select the kNN? by incorporating the Riemannian distance How to compute the matrix representing the local geometry? 17 17

Riemannian Calculation of M for LLE LLE involves writing each data point as a linear combination of its neighbors. Riemannian case: interpolation problem on the manifold. How should the data points be interpolated? What cost function should be minimized? as in the Euclidean case. 18 18

Riemannian Calculation of M for LE and HLLE Extending LE is straightforward and involves replacing the Euclidean metric by the Riemannian metric. Construct the weight matrix as We have where as before. HLLE involves finding the mean and the tangent space at each point. The mean at on the manifold is the Frechet mean of its k-NN . found as the solution to The basis for is found via an eigenanalysis of the covariance matrix using Principal Geodesic Analysis (Fletcher and Joshi 04). 19 19

Local Riemannian Manifold Clustering Let be a set of points drawn from a k-disconnected union of k-connected submanifolds with dimensions of a Riemannian manifold . When the assumption of separated submanifolds is violated, we have and , respectively. Applying K-means to gives the clustering of the data.

Unsupervised Clustering of PDFs Making use of the closed-form formula under the square-root representation and the algorithm for local Riemannian manifold clustering, we can perform unsupervised clustering of pdfs.

Synthetic Experiments Applying Riemannian LLE to two groups of uniform pdfs Clustering two groups of pdfs

Texture Clustering In order to construct a histogram that reflects the texture statistics in an image, we will calculate textons. This is done by first applying a filter bank to all images in the training set. Filters used to generate the histogram of texture This will provide us with a feature vector of dimension 13 at each pixel. Next, we apply k-means to all the vectors in the entire dataset to get 30 cluster centers. For each image, compute a histogram that contains the number of pixels corresponding to each one of these 30 bins.

Texture Clustering (two classes)

Texture Clustering (three classes)

Conclusion Presented an algorithm for grouping families of probability density functions. Exploit the fact that under the square root re-parametrization, the space of pdfs form a convex subset of the unit Hilbert sphere. The problem of clustering pdfs reduces to clustering multiple submanifolds on the unit Hilbert sphere. Results on synthetic and real data are encouraging.

Acknowledgments Funding from Startup funds from JHU, NSF CAREER IIS-0447739, NSF EHS-0509101, ONR N00014-05-10836, JHU APL-934652. The authors thank Rizwan Chaudhry for useful discussions. Vision, Dynamics and Learning Lab @ Johns Hopkins University Thank You!