Multidimensional Scaling Vuokko Vuori 20.10.1999 Based on: Data Exploration Using Self-Organizing Maps, Samuel Kaski, Ph.D. Thesis, 1997 Multivariate Statistical.

Slides:



Advertisements
Similar presentations
What we Measure vs. What we Want to Know
Advertisements

Text mining Gergely Kótyuk Laboratory of Cryptography and System Security (CrySyS) Budapest University of Technology and Economics
Clustering Clustering of data is a method by which large sets of data is grouped into clusters of smaller sets of similar data. The example below demonstrates.
Nonmetric Multidimensional Scaling input data are ranks: most similar pair AB
Self-Organizing Maps Projection of p dimensional observations to a two (or one) dimensional grid space Constraint version of K-means clustering –Prototypes.
CLUSTERING PROXIMITY MEASURES
Dimension reduction (1)
TEMPLATE DESIGN © Self Organized Neural Networks Applied to Animal Communication Abstract Background Objective The main.
Presented by: Mingyuan Zhou Duke University, ECE April 3, 2009
Non-linear Dimensionality Reduction CMPUT 466/551 Nilanjan Ray Prepared on materials from the book Non-linear dimensionality reduction By Lee and Verleysen,
X0 xn w0 wn o Threshold units SOM.
University of Joensuu Dept. of Computer Science P.O. Box 111 FIN Joensuu Tel fax Isomap Algorithm.
Making Sense of Complicated Microarray Data Part II Gene Clustering and Data Analysis Gabriel Eichler Boston University Some slides adapted from: MeV documentation.
Two Technique Papers on High Dimensionality Allan Rempel December 5, 2005.
Multivariate Methods Pattern Recognition and Hypothesis Testing.
Semi-Supervised Classification by Low Density Separation Olivier Chapelle, Alexander Zien Student: Ran Chang.
Mean-Shift Algorithm and Its Application Bohyung Han
Multivariate Data Analysis Chapter 10 - Multidimensional Scaling
Dimension reduction : PCA and Clustering Slides by Agnieszka Juncker and Chris Workman.
1 Numerical geometry of non-rigid shapes Spectral Methods Tutorial. Spectral Methods Tutorial 6 © Maks Ovsjanikov tosca.cs.technion.ac.il/book Numerical.
Switch to Top-down Top-down or move-to-nearest Partition documents into ‘k’ clusters Two variants “Hard” (0/1) assignment of documents to clusters “soft”
Dimension reduction : PCA and Clustering Christopher Workman Center for Biological Sequence Analysis DTU.
A Global Geometric Framework for Nonlinear Dimensionality Reduction Joshua B. Tenenbaum, Vin de Silva, John C. Langford Presented by Napat Triroj.
Atul Singh Junior Undergraduate CSE, IIT Kanpur.  Dimension reduction is a technique which is used to represent a high dimensional data in a more compact.
NonLinear Dimensionality Reduction or Unfolding Manifolds Tennenbaum|Silva|Langford [Isomap] Roweis|Saul [Locally Linear Embedding] Presented by Vikas.
Lightseminar: Learned Representation in AI An Introduction to Locally Linear Embedding Lawrence K. Saul Sam T. Roweis presented by Chan-Su Lee.
Lecture 09 Clustering-based Learning
Nonlinear Dimensionality Reduction by Locally Linear Embedding Sam T. Roweis and Lawrence K. Saul Reference: "Nonlinear dimensionality reduction by locally.
Projection methods in chemistry Autumn 2011 By: Atefe Malek.khatabi M. Daszykowski, B. Walczak, D.L. Massart* Chemometrics and Intelligent Laboratory.
Manifold learning: Locally Linear Embedding Jieping Ye Department of Computer Science and Engineering Arizona State University
The Tutorial of Principal Component Analysis, Hierarchical Clustering, and Multidimensional Scaling Wenshan Wang.
KOHONEN SELF ORGANISING MAP SEMINAR BY M.V.MAHENDRAN., Reg no: III SEM, M.E., Control And Instrumentation Engg.
Presented by Tienwei Tsai July, 2005
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology On multidimensional scaling and the embedding of self-organizing.
CZ5225: Modeling and Simulation in Biology Lecture 5: Clustering Analysis for Microarray Data III Prof. Chen Yu Zong Tel:
University of Joensuu Dept. of Computer Science P.O. Box 111 FIN Joensuu Tel fax Speech and.
More on Microarrays Chitta Baral Arizona State University.
Self organizing maps 1 iCSC2014, Juan López González, University of Oviedo Self organizing maps A visualization technique with data dimension reduction.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology Adaptive nonlinear manifolds and their applications to pattern.
A two-stage approach for multi- objective decision making with applications to system reliability optimization Zhaojun Li, Haitao Liao, David W. Coit Reliability.
1 SUPPORT VECTOR MACHINES İsmail GÜNEŞ. 2 What is SVM? A new generation learning system. A new generation learning system. Based on recent advances in.
Data Reduction. 1.Overview 2.The Curse of Dimensionality 3.Data Sampling 4.Binning and Reduction of Cardinality.
Self-organizing maps (SOMs) and k-means clustering: Part 1 Steven Feldstein The Pennsylvania State University Trieste, Italy, October 21, 2013 Collaborators:
Neighbourhood relation preservation (NRP) A rank-based data visualisation quality assessment criterion Jigang Sun PhD studies finished in July 2011 PhD.
Computer Vision Lab. SNU Young Ki Baik Nonlinear Dimensionality Reduction Approach (ISOMAP, LLE)
Slide 14.1 Nonmetric Scaling MathematicalMarketing Chapter 14 Nonmetric Scaling Measurement, perception and preference are the main themes of this section.
Clustering.
Chapter 2: Getting to Know Your Data
Manifold learning: MDS and Isomap
Tony Jebara, Columbia University Advanced Machine Learning & Perception Instructor: Tony Jebara.
381 Self Organization Map Learning without Examples.
Cluster Analysis.
Semiconductors, BP&A Planning, DREAM PLAN IDEA IMPLEMENTATION.
Jin Yan Embedded and Pervasive Computing Center
Data Mining Course 2007 Eric Postma Clustering. Overview Three approaches to clustering 1.Minimization of reconstruction error PCA, nlPCA, k-means clustering.
Unsupervised Learning
Using multidimensional scaling and kernel principal component analysis to interpret seismic signatures of thin shaly-sand reservoirs Piyapa Dejtrakulwong1,
1 E.V. Myasnikov 2007 Digital image collection navigation based on automatic classification methods Samara State Aerospace University RCDL 2007Интернет-математика.
Extending metric multidimensional scaling with Bregman divergences Mr. Jigang Sun Supervisor: Prof. Colin Fyfe Nov 2009.
ViSOM - A Novel Method for Multivariate Data Projection and Structure Visualization Advisor : Dr. Hsu Graduate : Sheng-Hsuan Wang Author : Hujun Yin.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology A Nonlinear Mapping for Data Structure Analysis John W.
Out of sample extension of PCA, Kernel PCA, and MDS WILSON A. FLORERO-SALINAS DAN LI MATH 285, FALL
Data Mining, Neural Network and Genetic Programming
Multivariate community analysis
Clustering and Multidimensional Scaling
Dimension reduction : PCA and Clustering
What Is Good Clustering?
Multidimensional Space,
Introduction to Cluster Analysis
NonLinear Dimensionality Reduction or Unfolding Manifolds
Presentation transcript:

Multidimensional Scaling Vuokko Vuori Based on: Data Exploration Using Self-Organizing Maps, Samuel Kaski, Ph.D. Thesis, 1997 Multivariate Statistical Analysis, A Conceptual Introduction, Kachigan Pattern Recognition and Neural Networks, B. D. Ripley

Contents l Motivations l Dissimilarity matrix l Multidimensional scaling (MDS) l Sammon’s mapping l Self-Organizing maps Comparison between MDS, Sammon’s mapping, and SOM

Motivations MDS attempts to l Identify abstract variables which have generated the inter-object similarity measures l Reduce the dimension of the data in a non-linear fashion l Reproduce non-linear higher-dimen- sional structures on a lower-dimen- sional display

Dissimilarity Matrix In MDS, the dissimilarities between every pair of observations are given l Genuine distances (continuos data) l Simple matching coefficients, Jaccard coefficients (categorical data) l Scaled ranks (ordinal data) l Gower’s dissimilarity for mixed data:

Multidimensional Scaling Metric MDS: l Distances between data items are given, a configuration of points which gives rise to those distances is sought l Can be used for non-linear projection l Objective function which is minimized:

Nonmetric MDS: l Only the rank order of the distances is important l A monotonically increasing function that acts on the original distances is introduced: the rank order can be better preserved l Normalized objective function: l For given projection, is always chosen to minimize

Sammon’s Mapping l Closely related to metric MDS l Tries to preserve pairwise distances l Errors in distance preservation are normalized with the original distance l Objective function:

Self-Organizing Maps l Algorithm that performs clustering and non-linear projection onto lower dimen- sion at the same time l Finds and orders a set of reference vectors located on a discrete lattice l Learning rule: l Objective function: (Discrete data, fixed neighbourhood kernel)

Comparison Between MDS, Sammon’s Mapping and SOM l MDS tries to preserve the metric (ordering relations) of the original space, long distances dominate over the shorter ones l SOM tries to preserve the topology (local neighbourhood relations), items projected to nearby locations are similar l Sammon’s lies in the middle: it is like MDS but puts more emphasis on small distances