Iterative K-Means Algorithm Based on Fisher Discriminant UNIVERSITY OF JOENSUU DEPARTMENT OF COMPUTER SCIENCE JOENSUU, FINLAND Mantao Xu to be presented.

Slides:



Advertisements
Similar presentations
Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
Advertisements

Support Vector Machines
Assessment. Schedule graph may be of help for selecting the best solution Best solution corresponds to a plateau before a high jump Solutions with very.
Gene selection using Random Voronoi Ensembles Stefano Rovetta Department of Computer and Information Sciences, University of Genoa, Italy Francesco masulli.
Software Quality Ranking: Bringing Order to Software Modules in Testing Fei Xing Michael R. Lyu Ping Guo.
EE 290A: Generalized Principal Component Analysis Lecture 6: Iterative Methods for Mixture-Model Segmentation Sastry & Yang © Spring, 2011EE 290A, University.
Clustering II.
Iterative Optimization of Hierarchical Clusterings Doug Fisher Department of Computer Science, Vanderbilt University Journal of Artificial Intelligence.
© University of Minnesota Data Mining for the Discovery of Ocean Climate Indices 1 CSci 8980: Data Mining (Fall 2002) Vipin Kumar Army High Performance.
1 Abstract This paper presents a novel modification to the classical Competitive Learning (CL) by adding a dynamic branching mechanism to neural networks.
University of CreteCS4831 The use of Minimum Spanning Trees in microarray expression data Gkirtzou Ekaterini.
Prénom Nom Document Analysis: Data Analysis and Clustering Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Semi-Supervised Clustering Jieping Ye Department of Computer Science and Engineering Arizona State University
1 Context-Aware Clustering Junsong Yuan and Ying Wu EECS Dept., Northwestern University.
What is Cluster Analysis?
Techniques for CBIR 03/10/16陳慶鋒. Outline Iteration-free clustering algorithm for nonstationary image database Iteration-free clustering algorithm for.
Part 3 Vector Quantization and Mixture Density Model CSE717, SPRING 2008 CUBS, Univ at Buffalo.
Parallel K-Means Clustering Based on MapReduce The Key Laboratory of Intelligent Information Processing, Chinese Academy of Sciences Weizhong Zhao, Huifang.
Dimensionality Reduction
Clustering with Bregman Divergences Arindam Banerjee, Srujana Merugu, Inderjit S. Dhillon, Joydeep Ghosh Presented by Rohit Gupta CSci 8980: Machine Learning.
Comparing Kernel-based Learning Methods for Face Recognition Zhiguo Li
Dimensionality Reduction. Multimedia DBs Many multimedia applications require efficient indexing in high-dimensions (time-series, images and videos, etc)
Introduction to Bioinformatics Algorithms Clustering and Microarray Analysis.
Machine Learning Usman Roshan Dept. of Computer Science NJIT.
Evaluating Performance for Data Mining Techniques
CSC 4510 – Machine Learning Dr. Mary-Angela Papalaskari Department of Computing Sciences Villanova University Course website:
Summarized by Soo-Jin Kim
嵌入式視覺 Pattern Recognition for Embedded Vision Template matching Statistical / Structural Pattern Recognition Neural networks.
Unsupervised Learning. CS583, Bing Liu, UIC 2 Supervised learning vs. unsupervised learning Supervised learning: discover patterns in the data that relate.
Clustering methods Course code: Pasi Fränti Speech & Image Processing Unit School of Computing University of Eastern Finland Joensuu,
COMMON EVALUATION FINAL PROJECT Vira Oleksyuk ECE 8110: Introduction to machine Learning and Pattern Recognition.
CSIE Dept., National Taiwan Univ., Taiwan
Combined Central and Subspace Clustering for Computer Vision Applications Le Lu 1 René Vidal 2 1 Computer Science Department, Johns Hopkins University,
Clustering Supervised vs. Unsupervised Learning Examples of clustering in Web IR Characteristics of clustering Clustering algorithms Cluster Labeling 1.
Basic Machine Learning: Clustering CS 315 – Web Search and Data Mining 1.
Genetic Algorithm Using Iterative Shrinking for Solving Clustering Problems UNIVERSITY OF JOENSUU DEPARTMENT OF COMPUTER SCIENCE FINLAND Pasi Fränti and.
Exploratory Learning Semi-supervised Learning in the presence of unanticipated classes Bhavana Dalvi, William W. Cohen, Jamie Callan School Of Computer.
Dimension reduction : PCA and Clustering Slides by Agnieszka Juncker and Chris Workman modified by Hanne Jarmer.
Chapter 9 DTW and VQ Algorithm  9.1 Basic idea of DTW  9.2 DTW algorithm  9.3 Basic idea of VQ  9.4 LBG algorithm  9.5 Improvement of VQ.
Clustering I. 2 The Task Input: Collection of instances –No special class label attribute! Output: Clusters (Groups) of instances where members of a cluster.
Optimal Component Analysis Optimal Linear Representations of Images for Object Recognition X. Liu, A. Srivastava, and Kyle Gallivan, “Optimal linear representations.
Cluster Analysis Potyó László. Cluster: a collection of data objects Similar to one another within the same cluster Similar to one another within the.
Clustering Clustering is a technique for finding similarity groups in data, called clusters. I.e., it groups data instances that are similar to (near)
Feature Selection in k-Median Clustering Olvi Mangasarian and Edward Wild University of Wisconsin - Madison.
Machine Learning Queens College Lecture 7: Clustering.
Data Mining Course 2007 Eric Postma Clustering. Overview Three approaches to clustering 1.Minimization of reconstruction error PCA, nlPCA, k-means clustering.
Faculty of Information Engineering, Shenzhen University Liao Huilian SZU TI-DSPs LAB Aug 27, 2007 Optimizer based on particle swarm optimization and LBG.
Optimal Reverse Prediction: Linli Xu, Martha White and Dale Schuurmans ICML 2009, Best Overall Paper Honorable Mention A Unified Perspective on Supervised,
Data Mining By Farzana Forhad CS 157B. Agenda Decision Tree and ID3 Rough Set Theory Clustering.
2D-LDA: A statistical linear discriminant analysis for image matrix
6.S093 Visual Recognition through Machine Learning Competition Image by kirkh.deviantart.com Joseph Lim and Aditya Khosla Acknowledgment: Many slides from.
1 Microarray Clustering. 2 Outline Microarrays Hierarchical Clustering K-Means Clustering Corrupted Cliques Problem CAST Clustering Algorithm.
Multilevel thresholding by fast PNN based algorithm UNIVERSITY OF JOENSUU DEPARTMENT OF COMPUTER SCIENCE FINLAND Olli Virmajoki and Pasi Fränti.
1 Cluster Analysis – 2 Approaches K-Means (traditional) Latent Class Analysis (new) by Jay Magidson, Statistical Innovations based in part on a presentation.
Given a set of data points as input Randomly assign each point to one of the k clusters Repeat until convergence – Calculate model of each of the k clusters.
Clustering Approaches Ka-Lok Ng Department of Bioinformatics Asia University.
Debrup Chakraborty Non Parametric Methods Pattern Recognition and Machine Learning.
K -means clustering via Principal Component Analysis (Chris Ding and Xiaofeng He, ICML 2004) 03 March 2011 Kwak, Namju 1.
Non-parametric Methods for Clustering Continuous and Categorical Data Steven X. Wang Dept. of Math. and Stat. York University May 13, 2010.
Intelligent Database Systems Lab Advisor : Dr. Hsu Graduate : Jian-Lin Kuo Author : Aristidis Likas Nikos Vlassis Jakob J.Verbeek 國立雲林科技大學 National Yunlin.
Cluster Analysis What is Cluster Analysis? Types of Data in Cluster Analysis A Categorization of Major Clustering Methods Partitioning Methods.
Sparse nonnegative matrix factorization for protein sequence motifs information discovery Presented by Wooyoung Kim Computer Science, Georgia State University.
Clustering Machine Learning Unsupervised Learning K-means Optimization objective Random initialization Determining Number of Clusters Hierarchical Clustering.
Machine Learning Usman Roshan Dept. of Computer Science NJIT.
Semi-Supervised Clustering
Machine Learning Clustering: K-means Supervised Learning
Glenn Fung, Murat Dundar, Bharat Rao and Jinbo Bi
KMeans Clustering on Hadoop Fall 2013 Elke A. Rundensteiner
Presented by: Chang Jia As for: Pattern Recognition
Concave Minimization for Support Vector Machine Classifiers
Presentation transcript:

Iterative K-Means Algorithm Based on Fisher Discriminant UNIVERSITY OF JOENSUU DEPARTMENT OF COMPUTER SCIENCE JOENSUU, FINLAND Mantao Xu to be presented at: Infomation Fusion 2004

Problem Formulation Given N data samples X={x 1, x 2, …, x N }, construct the codebook C = {c 1, c 2, …, c M } such that mean-square-error is minimized. The class membership p (i) is

Traditional K-Means Algorithm Iterations of two steps: assignment of each data vector with a class label computation of cluster centroid by averaging all data vectors that are assigned to it Characteristics: Randomized initial partition or codebook Convergence to a local minimum Use of L 2, L 1 and L  distance Fast and easy implementation Extensions: Kernel Kmeans algorithm EM algorithm K-median algorithm

Motivation Investigation on a clustering algorithm that : iteratively performs the regular K-Means algorithm in searching a solution close to the global optima estimates the initial partition close to the optimal solution by at each iteration applies a dissimilarity function based on the current partition instead of L 2 distance

Selecting K-Means initial partiton Selection of initial partion based on Fisher discriminant and dynamic programming: A suboptimal partition is estimated by dynamic programming in some one- dimenaional subspace of feature space. The one-dimensional subspace is constructed through linear multi-class Fisher discriminant analysis The output class of K-Means in each iteration is selected as the input class of discriminant analysis for next iteration

The multi-class Fisher discriminant analysis The separation of input classes in the discriminant direction w can be measured by F-ratio validity index, F(w) The multi-class linear Fisher discriminant w is the minimization of F-ratio validity index

The dynamic programming in discriminant direction The optimal convex partition Q k ={(q j-1,q j ]| j=1, ,n} in the discriminant direction w can be estimated by dynamic promgramming in terms of MSE distortion on the discriminant subspace or in terms of MSE distortion on original feature space (2) (1)

Application of Delta-MSE Dissimilarity x 1 x 2 x 3 y 1 y 2 y 3 x 4 G 1 G 2 Delta-MSE(x 4,G,G 1 )=RemovalVariance Delta-MSE(x 4,G,G 2 )=AddVariance Move vector x from cluster i to cluster j, the change of the MSE function [10] caused by this move is:

Pseudocodes of Iterative K-Means algorithm

Four K-Means algorithms conducted in experimental tests K-D tree based K-Means: selects its initial cluster centroids from the k-bucket centers of a kd-tree structure that is recursively built by principal component analysis PCA based K-Means: an intuitive approach to estimate a sub-optimal initial partition by applying the dynamic programming in the principal component direction LFD-I: the proposed iterative K-Means algorithm based on the dynamic programming criterion (1) LFD-II: the proposed iterative K-Means algorithm based on the dynamic programming criterion (2)

Comparisons of the four K-Means algorithms Table 1: Performance comparisons (in F-ratio validity indices ) of the four K-Means algorithms on the practical numbers of clusters

F-ratios produced by the four K- Means clusterings on dataset glass

F-ratios produced by the four K- Means clusterings on dataset heart

F-ratios produced by the four K- Means clusterings on dataset image

Conclusions A new approach to the k-center clustering problem by iteratively incorporating the Fisher discriminant analysis and the dynamic programming technique The proposed approach in general outperforms the two other algorithms: the PCA based K-Means algorithm and the kd-tree based K-Means algorithm The classification performance gains of the proposed approach over the two others is increased with the number of clusters

Further Work Sovling the k-center clustering problem by iteratively incorporating the kernel Fisher discriminant analysis and the dynamic programming technique Sovling the k-center clustering problem by incorporating the kernel PCA technique and the dynamic programming technique