Clustering by soft-constraint affinity propagation: applications to gene- expression data Michele Leone, Sumedha and Martin Weight Bioinformatics, 2007.

Slides:



Advertisements
Similar presentations
Algorithms and applications
Advertisements


Outline Data with gaps clustering on the basis of neuro-fuzzy Kohonen network Adaptive algorithm for probabilistic fuzzy clustering Adaptive probabilistic.
Principal Component Analysis (PCA) for Clustering Gene Expression Data K. Y. Yeung and W. L. Ruzzo.
Introduction to Bioinformatics
UNSUPERVISED ANALYSIS GOAL A: FIND GROUPS OF GENES THAT HAVE CORRELATED EXPRESSION PROFILES. THESE GENES ARE BELIEVED TO BELONG TO THE SAME BIOLOGICAL.
1 Fast Primal-Dual Strategies for MRF Optimization (Fast PD) Robot Perception Lab Taha Hamedani Aug 2014.
A Bayesian Approach to Joint Feature Selection and Classifier Design Balaji Krishnapuram, Alexander J. Hartemink, Lawrence Carin, Fellow, IEEE, and Mario.
Clustering short time series gene expression data Jason Ernst, Gerard J. Nau and Ziv Bar-Joseph BIOINFORMATICS, vol
Radial Basis Functions
Prénom Nom Document Analysis: Linear Discrimination Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
© University of Minnesota Data Mining for the Discovery of Ocean Climate Indices 1 CSci 8980: Data Mining (Fall 2002) Vipin Kumar Army High Performance.
Predictive Automatic Relevance Determination by Expectation Propagation Yuan (Alan) Qi Thomas P. Minka Rosalind W. Picard Zoubin Ghahramani.
Cluster Analysis.  What is Cluster Analysis?  Types of Data in Cluster Analysis  A Categorization of Major Clustering Methods  Partitioning Methods.
Ensemble Learning: An Introduction
Clustering (Part II) 10/07/09. Outline Affinity propagation Quality evaluation.
Classification with reject option in gene expression data Blaise Hanczar and Edward R Dougherty BIOINFORMATICS Vol. 24 no , pages
L15:Microarray analysis (Classification). The Biological Problem Two conditions that need to be differentiated, (Have different treatments). EX: ALL (Acute.
Clustering Color/Intensity
Clustering (Part II) 11/26/07. Spectral Clustering.
Microarray analysis 2 Golan Yona. 2) Analysis of co-expression Search for similarly expressed genes experiment1 experiment2 experiment3 ……….. Gene i:
Clustering Ram Akella Lecture 6 February 23, & 280I University of California Berkeley Silicon Valley Center/SC.
Introduction to Bioinformatics Algorithms Clustering and Microarray Analysis.
JAVED KHAN ET AL. NATURE MEDICINE – Volume 7 – Number 6 – JUNE 2001
Clustering Unsupervised learning Generating “classes”
Principal Component Analysis (PCA) for Clustering Gene Expression Data K. Y. Yeung and W. L. Ruzzo.
Graph-based consensus clustering for class discovery from gene expression data Zhiwen Yum, Hau-San Wong and Hongqiang Wang Bioinformatics, 2007.
Binary Variables (1) Coin flipping: heads=1, tails=0 Bernoulli Distribution.
Computational Stochastic Optimization: Bridging communities October 25, 2012 Warren Powell CASTLE Laboratory Princeton University
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum.
10/18/ Support Vector MachinesM.W. Mak Support Vector Machines 1. Introduction to SVMs 2. Linear SVMs 3. Non-linear SVMs References: 1. S.Y. Kung,
An Introduction to Support Vector Machine (SVM) Presenter : Ahey Date : 2007/07/20 The slides are based on lecture notes of Prof. 林智仁 and Daniel Yeung.
Classification and diagnostic prediction of cancers using gene expression profiling and artificial neural networks From Nature Medicine 7(6) 2001 By Javed.
First topic: clustering and pattern recognition Marc Sobel.
Dimension reduction : PCA and Clustering Slides by Agnieszka Juncker and Chris Workman modified by Hanne Jarmer.
Computational Intelligence: Methods and Applications Lecture 23 Logistic discrimination and support vectors Włodzisław Duch Dept. of Informatics, UMK Google:
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: ML and Simple Regression Bias of the ML Estimate Variance of the ML Estimate.
Evolutionary Algorithms for Finding Optimal Gene Sets in Micro array Prediction. J. M. Deutsch Presented by: Shruti Sharma.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Reestimation Equations Continuous Distributions.
Visual Categorization With Bags of Keypoints Original Authors: G. Csurka, C.R. Dance, L. Fan, J. Willamowski, C. Bray ECCV Workshop on Statistical Learning.
Clustering.
SemiBoost : Boosting for Semi-supervised Learning Pavan Kumar Mallapragada, Student Member, IEEE, Rong Jin, Member, IEEE, Anil K. Jain, Fellow, IEEE, and.
Guest lecture: Feature Selection Alan Qi Dec 2, 2004.
Radial Basis Function ANN, an alternative to back propagation, uses clustering of examples in the training set.
1 Lecture 6 Neural Network Training. 2 Neural Network Training Network training is basic to establishing the functional relationship between the inputs.
Molecular Classification of Cancer Class Discovery and Class Prediction by Gene Expression Monitoring.
Jen-Tzung Chien, Meng-Sung Wu Minimum Rank Error Language Modeling.
Evaluation of gene-expression clustering via mutual information distance measure Ido Priness, Oded Maimon and Irad Ben-Gal BMC Bioinformatics, 2007.
1 Microarray Clustering. 2 Outline Microarrays Hierarchical Clustering K-Means Clustering Corrupted Cliques Problem CAST Clustering Algorithm.
1 An Efficient Optimal Leaf Ordering for Hierarchical Clustering in Microarray Gene Expression Data Analysis Jianting Zhang Le Gruenwald School of Computer.
CZ5211 Topics in Computational Biology Lecture 4: Clustering Analysis for Microarray Data II Prof. Chen Yu Zong Tel:
Example Apply hierarchical clustering with d min to below data where c=3. Nearest neighbor clustering d min d max will form elongated clusters!
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Reestimation Equations Continuous Distributions.
Clustering Approaches Ka-Lok Ng Department of Bioinformatics Asia University.
CIVET seminar Presentation day: Presenter : Park, GilSoon.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Mixture Densities Maximum Likelihood Estimates.
DATA MINING: CLUSTER ANALYSIS Instructor: Dr. Chun Yu School of Statistics Jiangxi University of Finance and Economics Fall 2015.
Topics In Social Computing (67810) Module 1 (Structure) Centrality Measures, Graph Clustering Random Walks on Graphs.
Clustering (2) Center-based algorithms Fuzzy k-means Density-based algorithms ( DBSCAN as an example ) Evaluation of clustering results Figures and equations.
Data Mining K-means Algorithm
Clustering (3) Center-based algorithms Fuzzy k-means
Latent Variables, Mixture Models and EM
Jianping Fan Dept of CS UNC-Charlotte
Jianping Fan Dept of Computer Science UNC-Charlotte
Clustering.
CS623: Introduction to Computing with Neural Nets (lecture-15)
LECTURE 21: CLUSTERING Objectives: Mixture Densities Maximum Likelihood Estimates Application to Gaussian Mixture Models k-Means Clustering Fuzzy k-Means.
LECTURE 15: REESTIMATION, EM AND MIXTURES
Clustering.
Presentation transcript:

Clustering by soft-constraint affinity propagation: applications to gene- expression data Michele Leone, Sumedha and Martin Weight Bioinformatics, 2007

Outline Introduction The Algorithm and Method Analysis Experimental results Discussion 2

Introduction Affinity Propagation seeks to identify each cluster by one of its elements, exemplar. – each point in the cluster refers to this exemplar. – each exemplar is required to refer to itself as a self-exemplar. However, it forces clusters to appear as stars. 3 There’s only one central node, and all other nodes are directly connected to it.

Introduction Some drawbacks in Affinity Propagation: – The hard constraint in AP relies strongly on cluster-shape regularity. – All information about the internal structure and the hierarchical merging/dissociation of cluster is lost. – AP has robustness limitations. – AP forces each exemplar to point to itself. 4

Introduction How to improve it? The hard constraint: exemplars would be self- exemplars. We relax the hard constraint by introducing a finite penalty term for each constraint violation. 5

The Algorithm and Method Analysis The Soft Constraint Affinity Propagation(SCAP) equations. Efficient implementation of the algorithm. Extracting cluster signatures. 6

The SCAP equations We write the constraint attached to a given data point as follows, with : The first case assigns a penalty if data point is chosen as exemplar by some other data point, without being a self-exemplar. 7

The SCAP equations The penalty presents a compromise between the minimization the cost function and the search of compact clusters. Then, we introduce a positive real-valued parameter weighing the relative importance of the cost minimization with respect to the constraints. 8

The SCAP equations So, we can define the probability of an arbitrary clustering as: Original AP is recovered by taking since any violated constraint sets to zero. 9

The SCAP equations For general, the optimal clustering can be determined by maximizing the marginal probabilities for all data points : 10

The SCAP equations Assume, we find the SCAP equations: The exemplar of any data point can be computed as: 11

The SCAP equations Compared to original AP, SCAP amounts to an additional threshold on the self-availabilities and the self-responsibilities. For small enough, in many case. The self-responsibility is substituted with. For (i.e. ), the original AP equations are recovered. 12

The SCAP equations This means that variables are discouraged to be self-exemplars beyond a given threshold, even in the case someone is already pointing at them. 13

Efficient implementation The iterative solution: 14

Efficient implementation Difference between the original AP: – Step 3 is formulated as a sequential update. – The original AP used damped parallel update. 15

Extracting cluster signatures Only a few components carry useful information about the cluster structure, they are called cluster signatures. We assume the similarity between data points and to be additive in single-gene contributions: 16

Extracting cluster signatures Having found a clustering given by the exemplar selection, we can calculate the similarity of a cluster C defined as a connected component of the directed graph: as a sum over single-gene contributions 17

Extracting cluster signatures Then, we compare to random exemplar choices which are characterized by their mean: and variance 18

Extracting cluster signatures The relevance of a gene can be ranked by which measures the distance of the actual from the distribution of random exemplar mappings. Genes can be ranked according to, highest-ranking genes are considered a cluster signature. 19

Experimental results Iris data Brain cancer data Other benchmark cancer data – Lymphoma cancer data – SRBCT cancer data – Leukemia 20

Iris data Three clusters: setosa, versicolor, virginica. Four features for 150 flowers: – sepal length – sepal width – petal length – petal width 21

Iris data Experimental results: – Affinity Propagation: 16 errors. – SCAP: 9 errors with Manhattan distance measure for the similarity. On increasing the value of, the clusters for Versicolor and Virginica merge with each other, reflecting the fact that they are closer to each other than to Setosa. 22

Brain cancer data Five diagnosis types for 42 patients: – 10 medulloblastoma – 10 malignant glioma – 10 atypical teratoid/rhabdoid tumors – 4 normal cerebella – 8 primitive neuroectodermal tumors – PNET 23

Brain cancer data Clustering with AP(for ): 24 There are three well- distinguishable clusters. Five clusters for lowest errors.

Brain cancer data Clustering with SCAP: 25 The SCAP identifies four clusters with 8 errors.

Brain cancer data Eight errors are due to misclassifications of the fifth diagnosis(PNET). We use the procedure to extract cluster signatures in the case of four clusters: No. 34~41 are the fifth diagnosis. 26

Other benchmark cancer data Lymphoma cancer data – Three diagnoses for 62 patients. SRBCT cancer data – Four expression diagnosis patterns for 63 samples. Leukemia – Two diagnoses for 72 samples. 27

Other benchmark cancer data Lymphoma cancer data – AP: 3 errors with 3 clusters. – SCAP: 1 error with 3 clusters. SRBCT cancer data – AP: 22 errors with 5 clusters. – SCAP: 7 errors with 4 clusters. Leukemia – AP: 4 errors with 2 clusters. – SCAP: 2 errors with 2 clusters. 28

Discussion If clusters cannot be well represented by a single cluster exemplar, AP has to fail. SCAP is more efficient than AP in particular in the case of noisy, irregularly organized data and thus in biological applications concerning microarray data. The cluster structure can be efficiently probed. 29