Presentation is loading. Please wait.

Presentation is loading. Please wait.

Introduction to Hierarchical Clustering Analysis Pengyu Hong 09/16/2005.

Similar presentations


Presentation on theme: "Introduction to Hierarchical Clustering Analysis Pengyu Hong 09/16/2005."— Presentation transcript:

1 Introduction to Hierarchical Clustering Analysis Pengyu Hong 09/16/2005

2 Background Cell/Tissue 1 Cell/Tissue 2 Cell/Tissue N … Data 1 Data 2 Data N … Put similar samples/entries together.

3 Background  Clustering is one of the most important unsupervised learning processes that organizing objects into groups whose members are similar in some way.  Clustering finds structures in a collection of unlabeled data.  A cluster is a collection of objects which are similar between them and are dissimilar to the objects belonging to other clusters.

4 Motivation I Microarray data quality checking –Does replicates cluster together? –Does similar conditions, time points, tissue types cluster together?

5 Data: Rat Schizophrenia Data (Allen Fienberg and Mayetri Gupta)  Two time points:35 days (PD 35) and 60 days (PD60) past birth.  Two brain regions: Prefrontal cortex (PFC) and Nucleus accumbens (NA).  Two replicates (Samples are from the same set of tissue split into different tubes so that replicates should be in close agreement.)  dChip was used to normalize the data and get model-based expression values, using the full PM/MM model. How to read this clustering result? Heat map Link length Sample IDs Gene IDs Clustering results Problem?

6 Motivation II Cluster genes  Prediction of functions of unknown genes by known ones

7 Functional significant gene clusters Two-way clustering Gene clusters Sample clusters

8 Motivation II Cluster genes  Prediction of functions of unknown genes by known ones Cluster samples  Discover clinical characteristics (e.g. survival, marker status) shared by samples.

9 Bhattacharjee et al. (2001) Human lung carcinomas mRNA expression profiling reveals distinct adenocarcinoma subclasses. Proc. Natl. Acad. Sci. USA, Vol. 98, 13790- 13795.

10 Motivation II Cluster genes  Prediction of functions of unknown genes by known ones Cluster samples  Discover clinical characteristics (e.g. survival, marker status) shared by samples Promoter analysis of commonly regulated genes

11 David J. Lockhart & Elizabeth A. Winzeler, NATURE | VOL 405 | 15 JUNE 2000, p827 Promoter analysis of commonly regulated genes

12 Clustering Algorithms Start with a collection of n objects each represented by a p–dimensional feature vector x i, i=1, …n. The goal is to divide these n objects into k clusters so that objects within a clusters are more “similar” than objects between clusters. k is usually unknown. Popular methods: hierarchical, k-means, SOM, mixture models, etc.

13 Hierarchical Clustering Dendrogram Venn Diagram of Clustered Data From http://www.stat.unc.edu/postscript/papers/marron/Stat321FDA/RimaIzempresentation.ppt

14 Hierarchical Clustering (Cont.) Multilevel clustering: level 1 has n clusters  level n has one cluster. Agglomerative HC: starts with singleton and merge clusters. Divisive HC: starts with one sample and split clusters.

15 Nearest Neighbor Algorithm Nearest Neighbor Algorithm is an agglomerative approach (bottom-up). Starts with n nodes (n is the size of our sample), merges the 2 most similar nodes at each step, and stops when the desired number of clusters is reached. From http://www.stat.unc.edu/postscript/papers/marron/Stat321FDA/RimaIzempresentation.ppt

16 Nearest Neighbor, Level 2, k = 7 clusters. From http://www.stat.unc.edu/postscript/papers/marron/Stat321FDA/RimaIzempresentation.ppt

17 Nearest Neighbor, Level 3, k = 6 clusters.

18 Nearest Neighbor, Level 4, k = 5 clusters.

19 Nearest Neighbor, Level 5, k = 4 clusters.

20 Nearest Neighbor, Level 6, k = 3 clusters.

21 Nearest Neighbor, Level 7, k = 2 clusters.

22 Nearest Neighbor, Level 8, k = 1 cluster.

23 Calculate the similarity between all possible combinations of two profiles Two most similar clusters are grouped together to form a new cluster Calculate the similarity between the new cluster and all remaining clusters. Hierarchical Clustering Keys Similarity Clustering

24 Similarity Measurements Pearson Correlation Two profiles (vectors) and +1  Pearson Correlation  – 1

25 Similarity Measurements Pearson Correlation: Trend Similarity

26 Similarity Measurements Euclidean Distance

27 Similarity Measurements Euclidean Distance: Absolute difference

28 Similarity Measurements Cosine Correlation +1  Cosine Correlation  – 1

29 Similarity Measurements Cosine Correlation: Trend + Mean Distance

30 Similarity Measurements

31 Similar?

32 Clustering C1C1 C2C2 C3C3 Merge which pair of clusters?

33 + + Clustering Single Linkage C1C1 C2C2 Dissimilarity between two clusters = Minimum dissimilarity between the members of two clusters Tend to generate “long chains”

34 + + Clustering Complete Linkage C1C1 C2C2 Dissimilarity between two clusters = Maximum dissimilarity between the members of two clusters Tend to generate “clumps”

35 + + Clustering Average Linkage C1C1 C2C2 Dissimilarity between two clusters = Averaged distances of all pairs of objects (one from each cluster).

36 + + Clustering Average Group Linkage C1C1 C2C2 Dissimilarity between two clusters = Distance between two cluster means.

37 Considerations What genes are used to cluster samples? –Expression variation –Inherent variation –Prior knowledge (irrelevant genes) –Etc.

38 Take Home Questions Which clustering method is better? How to cut the clustering tree to get relatively tight clusters of genes or samples?


Download ppt "Introduction to Hierarchical Clustering Analysis Pengyu Hong 09/16/2005."

Similar presentations


Ads by Google