Presentation is loading. Please wait.

Presentation is loading. Please wait.

Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology 1 Hierarchical model-based clustering of large datasets.

Similar presentations


Presentation on theme: "Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology 1 Hierarchical model-based clustering of large datasets."— Presentation transcript:

1 Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology 1 Hierarchical model-based clustering of large datasets through fractionation and refractionation Advisor : Dr. Hsu Presenter : Zih-Hui Lin Author :Jeremy Tantrum *,1, Alejandro Murua, Werner Stuetzle 2 Information Systems, Volume 29, Issue 4, June 2004, pages: 315 - 326

2 Intelligent Database Systems Lab N.Y.U.S.T. I. M. 2  Motivation  Objective  Introduction  Method  Conclusions Outline

3 Intelligent Database Systems Lab N.Y.U.S.T. I. M. 3 Motivation  Compared to non-parametric clustering methods like complete linkage, hierarchical model-based clustering has the advantage of offering a way to estimate the number of groups present in the data. ─ But computational cost is quadratic in the number of items to be clustered, and it is therefore not applicable to large problems.

4 Intelligent Database Systems Lab N.Y.U.S.T. I. M. 4 Objective  This paper is proposed a method which can deal with the problem.

5 Intelligent Database Systems Lab N.Y.U.S.T. I. M. 5 p g are multivariate Gaussian with mean μ g and covariance matrix Sg: Model-based clustering ∑ g is the prior probability that a randomly chosen observation belongs to group g

6 Intelligent Database Systems Lab N.Y.U.S.T. I. M. 6 1234 567

7 Intelligent Database Systems Lab N.Y.U.S.T. I. M. 7 2.Cluster each fraction into a fixed number α M of clusters, with α< 1. Summarize each cluster by its mean. We refer to these cluster means as meta- observations. 1.Split the data into subsets or fractions of size M. 1.We randomly split the data into four fractions of 100 observations each 2.then use model-based hierarchical clustering to cluster each fraction into M/10 =10 clusters 3.If the total number of meta- observations is greater than M; return to step (1), with the meta-observations taking the place of the original data 3.The number of meta-observations produced by clustering the fractions in this case is 40 which is less than M =100 Method

8 Intelligent Database Systems Lab N.Y.U.S.T. I. M. 8 4.a. Cluster the meta-observations into G clusters, where G is determined by the BIC. 4.a. Clustering the 40 meta-observations into 25 clusters produces the mixture model whose component densities are shown in Fig. 4. 4.b.Clustering the 40 meta-observations into new fractions results in fraction sizes of 97, 108, 104, and 91. 4.b.Define the fractions for the ith pass: as soon as a cluster formed during the merging represents more than M observations, make those observations into a fraction and remove the cluster from the merge process. Method

9 Intelligent Database Systems Lab N.Y.U.S.T. I. M. 9 Method 1. Split the data into subsets or fractions of size M. 2. Cluster each fraction into a fixed number α M of clusters, with α< 1. Summarize each cluster by its mean. We refer to these cluster means as meta-observations. 3. If the total number of meta-observations is greater than M; return to step (1), with the meta-observations taking the place of the original data. 4. Cluster the meta-observations into G clusters, where G is determined by the BIC. 5. Define the fractions for the ith pass: as soon as a cluster formed during the merging represents more than M observations, make those observations into a fraction and remove the cluster from the merge process. 6. Assign each individual observation to the cluster with the closest mean.

10 Intelligent Database Systems Lab N.Y.U.S.T. I. M. 10 Measuring the agreement between groups and clusters  The index is the geometric mean of two probabilities: ─ the probability that two randomly chosen observations are in the same cluster given that they are in the same group ─ the probability that two randomly chosen observations are in the same group given that they are in the same cluster.

11 Intelligent Database Systems Lab N.Y.U.S.T. I. M. 11 The TDT dataset  The documents are news stories from Reuters and CNN for the year from July 1st 1994–June 30 th 1995, and they are ordered according to the time and date of the story. ─ There are 15, 863 documents in 25, 431 dimensions. 25 topics were chosen and all documents associated to these topics were manually labeled—there are 1, 131 labeled documents in the collection.

12 Intelligent Database Systems Lab N.Y.U.S.T. I. M. 12 Model-based clustering of the 1100 documents→19 clusters 。 Fowlkes-Mallows index= 0.65 Example-1 Example-3

13 Intelligent Database Systems Lab N.Y.U.S.T. I. M. 13 Example-5

14 Intelligent Database Systems Lab N.Y.U.S.T. I. M. 14 Conclusions  This paper is proposed a method which can deal with the large dataset  Model-based Refractionation does not require that the number of groups in the data be known a priori; it can be estimated from the data.

15 Intelligent Database Systems Lab N.Y.U.S.T. I. M. 15 My opinion Advantage The method can process the large dataset. Drawback ….. Application …..


Download ppt "Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology 1 Hierarchical model-based clustering of large datasets."

Similar presentations


Ads by Google