TOWARDS HIERARCHICAL CLUSTERING

Slides:



Advertisements
Similar presentations
Clustering II.
Advertisements

SEEM Tutorial 4 – Clustering. 2 What is Cluster Analysis?  Finding groups of objects such that the objects in a group will be similar (or.
Clustering.
Hierarchical Clustering
1 CSE 980: Data Mining Lecture 16: Hierarchical Clustering.
Hierarchical Clustering. Produces a set of nested clusters organized as a hierarchical tree Can be visualized as a dendrogram – A tree-like diagram that.
Learning Trajectory Patterns by Clustering: Comparative Evaluation Group D.
Introduction to Network Theory: Modern Concepts, Algorithms
Data Mining Cluster Analysis: Basic Concepts and Algorithms
Assessment. Schedule graph may be of help for selecting the best solution Best solution corresponds to a plateau before a high jump Solutions with very.
What is Cluster Analysis? Finding groups of objects such that the objects in a group will be similar (or related) to one another and different from (or.
COURSE “SYSTEM DESIGN: STRUCTURAL APPROACH” DETC Inst. for Information Transmission Problems Russian Academy of Sciences, Moscow , Russia.
Clustering II.
Clustering… in General In vector space, clusters are vectors found within  of a cluster vector, with different techniques for determining the cluster.
Region Segmentation. Find sets of pixels, such that All pixels in region i satisfy some constraint of similarity.
L16: Micro-array analysis Dimension reduction Unsupervised clustering.
Distance methods. UPGMA: similar to hierarchical clustering but not additive Neighbor-joining: more sophisticated and additive What is additivity?
Segmentation Graph-Theoretic Clustering.
Cluster Analysis: Basic Concepts and Algorithms
Graph, Search Algorithms Ka-Lok Ng Department of Bioinformatics Asia University.
CS541 Advanced Networking 1 Routing and Shortest Path Algorithms Neil Tang 2/18/2009.
Cluster Analysis (1).
© University of Minnesota Data Mining for the Discovery of Ocean Climate Indices 1 CSci 8980: Data Mining (Fall 2002) Vipin Kumar Army High Performance.
CSE 550 Computer Network Design Dr. Mohammed H. Sqalli COE, KFUPM Spring 2007 (Term 062)
Application of Graph Theory to OO Software Engineering Alexander Chatzigeorgiou, Nikolaos Tsantalis, George Stephanides Department of Applied Informatics.
Clustering Unsupervised learning Generating “classes”
LECTURE 1. Course: “Design of Systems: Structural Approach” Dept. “Communication Networks &Systems”, Faculty of Radioengineering & Cybernetics Moscow Inst.
LECTURE 5-6. Course: “Design of Systems: Structural Approach” Dept. “Communication Networks &Systems”, Faculty of Radioengineering & Cybernetics Moscow.
LECTURE 28. Course: “Design of Systems: Structural Approach” Dept. “Communication Networks &Systems”, Faculty of Radioengineering & Cybernetics Moscow.
MODULAR SYSTEMS & COMBINATORIAL OPTIMIZATION (based on course “System design”, 2004…2008, MIPT) Mark Sh. Levin Inst. for Inform. Transmission Problems,
Approximating Minimum Bounded Degree Spanning Tree (MBDST) Mohit Singh and Lap Chi Lau “Approximating Minimum Bounded DegreeApproximating Minimum Bounded.
LECTURE 8-9. Course: “Design of Systems: Structural Approach” Dept. “Communication Networks &Systems”, Faculty of Radioengineering & Cybernetics Moscow.
LECTURE (compressed version). Course: “Design of Systems: Structural Approach” Dept. “Communication Networks &Systems”, Faculty of Radioengineering.
LECTURE 19. Course: “Design of Systems: Structural Approach” Dept. “Communication Networks &Systems”, Faculty of Radioengineering & Cybernetics Moscow.
LECTURE Course: “Design of Systems: Structural Approach” Dept. “Communication Networks &Systems”, Faculty of Radioengineering & Cybernetics Moscow.
LECTURE 13. Course: “Design of Systems: Structural Approach” Dept. “Communication Networks &Systems”, Faculty of Radioengineering & Cybernetics Moscow.
Towards Communication Network Development (structural systems issues, combinatorial models) Mark Sh. Levin Inst. for Inform. Transmission Problems, Russian.
Ground Truth Free Evaluation of Segment Based Maps Rolf Lakaemper Temple University, Philadelphia,PA,USA.
Memory Allocation of Multi programming using Permutation Graph By Bhavani Duggineni.
LECTURE 4. Course: “Design of Systems: Structural Approach” Dept. “Communication Networks &Systems”, Faculty of Radioengineering & Cybernetics Moscow Inst.
CSE5334 DATA MINING CSE4334/5334 Data Mining, Fall 2014 Department of Computer Science and Engineering, University of Texas at Arlington Chengkai Li (Slides.
LECTURE 16. Course: “Design of Systems: Structural Approach” Dept. “Communication Networks &Systems”, Faculty of Radioengineering & Cybernetics Moscow.
Module #9: Matrices Rosen 5 th ed., §2.7 Now we are moving on to matrices, section 7.
Discrete Mathematics 1 Kemal Akkaya DISCRETE MATHEMATICS Lecture 16 Dr. Kemal Akkaya Department of Computer Science.
STUDENT RESEARCH PROJECTS IN SYSTEM DESIGN Inst. for Information Transmission Problems Russian Academy of Sciences, Moscow , Russia
Theoretical Foundations of Clustering – CS497 Talk Shai Ben-David February 2007.
LECTURE Course: “Design of Systems: Structural Approach” Dept. “Communication Networks &Systems”, Faculty of Radioengineering & Cybernetics Moscow.
LECTURE 2-3. Course: “Design of Systems: Structural Approach” Dept. “Communication Networks &Systems”, Faculty of Radioengineering & Cybernetics Moscow.
Lecture 3 1.Different centrality measures of nodes 2.Hierarchical Clustering 3.Line graphs.
Information Retrieval and Organisation Chapter 16 Flat Clustering Dell Zhang Birkbeck, University of London.
CS 8751 ML & KDDData Clustering1 Clustering Unsupervised learning Generating “classes” Distance/similarity measures Agglomerative methods Divisive methods.
Basic Linear Algebra Subroutines (BLAS) – 3 levels of operations Memory hierarchy efficiently exploited by higher level BLAS BLASMemor y Refs. FlopsFlops/
Relation. Combining Relations Because relations from A to B are subsets of A x B, two relations from A to B can be combined in any way two sets can be.
Hierarchical Clustering Produces a set of nested clusters organized as a hierarchical tree Can be visualized as a dendrogram – A tree like diagram that.
Definition Finding groups of objects such that the objects in a group will be similar (or related) to one another and different from (or unrelated to)
LECTURE 10. Course: “Design of Systems: Structural Approach” Dept. “Communication Networks &Systems”, Faculty of Radioengineering & Cybernetics Moscow.
Basic Machine Learning: Clustering CS 315 – Web Search and Data Mining 1.
LECTURE 27. Course: “Design of Systems: Structural Approach” Dept. “Communication Networks &Systems”, Faculty of Radioengineering & Cybernetics Moscow.
LECTURE Course: “Design of Systems: Structural Approach” Dept. “Communication Networks &Systems”, Faculty of Radioengineering & Cybernetics Moscow.
© Tan,Steinbach, Kumar Introduction to Data Mining 4/18/ Data Mining: Cluster Analysis This lecture node is modified based on Lecture Notes for Chapter.
Community structure in graphs Santo Fortunato. More links “inside” than “outside” Graphs are “sparse” “Communities”
LECTURE 7. Course: “Design of Systems: Structural Approach” Dept. “Communication Networks &Systems”, Faculty of Radioengineering & Cybernetics Moscow Inst.
Example Apply hierarchical clustering with d min to below data where c=3. Nearest neighbor clustering d min d max will form elongated clusters!
Data Mining Cluster Analysis: Basic Concepts and Algorithms Lecture Notes Introduction to Data Mining by Tan, Steinbach, Kumar © Tan,Steinbach, Kumar Introduction.
TOWARDS FOUR-LAYER FRAMEWORK OF COMBINATORIAL PROBLEMS
Graph clustering to detect network modules
Local Feature Extraction Using Scale-Space Decomposition
Scale-Space Representation for Matching of 3D Models
Hierarchical Clustering
Clustering.
Presentation transcript:

TOWARDS HIERARCHICAL CLUSTERING Mark Sh. Levin Inst. for Inform. Transm. Problems, Russian Acad. of Sci. Email: mslevin@acm.org Http://www.iitp.ru/mslevin/ PLAN: 1.Basic agglomerative algorithm for hierarchical clustering 2.Multicriteria decision making (DM) approach to proximity of objects 3.Integration of objects into several groups/clusters (i.e., clustering with intersection): algorithms & applications 4.Towards resultant performance (quality of results) 5.Conclusion CSR’2007, Ural State University, Ekaterinburg, Russia, Sept. 4, 2007

Hierarchical clustering: agglomerative algorithm Objects (alternatives) Criteria (characteristics) C1 … Cj … Cm A1 = ( z11 , … , z1j , … , z1m ) … … Ai = ( zi1 , … , zij , … , zim ) An = ( zn1 , … , znj , … , znm ) Matrix Z “Distance” A1 … Ai … An A1 = ( d11 , … , d1i , … , d1n ) … … Ai = ( di1 , … , dii , … , din ) An = ( dn1 , … , dni , … , dnn ) Matrix D

Hierarchical clustering: agglomerative algorithm Objects (alternatives) Criteria (characteristics) C1 … Cj … Cm A1 = ( z11 , … , z1j , … , z1m ) … … Ai = ( zi1 , … , zij , … , zim ) An = ( zn1 , … , znj , … , znm ) Matrix Z “Distance” A1 … Ai … An A1 = ( 0 , … , d1i , … , d1n ) … … Ai = ( di1 , … , 0 , … , din ) An = ( dn1 , … , dni , … , 0 ) Matrix D

Hierarchical clustering: agglomerative algorithm Matrix D: dil = sqrt ( ∑j=1 m ( zij – zlj )2 ) Scale for D min max

Hierarchical clustering: agglomerative algorithm Stage 1.Computing matrix D=| dil | (pair “distances”) Stage 2.Revelation of the smallest pair “distance” (i.e., the minimal pair “distance”, the minimal element in matrix D) and integration of the corresponding elements (Ax, Ay) (objects) into a new joint (integrated) object A=Ax*Ay Stage 3.Stopping the process or re-computing the matrix D and GOTO Stage 2.

Hierarchical clustering: agglomerative algorithm Pair of objects Ax and Ay Ax = ( zx1 , … , zxj , … , zxm ) Ay = ( zy1 , … , zyj , … , zym ) Integrated object A = ( Ax * Ay )  j (j = 1,…,m) zi (A) = ( zxj + zyj ) / 2

Hierarchical clustering: agglomerative algorithm Stage 6: (1*2*3*4*5*6*7) . . . Stage 4: 2 (1*3*4) (5*6*7) Stage 3: 2 (1*3*4) 5 (6*7) Stage 2: 1 2 (3*4) 5 (6*7) Stage 1: 1 2 (3*4) 5 6 7 Stage 0: 1 2 3 4 5 6 7

Hierarchical clustering: agglomerative algorithm ILLUSTRATIVE EXAMPLE Cluster F4 Cluster F2 Cluster F3 Cluster F5 Cluster F1 Cluster F6

Hierarchical clustering: agglomerative algorithm First, Complexity of agglomerative algorithm: 1.Number of stages (each stage – one integration): (n-1) stages 2.Each stage: (a)computing “distances” (n2 * m operations) THUS: Operations: O(m n3 ) Memory: O(n(n+m)) Second, we have got the TREE-LIKE STRUCTURE

Hierarchical clustering: IMPROVEMENTS (to do better) Question 1: What we can do better in the algorithm?

Hierarchical clustering: IMPROVEMENTS (to do better) Question 1: What we can do better in the algorithm? Question 2: What is needed in practice (e.g., applications)? What we can do for applications?

Hierarchical clustering: IMPROVEMENTS (to do better) Question 1: What we can do better? 1.Computing: pair “distance” (pair proximity) Usage of more “correct” approaches from multicriteria decision making, e.g., Revelation of Pareto-layers and usage of an ordinal scale for pair proximity 2.Complexity: decrease the number of stages: Integration of several pair of objects at each stage

Hierarchical clustering: IMPROVEMENTS (to do better) 2.Complexity: decrease the number of stages: Integration of several pair of objects at each stage Usage of an ordinal scale: max{dxy}

Hierarchical clustering: IMPROVEMENTS (to do better) 2.Complexity: decrease the number of stages: Integration of several pair of objects at each stage Usage of an ordinal scale: max{dxy} To divide the interval [0,max{dxy}] to get an ordinal scale

Hierarchical clustering: IMPROVEMENTS (to do better) 2.Complexity: decrease the number of stages: Integration of several pair of objects at each stage Usage of an ordinal scale: interval 0 interval 1 interval k max{dxy} To divide the interval [0,max{dxy}] to get an ordinal scale

Hierarchical clustering: IMPROVEMENTS (to do better) 2.Complexity: decrease the number of stages: Integration of several pair of objects at each stage Usage of an ordinal scale: dab duv dpq dgh interval 0 interval 1 interval k max{dxy} To divide the interval [0,max{dxy}] to get an ordinal scale Example: pairs of objects: (a,b), (u,v), (p,q), (g,h)

Hierarchical clustering: IMPROVEMENTS (to do better) 2.Complexity: decrease the number of stages: Integration of several pair of objects at each stage RESULT: dab = 0 duv = 0 dpq = 1 dgh = 1 Usage of an ordinal scale: dab duv dpq dgh interval 0 interval 1 interval k max{dxy} To divide the interval [0,max{dxy}] to get an ordinal scale

Hierarchical clustering: IMPROVEMENTS (to do better) 1.Computing: pair “distance” (pair proximity) Usage of more “correct” approaches from multicriteria decision making, e.g., Revelation of Pareto-layers and usage of an ordinal scale for pair proximity

Hierarchical clustering: IMPROVEMENTS (to do better) 1.Computing: pair “distance” (pair proximity) Usage of more “correct” approaches from multicriteria decision making, e.g., Revelation of Pareto-layers and usage of an ordinal scale for pair proximity Objects: {A1, … , Ai, … An } Ax -> (zx1, … , zxj , ... , zxm) Ay -> (zy1, … , zyj , … , zym)

Hierarchical clustering: IMPROVEMENTS (to do better) 1.Computing: pair “distance” (pair proximity) Usage of more “correct” approaches from multicriteria decision making, e.g., Revelation of Pareto-layers and usage of an ordinal scale for pair proximity Objects: {A1, … , Ai, … An } Ax -> (zx1, … , zxj , ... , zxm) Ay -> (zy1, … , zyj , … , zym) Vector of “differences” for Ax , Ay: ( (zx1 - zy1) , … , (zxj - zyj), … , (zxm - zym) )

Hierarchical clustering: IMPROVEMENTS (to do better) 1.Computing: pair “distance” (pair proximity) Usage of more “correct” approaches from multicriteria decision making, e.g., Revelation of Pareto-layers and usage of an ordinal scale for pair proximity Objects: {A1, … , Ai, … An } Ax -> (zx1, … , zxj , ... , zxm) Ay -> (zy1, … , zyj , … , zym) Vector of “differences” for Ax , Ay: ( (zx1 - zy1) , … , (zxj - zyj), … , (zxm - zym) ) Space of the vectors

Hierarchical clustering: IMPROVEMENTS (to do better) 1.Computing: pair “distance” (pair proximity) Usage of more “correct” approaches from multicriteria decision making, e.g., Revelation of Pareto-layers and usage of an ordinal scale for pair proximity Objects: {A1, … , Ai, … An } Ax -> (zx1, … , zxj , ... , zxm) Ay -> (zy1, … , zyj , … , zym) Vector of “differences” for Ax , Ay: ( (zx1 - zy1) , … , (zxj - zyj), … , (zxm - zym) ) Space of the vectors =>ordinal scale&ordinal proximity

Hierarchical clustering: IMPROVEMENTS (to do better) 1.Computing: pair “distance” (pair proximity) Usage of more “correct” approaches from multicriteria decision making, e.g., Revelation of Pareto-layers and usage of an ordinal scale for pair proximity C1 Pareto-effective Layer (1) Layer 2 Ideal point (equal objects) C2 Space of the vectors =>ordinal scale&ordinal proximity

Hierarchical clustering: IMPROVEMENTS (practice) Question 2: What is needed in practice (e.g., applications)? What we can do for applications?

Hierarchical clustering: IMPROVEMENTS (practice) Question 2: What is needed in practice (e.g., applications)? What we can do for applications? Integration of objects into several groups (clusters) to obtain more rich resultant structure (tree => hierarchy, i.e., clusters with intersection) Examples of applied domains: 1.Engineering: structures of complex systems 2.CS: structures of software/hardware 3.Communication networks (topology) 4.Biology 5.Others

Hierarchical clustering: IMPROVEMENTS (practice) Question 2: What is needed in practice (e.g., applications)? What we can do for applications? Clustering with intersection Cluster F3 Cluster F4 Cluster F2 Cluster F1 Cluster F5 Cluster F6

Hierarchical clustering: IMPROVEMENTS (practice) Stage 4: (1*2*3*4*5*6*7) 2 Stage 3: (3*4*5*6*7) (1*2*3*4) Stage 2: 1 (2*3*4) (6*7) (3*4*5*6) Stage 1: 1 (2*3) (3*4) (5*6) (6*7) Stage 0: 1 2 3 4 5 6 7

Hierarchical clustering: IMPROVEMENTS (practice) Resultant structure (1*2*3*4*5*6*7) Stage 3: Stage 2: Stage 1: Stage 0: 1 2 3 4 5 6 7

Hierarchical clustering: IMPROVEMENTS (practice) Example from biology (evolution) Traditional evolution process as tree

Hierarchical clustering: IMPROVEMENTS (practice) Example from biology (evolution) Hierarchical structure

Hierarchical clustering: IMPROVEMENTS (practice) Algorithm 1. The number of inclusion for each object is not limited): (i)initial set of objects -> vertices (ii)”small” proximity -> edges Thus: a graph Problem: to reveal cliques in the graph (It is NP-hard problem) Algorithm 2. The number of the inclusion is limited by t (e.g., t=2/3/4). Here complexity is polynomial.

Hierarchical clustering: performance (i.e., quality) Performance (i.e., quality) of clustering procedures: 1.Issues of complexity 2.Quality of results (??) Some traditional approaches: (a)computing a clustering quality InterCluster Distance / IntraCluster Distance (b)Coverage, Diversity Our case: research procedure (for investigation and problem structuring)

Basic DM problems: choice, ranking, Hierarchical clustering: performance (i.e., quality) Decision Making Paradigm (stages) by Herbert A. Simon 1.Analysis of an applied problem (to understand the problem: main contradictions, etc.) 2.Structuring the problem: 2.1.Generation of alternatives 2.2.Design of criteria 2.3.Design of scales for assessment of alternatives upon criteria 3.Evaluation of alternatives upon criteria 4.Selection of the best alternative (s) 5.Analysis of results Basic DM problems: choice, ranking,

THUS: we have got some prospective RESEARCH RPOCEDURES Hierarchical clustering: performance (i.e., quality) FOR CLUSTERING: 1.Analysis of an applied problem (to understand the problem: main contradictions, etc.) 2.Structuring the problem: 2.1.Generation of alternatives 2.2.Design of criteria 2.3.Design of scales for assessment of alternatives upon criteria 3.Evaluation of alternatives upon criteria 4.Design of CLUSTERS and STRUCTURE OF CLUSTERING PROCESS 5.Analysis of results THUS: we have got some prospective RESEARCH RPOCEDURES

1.Algorithms, procedures & their analysis 2.New approaches to CONCLUSION 1.Algorithms, procedures & their analysis 2.New approaches to performance/quality for research procedures 3.Various applied examples 4.Usage in education

That’s All Thanks! http://www.iitp.ru/mslevin/ Mark Sh. Levin