Roberto Battiti, Mauro Brunato

Slides:



Advertisements
Similar presentations
Clustering II.
Advertisements

SEEM Tutorial 4 – Clustering. 2 What is Cluster Analysis?  Finding groups of objects such that the objects in a group will be similar (or.
Clustering.
Hierarchical Clustering
1 CSE 980: Data Mining Lecture 16: Hierarchical Clustering.
Hierarchical Clustering. Produces a set of nested clusters organized as a hierarchical tree Can be visualized as a dendrogram – A tree-like diagram that.
Hierarchical Clustering, DBSCAN The EM Algorithm
O(N 1.5 ) divide-and-conquer technique for Minimum Spanning Tree problem Step 1: Divide the graph into  N sub-graph by clustering. Step 2: Solve each.
PARTITIONAL CLUSTERING
R OBERTO B ATTITI, M AURO B RUNATO. The LION Way: Machine Learning plus Intelligent Optimization. LIONlab, University of Trento, Italy, Feb 2014.
R OBERTO B ATTITI, M AURO B RUNATO The LION Way: Machine Learning plus Intelligent Optimization. LIONlab, University of Trento, Italy, Feb 2014.
R OBERTO B ATTITI, M AURO B RUNATO. The LION Way: Machine Learning plus Intelligent Optimization. LIONlab, University of Trento, Italy, Feb 2014.
Machine Learning and Data Mining Clustering
Introduction to Bioinformatics
Clustering (1) Clustering Similarity measure Hierarchical clustering Model-based clustering Figures from the book Data Clustering by Gan et al.
Unsupervised learning: Clustering Ata Kaban The University of Birmingham
Clustering II.
Image Segmentation Chapter 14, David A. Forsyth and Jean Ponce, “Computer Vision: A Modern Approach”.
Clustering Petter Mostad. Clustering vs. class prediction Class prediction: Class prediction: A learning set of objects with known classes A learning.
What is Cluster Analysis?
Neural Network Homework Report: Clustering of the Self-Organizing Map Professor : Hahn-Ming Lee Student : Hsin-Chung Chen M IEEE TRANSACTIONS ON.
Clustering Ram Akella Lecture 6 February 23, & 280I University of California Berkeley Silicon Valley Center/SC.
Ulf Schmitz, Pattern recognition - Clustering1 Bioinformatics Pattern recognition - Clustering Ulf Schmitz
R OBERTO B ATTITI, M AURO B RUNATO. The LION Way: Machine Learning plus Intelligent Optimization. LIONlab, University of Trento, Italy, Feb 2014.
R OBERTO B ATTITI, M AURO B RUNATO. The LION Way: Machine Learning plus Intelligent Optimization. LIONlab, University of Trento, Italy, Feb 2014.
Clustering Unsupervised learning Generating “classes”
R OBERTO B ATTITI, M AURO B RUNATO. The LION Way: Machine Learning plus Intelligent Optimization. LIONlab, University of Trento, Italy, Feb 2014.
R OBERTO B ATTITI, M AURO B RUNATO. The LION Way: Machine Learning plus Intelligent Optimization. LIONlab, University of Trento, Italy, Feb 2014.
CSE5334 DATA MINING CSE4334/5334 Data Mining, Fall 2014 Department of Computer Science and Engineering, University of Texas at Arlington Chengkai Li (Slides.
MACHINE LEARNING 8. Clustering. Motivation Based on E ALPAYDIN 2004 Introduction to Machine Learning © The MIT Press (V1.1) 2  Classification problem:
Chapter 12 Object Recognition Chapter 12 Object Recognition 12.1 Patterns and pattern classes Definition of a pattern class:a family of patterns that share.
Data Mining Practical Machine Learning Tools and Techniques By I. H. Witten, E. Frank and M. A. Hall 6.8: Clustering Rodney Nielsen Many / most of these.
Radial Basis Function ANN, an alternative to back propagation, uses clustering of examples in the training set.
Hierarchical Clustering Produces a set of nested clusters organized as a hierarchical tree Can be visualized as a dendrogram – A tree like diagram that.
Hierarchical Clustering
© Tan,Steinbach, Kumar Introduction to Data Mining 4/18/ Data Mining: Cluster Analysis This lecture node is modified based on Lecture Notes for Chapter.
Data Mining Cluster Analysis: Basic Concepts and Algorithms Lecture Notes Introduction to Data Mining by Tan, Steinbach, Kumar © Tan,Steinbach, Kumar Introduction.
Clustering Machine Learning Unsupervised Learning K-means Optimization objective Random initialization Determining Number of Clusters Hierarchical Clustering.
Data Science Practical Machine Learning Tools and Techniques 6.8: Clustering Rodney Nielsen Many / most of these slides were adapted from: I. H. Witten,
Clustering (1) Clustering Similarity measure Hierarchical clustering
Roberto Battiti, Mauro Brunato
Data Mining: Basic Cluster Analysis
Unsupervised Learning: Clustering
Roberto Battiti, Mauro Brunato
Hierarchical Clustering
Unsupervised Learning: Clustering
Machine Learning for the Quantified Self
PREDICT 422: Practical Machine Learning
CSE 4705 Artificial Intelligence
Clustering CSC 600: Data Mining Class 21.
Chapter 15 – Cluster Analysis
Hierarchical Clustering
Data Clustering Michael J. Watts
Roberto Battiti, Mauro Brunato
Roberto Battiti, Mauro Brunato
Roberto Battiti, Mauro Brunato
Roberto Battiti, Mauro Brunato
Roberto Battiti, Mauro Brunato
John Nicholas Owen Sarah Smith
Roberto Battiti, Mauro Brunato
Roberto Battiti, Mauro Brunato
Roberto Battiti, Mauro Brunato
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
CSCI N317 Computation for Scientific Applications Unit Weka
Text Categorization Berlin Chen 2003 Reference:
Hierarchical Clustering
Clustering Techniques
SEEM4630 Tutorial 3 – Clustering.
Hierarchical Clustering
Presentation transcript:

Roberto Battiti, Mauro Brunato Roberto Battiti, Mauro Brunato. The LION Way: Machine Learning plus Intelligent Optimization. LIONlab, University of Trento, Italy, Apr 2015 http://intelligent-optimization.org/LIONbook © Roberto Battiti and Mauro Brunato , 2015, all rights reserved. Slides can be used and modified for classroom usage, provided that the attribution (link to book website) is kept.

Bottom-up (agglomerative) clustering Birds of a feather flock together.

Agglomerative clustering: definition Hierarchical algorithms find successive clusters by merging previously established smaller clusters, Begin with each element as a separate cluster. At each step the most similar clusters are merged. Note: This requires a measure of similarity between two clusters. (or between a cluster and a single element).

Distance between clusters A distance between clusters can be derived from a distance between elements Three main choices: C,D are two clusters, δ(x,y) is the distance between the elements x and y.

Merging algorithm Given a set of clusters C, proceed as follows: Find clusters C and D in C with minimum distance substitute C and D with their union C U D, and register δ* as the distance for which the specific merging occurred; until a single cluster containing all entities is obtained

Dendrogram Hierarchical merging can be visualized through dendrograms, where the original entities are at the bottom and each merging action is represented with a horizontal line connecting the two fused clusters (at a level given by the dissimilarity)

A dendrogram for vehicles

Mahalanobis distance Prompted by the problem of identifying similarities of skulls based on measurements in 1927 Given a point x, how do we define the probability for the point to belong to a cluster? For a symmetric spherical distribution: Evaluate the standard deviation σ of the distances of the sample points Evaluate its distance from the average of points (center of mass) Define the normalized distance |x-μ|/σ Derive the probability of the test point belonging to the set from the normal distribution

Mahalanobis distance (2) If the distribution is highly non-spherical the probability of the test point belonging to the set will depend also on the direction of the vector (x- μ)

Mahalanobis distance (3) In the case on the left we can use the Euclidean distance as a dissimilarity measure, while in the other case we need to refer to the Mahalanobis distance, because the data are distributed in an ellipsoidal shape.

Mahalanobis distance (4) The ellipsoid that best represents the set’s probability distribution can be estimated by building the covariance matrix of the samples The center p of the cluster is the mean value: Let the covariance matrix components be defined as:

Mahanalobis distance (5) The Mahalanobis distance of a vector x from a set of values with mean and covariance matrix S is defined as: It is the distance of the test point from the center of mass divided by the width of the ellipsoid in the direction of the test point

Clustering visualization

Gist Agglomerative clustering builds a tree (a hierarchical organization) containing bigger and bigger clusters It is a “bottom up” method: first nearby points are merged, then similar sets are merged, until a single set is obtained. The number of clusters is not specified at the beginning a proper number can be obtained by cutting the tree (a.k.a. dendrogram) at an appropriate level of similarity