Unsupervised Learning and Clustering k-means clustering Sum-of-Squared Errors Competitive Learning SOM Pre-processing and Post-processing techniques.

Slides:



Advertisements
Similar presentations
Random Forest Predrag Radenković 3237/10
Advertisements

PARTITIONAL CLUSTERING
2806 Neural Computation Self-Organizing Maps Lecture Ari Visa.
Neural Networks Dr. Peter Phillips. The Human Brain (Recap of week 1)
1 Machine Learning: Lecture 10 Unsupervised Learning (Based on Chapter 9 of Nilsson, N., Introduction to Machine Learning, 1996)
CS Clustering1 Unsupervised Learning and Clustering In unsupervised learning you are given a data set with no output classifications Clustering is.
Neural Networks Chapter 9 Joost N. Kok Universiteit Leiden.
Self Organization: Competitive Learning
Kohonen Self Organising Maps Michael J. Watts
Unsupervised Networks Closely related to clustering Do not require target outputs for each input vector in the training data Inputs are connected to a.
X0 xn w0 wn o Threshold units SOM.
RBF Neural Networks x x1 Examples inside circles 1 and 2 are of class +, examples outside both circles are of class – What NN does.
© University of Minnesota Data Mining for the Discovery of Ocean Climate Indices 1 CSci 8980: Data Mining (Fall 2002) Vipin Kumar Army High Performance.
Clustering.
Three kinds of learning
KNN, LVQ, SOM. Instance Based Learning K-Nearest Neighbor Algorithm (LVQ) Learning Vector Quantization (SOM) Self Organizing Maps.
Introduction to Bioinformatics - Tutorial no. 12
What is Cluster Analysis?
1 Kunstmatige Intelligentie / RuG KI2 - 7 Clustering Algorithms Johan Everts.
1 Study of Topographic and Equiprobable Mapping with Clustering for Fault Classification Ashish Babbar EE645 Final Project.
Microarray analysis 2 Golan Yona. 2) Analysis of co-expression Search for similarly expressed genes experiment1 experiment2 experiment3 ……….. Gene i:
Clustering Ram Akella Lecture 6 February 23, & 280I University of California Berkeley Silicon Valley Center/SC.
A Hybrid Self-Organizing Neural Gas Network James Graham and Janusz Starzyk School of EECS, Ohio University Stocker Center, Athens, OH USA IEEE World.
Lecture 09 Clustering-based Learning
Radial-Basis Function Networks
Face Recognition Using Neural Networks Presented By: Hadis Mohseni Leila Taghavi Atefeh Mirsafian.
Introduction to undirected Data Mining: Clustering
Clustering Unsupervised learning Generating “classes”
Evaluating Performance for Data Mining Techniques
 C. C. Hung, H. Ijaz, E. Jung, and B.-C. Kuo # School of Computing and Software Engineering Southern Polytechnic State University, Marietta, Georgia USA.
Methods in Medical Image Analysis Statistics of Pattern Recognition: Classification and Clustering Some content provided by Milos Hauskrecht, University.
Unsupervised Learning. CS583, Bing Liu, UIC 2 Supervised learning vs. unsupervised learning Supervised learning: discover patterns in the data that relate.
KOHONEN SELF ORGANISING MAP SEMINAR BY M.V.MAHENDRAN., Reg no: III SEM, M.E., Control And Instrumentation Engg.
Self Organizing Maps (SOM) Unsupervised Learning.
CZ5225: Modeling and Simulation in Biology Lecture 5: Clustering Analysis for Microarray Data III Prof. Chen Yu Zong Tel:
Self-organizing Maps Kevin Pang. Goal Research SOMs Research SOMs Create an introductory tutorial on the algorithm Create an introductory tutorial on.
Artificial Neural Networks Dr. Abdul Basit Siddiqui Assistant Professor FURC.
COMMON EVALUATION FINAL PROJECT Vira Oleksyuk ECE 8110: Introduction to machine Learning and Pattern Recognition.
A Scalable Self-organizing Map Algorithm for Textual Classification: A Neural Network Approach to Thesaurus Generation Dmitri G. Roussinov Department of.
Stephen Marsland Ch. 9 Unsupervised Learning Stephen Marsland, Machine Learning: An Algorithmic Perspective. CRC 2009 based on slides from Stephen.
Machine Learning Neural Networks (3). Understanding Supervised and Unsupervised Learning.
CLUSTERING. Overview Definition of Clustering Existing clustering methods Clustering examples.
Clustering What is clustering? Also called “unsupervised learning”Also called “unsupervised learning”
Neural Networks - Lecture 81 Unsupervised competitive learning Particularities of unsupervised learning Data clustering Neural networks for clustering.
IE 585 Competitive Network – Learning Vector Quantization & Counterpropagation.
Clustering.
Computational Biology Clustering Parts taken from Introduction to Data Mining by Tan, Steinbach, Kumar Lecture Slides Week 9.
Semiconductors, BP&A Planning, DREAM PLAN IDEA IMPLEMENTATION.
Example Apply hierarchical clustering with d min to below data where c=3. Nearest neighbor clustering d min d max will form elongated clusters!
Debrup Chakraborty Non Parametric Methods Pattern Recognition and Machine Learning.
Artificial Neural Networks for Data Mining. Copyright © 2011 Pearson Education, Inc. Publishing as Prentice Hall 6-2 Learning Objectives Understand the.
SUPERVISED AND UNSUPERVISED LEARNING Presentation by Ege Saygıner CENG 784.
DATA MINING: CLUSTER ANALYSIS Instructor: Dr. Chun Yu School of Statistics Jiangxi University of Finance and Economics Fall 2015.
Rodney Nielsen Many of these slides were adapted from: I. H. Witten, E. Frank and M. A. Hall Data Science Algorithms: The Basic Methods Clustering WFH:
Machine Learning 12. Local Models.
Chapter 5 Unsupervised learning
Self-Organizing Network Model (SOM) Session 11
Semi-Supervised Clustering
Data Mining, Neural Network and Genetic Programming
Data Mining K-means Algorithm
Other Applications of Energy Minimzation
Clustering (3) Center-based algorithms Fuzzy k-means
Dr. Unnikrishnan P.C. Professor, EEE
Lecture 22 Clustering (3).
Data Mining 資料探勘 分群分析 (Cluster Analysis) Min-Yuh Day 戴敏育
Feature mapping: Self-organizing Maps
Artificial Neural Networks
Unsupervised Networks Closely related to clustering
Presentation transcript:

Unsupervised Learning and Clustering k-means clustering Sum-of-Squared Errors Competitive Learning SOM Pre-processing and Post-processing techniques

K-means clustering This an elementary but very popular method for clustering. Our goal is to find the k mean vectors or “cluster centers”.  Initialize k, m1, m2, …, mk  Repeat  Classify samples according to its nearest mi  Recompute mi  Until there is no change in mi  Return m1, m2, …, mk

Complexity The computational complexity of the algorithm is defined as follows: O( n d c T ) Where d is the number of features, n is the number of examples, c is the number of clusters, and T is the number of iterations. The number of iterations is normally much less than the number of examples.

Figure 10.3

K-means clustering Disadvantage 1: Prone to fall into local minima. This can be solved with more computational power by running the algorithm many times with different initial means. Disadvantage 2: Susceptible to outliers. One solution is to replace the mean with the median.

K-means clustering Hugo Steinhaus Born in January 14, 1887 (Austria-Hungary). Professor at the University of Wroclaw, Notre Dame, and Sussex. Authored over 170 works in mathematics. First one to use k-means clustering

Unsupervised Learning and Clustering k-means clustering Sum-of-Squared Errors Competitive Learning SOM Pre-processing and Post-processing techniques

The Sum-of-Squared Error We can now define the goal of clustering: Goal: To divide a dataset of examples into c disjoint subsets D1, D2, …, Dc, so that the distance between examples within the same partition is small compared to the distance between examples on different partitions. To achieve this, we define the c means by looking to minimize a metric.

Metric Let mi be the mean of examples on partition Di: mi = (1 / ni) Σ x (for all x in Di) Then the metric to minimize is the sum-of-squared errors: Je = Σi Σx || x – mi || 2 For all x in Di where index i goes along the clusters.

Figure 10.10

Others  Hierarchical clustering Clusters have subclusters which also have subclusters and so on.  Online clustering As time goes on new information may call for restructuring the clusters (plasticity). But we don’t want this to happen very often (stability).

Figure 10.11

Unsupervised Learning and Clustering k-means clustering Sum-of-Squared Errors Competitive Learning SOM Pre-processing and Post-processing techniques

Vector Quantisation Data will be represented with prototype vectors.

Feature Mapping Input Nodes

Feature Mapping Input Nodes [ x1, x2, x3, x4 ] T [ w1, w2, w3, w4 ] T w1 w2 w3 w4

Feature Mapping Weight vector will be mapped into the feature space. [ w1, w2, w3, w4 ] T [ x1, x2, x3, x4 ] T

SOM Algorithm Initialization Select the number of neurons in the map Choose random values for all weights Learning Repeat For each example, find the neuron closest to the point: min || x - w ||

SOM Algorithm Winner takes all Input Nodes Update weights of winner only (and neighbors)

SOM Algorithm Update Weights Update weights for the closest neuron and neighbors: w t+1 = w t +  A(x,w) (x – w) where  is the learning rate Function A defines a neighboring function.

SOM Algorithm The neighboring function A:

SOM Algorithm Usage For every test point Select the closest neuron using minimum Euclidean distance: min || x - w ||

Mapping a Grid to a Grid

SOM Algorithm Comments Neighborhoods should be large at the beginning but short as the nodes gain a specific ordering Global ordering comes naturally (complexity theory) Architecture of the map: Few nodes: underfitting Many nodes: overfitting

Teuvo Kohonen Born in 1934, Finland He has several books and over 300 papers His most famous work is in Self Organizing Maps Member of the Academy of Finland Awards: IEEE Neural Networks Council Pioneer Award, 1991 Technical Achievement Award of IEEE, 1995 Frank Rosenblatt Technical Field Award, 2008

Unsupervised Learning and Clustering k-means clustering Sum-of-Squared Errors Competitive Learning SOM Pre-processing and Post-processing techniques

Cluster Tendency Cluster tendency is a preprocessing step that indicates when data objects exhibit a clustering structure; it precludes using clustering when the data appears randomly generated under the uniform distribution over a sample window of interest in the attribute space

Example Cluster Tendency Clustering captures inherent data groups. Clustering does not capture groups; Results come from random variations.

Example Cluster Tendency Problem: How do we choose the sampling window? Rule of thumb: Create a window centered at the mean that captures half the total number of examples.

Cluster Validation Cluster validation is used to assess the value of the output of a clustering algorithm. Internal Statistics are devised to capture the quality of the induced clusters using the available data objects. External If the validation is performed by gathering statistics comparing the induced clusters against an external and independent classification of objects, the validation is called external.

Example Cluster Validation

Metrics Cluster Validation One type of statistical metrics is defined in terms of a 2 x2 table where each entry counts the number of object pairs that agree or disagree with the class and cluster to which they belong: E11 E12 E21 E22 Same class; Same cluster Different class; Different cluster Same class; Different cluster Different class; Same cluster

Examples Metrics Cluster Validation Rand: [ E11 + E22 ] / [ E11 + E12 + E21 + E22 ] Jaccard: E11 / [ E11 + E12 + E21 ]