Intro. ANN & Fuzzy Systems Lecture 20 Clustering (1)

Slides:



Advertisements
Similar presentations
1 Machine Learning: Lecture 10 Unsupervised Learning (Based on Chapter 9 of Nilsson, N., Introduction to Machine Learning, 1996)
Advertisements

Neural Networks Chapter 9 Joost N. Kok Universiteit Leiden.
Self Organization: Competitive Learning
Kohonen Self Organising Maps Michael J. Watts
Unsupervised Learning with Artificial Neural Networks The ANN is given a set of patterns, P, from space, S, but little/no information about their classification,
Intro. ANN & Fuzzy Systems Lecture 8. Learning (V): Perceptron Learning.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
November 9, 2010Neural Networks Lecture 16: Counterpropagation 1 Unsupervised Learning So far, we have only looked at supervised learning, in which an.
Prénom Nom Document Analysis: Data Analysis and Clustering Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Slide 1 EE3J2 Data Mining Lecture 16 Unsupervised Learning Ali Al-Shahib.
Clustering.
Clustering. 2 Outline  Introduction  K-means clustering  Hierarchical clustering: COBWEB.
Semi-Supervised Clustering Jieping Ye Department of Computer Science and Engineering Arizona State University
What is Cluster Analysis?
Intro. ANN & Fuzzy Systems Lecture 21 Clustering (2)
Lecture 09 Clustering-based Learning
Introduction to undirected Data Mining: Clustering
Clustering Unsupervised learning Generating “classes”
Evaluating Performance for Data Mining Techniques
KOHONEN SELF ORGANISING MAP SEMINAR BY M.V.MAHENDRAN., Reg no: III SEM, M.E., Control And Instrumentation Engg.
Unsupervised Learning and Clustering k-means clustering Sum-of-Squared Errors Competitive Learning SOM Pre-processing and Post-processing techniques.
CZ5225: Modeling and Simulation in Biology Lecture 5: Clustering Analysis for Microarray Data III Prof. Chen Yu Zong Tel:
COMMON EVALUATION FINAL PROJECT Vira Oleksyuk ECE 8110: Introduction to machine Learning and Pattern Recognition.
Chapter 8 The k-Means Algorithm and Genetic Algorithm.
Apache Mahout. Mahout Introduction Machine Learning Clustering K-means Canopy Clustering Fuzzy K-Means Conclusion.
CSC321: Neural Networks Lecture 13: Learning without a teacher: Autoencoders and Principal Components Analysis Geoffrey Hinton.
© by Yu Hen Hu 1 ECE533 Digital Image Processing Image Segmentation.
1 Chapter 8: Introduction to Pattern Discovery 8.1 Introduction 8.2 Cluster Analysis 8.3 Market Basket Analysis (Self-Study)
Intro. ANN & Fuzzy Systems Lecture 23 Clustering (4)
Intro. ANN & Fuzzy Systems Lecture 14. MLP (VI): Model Selection.
Clustering. 2 Outline  Introduction  K-means clustering  Hierarchical clustering: COBWEB.
Radial Basis Function ANN, an alternative to back propagation, uses clustering of examples in the training set.
Clustering Instructor: Max Welling ICS 178 Machine Learning & Data Mining.
Machine Learning Queens College Lecture 7: Clustering.
CHAPTER 14 Competitive Networks Ming-Feng Yeh.
Lloyd Algorithm K-Means Clustering. Gene Expression Susumu Ohno: whole genome duplications The expression of genes can be measured over time. Identifying.
Intro. ANN & Fuzzy Systems Lecture 15. Pattern Classification (I): Statistical Formulation.
Intro. ANN & Fuzzy Systems Lecture 13. MLP (V): Speed Up Learning.
Intro. ANN & Fuzzy Systems Lecture 38 Mixture of Experts Neural Network.
Example Apply hierarchical clustering with d min to below data where c=3. Nearest neighbor clustering d min d max will form elongated clusters!
Intro. ANN & Fuzzy Systems Lecture 16. Classification (II): Practical Considerations.
Given a set of data points as input Randomly assign each point to one of the k clusters Repeat until convergence – Calculate model of each of the k clusters.
Debrup Chakraborty Non Parametric Methods Pattern Recognition and Machine Learning.
Intro. ANN & Fuzzy Systems Lecture 37 Genetic and Random Search Algorithms (2)
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Mixture Densities Maximum Likelihood Estimates.
Cluster Analysis What is Cluster Analysis? Types of Data in Cluster Analysis A Categorization of Major Clustering Methods Partitioning Methods.
DATA MINING: CLUSTER ANALYSIS Instructor: Dr. Chun Yu School of Statistics Jiangxi University of Finance and Economics Fall 2015.
Supervised Learning – Network is presented with the input and the desired output. – Uses a set of inputs for which the desired outputs results / classes.
Clustering Machine Learning Unsupervised Learning K-means Optimization objective Random initialization Determining Number of Clusters Hierarchical Clustering.
Machine Learning Lecture 4: Unsupervised Learning (clustering) 1.
Intro. ANN & Fuzzy Systems Lecture 11. MLP (III): Back-Propagation.
Lecture 39 Hopfield Network
Chapter 5 Unsupervised learning
Self-Organizing Network Model (SOM) Session 11
Semi-Supervised Clustering
Clustering and Segmentation
Classification of unlabeled data:
Lecture 12. MLP (IV): Programming & Implementation
Creating fuzzy rules from numerical data using a neural network
Lecture 12. MLP (IV): Programming & Implementation
Lecture 9 MLP (I): Feed-forward Model
Lecture 22 Clustering (3).
Outline Single neuron case: Nonlinear error correcting learning
Competitive Networks.
Lecture 7. Learning (IV): Error correcting Learning and LMS Algorithm
Competitive Networks.
LECTURE 21: CLUSTERING Objectives: Mixture Densities Maximum Likelihood Estimates Application to Gaussian Mixture Models k-Means Clustering Fuzzy k-Means.
CS 621 Artificial Intelligence Lecture 27 – 21/10/05
Introduction to Machine learning
Lecture 16. Classification (II): Practical Considerations
Presentation transcript:

Intro. ANN & Fuzzy Systems Lecture 20 Clustering (1)

Intro. ANN & Fuzzy Systems (C) 2001 by Yu Hen Hu 2 Outline Unsupervised learning (Competitive Learning) and Clustering K-Means Clustering Algorithm

Intro. ANN & Fuzzy Systems (C) 2001 by Yu Hen Hu 3 Unsupervised Learning Data Mining –Understand internal/hidden structure of data distribution Labeling (Target value, teaching input) Cost is High –Large amount of feature vectors –Sampling may involve costly experiments –Data label may not be available at all Pre-processing for classification –features within the same cluster are similar, and –often belong to the same class

Intro. ANN & Fuzzy Systems (C) 2001 by Yu Hen Hu 4 Competitive Learning A form of unsupervised learning. Neurons compete against each other with their activation values. The winner(s) reserve the privilege to update their weights. The losers may even be punished by updating their weights in opposite direction. Competitive and Cooperative Learning: Competitive: Only one neuron's activation can be reinforced. Cooperative: Several neurons' activation can be reinforced.

Intro. ANN & Fuzzy Systems (C) 2001 by Yu Hen Hu 5 Competitive Learning Rule A neuron WINS the competition if its output is largest among all neurons for the same input x(n). The weights of the winning neuron (k-th) is adjusted:  w k (n)  [x(n) – w k (n)] The positions of losing neurons remain unchanged. If the weights of a neuron represents its POSITION. If the output of a neuron is inversely proportional to the distance between x(n) and w k (n), then Competitive Learning = CLUSTERING!

Intro. ANN & Fuzzy Systems (C) 2001 by Yu Hen Hu 6 Competitive Learning Example learncl1.m

Intro. ANN & Fuzzy Systems (C) 2001 by Yu Hen Hu 7 What is “Clustering”? What can we learn from these “unlabeled” data samples? –Structures: Some samples are closer to each other than other samples –The closeness between samples are determined using a “similarity measure” –The number of samples per unit volume is related to the concept of “density” or “distribution”

Intro. ANN & Fuzzy Systems (C) 2001 by Yu Hen Hu 8 Clustering Problem Statement Given a set of vectors {x k ; 1  k  K}, find a set of M clustering centers {w(i); 1  i  c} such that each x k is assigned to a cluster, say, w(i*), according to a distance (distortion, similarity) measure d(x k, w(i)) such that the average distortion is minimized. I(x k,i) = 1 if x is assigned to cluster i with cluster center w(I); and = 0 otherwise -- indicator function.

Intro. ANN & Fuzzy Systems (C) 2001 by Yu Hen Hu 9 k-means Clustering Algorithm Initialization: Initial cluster center w(i); 1  i  c, D(–1)= 0, I(x k,i) = 0, 1  i  c, 1  k  K; Repeat (A) Assign cluster membership (Expectation step) Evaluate d(x k, w(i)); 1  i  c, 1  k  K I(x k,i) = 1 if d(x k, w(i)) < d(x k, w(j)), j  i; = 0; otherwise. 1  k  K (B) Evaluate distortion D: (C) Update code words according to new assignment (Maximization) (D) Check for convergence if 1–D(Iter–1)/D(Iter) < , then convergent = TRUE,

Intro. ANN & Fuzzy Systems (C) 2001 by Yu Hen Hu 10 A Numerical Example x = {-1, -2,0,2,3,4}, W={2.1, 2.3} 1.Assign membership 2.1: {-1, -2, 0, 2} 2.3: {3, 4} 2.Distortion D = (-1-2.1) 2 + (-2-2.1) 2 + (0-2.1) 2 + (2-2.1) 2 + (3-2.3) 2 + (4-2.3) 2 3. Update W to minimize distortion W 1 = ( )/4 = -.25 W 2 = (3+4)/2 = Reassign membership -.25: {-1, -2, 0} 3.5: {2, 3, 4} 5.Update W: w 1 = (-1-2+0)/3 = -1 w 2 = (2+3+4)/3 = 3. Converged.

Intro. ANN & Fuzzy Systems (C) 2001 by Yu Hen Hu 11 Kmeans Algorithm Demonstration Clusterdemo.m