Neural Network Homework Report: Clustering of the Self-Organizing Map Professor : Hahn-Ming Lee Student : Hsin-Chung Chen M9315928 IEEE TRANSACTIONS ON.

Slides:



Advertisements
Similar presentations
A New Evolving Tree for Text Document Clustering and Visualization 1 Wui Lee Chang, 1* Kai Meng Tay, 2 Chee Peng Lim 1 Faculty of Engineering, Universiti.
Advertisements

Ch2 Data Preprocessing part3 Dr. Bernard Chen Ph.D. University of Central Arkansas Fall 2009.
Learning Trajectory Patterns by Clustering: Comparative Evaluation Group D.
1 Machine Learning: Lecture 10 Unsupervised Learning (Based on Chapter 9 of Nilsson, N., Introduction to Machine Learning, 1996)
Intelligent Database Systems Lab Advisor : Dr.Hsu Graduate : Keng-Wei Chang Author : Gianfranco Chicco, Roberto Napoli Federico Piglione, Petru Postolache.
DNA Microarray Bioinformatics - #27611 Program Normalization exercise (from last week) Dimension reduction theory (PCA/Clustering) Dimension reduction.
1 Learning to Detect Objects in Images via a Sparse, Part-Based Representation S. Agarwal, A. Awan and D. Roth IEEE Transactions on Pattern Analysis and.
1 Lecture 5: Automatic cluster detection Lecture 6: Artificial neural networks Lecture 7: Evaluation of discovered knowledge Brief introduction to lectures.
© University of Minnesota Data Mining for the Discovery of Ocean Climate Indices 1 CSci 8980: Data Mining (Fall 2002) Vipin Kumar Army High Performance.
Dimension reduction : PCA and Clustering Agnieszka S. Juncker Slides: Christopher Workman and Agnieszka S. Juncker Center for Biological Sequence Analysis.
Dimension reduction : PCA and Clustering by Agnieszka S. Juncker
Dimension reduction : PCA and Clustering Slides by Agnieszka Juncker and Chris Workman.
Dimension reduction : PCA and Clustering Christopher Workman Center for Biological Sequence Analysis DTU.
Semi-Supervised Clustering Jieping Ye Department of Computer Science and Engineering Arizona State University
Introduction to Bioinformatics - Tutorial no. 12
Microarray analysis 2 Golan Yona. 2) Analysis of co-expression Search for similarly expressed genes experiment1 experiment2 experiment3 ……….. Gene i:
Dimension reduction : PCA and Clustering by Agnieszka S. Juncker Part of the slides is adapted from Chris Workman.
Hazırlayan NEURAL NETWORKS Radial Basis Function Networks I PROF. DR. YUSUF OYSAL.
Revision (Part II) Ke Chen COMP24111 Machine Learning Revision slides are going to summarise all you have learnt from Part II, which should be helpful.
Lecture 09 Clustering-based Learning
Kansas State University Department of Computing and Information Sciences KDD Group Presentation #7 Fall’01 Friday, November 9, 2001 Cecil P. Schmidt Department.
Clustering methods Course code: Pasi Fränti Speech & Image Processing Unit School of Computing University of Eastern Finland Joensuu,
Unsupervised Learning and Clustering k-means clustering Sum-of-Squared Errors Competitive Learning SOM Pre-processing and Post-processing techniques.
Data mining and machine learning A brief introduction.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology Advisor : Dr. Hsu Student : Sheng-Hsuan Wang Department.
Presented by Tienwei Tsai July, 2005
Self-organizing Maps Kevin Pang. Goal Research SOMs Research SOMs Create an introductory tutorial on the algorithm Create an introductory tutorial on.
Self-organizing map Speech and Image Processing Unit Department of Computer Science University of Joensuu, FINLAND Pasi Fränti Clustering Methods: Part.
Artificial Neural Networks Dr. Abdul Basit Siddiqui Assistant Professor FURC.
Self organizing maps 1 iCSC2014, Juan López González, University of Oviedo Self organizing maps A visualization technique with data dimension reduction.
Hierarchical Clustering
Partitional and Hierarchical Based clustering Lecture 22 Based on Slides of Dr. Ikle & chapter 8 of Tan, Steinbach, Kumar.
Text Clustering.
Dimension reduction : PCA and Clustering Slides by Agnieszka Juncker and Chris Workman modified by Hanne Jarmer.
MACHINE LEARNING 8. Clustering. Motivation Based on E ALPAYDIN 2004 Introduction to Machine Learning © The MIT Press (V1.1) 2  Classification problem:
Clustering.
Prepared by: Mahmoud Rafeek Al-Farra
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology 1 The Evolving Tree — Analysis and Applications Advisor.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology A modified version of the K-means algorithm with a distance.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology Rival-Model Penalized Self-Organizing Map Yiu-ming Cheung.
Computational Biology Clustering Parts taken from Introduction to Data Mining by Tan, Steinbach, Kumar Lecture Slides Week 9.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology O( ㏒ 2 M) Self-Organizing Map Algorithm Without Learning.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology 1 A self-organizing map for adaptive processing of structured.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology 1 Adaptive FIR Neural Model for Centroid Learning in Self-Organizing.
Unsupervised Learning
Self-organizing maps applied to information retrieval of dissertations and theses from BDTD-UFPE Bruno Pinheiro Renato Correa
Roberto Battiti, Mauro Brunato
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology 1 Growing Hierarchical Tree SOM: An unsupervised neural.
Example Apply hierarchical clustering with d min to below data where c=3. Nearest neighbor clustering d min d max will form elongated clusters!
Data Mining Cluster Analysis: Basic Concepts and Algorithms Lecture Notes Introduction to Data Mining by Tan, Steinbach, Kumar © Tan,Steinbach, Kumar Introduction.
Data Mining and Text Mining. The Standard Data Mining process.
Data Mining: Basic Cluster Analysis
Semi-Supervised Clustering
Data Mining, Neural Network and Genetic Programming
Dimension reduction : PCA and Clustering by Agnieszka S. Juncker
Basic machine learning background with Python scikit-learn
John Nicholas Owen Sarah Smith
Lecture 22 Clustering (3).
CSE572, CBS598: Data Mining by H. Liu
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
The BIRCH Algorithm Davitkov Miroslav, 2011/3116
CSE572, CBS572: Data Mining by H. Liu
Dimension reduction : PCA and Clustering
Source: Pattern Recognition Vol. 38, May, 2005, pp
Introduction to Cluster Analysis
Clustering The process of grouping samples so that the samples are similar within each group.
CSE572: Data Mining by H. Liu
Artificial Neural Networks
BIRCH: Balanced Iterative Reducing and Clustering using Hierarchies
Presentation transcript:

Neural Network Homework Report: Clustering of the Self-Organizing Map Professor : Hahn-Ming Lee Student : Hsin-Chung Chen M IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL.11, NO.3, MAY 2000

OUTLINE  INTRODUCTION  CLUSTERING  SOM CLUSTERING  EXPERIMENTS  CONCLUSION

INTRODUCTION  DATA mining processes problem definition data acquisition. data preprocessing and survey data modeling evaluation. knowledge deployment.  Self-organization map feature: Dimensionality reduction of unsupervised learning Can applied in deal huge amounts of sample The original data set is represented using a smaller set of prototype vectors not to find an optimal clustering but to get good

CLUSTERING  two main ways approaches hierarchical approaches  agglomerative algorithm : bottom-up strategies to build a hierarchical clustering tree  divisive algorithm : top-down strategies to build a hierarchical clustering tree partitive approaches  k-means  optimal clustering is a partitioning minimizes distances within maximizes distances between clusters

CLUSTERING(cont.)

SOM CLUSTERING  SOM training first to find the best matching unit (BMU) the prototype vectors are updated.

 The SOM algorithm characteristic : applicable to large data sets. The computational complexity scales linearly with the number of data samples it does not require huge amounts of memory that basically just the prototype vectors and the current training vector.

EXPERIMENTS  Tools: SOM_ToolBox 2.0 :  Data set: clown.datclown.dat Data set ( “ clown.data ” ) consisted of D samples.  cluster with three subclusters (right eye)  spherical cluster (left eye)  elliptical cluster (nose)  nonspherical cluster (U-shaped: mouth)  large and sparse cluster (body)  noise.(such as black x)

 Methods and Parameters: Cluster step 1:  Training Parameters of the SOM's  Map size: 19x17 Initial Neighborhood Widths: Rough Phases σ1(0): 10 Fine-Tuning Phases σ2(0): 2 learning rates:(The learning rate decreased linearly to zero during the training) Rough Phases : 0.5 Fine-Tuning Phases 0.05 Cluster step 2:  Method: K-Means Using 100 Runs

Experimental Results

single linkage dendrogram of 323 SOM Map unit

SOM Map average linkage dendrogram of 323 SOM Map unit

complete linkage dendrogram of 323 SOM Map unit

Conclusion