Revision (Part II) Ke Chen COMP24111 Machine Learning Revision slides are going to summarise all you have learnt from Part II, which should be helpful.

Slides:



Advertisements
Similar presentations
K-means Clustering Ke Chen.
Advertisements

Albert Gatt Corpora and Statistical Methods Lecture 13.
Clustering Clustering of data is a method by which large sets of data is grouped into clusters of smaller sets of similar data. The example below demonstrates.
What we will cover here What is a classifier
Naïve Bayes Classifier
Data Mining Techniques: Clustering
On Discriminative vs. Generative classifiers: Naïve Bayes
Introduction to Bioinformatics
Lecture 17: Supervised Learning Recap Machine Learning April 6, 2010.
Clustering II.
Clustering… in General In vector space, clusters are vectors found within  of a cluster vector, with different techniques for determining the cluster.
Dimension reduction : PCA and Clustering Agnieszka S. Juncker Slides: Christopher Workman and Agnieszka S. Juncker Center for Biological Sequence Analysis.
Today Unsupervised Learning Clustering K-means. EE3J2 Data Mining Lecture 18 K-means and Agglomerative Algorithms Ali Al-Shahib.
Slide 1 EE3J2 Data Mining Lecture 16 Unsupervised Learning Ali Al-Shahib.
Unsupervised Learning and Data Mining
Clustering. 2 Outline  Introduction  K-means clustering  Hierarchical clustering: COBWEB.
Semi-Supervised Clustering Jieping Ye Department of Computer Science and Engineering Arizona State University
Ranking by Odds Ratio A Probability Model Approach let be a Boolean random variable: document d is relevant to query q otherwise Consider document d as.
Clustering Unsupervised learning Generating “classes”
Naïve Bayes Classifier Ke Chen Extended by Longin Jan Latecki COMP20411 Machine Learning.
Methods in Medical Image Analysis Statistics of Pattern Recognition: Classification and Clustering Some content provided by Milos Hauskrecht, University.
CSE 185 Introduction to Computer Vision Pattern Recognition.
Data mining and machine learning A brief introduction.
Naïve Bayes Classifier Ke Chen Modified and extended by Longin Jan Latecki
Hierarchical Clustering
Ch10 Machine Learning: Symbol-Based
Text Clustering.
Clustering Supervised vs. Unsupervised Learning Examples of clustering in Web IR Characteristics of clustering Clustering algorithms Cluster Labeling 1.
Basic Machine Learning: Clustering CS 315 – Web Search and Data Mining 1.
Data Mining Practical Machine Learning Tools and Techniques Chapter 4: Algorithms: The Basic Methods Section 4.6: Linear Models Rodney Nielsen Many of.
Naïve Bayes Classifier Ke Chen Modified and extended by Longin Jan Latecki
Dimension reduction : PCA and Clustering Slides by Agnieszka Juncker and Chris Workman modified by Hanne Jarmer.
Artificial Intelligence 8. Supervised and unsupervised learning Japan Advanced Institute of Science and Technology (JAIST) Yoshimasa Tsuruoka.
Clustering.
Prepared by: Mahmoud Rafeek Al-Farra
K-Means Algorithm Each cluster is represented by the mean value of the objects in the cluster Input: set of objects (n), no of clusters (k) Output:
Data Mining Practical Machine Learning Tools and Techniques Chapter 4: Algorithms: The Basic Methods Section 4.8: Clustering Rodney Nielsen Many of these.
INTRODUCTION TO MACHINE LEARNING 3RD EDITION ETHEM ALPAYDIN © The MIT Press, Lecture.
Mehdi Ghayoumi MSB rm 132 Ofc hr: Thur, a Machine Learning.
CS 8751 ML & KDDData Clustering1 Clustering Unsupervised learning Generating “classes” Distance/similarity measures Agglomerative methods Divisive methods.
Radial Basis Function ANN, an alternative to back propagation, uses clustering of examples in the training set.
Machine Learning Queens College Lecture 7: Clustering.
Slide 1 EE3J2 Data Mining Lecture 18 K-means and Agglomerative Algorithms.
Basic Machine Learning: Clustering CS 315 – Web Search and Data Mining 1.
1 Machine Learning Lecture 9: Clustering Moshe Koppel Slides adapted from Raymond J. Mooney.
COMP24111 Machine Learning Naïve Bayes Classifier Ke Chen.
Information Retrieval Search Engine Technology (8) Prof. Dragomir R. Radev.
Introduction to Data Mining Clustering & Classification Reference: Tan et al: Introduction to data mining. Some slides are adopted from Tan et al.
Data Mining and Text Mining. The Standard Data Mining process.
COMP24111 Machine Learning Naïve Bayes Classifier Ke Chen.
COMP24111 Machine Learning K-means Clustering Ke Chen.
Rodney Nielsen Many of these slides were adapted from: I. H. Witten, E. Frank and M. A. Hall Data Science Algorithms: The Basic Methods Clustering WFH:
Semi-Supervised Clustering
Machine Learning Clustering: K-means Supervised Learning
Hierarchical Clustering
Ke Chen Reading: [7.3, EA], [9.1, CMB]
Constrained Clustering -Semi Supervised Clustering-
Data Mining K-means Algorithm
K-means and Hierarchical Clustering
Clustering.
John Nicholas Owen Sarah Smith
Hierarchical and Ensemble Clustering
Revision (Part II) Ke Chen
Ke Chen Reading: [7.3, EA], [9.1, CMB]
Data Mining 資料探勘 分群分析 (Cluster Analysis) Min-Yuh Day 戴敏育
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Revision (Part II) Ke Chen
Naïve Bayes Classifier
Hierarchical and Ensemble Clustering
Text Categorization Berlin Chen 2003 Reference:
Presentation transcript:

Revision (Part II) Ke Chen COMP24111 Machine Learning Revision slides are going to summarise all you have learnt from Part II, which should be helpful for you to prepare your exam in January along with those non-assessed exercises.

COMP24111 Machine Learning 2 Naïve Bayes Classifier Probabilistic Classifiers –discriminative vs. generative classifiers –Bayesian rule used to convert generative to discriminative Naïve Bayesian Assumption –Conditionally independent assumption on input attributes Naïve Bayes classification algorithm –Estimate conditional probabilities for each attribute given a class label and prior probabilities for each class label –MAP decision rule Relevant issues –Zero conditional probability due to short of training examples –Applicability to problems violating the naïve Bayes assumption

COMP24111 Machine Learning 3 Clustering Analysis Basics Clustering Analysis Task –discover the “natural” clustering number –properly grouping objects into sensible clusters Data type and representation –Data type: continuous vs. discrete (binary, ranking, …) –Data matrix and distance matrix Distance Measure –Minkowski distance (Manhattan, Euclidean …) for continuous –Cosine measure for nonmetric –Distance for binary: contingency table, symmetric vs. asymmetric Major Clustering Approach –Partitioning, hierarchical, density-based, graph-based, ensemble

COMP24111 Machine Learning 4 K-Means Clustering Principle –A typical partitioning clustering approach with an iterative process to minimise the square distance in each cluster K-means algorithm 1)Initialisation: choose K centroids (seed points) 2)Assign each data object to the cluster whose centroid is nearest 3)Re-calculate the mean for each cluster to get a updated centroid 4)Repeat 2) and 3) until no new assignment Relevant issues –Efficiency: O(tkn) where t, k <<n –Sensitive to initialisation and converge to local optimum –Other weakness and limitations –Clustering validation

COMP24111 Machine Learning 5 Hierarchical Clustering Hierarchical clustering –Principle: partitioning data set sequentially –Strategy: divisive (top-down) vs. agglomerative (bottom-up) Cluster Distance –Single-link, complete-link and averaging-link Agglomerative algorithm 1)Convert object attributes to distance matrix 2)Repeat until number of cluster is one o Merge two closest clusters o Update distance matrix with cluster distance Relevant concepts and techniques –Construct a dendrogram tree –Life-time of clusters achieved from a dendrogram tree –Determine the number of clusters with maximum k life-time

COMP24111 Machine Learning 6 Cluster Validation –Evaluate the results of clustering in a quantitative and objective fashion –Performance evaluation, clustering comparison, find cluster num. Two different types of cluster validation methods –Internal indexes No ground truth available and sometimes named “relative index” Defined based on “common sense” or “a priori knowledge” Variance-based validity indexes Application: finding the “proper” number of clusters, … –External indexes ground truth known or reference given Rand Index Application: performance evaluation of clustering, clustering comparison... –There are many validity indexes, still an active research area in unsupervised learning

COMP24111 Machine Learning 7 Examination Information Three Sections (total 50 marks) –Section 1 (20 marks) o 20 multiple choice questions totally o Questions relevant to Part II –Section 2 (15 marks) o One compulsory question relevant to Part I –Section 3 (15 marks) o One compulsory question relevant to Part II Length: two hours Calculator (without memory) allowed