Data-intensive Computing Algorithms: Classification

Slides:



Advertisements
Similar presentations
Pat Langley Computational Learning Laboratory Center for the Study of Language and Information Stanford University, Stanford, California
Advertisements

COMPUTER AIDED DIAGNOSIS: CLASSIFICATION Prof. Yasser Mostafa Kadah –
Naïve-Bayes Classifiers Business Intelligence for Managers.
Data Mining Classification: Alternative Techniques
Machine learning continued Image source:
An Overview of Machine Learning
Ch5 Stochastic Methods Dr. Bernard Chen Ph.D. University of Central Arkansas Spring 2011.
CS292 Computational Vision and Language Pattern Recognition and Classification.
Learning from Observations Chapter 18 Section 1 – 4.
CS 590M Fall 2001: Security Issues in Data Mining Lecture 3: Classification.
Data-intensive Computing Algorithms: Classification Ref: Algorithms for the Intelligent Web 6/26/20151.
1 MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING By Kaan Tariman M.S. in Computer Science CSCI 8810 Course Project.
KNN, LVQ, SOM. Instance Based Learning K-Nearest Neighbor Algorithm (LVQ) Learning Vector Quantization (SOM) Self Organizing Maps.
Pattern Recognition. Introduction. Definitions.. Recognition process. Recognition process relates input signal to the stored concepts about the object.
ML ALGORITHMS. Algorithm Types Classification (supervised) Given -> A set of classified examples “instances” Produce -> A way of classifying new examples.
Chapter 5 Data mining : A Closer Look.
Introduction to machine learning
Digital Camera and Computer Vision Laboratory Department of Computer Science and Information Engineering National Taiwan University, Taipei, Taiwan, R.O.C.
Methods in Medical Image Analysis Statistics of Pattern Recognition: Classification and Clustering Some content provided by Milos Hauskrecht, University.
Data Mining Joyeeta Dutta-Moscato July 10, Wherever we have large amounts of data, we have the need for building systems capable of learning information.
Processing of large document collections Part 2 (Text categorization) Helena Ahonen-Myka Spring 2006.
Bayesian Networks. Male brain wiring Female brain wiring.
INTRODUCTION TO MACHINE LEARNING. $1,000,000 Machine Learning  Learn models from data  Three main types of learning :  Supervised learning  Unsupervised.
6/28/2014 CSE651C, B. Ramamurthy1.  Classification is placing things where they belong  Why? To learn from classification  To discover patterns  To.
Digital Camera and Computer Vision Laboratory Department of Computer Science and Information Engineering National Taiwan University, Taipei, Taiwan, R.O.C.
Machine Learning1 Machine Learning: Summary Greg Grudic CSCI-4830.
Data Mining Practical Machine Learning Tools and Techniques Chapter 4: Algorithms: The Basic Methods Section 4.6: Linear Models Rodney Nielsen Many of.
Today Ensemble Methods. Recap of the course. Classifier Fusion
Bayesian Classification. Bayesian Classification: Why? A statistical classifier: performs probabilistic prediction, i.e., predicts class membership probabilities.
Classification Techniques: Bayesian Classification
Pattern Recognition April 19, 2007 Suggested Reading: Horn Chapter 14.
Map of the Great Divide Basin, Wyoming, created using a neural network and used to find likely fossil beds See:
Digital Camera and Computer Vision Laboratory Department of Computer Science and Information Engineering National Taiwan University, Taipei, Taiwan, R.O.C.
1Ellen L. Walker Category Recognition Associating information extracted from images with categories (classes) of objects Requires prior knowledge about.
DATA MINING WITH CLUSTERING AND CLASSIFICATION Spring 2007, SJSU Benjamin Lam.
Chapter 20 Classification and Estimation Classification – Feature selection Good feature have four characteristics: –Discrimination. Features.
Data Mining and Decision Support
Learning Kernel Classifiers 1. Introduction Summarized by In-Hee Lee.
Naïve Bayes Classifier April 25 th, Classification Methods (1) Manual classification Used by Yahoo!, Looksmart, about.com, ODP Very accurate when.
BAYESIAN LEARNING. 2 Bayesian Classifiers Bayesian classifiers are statistical classifiers, and are based on Bayes theorem They can calculate the probability.
SUPERVISED AND UNSUPERVISED LEARNING Presentation by Ege Saygıner CENG 784.
Data-intensive Computing Algorithms: Classification Ref: Algorithms for the Intelligent Web 7/10/20161.
Rodney Nielsen Many of these slides were adapted from: I. H. Witten, E. Frank and M. A. Hall Data Science Algorithms: The Basic Methods Clustering WFH:
Business Intelligence and Decision Support Systems (9 th Ed., Prentice Hall) Chapter 6: Artificial Neural Networks for Data Mining.
Combining Models Foundations of Algorithms and Machine Learning (CS60020), IIT KGP, 2017: Indrajit Bhattacharya.
Data Mining Practical Machine Learning Tools and Techniques
Big data classification using neural network
Machine Learning with Spark MLlib
Naïve Bayes CSE651C, B. Ramamurthy 6/28/2014.
Semi-Supervised Clustering
Machine Learning – Classification David Fenyő
Table 1. Advantages and Disadvantages of Traditional DM/ML Methods
Map of the Great Divide Basin, Wyoming, created using a neural network and used to find likely fossil beds See:
Data Mining Lecture 11.
Self organizing networks
Machine Learning & Data Science
Classification Techniques: Bayesian Classification
Naïve Bayes CSE487/587 Spring /17/2018.
Naïve Bayes CSE651 6/7/2014.
Prepared by: Mahmoud Rafeek Al-Farra
Prepared by: Mahmoud Rafeek Al-Farra
Clustering.
Revision (Part II) Ke Chen
Computer Vision Chapter 4
MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING
MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING
The Naïve Bayes (NB) Classifier
Prepared by: Mahmoud Rafeek Al-Farra
Naïve Bayes CSE487/587 Spring2017 4/4/2019.
Chapter 4, Doing Data Science
Presentation transcript:

Data-intensive Computing Algorithms: Classification Ref: Algorithms for the Intelligent Web 5/10/2018

Goals Study important classification algorithms with the idea of transforming them into parallel algorithms exploiting MR, Pig and related Hadoop-based application suite. Classification is placing things where they belong To learn from classification To discover patterns It is a form of machine learning 5/10/2018

Learning approaches Goal: label the elements in a collection Supervised: train set, validation set, production set to be labeled. You model a classifier and use that that classifier to label your input Example: Recommender systems Unsupervised: unsupervised learning refers to the problem of trying to find hidden structure in unlabeled data. Ex: epidemiology: London Cholera clusters Semi-supervised: in between the above; some of the input data set is labeled. Ex: astronomy labeling items in an image 5/10/2018

Classification Classification relies on a priori reference structures that divide the space of all possible data points into a set of classes that are not overlapping. (what do you do the data points overlap?) The term ontology is typically used for a reference structure that constitutes a knowledge representation of the part of the world that is of interest in our application. What are the problems it (classification) can solve? What are some of the common classification methods? Which one is better for a given situation? (meta classifier) 5/10/2018

Classification examples in daily life Restaurant menu: appetizers, salads, soups, entrée, dessert, drinks,… Library of congress (LIC) system classifies books according to a standard scheme Injuries and diseases classification is physicians and healthcare workers Classification of all living things: eg., Home Sapiens (genus, species) 5/10/2018

An Ontology An ontology consists of three main things: concepts, attributes, and instances Classification maps an instance to a concept based on the attribute values. More the number of attributes, more finer the classification. This also leads to curse of dimensionality 5/10/2018

A rudimentary ontology 5/10/2018

Categories of classification algorithms With respect to underlying technique two broad categories: Statistical algorithms Regression for forecasting Bayes classifier depicts the dependency of the various attributes of the classification problem. Structural algorithms Rule-based algorithms: if-else, decision trees Distance-based algorithm: similarity, nearest neighbor Neural networks 5/10/2018

Classifiers 5/10/2018

Advantages and Disadvantages Decision tree, simple and powerful, works well for discrete (0,1- yes-no)rules; Neural net: black box approach, hard to interpret results Distance-based ones work well for low-dimensionality space .. 5/10/2018

Naïve Bayes Naïve Bayes classifier One of the most celebrated and well-known classification algorithms of all time. Probabilistic algorithm Typically applied and works well with the assumption of independent attributes, but also found to work well even with some dependencies. 5/10/2018

Naïve Bayes Example Reference: http://en.wikipedia.org/wiki/Bayes_Theorem Suppose there is a school with 60% boys and 40% girls as its students. The female students wear trousers or skirts in equal numbers; the boys all wear trousers. An observer sees a (random) student from a distance, and what the observer can see is that this student is wearing trousers. What is the probability this student is a girl? The correct answer can be computed using Bayes' theorem. The event A is that the student observed is a girl, and the event B is that the student observed is wearing trousers. To compute P(A|B), we first need to know: P(A), or the probability that the student is a girl regardless of any other information. Since the observer sees a random student, meaning that all students have the same probability of being observed, and the fraction of girls among the students is 40%, this probability equals 0.4. P(B|A), or the probability of the student wearing trousers given that the student is a girl. Since they are as likely to wear skirts as trousers, this is 0.5. P(B), or the probability of a (randomly selected) student wearing trousers regardless of any other information. Since half of the girls and all of the boys are wearing trousers, this is 0.5×0.4 + 1.0×0.6 = 0.8. Given all this information, the probability of the observer having spotted a girl given that the observed student is wearing trousers can be computed by substituting these values in the formula: P(A|B) = P(B|A)P(A)/P(B) = 0.5 * 0.4 / 0.8 = 0.25 5/10/2018

From the book machine learning in action 5/10/2018

Life Cycle of a classifier: training, testing and production 5/10/2018

Training Stage Provide classifier with data points for which we have already assigned an appropriate class. Purpose of this stage is to determine the parameters 5/10/2018

Validation Stage Testing or validation stage we validate the classifier to ensure credibility for the results. Primary goal of this stage is to determine the classification errors. Quality of the results should be evaluated using various metrics Training and testing stages may be repeated several times before a classifier transitions to the production stage. We could evaluate several types of classifiers and pick one or combine all classifiers into a metaclassifier scheme. 5/10/2018

Production stage The classifier(s) is used here in a live production system. It is possible to enhance the production results by allowing human-in-the-loop feedback. The three steps are repeated as we get more data from the production system. 5/10/2018

Unsupervised K-means Bayes We will discuss K-means with an example. Sample data set: people of all demographics 5/10/2018

Data Name Age Albert 23 Babis 21 Athena 24 Carl 30 Elena 38 Constantine 37 Catherine 31 Bob 32 Bill Charlie Aurora Alexandra 25 Maria 43 Dmitry 35 George 42 Eric Frank 39 Jack Lucas 45 John 5/10/2018