TOPICS IN BUSINESS INTELLIGENCE K-NN & Naive Bayes – GROUP 1 Isabel van der Lijke Nathan Bok Gökhan Korkmaz.

Slides:



Advertisements
Similar presentations
Combining Classification and Model Trees for Handling Ordinal Problems D. Anyfantis, M. Karagiannopoulos S. B. Kotsiantis, P. E. Pintelas Educational Software.
Advertisements

Learning Algorithm Evaluation
Chapter 5 – Evaluating Classification & Predictive Performance
Chapter 4 – Evaluating Classification & Predictive Performance © Galit Shmueli and Peter Bruce 2008 Data Mining for Business Intelligence Shmueli, Patel.
Indian Statistical Institute Kolkata
Data preprocessing before classification In Kennedy et al.: “Solving data mining problems”
Chapter 7 – Classification and Regression Trees
Chapter 7 – Classification and Regression Trees
Introduction to Data Mining with XLMiner
x – independent variable (input)
Chapter 7 – K-Nearest-Neighbor
1 Application of Metamorphic Testing to Supervised Classifiers Xiaoyuan Xie, Tsong Yueh Chen Swinburne University of Technology Christian Murphy, Gail.
CS 590M Fall 2001: Security Issues in Data Mining Lecture 3: Classification.
ROC Curves.
5/30/2006EE 148, Spring Visual Categorization with Bags of Keypoints Gabriella Csurka Christopher R. Dance Lixin Fan Jutta Willamowski Cedric Bray.
Data Mining: A Closer Look Chapter Data Mining Strategies (p35) Moh!
z-Scores What is a z-Score? How Are z-Scores Useful? Distributions of z-Scores Standard Normal Curve.
Bayesian Classification with a brief introduction to pattern recognition Modified from slides by Michael L. Raymer, Ph.D.
On Comparing Classifiers: Pitfalls to Avoid and Recommended Approach Published by Steven L. Salzberg Presented by Prakash Tilwani MACS 598 April 25 th.
Chapter 5 Data mining : A Closer Look.
CSCI 347 / CS 4206: Data Mining Module 06: Evaluation Topic 07: Cost-Sensitive Measures.
Evaluating Classifiers
Overview DM for Business Intelligence.
Evaluation – next steps
MULTICLASS CONTINUED AND RANKING David Kauchak CS 451 – Fall 2013.
by B. Zadrozny and C. Elkan
ROC 1.Medical decision making 2.Machine learning 3.Data mining research communities A technique for visualizing, organizing, selecting classifiers based.
Chi-square Test of Independence Steps in Testing Chi-square Test of Independence Hypotheses.
Chapter 9 – Classification and Regression Trees
Chapter 12 – Discriminant Analysis © Galit Shmueli and Peter Bruce 2010 Data Mining for Business Intelligence Shmueli, Patel & Bruce.
GA-Based Feature Selection and Parameter Optimization for Support Vector Machine Cheng-Lung Huang, Chieh-Jen Wang Expert Systems with Applications, Volume.
K Nearest Neighbors Saed Sayad 1www.ismartsoft.com.
XLMiner – a Data Mining Toolkit QuantLink Solutions Pvt. Ltd.
Chapter 4: Pattern Recognition. Classification is a process that assigns a label to an object according to some representation of the object’s properties.
MKT 700 Business Intelligence and Decision Models Algorithms and Customer Profiling (1)
Evaluating Results of Learning Blaž Zupan
Chapter 6 – Three Simple Classification Methods © Galit Shmueli and Peter Bruce 2008 Data Mining for Business Intelligence Shmueli, Patel & Bruce.
COMP 2208 Dr. Long Tran-Thanh University of Southampton K-Nearest Neighbour.
Feature Extraction Artificial Intelligence Research Laboratory Bioinformatics and Computational Biology Program Computational Intelligence, Learning, and.
Evaluating Classification Performance
Discriminant Analysis Dr. Satyendra Singh Professor and Director University of Winnipeg, Canada
Bayesian decision theory: A framework for making decisions when uncertainty exit 1 Lecture Notes for E Alpaydın 2010 Introduction to Machine Learning 2e.
Validation methods.
Typically, classifiers are trained based on local features of each site in the training set of protein sequences. Thus no global sequence information is.
Chapter 5: Credibility. Introduction Performance on the training set is not a good indicator of performance on an independent set. We need to predict.
A Brief Introduction and Issues on the Classification Problem Jin Mao Postdoc, School of Information, University of Arizona Sept 18, 2015.
Competition II: Springleaf Sha Li (Team leader) Xiaoyan Chong, Minglu Ma, Yue Wang CAMCOS Fall 2015 San Jose State University.
Evaluating Classifiers Reading: T. Fawcett, An introduction to ROC analysis, Sections 1-4, 7 (linked from class website)An introduction to ROC analysis.
Chapter 5 – Evaluating Predictive Performance Data Mining for Business Analytics Shmueli, Patel & Bruce.
Chapter 8 – Naïve Bayes DM for Business Intelligence.
Clustering Algorithms Minimize distance But to Centers of Groups.
Evaluating Classifiers. Reading for this topic: T. Fawcett, An introduction to ROC analysis, Sections 1-4, 7 (linked from class website)
Describing a Score’s Position within a Distribution Lesson 5.
2011 Data Mining Industrial & Information Systems Engineering Pilsung Kang Industrial & Information Systems Engineering Seoul National University of Science.
SUPERVISED AND UNSUPERVISED LEARNING Presentation by Ege Saygıner CENG 784.
Chapter 12 – Discriminant Analysis
Data Mining, Machine Learning, Data Analysis, etc. scikit-learn
Evolving Decision Rules (EDR)
Evaluating Classifiers
Table 1. Advantages and Disadvantages of Traditional DM/ML Methods
Evaluating Results of Learning
Performance Measures II
Chapter 7 – K-Nearest-Neighbor
Features & Decision regions
SEG 4630 E-Commerce Data Mining — Final Review —
TED Talks – A Predictive Analysis Using Classification Algorithms
Classification of highly unbalanced data using deep learning techniques
CSCI N317 Computation for Scientific Applications Unit Weka
Model Evaluation and Selection
Data Mining, Machine Learning, Data Analysis, etc. scikit-learn
Presentation transcript:

TOPICS IN BUSINESS INTELLIGENCE K-NN & Naive Bayes – GROUP 1 Isabel van der Lijke Nathan Bok Gökhan Korkmaz

INTRODUCTION K-NN  k-NN Classifier (Categorical Outcome)  Determining Neighbors  Classification Rule  Example: Riding Mowers  Choosing k  Setting the Cutoff Value  Advantages and shortcomings of k-NN algorithms 2

INTRODUCTION NAIVE BAYES  Basic Classification Procedure  Cutoff Probability Method  Conditional Probability  Naive Bayes  Advantages and shortcomings of the naive Bayes classifier 3

SIMPLE CASE APPLICATION  Depression 4

SIMPLE CASE APPLICATION  Fruits Example: P(Banana) = 500 / 1000 = 0,5 1-0,5 = 0,5 (Not banana) New fruit  compute all the chances 5 Sweet Not sweet Total Banana Orange Other fruit Total

REAL-LIFE APPLICATION NAIVE BAYES  Medical Data Classification with Naive Bayes Approach  Introduction  Requirements for systems dealing with medical data  An empirical comparison  Tables  Conclusion 6

TABLE 2:COMPARATIVE ANALYSIS BASED ON PREDICTIVE ACCURACY 7

TABLE 3:COMPARATIVE ANALYSIS BASED ON AREA UNDER ROC CURVE (AUC) 8

REAL-LIFE APPLICATION K-NN  Used to help health care professionals in diagnosing heart disease.  Useful for pattern recognition and classification.  Euclidean distance:  Often normalized data due to different variable formats. 9

CASE STUDY  “Our customer is a Dutch charity organization that wants to be able to classify it's supporters to donators and non-donators. The non-donators are sent a single marketing mail a year, whereas the donators receive multiple ones (up to 4).”  Who are the donators?  Who are the non-donators?  Application of K-NN & Naive Bayes to training and test dataset.  4000 customers.  SPSS, Excel, XLMiner 10

CLEAN-UP  No missing values  1-dimensional outliers removed through sorting (regarding annual & average donation)  2-dimensional outliers removed through scatterplot 11

12

Variables Kept Average donation Frequency of Response Median Time of Response Time as client Variables removed Annual donation Last donation Time since last response. 13

 Normalization of scores into z-scores.  Nominal categorization of data  Classification through percentiles of z-score & by manually processing values within the variables. 14

ANALYSIS OF CASE STUDY – K-NN 15  Xlminer  Partition data  Models created:  M1 = Zavgdon & Zfrqres  M2 = ZtimeCl, Zfrqres & Zavgdon  M3 = Zmedtor, Zfrqres & Zavgdon  ZtimeCl, Zfrqres, Zmedtor & Zavgdon

 Validation Data Scoring - Summary Report (for k = 13) 16 Error Report Class# Cases# Errors% Error , , Overall , Classification Confusion Matrix Predicted Class Actual Class

CHOOSING MODEL FOR K-NN  Accuracy: Proportion of correctly classified instances.  Error rate: (1 – Accuracy)  Sensitivity: Sensitivity is the proportion of actual positives which are correctly identified as positives by the classifier.  Specificity: Like sensitivity, but for the negatives. 17

18

M1M2 Selecting everyone in validation data €711.20€ Selecting while correcting for sensitivity and specificity €583.60€

APPLICATION OF MODEL ON TEST DATA Classification Confusion Matrix Predicted Class Actual Class Error Report Class# Cases# Errors% Error , ,5812 Overall ,65415

21

ANALYSIS OF THE CASE STUDY – NAIVE BAYES 22 Classification Confusion Matrix Predicted Class Actual Class Error Report Class# Cases# Errors% Error , ,13483 Overall ,042  M1 = Cfrqres & Cavgdon  M2 = Cfrqresp, Cavgdon & Cmedtor Classes --> Input Variab les 01 ValueProbValueProb CFRQR ES 10, , , , , , , , CAVGD ON 10, , , , , , , ,066362

23 Model 1Model 2 Selecting everyone€1072€1006 Selecting by class€2460,82€

APPLICATION OF MODEL ON TEST DATA Classes --> Input Variabl es 01 ValueProbValueProb CFRQR ES 10, , , , , , , , CAVGD ON 10, , , , , , , ,

25 Classification Confusion Matrix Predicted Class Actual Class Error Report Class# Cases# Errors% Error , ,59829 Overall ,61858

LOOKING AT BOTH MODELS 26

27

QUESTIONS? 28