1 Knowledge Discovery Transparencies prepared by Ho Tu Bao [JAIST] ITCS 6162.

Slides:



Advertisements
Similar presentations
COMP3740 CR32: Knowledge Management and Adaptive Systems
Advertisements

Data Mining Techniques Association Rule
1 Machine Learning: Lecture 3 Decision Tree Learning (Based on Chapter 3 of Mitchell T.., Machine Learning, 1997)
Decision Trees Decision tree representation ID3 learning algorithm
Decision Tree Approach in Data Mining
Data Mining Classification: Basic Concepts, Decision Trees, and Model Evaluation Lecture Notes for Chapter 4 Part I Introduction to Data Mining by Tan,
ICS320-Foundations of Adaptive and Learning Systems
Deriving rules from data Decision Trees a.j.m.m (ton) weijters.
Classification Techniques: Decision Tree Learning
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
1 Lecture 5: Automatic cluster detection Lecture 6: Artificial neural networks Lecture 7: Evaluation of discovered knowledge Brief introduction to lectures.
CS 590M Fall 2001: Security Issues in Data Mining Lecture 4: ID3.
Tree-based methods, neutral networks
Classification Continued
Data Mining with Decision Trees Lutz Hamel Dept. of Computer Science and Statistics University of Rhode Island.
INTRODUCTION TO Machine Learning ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Learning….in a rather broad sense: improvement of performance on the basis of experience Machine learning…… improve for task T with respect to performance.
Classification.
Data Mining – Intro.
Machine Learning Lecture 10 Decision Trees G53MLE Machine Learning Dr Guoping Qiu1.
Data Mining : Introduction Chapter 1. 2 Index 1. What is Data Mining? 2. Data Mining Functionalities 1. Characterization and Discrimination 2. MIning.
Kansas State University Department of Computing and Information Sciences CIS 830: Advanced Topics in Artificial Intelligence From Data Mining To Knowledge.
DATA MINING : CLASSIFICATION. Classification : Definition  Classification is a supervised learning.  Uses training sets which has correct answers (class.
Fall 2004 TDIDT Learning CS478 - Machine Learning.
Machine Learning Chapter 3. Decision Tree Learning
Short Introduction to Machine Learning Instructor: Rada Mihalcea.
2 Outline of the presentation Objectives, Prerequisite and Content Brief Introduction to Lectures Discussion and Conclusion Objectives, Prerequisite and.
Inductive learning Simplest form: learn a function from examples
Mohammad Ali Keyvanrad
1 SHIM 413 Database Applications for Healthcare Fall 2006 Slides by H. T. Bao.
Introduction to machine learning and data mining 1 iCSC2014, Juan López González, University of Oviedo Introduction to machine learning Juan López González.
Lecture 7. Outline 1. Overview of Classification and Decision Tree 2. Algorithm to build Decision Tree 3. Formula to measure information 4. Weka, data.
Basic Data Mining Technique
Decision-Tree Induction & Decision-Rule Induction
Copyright © 2004 Pearson Education, Inc.. Chapter 27 Data Mining Concepts.
Data Mining – Intro. Course Overview Spatial Databases Temporal and Spatio-Temporal Databases Multimedia Databases Data Mining.
CS690L Data Mining: Classification
Chapter 6 Classification and Prediction Dr. Bernard Chen Ph.D. University of Central Arkansas.
CS 5751 Machine Learning Chapter 3 Decision Tree Learning1 Decision Trees Decision tree representation ID3 learning algorithm Entropy, Information gain.
Decision Trees Binary output – easily extendible to multiple output classes. Takes a set of attributes for a given situation or object and outputs a yes/no.
CS Data Mining1 Data Mining The Extraction of useful information from data The automated extraction of hidden predictive information from (large)
Data Mining: Knowledge Discovery in Databases Peter van der Putten ALP Group, LIACS Pre-University College Bio Informatics January
Decision Tree Learning
An Introduction Student Name: Riaz Ahmad Program: MSIT( ) Subject: Data warehouse & Data Mining.
Decision Tree Learning Presented by Ping Zhang Nov. 26th, 2007.
Data Mining and Decision Support
1 Classification: predicts categorical class labels (discrete or nominal) classifies data (constructs a model) based on the training set and the values.
Data Mining By Farzana Forhad CS 157B. Agenda Decision Tree and ID3 Rough Set Theory Clustering.
WHAT IS DATA MINING?  The process of automatically extracting useful information from large amounts of data.  Uses traditional data analysis techniques.
Outline Decision tree representation ID3 learning algorithm Entropy, Information gain Issues in decision tree learning 2.
WHAT IS DATA MINING?  The process of automatically extracting useful information from large amounts of data.  Uses traditional data analysis techniques.
Friday’s Deliverable As a GROUP, you need to bring 2N+1 copies of your “initial submission” –This paper should be a complete version of your paper – something.
Dr. Chen, Data Mining  A/W & Dr. Chen, Data Mining Chapter 3 Basic Data Mining Techniques Jason C. H. Chen, Ph.D. Professor of MIS School of Business.
Chapter 3 Data Mining: Classification & Association Chapter 4 in the text box Section: 4.3 (4.3.1),
DATA MINING TECHNIQUES (DECISION TREES ) Presented by: Shweta Ghate MIT College OF Engineering.
Decision Tree Learning DA514 - Lecture Slides 2 Modified and expanded from: E. Alpaydin-ML (chapter 9) T. Mitchell-ML.
Review of Decision Tree Learning Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.
CSE573 Autumn /11/98 Machine Learning Administrative –Finish this topic –The rest of the time is yours –Final exam Tuesday, Mar. 17, 2:30-4:20.
Data Mining – Intro.
Prepared by: Mahmoud Rafeek Al-Farra
Decision Tree Saed Sayad 9/21/2018.
Prepared by: Mahmoud Rafeek Al-Farra
Machine Learning Chapter 3. Decision Tree Learning
Machine Learning: Lecture 3
Classification and Prediction
Machine Learning Chapter 3. Decision Tree Learning
CSCI N317 Computation for Scientific Applications Unit Weka
A task of induction to find patterns
A task of induction to find patterns
Presentation transcript:

1 Knowledge Discovery Transparencies prepared by Ho Tu Bao [JAIST] ITCS 6162

2 “Are there clusters of similar cells?” Light color with 1 nucleus Dark color with 2 tails 2 nuclei 1 nucleus and 1 tail Dark color with 1 tail and 2 nuclei Clustering

3 Task: Discovering association rules among items in a transaction database. An association among two items A and B means that the presence of A in a record implies the presence of B in the same record: A => B. In general: A 1, A 2, … => B Association Rule Discovery Association Rule Discovery

4 “Are there any associations between the characteristics of the cells?” If color = light and # nuclei = 1 then # tails = 1 (support = 12.5%; confidence = 50%) If # nuclei = 2 and Cell = Cancerous then # tails = 2 (support = 25%; confidence = 100%) If # tails = 1 then Color = light (support = 37.5%; confidence = 75%) Association Rule Discovery Association Rule Discovery

5 Genetic Algorithms Statistics Bayesian Networks Rough Sets Time Series Many Other Data Mining Techniques Text Mining

6 Lecture 1: Overview of KDD 1. What is KDD and Why ? 2. The KDD Process 4. Data Mining Methods 3. KDD Applications 5. Challenges for KDD

7 Challenges and Influential Aspects Handling of different types of data with different degree of supervision Changing data and knowledge Understandability of patterns, various kinds of requests and results (decision lists, inference networks, concept hierarchies, etc.) Interactive, Visualization Knowledge Discovery Different sources of data (distributed, heterogeneous databases, noise and missing, irrelevant data, etc. ) Massive data sets, high dimensionality (efficiency, scalability)

8 High dimensionality increases exponentially the size of the space of patterns. Massive Data Sets and High Dimensionality With p attributes each has d discrete values in average, the space of possible instances has the size of d p. Classes: {Cancerous, Healthy} Attributes: Cell body: {dark, light} #nuclei: {1, 2} #tails: {1, 2} Healthy Cell C1 C2 C3 h1 h2 Cancerous Cell (# instances = 2 3 = 8) 38 attributes, each 10 values # instances = # attributes ?

9 Attribute Numerical Symbolic No structure Places, Color Ordinal structure Ring structure Rank, Resemblance Age, Temperature, Taste, Income, Length Nominal (categorical) Ordinal Measurable Different Types of Data Combinatorical search in hypothesis spaces (machine learning) Often matrix-based computation (multivariate data analysis)

10 Lecture 1: Overview of KDD Lecture 2: Preparing data Lecture 3: Decision tree induction Lecture 4: Mining association rules Lecture 5: Automatic cluster detection Lecture 6: Artificial neural networks Lecture 7: Evaluation of discovered knowledge Brief introduction to lectures

11 Lecture 2: Preparing Data As much as 80% of KDD is about preparing data, and the remaining 20% is about mining Content of the lecture 1. Data cleaning 2. Data transformations 3. Data reduction 4. Software and case-studies Prerequisite: Nothing special but expected some understanding of statistics

12 Data Preparation The design and organization of data, including the setting of goals and the composition of features, is done by humans. There are two central goals for the preparation of data: To organize data into a standard form that is ready for processing by data mining programs. To prepare features that lead to the best data mining performance.

13 Lecture 1: Overview of KDD Lecture 2: Preparing data Lecture 3: Decision tree induction Lecture 4: Mining association rules Lecture 5: Automatic cluster detection Lecture 6: Artificial neural networks Lecture 7: Evaluation of discovered knowledge Brief introduction to lectures

14 Lecture 3: Decision Tree Induction One of the most widely used KDD classification techniques for supervised data. Content of the lecture 1. Decision tree representation and framework 2. Attribute selection 3. Pruning trees 4. Software C4.5, CABRO and case-studies Prerequisite: Nothing special

15 Decision Trees Decision Tree is a classifier in the form of a tree structure that is either: A decision tree can be used to classify an instance by starting at the root of the tree and moving through it until a leaf is met. a leaf node, indicating a class of instances, or a decision node that specifies some test to be carried out on a single attribute value, with one branch and subtree for each possible outcome of the test

16 General Framework of Decision Tree Induction 1. Choose the “best” attribute by a given selection measure 2. Extend tree by adding new branch for each attribute value 3. Sorting training examples to leaf nodes 4. If examples unambiguously classified Then Stop Else Repeat steps 1-4 for leaf nodes Headache Temperature Flu e1 yes normal no e2 yes high yes e3 yes very high yes e4 no normal no e5 no high no e6 no very high no Temperature yes Headache normal high very high Headache no yesno yes {e2} no {e5} yes {e3} no {e6} {e1, e4} {e2, e5} {e3,e6} 5. Pruning unstable leaf nodes

17 Some Attribute Selection Measures

18 Avoiding Overfitting How can we avoid overfitting? Stop growing when data split not statistically significant (pre-pruning) Grow full tree, then post-prune (post-pruning) Two post-pruning techniques Reduce-Error Pruning Rule Post-Pruning

19 Converting A Tree to Rules sunny o’cast rain outlook high normal humidity noyes true false wind noyes IF (Outlook = Sunny) and (Humidity = High) THEN PlayTennis = No IF (Outlook = Sunny) and (Humidity = Normal) THEN PlayTennis = Yes

20 Matching path and the class that matches the unknown instance Discovered decision tree Unknown case CABRO: Decision Tree Induction CABRO based on R-measure, a measure for attribute dependency stemmed from rough sets theory.

21 Lecture 1: Overview of KDD Lecture 2: Preparing data Lecture 3: Decision tree induction Lecture 4: Mining association rules Lecture 5: Automatic cluster detection Lecture 6: Artificial neural networks Lecture 7: Evaluation of discovered knowledge Brief introduction to lectures

22 Lecture 4: Mining Association Rules A new technique and attractive topic. It allows discovering the important associations among items of transactions in a database. Content of the lecture Prerequisite: Nothing special 3. The Basket Analysis Program 1. Basic Definitions 2. Algorithm Apriori

23 There are two measures of rule strength: Support of (A => B) = [AB] / N, where N is the number of records in the database. The support of a rule is the proportion of times the rule applies. Confidence of (A => B) = [AB] / [A] The confidence of a rule is the proportion of times the rule is correct. Measures of Association Rules Measures of Association Rules

24 The task of mining association rules is mainly to discover strong association rules (high confidence and strong support) in large databases. Algorithm Apriori TID Items 1000 A, B, C 2000 A, C 3000 A, D 4000 B, E, F {A} 75% {B} 50% {C} 50% {A,C} 50% S = 40% Large support items Mining association rules is composed of two steps: 1. discover the large items, i.e., the sets of itemsets that have transaction support above a predetermined minimum support s. 2. Use the large itemsets to generate the association rules

25 Algorithm Apriori: Illustration TID Items 100 A, C, D 200 B, C, E 300 A, B, C, E 400 B, E Database D {A} {B} {C} {D} {E} Itemset Count {A} 2 {B} 3 {C} 3 {E} 3 Itemset Count {A, B} {A, C} {A, E} {B, C} {B, E} {C, E} Itemset {A,B} {A,C} {A,E} {B,C} {B,E} {C,E} Itemset Count {A, C} 2 {B, C} 2 {B, E} 3 {C, E} 2 Itemset Count {B, C, E} Itemset {B, C, E} 2 Itemset Count {B, C, E} 2 Itemset Count C1 L1 C2 L2 C2 C3 L3C3 Scan D Scan D Scan D S = 40%