Download presentation

Presentation is loading. Please wait.

Published byBailey Parris Modified over 2 years ago

1
Data Mining in Clinical Databases by using Association Rules Department of Computing Charles Lo

2
Outline What is Association Rule ? Previous Works Target Problems Methodology and Algorithm Experiment and Discussion Q & A

3
What is Association Rule ? (1) It was introduced in “Agrawal, Imielinski, & Swami 1993”. Database A, BC 30% of the transactions that contain A and B also contain C, 5% of all the transactions contain all of them.

4
What is Association Rule (2) In a supermarket, 20% of transactions that contain Coke Cola also contain Pepsi, 3% of all transactions contain both items. –20% is the confidence of the rule –3% is the support of the rule Association rule can be applied in –Decision Support –Market Strategy –Financial Forecast

5
Related Work (1) In 1993, Agrawal, Imielinski and Swami Generate all significant association rules between items Algorithm Apriori –Pruning Techniques –Buffer management if support > min support if confidence min confidence Significant association rule

6
Related Work (2) Pruning Technique –Frequency Constraint Memory Management –Memory to store any itemset and all its 1-extensions

7
Related Work (3) In 1997, Srikant, Vu and Agrawal Consider constraints that are boolean expression over the presence or absence of items in the rules Incomplete candidate generation The boolean constraint: (B C) (X Y)

8
Related Work (4) Selected Items approaches 1. generate a set of selected items for B= (1 2) 3 2. only count candidates that contain selected items 3.Discard frequent itemsets that do not satisfy the boolean expression 1,32,3 1,2,3,4,5 any (non-empty) itemset that satisfies B will contain an item from this set

9
Related Work (5) In 1998, Ng, Lakshmanan, Han and Pang Achieved a maximized degree of pruning for different categories of constraints. Two critical properties to pruning –Anti-monotonicity –Succinctness Algorithm CAP 1. Both anti-monotone and succinct 2. Succinct but Non-anti-monotone 3.Anti-monotone and Non-Succinct 4.Non-anti-monotone and Non-succinct

10
Related Work (6) Anti-Monotone Constraint –S S’ & S satisfied C S’ satisfied C Domain ConstraintAggregate Constraint S = v, S v, S v,S V min(S) v, max(S) v, count(s) v, sum(s) v

11
Related Work (7) Succinct Constraint –pruning can be done once-and-for-all before any iteration take place S = v, S v, S v, S V min(S) v, min(S) v, max(S) v, max(S) v, count(s) v, sum(s) v Domain ConstraintAggregate Constraint S V

12
Target Problems (1) Association of quantitative items satisfy a given inequality constraint which are composed of either (+, -) or (*, /) –( I i1 I i2 ... I im ) ( I j1 I j2 ... I jn ) C 1. size m 2.size n 3. + ( * ) 4. - [ /] 5. (, =, ] 6.constant C –(3,2,+,-,>,100) –(1,1,0,/,=,2)

13
Target Problems (2) Temporal aspect of the data Hierarchies over the data A B C ABC AB CD Serial patternParallel patternSequence pattern

14
Problem Statement V= I 1 I 2,..., I M, a set of quantitative items T, the transactions of a database D t[k] > 0 means t contain item I k t[k] = 0 means I k does not exist Association of items which satisfy ( I i1 I i2 ... I im ) ( I j1 I j2 ... I jn ) C where is + ( * ), is - [ /], is (, =, ] and c is a scalar value

15
Application in Clinical Database Relationship between the treatments and clinical diagonsis –nursing : 100, clinical test : 30, pharmacies : 165,... –nursing : 120, injection : 130, pharmacies : 100,... –Operation : 220, injection : 542, clinical test : 60,... (X + Y ) - Z> 100 X / Y = 2

16
QMIC (1) QMIC (Quantitative Mining under Inequality Constraints) –Candidate generation reduce the number of itemsets Max_Min pruning –Support counting reduce the iteration of database scanning Generation sequence Memory requirement –limitation of the available memory

17
QMIC (2) Skip generation steps by the pre-defined size m and n Generation Steps –Algorithm Apriori : L k-1 L k –Algorithm QMIC : L K/2 L k

18
QMIC (3) Candidate itemsets generation

19
QMIC (4) why in this sequence ? –How about using 3, 4 or larger factor ? –Or even the power series ? Memory Management –keep the previous L’s to generate next level of large itemsets –Only limited memory is available –In QMIC, only three previous L’s are need in order to generate the next level of large itemsets in the generation sequence.

20
QMIC (5) What is the trade off of generation sequence ? –more number of candidate itemsets –longer process time in pruning Max_Min Pruning –involve the inequality constraint to the pruning –Maximum value itemset list (maxlst) Sorted list in a descending order according to the maximum value of sum (product) –Minimum value itemset list (minlst] Sorted list in an ascending order according to the minimum value of sum (product)

21
QMIC (6) Max_Pruning – = { , >} –A B C where A = ( I i1 I i2 ... I im ), B= ( I j1 I j2 ... I jn ) –Minimum value of A Over pruning ? –Using maxlst –Sliding Window with size m+ Window of maxlst 1 stop sliding if total sum of inside items is smaller than C

22
QMIC (7) Max_Pruning procedure

23
Experiments (1) Number of items

24
Experiments (2) Number of transactions

25
Future Plan Association Rules of Sequence Patterns –Time constraint Association Rules of Multi-layer data

Similar presentations

OK

2/8/00CSE 711 data mining: Apriori Algorithm by S. Cha 1 CSE 711 Seminar on Data Mining: Apriori Algorithm By Sung-Hyuk Cha.

2/8/00CSE 711 data mining: Apriori Algorithm by S. Cha 1 CSE 711 Seminar on Data Mining: Apriori Algorithm By Sung-Hyuk Cha.

© 2017 SlidePlayer.com Inc.

All rights reserved.

Ads by Google