Presentation is loading. Please wait.

Presentation is loading. Please wait.

Authors: Yutaka Matsuo & Mitsuru Ishizuka Designed by CProDM Team.

Similar presentations


Presentation on theme: "Authors: Yutaka Matsuo & Mitsuru Ishizuka Designed by CProDM Team."— Presentation transcript:

1 Authors: Yutaka Matsuo & Mitsuru Ishizuka Designed by CProDM Team

2 Introduction Algorithm implement Evaluation Outline Study Algorithm

3 Introduction Discard stop words Stem Extract frequency Select frequent term Clustering Expected probability Calculate X’ 2 value Output

4 Study Algorithm Preprocessing Goal: - Remove unnecessary words in document. - Get terms which are candidate keywords. Stop word: the function words and, the, and of, or other words with minimal lexical meaning. Stem: remove suffixes from words

5 Discard stop words It might be urged that when playing the “imitation game" the best strategy for the machine may possibly be something other than imitation of the behaviour of a man. This may be, but I think it is unlikely that there is any great effect of this kind. In any case there is no intention to investigate here the theory of the 2 game, and it will be assumed that the best strategy is to try to provide answers that would naturally be given by a man. urged playing “imitation game" best strategy machine possibly imitation behaviour man think unlikely great effect kind. case intention investigate theory 2 game, assumed best strategy try provide answers naturally given man. Study Algorithm Preprocessing

6 urged playing “imitation game" best strategy machine possibly imitation behaviour man think unlikely great effect kind. case intention investigate theory game, assumed best strategy try provide answers naturally given man. Stem urge play “imitation game" best strategi machine possible imitation behaviour man think unlike great effect kind. case intention investigate theory game, assum best strategi try provide answers natural give man. Study Algorithm Preprocessing

7 imitation best strategi man best strategi Extract frequency urge play “imitation game" best strategi machine possible imitation behaviour man think unlike great effect kind. case intention investigate theory game, assum best strategi try provide answers natural give man. Study Algorithm Preprocessing

8 Study Algorithm Term Co-occurrence and Importance the top ten frequent terms (denoted as ) and the probability of occurrence, normalized so that the sum is to be 1

9 Study Algorithm Term Co-occurrence and Importance Two terms in a sentence are considered to co-occur once.

10 co-occurrence probability distribution of some terms and the frequent terms. Study Algorithm Term Co-occurrence and Importance

11 The statistical value of χ2 is defined as P g Unconditional probability of a frequent term g ∈ G (the expected probability) N w The total number of co-occurrence of term w and frequent terms G freq (w, g) Frequency of co-occurrence of term w and term g Study Algorithm Term Co-occurrence and Importance

12 Study Algorithm Term Co-occurrence and Importance

13 Study Algorithm Algorithm improvement P g (the sum of the total number of terms in sentences where g appears) divided by (the total number of terms in the document) N w The total number of terms in the sentences where w appears including w If a term appears in a long sentence, it is likely to co-occur with many terms; if a term appears in a short sentence, it is less likely to co-occur with other terms. We consider the length of each sentence and revise our definitions

14 the following function to measure robustness of bias values Study Algorithm Algorithm improvement

15 To improve extracted keyword quality, we will cluster terms Two major approaches (Hofmann & Puzicha 1998) are:  Similarity-based clustering If terms w1 and w2 have similar distribution of co- occurrence with other terms, w1 and w2 are considered to be the same cluster.  Pairwise clustering If terms w1 and w2 co-occur frequently, w1 and w2 are considered to be the same cluster. Study Algorithm Algorithm improvement

16 Similarity-based clustering centers upon Red Circles Pairwise clustering focuses on Yellow Circles Study Algorithm Algorithm improvement

17 Where: Similarity-based clustering Cluster a pair of terms whose Jensen-Shannon divergence is and: Study Algorithm Algorithm improvement

18 Cluster a pair of terms whose mutual information is Pairwise clustering Where: Study Algorithm Algorithm improvement

19 Study Algorithm Algorithm improvement

20 Algorithm Implement

21 Discard stop words Stem Extract frequency Algorithm Implement Step 1: Preprocessing

22 Algorithm Implement Step 2: Selection of frequent terms Select the top frequent terms up to 30% of the number of running terms as a standard set of terms Count number of terms in document (Ntotal )

23 Algorithm Implement Step 3: Clustering frequent terms Similarity-base clustering Pairwise clustering

24 Algorithm Implement Step 4: Calculate expected probability Count the number of terms co-occurring with c ∈ C, denoted as n c, to yield the expected probability

25 Algorithm Implement Step 5: Calculate χ’2 value Where: the number of co-occurrence frequency with c ∈ C the total number of terms in the sentences including w

26 Algorithm Implement Step 6: Output keywords

27 Evaluation

28 In this paper, we developed an algorithm to extract keywords from a single document. Main advantages of our method are its simplicity without requiring use of a corpus and its high performance comparable to tfidf algorithm. As more electronic documents become available, we believe our method will be useful in many applications, especially for domain- independent keyword extraction.

29 Thank for your attention


Download ppt "Authors: Yutaka Matsuo & Mitsuru Ishizuka Designed by CProDM Team."

Similar presentations


Ads by Google