Presentation is loading. Please wait.

Presentation is loading. Please wait.

Tuning Before Feedback: Combining Ranking Discovery and Blind Feedback for Robust Retrieval* Weiguo Fan, Ming Luo, Li Wang, Wensi Xi, and Edward A. Fox.

Similar presentations


Presentation on theme: "Tuning Before Feedback: Combining Ranking Discovery and Blind Feedback for Robust Retrieval* Weiguo Fan, Ming Luo, Li Wang, Wensi Xi, and Edward A. Fox."— Presentation transcript:

1 Tuning Before Feedback: Combining Ranking Discovery and Blind Feedback for Robust Retrieval* Weiguo Fan, Ming Luo, Li Wang, Wensi Xi, and Edward A. Fox Digital Library Research Laboratory, Virginia Tech *This research is supported by the National Science Foundation under Grant Numbers IIS-0325579, DUE-0136690 and DUE-0333531

2 Outline Introduction Research Questions Approach: Ranking Tuning + Blind Fdbk Experiment Results Conclusion

3 Introduction Ranking functions play an important role in IR performance Blind feedback (pseudo-relevance feedback) has been found very useful for ad hoc retrieval Why not combine ranking function optimization with blind feedback to improve robustness?

4 Research Questions Does blind feedback work even better on fine- tuned ranking functions as compared to on traditional ranking functions such as Okapi BM25? Does the type of query (very short vs. very long) have any impact on the combination approach? Can the ranking function discovered, in combination with blind feedback, extrapolate well for new unseen queries?

5 Our Approach Use ARRANGER a Genetic Programming-based discovery engine to perform the ranking function tuning [Fan 2003tkde, Fan 2004ip&m, Fan 2004jasist] Combine ranking tuning and feedback Test on different types of queries

6 RF Discovery Problem Feedback Training Data Input Ranking Function Discovery Ranking Function f Output

7 Ranking Function Optimization Ranking Function Tuning is an art! – Paul Kantor Why not adaptively discover RF by Genetic Programming? Huge search space Discrete objective function Modeling advantage What is GP? Problem solving system designed based on principles of evolution and heredity Widely used for structure discovery, functional form discovery, other data mining and optimization tasks

8 Genetic Algorithms/Programming Representation: Vector of bit strings or real numbers for GA Complex data structures: trees, arrays for GP Genetic transformation Reproduction Crossover Mutation IR application [Gordon’88, ’91], [Chen’98a, ’98b], [Pathak’00], etc.

9 Essential GP Components ComponentsMeaning TerminalsLeaf nodes in the tree structure (i.e., x, y). FunctionsNon-leaf nodes used to combine the leaf nodes. Commonly, numerical operations: +, -, *, /, log, sqrt. Fitness function The objective function GP aims to optimize. ReproductionA genetic operator that copies the individuals with the best fitness values directly into the population of the next generation without going through the crossover operation. CrossoverA genetic operator aiming to improve the diversity as well as the genetic fitness of the population. See details in next slide.

10 Example of Crossover in GP tf*(tf+df) tf*(N/df) + df * tf * + df Crossover Parent1 Parent2 Child1 Child2 N/df+df (tf*df)+df N / df tf + Generation: N Generation: N+1 N / df tf +

11 The ARRANGER Engine 1.Split the training data into training and validation 2.Generate an initial population of random “ranking functions” 3.Evaluate the fitness of each “ranking function” in the population and record 10 best ones 4.If stopping criteria is not met, generate the next generation of population by genetic transformation, go to Step 3. 5.Validate the recorded best “ranking functions” and select the best one as the RF 1 23 48 49 50 Start Initialize Population Evaluate Fitness Apply Crossover Stop? Validate and Output End 4849 50 123 0.40.3 0.4 0.8 0.3 0.4

12 The ARRANGER Engine 1.Split the training data into training and validation 2.Generate an initial population of random “ranking functions” 3.Evaluate the fitness of each “ranking function” in the population and record 10 best ones 4.If stopping criteria is not met, generate the next generation of population by genetic transformation, go to Step 3. 5.Validate the recorded best “ranking functions” and select the best one as the RF

13 The ARRANGER Engine 1 23 48 49 50 Start Initialize Population Evaluate Fitness Apply Crossover Stop? Validate and Output End 4849 50 123 0.40.3 0.4 0.8 0.3 0.4

14 Blind Feedback Automatically adds more terms to a user’s query to enhance the performance of search engines by assuming top ranked docs relevant Some examples Rocchio (performs best in our experiment) Dec-Hi Kullback-Leibler Divergence (KLD) Chi-Square

15 An Integrated Model

16 Experiment Setting Data 2003 Robust Track data (from TREC 6, 7, 8) Training Queries 150 old queries from TREC 6, 7, 8 Test Questions 50 very hard queries + 50 new queries

17 The Results on 150 Training Queries Run No.DescShort Okapi without BF (Baseline) 0.18800.2194 Okapi with BF0.2076 (+10.4%)0.2385 (+8.7%) RF 1 without BF0.2173 (+15.6%)0.2394 (+9.1%) RF 1 with BF0.2422 (+28.8%)0.2661 (+21.3%)

18 Results on Test Queries (1)

19 Results on Test Queries (2)

20 Conclusions Blind feedback works well on GP trained queries. Ranking function combined with blind feedback works with new queries Two stage model responds differently to Desc query (slightly better) and Long query

21 Thank You! Q&A?


Download ppt "Tuning Before Feedback: Combining Ranking Discovery and Blind Feedback for Robust Retrieval* Weiguo Fan, Ming Luo, Li Wang, Wensi Xi, and Edward A. Fox."

Similar presentations


Ads by Google