Automating Readers’ Advisory to Make Book Recommendations for K-12 Readers by Alicia Wood.

Slides:



Advertisements
Similar presentations
Recommender System A Brief Survey.
Advertisements

Image Retrieval With Relevant Feedback Hayati Cam & Ozge Cavus IMAGE RETRIEVAL WITH RELEVANCE FEEDBACK Hayati CAM Ozge CAVUS.
Pseudo-Relevance Feedback For Multimedia Retrieval By Rong Yan, Alexander G. and Rong Jin Mwangi S. Kariuki
You can use this presentation to: Gain an overall understanding of the purpose of the revised tool Learn about the changes that have been made Find advice.
1 Evaluation Rong Jin. 2 Evaluation  Evaluation is key to building effective and efficient search engines usually carried out in controlled experiments.
1.Accuracy of Agree/Disagree relation classification. 2.Accuracy of user opinion prediction. 1.Task extraction performance on Bing web search log with.
Jean-Eudes Ranvier 17/05/2015Planet Data - Madrid Trustworthiness assessment (on web pages) Task 3.3.
GENERATING AUTOMATIC SEMANTIC ANNOTATIONS FOR RESEARCH DATASETS AYUSH SINGHAL AND JAIDEEP SRIVASTAVA CS DEPT., UNIVERSITY OF MINNESOTA, MN, USA.
A New Suffix Tree Similarity Measure for Document Clustering Hung Chim, Xiaotie Deng City University of Hong Kong WWW 2007 Session: Similarity Search April.
Sentiment Lexicon Creation from Lexical Resources BIS 2011 Bas Heerschop Erasmus School of Economics Erasmus University Rotterdam
Gimme’ The Context: Context- driven Automatic Semantic Annotation with CPANKOW Philipp Cimiano et al.
Learning to Advertise. Introduction Advertising on the Internet = $$$ –Especially search advertising and web page advertising Problem: –Selecting ads.
ReQuest (Validating Semantic Searches) Norman Piedade de Noronha 16 th July, 2004.
Latent Semantic Analysis (LSA). Introduction to LSA Learning Model Uses Singular Value Decomposition (SVD) to simulate human learning of word and passage.
Xiaomeng Su & Jon Atle Gulla Dept. of Computer and Information Science Norwegian University of Science and Technology Trondheim Norway June 2004 Semantic.
1 UCB Digital Library Project An Experiment in Using Lexical Disambiguation to Enhance Information Access Robert Wilensky, Isaac Cheng, Timotius Tjahjadi,
Session 6: Writing from Sources Audience: K-5 Teachers.
Knowledge Science & Engineering Institute, Beijing Normal University, Analyzing Transcripts of Online Asynchronous.
DOG I : an Annotation System for Images of Dog Breeds Antonis Dimas Pyrros Koletsis Euripides Petrakis Intelligent Systems Laboratory Technical University.
A Privacy Preserving Efficient Protocol for Semantic Similarity Join Using Long String Attributes Bilal Hawashin, Farshad Fotouhi Traian Marius Truta Department.
Combining Content-based and Collaborative Filtering Department of Computer Science and Engineering, Slovak University of Technology
Slide Image Retrieval: A Preliminary Study Guo Min Liew and Min-Yen Kan National University of Singapore Web IR / NLP Group (WING)
Including a detailed description of the Colorado Growth Model 1.
Copyright R. Weber Machine Learning, Data Mining ISYS370 Dr. R. Weber.
Tree Kernels for Parsing: (Collins & Duffy, 2001) Advanced Statistical Methods in NLP Ling 572 February 28, 2012.
An Integrated Approach to Extracting Ontological Structures from Folksonomies Huairen Lin, Joseph Davis, Ying Zhou ESWC 2009 Hyewon Lim October 9 th, 2009.
Semantic Similarity over Gene Ontology for Multi-label Protein Subcellular Localization Shibiao WAN and Man-Wai MAK The Hong Kong Polytechnic University.
1 Wikification CSE 6339 (Section 002) Abhijit Tendulkar.
An Automated Approach to Predict Effectiveness of Fault Localization Tools Tien-Duy B. Le, and David Lo School of Information Systems Singapore Management.
Mining the Semantic Web: Requirements for Machine Learning Fabio Ciravegna, Sam Chapman Presented by Steve Hookway 10/20/05.
Learning Object Metadata Mining Masoud Makrehchi Supervisor: Prof. Mohamed Kamel.
Reyyan Yeniterzi Weakly-Supervised Discovery of Named Entities Using Web Search Queries Marius Pasca Google CIKM 2007.
©2008 Srikanth Kallurkar, Quantum Leap Innovations, Inc. All rights reserved. Apollo – Automated Content Management System Srikanth Kallurkar Quantum Leap.
PAUL ALEXANDRU CHIRITA STEFANIA COSTACHE SIEGFRIED HANDSCHUH WOLFGANG NEJDL 1* L3S RESEARCH CENTER 2* NATIONAL UNIVERSITY OF IRELAND PROCEEDINGS OF THE.
User Profiling based on Folksonomy Information in Web 2.0 for Personalized Recommender Systems Huizhi (Elly) Liang Supervisors: Yue Xu, Yuefeng Li, Richi.
Exploring Online Social Activities for Adaptive Search Personalization CIKM’10 Advisor : Jia Ling, Koh Speaker : SHENG HONG, CHUNG.
Use of Hierarchical Keywords for Easy Data Management on HUBzero HUBbub Conference 2013 September 6 th, 2013 Gaurav Nanda, Jonathan Tan, Peter Auyeung,
GAUSSIAN PROCESS FACTORIZATION MACHINES FOR CONTEXT-AWARE RECOMMENDATIONS Trung V. Nguyen, Alexandros Karatzoglou, Linas Baltrunas SIGIR 2014 Presentation:
TEMPLATE DESIGN © Zhiyao Duan 1,2, Lie Lu 1, and Changshui Zhang 2 1. Microsoft Research Asia (MSRA), Beijing, China.2.
Weighting and Matching against Indices. Zipf’s Law In any corpus, such as the AIT, we can count how often each word occurs in the corpus as a whole =
Developing Trust Networks based on User Tagging Information for Recommendation Making Touhid Bhuiyan et al. WISE May 2012 SNU IDB Lab. Hyunwoo Kim.
Automatic Syllabus Classification JCDL – Vancouver – 22 June 2007 Edward A. Fox (presenting co-author), Xiaoyan Yu, Manas Tungare, Weiguo Fan, Manuel Perez-Quinones,
CONCLUSION & FUTURE WORK Normally, users perform search tasks using multiple applications in concert: a search engine interface presents lists of potentially.
Contextual Ranking of Keywords Using Click Data Utku Irmak, Vadim von Brzeski, Reiner Kraft Yahoo! Inc ICDE 09’ Datamining session Summarized.
A Scalable Machine Learning Approach for Semi-Structured Named Entity Recognition Utku Irmak(Yahoo! Labs) Reiner Kraft(Yahoo! Inc.) WWW 2010(Information.
New NoveList Training. Why the new interface? We’ve spent the two years since our last interface change listening to your feedback, conducting focus groups,
A Content-Based Approach to Collaborative Filtering Brandon Douthit-Wood CS 470 – Final Presentation.
Flickr Tag Recommendation based on Collective Knowledge BÖrkur SigurbjÖnsson, Roelof van Zwol Yahoo! Research WWW Summarized and presented.
Vector Space Models.
Performance Measures. Why to Conduct Performance Evaluation? 2 n Evaluation is the key to building effective & efficient IR (information retrieval) systems.
Pairwise Preference Regression for Cold-start Recommendation Speaker: Yuanshuai Sun
A Novel Visualization Model for Web Search Results Nguyen T, and Zhang J IEEE Transactions on Visualization and Computer Graphics PAWS Meeting Presented.
Improved Video Categorization from Text Metadata and User Comments ACM SIGIR 2011:Research and development in Information Retrieval - Katja Filippova -
Generating Query Substitutions Alicia Wood. What is the problem to be solved?
Term Weighting approaches in automatic text retrieval. Presented by Ehsan.
PERSONALIZED DIVERSIFICATION OF SEARCH RESULTS Date: 2013/04/15 Author: David Vallet, Pablo Castells Source: SIGIR’12 Advisor: Dr.Jia-ling, Koh Speaker:
Virtual Examples for Text Classification with Support Vector Machines Manabu Sassano Proceedings of the 2003 Conference on Emprical Methods in Natural.
A Supervised Machine Learning Algorithm for Research Articles Leonidas Akritidis, Panayiotis Bozanis Dept. of Computer & Communication Engineering, University.
Argumentative Writing Grades College and Career Readiness Standards for Writing Text Types and Purposes arguments 1.Write arguments to support a.
Improving the Classification of Unknown Documents by Concept Graph Morteza Mohagheghi Reza Soltanpour
Identifying “Best Bet” Web Search Results by Mining Past User Behavior Author: Eugene Agichtein, Zijian Zheng (Microsoft Research) Source: KDD2006 Reporter:
User Modeling and Recommender Systems: recommendation algorithms
A Framework to Predict the Quality of Answers with Non-Textual Features Jiwoon Jeon, W. Bruce Croft(University of Massachusetts-Amherst) Joon Ho Lee (Soongsil.
Multi-Class Sentiment Analysis with Clustering and Score Representation Yan Zhu.
A Document-Level Sentiment Analysis Approach Using Artificial Neural Network and Sentiment Lexicons Yan Zhu.
Personalized Ontology for Web Search Personalization S. Sendhilkumar, T.V. Geetha Anna University, Chennai India 1st ACM Bangalore annual Compute conference,
Personalized Social Image Recommendation
Multimedia Information Retrieval
Learning to Rank with Ties
Bug Localization with Combination of Deep Learning and Information Retrieval A. N. Lam et al. International Conference on Program Comprehension 2017.
Presentation transcript:

Automating Readers’ Advisory to Make Book Recommendations for K-12 Readers by Alicia Wood

Problem Existing book recommenders failed to offer adequate choices for K-12 readers Important to provide good reading material among K-12 students Not easy to find the right books for the right audience

Who cares? Parents – 32% of American 4 th graders proficient in reading Children

Previous Work Previous book recommenders Extract features, opinion, feature/opinion pairs – Bootstrapping, NLP, ML, extraction rules, latent semantic analysis, statistical analysis, and information retrieval Information extraction approaches on product reviews Amazon Require historical data Require an ontology Don’t consider readability level of users

Proposed Solution Rabbit Multi-dimensional approach No feedback from users ABET (Appeal-based extraction tool)

Readers’ Advisory Offers materials of potential interest with “the help of knowledgeable and non-judgmental library staff” Based on: – Reasons behind preferences – Topical areas – Content descriptions – Appeal factors (pacing, description of characters, tone, etc.)

Appeal Factors & Terms

ABET Extracts appeal-term descriptions of books from reviews available Imperative to properly associate appeal terms and appeal factors – pairs can be correctly extracted to generate accurate appeal-term description for the book

Extraction Rules for ABET

Example SA = “The narrative of the book is dramatic” – Subject: narrative SB = “He creates believable characters” – Object: character (AF) If subject/object is an appeal factor, the word semantically linked to that subject/object is often an appeal term Rules 1 + 2

Example “The characters are not simple – Rule 4 – negation

ABET Creates the appeal-term description for a book applying the rules Frequency of occurrence – degree of significance

Rabbit 1.Analyze profile of a reader 2.Determine readability level 3.Select books 4.Compute ranking score

Candidate Books CB – candidate book available at a book repository Rep PB each book in R’s profile |P| - # of books in R’s profile TRoLL(CB), TRoLL(PB) – grade level of CB/PB determined by TRoLL

Topical Similarity Measure CB – vector of weights of CB if subject heading is of CB P – vector of weights of Pi (proportion between number of books in P that have been assigned Pi)

Content Similarity Enhanced version of cosine similarity CB = vector of Wcb1…Wcbn P = vector of Wp1…Wpn Wpi, Wcbi = weights of keywords Pi and Cb

Appeal Term Similarity F = set of appeal factors in appeal term descriptions CBf and Pf = n dimensional vector representation of appeal term distribution of an appeal factor (f)

Ranking Candidate Books Multiple linear regression Train using Ordinary Least Squares method – T set dataset

Experimental Results 1.Compute precision and recall of appeal factor-appeal term pairs extracted from book reviews 2.Analyze correctness of appeal-term descriptions created by ABET 3.Compare appeal-term descriptions generated by ABET with respect to ones extracted from Novelist

Experimental Results – books Manually annotated and compared with ABET Precision: 0.85 Recall: 0.82 F-measure: 0.83 High accuracy for ABET in generating appeal factor-appeal term pairs

Experimental Results - 2 Surveys to determine correctness Overall 94% accuracy

Experimental Results - 3 Surveys to determine comparison Appeal-term descriptions provided by ABET favored over Novelist

Validation Computed empirical studies Assessed performance of Rabbit using Eset in terms of Normalized Discounted Cumulative Gain Rabbit locates more relevant books Rabbit outperforms GoodReads and Novelist