Content-based Recommendation Systems

Slides:



Advertisements
Similar presentations
Recommender Systems & Collaborative Filtering
Advertisements

Pseudo-Relevance Feedback For Multimedia Retrieval By Rong Yan, Alexander G. and Rong Jin Mwangi S. Kariuki
Text Categorization.
Improvements and extras Paul Thomas CSIRO. Overview of the lectures 1.Introduction to information retrieval (IR) 2.Ranked retrieval 3.Probabilistic retrieval.
Prediction Modeling for Personalization & Recommender Systems Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.
DECISION TREES. Decision trees  One possible representation for hypotheses.
Chapter 5: Introduction to Information Retrieval
Introduction to Information Retrieval
Data Mining Classification: Alternative Techniques
Bag-of-Words Methods for Text Mining CSCI-GA.2590 – Lecture 2A
George Lee User Context-based Service Control Group
Learning for Text Categorization
Recommender Systems Aalap Kohojkar Yang Liu Zhan Shi March 31, 2008.
CS345 Data Mining Recommendation Systems Netflix Challenge Anand Rajaraman, Jeffrey D. Ullman.
Database Management Systems, R. Ramakrishnan1 Computing Relevance, Similarity: The Vector Space Model Chapter 27, Part B Based on Larson and Hearst’s slides.
Modern Information Retrieval Chapter 2 Modeling. Can keywords be used to represent a document or a query? keywords as query and matching as query processing.
Modeling Modern Information Retrieval
Agent Technology for e-Commerce
Recommender systems Ram Akella February 23, 2011 Lecture 6b, i290 & 280I University of California at Berkeley Silicon Valley Center/SC.
Vector Space Model CS 652 Information Extraction and Integration.
1 Introduction to Recommendation System Presented by HongBo Deng Nov 14, 2006 Refer to the PPT from Stanford: Anand Rajaraman, Jeffrey D. Ullman.
Recommender Systems; Social Information Filtering.
Modern Information Retrieval Chapter 2 Modeling. Can keywords be used to represent a document or a query? keywords as query and matching as query processing.
Recommender systems Ram Akella November 26 th 2008.
Information Retrieval
12 -1 Lecture 12 User Modeling Topics –Basics –Example User Model –Construction of User Models –Updating of User Models –Applications.
Chapter 5: Information Retrieval and Web Search
1 Text Categorization  Assigning documents to a fixed set of categories  Applications:  Web pages  Recommending pages  Yahoo-like classification hierarchies.
Modeling (Chap. 2) Modern Information Retrieval Spring 2000.
Content-Based Recommendation Systems Michael J. Pazzani and Daniel Billsus Rutgers University and FX Palo Alto Laboratory By Vishal Paliwal.
Processing of large document collections Part 2 (Text categorization) Helena Ahonen-Myka Spring 2006.
Distributed Networks & Systems Lab. Introduction Collaborative filtering Characteristics and challenges Memory-based CF Model-based CF Hybrid CF Recent.
Chapter 7 Web Content Mining Xxxxxx. Introduction Web-content mining techniques are used to discover useful information from content on the web – textual.
Processing of large document collections Part 2 (Text categorization, term selection) Helena Ahonen-Myka Spring 2005.
K Nearest Neighborhood (KNNs)
The identification of interesting web sites Presented by Xiaoshu Cai.
Xiaoying Gao Computer Science Victoria University of Wellington Intelligent Agents COMP 423.
Recommendation system MOPSI project KAROL WAGA
Data Mining Chapter 1 Introduction -- Basic Data Mining Tasks -- Related Concepts -- Data Mining Techniques.
Xiaoying Gao Computer Science Victoria University of Wellington Intelligent Agents COMP 423.
Presented By :Ayesha Khan. Content Introduction Everyday Examples of Collaborative Filtering Traditional Collaborative Filtering Socially Collaborative.
Toward the Next generation of Recommender systems
Chapter 6: Information Retrieval and Web Search
NewsWeeder: Learning to Filter Netnews By: Ken Lang Presented by Salah Omer.
1 Computing Relevance, Similarity: The Vector Space Model.
CPSC 404 Laks V.S. Lakshmanan1 Computing Relevance, Similarity: The Vector Space Model Chapter 27, Part B Based on Larson and Hearst’s slides at UC-Berkeley.
Data Management and Database Technologies 1 DATA MINING Extracting Knowledge From Data Petr Olmer CERN
Bag-of-Words Methods for Text Mining CSCI-GA.2590 – Lecture 2A Ralph Grishman NYU.
Data Mining: Knowledge Discovery in Databases Peter van der Putten ALP Group, LIACS Pre-University College LAPP-Top Computer Science February 2005.
USE RECIPE INGREDIENTS TO PREDICT THE CATEGORY OF CUISINE Group 7 – MEI, Yan & HUANG, Chenyu.
Exploring in the Weblog Space by Detecting Informative and Affective Articles Xiaochuan Ni, Gui-Rong Xue, Xiao Ling, Yong Yu Shanghai Jiao-Tong University.
KNN & Naïve Bayes Hongning Wang Today’s lecture Instance-based classifiers – k nearest neighbors – Non-parametric learning algorithm Model-based.
User Modeling and Recommender Systems: Introduction to recommender systems Adolfo Ruiz Calleja 06/09/2014.
A System for Automatic Personalized Tracking of Scientific Literature on the Web Tzachi Perlstein Yael Nir.
User Modeling and Recommender Systems: recommendation algorithms
Xiaoying Gao Computer Science Victoria University of Wellington COMP307 NLP 4 Information Retrieval.
Eick: kNN kNN: A Non-parametric Classification and Prediction Technique Goals of this set of transparencies: 1.Introduce kNN---a popular non-parameric.
Recommendation Systems By: Bryan Powell, Neil Kumar, Manjap Singh.
1 Text Categorization  Assigning documents to a fixed set of categories  Applications:  Web pages  Recommending pages  Yahoo-like classification hierarchies.
KNN & Naïve Bayes Hongning Wang
Lecture-6 Bscshelp.com. Todays Lecture  Which Kinds of Applications Are Targeted?  Business intelligence  Search engines.
Automated Information Retrieval
Data Mining: Concepts and Techniques
Recommender Systems & Collaborative Filtering
Data Mining Lecture 11.
Information Retrieval
Author: Kazunari Sugiyama, etc. (WWW2004)
Text Categorization Assigning documents to a fixed set of categories
Authors: Wai Lam and Kon Fan Low Announcer: Kyu-Baek Hwang
Recommendation Systems
Presentation transcript:

Content-based Recommendation Systems Group: Tippy

Group Members Nerin George Deepan Murugan Thach Tran Goal Models + Presentation Deepan Murugan Domain Models + Presentation Thach Tran Strategies + Presentation

Outline Introduction Item Representation User Profiles Manual Recommendation Methods Learning A User Model Classification Learning Algorithms Decision Trees and Rule Induction Nearest Neighbour Methods Conclusions Q & A

Introduction The WWW is growing exponentially. Many websites become enormous in term of size and complexity Users need help in finding items that are in accordance with their interests Recommendation Content-based recommendation: recommend an item to a user based upon a description of the item and a profile of the user’s interests

Introduction Pazzani, M. J., & Billsus, D. (2007). Content- Based Recommendation Systems. Lecture Notes in Computer Science. (4321), 325-341.

Related Research Recommender systems present items (e.g., movies, books, music, images, web pages, news, etc.) that are likely of interest to the user compare the user’s profile to some reference characteristics to predict whether the user would be interested in an unseen item Reference characteristics Information about the unseen item  content-based approach User’s social environment  collaborative filtering approach

Item Representation Items stored in a database table Structured data Small number of attributes Each item is described by the same set of attributes Known set of values that the attributes may have Straightforward to work with User’s profile contains positive rating for 1001, 1002, 1003 Would the user be interested in say Oscars (French cuisine, table service)? ID Name Cuisine Service Cost 1001 Mike’s Pizza Italian Counter Low 1002 Chris’s Café French Table Medium 1003 Jacques Bistro High

Item Representation Information about item could also be free text; e.g., text description or review of the restaurant, or news articles Unstructured data No attribute names with well-defined values Natural language complexity Same word with different meanings Different words with same meaning Need to impose structure on free text before it can be used in recommendation algorithm

TF*IDF Weighting First, stemming is applied to get the root forms of words “compute”, “computation”, “computer”, “computes”, etc., are represented by one term Compute a weight for each term that represents the importance or relevance of that term

TF*IDF Weighting Term frequency tft,d of a term t in a document d Inverse document frequency idft of a term t TF*IDF weighting nt,d is term count of t in d N is number of documents in the collection dft is number of documents that contains term t

TF*IDF Weighting The term with highest weight occur more often in that document than in other documents  more central to the topic of the document Limitations This method does not capture the context in which a word is used “This restaurant does not serve vegetarian dishes” The term “vegetarian” might still have significant weight according to the method and the restaurant might get classified into a group of restaurants which serve vegetarian food.

User Profiles A profile of the user’s interests is used by most recommendation systems This profile consists of two main types of information A model of the user’s preferences. E.g., a function that for any item predicts the likelihood that the user is interested in that item User’s interaction history. E.g., items viewed by a user, items purchased by a user, search queries, etc.

User Profiles User’s history will be used as training data for a machine learning algorithm that creates a user model “Manual” recommending approaches User customisation Provide “check box” interface that let the users construct their own profiles of interests A simple database matching process is used to find items that meet the specified criteria and recommend these to users.

User Profiles Limitations Require efforts from users Cannot cope with changes in user’s interests Do not provide a way to determine order among recommending items

User Profiles “Manual” recommending approaches Rule-based Recommendation The system has rules to recommend other products based on user history Rule to recommend sequel to a book or movie to customers who purchased the previous item in the series Can capture common reasons for making recommendations

Learning a User Model Creating a model of the user’s preference from the user history is a form of classification learning The training data (i.e., user’s history) could be captured through explicit feedback (e.g., user rates items) or implicit observing of user’s interactions (e.g., user bought an item and later returned it is a sign of user doesn’t like the item) Implicit method can collect large amount of data but could contains noise while data collected through explicit method is perfect but the amount collected could be limited

Learning a User Model Next, a number of classification learning algorithms are reviewed The main goal of these classification learning algorithms is to learn a function that model the user’s interests Applying the function on a new item can give the probability that a user will like this item or a numeric value indicating the degree of interest in this item

Decision Trees and Rule Induction Given the history of user’s interests as training data, build a decision tree which represents the user’s profile of interest Will the user like an inexpensive Mexican restaurant? Cuisine Service Cost Rating Italian Counter Low Negative French Table Med Positive …

Decision Trees and Rule Induction Well-suited for structured data In unstructured data, the number of attributes becomes too enormous and consequently, the tree becomes too large to provide sufficient performance RIPPER: a rule induction algorithm based on the same principles but provide better performance in classifying text RIPPER is a rule induction algorithm closely related to decision trees that operates in a similar fashion to the recursive data partitioning approach described above. Despite the problematic inductive bias, however, RIPPER performs competitively with other state-of-the-art text classification algorithms. In part, the performance can be attributed to a sophisticated post-pruning algorithm that optimizes the fit of the induced rule set with respect to the training data as a whole. Furthermore, RIPPER supports multi-valued attributes, which leads to a natural representation for text classification tasks, i.e., the individual words of a text document can be represented as multiple feature values for a single feature. While this is essentially a representational convenience if rules are to be learned from unstructured text documents, the approach can lead to more powerful classifiers for semi-structured text documents. For example, the text contained in separate fields of an email message, such as sender, subject, and body text, can be represented as separate multi-valued features, which allows the algorithm to take advantage of the document’s structure in a natural fashion.

Nearest Neighbour Methods Simply store all the training data in memory To classify a new item, compare it to all stored items using a similarity function and determine the “nearest neighbour” or the k nearest neighbours. The class or numeric score of the previously unseen item can then be derived from the class of the nearest neighbour.

Nearest Neighbour Methods unseen item needed to be classified positive rated items negative rated items k = 3: negative k = 5: positive

Nearest Neighbour Methods The similarity function depends on the type of data Structured data: Euclidean distance metric Unstructured data (i.e., free text): cosine similarity function

Euclidean Distance Metric Distance between A and B Attributes which are not measured quantitatively need to be labeled by numbers representing their categories Cuisine attribute: 1=Frech, 2=Italian, 3=Mexican. Item Attr. X Attr. Y Attr. Z A XA YA ZA B XB YB ZB

Cosine Similarity Function Vector space model An item or a document d is represented as a vector wt,d is the tf*idf weight of a term t in a document d The similarity between two items can then be computed by the cosine of the angle between two vectors

Nearest Neighbour Methods Despite the simplicity of the algorithm, its performance has been shown to be competitive with more complex algorithms

Other Classification Learning Algorithms Relevance Feedback and Rocchio’s Algorithm Linear Classifiers Probabilistic Methods and Naïve Bayes These are more complex methods which have been described in the paper but we don’t have time to cover them in this presentation

Conclusions Can only be effective in limited circumstances. It is not straightforward to recognise the subtleties in content Depend entirely on previous selected items and therefore cannot make predictions about future interests of users These shortcomings can be addressed by collaborative filtering (CF) techniques CF is the dominant technique nowadays thanks to the popularity of Web 2.0/Social Web concept Many recommendation system utilise a hybrid of content-based and collaborative filtering approaches

Summary Content-based Recommendation Item Representation User Profiles Manual Recommendation Methods Learning A User Model Decision Trees and Rule Induction Nearest Neighbour Methods

Q & A