Download presentation

1
**Online Recommendations**

The UV Decomposition AlgoritHm

2
Motivation It is now common to get “personal recommendations” when we visit a website. News articles Product recommendations Advertisements Why ? Unlike paper newspapers or brick and mortar stores, there is no limit [in terms of space/inventory] what can be shown or sold on a web-site.. Long-Tail effect Large part of the income comes from the tail [example in search revenue]

3
The Netflix challenge Netflix is a (online) US company from where people can rent movies Netflix would like to recommend movies to users. Netflix challenge (2006) – one million dollars prize who could beat their movie recommendation by 10% After three years the prize was awarded.. We will discuss one of the “main ideas” behind the winning entry [We will follow the discussion from the “Mining of Massive Data Sets”]

4
**The Data Data can be arranged in the form… M1 M2 M3 M4 M5 M6 M7 U1 3 2**

Users have given rating (between 1 and 5) from a database of 7 movie .. What should be recommended to U2 ?

5
**First ideas Do missing entries mean a rating of “0” ?**

How about simple dot product. Other ideas…? Clustering Association Rules ?

6
**Another basic idea Let the user “u” rate movie “m” as follow:**

Take the average of the following two numbers. Average of user’s “u” ratings. Average of all ratings given to movie “m” by all users’ who rated “m” This was only 3% worse than than the Netflix algorithm (called CineMatch)

7
Latent Modeling A big breakthrough is the idea of “latent variable modeling” The data we observe is a result of another variable or set of variables which are “latent” (not observable)… The latent variables generate the observed data… In our case…the latent variables control the generation of the ratings.. So the challenge is to “infer” the latent variables from the observed data… clustering Bayes Theorem

8
Cognitive Science/AI Scientists working in AI/Cognitive Science have drawn the following analogy.. Mind Computer Mental Representation Programs/Theories Thinking Computational Process/Algorithms Practical Outcome: Infer Latent Structures

9
**The Key Idea Decompose the User x Rating matrix into:**

User x Rating = ( User x Genre ) x (Genre x Movies) Number of Genres is typically small Or R =~ UV Find U and V such that ||R – UV|| is minimized… Almost like k-means clustering…why ?

10
**Example of UV Decomposition**

The criterion used to select U and V is the Root Mean Square Error (RMSE)

11
RMSE Example

12
Example..Continued Results in an RMSE of 1.8

13
UV Computation

14
**UV Computation..continued**

Same calculation as before…..

15
UV Computation….

16
**UV Decomposition The above process can be generalized to any entry…**

Continue the process until RMSE settles into a local optimal… So in spirit very similar to k-means..

17
References Mining of Massive Data Sets Rajaraman, Leskovic, Ullman

Similar presentations

OK

CIS 2200 Kannan Mohan Department of CIS Zicklin School of Business, Baruch College.

CIS 2200 Kannan Mohan Department of CIS Zicklin School of Business, Baruch College.

© 2018 SlidePlayer.com Inc.

All rights reserved.

Ads by Google

Ppt on old age homes Ppt on csr as per companies act 2013 Ppt on area of circle Ppt on constructive disposal of computer waste products Ppt on number system for class 7 Two dimensional viewing ppt on ipad Ppt on solar power satellites Ppt on human chromosomes types Ppt on ideal gas law r Ppt on contract labour act 1970 malaysia