Rachid Guerraoui, EPFL. What is a good recommendation system? Recommendation systems are good.

Slides:



Advertisements
Similar presentations
Shortest Vector In A Lattice is NP-Hard to approximate
Advertisements

Approximate List- Decoding and Hardness Amplification Valentine Kabanets (SFU) joint work with Russell Impagliazzo and Ragesh Jaiswal (UCSD)
Price Of Anarchy: Routing
Knapsack Problem Section 7.6. Problem Suppose we have n items U={u 1,..u n }, that we would like to insert into a knapsack of size C. Each item u i has.
Chapter 6 The Structural Risk Minimization Principle Junping Zhang Intelligent Information Processing Laboratory, Fudan University.
Theory of Computing Lecture 3 MAS 714 Hartmut Klauck.
Study Group Randomized Algorithms 21 st June 03. Topics Covered Game Tree Evaluation –its expected run time is better than the worst- case complexity.
Critical Thinking: Does It Mean What You Think?. On Conjugation of Cyclically-Generated Banach Spaces.
Bounds on Code Length Theorem: Let l ∗ 1, l ∗ 2,..., l ∗ m be optimal codeword lengths for a source distribution p and a D-ary alphabet, and let L ∗ be.
Games of Prediction or Things get simpler as Yoav Freund Banter Inc.
CPSC 689: Discrete Algorithms for Mobile and Wireless Systems Spring 2009 Prof. Jennifer Welch.
A (1+  )-Approximation Algorithm for 2-Line-Center P.K. Agarwal, C.M. Procopiuc, K.R. Varadarajan Computational Geometry 2003.
Christos alatzidis constantina galbogini.  The Complexity of Computing a Nash Equilibrium  Constantinos Daskalakis  Paul W. Goldberg  Christos H.
Infinite Horizon Problems
Planning under Uncertainty
Greedy vs Dynamic Programming Approach
Complexity 15-1 Complexity Andrei Bulatov Hierarchy Theorem.
Models and Security Requirements for IDS. Overview The system and attack model Security requirements for IDS –Sensitivity –Detection Analysis methodology.
Clustering short time series gene expression data Jason Ernst, Gerard J. Nau and Ziv Bar-Joseph BIOINFORMATICS, vol
Defending Complex System Against External Impacts Gregory Levitin (IEC, UESTC)
CPSC 689: Discrete Algorithms for Mobile and Wireless Systems Spring 2009 Prof. Jennifer Welch.
Tirgul 8 Universal Hashing Remarks on Programming Exercise 1 Solution to question 2 in theoretical homework 2.
September 21, 2010Neural Networks Lecture 5: The Perceptron 1 Supervised Function Approximation In supervised learning, we train an ANN with a set of vector.
Preference Analysis Joachim Giesen and Eva Schuberth May 24, 2006.
Foundations of Privacy Lecture 11 Lecturer: Moni Naor.
Experts and Boosting Algorithms. Experts: Motivation Given a set of experts –No prior information –No consistent behavior –Goal: Predict as the best expert.
Learning and Planning for POMDPs Eyal Even-Dar, Tel-Aviv University Sham Kakade, University of Pennsylvania Yishay Mansour, Tel-Aviv University.
Maximum likelihood (ML)
Foundations of Cryptography Lecture 9 Lecturer: Moni Naor.
Computational aspects of stability in weighted voting games Edith Elkind (NTU, Singapore) Based on joint work with Leslie Ann Goldberg, Paul W. Goldberg,
Hamming Codes 11/17/04. History In the late 1940’s Richard Hamming recognized that the further evolution of computers required greater reliability, in.
Information theory, fitness and sampling semantics colin johnson / university of kent john woodward / university of stirling.
Dr. Hala Moushir Ebied Faculty of Computers & Information Sciences
Quality Indicators (Binary ε-Indicator) Santosh Tiwari.
The Hat Game 11/19/04 James Fiedler. References Hendrik W. Lenstra, Jr. and Gadiel Seroussi, On Hats and Other Covers, preprint, 2002,
The Multiplicative Weights Update Method Based on Arora, Hazan & Kale (2005) Mashor Housh Oded Cats Advanced simulation methods Prof. Rubinstein.
Section 8.1 Estimating  When  is Known In this section, we develop techniques for estimating the population mean μ using sample data. We assume that.
Extrapolation Models for Convergence Acceleration and Function ’ s Extension David Levin Tel-Aviv University MAIA Erice 2013.
CS433: Modeling and Simulation Dr. Anis Koubâa Al-Imam Mohammad bin Saud University 15 October 2010 Lecture 05: Statistical Analysis Tools.
Ch. 6 - Approximation via Reweighting Presentation by Eran Kravitz.
Approximation Algorithms for Knapsack Problems 1 Tsvi Kopelowitz Modified by Ariel Rosenfeld.
1 ECE-517 Reinforcement Learning in Artificial Intelligence Lecture 7: Finite Horizon MDPs, Dynamic Programming Dr. Itamar Arel College of Engineering.
Data Reduction. 1.Overview 2.The Curse of Dimensionality 3.Data Sampling 4.Binning and Reduction of Cardinality.
1 Online algorithms Typically when we solve problems and design algorithms we assume that we know all the data a priori. However in many practical situations.
1 Markov Decision Processes Infinite Horizon Problems Alan Fern * * Based in part on slides by Craig Boutilier and Daniel Weld.
The delta rule. Learn from your mistakes If it ain’t broke, don’t fix it.
Maximizing the Spread of Influence through a Social Network Authors: David Kempe, Jon Kleinberg, É va Tardos KDD 2003.
1 Markov Decision Processes Infinite Horizon Problems Alan Fern * * Based in part on slides by Craig Boutilier and Daniel Weld.
Exploiting Group Recommendation Functions for Flexible Preferences.
1  The Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
1  Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
CHAPTER 10 Widrow-Hoff Learning Ming-Feng Yeh.
1 CLUSTER VALIDITY  Clustering tendency Facts  Most clustering algorithms impose a clustering structure to the data set X at hand.  However, X may not.
Computational Thinking/CE21/CSPrinciples 1 What is Computational Thinking? Owen Astrachan.
1. 2 You should know by now… u The security level of a strategy for a player is the minimum payoff regardless of what strategy his opponent uses. u A.
Unique Games Approximation Amit Weinstein Complexity Seminar, Fall 2006 Based on: “Near Optimal Algorithms for Unique Games" by M. Charikar, K. Makarychev,
Rate Distortion Theory. Introduction The description of an arbitrary real number requires an infinite number of bits, so a finite representation of a.
Week 21 Order Statistics The order statistics of a set of random variables X 1, X 2,…, X n are the same random variables arranged in increasing order.
1 Distributed Vertex Coloring. 2 Vertex Coloring: each vertex is assigned a color.
Week 21 Statistical Model A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution that produced.
Supervise Learning Introduction. What is Learning Problem Learning = Improving with experience at some task – Improve over task T, – With respect to performance.
Estimating standard error using bootstrap
Tell Me Who I Am: An Interactive Recommendation System
Spectral Clustering.
Rank Aggregation.
Lecture 15: Least Square Regression Metric Embeddings
Unfolding with system identification
Information Theoretical Analysis of Digital Watermarking
Presentation transcript:

Rachid Guerraoui, EPFL

What is a good recommendation system? Recommendation systems are good

A good recommendation system is one that provides good recommendations What is a good recommendation?

You know it when you see it “ I shall not today attempt further to define the kinds of material I understand to be embraced within that shorthand description ["hard-core pornography"]; and perhaps I could never succeed in intelligibly doing so. But I know it when I see it, and the motion picture involved in this case is not that” Justice Potter Stewart, US Supreme Court, 1964

Ideally: Build and deploy your system Pragmatic: Transform past into future What is a good recommendation system ?

Example Members of program committee (20) want to evaluate the submitted papers (200) Nobody has enough time to read all papers Each researcher is assigned a subset of papers A recommendation system uses the scores to find the opinion of all members about all papers

What is a good recommendation? It depends on the correlation Theory to the rescue

n users k * n objects For each user and object: a grade – The grades of a user form his preference vector – The vectors of users form the preference matrix – Grades may be binary, discrete, continuous General recommendation model

Vectors of grades: v(p) (known partially to the players) Input? Vectors of grades: w(p) (seeking to approximate v(p)) Output?

Ideal output Target output w(p) = v(p) Minimize max |w(p)-v(p)| (Hamming distance)

Compare with a perfect on-line algorithm How to account for the level of correlation?

Shared billboard (1) All players know all partial vectors The perfect on-line algorithm

(2) Chooses elements of the partial vectors to fill (B budget) The player is initially indulgent (learning phase) The algorithm assigns initial papers

(3) Knows the level of correlation Hamming diameter of a set P The perfect on-line algorithm

20 pc members; 200 papers Every member can read 10 papers All have the same taste Perfect solution possible?

20 pc members; 200 papers Two clusters of 10 have the same taste Perfect solution possible? Every member needs to read 20

Assume player p can probe B objects n/B*k – 1 How many other players does p need to collaborate with to fill its vector?

20 pc members; 200 papers 4 clusters of 5 with diameter 8 Every member reads 20 What is the minimal error rate?

Ideal algorithm (k=1) A player p has to use ideas of (n/B)-1 other players to estimate her/his preferences In the worst case, p cannot do better The rate of error for p depends on the hamming distance between p and the other (n/B) players This is with a constant factor of the diameter of these n/B players

Claim For every B-algorithm, there is some distribution of preferences such that (with constant probability)

Proof (sketch) Consider a constant D > 2B Define a preference vector as follows: Let P be a set of players of size n/B - Let p in P with a random preference vector -Assign a random preference vector outside P Choose a set S of D objects. For every player q in P, v(q)=v(p) except in S which is random

Proof (sketch) Probes outside P provide no information to p Probes inside P provide no information to p w.r.t S Since p probes at most B objects and S contains D > 2B objects, there are at least D/2 objects for which p has no information No algorithm can do better than guess preferences in S The rate of error is at least D/4 and the diameter of P is less than D

Optimality An algorithm is (B,c)-optimal if for every input set of preferences

So what? The best we can do is find clusters of players that are - Small enough (small diameter) to provide “accurate” preferences And - Big enough to cover all objects Practically speaking? - Try different sizes of clusters

Optimality Assume each player can evaluate B objects. Given B, and the level of correlation among players, there is a minimum rate of error that can be achieved. There is an algorithm that obtains a constant approximation of this error-rate, and each player evalutes O(B.Polylog(n)) objects.

Definition of Optimality An algorithm is asymptotically optimal in terms of error rate, if for every player p we have: |w(p)-v(p)| n/B-1 cD(P) Where c is a constant and D(P) is the diameter of set P. P can be any set of players with size at least n/B.