Eliciting Honest Feedback 1.Eliciting Honest Feedback: The Peer-Prediction Model (Miller, Resnick, Zeckhauser) 2.Minimum Payments that Reward Honest Feedback.

Slides:



Advertisements
Similar presentations
Hansons Market Scoring Rules Robin Hanson, Logarithmic Market Scoring Rules for Modular Combinatorial Information Aggregation, Robin Hanson, Combinatorial.
Advertisements

6.896: Topics in Algorithmic Game Theory Lecture 20 Yang Cai.
Auction Theory Class 5 – single-parameter implementation and risk aversion 1.
CPS Bayesian games and their use in auctions Vincent Conitzer
Incentivize Crowd Labeling under Budget Constraint
Non myopic strategy Truth or Lie?. Scoring Rules One important feature of market scoring rules is that they are myopic strategy proof. That means that.
Computer-aided mechanism design Ye Fang, Swarat Chaudhuri, Moshe Vardi 1.
This Segment: Computational game theory Lecture 1: Game representations, solution concepts and complexity Tuomas Sandholm Computer Science Department Carnegie.
PROBABILITY. Uncertainty  Let action A t = leave for airport t minutes before flight from Logan Airport  Will A t get me there on time ? Problems :
DECISION MARKETS WITH GOOD INCENTIVES Yiling Chen (Harvard), Ian Kash (Harvard), Internet and Network Economics, 2011.
Imperfect commitment Santiago Truffa. Agenda 1.“Authority and communication in Organizations” Wouter Dessein, RES “Contracting for information.
NOGA ALON, MOSHE BABAIOFF, RON KARIDI, RON LAVI, MOSHE TENNENHOLTZ PRESENTED BY EREZ SHABAT Sequential Voting with Externalities: Herding in Social Networks.
Bundling Equilibrium in Combinatorial Auctions Written by: Presented by: Ron Holzman Rica Gonen Noa Kfir-Dahav Dov Monderer Moshe Tennenholtz.
Adverse Selection Asymmetric information is feature of many markets
A camper awakens to the growl of a hungry bear and sees his friend putting on a pair of running shoes, “You can’t outrun a bear,” scoffs the camper. His.
1 Only Valuable Experts Can be Valued Moshe Babaioff, Microsoft Research, Silicon Valley Liad Blumrosen, Hebrew U, Dept. of Economics Nicolas Lambert,
Algorithmic and Economic Aspects of Networks Nicole Immorlica.
Robust Mechanisms for Information Elicitation Aviv Zohar & Jeffrey S. Rosenschein.
A Game Theoretic Approach to Provide Incentive and Service Differentiation in P2P Networks John C.S. Lui The Chinese University of Hong Kong Joint work.
Marakas: Decision Support Systems, 2nd Edition © 2003, Prentice-Hall Chapter Chapter 4: Modeling Decision Processes Decision Support Systems in the.
Radu Jurca DIP internal workshop, EPFL, 2005 Trust, Reputation and Incentives in Online Environments Radu Jurca, Boi Faltings
CS522: Algorithmic and Economic Aspects of the Internet Instructors: Nicole Immorlica Mohammad Mahdian
Job Market Signaling (Spence model)
Robust Mechanisms for Information Elicitation Aviv Zohar & Jeffrey S. Rosenschein The Hebrew University.
Renegotiation Design with Unverifiable Info Dylan B. Minor.
DANSS Colloquium By Prof. Danny Dolev Presented by Rica Gonen
Competitive Analysis of Incentive Compatible On-Line Auctions Ron Lavi and Noam Nisan SISL/IST, Cal-Tech Hebrew University.
Games in the normal form- An application: “An Economic Theory of Democracy” Carl Henrik Knutsen 5/
Yang Cai Sep 15, An overview of today’s class Myerson’s Lemma (cont’d) Application of Myerson’s Lemma Revelation Principle Intro to Revenue Maximization.
Anonymizing Web Services Through a Club Mechanism With Economic Incentives Mamata Jenamani Leszek Lilien Bharat Bhargava Department of Computer Sciences.
Static Games of Incomplete Information.. Mechanism design Typically a 3-step game of incomplete info Step 1: Principal designs mechanism/contract Step.
Collective Revelation: A Mechanism for Self-Verified, Weighted, and Truthful Predictions Sharad Goel, Daniel M. Reeves, David M. Pennock Presented by:
Welcome! Econ A494 Math Econ & Advanced Micro Theory Spring 2013 Prof. Jim Murphy.
MAKING COMPLEX DEClSlONS
Mechanisms for Making Crowds Truthful Andrew Mao, Sergiy Nesterko.
The Multiplicative Weights Update Method Based on Arora, Hazan & Kale (2005) Mashor Housh Oded Cats Advanced simulation methods Prof. Rubinstein.
Derivative Action Learning in Games Review of: J. Shamma and G. Arslan, “Dynamic Fictitious Play, Dynamic Gradient Play, and Distributed Convergence to.
The Moral Hazard Problem Stefan P. Schleicher University of Graz
Nash equilibrium Nash equilibrium is defined in terms of strategies, not payoffs Every player is best responding simultaneously (everyone optimizes) This.
Trust-Aware Optimal Crowdsourcing With Budget Constraint Xiangyang Liu 1, He He 2, and John S. Baras 1 1 Institute for Systems Research and Department.
The Cost and Windfall of Manipulability Abraham Othman and Tuomas Sandholm Carnegie Mellon University Computer Science Department.
Mechanism Design CS 886 Electronic Market Design University of Waterloo.
Lecture 2 Economic Actors and Organizations: Motivation and Behavior.
Auction Theory תכנון מכרזים ומכירות פומביות Topic 7 – VCG mechanisms 1.
Regret Minimizing Equilibria of Games with Strict Type Uncertainty Stony Brook Conference on Game Theory Nathanaël Hyafil and Craig Boutilier Department.
Lecture 7 Course Summary The tools of strategy provide guiding principles that that should help determine the extent and nature of your professional interactions.
Paul Milgrom and Nancy Stokey Journal of Economic Thoery,1982.
Coordination with Local Information Munther Dahleh Alireza Tahbaz-Salehi, John Tsitsiklis Spyros Zoumpouli.
Lecture 5 Financial Incentives This lecture is paired with our previous one that discussed employee benefits. Here we focus on the reasons why monetary.
© 2010 Institute of Information Management National Chiao Tung University Chapter 7 Incentive Mechanism Principle-Agent Problem Production with Teams Competition.
Mechanism Design II CS 886:Electronic Market Design Sept 27, 2004.
Testing theories of fairness— Intentions matter Armin Falk, Ernst Fehr, Urs Fischbacher February 26, 2015.
Introduction Many organizations use decision rules to alleviate incentive problems as opposed to incentive contracts The principal typically retains some.
6.853: Topics in Algorithmic Game Theory Fall 2011 Constantinos Daskalakis Lecture 22.
Auctions serve the dual purpose of eliciting preferences and allocating resources between competing uses. A less fundamental but more practical reason.
Transient Unterdetermination and the Miracle Argument Paul Hoyningen-Huene Leibniz Universität Hannover Center for Philosophy and Ethics of Science (ZEWW)
Trust Signaling Maroš Servátka (University of Canterbury, NZ) Steven Tucker (University of Canterbury, NZ) Radovan Vadovič (ITAM, Mexico) June 2007 ESA,
Is Ignoring Public Information Best Policy? Reinforcement Learning In Information Cascade Toshiji Kawagoe Future University - Hakodate and Shinichi Sasaki.
On the Difficulty of Achieving Equilibrium in Interactive POMDPs Prashant Doshi Dept. of Computer Science University of Georgia Athens, GA Twenty.
Game Theory Georg Groh, WS 08/09 Verteiltes Problemlösen.
Contents of the Talk Preliminary Materials Motivation and Contribution
DESIGN ISSUES, CHOICES Between vs. within-subject design?
The Beta Reputation System
Implementation in Bayes-Nash equilibrium
Sequential Voting with Externalities: Herding in Social Networks
Implementation in Bayes-Nash equilibrium
Robust Mechanism Design with Correlated Distributions
Peer Prediction
Implementation in Bayes-Nash equilibrium
Auction Theory תכנון מכרזים ומכירות פומביות
Presentation transcript:

Eliciting Honest Feedback 1.Eliciting Honest Feedback: The Peer-Prediction Model (Miller, Resnick, Zeckhauser) 2.Minimum Payments that Reward Honest Feedback (Jurca, Faltings) Nikhil SrivastavaHao-Yuh Su

Eliciting Feedback Fundamental purpose of reputation systems Review: general setup

Eliciting Feedback Fundamental purpose of reputation systems Review: general setup –Information distributed among individuals about value of some item external goods (NetFlix, Amazon, ePinions, admissions) each other (eBay, PageRank) –Aggregated information valuable for individual or group decisions –How to gather and disseminate information?

Challenges Underprovision –inconvenience cost of contributing Dishonesty –niceness, fear of retaliation –conflicting motivations

Challenges Underprovision –inconvenience cost of contributing Dishonesty –niceness, fear of retaliation –conflicting motivations Reward systems –motivate participation, honest feedback –monetary (prestige, privilege, pure competition)

Overcoming Dishonesty Need to distinguish good from bad reports –explicit reward systems require objective outcome, public knowledge –stock, weather forecasting But what if … –subjective? (product quality/taste) –private? (breakdown frequency, seller reputability)

Overcoming Dishonesty Need to distinguish good from bad reports –explicit reward systems require objective outcome, public knowledge –stock, weather forecasting But what if … –subjective? (product quality/taste) –private? (breakdown frequency, seller reputability) Naive solution: reward peer agreement –Information cascade, herding

Peer Prediction Basic idea –reports determine probability distribution on other reports –reward based on predictive power of users report for a reference raters report –taking advantage of proper scoring rules, honest reporting is a Nash equilibrium

Outline Peer Prediction method –model, assumptions –result: –underlying intuition through example Extensions/Applications –primary practical concerns –(weak) assumptions: sequential reporting, continuous signals, risk aversion (Strong) assumptions Other limitations

Information Flow - Model PRODUCT type t CENTER (a) signal S announcement a transfer

Information Flow - Model PRODUCT type t CENTER (a) signal S transfer PRODUCT type t CENTER (a) signal S transfer announcement a

Information Flow - Example PLUMBER type = {H, L} hhlhhl signal = {h (high), l (low)} (a) agreement hhlhhl $1 $0 hhlhhl

Assumptions - Model PRODUCT type t {1, …, T} f (s | t) - fixed type - finite T - common prior: distribution p(t) - common knowledge: distribution f(s|t) - linear utility - stochastic relevance

Stochastic Relevance Informally –same product, so signals dependent certain observation (realization) should change posterior on type p(t), and thus on signal distribution f(s | t) Rolex v. Faux-lex –generically satisfied if different types yield different signal distributions

Stochastic Relevance Informally –same product, so signals dependent certain observation (realization) should change posterior on type p(t), and thus on signal distribution f(s) Rolex v. Faux-lex –generically satisfied if different types yield different signal distributions Formally –S i stochastically relevant for S j iff: distribution (S i | S j ) different for different realizations of S j there is s j such that

Assumptions - Example finite T: plumber is either H or L fixed type: plumber quality does not change common prior: p(t) –p(H) = p(L) =.5 stochastic relevance: need good plumbers signal distribution to be different than bads common knowledge: f(s|t) –p(h | H) =.85, p(h | L) =.45 –note this gives p(h), p(l)

Definitions - Model (a) T types, M signals, N raters signals: S = (S 1, …, S N ), where S i = {s 1, …, s M } announcements: a = (a 1, …, a N ), where a i = {s 1, …, s M } transfers: (a) = ( 1 (a), …, N (a)) announcement strategy for player i: a i = (a i 1, … a i M )

Definitions - Example PLUMBER type = {H, L} hhlhhl signal = {h, l} T(a) $1 $0 hhlhhl 2 types, 2 signals, 3 raters signals: S = (h, h, l) announcements: a = (h, h, l) transfers: (a) = ( 1 (a), 2 (a), 3 (a)) announcement strategy for player 2: a 2 = (a 2 h, a 2 l ) –total set of strategies: (h, h), (h, l), (l, h), (l, l)

Best Responses - Model Each player decides announcement strategy a i a i is a best-response to other strategies a -i if: Best-response strategy maximizes raters expected transfer with respect to other raters signals … conditional on S i = s m Nash equilibrium if equation holds for all i T(a)

Best Responses - Example PLUMBER S 1 = h S 2 = ? T(a) t 1 ( a 1, a 2 ) t 2 ( a 1, a 2 ) h or l a 2 Player 1 receives signal h Player 2s strategy is to report a 2 Player 1 reporting signal h is a best-response if

Peer Prediction Find reward mechanism that induces honest reporting –where a i = S i for all i is a Nash equilibrium Will need Proper Scoring Rules

Proper Scoring Rules Definition: –for two variables S i and S j, a scoring rule assigns to each announcement a i of S i a score for each realization of S j –R ( s j | a i ) –proper if score maximized by the announcement of the true realization of S i

Applying Scoring Rules Before: predictive markets (Hanson) –S i = S j = reality, a i = agent report –R ( reality | report ) –Proper scoring rules ensure honest reports: a i = S i stochastic relevance for private info and public signal automatically satisfied –What if theres no public signal?

Applying Scoring Rules Before: predictive markets (Hanson) –S i = S j = reality, a i = agent report –R ( reality | report ) –Proper scoring rules ensure honest reports: a i = S i stochastic relevance for private info and public signal automatically satisfied –What if theres no public signal? Use other reports Now: predictive peers –S i = my signal, S j = your signal, a i = my report –R ( your report | my report )

How it Works For each rater i, we choose a different reference rater r(i) Rater i is rewarded for predicting rater r(i)s announcement – * i (a i, a r(i) ) = R( a r(i), a i ) –based on updated beliefs about r(i)s announcement given is announcement Proposition: for any strictly proper scoring rule R, a reward system with transfers * i makes truthful reporting a strict Nash equilibrium

Proof of Proposition If player i receives signal s*, he seeks to maximize his expected transfer Since player r(i) is honestly reporting, his report a equals his signal s since R is proper, score is uniquely maximized by announcement of true realization of S i –uniquely maximized for a i = s* –S i is stochastically relevant for S r(i), and since r(i) reports honestly S i is stochastically relevant for a r(i) = s r(i)

Peer Prediction Example Player 1 observes low and must decide a 1 = {h, l} Assume logarithmic scoring –t 1 (a 1, a 2 ) = R(a 2 | a 1 ) = ln[ p(a 2 | a 1 )] What signal maximizes expected payoff? –Note that peer agreement would incentivize dishonesty (h) PLUMBER p(H) = p(L) =.5 S 1 = l S 2 = ? T(a) t 1 ( a 1, a 2 ) a 1 = {h, l} a 2 = s 2 p(h | H) =.85 p(h | L) =.45

Peer Prediction Example Player 1 observes low and must decide a 1 = {h, l} Assume logarithmic scoring –t 1 (a 1, a 2 ) = R(a 2 | a 1 ) = ln[ p(a 2 | a 1 )] PLUMBER p(H) = p(L) =.5 S 1 = l S 2 = ? T(a) t 1 ( a 1, a 2 ) p(h | H) =.85 p(h | L) =.45 a 1 = l (honest) yields expected transfer: -.69 a 1 = h (false) yields expected transfer: -.75 a 1 = {h, l} a 2 = s 2

Things to Note Players dont have to perform complicated Bayesian reasoning if they: –trust the center to accurately compute posteriors –believe other players will report honestly Not unique equilibrium –collusion

Primary Practical Concerns Examples –inducing effort: fixed cost c > 0 of reporting –better information: users seek multiple samples –participation constraints –budget balancing

Primary Practical Concerns Examples –inducing effort: fixed cost c > 0 of reporting –better information: users seek multiple samples –participation constraints –budget balancing Basic idea: –affine rescaling (a*x + b) to overcome obstacle –preserves honesty incentive –increases budget … see 2nd paper

Extensions to Model Sequential reporting –allows immediate use of reports –let rater i predict report of rater i + 1 –scoring rule must reflect changed beliefs about product type due to (1, …, i - 1) reports

Extensions to Model Sequential reporting –allows immediate use of reports –let rater i predict report of rater i + 1 –scoring rule must reflect changed beliefs about product type due to (1, …, i - 1) reports Continuous signals –analytic comparison of three common rules –eliciting coarse reports from exact information for two types, raters will choose closest bin (complicated)

Common Prior Assumption Practical concern - how do we know p(t),? –needed by center to compute payments –needed by raters to compute posteriors define types with respect to past products, signals types t = {1,…,9} where f(h | t) = t/10 for new product, p(t) based on past product signals update beliefs with new signals note f(s | t) given automatically

Common Prior Assumption Practical concern - how do we know p(t)? Theoretical concern - are p(t), f(s|t) public? –raters trust center to compute appropriate posterior distributions for reference raters signal –rater with private information has no guarantee center will not report true posterior beliefs rater might skew report to reflect appropriate posteriors report both private information and announcement two scoring mechanisms, one for distribution implied by private priors, another for distribution implied by announcement

Limitations Collusion –could a subset of raters gain higher transfers? higher balanced transfers? –can such strategies: overcome random pairings avoid suspicious patterns Multidimensional signals –economist with knowledge of computer science Understanding/trust in the system –complicated Bayesian reasoning, payoff rules rely on experts to ensure public confidence

Discussion Is the common priors assumption reasonable? –How might we relax it and keep some positive results? What are the most serious challenges to implementation? –Can you envision a(n online) system that rewards feedback? –How would the dynamics differ from a reward-less system? Is this paper necessary at all? –Predominance of honesty from fairness incentive cost of reporting, coarse reports, common priors, collusion, multidimensional signals, trust in system