Presentation on theme: "1 Field Experiment on Incentive of Communication, Compilation and Evaluation 1. Motivation for the study 2. What is the Field Experiment 3. Framework of."— Presentation transcript:
1 Field Experiment on Incentive of Communication, Compilation and Evaluation 1. Motivation for the study 2. What is the Field Experiment 3. Framework of the experiment 4. The flow of the experiment 5. Analysis of Results Takuya Nakaizumi 2009.6.10
2 Motivation for the study 1. Does monetary Incentive affect the conversation or editing? 1. Editors or writers have their own opinion.They will reject the opinion of others on their writing or compilation especially when they do not agree with or do not like it. 2. Even in such circumstances, incentives may make them edit in a way different from what they would have done otherwise. 2. In this experiment, we tested whether a monetary incentive affected the editing of a conversation in the BBS.
3 Evaluation of the editing 1. To give an editor adequate incentive, we need an evaluation of the performance on which the reward depends 2. The editing must have a true value. However it is quite difficult to estimate. 3. Thus in this study we employ a peer review system. The editing was evaluated by all the participants of the conversation in the BBS. And the rewards of the editor depended on the evaluation of the participants of the conversation of the BBS.
4 Evaluation Behavior 1.Participants of the conversation of the BBS, might appreciate the editing more if the editing attracts the participants. 2.The reward for the editor might affect just the effort of the editor. The reward itself for the editor might not affect the evaluation from the other participants. 3.Thus higher rewards lead to more attractive editing. → We check it by experiment
5 What is a Field Experiment? Harrison and List  : Our primary point is that dissecting the characteristics of field experiments helps define what might be better called an ideal experiment, in the sense that one is able to observe a subject in a controlled setting but where the subject does not perceive any of the controls as being unnatural and there is no deception being practiced. The distinction between the laboratory and the field 1. In a field experiment, sample populations are drawn from many different domains, while sample is student of ordinal laboratory experiment. 2. The laboratory environment might not be fully representative of the field environment. c.f. winner’s curse
6 Determine the Field context of an Experiment Harrison and List  : the subject (participant) pool, the information that the subjects bring to the task, the commodity, the task or trading rules applied, the stakes, the environment in which the subjects operate.
7 Taxonomy of Field Experiment Harrison and List  : 1. ‘artefactual’ field experiment: ‘ non-standard’ subjects, or experimental participants from the market of interest. 2. ‘framed field experiment’: as the same as an artefactual field experiment but with field context in the commodity, task, stakes, or information set of the subjects, the commodity, 3. ‘natural field experiment’: ‘natural field experiment’ is the same as a framed field experiment but where the environment is one where the subjects naturally undertake these tasks and where the subjects do not know that they are participants in an experiment. Our field experiment
8 The Flow of the Experiment 1 1. Pre-survey of the topic and attitude of the participants 2. Discussion on BBS on specific topic Religion, Politics, Marketing, Finance (Japan only) Discussion is rewarded if they contribute more than four times. 3. Choice of the editor of the discussion by Calculating Social Influence Score 4. Editing of the editor 5. Evaluation of the editing by the participants and the editor is then rewarded according to the evaluation.
9 The Flow of the Experiment 2 1. Pre-survey 1. Opinions of the participants on which topic they prefer to discuss. 2. Variance of the opinions in each domain 2. Discussion on BBS 1. Religion, Politics (Marketing, Finance in Japan only) 2. The most and least conflicting topic based on the variance of opinions in each domain.
10 The Flow of the Experiment 3 1. Calculate (SIS) Social Influence Score l Based on mutual evaluation l Like in Google, the SIS evaluation carries more weight if the evaluator has a higher SIS. l Long tail distribution 2. Editing/Summarizing Assigned to the discussant with the highest SIS Two types of reward for the editing, $20 or $80. If others’ assessment is low, then the rewards is reduced. “How satisfied are you with the summary by the selected editor? In order to guarantee the quality of the discussion summary, if the editor is rated less than 'not satisfactory' (A 2 out of a scale of 7) on average, then the bonus for editing will be reduced by half.”
11 Experimental conditions Domains Economic: Marketing, Investment (in Japan only) Non-Economic: Politics, Religions Difficulties of topics: variance in survey High: The most conflicting topic Low: The least conflicting topic Amount of reward to a editor High:$80 (¥8,000 in Japan) Low:$20 (¥2,000 in Japan)
12 Incentive to discuss and edit Discussion on BBS If they post more than four comments, they get the participant fee. Higher rewards for a editors becomes an incentive for participants to discuss the topic. Editing of BBS High:$80 (¥8,000 in Japan) or Low:$20 (¥2,000 in Japan) Higher rewards let the editor flatter the participants more and that may raise the actual score. Evaluation of Editing There is no monetary incentive to evaluate the editing!!
13 Summary of Experiment Data In Japan reward to each editor Doma- ins Difficulty of topics Total BBS Partici- pants In each group Partici- pants at the beginning Partici- pants who continue to the end ¥8,00042 types8×3=2410240 (24 editor) 148+ 20(editor) ¥2,00042 types8×3=2410240 (24 editor) 137 + 20(editor) total42 types4810480 (48 editor) 285+ 40(editor)
14 Summary of Experiment Data In the U.S.(and combined total) reward to each editor Domain religion, politics Difficulty of topics Total BBS Partici- pants In each group Partici- pants at the beginning Partici-pants who continue to the end $8022 types4×3=1220240 (12 editor) 138 + 12(editor) $2022 types4×3=1220240 (12 editor) 122 + 8(editor) total22 types2420480 (24 editor) 260 + 20(editor) U.S., and Japan 2,42 types24 or 48 20 or 10 840 (72 editor) 545 + 60(editor)
15 Basic Model (1) Evaluation depends on the quality of editing by editor 0, that is X 0 Assumption 1: i participant’s utility is Thus the evaluation score of i by j, When editor edit the conversation, the editor does not know the evaluation of the other participants. Thus she/he face uncertainty and we describe it by ε, ε 〜
16 Basic Model (2) Quality of editing is assumed to be the function of the effort of editor: X 0 =X 0 (x)=x and reward of the editor, W(s 0 ) is And Cost Function is αc (x), (α>0, c'(x)>0, c" (x)>0, ) Thus editor’s expected benefit of editing, E[B (x)] is
17 Basic Model(3) Then (1) From →Proposition 1There is interior solution of (1)
18 Basic Model(4) That means x 0 is non decreasing function of w And non-decreasing function of α and Thus and Hypothesis : 1) Higher rewards with easy topic lead to highest score. 2)Lower rewards with difficult topic lead to lowest score. 3)The score of Higher reward with difficult topic and Lower reward with easy topic are between them
19 Results of Experiment (1): Evaluation Score (1) Difficulty Reward for editing Difference of Topic: 20 80 (p-value of GWT) Easy 0.889 0.817 -0.072 1.5356 1.612 (0.1635) 117 131 Difficult 1.127 0.690 -0.437* 1.315 1.577 (0.0017) 142 155 Total 1.019 0.748 -0.271* 1.421 1.592 (0.03) 259 286
20 Results of Experiment (2): length of editing (effort)
21 Results of Experiment (3): Analysis of Evaluation Score(1) Higher reward with easy topic Ordinal incentive theory lower reward with difficult topic lower reward with easy topic higher reward with difficult topic Experimental results 1.127 lower with difficult 0.889 lower with easy 0.817 higher with easy 0.748 Higher with difficult
22 Results of Experiment (4): Analysis of Evaluation Score(2) Higher reward with easy topic incentive theory with spite lower reward with difficult topic lower reward with easy topic higher reward with difficult topic Experimental results 1.127 lower with difficult 0.889 lower with easy 0.817 higher with easy 0.748 Higher with difficult
23 Results of Experiment (5): Analysis of Evaluation Score(3) Higher reward with easy topic incentive theory with fairness evaluation lower reward with difficult topic lower reward with easy topic higher reward with difficult topic Experimental results 1.127 lower with difficult 0.889 lower with easy 0.817 higher with easy 0.748 Higher with difficult
24 Results of Experiment (6): Evaluation Score (2) Low reward ($20,or ¥2,000) High reward ($80,or ¥8,000) Difference (p-value of GWT) Japan (average, variance sample) 0.854 1.401 137 0.365 1.591 148 -0.489* The U.S.1.205 1.420 122 1.160 1.490 138 -0.045 Total1.019 1.421 259 0.748 1.591 286 -0.402* (0.03)
25 Results of Experiment (7): Evaluation Score of The U.S. Difficulty Reward for editing 20 80 Difference Easy 1.28 1.29 0.01 1.35 1.44 53 65 Difficult 1.14 1.04 0.1 1.48 1.53 69 73 Total 1.20 1.16 0.04 1.42 1.491 122 138
26 Results of Experiment (8): Evaluation Score of Japan Difficulty Reward for editing 20 80 Difference Easy 0.563 0.348 -0.215 1.553 1.653 41 68 Difficult 1.109 0.378 -0.731* 1.603 1.697 58 62 Total 0.854 0.365 -0.489* 1.578 1.669 99 130
27 Results of Experiment (9): Evaluation Score (3)
28 Results of Experiment (10): Evaluation Score of The U.S. Domain Reward for editing 20 80 Difference Religion 0.963 1.429 0.466 1.553 1.325 54 70 Politics 1.397 0.882 -0.515 * 1.603 1.697 68 66 Total 1.205 1.159 -0.046 1.578 1.669 122 138
29 Analysis of the results Hypothesis is rejected. Then How the participants evaluate the editing? →Spite or altruistic ? Valuation of the editing by outsiders or Can the market value of the editing be estimated? Evaluation of Evaluators
30 Possibility of the behavioral evaluation Suppose the evaluation function depends on not only the quality but also the rewards the editor gets. The evaluation score is based on both the effort function of the editor and the evaluation function of the other participants as an evaluator.
31 Future exercise How the participants evaluate the editing? Valuation of the editing by outsiders or Can the market value of the editing be estimated? Evaluation of Evaluators Spite or altruistic ?
Your consent to our cookies if you continue to use this website.