Presentation is loading. Please wait.

Presentation is loading. Please wait.

Exploiting Subjective Annotations Dennis Reidsma and Rieks op den Akker Human Media Interaction University of Twente

Similar presentations


Presentation on theme: "Exploiting Subjective Annotations Dennis Reidsma and Rieks op den Akker Human Media Interaction University of Twente"— Presentation transcript:

1 Exploiting Subjective Annotations Dennis Reidsma and Rieks op den Akker Human Media Interaction University of Twente http://hmi.ewi.utwente.nl

2 Types of content Annotation as a task of subjective judgments? Manifest content Pattern latent content Projective latent content Cf. Potter and Levine-Donnerstein 1999

3 Projective latent content Why annotate data as projective latent content? Because it cannot be defined exhaustively, whereas annotators have good `mental schema’s’ for it Because the data should be annotated in terms that fit with the understanding of `naïve users’

4 Inter-annotator agreement and projective content Disagreements may be caused by Errors by annotators Invalid scheme (no true label exists) Different annotators having different `truths’ in interpretation of behavior (subjectivity)

5 Subjective annotation People communicate in different ways, and therefore, as an observer, may also judge the behavior of others differently

6 Subjective annotation People communicate in different ways, and therefore, as an observer, may also judge the behavior of others differently Projective content may be especially vulnerable to this problem

7 Subjective annotation People communicate in different ways, and therefore, as an observer, may also judge the behavior of others differently Projective content may be especially vulnerable to this problem How to work with subjectively annotated data?

8 Subjective annotation How to work with subjectively annotated data? Unfortunately, it leads to low levels of agreement, and therefore usually would be avoided as `unproductive material’

9 I. Predicting agreement One way to work with subjective data is to try to find out in which contexts annotators would agree, and focus on those situations. Result: a classifier that will not always classify all instances, but if it does, it will do so with greater accuracy

10 II. Explicitly modeling intersubjectivity A second way: model different annotators separately, then find the cases where the models agree, and assume that those are the cases where the annotators would have agreed, too. Result: a classifier that tells you for which instances other annotators would most probably agree with its classification

11 Advantages Both solutions lead to `cautious classifiers’ that only render a judgment in those cases where annotators would have been expected to agree This may carry over to users, too… Neither solution needs to have all data multiply annotated for this

12 Time?

13 Pressing questions so far? (The remainder of the talk will give two case studies.)

14 Case studies I. Predicting agreement from information in other (easier) modalities: The case of contextual addressing II. Explicitly modeling intersubjectivity in dialog markup: The case of Voting Classifiers

15 Data used: The AMI Corpus 100h of recorded meetings, annotated with dialog acts, focus of attention, gestures, addressing, decision points, and other layers

16 I. Contextual addressing Addressing, and focus of attention. Agreement is highest for certain FOA contexts. In those contexts, the classifier also performed better. … more in paper

17 II. Modeling intersubjectivity Modeling single annotators, for `yeah’ utterances Data annotated non-overlapping, 3 annotators All data dsv Trn (3585) Tst (2289) Trn (1753) Tst (528) Trn (3500) Tst (1362)

18 II. Modeling intersubjectivity Cross annotator training and testing TST_dTST_sTST_vTST_all C_d69645263 C_s59684857 C_v63576663

19 II. Modeling intersubjectivity Building a voting classifier: Only classify an instance when all three annotator-specific expert classifiers agree

20 II. Modeling intersubjectivity In the unanimous voting context, performance is higher due to increased precision (avg 6%)

21 Conclusions Possible subjective aspects to annotation should be taken into account Agreement metrics are not designed to handle this We proposed two methods designed to cope with subjective data

22 Thank you! Questions?

23

24

25


Download ppt "Exploiting Subjective Annotations Dennis Reidsma and Rieks op den Akker Human Media Interaction University of Twente"

Similar presentations


Ads by Google