Microplanning (Sentence planning) Part 1 Kees van Deemter.

Slides:



Advertisements
Similar presentations
1 Knowledge Representation Introduction KR and Logic.
Advertisements

Kees van Deemter Matthew Stone Formal Issues in Natural Language Generation Lecture 4 Shieber 1993; van Deemter 2002.
A very short introduction to Natural Language Generation Kees van Deemter Computing Science University of Aberdeen.
Generation of Referring Expressions: Managing Structural Ambiguities I.H. KhanG. Ritchie K. van Deemter University of Aberdeen, UK.
Some common assumptions behind Computational Generation of Referring Expressions (GRE) (Introductory remarks at the start of the workshop)
Kees van Deemter Matthew Stone Formal Issues in Natural Language Generation Lecture 5 Stone, Doran,Webber, Bleam & Palmer.
Generation of Referring Expressions (GRE) Reading: Dale & Reiter (1995) (key paper in this area)
Conceptual coherence in the generation of referring expressions Albert Gatt & Kees van Deemter University of Aberdeen {agatt,
Generation of Referring Expressions: the State of the Art SELLC Summer School, Harbin 2010 Kees van Deemter Computing Science University of Aberdeen.
Charting the Potential of Description Logic for the Generation of Referring Expression SELLC, Guangzhou, Dec Yuan Ren, Kees van Deemter and Jeff.
Kees van Deemter Matthew Stone Formal Issues in Natural Language Generation Lecture 1 Overview + Reiter & Dale 1997.
Generation of Referring Expressions (GRE) The Incremental Algorithm (IA) Dale & Reiter (1995)
A small taste of inferential statistics
CS4018 Formal Models of Computation weeks Computability and Complexity Kees van Deemter (partly based on lecture notes by Dirk Nikodem)
Dr Casey Wilson, 2009 Panels and Reviews. 1 st year Panels Dr C. Wilson, 2009 Format: (check details with your Dept) Chair, supervisor(s) and at least.
Lecture 3: Salience and Relations Reading: Krahmer and Theune (2002), in Van Deemter and Kibble (Eds.) “Information Sharing: Reference and Presupposition.
Is it Mathematics? Linking to Content Standards. Some questions to ask when looking at student performance Is it academic? – Content referenced: reading,
Lecture 24 MAS 714 Hartmut Klauck
C O N T E X T - F R E E LANGUAGES ( use a grammar to describe a language) 1.
Week 11 Review: Statistical Model A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution.
December 2003CSA3050: Natural Language Generation 1 What is Natural Language Generation? When is NLG an Appropriate Technology? NLG System Architectures.
The Cooperative Principle
1 MODULE 2 Meaning and discourse in English COOPERATION, POLITENESS AND FACE Lecture 14.
EL1101E WEEK 10: PRAGMATICS Group members: Elaine Ong Ong Min Thakshayeni Skanthakumar Jeannie Poon.
On Status and Form of the Relevance Principle Anton Benz, ZAS Berlin Centre for General Linguistics, Typology and Universals Research.
Albert Gatt LIN1180 – Semantics Lecture 10. Part 1 (from last week) Theories of presupposition: the semantics- pragmatics interface.
Natural Language Generation: Discourse Planning
Chapter 20: Natural Language Generation Presented by: Anastasia Gorbunova LING538: Computational Linguistics, Fall 2006 Speech and Language Processing.
Natural Language Generation Research Presentation Presenter Shamima Mithun.
Generation of Referring Expressions: Modeling Partner Effects Surabhi Gupta Advisor: Amanda Stent Department of Computer Science.
1 Introduction to Computability Theory Lecture12: Reductions Prof. Amos Israeli.
Proof Points Key ideas when proving mathematical ideas.
People & Speech Interfaces CS 260 Wednesday, October 4, 2006.
Discourse Analysis The study of language inside conversations.
Tirgul 10 Rehearsal about Universal Hashing Solving two problems from theoretical exercises: –T2 q. 1 –T3 q. 2.
Natural Language Generation Ling 571 Fei Xia Week 8: 11/17/05.
1 The Sample Mean rule Recall we learned a variable could have a normal distribution? This was useful because then we could say approximately.
Lecture 1 Introduction: Linguistic Theory and Theories
Generating Referring Expressions (Dale & Reiter 1995) Ivana Kruijff-Korbayová (based on slides by Gardent&Webber, and Stone&van Deemter) Einfürung.
Homework – Day 1 Read all of Chapter 1. As you read, answer the following questions. 1. Define economics. 2. Explain the “economic way of thinking,” including.
Pragmatics.
Semantics 3rd class Chapter 5.
Copyright © 2010 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Some Basics of Algebra Algebraic Expressions and Their Use Translating to.
SOFTWARE ENGINEERING BIT-8 APRIL, 16,2008 Introduction to UML.
Game Theory and Grice’ Theory of Implicatures Anton Benz.
Lecture 2(b) Rational Choice and Demand. Why It Would Probably Be Ok to Sleep Through This Part of the Lecture The previous lecture described almost everything.
Department of English Introduction To Linguistics Level Four Dr. Mohamed Younis.
Natural Information and Conversational Implicatures Anton Benz.
Slide 1 Propositional Definite Clause Logic: Syntax, Semantics and Bottom-up Proofs Jim Little UBC CS 322 – CSP October 20, 2014.
Unit 8 Why don’t you get her a scarf?. Theoretical Basis The Analysis of Teaching Material Teaching Method and Studying Ways Teaching procedures Blackboard.
Generation of Referring Expressions (GRE) The Incremental Algorithm Dale & Reiter (1995)
Pragmatics.
Welcome to MM204! Unit 6 Seminar To resize your pods: Place your mouse here. Left mouse click and hold. Drag to the right to enlarge the pod. To maximize.
LECTURE 2: SEMANTICS IN LINGUISTICS
Sight Word List.
Corpus-based evaluation of Referring Expression Generation Albert Gatt Ielka van der Sluis Kees van Deemter Department of Computing Science University.
Welcome Back, Folks! We’re travelling to a littele bit far-end of Language in Use Studies EAA remains your faithful companion.
UNIT 2 - IMPLICATURE.
1Computer Sciences Department. Book: INTRODUCTION TO THE THEORY OF COMPUTATION, SECOND EDITION, by: MICHAEL SIPSER Reference 3Computer Sciences Department.
Optimal answers and their implicatures A game-theoretic approach Anton Benz April 18 th, 2006.
Unit 1 Lifelong Education. Objectives: express requests and offers fluently and properly; use proper words and expressions to talk about the plan for.
Charting the Potential of Description Logic for the Generation of Referring Expression SELLC, Guangzhou, Dec Yuan Ren, Kees van Deemter and Jeff.
Kees van Deemter Generation of Referring Expressions: a crash course Background information and Project HIT 2010.
Chapter 6 Guidelines for Modelling. 1. The Modelling Process 1. Modelling as a Transformation Process 2. Basic Modelling Activities 3. Types of Modelling.
Chapter 2 1. Chapter Summary Sets (This Slide) The Language of Sets - Sec 2.1 – Lecture 8 Set Operations and Set Identities - Sec 2.2 – Lecture 9 Functions.
Interpreting as Process
COOPERATIVE PRINCIPLE:
Truth tables.
Generation of Referring Expressions (GRE)
Presentation transcript:

Microplanning (Sentence planning) Part 1 Kees van Deemter

Natural Language Generation Taking some computer-readable gibberish Translating it into proper English Applications include –dialogue/chat systems –on-line help –summarisation, –document authoring

NLG Tasks (as explained by Anja): 1.Content determination: decide what to say; construct set of messages 2.Discourse planning: ordering, structuring concepts; rhetorical relationships 3.Sentence aggregation: divide content into sentences; construct sentence plans 4.Lexicalisation: map concepts and relations to lexemes (= words) 5.Referring expression generation: decide how to refer to objects 6.Linguistic realisation: put it all together in acceptable words and sentences

Modular structure of NLG systems (in theory!): Content determination Discourse planning Sentence aggregation Realisation Lexicalisation Referring expressions TEXT PLANNER REALISER SENTENCE PLANNER/ MICROPLANNER

Last week: Input to realisation message-id: msg02 relation: C_DEPARTURE departing-entity: C_CALEDON-EXPRESS args: departure-location: C_ABERDEEN departure-time: C_1000 departure-platform: C_7

Microplanning 1:Aggregation Distributing information over different sentences. Example: a. The Caledonian express departs Aberdeen at 10:00, from platform 7 b. The Caledonian express departs Aberdeen at 10:00. The Caledonia express departs from platform 7

Microplanning 2: GRE GRE = Generation of Referring Expressions Explaining which objects youre talking about a. The Caledonian express departs Aberdeen at 10:00, from platform 7 b. The Caledonian express departs -- at 10:00. The train departs from this platform

Microplanning 3: lexical choice Using different words for the same concept a. The Caledonian express departs Aberdeen at ten oclock, from platform 7 b. The Caledonian express departs Aberdeen at ten. The Caledonia express leaves from platform 7

In practice: tasks can be performed in different order Example: aggregation can be performed on messages:

message-id: msg02 relation: C_DEPARTURE_1 departing-entity: C_CALEDON-EXPRESS args: departure-location: C_ABERDEEN departure-time: C_1000 message-id: msg03 relation: C_DEPARTURE_2 args: departure-entity: C_CALEDON-EXPRESS departure-platform: C_7

Aggregation can also be performed later: [ The Caledonian express] departs Aberdeen [at 10:00] [from platform 7] ===> [The Caledonian express] departs Aberdeen [at 10:00]. [The Caledonia express] departs [from platform 7]

Lets focus on GRE, but... A little detour: NLG systems do not always work as youve been told Some practically deployed systems combine canned text with NLG One possibility: system has a library of language templates, with gaps that need to be filled. E.g.,

[TRAIN] departs [TOWN] at [TIME] [TRAIN] departs [TOWN] from [PLATFORM] We apologise for the fact that [TRAIN] is delayed by [AMOUNT] Gap filling: using canned text or GRE. Question: which of the other tasks are still relevant?

Lets move on to GRE Why/when is GRE useful?

1.The referent has a familiar name, but its not unique, e.g., John Smith 2.The referent has no familiar name: trains, furniture, trees, atomic particles, … ( Databases use keys, e.g., Smith$73527$, TRAIN-3821 ) 3. Similar: sets of objects 4. NL is too economical to have names for everything

Last week: Input to realisation message-id: msg02 relation: C_DEPARTURE departing-entity: C_CALEDON-EXPRESS args: departure-location: C_ABERDEEN departure-time: C_1000

Last week: Input to realisation message-id: msg02 relation: C_DEPARTURE departing-entity: C_CALEDON-EXPRESS args: departure-location: C_ABERDEEN departure-time: C_1000

This week: more realistic input message-id: msg02 relation: C_DEPARTURE departing-entity: C_34435 args: departure-location:..... departure-time:..... the caledonian (express), the Aberdeen-Glasgow express the blue train on your left, the train

Communication is about saying the truth... but thats not all there is to it Paul Grice (around 1970): principles of rational, cooperative communication GRE, it a good case study. (R.Dale and E.Reiter, Cognitive Science, 1995)

Grice: maxims of conversation Quality: only say what you know to be true Quantity: give enough but not too much information Relevance: be relevant Manner: be clear and brief (There is overlap between these four)

Maxims are two-edged sword: 1.They say how one should normally speak/write. Example: Yes, theres a gasoline station around the corner (when its no longer operational) quality: yes, its true quantity: probably yes relevance: no, not relevant to hearers intentions manner: its brief, clear, etc.

Maxims are two-edged sword: 2. They can also be exploited. Example: Asked to write academic reference: Kees always came to my lectures and hes a nice guy quality: yes, its true (lets assume) quantity: No -- How about academic achievements? relevance: yes manner: yes

Application to GRE Dale & Reiter: best description of an object fulfils the Gricean maxims. E.g., (Quality:) list properties truthfully (Quantity:) use properties that allow identification – without containing more info (Relevance:) use properties that are of interest in the situation (Manner:) be brief

D&Rs expectation: Violation of a maxim leads to implicatures. For example, – [Quantity] the pitbull (when there is only one dog). –[Manner] Get the cordless drill thats in the toolbox (Appelt). Theres just one problem: …

people dont always speak this way For example, –[Manner] the red chair (when there is only one red object in the domain). –[Manner/Quantity] I broke my arm (when I have two). General: empirical work shows much redundancy Similar for other maxims, e.g., –[Quality] the man with the martini (Donellan)

Example Situation a, £100 b, £150 c, £100 d, £150 e, £? SwedishItalian

Formalized in a KB Type: furniture (abcde), desk (ab), chair (cde) Origin: Sweden (ac), Italy (bde) Colours: dark (ade), light (bc), grey (a) Price: 100 (ac), 150 (bd), 250 ({}) Contains: wood ({}), metal (abcde), cotton(d) Assumption: all this is shared knowledge.

Game 1. Describe object a. 2. Describe object e. 3. Describe object d.

Game 1. Describe object a: {desk,sweden}, {grey} 2. Describe object e: no solution 3. Describe object d: {Italy, 150}

Violations of … Manner: *The £100 grey Swedish desk which is made of metal (Description of a) Relevance: The cotton chair is a fire hazard? ?Then why not buy the Swedish chair? (Descriptions of d and c respectively)

In fact, there is a second problem with Quantity/Manner. Consider the following formalization: Full Brevity: Never use more than the minimal number of properties required for identification (Dale 1989) An algorithm:

Dale 1989: 1.Check whether 1 property is enough 2.Check whether 2 properties is enough …. Etc., until success {minimal description is generated} or failure {no description is possible}

Problem: exponential complexity Worst-case, this algorithm would have to inspect all combinations of properties. n properties combinations. Recall: one grain of rice on square one; twice as many on any subsequent square. Some algorithms may be faster, but … Theoretical result: algorithm must be exponential in the number of properties.

D&R conclude that Full Brevity cannot be achieved in practice. They designed an algorithm that only approximates Full Brevity: the Incremental Algorithm.

Incremental Algorithm (informal): Properties are considered in a fixed order: P = A property is included if it is useful: true of target; false of some distractors Stop when done; so earlier properties have a greater chance of being included. (E.g., a perceptually salient property) Therefore called preference order.

r = individual to be described P = list of properties, in preference order P is a property L= properties in generated description (Recall: were not worried about realization today)

Back to the KB Type: furniture (abcde), desk (ab), chair (cde) Origin: Sweden (ac), Italy (bde) Colours: dark (ade), light (bc), grey (a) Price: 100 (ac), 150 (bd), 250 ({}) Contains: wood ({}), metal (abcde), cotton(d) Assumption: all this is shared knowledge.

Back to our game 1. Describe object a. 2. Describe object e. 3. Describe object d. Can you see room for improvement?