Generation of Referring Expressions (GRE) The Incremental Algorithm Dale & Reiter (1995)

Slides:



Advertisements
Similar presentations
Kees van Deemter Matthew Stone Formal Issues in Natural Language Generation Lecture 4 Shieber 1993; van Deemter 2002.
Advertisements

Generation of Referring Expressions: Managing Structural Ambiguities I.H. KhanG. Ritchie K. van Deemter University of Aberdeen, UK.
Some common assumptions behind Computational Generation of Referring Expressions (GRE) (Introductory remarks at the start of the workshop)
Kees van Deemter Matthew Stone Formal Issues in Natural Language Generation Lecture 5 Stone, Doran,Webber, Bleam & Palmer.
Generation of Referring Expressions (GRE) Reading: Dale & Reiter (1995) (key paper in this area)
Conceptual coherence in the generation of referring expressions Albert Gatt & Kees van Deemter University of Aberdeen {agatt,
Generation of Referring Expressions: the State of the Art SELLC Summer School, Harbin 2010 Kees van Deemter Computing Science University of Aberdeen.
Charting the Potential of Description Logic for the Generation of Referring Expression SELLC, Guangzhou, Dec Yuan Ren, Kees van Deemter and Jeff.
Generation of Referring Expressions: the State of the Art SELLC Winter School, Guangzhou 2010 Kees van Deemter Computing Science University of Aberdeen.
Generation of Referring Expressions (GRE) The Incremental Algorithm (IA) Dale & Reiter (1995)
Microplanning (Sentence planning) Part 1 Kees van Deemter.
CS4018 Formal Models of Computation weeks Computability and Complexity Kees van Deemter (partly based on lecture notes by Dirk Nikodem)
Lecture 3: Salience and Relations Reading: Krahmer and Theune (2002), in Van Deemter and Kibble (Eds.) “Information Sharing: Reference and Presupposition.
CHAPTER 2 ALGORITHM ANALYSIS 【 Definition 】 An algorithm is a finite set of instructions that, if followed, accomplishes a particular task. In addition,
Fast Algorithms For Hierarchical Range Histogram Constructions
Lecture 24 Coping with NPC and Unsolvable problems. When a problem is unsolvable, that's generally very bad news: it means there is no general algorithm.
Logic CPSC 386 Artificial Intelligence Ellen Walker Hiram College.
F22H1 Logic and Proof Week 7 Clausal Form and Resolution.
Infinite Horizon Problems
Planning under Uncertainty
CMPT 354, Simon Fraser University, Fall 2008, Martin Ester 52 Database Systems I Relational Algebra.
Proof methods Proof methods divide into (roughly) two kinds: –Application of inference rules Legitimate (sound) generation of new sentences from old Proof.
ECE 331 – Digital System Design
Constraint Logic Programming Ryan Kinworthy. Overview Introduction Logic Programming LP as a constraint programming language Constraint Logic Programming.
Tirgul 10 Rehearsal about Universal Hashing Solving two problems from theoretical exercises: –T2 q. 1 –T3 q. 2.
LEARNING FROM OBSERVATIONS Yılmaz KILIÇASLAN. Definition Learning takes place as the agent observes its interactions with the world and its own decision-making.
The Theory of NP-Completeness
CSE 830: Design and Theory of Algorithms
LEARNING DECISION TREES
Let remember from the previous lesson what is Knowledge representation

Monadic Predicate Logic is Decidable Boolos et al, Computability and Logic (textbook, 4 th Ed.)
Generating Referring Expressions (Dale & Reiter 1995) Ivana Kruijff-Korbayová (based on slides by Gardent&Webber, and Stone&van Deemter) Einfürung.
Information theory, fitness and sampling semantics colin johnson / university of kent john woodward / university of stirling.
Determining Sample Size
Called as the Interval Scheduling Problem. A simpler version of a class of scheduling problems. – Can add weights. – Can add multiple resources – Can ask.
Logics for Data and Knowledge Representation
Dr. Gary Blau, Sean HanMonday, Aug 13, 2007 Statistical Design of Experiments SECTION V SCREENING.
CS143 Review: Normalization Theory Q: Is it a good table design? We can start with an ER diagram or with a large relation that contain a sample of the.
Declarative vs Procedural Programming  Procedural programming requires that – the programmer tell the computer what to do. That is, how to get the output.
The Integers. The Division Algorithms A high-school question: Compute 58/17. We can write 58 as 58 = 3 (17) + 7 This forms illustrates the answer: “3.
Fall Week 4 CSCI-141 Scott C. Johnson.  Computers can process text as well as numbers ◦ Example: a news agency might want to find all the articles.
Slide 1 Propositional Definite Clause Logic: Syntax, Semantics and Bottom-up Proofs Jim Little UBC CS 322 – CSP October 20, 2014.
LOGIC AND ONTOLOGY Both logic and ontology are important areas of philosophy covering large, diverse, and active research projects. These two areas overlap.
RANLP, Borovets Sept Evaluating Algorithms for GRE (Going beyond Toy Domains) Ielka van der Sluis Albert Gatt Kees van Deemter University of.
1.1 CAS CS 460/660 Introduction to Database Systems Relational Algebra.
Albert Gatt LIN3021 Formal Semantics Lecture 4. In this lecture Compositionality in Natural Langauge revisited: The role of types The typed lambda calculus.
1 Markov Decision Processes Infinite Horizon Problems Alan Fern * * Based in part on slides by Craig Boutilier and Daniel Weld.
The Karnaugh Map.
Corpus-based evaluation of Referring Expression Generation Albert Gatt Ielka van der Sluis Kees van Deemter Department of Computing Science University.
MIPS ALU. Building from the adder to ALU ALU – Arithmetic Logic Unit, does the major calculations in the computer, including – Add – And – Or – Sub –
Machine Learning Concept Learning General-to Specific Ordering
1 Propositional Logic Limits The expressive power of propositional logic is limited. The assumption is that everything can be expressed by simple facts.
Chapter 11 Introduction to Computational Complexity Copyright © 2011 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. 1.
Charting the Potential of Description Logic for the Generation of Referring Expression SELLC, Guangzhou, Dec Yuan Ren, Kees van Deemter and Jeff.
Concept Learning and The General-To Specific Ordering
Computational Learning Theory Part 1: Preliminaries 1.
MIPS ALU. Exercise – Design a selector? I need a circuit that takes two input bits, a and b, and a selector bit s. The function is that if s=0, f=a. if.
Copyright © 2014 Curt Hill Algorithms From the Mathematical Perspective.
STROUD Worked examples and exercises are in the text Programme 10: Sequences PROGRAMME 10 SEQUENCES.
Kees van Deemter Generation of Referring Expressions: a crash course Background information and Project HIT 2010.
Lecture 9: Query Complexity Tuesday, January 30, 2001.
Computer Science 210 Computer Organization
Copyright © Cengage Learning. All rights reserved.
Algorithm Discovery and Design
Java Programming Loops
Truth tables.
Generation of Referring Expressions (GRE)
Presentation transcript:

Generation of Referring Expressions (GRE) The Incremental Algorithm Dale & Reiter (1995)

The task: GRE NLG can have different kinds of inputs: ‘Flat’ data (collections of atoms, e.g., in the tables of a database) Logically complex data In both cases, unfamiliar constants may be used, and this is sometimes unavoidable

No familiar constant available: 1. The referent has a familiar name, but it’s not unique, e.g., ‘John Smith’ 2. The referent has no familiar name: trains, furniture, trees, atomic particles, … ( In such cases, databases use database keys, e.g., ‘Smith$73527$’, ‘TRAIN-3821’ ) 3. Similar: sets of objects.

Natural Languages are too economic to have a proper name for everything Names may not even be most appropriate So, speakers/NLG systems have to invent ways of referring to things. E.g., ‘the 7:38 Trenton express’

Dale & Reiter: best description fulfils Gricean maxims. (Quality:) list properties truthfully (Quantity:) list sufficient properties to allow hearer to identify referent – but not more (Relevance:) use properties that are of interest in themselves * (Manner:) be brief * Slightly different from D&R 1995

D&R’s expectation: Violation of a maxim leads to implicatures. For example, [Quantity] ‘the black pitbull’ (when there is only one). [Manner] ‘Get the cordless drill that’s in the toolbox’ (Appelt). There’s just one problem: …

…people don’t speak this way For example, [Manner] ‘the red chair’ (when there is only one red object in the domain). [Quantity] ‘I broke my arm’ (when I have two). General: empirical work shows much redundancy Similar for other maxims, e.g., [Quality] ‘the man with the martini’ (Donellan)

Example Situation a, £100 b, £150 c, £100 d, £150 e, £? SwedishItalian

Formalized in a KB Type: furniture (abcde), desk (ab), chair (cde) Origin: Sweden (ac), Italy (bde) Colours: dark (ade), light (bc), grey (a) Price: 100 (ac), 150 (bd), 250 ({}) Contains: wood ({}), metal ({abcde}), cotton(d) Assumption: all this is shared knowledge.

Violations of … Manner: * ‘The £100 grey Swedish desk which is made of metal’ (Description of a) Relevance: ‘The cotton chair is a fire hazard? ?Then why not buy the Swedish chair?’ (Descriptions of d and c respectively)

Consider the following formalization: Full Brevity: Never use more than the minimal number of properties required for identification (Dale 1989) An algorithm:

Dale 1989: 1. Check whether 1 property is enough 2. Check whether 2 properties is enough …. Etc., until success {minimal description is generated} or failure {no description is possible}

Problem: exponential complexity Worst-case, this algorithm would have to inspect all combinations of properties. n properties combinations. Recall: one grain of rice on square one; twice as many on any subsequent square. Some algorithms may be faster, but … Theoretical result: algorithm must be exponential in the number of properties.

D&R conclude that Full Brevity cannot be achieved in practice. They designed an algorithm that only approximates Full Brevity: the Incremental Algorithm (IA).

Psycholinguistic inspiration behind IA (e.g. Pechmann 89; overview in Levelt 89) Speakers often include “unnecessary modifiers” in their referring expressions Speakers often start describing a referent before they have seen all distractors (as shown by eye-tracking experiments) Some Attributes (e.g. Colour) seem more likely to be noticed and used than others Some Attributes (e.g., Type) contribute strongly to a Gestalt. Gestalts help readers identify referents. (“The red thing” vs. “the red bird”) Let’s start with a simplified version of IA, which uses properties rather than pairs. Type and head nouns are ignored, for now.

Incremental Algorithm (informal): Properties are considered in a fixed order: P = A property is included if it is ‘useful’: true of target; false of some distractors Stop when done; so earlier properties have a greater chance of being included. (E.g., a perceptually salient property) Therefore called preference order.

r = individual to be described P = list of properties, in preference order P is a property L= properties in generated description (Recall: we’re not worried about realization today)

P = < desk (ab), chair (cde), Swedish (ac), Italian (bde), dark (ade), light (bc), grey (a), 100£ ({ac}), 150£(bd), 250£ ({}), wooden ({}), metal (abcde), cotton ({d}) > Domain = {a,b,c,d,e}. Now describe: a = d = e =

P = < desk (ab), chair (cde), Swedish (ac), Italian (bde), dark (ade), light (bc), grey (a), 100£ (ac),200£ (bd),250£ ({}), wooden ({}), metal (abcde), cotton (d) > Domain = {a,b,c,d,e}. Now describe: a = d = (Nonminimal) e = (Impossible)

Incremental Algorithm It’s a hillclimbing algorithm: ever better approximations of a successful description. ‘Incremental’ implies no backtracking. Not always the minimal number of properties.

Incremental Algorithm Logical completeness: A unique description is found in finite time if there exists one. (Given reasonable assumptions, see van Deemter 2002) Computational complexity: Assume that testing for usefulness takes constant time. Then worst-case time complexity is O(n p ) where n p is the number of properties in P.

Better approximation of Full Brevity (D&R 1995) Attribute + Value model: Properties grouped together as in original example: Origin: Sweden, Italy,... Colour: dark, grey,... Optimization within the set of properties based on the same Attribute

Incremental Algorithm, using Attributes and Values r = individual to be described A = list of Attributes, in preference order Def: = Value j of Attribute i L= properties in generated description

FindBestValue(r,A): - Find Values of A that are true of r, while removing some distractors (If these don’t exist, go to next Attribute) - Within this set, select the Value that removes the largest number of distractors - If there’s a tie, select the most general one - If there’s still a tie, select an arbitrary one

Example: D = {a,b,c,d,f,g} Type: furniture (abcd), desk (ab), chair (cd) Origin: Europe (bdfg), USA (ac), Italy (bd) Describe a: {desk, American} (furniture removes fewer distractors than desk) Describe b: {desk, European} (European is more general than Italian) N.B. This disregards relevance, etc.

Exercise on Logical Completeness: Construct an example where no description is found, although one exists. Hint: Let Attribute have Values whose extensions overlap.

Example: D = {a,b,c,d,e,f} Contains: wood (abe), plastic (acdf) Colour: grey (ab), yellow (cd) Describe a: {wood, grey,...} - Failure (wood removes more distractors than plastic) Compare: Describe a: {plastic, grey} - Success

Complexity of the algorithm n d = nr. of distractors n l = nr. of properties in the description n v = nr. of Values ( for all Attributes ) According to D&R: O(n d n l ) (Typical running time) Alternative assessment: O(n v ) (Worst-case running time)

Minor complication: Head nouns Another way in which human descriptions are nonminimal A description needs a Noun, but not all properties are expressed as Nouns Example: Suppose Colour was the most-preferred Attribute, and suppose target = a

Colours: dark (ade), light (bc), grey (a) Type: furniture (abcde), desk (ab), chair (cde) Origin: Sweden (ac), Italy (bde) Price: 100 (ac), 150 (bd), 250 ({}) Contains: wood ({}), metal ({abcde}), cotton(d) target = a Describe a: {grey} ‘The grey’ ? (Not in English)

D&R’s repair: Assume that Values of the Attribute Type can be expressed in a Noun. After the core algorithm: - check whether Type is represented. - if not, then add the best Value of the Type Attribute to the description

Versions of Dale and Reiter’s Incremental Algorithm (IA) have often been implemented Still the starting point for many new algorithms. But how human-like is the output of the IA really? The paper does not contain an evaluation of the algorithms discussed

Limitations of the algorithm 1. Redundancy exists, but not for principled reasons, e.g., for - marking topic changes, etc. (Corpus work by Pam Jordan et. al.) - making it easy to find the referent (Experimental work by Paraboni et al.)

Limitations of the algorithm 2. Targets are individual objects, never sets. What changes when target = {a,b,c} ? 3. Incremental algorithm uses only conjunctions of atomic properties. No negations, disjunctions, etc.

Limitations of D&R 4. No relations with other objects, e.g., ‘the orange on the table’. 5. Differences in salience are not taken into account. When we say “the dog”, does this mean that there is only one dog in the world? 6. Language realization is disregarded. 7. Logical completeness is violated when Attribute has overlapping Values

Limitations of D&R 8. Calculation of complexity is iffy Role of “Typical” run time and length of description is unclear Greedy Algorithm (GA) dismissed even though it has polynomial complexity GA: always choose the property that removes the maximum number of distractors

More fundamental assumptions Speaker and Hearer have shared knowledge This knowledge can be formalised using atomic statements (plus, implicitly, negations of atomic statements) The aim of GRE is to identify the target referent uniquely. (I.e., the aim is to construct a “distinguishing description” of the referent.) Linguistic Realisation comes after Content Determination

Discussion: How bad is it for a GRE algorithm to take exponential time choosing the best RE? How do human speakers cope? More complex types of referring expressions problem becomes even harder Restrict to combinations whose length is less than x problem not exponential. Example: descriptions containing at most n properties (Full Brevity)

Linguist’s view We don’t pretend to mirror psychologically correct processes. (It’s enough if GRE output is correct). So why worry if our algorithms are slow?

Mathematicians’ view structure of a problem becomes clear when no restrictions are put. Practical addition: What if the input does not conform with these restrictions? (GRE does not control its own input!)

A compromise view Compare with Description Logic: - Increasingly complex algorithms … - that tackle larger and larger fragments of logic … - and whose complexity is ‘conservative’ When looking at more complex phenomena, take care not to slow down generation of simple cases too much