Presentation is loading. Please wait.

Presentation is loading. Please wait.

David L. Chen Supervisor: Professor Raymond J. Mooney Ph.D. Dissertation Defense January 25, 2012 Learning Language from Ambiguous Perceptual Context.

Similar presentations


Presentation on theme: "David L. Chen Supervisor: Professor Raymond J. Mooney Ph.D. Dissertation Defense January 25, 2012 Learning Language from Ambiguous Perceptual Context."— Presentation transcript:

1 David L. Chen Supervisor: Professor Raymond J. Mooney Ph.D. Dissertation Defense January 25, 2012 Learning Language from Ambiguous Perceptual Context

2 Semantics of Language The meaning of words, phrases, etc Learning semantics of language is one of the ultimate goals in natural language processing The meanings of many words are grounded in our perception of the physical world: red, ball, cup, run, hit, fall, etc. [Harnad, 1990] Computer representation should also be grounded in real world perception 1

3 Grounding Language Spanish goalkeeper Casillas blocks the ball 2

4 Grounding Language Block(Casillas) Spanish goalkeeper Casillas blocks the ball 3

5 Natural Language and Meaning Representation Spanish goalkeeper Casillas blocks the ball Block(Casillas) 4

6 Natural Language and Meaning Representation Natural Language (NL) NL: A language that has evolved naturally, such as English, German, French, Chinese, etc Block(Casillas) Spanish goalkeeper Casillas blocks the ball 4

7 Natural Language and Meaning Representation NL: A language that has evolved naturally, such as English, German, French, Chinese, etc MRL: Formal languages such as logic or any computer-executable code Meaning Representation Language (MRL) Block(Casillas) Spanish goalkeeper Casillas blocks the ball Natural Language (NL) 4

8 Semantic Parsing and Surface Realization NL Semantic Parsing: maps a natural-language sentence to a complete, detailed semantic representation MRL Semantic Parsing (NL  MRL) Block(Casillas) Spanish goalkeeper Casillas blocks the ball 5

9 Semantic Parsing and Surface Realization NL Semantic Parsing: maps a natural-language sentence to a complete, detailed semantic representation Surface Realization: Generates a natural-language sentence from a meaning representation. MRL Semantic Parsing (NL  MRL) Surface Realization (NL  MRL) Block(Casillas) Spanish goalkeeper Casillas blocks the ball 5

10 Applications Natural language interface – Issue commands and queries in natural language – Computer responds with answer in natural language Knowledge acquisition Computer assisted tasks 6

11 Traditional Learning Approach Manually Annotated Training Corpora (NL/MRL pairs) Semantic Parser MRLNL Semantic Parser Learner 7

12 Example of Annotated Training Corpus Natural Language (NL) Meaning Representation Language (MRL) Alice passes the ball to BobPass(Alice, Bob) Bob turns the ball over to JohnTurnover(Bob, John) John passes to FredPass(John, Fred) Fred shoots for the goalKick(Fred) Paul blocks the ballBlock(Paul) Paul kicks off to NancyPass(Paul, Nancy) …… 8

13 Learning Language from Perceptual Context Constructing annotated corpora for language learning is difficult Children acquire language through exposure to linguistic input in the context of a rich, relevant, perceptual environment Ideally, a computer system can learn language in the same manner 9

14 Navigation Example 在餐廳的地方右轉 Alice: Bob 10

15 Navigation Example 在醫院的地方右轉 Alice: Bob 11

16 Navigation Example 在醫院的地方右轉 在餐廳的地方右轉 Scenario 1 Scenario 2 12

17 Navigation Example 在 的地方右轉 Scenario 1 Scenario 2 13

18 Navigation Example 在 的地方右轉 Scenario 1 Scenario 2 Make a right turn 13

19 Navigation Example 在醫院的地方右轉 在餐廳的地方右轉 Scenario 1 Scenario 2 14

20 Navigation Example 餐廳 Scenario 1 Scenario 2 15

21 Navigation Example 在醫院的地方右轉 在餐廳的地方右轉 Scenario 1 Scenario 2 16

22 Navigation Example 醫院 Scenario 1 Scenario 2 17

23 Thesis Contributions Learning from limited ambiguity – Sportscasting task [Chen & Mooney, ICML, 2008; Chen, Kim & Mooney, JAIR, 2010] – Collected English corpus of 2000 sentences that has been adopted by other researchers [Liang et al., 2009; Bordes, Usunier, & Weston, 2010; Angeli et al., 2010; Hajishirzi, Hockenmaier, Muellar, & Amir, 2011] Learning from ambiguously supervised relational data – Navigation task [Chen & Mooney, AAAI, 2011] – Organized and semantically annotated corpus from MacMahon et al. (2006) 18

24 Overview Background Learning from limited ambiguity Learning from ambiguous relational data Future directions Conclusions 19

25 Semantic Parser Learners Learn a function from NL to MR NL: “Purple3 passes the ball to Purple5” MR: Pass ( Purple3, Purple5 ) Semantic Parsing (NL  MR) Surface Realization (MR  NL) We experiment with two semantic parser learners –WASP [Wong & Mooney, 2006; 2007] –KRISP [Kate & Mooney, 2006] 20

26 Uses statistical machine translation techniques – Synchronous context-free grammars (SCFG) [Wu, 1997; Melamed, 2004; Chiang, 2005] – Word alignments [Brown et al., 1993; Och & Ney, 2003] Capable of both semantic parsing and surface realization WASP: Word Alignment-based Semantic Parsing 21

27 KRISP: Kernel-based Robust Interpretation by Semantic Parsing Productions of MR language are treated like semantic concepts SVM classifier is trained for each production with string subsequence kernel These classifiers are used to compositionally build MRs of the sentences More resistant to noisy supervision but incapable of language generation 22

28 Overview Background Learning from limited ambiguity Learning from ambiguous relational data Future directions Conclusions 23

29 Tractable Challenge Problem: Learning to Be a Sportscaster Goal: Learn from realistic data of natural language used in a representative context while avoiding difficult issues in computer perception (i.e. speech and vision). Solution: Learn from textually annotated traces of activity in a simulated environment. Example: Traces of games in the RoboCup simulator paired with textual sportscaster commentary. 24

30 Sample Human Sportscast in Korean 25

31 Robocup Sportscaster Trace Natural Language CommentaryMeaning Representation Purple goalie turns the ball over to Pink8 badPass ( Purple1, Pink8 ) Pink11 looks around for a teammate Pink8 passes the ball to Pink11 Purple team is very sloppy today Pink11 makes a long pass to Pink8 Pink8 passes back to Pink11 turnover ( Purple1, Pink8 ) pass ( Pink11, Pink8 ) pass ( Pink8, Pink11 ) ballstopped pass ( Pink8, Pink11 ) kick ( Pink11 ) kick ( Pink8) kick ( Pink11 ) kick ( Pink8 ) 26

32 Robocup Sportscaster Trace Natural Language CommentaryMeaning Representation Purple goalie turns the ball over to Pink8 badPass ( Purple1, Pink8 ) Pink11 looks around for a teammate Pink8 passes the ball to Pink11 Purple team is very sloppy today Pink11 makes a long pass to Pink8 Pink8 passes back to Pink11 turnover ( Purple1, Pink8 ) pass ( Pink11, Pink8 ) pass ( Pink8, Pink11 ) ballstopped pass ( Pink8, Pink11 ) kick ( Pink11 ) kick ( Pink8) kick ( Pink11 ) kick ( Pink8 ) 26

33 Robocup Sportscaster Trace Natural Language CommentaryMeaning Representation Purple goalie turns the ball over to Pink8 badPass ( Purple1, Pink8 ) Pink11 looks around for a teammate Pink8 passes the ball to Pink11 Purple team is very sloppy today Pink11 makes a long pass to Pink8 Pink8 passes back to Pink11 turnover ( Purple1, Pink8 ) pass ( Pink11, Pink8 ) pass ( Pink8, Pink11 ) ballstopped pass ( Pink8, Pink11 ) kick ( Pink11 ) kick ( Pink8) kick ( Pink11 ) kick ( Pink8 ) 26

34 Robocup Sportscaster Trace Natural Language CommentaryMeaning Representation Purple goalie turns the ball over to Pink8 P6 ( C1, C19 ) Pink11 looks around for a teammate Pink8 passes the ball to Pink11 Purple team is very sloppy today Pink11 makes a long pass to Pink8 Pink8 passes back to Pink11 P5 ( C1, C19 ) P2 ( C22, C19 ) P2 ( C19, C22 ) P0 P2 ( C19, C22 ) P1 ( C22 ) P1( C19 ) P1 ( C22 ) P1 ( C19 ) 27

35 Robocup Data Collected human textual commentary for the 4 Robocup championship games from 2001-2004. –Avg # events/game = 2,613 –Avg # English sentences/game = 509 –Avg # Korean sentences/game = 499 Each sentence matched to all events within previous 5 seconds. –Avg # MRs/sentence = 2.5 (min 1, max 12) Manually annotated with correct matchings of sentences to MRs (for evaluation purposes only). 28

36 Machine Sportscast in English 29

37 Overview Background Learning from limited ambiguity Learning from ambiguous relational data Future directions Conclusions 30

38 Navigation Task Learn to interpret and follow free-form navigation instructions – e.g. Go down this hall and make a right when you see an elevator to your left Assume no prior linguistic knowledge Learn by observing how humans follow instructions Use virtual worlds and instructor/follower data from MacMahon et al. (2006) 31

39 H C L S S B C H E L E Environment H – Hat Rack L – Lamp E – Easel S – Sofa B – Barstool C - Chair 32

40 Environment 33

41 Sample Instructions Take your first left. Go all the way down until you hit a dead end. Go towards the coat hanger and turn left at it. Go straight down the hallway and the dead end is position 4. Walk to the hat rack. Turn left. The carpet should have green octagons. Go to the end of this alley. This is p-4. Walk forward once. Turn left. Walk forward twice. Start 3 3 H H 4 4 End 34

42 Sample Instructions 3 3 H H 4 4 Take your first left. Go all the way down until you hit a dead end. Go towards the coat hanger and turn left at it. Go straight down the hallway and the dead end is position 4. Walk to the hat rack. Turn left. The carpet should have green octagons. Go to the end of this alley. This is p-4. Walk forward once. Turn left. Walk forward twice. Observed primitive actions: Forward, Left, Forward, Forward Start End 34

43 Formal Problem Definition Given: { (e 1, a 1, w 1 ), (e 2, a 2, w 2 ), …, (e n, a n, w n ) } e i – A natural language instruction a i – An observed action sequence w i – A world state Goal: Build a system that produces the correct a j given a previously unseen (e j, w j ). 35

44 Challenges Many different instructions for the same task – Describe different actions – Use different parameters – Different ways to describe the same parameters Hidden plans – Needs to infer the plans from observed actions Exponential number of possible plans 36

45 Observation Instruction World State Training Action Trace 37

46 Learning system for parsing navigation instructions Learning system for parsing navigation instructions Observation Instruction World State Training Action Trace Navigation Plan Constructor 37

47 Learning system for parsing navigation instructions Learning system for parsing navigation instructions Observation Instruction World State Training Action Trace Navigation Plan Constructor Semantic Parser Learner 37

48 Learning system for parsing navigation instructions Learning system for parsing navigation instructions Observation Instruction World State Training Action Trace Navigation Plan Constructor Semantic Parser Learner Plan Refinement 37

49 Learning system for parsing navigation instructions Learning system for parsing navigation instructions Observation Instruction World State Instruction World State Training Testing Action Trace Navigation Plan Constructor Semantic Parser Learner Plan Refinement 37

50 Learning system for parsing navigation instructions Learning system for parsing navigation instructions Observation Instruction World State Instruction World State Training Testing Action Trace Navigation Plan Constructor Semantic Parser Learner Plan Refinement Semantic Parser 37

51 Learning system for parsing navigation instructions Learning system for parsing navigation instructions Observation Instruction World State Execution Module (MARCO) Instruction World State Training Testing Action Trace Navigation Plan Constructor Semantic Parser Learner Plan Refinement Semantic Parser Action Trace 37

52 Learning system for parsing navigation instructions Learning system for parsing navigation instructions Observation Instruction World State Execution Module (MARCO) Instruction World State Training Testing Action Trace Navigation Plan Constructor Semantic Parser Learner Plan Refinement Semantic Parser Action Trace 37

53 Constructing Navigation Plans Landmarks plan: Add interleaving verification steps Basic plan: Directly model the observed actions Verify Travel Turn Verify front: BLUE HALL front: BLUE HALL steps: 1 at: SOFA LEFT Travel Turn steps: 1 LEFT Instruction: Walk to the couch and turn left Action Trace: Forward, Left 38

54 Plan Refinement There are generally many ways to describe the same route. The plan refinement step aims to find the correct plan that corresponds to the instruction. Verify Travel Turn Verify LEFT steps: 2 at: SOFA front: SOFA front: BLUE HALL front: BLUE HALL 39

55 Plan Refinement Turn and walk to the couch Verify Travel Turn Verify LEFT steps: 2 at: SOFA front: SOFA front: BLUE HALL front: BLUE HALL 40

56 Plan Refinement Face the blue hall and walk 2 steps Verify Travel Turn Verify LEFT steps: 2 at: SOFA front: SOFA front: BLUE HALL front: BLUE HALL 41

57 Plan Refinement Turn left. Walk forward twice. Verify Travel Turn Verify LEFT steps: 2 at: SOFA front: SOFA front: BLUE HALL front: BLUE HALL 42

58 Plan Refinement Remove extraneous details in the plans First learn the meaning of words and short phrases Use the learned lexicon to remove parts of the plans unrelated to the instructions 43

59 Verify Travel Turn Verify LEFT steps: 2 at: SOFA front: SOFA front: BLUE HALL front: BLUE HALL Subgraph Generation Online Lexicon Learning (SGOLL) Turn and walk to the couch 1. As an example comes in, break down the sentence and the graph into n-grams and connected subgraphs 44

60 Verify Travel Turn Verify LEFT step s: 2 at: SOF A front : SOF A front : BLUE HALL front : BLUE HALL Subgraph Generation Online Lexicon Learning (SGOLL) Turn and walk to the couch turn, and, walk, to, the, couch 1-gram 2-gram 3-gram turn and, and walk, walk to, to the, the couch turn and walk, and walk to, walk to the, to the couch … Connected subgraph of size 1 Connected subgraph of size 2 Turn LEFT Verify … … Turn LEFT Verify Turn 45

61 Verify Travel Turn Verify LEFT step s: 2 at: SOF A front : SOF A front : BLUE HALL front : BLUE HALL Subgraph Generation Online Lexicon Learning (SGOLL) Turn and walk to the couch turn, and, walk, to, the, couch 1-gram 2-gram 3-gram turn and, and walk, walk to, to the, the couch turn and walk, and walk to, walk to the, to the couch … Turn LEFT Verify … … Turn LEFT Verify Turn Connected subgraph of size 1 Connected subgraph of size 2 45

62 Verify Travel Turn Verify LEFT step s: 2 at: SOF A front : SOF A front : BLUE HALL front : BLUE HALL Subgraph Generation Online Lexicon Learning (SGOLL) Turn and walk to the couch turn, and, walk, to, the, couch 1-gram 2-gram 3-gram turn and, and walk, walk to, to the, the couch turn and walk, and walk to, walk to the, to the couch … Turn LEFT Verify … … Turn LEFT Verify Turn Connected subgraph of size 1 Connected subgraph of size 2 45

63 Verify Travel Turn Verify LEFT step s: 2 at: SOF A front : SOF A front : BLUE HALL front : BLUE HALL Subgraph Generation Online Lexicon Learning (SGOLL) Turn and walk to the couch turn, and, walk, to, the, couch 1-gram 2-gram 3-gram turn and, and walk, walk to, to the, the couch turn and walk, and walk to, walk to the, to the couch … Turn LEFT Verify … … Turn LEFT Verify Turn Connected subgraph of size 1 Connected subgraph of size 2 45

64 Verify Travel Turn Verify LEFT step s: 2 at: SOF A front : SOF A front : BLUE HALL front : BLUE HALL Subgraph Generation Online Lexicon Learning (SGOLL) Turn and walk to the couch turn, and, walk, to, the, couch 1-gram 2-gram 3-gram turn and, and walk, walk to, to the, the couch turn and walk, and walk to, walk to the, to the couch … Turn LEFT Verify … … Turn LEFT Verify Turn Connected subgraph of size 1 Connected subgraph of size 2 45

65 Verify Travel Turn Verify LEFT step s: 2 at: SOF A front : SOF A front : BLUE HALL front : BLUE HALL Subgraph Generation Online Lexicon Learning (SGOLL) Turn and walk to the couch turn, and, walk, to, the, couch 1-gram 2-gram 3-gram turn and, and walk, walk to, to the, the couch turn and walk, and walk to, walk to the, to the couch … Turn LEFT Verify … … Turn LEFT Verify Turn Connected subgraph of size 1 Connected subgraph of size 2 45

66 Verify Travel Turn Verify LEFT step s: 2 at: SOF A front : SOF A front : BLUE HALL front : BLUE HALL Subgraph Generation Online Lexicon Learning (SGOLL) Turn and walk to the couch turn, and, walk, to, the, couch 1-gram 2-gram 3-gram turn and, and walk, walk to, to the, the couch turn and walk, and walk to, walk to the, to the couch … Turn LEFT Verify … … Turn LEFT Verify Turn Connected subgraph of size 1 Connected subgraph of size 2 45

67 Verify Travel Turn Verify LEFT step s: 2 at: SOF A front : SOF A front : BLUE HALL front : BLUE HALL Subgraph Generation Online Lexicon Learning (SGOLL) Turn and walk to the couch turn, and, walk, to, the, couch 1-gram 2-gram 3-gram turn and, and walk, walk to, to the, the couch turn and walk, and walk to, walk to the, to the couch … Turn LEFT Verify … … Turn LEFT Verify Turn Connected subgraph of size 1 Connected subgraph of size 2 45

68 Verify Travel Turn Verify LEFT step s: 2 at: SOF A front : SOF A front : BLUE HALL front : BLUE HALL Subgraph Generation Online Lexicon Learning (SGOLL) Turn and walk to the couch turn, and, walk, to, the, couch 1-gram 2-gram 3-gram turn and, and walk, walk to, to the, the couch turn and walk, and walk to, walk to the, to the couch … Turn LEFT Verify … … Turn LEFT Verify Turn Connected subgraph of size 1 Connected subgraph of size 2 45

69 Verify Travel Turn Verify LEFT step s: 2 at: SOF A front : SOF A front : BLUE HALL front : BLUE HALL Subgraph Generation Online Lexicon Learning (SGOLL) Turn and walk to the couch turn, and, walk, to, the, couch 1-gram 2-gram 3-gram turn and, and walk, walk to, to the, the couch turn and walk, and walk to, walk to the, to the couch … Turn LEFT Verify … … Turn LEFT Verify Turn Connected subgraph of size 1 Connected subgraph of size 2 45

70 Verify Travel Turn Verify LEFT step s: 2 at: SOF A front : SOF A front : BLUE HALL front : BLUE HALL Subgraph Generation Online Lexicon Learning (SGOLL) Turn and walk to the couch turn, and, walk, to, the, couch 1-gram 2-gram 3-gram turn and, and walk, walk to, to the, the couch turn and walk, and walk to, walk to the, to the couch … Turn LEFT Verify … … Turn LEFT Verify Turn Connected subgraph of size 1 Connected subgraph of size 2 45

71 Verify Travel Turn Verify LEFT step s: 2 at: SOF A front : SOF A front : BLUE HALL front : BLUE HALL Subgraph Generation Online Lexicon Learning (SGOLL) Turn and walk to the couch turn, and, walk, to, the, couch 1-gram 2-gram 3-gram turn and, and walk, walk to, to the, the couch turn and walk, and walk to, walk to the, to the couch … Turn LEFT Verify … … Turn LEFT Verify Turn Connected subgraph of size 1 Connected subgraph of size 2 45

72 Verify Travel Turn Verify LEFT step s: 2 at: SOF A front : SOF A front : BLUE HALL front : BLUE HALL Subgraph Generation Online Lexicon Learning (SGOLL) Turn and walk to the couch turn, and, walk, to, the, couch 1-gram 2-gram 3-gram turn and, and walk, walk to, to the, the couch turn and walk, and walk to, walk to the, to the couch … Turn LEFT Verify … … Turn LEFT Verify Turn Connected subgraph of size 1 Connected subgraph of size 2 45

73 Verify Travel Turn Verify LEFT step s: 2 at: SOF A front : SOF A front : BLUE HALL front : BLUE HALL Subgraph Generation Online Lexicon Learning (SGOLL) Turn and walk to the couch turn, and, walk, to, the, couch 1-gram 2-gram 3-gram turn and, and walk, walk to, to the, the couch turn and walk, and walk to, walk to the, to the couch … Turn LEFT Verify … … Turn LEFT Verify Turn Connected subgraph of size 1 Connected subgraph of size 2 45

74 Subgraph Generation Online Lexicon Learning (SGOLL) turn Turn … LEFT 2. Increase the counts and co-occurrence count of each n-gram, connected-subgraph pair. Hash the connected-subgraphs for efficient update. 48 22 26 Turn RIGHT 46

75 Subgraph Generation Online Lexicon Learning (SGOLL) turn Turn … LEFT 2. Increase the counts and co-occurrence count of each n-gram, connected-subgraph pair. Hash the connected-subgraphs for efficient update. 48+1 22+1 26 Turn RIGHT 46

76 Subgraph Generation Online Lexicon Learning (SGOLL) turn Turn … LEFT 2. Increase the counts and co-occurrence count of each n-gram, connected-subgraph pair. Hash the connected-subgraphs for efficient update. 49 23 26 Turn RIGHT 46

77 Subgraph Generation Online Lexicon Learning (SGOLL) turn Turn … LEFT 3. Rank the entries by the scoring function 0.56 0.31 0.27 Turn RIGHT 47

78 Refining Plans Using A Lexicon Verify Travel Turn Verify front: BLUE HALL front: BLUE HALL steps: 1 at: SOFA LEFT Walk to the couch and turn left Look for the highest-scoring lexical entry that matches both the words in the instruction and the nodes in the graph. 48

79 Refining Plans Using A Lexicon Walk to the couch and turn left Lexicon entry: turn left (score: 0.84) Turn LEFT Verify Travel Turn Verify front: BLUE HALL front: BLUE HALL steps: 1 at: SOFA LEFT 49

80 Refining Plans Using A Lexicon Walk to the couch and Lexicon entry: walk to the couch (score: 0.75) Verify Travel at: SOFA Verify Travel Turn Verify front: BLUE HALL front: BLUE HALL steps: 1 at: SOFA LEFT 50

81 Refining Plans Using A Lexicon and Lexicon exhausted Verify Travel Turn Verify front: BLUE HALL front: BLUE HALL steps: 1 at: SOFA LEFT 51

82 Refining Plans Using A Lexicon and Remove all unmarked nodes Verify Travel Turn at: SOFA LEFT 52

83 Refining Plans Using A Lexicon Walk to the couch and turn left Verify Travel Turn at: SOFA LEFT 53

84 Evaluation Data Statistics 3 maps, 6 instructors, 1-15 followers/direction Hand-segmented into single sentence steps Take the wood path towards the easel. At the easel, go left and then take a right on the the blue path at the corner. Follow the blue path towards the chair and at the chair, take a right towards the stool. When you reach the stool, you are at 7. Paragraph Single sentence Take the wood path towards the easel. At the easel, go left and then take a right on the the blue path at the corner. … 54 Turn, Forward, Turn left, Forward, Turn right, Forward x 3, Turn right, Forward Turn Forward, Turn left, Forward, Turn right

85 Evaluation Data Statistics ParagraphSingle-Sentence # Instructions7063236 Avg. # sentences5.0 (±2.8)1.0 (±0) Avg. # words37.6 (±21.1)7.8 (±5.1) Avg. # actions10.4 (±5.7)2.1 (±2.4) 55

86 Plan Construction Test how well the system infers the correct navigation plans Gold-standard plans annotated manually Use partial parse accuracy as metric – Credit for the correct action type (e.g. Turn) – Additional credit for correct arguments (e.g. LEFT) Lexicon learned and tested on the same data from two maps out of three 56

87 Plan Construction PrecisionRecallF1 Basic Plans81.4655.8866.27 Landmarks Plans45.4285.4659.29 Refined Landmarks Plans 87.3272.9679.49 57

88 Parse Accuracy Examine how well the learned semantic parsers can parse novel sentences Train semantic parsers on the inferred plans using KRISP [Kate and Mooney, 2006] Leave-one-map-out approach Use partial parse accuracy as metric Upper baseline: train on human annotated plans 58

89 Parse Accuracy PrecisionRecallF1 Basic Plans 85.3950.9563.76 Landmarks Plans 52.2050.4851.11 Refined Landmarks Plans 88.3657.0369.31 Human Annotated Plans 90.0976.2982.62 59

90 End-to-end Execution Test how well the system can perform the overall navigation task Leave-one-map-out approach Strict metric: Only successful if the final position matches exactly 60

91 End-to-end Execution Lower baseline – A simple generative model based on the frequency of actions alone Upper baselines – Training with human annotated gold plans – Complete MARCO system [MacMahon, 2006] – Humans 61

92 End-to-end Execution Single SentencesParagraphs Simple Generative Model 11.08%2.15% Basic Plans 58.72%16.21% Landmarks Plans 18.67%2.66% Refined Landmarks Plans 57.28%19.18% Human Annotated Plans 62.67%29.59% MARCO 77.87%55.69% Human Followers N/A69.64% 62

93 End-to-end Execution Single SentencesParagraphs Simple Generative Model 11.08%2.15% Basic Plans 58.72%16.21% Landmarks Plans 18.67%2.66% Refined Landmarks Plans 57.28%19.18% Human Annotated Plans 62.67%29.59% MARCO 77.87%55.69% Human Followers N/A69.64% 62

94 End-to-end Execution Single SentencesParagraphs Simple Generative Model 11.08%2.15% Basic Plans 58.72%16.21% Landmarks Plans 18.67%2.66% Refined Landmarks Plans 57.28%19.18% Human Annotated Plans 62.67%29.59% MARCO 77.87%55.69% Human Followers N/A69.64% 62

95 End-to-end Execution Single SentencesParagraphs Simple Generative Model 11.08%2.15% Basic Plans 58.72%16.21% Landmarks Plans 18.67%2.66% Refined Landmarks Plans 57.28%19.18% Human Annotated Plans 62.67%29.59% MARCO 77.87%55.69% Human Followers N/A69.64% 62

96 Example Parse Instruction: “Place your back against the wall of the ‘T’ intersection. Turn left. Go forward along the pink-flowered carpet hall two segments to the intersection with the brick hall. This intersection contains a hatrack. Turn left. Go forward three segments to an intersection with a bare concrete hall, passing a lamp. This is Position 5.” Parse:Turn ( ), Verify ( back: WALL ), Turn ( LEFT ), Travel ( ), Verify ( side: BRICK HALLWAY ), Turn ( LEFT ), Travel ( steps: 3 ), Verify ( side: CONCRETE HALLWAY ) 63

97 Mandarin Chinese Experiment Translated all the instructions from English to Chinese Train and test in the same way Chinese does not include word boundaries (spaces) Naively segment each character Use a trained Chinese word segmenter [Chang, Galley & Manning, 2008] 64

98 Mandarin Chinese Experiment Single SentencesParagraphs Refined Landmarks Plans (segment by character) 58.54%16.11% Refined Landmarks Plans (segment by Stanford segmenter) 58.70%20.13% 65

99 Collect Additional Training Data Using Mechanical Turk Needs 2 types of data – Navigation instructions – Action traces corresponding to those instructions Design 2 Human Intelligence Tasks (HITs) to collect them from Amazon’s Mechanical Turk 66

100 Instructor Task Shown a randomly generated simulation that contains 1 to 4 actions Workers asked to write an instruction that would lead someone to perform those actions – Intended to be single sentences, but not enforced 67

101 Follower Task Given an instruction and placed at the starting position Worker asked to follow the instruction as best as they can 68

102 Data Statistics Two-month period, 653 workers, $277.92 Over 2,400 instructions and 14,000 follower traces Filter so that instructions need to have > 5 followers and majority agreement on a single path 69 MTurkOriginal Data # Instructions10113236 Vocabulary size590629 Avg. # words7.69 (±7.12)7.8 (±5.1) Avg. # actions1.84 (±1.24)2.1 (±2.4)

103 Mechanical Turk Training Data Experiment Single SentencesParagraphs Refined Landmarks Plans 57.28%19.18% Refined Landmarks Plans (with additional training data from Mechanical Turk) 57.62%20.64% 70

104 Overview Background Learning from limited ambiguity Learning from ambiguous relational data Future directions Conclusions 71

105 System Improvements Tighter integration of the components Learning higher-order concepts Incorporating syntactic information Scaling the system Sportscasting task – More detailed meaning representation – Entertainment (variability, emotion, etc) Navigation task – Generating instructions – Dialogue system for more interactions 72

106 Long-term Directions Data acquisition framework – Let users of the system also provide training data Real perceptions – Use computer vision to recognize objects and events Machine Translation – Learn translation models from comparable corpora 73

107 Conclusion Traditional language learning approach relies on expensive, annotated training data. We have developed language learning systems that can learn by observing language used in a relevant, perceptual environment. We have devised two test applications that address two different scenarios: – Learning to sportscast (limited ambiguity) – Learning to follow navigation instructions (exponential ambiguity) 74


Download ppt "David L. Chen Supervisor: Professor Raymond J. Mooney Ph.D. Dissertation Defense January 25, 2012 Learning Language from Ambiguous Perceptual Context."

Similar presentations


Ads by Google