Course Overview  What is AI?  What are the Major Challenges?  What are the Main Techniques?  Where are we failing, and why?  Step back and look at.

Slides:



Advertisements
Similar presentations
A very short introduction to Natural Language Generation Kees van Deemter Computing Science University of Aberdeen.
Advertisements

The following are badly translated signs from around the world.
Grammar: Meaning and Contexts * From Presentation at NCTE annual conference in Pittsburgh, 2005.
The Chinese Room: Understanding and Correcting Machine Translation This work has been supported by NSF Grants IIS Solution: The Chinese Room Conclusions.
Helping Your Child Learn to Read
What ’ s New? Acquiring New Information as a Process in Comprehension Suan E. Haviland & Herbert H. Clark.
Unit 8. Find a new word in this unit to match the definition given.
2-May-15 GUI Design. 2 HMI design There are entire college courses taught on HMI (Human-Machine Interface) design This is just a very brief presentation.
Rhee Dong Gun. Chapter The speaking process The differences between spoken and written language Speaking skills Speaking in the classroom Feedback.
People & Speech Interfaces CS 260 Wednesday, October 4, 2006.
Natural Language Processing AI - Weeks 19 & 20 Natural Language Processing Lee McCluskey, room 2/07
Inducing Information Extraction Systems for New Languages via Cross-Language Projection Ellen Riloff University of Utah Charles Schafer, David Yarowksy.
Language, Mind, and Brain by Ewa Dabrowska Chapter 2: Language processing: speed and flexibility.
Resources Primary resources – Lexicons, structured vocabularies – Grammars (in widest sense) – Corpora – Treebanks Secondary resources – Designed for a.
Comments on Guillaume Pitel: “Using bilingual LSA for FrameNet annotation of French text from generic resources” Gerd Fliedner Computational Linguistics.
Fundamentals: Linguistic principles
Lecture 1 Introduction: Linguistic Theory and Theories
Programming. Software is made by programmers Computers need all kinds of software, from operating systems to applications People learn how to tell the.
thinking hats Six of Prepared by Eman A. Al Abdullah ©
Cursive Cursor ™ We read things So you don’t have to.
Machine translation Context-based approach Lucia Otoyo.
Classical Method : The very boring-sounding method of language teaching described above is the Classical Method, also known as the Grammar Translation.
Machine Translation Dr. Radhika Mamidi. What is Machine Translation? A sub-field of computational linguistics It investigates the use of computer software.
Course Overview  What is AI?  What are the Major Challenges?  What are the Main Techniques?  Where are we failing, and why?  Step back and look at.
Comprehensible Input SIOP Component #3.
CSCI 4410 Introduction to Artificial Intelligence.
Turing Test and other amusements. Read this! The Actual Article by Turing.
Lecture 12: 22/6/1435 Natural language processing Lecturer/ Kawther Abas 363CS – Artificial Intelligence.
Supervisor: Dr. Eddie Jones Electronic Engineering Department Final Year Project 2008/09 Development of a Speaker Recognition/Verification System for Security.
Historical linguistics Historical linguistics (also called diachronic linguistics) is the study of language change. Diachronic: The study of linguistic.
1 Computational Linguistics Ling 200 Spring 2006.
201 English as a Global Language By Don L. F. Nilsen (pictures from )
TMF1013 : Introduction To Computing Lecture 1 : Fundamental of Computer ComputerFoudamentals.
Unit 2 A Flat World.  Objectives Objectives  FocusFocus  Warming up Warming up  7.1 Asking people to do things 7.1 Asking people to do things  7.2.
The Great Vowel Shift Continued The reasons behind this shift are something of a mystery, and linguists have been unable to account for why it took place.
CSE 219 Computer Science III Program Design Principles.
Course Overview  What is AI?  What are the Major Challenges?  What are the Main Techniques?  Where are we failing, and why?  Step back and look at.
Machine Translation  Machine translation is of one of the earliest uses of AI  Two approaches:  Traditional approach using grammars, rewrite rules,
English II March 18 – March 22. Daily Grammar Practice – Monday Notes Write out the sentence and identify parts of speech (noun, verb, adjective, etc.)
Early Behaviours and What to Look For EARLY READING BEHAVIOURS…
SPEECH AND WRITING. Spoken language and speech communication In a normal speech communication a speaker tries to influence on a listener by making him:
CMPF144 FUNDAMENTALS OF COMPUTING THEORY Module 10: Applied Mathematical - Logical Knowledge Riddle/Puzzle.
Intermediate 2 Software Development Process. Software You should already know that any computer system is made up of hardware and software. The term hardware.
Are you ready to play…. Deal or No Deal? Deal or No Deal?
Close Reading Intermediate 2. Time The Close Reading exam paper lasts for one hour. (Date and time for 2011: Friday 13 May, 1.00pm to 2.00pm.) NAB: Friday.
Introduction to Computational Linguistics
Data Analysis Econ 176, Fall Populations When we run an experiment, we are always measuring an outcome, x. We say that an outcome belongs to some.
The Software Development Process
Artificial Intelligence: Natural Language
Unit 8 LANGUAGE FOCUS. Content  Word study  Word used in Computing and Telephoning  Grammar  Pronoun  Indirect speech with conditional sentences.
ACE TESOL Diploma Program – London Language Institute OBJECTIVES You will understand: 1. The terminology and concepts of semantics, pragmatics and discourse.
1 Introduction to Natural Language Processing ( ) Language Modeling (and the Noisy Channel) AI-lab
Fall 2015 ECEn 490 Lecture #8 1 Effective Presentations How to communicate effectively with your audience.
Clustering Prof. Ramin Zabih
Topic and the Representation of Discourse Content
For Friday Finish chapter 23 Homework –Chapter 23, exercise 15.
Course Overview  What is AI?  What are the Major Challenges?  What are the Main Techniques?  Where are we failing, and why?  Step back and look at.
Questioning as Formative Assessment: GRECC Math Alliance February 4 th - 7 th, 2008.
A Simple English-to-Punjabi Translation System By : Shailendra Singh.
1 Instructing the English Language Learner (ELL) in the Regular Classroom.
Tests of Significance We use test to determine whether a “prediction” is “true” or “false”. More precisely, a test of significance gets at the question.
The Law of Averages. What does the law of average say? We know that, from the definition of probability, in the long run the frequency of some event will.
By: Megan Funk. I will: 1. Explain the binary number system How to: -Generate binary from a number -Add binary 2. Explain the base-b number system 3.
ACT REVIEW. RUN-ONS A complete sentence contains a subject, a verb, and a complete thought. If any of the three is lacking, the sentence is called a.
Methods of Training.
Mrs Joslyn Fox.  TIME MANAGEMENT: Don’t leave everything until the last minute!!!
Arnar Thor Jensson Koji Iwano Sadaoki Furui Tokyo Institute of Technology Development of a Speech Recognition System For Icelandic Using Machine Translated.
Done Done Course Overview What is AI? What are the Major Challenges?
Teaching Listening Based on Active Learning.
Computational Thinking for KS3
Presentation transcript:

Course Overview  What is AI?  What are the Major Challenges?  What are the Main Techniques?  Where are we failing, and why?  Step back and look at the Science  Step back and look at the History of AI  What are the Major Schools of Thought?  What of the Future? What are we trying to do? How far have we got?  Natural language (text & speech) (continued…)  Robotics  Computer vision  Problem solving  Learning  Board games  Applied areas: Video games, healthcare, … What has been achieved, and not achieved, and why is it hard?

Text Language Technology Natural Language Understanding Natural Language Generation Speech Recognition Speech Synthesis Text Meaning Speech Hard! Interlingua open 1 open 2 Cheaters’ shortcut

Modern Machine Translation  Prevalent approach uses statistics, following an idea by Warren Weaver (conceived as early as 1947)  View translation as a form of decoding: “Dutch is just coded English” (or the other way round)  i.e. look at the problem from the computer’s point of view  Deciphering coded text, which replaces each English word with a coded word  Suppose you have a large English text, and an even larger corpus of English  You guess the correct version of a coded word by comparing the frequency of that word in the corpus with the frequency of all the words in your text  E.g., most frequent word must be ‘the’, so the most frequent word in the corpus may be code for ‘the’. (Just a guess!)  Check whether this combination of guesses is a proper English text; change where necessary  Can try with Google : “wound will cure/heal” “served his sorrow/sentence”

Modern Machine Translation  But of course, Dutch is not just coded English. (For example, the right translation for “open” may depend on the words surrounding “open”.)  How do we find out how sentences in the two languages are related?  To get a good starting point, Machine Translation uses huge bilingual corpora (usually based on human translation)  Example: Canadian Hansard corpus, bilingual French/English parliament proceedings, also Hong Kong  (But I’ll use Dutch as an example)

 Here we will not explain the statistical techniques used  Just observe: Guess how expressions line up across two languages  Based on pure pattern matching. No knowledge of Dutch or English is required  NB: in statistical translation program, no longer easy to see understanding followed by generation Modern Machine Translation

 perform various preparatory operations (e.g., match corresponding sentences with each other)  hypothesise ways of matching smaller expressions with each other. Example 1:  E: ‘that Blair responded’  N: ‘dat Blair antwoordde’  E: ‘whether Kennedy responded’  N: ‘of Kennedy antwoordde’ Modern Machine Translation

 Here is a more interesting example, involving differences in word order between the two languages:  Example 2:  E: that Blair responded to the question  N: dat Blair op de vraag antwoordde  E: whether Kennedy responded to the challenge  N: of Kennedy op de uitdaging antwoordde  Need offsets in translation model Modern Machine Translation

 Need a translation model and a language model  Translation model tells us likely translations (roughly)  Language model tells us how good those sentences are in the target language  Language model  Ideally we would like to know how common any sentence is  We will settle for pairs (bigrams)  Translation model  Often use something quite crude, like word by word  Correct positions with offsets  Good language model can save bad translation model Models, Unigrams, Bigrams, Trigrams

How far has Machine Translation advanced? National Institute of Standards and Technology (NIST) Regular competitions between MT systems (Source: K.Knight, Statistical MT Tutorial, Aberdeen 2005)

winner 2002 insistent Wednesday may recurred her trips to Libya tomorrow for flying Cairo 6-4 (AFP) – an official announced today in the Egyptian lines company for flying Tuesday is a company “insistent for flying” may resumed a consideration of a day wednesday tomorrow her trips to Libya of security council decision trace international the imposed ban comment winner 2003 Egyptair Has Tomorrow to Resume its flights to Libya Cairo 4-6 (AFP) – Said an official at the Egyptian Aviation Company today that the company egyptair may resume as of tomorrow, Wednesday its flights to Libya after the International Security Council resolution to the suspension of the embargo imposed on Libya.

Conclusion on Statistical MT  This approach to MT relies on massive parallel corpora; these are not yet available for all language pairs  The MT system does not “understand” the content of the sentences  Perhaps progress using statistical methods will flatten in future  but they are starting to be combined with “higher-level” information

Practical Machine Translation Types of translation  Rough translation  Could perhaps be post-edited by a monolingual human (cheaper)  Restricted source translation  Subject and form restricted, e.g. weather forecast  Pre-edited translation  Human pre-edits, e.g. Caterpillar English  Can improve original too  Literary

Summing up Modern Translation  Deep vs. Shallow?  Deep - comprehensive knowledge of the word.  Shallow - no knowledge.  So far, shallow approaches more successful.  Deep can be better on a particular domain if a lot of expert effort is put into building models  Shallow approach is much easier  Similar story in other areas of AI  Each of these programs on its own is highly specialised (i.e., limited)

On the other hand… Humans don’t always get it right either!  French hotel: “Please leave your values at the front desk.”  Athens hotel: “We expect our visitors to complain daily at the office between the hours of 9 and 11 a.m.”  Tokyo hotel room: “The flattening of underwear is the job of the chambermaid - get it done, turn her on.”  Hong Kong tailor shop: “Order your summer suit now. Because of big rush we execute customers in strict rotation.”  Men's room at Mexican golf course/resort: “Guests are requested not to wash their balls in the hand basins.”  Budapest elevator: “due to out of order we regret that you are unbearable”  Bangkok Dry Cleaner's: “Drop Trousers Here for Best Results”  Tokyo hotel room: “Please take advantage of our chambermaids.” They do understand, but they may make the wrong choices in the target language

Speech Recognition  Signal processing to recognise features  Coarticulation: model how each sound (“phone”) depends on neighbours  Dialect: different possible pronunciations  To recognise isolated words use unigram language model again  Continuous speech: use bigram or trigram model  Try:  “eat I scream” vs. “eat ice cream”  “eat a banana” vs. “eat a bandana”

Speech Recognition  Humans are remarkably good because of high level knowledge  Computers:  No background noise, single speaker, vocabulary few thousand words: >99%  In general with good acoustics: 60-80%  On noisy phone terrible

Natural Language Generation (NLG) Natural Language Generation is better than having people write texts when:  There are many potential documents to be written, differing according to the context (user, situation, language)  There are some general principles behind document design

Example: Noun Phrase design A noun phrase can convey an arbitrary amount of information:  Someone vs.  a designer vs.  an old designer vs.  an old designer with red hair … How much information should we “pack into” a given Noun Phrase? This is normally considered part of the aggregation task.

Some Issues to Consider  Preferred ordering within the text (e.g. most important first)  Readability of the Noun Phrase,  Flow of “focus”,  Successful use of pronouns and abbreviated references

Example Content (NB we assume that words, basic syntax etc have been chosen) This T-shirt was made by James Sportler. Sportler is a famous British designer. He drives an ancient pink Jaguar. He works in London with Thomas Wendsop. Wendsop won the first prize in the FWJG awards. Can/should we add more to the Noun Phrase?

One possible addition This T-shirt was made by James Sportler, who works in London with Thomas Wendsop. Sportler is a famous British designer. He drives an ancient pink Jaguar. Wendsop won the first prize in the FWJG awards.  Facts about Wendsop are now separated from one another (focus).  Wendsop now has greater prominence in the text (ordering)

Another possible addition This T-shirt was made by James Sportler, a famous British designer who works in London with Thomas Wendsop, who won the first prize in the FWJG awards. Sportler drives an ancient pink Jaguar.  The Noun Phrase is now very complex (readability)  “He” now doesn’t seem to work in the second sentence (pronouns)

Another possible addition This T-shirt was made by James Sportler, a famous British designer. He drives an ancient pink Jaguar. He works in London with Thomas Wendsop. Wendsop won the first prize in the FWJG awards.  Possibly the best solution, but is this better than the original “text”?

Why is Natural Language Generation hard?  Natural Language Generation involves making many choices, e.g. which content to include, what order to say it in, what words and syntactic constructions to use.  Linguistics does not provide us with a ready-made, precise theory about how to make such choices to produce coherent text  The choices to be made interact with one another in complex ways  Many results of choices (e.g. text length) are only visible at the end of the process  There doesn’t seem to be any simple and reliable way to order the choices