Machine Translation  Machine translation is of one of the earliest uses of AI  Two approaches:  Traditional approach using grammars, rewrite rules,

Slides:



Advertisements
Similar presentations
CILC2011 A framework for structured knowledge extraction and representation from natural language via deep sentence analysis Stefania Costantini Niva Florio.
Advertisements

Building an ASR using HTK CS4706
Speech Recognition Part 3 Back end processing. Speech recognition simplified block diagram Speech Capture Speech Capture Feature Extraction Feature Extraction.
SEARCHING QUESTION AND ANSWER ARCHIVES Dr. Jiwoon Jeon Presented by CHARANYA VENKATESH KUMAR.
Probabilistic Language Processing Chapter 23. Probabilistic Language Models Goal -- define probability distribution over set of strings Unigram, bigram,
For Friday No reading Homework –Chapter 23, exercises 1, 13, 14, 19 –Not as bad as it sounds –Do them IN ORDER – do not read ahead here.
December 2006 Cairo University Faculty of Computers and Information HMM Based Speech Synthesis Presented by Ossama Abdel-Hamid Mohamed.
Hidden Markov Models Bonnie Dorr Christof Monz CMSC 723: Introduction to Computational Linguistics Lecture 5 October 6, 2004.
Natural Language Processing - Speech Processing -
Speech Translation on a PDA By: Santan Challa Instructor Dr. Christel Kemke.
Application of HMMs: Speech recognition “Noisy channel” model of speech.
ITCS 6010 Spoken Language Systems: Architecture. Elements of a Spoken Language System Endpointing Feature extraction Recognition Natural language understanding.
Big Ideas in Cmput366. Search Blind Search Iterative deepening Heuristic Search A* Local and Stochastic Search Randomized algorithm Constraint satisfaction.
Grammar induction by Bayesian model averaging Guy Lebanon LARG meeting May 2001 Based on Andreas Stolcke’s thesis UC Berkeley 1994.
Machine Translation Prof. Alexandros Potamianos Dept. of Electrical & Computer Engineering Technical University of Crete, Greece May 2003.
COMP 4060 Natural Language Processing Speech Processing.
Why is ASR Hard? Natural speech is continuous
1 Statistical NLP: Lecture 13 Statistical Alignment and Machine Translation.
Natural Language Understanding
Representing Acoustic Information
Lecture 1, 7/21/2005Natural Language Processing1 CS60057 Speech &Natural Language Processing Autumn 2005 Lecture 1 21 July 2005.
Audio Processing for Ubiquitous Computing Uichin Lee KAIST KSE.
Introduction to Automatic Speech Recognition
Machine translation Context-based approach Lucia Otoyo.
1 7-Speech Recognition (Cont’d) HMM Calculating Approaches Neural Components Three Basic HMM Problems Viterbi Algorithm State Duration Modeling Training.
Artificial Intelligence 2004 Speech & Natural Language Processing Natural Language Processing written text as input sentences (well-formed) Speech.
Midterm Review Spoken Language Processing Prof. Andrew Rosenberg.
7-Speech Recognition Speech Recognition Concepts
1 Computational Linguistics Ling 200 Spring 2006.
A brief overview of Speech Recognition and Spoken Language Processing Advanced NLP Guest Lecture August 31 Andrew Rosenberg.
By: Meghal Bhatt.  Sphinx4 is a state of the art speaker independent, continuous speech recognition system written entirely in java programming language.
Hierarchical Dirichlet Process (HDP) A Dirichlet process (DP) is a discrete distribution that is composed of a weighted sum of impulse functions. Weights.
Speech, Perception, & AI Artificial Intelligence CMSC March 5, 2002.
17.0 Distributed Speech Recognition and Wireless Environment References: 1. “Quantization of Cepstral Parameters for Speech Recognition over the World.
Jacob Zurasky ECE5526 – Spring 2011
Entropy & Hidden Markov Models Natural Language Processing CMSC April 22, 2003.
LML Speech Recognition Speech Recognition Introduction I E.M. Bakker.
Artificial Intelligence 2004 Speech & Natural Language Processing Natural Language Processing written text as input sentences (well-formed) Speech.
Approaches to Machine Translation CSC 5930 Machine Translation Fall 2012 Dr. Tom Way.
Speech, Perception, & AI Artificial Intelligence CMSC February 13, 2003.
Artificial Intelligence 2004 Speech & Natural Language Processing Speech Recognition acoustic signal as input conversion into written words Natural.
For Monday Read chapter 24, sections 1-3 Homework: –Chapter 23, exercise 8.
For Friday Finish chapter 24 No written homework.
For Monday Read chapter 26 Last Homework –Chapter 23, exercise 7.
Speech Recognition with CMU Sphinx Srikar Nadipally Hareesh Lingareddy.
Probabilistic reasoning over time Ch. 15, 17. Probabilistic reasoning over time So far, we’ve mostly dealt with episodic environments –Exceptions: games.
For Friday Finish chapter 23 Homework –Chapter 23, exercise 15.
Performance Comparison of Speaker and Emotion Recognition
BY KALP SHAH Sentence Recognizer. Sphinx4 Sphinx4 is the best and versatile recognition system. Sphinx4 is a speech recognition system which is written.
Chunk Parsing II Chunking as Tagging. Chunk Parsing “Shallow parsing has become an interesting alternative to full parsing. The main goal of a shallow.
Chapter 7 Speech Recognition Framework  7.1 The main form and application of speech recognition  7.2 The main factors of speech recognition  7.3 The.
Natural Language Processing Group Computer Sc. & Engg. Department JADAVPUR UNIVERSITY KOLKATA – , INDIA. Professor Sivaji Bandyopadhyay
Natural Language Processing (NLP)
For Monday Read chapter 26 Homework: –Chapter 23, exercises 8 and 9.
1 7-Speech Recognition Speech Recognition Concepts Speech Recognition Approaches Recognition Theories Bayse Rule Simple Language Model P(A|W) Network Types.
23.3 Information Extraction More complicated than an IR (Information Retrieval) system. Requires a limited notion of syntax and semantics.
A Simple English-to-Punjabi Translation System By : Shailendra Singh.
Automatic Speech Recognition
CSC 594 Topics in AI – Natural Language Processing
Approaches to Machine Translation
ARTIFICIAL NEURAL NETWORKS
Natural Language Processing (NLP)
Speech Processing Speech Recognition
Approaches to Machine Translation
CS4705 Natural Language Processing
Natural Language Processing (NLP)
Artificial Intelligence 2004 Speech & Natural Language Processing
Information Retrieval
Measuring the Similarity of Rhythmic Patterns
Natural Language Processing (NLP)
Presentation transcript:

Machine Translation  Machine translation is of one of the earliest uses of AI  Two approaches:  Traditional approach using grammars, rewrite rules, and lexicons  May be shallow or deep translation  May translate directly between two languages, or from one language into interlingua and then into second language  Statistical machine translation

Shallow Translation  Transfer model: Keep a database of translation rules or examples. When rule matches, translate directly  Could operate on lexical, syntactic, or semantic level  Example: technical manuals (Siemens)  Doesn't deal with context

Deep Translation  One method uses language independent representation of information – interlingua  Three problems:  Knowledge representation  Parsing into that representation  Generation from representation  Alternate approach is directly from one language to another

Statistical Machine Translation  Translation model is learned from a bilingual corpus  One system:  Break the original sentences in phrases  Choose a corresponding phrase in the target language  Choose a permutation of the phrases  Select the most probable translation

Efficiency  Instead of examining all permutations (n!), use the concept of distortion  Distortion d i is the number of words that the phrase f i has moved with respect to f i-1, positive if moved to right, negative if moved to the left  Find a probability distribution for d  Each distortion is independent of the others

Efficiency (cont'd)  Still exponential over the number of phrases  Use beam search with a heuristic that estimates probability to find nearly-most-probably translation

Speech Recognition  Challenges:  Little segmentation, unlike written text  Coarticulation: sound at end of word runs into sound at beginning of next word  Homophones  Solutions:  Find acoustic model: P(sound i:t | word 1:t )  Find language model: P(word 1:t )  Use HMM and Viterbi algorithm

Acoustical Processing  Sample analog signal: sampling rate, quantization factor matter  Divide into frames  Extract features from each frame:  Use Fourier Transform to measure acoustic energy at about a dozen frequencies  Computer the mel frequency cepstral coefficient (mfcc) for each frequency  Yields thirteen features

Processing (cont'd)  Each phone has a onset, middle, and end  The phone models are strung together to form a pronunciation model for each word  Words can have a coarticulation model: “tomato” vs. “tomahto”  Language model can be an n-gram model learned from a corpus of text