Presentation is loading. Please wait.

Presentation is loading. Please wait.

Latent Topic Modeling of Word Vicinity Information for Speech Recognition Kuan-Yu Chen, Hsuan-Sheng Chiu, Berlin Chen ICASSP 2010 Hao-Chin Chang Department.

Similar presentations


Presentation on theme: "Latent Topic Modeling of Word Vicinity Information for Speech Recognition Kuan-Yu Chen, Hsuan-Sheng Chiu, Berlin Chen ICASSP 2010 Hao-Chin Chang Department."— Presentation transcript:

1 Latent Topic Modeling of Word Vicinity Information for Speech Recognition Kuan-Yu Chen, Hsuan-Sheng Chiu, Berlin Chen ICASSP 2010 Hao-Chin Chang Department of Computer Science & Information Engineering National Taiwan Normal University

2 2 Outline Introduction Related work Experiment Conclusion

3 Introduction A new topic language model, named word vicinity model (WVM), is proposed to explore the co-occurrence relationship between words, as well as the long-span latent topical information for language model adaptation. WVM is close in spirit to WTM, but has a more concise parameterization leading to more reliable model estimation, particularly when available training data is of limited size We present a new modeling approach to WTM, and also design an efficient way to simultaneously extract “word-document” and “word-word” co-occurrence dependence inherent in the adaptation corpus through the sharing of a common set of latent topics. 3

4 DTM We can interpret each of its corresponding search histories Hwi as a document (or history) topic model used for predicting the occurrence probability of Wi The probabilities of PLSA and the background n-gram (e.g: trigram) language model can be combined through a simple linear interpolation 4

5 WTM 5 The WTM model of each word wj for predicting the occurrence of a particular word wi can be expressed by During speech recognition, for a search history Hwi=w1,w2,…,wi-1 of a decoded word wi, we can linearly combine the associated WTM models of the words occurring in Hwi to form a composite WTM model for predicting wi

6 WVM The relationship between words, projected into a low dimensional probability space characterized by the shared set of topic distributions Historical word wj predicting another decoded word wi 6

7 Comparison 7 DTMWTM WVM 1. Relations Word and DoucmentWord and Word 2. Topic weight Expectation Maximization Basic Topic DTM WTMWVM 3. Parameterization ComplexConcise

8 Hybrid We present a novel approach to simultaneously capturing “word- document” and “word-word” co-occurrence relationship inherent in the corpus through using a common set of latent topics During speech recognition, we can linearly combine PLSA and WTM to form a hybrid topic language model: 8

9 Experiment 9 The speech corpus consists of about 200 hours of MATBN Mandarin broadcast news A subset of 25-hour speech compiled during November 2001 to December 2002 was used to bootstrap the acoustic training with the minimum phone error rate (MPE) criterion and the training data selection scheme Another subset of 3-hour speech data collected within 2003 is reserved for development (1.5 hours) and testing (1.5 hours) The vocabulary size is about 72 thousand words The n-gram language models used in this paper consist of trigram and bigram models, which a background text corpus consisting of 170 million Chinese characters collected from Central News Agency (CNA) in 2001 and 2002 (the Chinese Gigaword Corpus released by LDC) using the SRI Language Modeling Toolkit (SRILM) The adaptation text corpus used for training PLSA, LDA, WTM and WVM was collected from MATBN 2001, 2002, and 2003, which consists of one million Chinese characters of the orthographic broadcast news transcripts.

10 Experimental ReductionPLSALDAWTMWVM Perplexity23.1%21.8%26.1%24.2% Character Error rate 4.1%5.5% 4.6% 10

11 11 Conclusion 1. Investigate more elaborate window functions for WVM and WTM 2. Seek the use of discriminative training algorithms for training WVM and WTM 3. Apply WVM to speech retrieval and summarization tasks


Download ppt "Latent Topic Modeling of Word Vicinity Information for Speech Recognition Kuan-Yu Chen, Hsuan-Sheng Chiu, Berlin Chen ICASSP 2010 Hao-Chin Chang Department."

Similar presentations


Ads by Google