VARIATIONAL APPROXIMATION OF LONG-SPAN LANGUAGE MODELS FOR LVCSR
Language Modeling
Přednášející: Sanjeev Khudanpur, Autoři: Anoop Deoras, Center for Language and Speech Processing, United States; Tomáš Mikolov, Stefan Kombrink, Martin Karafiát, Brno University of Technology, Czech Republic; Sanjeev Khudanpur, Center for Language and Speech Processing, United States
Long-span language models that capture syntax and semantics are seldom used in the first pass of large vocabulary continuous speech recognition systems due to the prohibitive search-space of sentence-hypotheses. Instead, an N-best list of hypotheses is created using tractable n-gram models, and rescored using the long-span models. It is shown in this paper that computationally tractable variational approximations of the long-span models are a better choice than standard n-gram models for first pass decoding. They not only result in a better first pass output, but also produce a lattice with a lower oracle word error rate, and rescoring the N-best list from such lattices with the long-span models requires a smaller N to attain the same accuracy. Empirical results on the WSJ, MIT Lectures, NIST 2007 Meeting Recognition and NIST 2001 Conversational Telephone Recognition data sets are presented to support these claims.
Informace o přednášce
Nahráno: | 2011-05-25 17:35 - 17:55, Club H |
---|---|
Přidáno: | 9. 6. 2011 09:38 |
Počet zhlédnutí: | 41 |
Rozlišení videa: | 1024x576 px, 512x288 px |
Délka videa: | 0:25:02 |
Audio stopa: | MP3 [8.56 MB], 0:25:02 |
Komentáře