STRUCTURED OUTPUT LAYER NEURAL NETWORK LANGUAGE MODEL
Language Modeling
Presented by: Ilya Oparin, Author(s): Hai Son Le, LIMSI CNRS / Uni. Paris-Sud, France; Ilya Oparin, LIMSI CNRS, France; Alexandre Allauzen, LIMSI CNRS / Uni. Paris-Sud, France; Jean-Luc Gauvain, LIMSI CNRS, France; Francois Yvon, LIMSI CNRS / Uni. Paris-Sud, France
This paper introduces a new neural network language model (NNLM) based on word clustering to structure the output vocabulary: Structured Output Layer NNLM. This model is able to handle vocabularies of arbitrary size, hence dispensing with the design of short-lists that are commonly used in NNLMs. Several softmax layers replace the standard output layer in this model. The output structure depends on the word clustering which uses the continuous word representation induced by a NNLM. The GALE Mandarin data was used to carry out the speech-to-text experiments and evaluate the NNLMs. On this data the well tuned baseline system has a character error rate under 10%. Our model achieves consistent improvements over the combination of an n-gram model and classical short-list NNLMs both in terms of perplexity and recognition accuracy.
Lecture Information
Recorded: | 2011-05-25 16:55 - 17:15, Club H |
---|---|
Added: | 9. 6. 2011 02:30 |
Number of views: | 49 |
Video resolution: | 1024x576 px, 512x288 px |
Video length: | 0:22:23 |
Audio track: | MP3 [7.65 MB], 0:22:23 |
Comments