TY - JOUR
T1 - A survey on the application of recurrent neural networks to statistical language modeling
AU - De Mulder, Wim
AU - Bethard, Steven
AU - Moens, Marie Francine
N1 - Funding Information:
The research was financed by the EU FP7-296703 (FET-open call) project MUSE (Machine Understanding for interactive StorytElling). We thank the anonymous reviewers for their valuable remarks and suggestions.
Publisher Copyright:
© 2014 The Authors. Published by Elsevier Ltd.
PY - 2015/3
Y1 - 2015/3
N2 - In this paper, we present a survey on the application of recurrent neural networks to the task of statistical language modeling. Although it has been shown that these models obtain good performance on this task, often superior to other state-of-the-art techniques, they suffer from some important drawbacks, including a very long training time and limitations on the number of context words that can be taken into account in practice. Recent extensions to recurrent neural network models have been developed in an attempt to address these drawbacks. This paper gives an overview of the most important extensions. Each technique is described and its performance on statistical language modeling, as described in the existing literature, is discussed. Our structured overview makes it possible to detect the most promising techniques in the field of recurrent neural networks, applied to language modeling, but it also highlights the techniques for which further research is required.
AB - In this paper, we present a survey on the application of recurrent neural networks to the task of statistical language modeling. Although it has been shown that these models obtain good performance on this task, often superior to other state-of-the-art techniques, they suffer from some important drawbacks, including a very long training time and limitations on the number of context words that can be taken into account in practice. Recent extensions to recurrent neural network models have been developed in an attempt to address these drawbacks. This paper gives an overview of the most important extensions. Each technique is described and its performance on statistical language modeling, as described in the existing literature, is discussed. Our structured overview makes it possible to detect the most promising techniques in the field of recurrent neural networks, applied to language modeling, but it also highlights the techniques for which further research is required.
KW - Language modeling
KW - Machine translation
KW - Natural language processing
KW - Recurrent neural networks
KW - Speech recognition
UR - http://www.scopus.com/inward/record.url?scp=84913601027&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=84913601027&partnerID=8YFLogxK
U2 - 10.1016/j.csl.2014.09.005
DO - 10.1016/j.csl.2014.09.005
M3 - Review article
AN - SCOPUS:84913601027
VL - 30
SP - 61
EP - 98
JO - Computer Speech and Language
JF - Computer Speech and Language
SN - 0885-2308
IS - 1
ER -