Multi-stream long short-term memory neural network language model
MetadataShow full item record
CitationArisoy, E., Saraclar, M., (2015). Multi-stream long short-term memory neural network language model. Conference: 16th Annual Conference of the International-Speech-Communication-Association (INTERSPEECH 2015) Location: Dresden, GERMANY, vol: 1-5. p. 1413-1417.
Long Short-Term Memory (LSTM) neural networks are recurrent neural networks that contain memory units that can store contextual information from past inputs for arbitrary amounts of time. A typical LSTM neural network language model is trained by feeding an input sequence. i.e., a stream of words, to the input layer of the network and the output layer predicts the probability of the next word given the past inputs in the sequence. In this paper we introduce a multi-stream LSTM neural network language model where multiple asynchronous input sequences are fed to the network as parallel streams while predicting the output word sequence. For our experiments, we use a sub-word sequence in addition to a word sequence as the input streams, which allows joint training of the LSTM neural network language model using both information sources.