Back to Search Start Over

Markov Recurrent Neural Network Language Model

Authors :
Che-Yu Kuo
Jen-Tzung Chien
Source :
ASRU
Publication Year :
2019
Publisher :
IEEE, 2019.

Abstract

Recurrent neural network (RNN) has achieved a great success in language modeling where the temporal information based on deterministic state is continuously extracted and evolved through time. Such a simple deterministic transition function using input-to-hidden and hidden-to-hidden weights is usually insufficient to reflect the diversities and variations of latent variable structure behind the heterogeneous natural language. This paper presents a new stochastic Markov RNN (MRNN) to strengthen the learning capability in language model where the trajectory of word sequences is driven by a neural Markov process with Markov state transitions based on a K-state long short-term memory model. A latent state machine is constructed to characterize the complicated semantics in the structured lexical patterns. Gumbel-softmax is introduced to implement the stochastic backpropatation algorithm with discrete states. The parallel computation for rapid realization of MRNN is presented. The variational Bayesian learning procedure is implemented. Experiments demonstrate the merits of stochastic and diverse representation using MRNN language model where the overhead of parameters and computations is limited.

Details

Database :
OpenAIRE
Journal :
2019 IEEE Automatic Speech Recognition and Understanding Workshop (ASRU)
Accession number :
edsair.doi...........0b42cc57c8fcb0e6dd8a23b4d1ff5896
Full Text :
https://doi.org/10.1109/asru46091.2019.9003850