Back to Search
Start Over
Efficient Language Modeling with Automatic Relevance Determination in Recurrent Neural Networks
- Source :
- RepL4NLP@ACL, Scopus-Elsevier
- Publication Year :
- 2019
- Publisher :
- Association for Computational Linguistics, 2019.
-
Abstract
- Reduction of the number of parameters is one of the most important goals in Deep Learning. In this article we propose an adaptation of Doubly Stochastic Variational Inference for Automatic Relevance Determination (DSVI-ARD) for neural networks compression. We find this method to be especially useful in language modeling tasks, where large number of parameters in the input and output layers is often excessive. We also show that DSVI-ARD can be applied together with encoder-decoder weight tying allowing to achieve even better sparsity and performance. Our experiments demonstrate that more than 90% of the weights in both encoder and decoder layers can be removed with a minimal quality loss.
- Subjects :
- 0209 industrial biotechnology
Artificial neural network
Computer science
business.industry
Deep learning
Inference
02 engineering and technology
Machine learning
computer.software_genre
Reduction (complexity)
020901 industrial engineering & automation
Recurrent neural network
0202 electrical engineering, electronic engineering, information engineering
020201 artificial intelligence & image processing
Relevance (information retrieval)
Artificial intelligence
Language model
business
Encoder
computer
Subjects
Details
- Database :
- OpenAIRE
- Journal :
- Proceedings of the 4th Workshop on Representation Learning for NLP (RepL4NLP-2019)
- Accession number :
- edsair.doi.dedup.....8aab9417fbc8aa306598ce639e0797d3
- Full Text :
- https://doi.org/10.18653/v1/w19-4306