Recurrent Neural Networks and Morphological Features in Language Modeling for Serbian

Author(s):  
Edvin T. Pakoci ◽  
Branislav Z. Popovic
2017 ◽  
Author(s):  
Zhe Gan ◽  
Chunyuan Li ◽  
Changyou Chen ◽  
Yunchen Pu ◽  
Qinliang Su ◽  
...  

2020 ◽  
Vol 34 (04) ◽  
pp. 4989-4996
Author(s):  
Ekaterina Lobacheva ◽  
Nadezhda Chirkova ◽  
Alexander Markovich ◽  
Dmitry Vetrov

One of the most popular approaches for neural network compression is sparsification — learning sparse weight matrices. In structured sparsification, weights are set to zero by groups corresponding to structure units, e. g. neurons. We further develop the structured sparsification approach for the gated recurrent neural networks, e. g. Long Short-Term Memory (LSTM). Specifically, in addition to the sparsification of individual weights and neurons, we propose sparsifying the preactivations of gates. This makes some gates constant and simplifies an LSTM structure. We test our approach on the text classification and language modeling tasks. Our method improves the neuron-wise compression of the model in most of the tasks. We also observe that the resulting structure of gate sparsity depends on the task and connect the learned structures to the specifics of the particular tasks.


Sign in / Sign up

Export Citation Format

Share Document