Improving handwritten Chinese text recognition using neural network language models and convolutional neural network shape models

2017 ◽  
Vol 65 ◽  
pp. 251-264 ◽  
Author(s):  
Yi-Chao Wu ◽  
Fei Yin ◽  
Cheng-Lin Liu
2016 ◽  
Author(s):  
Ngoc-Quan Pham ◽  
Germán Kruszewski ◽  
Gemma Boleda

2018 ◽  
Vol 28 (09) ◽  
pp. 1850007
Author(s):  
Francisco Zamora-Martinez ◽  
Maria Jose Castro-Bleda

Neural Network Language Models (NNLMs) are a successful approach to Natural Language Processing tasks, such as Machine Translation. We introduce in this work a Statistical Machine Translation (SMT) system which fully integrates NNLMs in the decoding stage, breaking the traditional approach based on [Formula: see text]-best list rescoring. The neural net models (both language models (LMs) and translation models) are fully coupled in the decoding stage, allowing to more strongly influence the translation quality. Computational issues were solved by using a novel idea based on memorization and smoothing of the softmax constants to avoid their computation, which introduces a trade-off between LM quality and computational cost. These ideas were studied in a machine translation task with different combinations of neural networks used both as translation models and as target LMs, comparing phrase-based and [Formula: see text]-gram-based systems, showing that the integrated approach seems more promising for [Formula: see text]-gram-based systems, even with nonfull-quality NNLMs.


Sign in / Sign up

Export Citation Format

Share Document