Language identification of code Switching sentences and multilingual sentences of under-resourced languages by using multi structural word information

Author(s):  
Yin-Lai Yeong ◽  
Tien-Ping Tan
2019 ◽  
Author(s):  
Manuel Mager ◽  
Özlem Çetinoğlu ◽  
Katharina Kann

2021 ◽  
Vol 11 (19) ◽  
pp. 9106
Author(s):  
Zheying Huang ◽  
Pei Wang ◽  
Jian Wang ◽  
Haoran Miao ◽  
Ji Xu ◽  
...  

A Recurrent Neural Networks (RNN) based attention model has been used in code-switching speech recognition (CSSR). However, due to the sequential computation constraint of RNN, there are stronger short-range dependencies and weaker long-range dependencies, which makes it hard to immediately switch languages in CSSR. Firstly, to deal with this problem, we introduce the CTC-Transformer, relying entirely on a self-attention mechanism to draw global dependencies and adopting connectionist temporal classification (CTC) as an auxiliary task for better convergence. Secondly, we proposed two multi-task learning recipes, where a language identification (LID) auxiliary task is learned in addition to the CTC-Transformer automatic speech recognition (ASR) task. Thirdly, we study a decoding strategy to combine the LID into an ASR task. Experiments on the SEAME corpus demonstrate the effects of the proposed methods, achieving a mixed error rate (MER) of 30.95%. It obtains up to 19.35% relative MER reduction compared to the baseline RNN-based CTC-Attention system, and 8.86% relative MER reduction compared to the baseline CTC-Transformer system.


Sign in / Sign up

Export Citation Format

Share Document