Conversational Speech Emotion Recognition From Indonesian Spoken Language Using Recurrent Neural Network-Based Model

Author(s):  
Aisyah Nurul Izzah Adma ◽  
Dessi Puji Lestari
IEEE Access ◽  
2019 ◽  
Vol 7 ◽  
pp. 90368-90377 ◽  
Author(s):  
Pengxu Jiang ◽  
Hongliang Fu ◽  
Huawei Tao ◽  
Peizhi Lei ◽  
Li Zhao

IEEE Access ◽  
2020 ◽  
Vol 8 ◽  
pp. 199909-199919
Author(s):  
Xusheng Ai ◽  
Victor S. Sheng ◽  
Wei Fang ◽  
Charles X. Ling ◽  
Chunhua Li

2021 ◽  
Vol 11 (4) ◽  
pp. 1890
Author(s):  
Sung-Woo Byun ◽  
Seok-Pil Lee

The goal of the human interface is to recognize the user’s emotional state precisely. In the speech emotion recognition study, the most important issue is the effective parallel use of the extraction of proper speech features and an appropriate classification engine. Well defined speech databases are also needed to accurately recognize and analyze emotions from speech signals. In this work, we constructed a Korean emotional speech database for speech emotion analysis and proposed a feature combination that can improve emotion recognition performance using a recurrent neural network model. To investigate the acoustic features, which can reflect distinct momentary changes in emotional expression, we extracted F0, Mel-frequency cepstrum coefficients, spectral features, harmonic features, and others. Statistical analysis was performed to select an optimal combination of acoustic features that affect the emotion from speech. We used a recurrent neural network model to classify emotions from speech. The results show the proposed system has more accurate performance than previous studies.


Sign in / Sign up

Export Citation Format

Share Document