Gammatone spectral latitude features extraction for pathological voice detection and classification

2022 ◽  
Vol 185 ◽  
pp. 108417
Author(s):  
Changwei Zhou ◽  
Yuanbo Wu ◽  
Ziqi Fan ◽  
Xiaojun Zhang ◽  
Di Wu ◽  
...  
IEEE Access ◽  
2020 ◽  
Vol 8 ◽  
pp. 66749-66776
Author(s):  
Rumana Islam ◽  
Mohammed Tarique ◽  
Esam Abdel-Raheem

2021 ◽  
Author(s):  
Zhang Yihua ◽  
Zhu Xincheng ◽  
Wu Yuanbo ◽  
Zhang Xiaojun ◽  
Xu Yishen ◽  
...  

IEEE Access ◽  
2020 ◽  
Vol 8 ◽  
pp. 67745-67755 ◽  
Author(s):  
N. P. Narendra ◽  
Paavo Alku

2021 ◽  
Vol 11 (15) ◽  
pp. 7149
Author(s):  
Ji-Yeoun Lee

This work is focused on deep learning methods, such as feedforward neural network (FNN) and convolutional neural network (CNN), for pathological voice detection using mel-frequency cepstral coefficients (MFCCs), linear prediction cepstrum coefficients (LPCCs), and higher-order statistics (HOSs) parameters. In total, 518 voice data samples were obtained from the publicly available Saarbruecken voice database (SVD), comprising recordings of 259 healthy and 259 pathological women and men, respectively, and using /a/, /i/, and /u/ vowels at normal pitch. Significant differences were observed between the normal and the pathological voice signals for normalized skewness (p = 0.000) and kurtosis (p = 0.000), except for normalized kurtosis (p = 0.051) that was estimated in the /u/ samples in women. These parameters are useful and meaningful for classifying pathological voice signals. The highest accuracy, 82.69%, was achieved by the CNN classifier with the LPCCs parameter in the /u/ vowel in men. The second-best performance, 80.77%, was obtained with a combination of the FNN classifier, MFCCs, and HOSs for the /i/ vowel samples in women. There was merit in combining the acoustic measures with HOS parameters for better characterization in terms of accuracy. The combination of various parameters and deep learning methods was also useful for distinguishing normal from pathological voices.


2021 ◽  
Vol 0 (0) ◽  
Author(s):  
Lei Geng ◽  
Hongfeng Shan ◽  
Zhitao Xiao ◽  
Wei Wang ◽  
Mei Wei

Abstract Automatic voice pathology detection and classification plays an important role in the diagnosis and prevention of voice disorders. To accurately describe the pronunciation characteristics of patients with dysarthria and improve the effect of pathological voice detection, this study proposes a pathological voice detection method based on a multi-modal network structure. First, speech signals and electroglottography (EGG) signals are mapped from the time domain to the frequency domain spectrogram via a short-time Fourier transform (STFT). The Mel filter bank acts on the spectrogram to enhance the signal’s harmonics and denoise. Second, a pre-trained convolutional neural network (CNN) is used as the backbone network to extract sound state features and vocal cord vibration features from the two signals. To obtain a better classification effect, the fused features are input into the long short-term memory (LSTM) network for voice feature selection and enhancement. The proposed system achieves 95.73% for accuracy with 96.10% F1-score and 96.73% recall using the Saarbrucken Voice Database (SVD); thus, enabling a new method for pathological speech detection.


Sign in / Sign up

Export Citation Format

Share Document