scholarly journals Generalized locally recurrent probabilistic neural networks for text-independent speaker verification

Author(s):  
T. Ganchev ◽  
N. Fakotakis ◽  
D.K. Tasoulis ◽  
M.N. Vrahatis
2007 ◽  
Vol 70 (7-9) ◽  
pp. 1424-1438 ◽  
Author(s):  
Todor D. Ganchev ◽  
Dimitris K. Tasoulis ◽  
Michael N. Vrahatis ◽  
Nikos D. Fakotakis

2009 ◽  
Vol 18 (06) ◽  
pp. 853-881 ◽  
Author(s):  
TODOR GANCHEV

In the present contribution we propose an integral training procedure for the Locally Recurrent Probabilistic Neural Networks (LR PNNs). Specifically, the adjustment of the smoothing factor "sigma" in the pattern layer of the LR PNN and the training of the recurrent layer weights are integrated in an automatic process that iteratively estimates all adjustable parameters of the LR PNN from the available training data. Furthermore, in contrast to the original LR PNN, whose recurrent layer was trained to provide optimum separation among the classes on the training dataset, while striving to keep a balance between the learning rates for all classes, here the training strategy is oriented towards optimizing the overall classification accuracy, straightforwardly. More precisely, the new training strategy directly targets at maximizing the posterior probabilities for the target class and minimizing the posterior probabilities estimated for the non-target classes. The new fitness function requires fewer computations for each evaluation, and therefore the overall computational demands for training the recurrent layer weights are reduced. The performance of the integrated training procedure is illustrated on three different speech processing tasks: emotion recognition, speaker identification and speaker verification.


THE BULLETIN ◽  
2020 ◽  
Vol 5 (387) ◽  
pp. 6-15
Author(s):  
O. Mamyrbayev ◽  
◽  
A. Akhmediyarova ◽  
A. Kydyrbekova ◽  
N. O. Mekebayev ◽  
...  

Biometrics offers more security and convenience than traditional methods of identification. Recently, DNN has become a means of a more reliable and efficient authentication scheme. In this work, we compare two modern teaching methods: these two methods are methods based on the Gaussian mixture model (GMM) (denoted by the GMM i-vector) and methods based on deep neural networks (DNN) (denoted as the i-vector DNN). The results show that the DNN system with an i-vector is superior to the GMM system with an i-vector for various durations (from full length to 5s). DNNs have proven to be the most effective features for text-independent speaker verification in recent studies. In this paper, a new scheme is proposed that allows using DNN when checking text using hints in a simple and effective way. Experiments show that the proposed scheme reduces EER by 24.32% compared with the modern method and is evaluated for its reliability using noisy data, as well as data collected in real conditions. In addition, it is shown that the use of DNN instead of GMM for universal background modeling leads to a decrease in EER by 15.7%.


2002 ◽  
Vol 12 (05) ◽  
pp. 381-397 ◽  
Author(s):  
K. K. YIU ◽  
M. W. MAK ◽  
S. Y. KUNG

This paper compares kernel-based probabilistic neural networks for speaker verification based on 138 speakers of the YOHO corpus. Experimental evaluations using probabilistic decision-based neural networks (PDBNNs), Gaussian mixture models (GMMs) and elliptical basis function networks (EBFNs) as speaker models were conducted. The original training algorithm of PDBNNs was also modified to make PDBNNs appropriate for speaker verification. Results show that the equal error rate obtained by PDBNNs and GMMs is less than that of EBFNs (0.33% vs. 0.48%), suggesting that GMM- and PDBNN-based speaker models outperform the EBFN ones. This work also finds that the globally supervised learning of PDBNNs is able to find decision thresholds that not only maintain the false acceptance rates to a low level but also reduce their variation, whereas the ad-hoc threshold-determination approach used by the EBFNs and GMMs causes a large variation in the error rates. This property makes the performance of PDBNN-based systems more predictable.


Sign in / Sign up

Export Citation Format

Share Document