A comparison of a priori threshold setting procedures for speaker verification in the CAVE project

Author(s):  
J.-B. Pierrot ◽  
J. Lindberg ◽  
J. Koolwaaij ◽  
H.-P. Hutter ◽  
D. Genoud ◽  
...  
Author(s):  
Сергей Клавдиевич Абрамов ◽  
Виктория Валерьевна Абрамова ◽  
Сергей Станиславович Кривенко ◽  
Владимир Васильевич Лукин

The article deals with the analysis of the efficiency and expedience of applying filtering based on the discrete cosine transform (DCT) for one-dimensional signals distorted by white Gaussian noise with a known or a priori estimated variance. It is shown that efficiency varies in wide limits depending upon the input ratio of signal-to-noise and degree of processed signal complexity. It is offered a method for predicting filtering efficiency according to the traditional quantitative criteria as the ratio of mean square error to the variance of additive noise and improvement of the signal-to-noise ratio. Forecasting is performed based on dependences obtained by regression analysis. These dependencies can be described by simple functions of several types parameters of which are determined as the result of least mean square fitting. It is shown that for sufficiently accurate prediction, only one statistical parameter calculated in the DCT domain can be preliminarily evaluated (before filtering), and this parameter can be calculated in a relatively small number of non-overlapping or partially overlapping blocks of standard size (for example, 32 samples). It is analyzed the variations of efficiency criteria variations for a set of realizations; it is studied factors that influence prediction accuracy. It is demonstrated that it is possible to carry out the forecasting of filtering efficiency for several possible values of the DCT-filter parameter used for threshold setting and, then, to recommend the best value for practical use. An example of using such an adaptation procedure for the filter parameter setting for processing the ECG signal that has not been used in the determination of regression dependences is given. As a result of adaptation, the efficiency of filtering can be essentially increased – benefit can reach 0.5-1 dB. An advantage of the proposed procedures of adaptation and prediction is their universality – they can be applied for different types of signals and different ratios of signal-to-noise.


Author(s):  
Tuan Pham ◽  
◽  
Michael Wagner ◽  

Most speaker verification systems are based on similarity or likelihood normalization techniques as they help to better cope with speaker variability. In the conventional normalization, the it a priori probabilities of the cohort speakers are assumed to be equal. From this standpoint, we apply the fuzzy integral and genetic algorithms to combine the likelihood values of the cohort speakers in which the assumption of equal <I>a priori</I> probabilities is relaxed. This approach replaces the conventional normalization term by the fuzzy integral which acts as a non-linear fusion of the similarity measures of an utterance assigned to the cohort speakers. Furthermore, genetic algorithms are applied to find optimal fuzzy densities which are very important for the fuzzy fusion. We illustrate the performance of the proposed approach by testing the speaker verification system with both the conventional and the proposed algorithms using the commercial speech corpus TI46. The results in terms of the equal error rates show that the speaker verification system using the fuzzy integral is more favorable than the conventional normalization method.


Author(s):  
D. E. Luzzi ◽  
L. D. Marks ◽  
M. I. Buckett

As the HREM becomes increasingly used for the study of dynamic localized phenomena, the development of techniques to recover the desired information from a real image is important. Often, the important features are not strongly scattering in comparison to the matrix material in addition to being masked by statistical and amorphous noise. The desired information will usually involve the accurate knowledge of the position and intensity of the contrast. In order to decipher the desired information from a complex image, cross-correlation (xcf) techniques can be utilized. Unlike other image processing methods which rely on data massaging (e.g. high/low pass filtering or Fourier filtering), the cross-correlation method is a rigorous data reduction technique with no a priori assumptions.We have examined basic cross-correlation procedures using images of discrete gaussian peaks and have developed an iterative procedure to greatly enhance the capabilities of these techniques when the contrast from the peaks overlap.


Author(s):  
H.S. von Harrach ◽  
D.E. Jesson ◽  
S.J. Pennycook

Phase contrast TEM has been the leading technique for high resolution imaging of materials for many years, whilst STEM has been the principal method for high-resolution microanalysis. However, it was demonstrated many years ago that low angle dark-field STEM imaging is a priori capable of almost 50% higher point resolution than coherent bright-field imaging (i.e. phase contrast TEM or STEM). This advantage was not exploited until Pennycook developed the high-angle annular dark-field (ADF) technique which can provide an incoherent image showing both high image resolution and atomic number contrast.This paper describes the design and first results of a 300kV field-emission STEM (VG Microscopes HB603U) which has improved ADF STEM image resolution towards the 1 angstrom target. The instrument uses a cold field-emission gun, generating a 300 kV beam of up to 1 μA from an 11-stage accelerator. The beam is focussed on to the specimen by two condensers and a condenser-objective lens with a spherical aberration coefficient of 1.0 mm.


2019 ◽  
Vol 4 (5) ◽  
pp. 878-892
Author(s):  
Joseph A. Napoli ◽  
Linda D. Vallino

Purpose The 2 most commonly used operations to treat velopharyngeal inadequacy (VPI) are superiorly based pharyngeal flap and sphincter pharyngoplasty, both of which may result in hyponasal speech and airway obstruction. The purpose of this article is to (a) describe the bilateral buccal flap revision palatoplasty (BBFRP) as an alternative technique to manage VPI while minimizing these risks and (b) conduct a systematic review of the evidence of BBFRP on speech and other clinical outcomes. A report comparing the speech of a child with hypernasality before and after BBFRP is presented. Method A review of databases was conducted for studies of buccal flaps to treat VPI. Using the principles of a systematic review, the articles were read, and data were abstracted for study characteristics that were developed a priori. With respect to the case report, speech and instrumental data from a child with repaired cleft lip and palate and hypernasal speech were collected and analyzed before and after surgery. Results Eight articles were included in the analysis. The results were positive, and the evidence is in favor of BBFRP in improving velopharyngeal function, while minimizing the risk of hyponasal speech and obstructive sleep apnea. Before surgery, the child's speech was characterized by moderate hypernasality, and after surgery, it was judged to be within normal limits. Conclusion Based on clinical experience and results from the systematic review, there is sufficient evidence that the buccal flap is effective in improving resonance and minimizing obstructive sleep apnea. We recommend BBFRP as another approach in selected patients to manage VPI. Supplemental Material https://doi.org/10.23641/asha.9919352


Sign in / Sign up

Export Citation Format

Share Document