Accuracy of Adaptive Procedure Estimates of PB-Max Level

1983 ◽  
Vol 48 (2) ◽  
pp. 202-209 ◽  
Author(s):  
Candace A. Kamm ◽  
Donald E. Morgan ◽  
Donald D. Dirks

A measure of maximum speech recognition ability (PB max) is often obtained in the clinical audiologic evaluation of patients. However, without developing a complete performance-intensity function (PIF), it is often difficult to determine an appropriate intensity level for measuring PB max. Levitt (1978) has described an adaptive procedure designed to estimate an intensity level at which maximum speech recognition performance could be measured. We have investigated the accuracy of this procedure by comparing speech recognition performance on a CNC word list presented at the level estimated by the adaptive procedure with maximum performance as measured on the listener's performance-intensity function. Using the presentation level indicated by the adaptive procedure, PB max scores were obtained for 12 of 16 normally hearing subjects and 19 of 25 listeners with mild-to-moderate sensorineural hearing loss. Similar accuracy was obtained for the hearing-impaired subjects using a single presentation level of 95 dB SPL, while use of a 40 dB SL speech level would have resulted in a lower proportion of PB-max measurements. Although PB max is most accurately estimated by delineating the entire performance-intensity function, the current results suggest that, if speech recognition is to be measured only at a single level, scores obtained at the level estimated by the adaptive procedure or at 95 dB SPL may serve as reasonable estimates of PB max for listeners with primarily cochlear hearing losses of ⩽ 50 dB.

Author(s):  
S. B. Rathna Kumar ◽  
Madhu Sudharshan Reddy. B ◽  
Sale Kranthi

<p class="abstract"><strong>Background:</strong> <span lang="EN-IN">The present study aimed to develop word lists in Telugu for assessing speech recognition threshold which might serve as equivalent and alternative forms to the existing word lists. </span></p><p class="abstract"><strong>Methods:</strong> <span lang="EN-IN">A total of two word lists were developed using compound words with each list consisting of 25 words. Equivalence analysis and performance-intensity function testing was carried out using two word lists on a total of 75 native speakers of Telugu who were equally divided into three groups.  </span></p><p class="abstract"><strong>Results:</strong> <span lang="EN-IN">The results revealed that there was no statistically significant difference (p&gt;0.05) in the speech recognition performance between three groups for each word list, and between two word lists for each group. Hence, the two word lists developed were found to be equally difficult for all the groups and can be used interchangeably. The performance-intensity (PI) function curve showed semi-linear function, and the subjects reached the beginning of the plateau at 3 dBSL where they reached more than 90% speech recognition score for two word lists, and reached 100% speech recognition score at 6 dBSL. The 50% speech recognition score which corresponds to SRT was obtained at less than 1.5 dBSL for two word lists suggesting good agreement between PTA and SRT. </span></p><p class="abstract"><strong>Conclusions:</strong> <span lang="EN-IN">The findings of the study are similar to the findings of existing word lists in Telugu. Thus the developed word lists in the present study can be considered equivalent and alternative forms to existing word lists in Telugu.</span></p>


1988 ◽  
Vol 31 (4) ◽  
pp. 681-695 ◽  
Author(s):  
Faith C. Loven ◽  
M. Jane Collins

The purpose of this investigation was to describe the interactive effects of four signal modifications typically encountered in everyday communication settings. These modifications included reverberation, masking, filtering, and fluctuation in speech intensity. The relationship between recognition performance and spectral changes to the speech signal due to the presence of these signal alterations was also studied. The interactive effects of these modifications were evaluated by obtaining indices of nonsense syllable recognition ability from normally hearing listeners for systematically varied combinations of the four signal parameters. The results of this study were in agreement with previous studies concerned with the effect of these variables in isolation on speech recognition ability. When present in combination, the direction of each variable's effect on recognition performance is maintained; however, the magnitude of the effect increases. The results of this investigation are reasonably accounted for by a spectral theory of speech recognition.


2018 ◽  
Vol 61 (4) ◽  
pp. 936-944 ◽  
Author(s):  
Robert S. Schlauch ◽  
Edward Carney

Purpose Computer simulation was used to estimate the statistical properties of searches for maximum word recognition ability (PB max). These involve presenting multiple lists and discarding all scores but that of the 1 list that produced the highest score. The simulations, which model limitations inherent in the precision of word recognition scores, were done to inform clinical protocols. A secondary consideration was a derivation of 95% confidence intervals for significant changes in score from phonemic scoring of a 50-word list. Method The PB max simulations were conducted on a “client” with flat performance intensity functions. The client's performance was assumed to be 60% initially and 40% for a second assessment. Thousands of estimates were obtained to examine the precision of (a) single lists and (b) multiple lists using a PB max procedure. This method permitted summarizing the precision for assessing a 20% drop in performance. Results A single 25-word list could identify only 58.4% of the cases in which performance fell from 60% to 40%. A single 125-word list identified 99.8% of the declines correctly. Presenting 3 or 5 lists to find PB max produced an undesirable finding: an increase in the word recognition score. Conclusions A 25-word list produces unacceptably low precision for making clinical decisions. This finding holds in both single and multiple 25-word lists, as in a search for PB max. A table is provided, giving estimates of 95% critical ranges for successive presentations of a 50-word list analyzed by the number of phonemes correctly identified.


2021 ◽  
Vol 15 ◽  
Author(s):  
Jing Chen ◽  
Zhe Wang ◽  
Ruijuan Dong ◽  
Xinxing Fu ◽  
Yuan Wang ◽  
...  

Objective: This study was aimed at evaluating improvements in speech-in-noise recognition ability as measured by signal-to-noise ratio (SNR) with the use of wireless remote microphone technology. These microphones transmit digital signals via radio frequency directly to hearing aids and may be a valuable assistive listening device for the hearing-impaired population of Mandarin speakers in China.Methods: Twenty-three adults (aged 19–80 years old) and fourteen children (aged 8–17 years old) with bilateral sensorineural hearing loss were recruited. The Mandarin Hearing in Noise Test was used to test speech recognition ability in adult subjects, and the Mandarin Hearing in Noise Test for Children was used for children. The subjects’ perceived SNR was measured using sentence recognition ability at three different listening distances of 1.5, 3, and 6 m. At each distance, SNR was obtained under three device settings: hearing aid microphone alone, wireless remote microphone alone, and hearing aid microphone and wireless remote microphone simultaneously.Results: At each test distance, for both adult and pediatric groups, speech-in-noise recognition thresholds were significantly lower with the use of the wireless remote microphone in comparison with the hearing aid microphones alone (P &lt; 0.05), indicating better SNR performance with the wireless remote microphone. Moreover, when the wireless remote microphone was used, test distance had no effect on speech-in-noise recognition for either adults or children.Conclusion: Wireless remote microphone technology can significantly improve speech recognition performance in challenging listening environments for Mandarin speaking hearing aid users in China.


2005 ◽  
Vol 16 (01) ◽  
pp. 054-062
Author(s):  
Rochelle Cherry ◽  
Adrienne Rubinstein

Single-case design with the randomization test (RT) has been proposed as an alternative to the binomial distribution (BD) tables of Thornton and Raffin (1978) to assess changes in speech recognition performance in individual subjects. The present study investigated whether data analyzed using both approaches would result in similar outcomes. Sixty-two adults with normal hearing were evaluated using phoneme scoring and a restricted alternating treatments design under two signal-to-noise conditions. Results revealed a significant correlation between the RT and a BD analysis using at least 50-word lists, although the BD analysis was a more sensitive measure. The absence of correlation between the RT with a BD analysis using 25-word lists challenges the common clinical practice of using reduced list size, and supports the use of phoneme scoring and other attempts to find clinically acceptable yet evidence-based solutions to overcome the conflict between time and accuracy.


2020 ◽  
Vol 24 ◽  
pp. 233121652098029
Author(s):  
Allison Trine ◽  
Brian B. Monson

Several studies have demonstrated that extended high frequencies (EHFs; >8 kHz) in speech are not only audible but also have some utility for speech recognition, including for speech-in-speech recognition when maskers are facing away from the listener. However, the contribution of EHF spectral versus temporal information to speech recognition is unknown. Here, we show that access to EHF temporal information improved speech-in-speech recognition relative to speech bandlimited at 8 kHz but that additional access to EHF spectral detail provided an additional small but significant benefit. Results suggest that both EHF spectral structure and the temporal envelope contribute to the observed EHF benefit. Speech recognition performance was quite sensitive to masker head orientation, with a rotation of only 15° providing a highly significant benefit. An exploratory analysis indicated that pure-tone thresholds at EHFs are better predictors of speech recognition performance than low-frequency pure-tone thresholds.


2012 ◽  
Vol 23 (08) ◽  
pp. 577-589 ◽  
Author(s):  
Mary Rudner ◽  
Thomas Lunner ◽  
Thomas Behrens ◽  
Elisabet Sundewall Thorén ◽  
Jerker Rönnberg

Background: Recently there has been interest in using subjective ratings as a measure of perceived effort during speech recognition in noise. Perceived effort may be an indicator of cognitive load. Thus, subjective effort ratings during speech recognition in noise may covary both with signal-to-noise ratio (SNR) and individual cognitive capacity. Purpose: The present study investigated the relation between subjective ratings of the effort involved in listening to speech in noise, speech recognition performance, and individual working memory (WM) capacity in hearing impaired hearing aid users. Research Design: In two experiments, participants with hearing loss rated perceived effort during aided speech perception in noise. Noise type and SNR were manipulated in both experiments, and in the second experiment hearing aid compression release settings were also manipulated. Speech recognition performance was measured along with WM capacity. Study Sample: There were 46 participants in all with bilateral mild to moderate sloping hearing loss. In Experiment 1 there were 16 native Danish speakers (eight women and eight men) with a mean age of 63.5 yr (SD = 12.1) and average pure tone (PT) threshold of 47. 6 dB (SD = 9.8). In Experiment 2 there were 30 native Swedish speakers (19 women and 11 men) with a mean age of 70 yr (SD = 7.8) and average PT threshold of 45.8 dB (SD = 6.6). Data Collection and Analysis: A visual analog scale (VAS) was used for effort rating in both experiments. In Experiment 1, effort was rated at individually adapted SNRs while in Experiment 2 it was rated at fixed SNRs. Speech recognition in noise performance was measured using adaptive procedures in both experiments with Dantale II sentences in Experiment 1 and Hagerman sentences in Experiment 2. WM capacity was measured using a letter-monitoring task in Experiment 1 and the reading span task in Experiment 2. Results: In both experiments, there was a strong and significant relation between rated effort and SNR that was independent of individual WM capacity, whereas the relation between rated effort and noise type seemed to be influenced by individual WM capacity. Experiment 2 showed that hearing aid compression setting influenced rated effort. Conclusions: Subjective ratings of the effort involved in speech recognition in noise reflect SNRs, and individual cognitive capacity seems to influence relative rating of noise type.


1981 ◽  
Vol 10 (4) ◽  
pp. 239-246 ◽  
Author(s):  
D. D. Dirks ◽  
C. A. Kamm ◽  
J. R. Dubno ◽  
T. M. Velde

Author(s):  
Jiahao Chen ◽  
Ryota Nishimura ◽  
Norihide Kitaoka

Many end-to-end, large vocabulary, continuous speech recognition systems are now able to achieve better speech recognition performance than conventional systems. Most of these approaches are based on bidirectional networks and sequence-to-sequence modeling however, so automatic speech recognition (ASR) systems using such techniques need to wait for an entire segment of voice input to be entered before they can begin processing the data, resulting in a lengthy time-lag, which can be a serious drawback in some applications. An obvious solution to this problem is to develop a speech recognition algorithm capable of processing streaming data. Therefore, in this paper we explore the possibility of a streaming, online, ASR system for Japanese using a model based on unidirectional LSTMs trained using connectionist temporal classification (CTC) criteria, with local attention. Such an approach has not been well investigated for use with Japanese, as most Japanese-language ASR systems employ bidirectional networks. The best result for our proposed system during experimental evaluation was a character error rate of 9.87%.


2008 ◽  
Vol 19 (02) ◽  
pp. 120-134 ◽  
Author(s):  
Kate Gfeller ◽  
Jacob Oleson ◽  
John F. Knutson ◽  
Patrick Breheny ◽  
Virginia Driscoll ◽  
...  

The research examined whether performance by adult cochlear implant recipients on a variety of recognition and appraisal tests derived from real-world music could be predicted from technological, demographic, and life experience variables, as well as speech recognition scores. A representative sample of 209 adults implanted between 1985 and 2006 participated. Using multiple linear regression models and generalized linear mixed models, sets of optimal predictor variables were selected that effectively predicted performance on a test battery that assessed different aspects of music listening. These analyses established the importance of distinguishing between the accuracy of music perception and the appraisal of musical stimuli when using music listening as an index of implant success. Importantly, neither device type nor processing strategy predicted music perception or music appraisal. Speech recognition performance was not a strong predictor of music perception, and primarily predicted music perception when the test stimuli included lyrics. Additionally, limitations in the utility of speech perception in predicting musical perception and appraisal underscore the utility of music perception as an alternative outcome measure for evaluating implant outcomes. Music listening background, residual hearing (i.e., hearing aid use), cognitive factors, and some demographic factors predicted several indices of perceptual accuracy or appraisal of music. La investigación examinó si el desempeño, por parte de adultos receptores de un implante coclear, sobre una variedad de pruebas de reconocimiento y evaluación derivadas de la música del mundo real, podrían predecirse a partir de variables tecnológicas, demográficas y de experiencias de vida, así como de puntajes de reconocimiento del lenguaje. Participó una muestra representativa de 209 adultos implantados entre 1965 y el 2006. Usando múltiples modelos de regresión lineal y modelos mixtos lineales generalizados, se seleccionaron grupos de variables óptimas de predicción, que pudieran predecir efectivamente el desempeño por medio de una batería de pruebas que permitiera evaluar diferentes aspectos de la apreciación musical. Estos análisis establecieron la importancia de distinguir entre la exactitud en la percepción musical y la evaluación de estímulos musicales cuando se utiliza la apreciación musical como un índice de éxito en la implantación. Importantemente, ningún tipo de dispositivo o estrategia de procesamiento predijo la percepción o la evaluación musical. El desempeño en el reconocimiento del lenguaje no fue un elemento fuerte de predicción, y llegó a predecir primariamente la percepción musical cuando los estímulos de prueba incluyeron las letras. Adicionalmente, las limitaciones en la utilidad de la percepción del lenguaje a la hora de predecir la percepción y la evaluación musical, subrayan la utilidad de la percepción de la música como una medida alternativa de resultado para evaluar la implantación coclear. La música de fondo, la audición residual (p.e., el uso de auxiliares auditivos), los factores cognitivos, y algunos factores demográficos predijeron varios índices de exactitud y evaluación perceptual de la música.


Sign in / Sign up

Export Citation Format

Share Document