consonant recognition
Recently Published Documents


TOTAL DOCUMENTS

139
(FIVE YEARS 11)

H-INDEX

27
(FIVE YEARS 0)

2021 ◽  
Vol 15 ◽  
Author(s):  
Zhong Zheng ◽  
Keyi Li ◽  
Gang Feng ◽  
Yang Guo ◽  
Yinan Li ◽  
...  

Objectives: Mandarin-speaking users of cochlear implants (CI) perform poorer than their English counterpart. This may be because present CI speech coding schemes are largely based on English. This study aims to evaluate the relative contributions of temporal envelope (E) cues to Mandarin phoneme (including vowel, and consonant) and lexical tone recognition to provide information for speech coding schemes specific to Mandarin.Design: Eleven normal hearing subjects were studied using acoustic temporal E cues that were extracted from 30 continuous frequency bands between 80 and 7,562 Hz using the Hilbert transform and divided into five frequency regions. Percent-correct recognition scores were obtained with acoustic E cues presented in three, four, and five frequency regions and their relative weights calculated using the least-square approach.Results: For stimuli with three, four, and five frequency regions, percent-correct scores for vowel recognition using E cues were 50.43–84.82%, 76.27–95.24%, and 96.58%, respectively; for consonant recognition 35.49–63.77%, 67.75–78.87%, and 87.87%; for lexical tone recognition 60.80–97.15%, 73.16–96.87%, and 96.73%. For frequency region 1 to frequency region 5, the mean weights in vowel recognition were 0.17, 0.31, 0.22, 0.18, and 0.12, respectively; in consonant recognition 0.10, 0.16, 0.18, 0.23, and 0.33; in lexical tone recognition 0.38, 0.18, 0.14, 0.16, and 0.14.Conclusion: Regions that contributed most for vowel recognition was Region 2 (502–1,022 Hz) that contains first formant (F1) information; Region 5 (3,856–7,562 Hz) contributed most to consonant recognition; Region 1 (80–502 Hz) that contains fundamental frequency (F0) information contributed most to lexical tone recognition.


2021 ◽  
Vol 12 ◽  
Author(s):  
Yang-Soo Yoon

In this paper, the effects of intensifying useful frequency and time regions (target frequency and time ranges) and the removal of detrimental frequency and time regions (conflicting frequency and time ranges) for consonant enhancement were determined. Thirteen normal-hearing (NH) listeners participated in two experiments. In the first experiment, the target and conflicting frequency and time ranges for each consonant were identified under a quiet, dichotic listening condition by analyzing consonant confusion matrices. The target frequency range was defined as the frequency range that provided the highest performance and was decreased 40% from the peak performance from both high-pass filtering (HPF) and low-pass filtering (LPF) schemes. The conflicting frequency range was defined as the frequency range that yielded the peak errors of the most confused consonants and was 20% less than the peak error from both filtering schemes. The target time range was defined as a consonant segment that provided the highest performance and was decreased 40% from that peak performance when the duration of the consonant was systematically truncated from the onset. The conflicting time ranges were defined on the coincided target time range because, if they temporarily coincide, the conflicting frequency ranges would be the most detrimental factor affecting the target frequency ranges. In the second experiment, consonant recognition was binaurally measured in noise under three signal processing conditions: unprocessed, intensified target ranges by a 6-dB gain (target), and combined intensified target and removed conflicting ranges (target-conflicting). The results showed that consonant recognition improved significantly with the target condition but greatly deteriorated with a target-conflicting condition. The target condition helped transmit voicing and manner cues while the target-conflicting condition limited the transmission of these cues. Confusion analyses showed that the effect of the signal processing on consonant improvement was consonant-specific: the unprocessed condition was the best for /da, pa, ma, sa/; the target condition was the best for /ga, fa, va, za, ʒa/; and the target-conflicting condition was the best for /na, ʃa/. Perception of /ba, ta, ka/ was independent of the signal processing. The results suggest that enhancing the target ranges is an efficient way to improve consonant recognition while the removal of conflicting ranges negatively impacts consonant recognition.


2021 ◽  
Vol 32 (08) ◽  
pp. 521-527
Author(s):  
Yang-Soo Yoon ◽  
George Whitaker ◽  
Yune S. Lee

Abstract Background Cochlear implant technology allows for acoustic and electric stimulations to be combined across ears (bimodal) and within the same ear (electric acoustic stimulation [EAS]). Mechanisms used to integrate speech acoustics may be different between the bimodal and EAS hearing, and the configurations of hearing loss might be an important factor for the integration. Thus, differentiating the effects of different configurations of hearing loss on bimodal or EAS benefit in speech perception (differences in performance with combined acoustic and electric stimulations from a better stimulation alone) is important. Purpose Using acoustic simulation, we determined how consonant recognition was affected by different configurations of hearing loss in bimodal and EAS hearing. Research Design A mixed design was used with one between-subject variable (simulated bimodal group vs. simulated EAS group) and one within-subject variable (acoustic stimulation alone, electric stimulation alone, and combined acoustic and electric stimulations). Study Sample Twenty adult subjects (10 for each group) with normal hearing were recruited. Data Collection and Analysis Consonant perception was unilaterally or bilaterally measured in quiet. For the acoustic stimulation, four different simulations of hearing loss were created by band-pass filtering consonants with a fixed lower cutoff frequency of 100 Hz and each of the four upper cutoff frequencies of 250, 500, 750, and 1,000 Hz. For the electric stimulation, an eight-channel noise vocoder was used to generate a typical spectral mismatch by using fixed input (200–7,000 Hz) and output (1,000–7,000 Hz) frequency ranges. The effects of simulated hearing loss on consonant recognition were compared between the two groups. Results Significant bimodal and EAS benefits occurred regardless of the configurations of hearing loss and hearing technology (bimodal vs. EAS). Place information was better transmitted in EAS hearing than in bimodal hearing. Conclusion These results suggest that configurations of hearing loss are not a significant factor for integrating consonant information between acoustic and electric stimulations. The results also suggest that mechanisms used to integrate consonant information may be similar between bimodal and EAS hearing.


2021 ◽  
Vol 15 (3) ◽  
pp. 253-265
Author(s):  
Kyung-ah Yu

The purpose of this study is intended to empirically examine whether or not communication-oriented pronunciation teaching affects the English segmental and suprasegmental pronunciation improvements of Korean college learners, as well as their English speaking ability. The students who participated in this study were 26 college students, all of whom were taking an English pronunciation clinic course at a local university. Three types of pre and post-tests were conducted to find out how communication-oriented pronunciation teaching influenced the learners’ pronunciation. Namely, tests of consonant and vowel sound recognition, reading aloud tests, and picture description tests.First, a total of 40 questions were given as pre-and post-pronunciation recognition evaluations to find out how accurately the students perceived the sounds of consonants and vowels. In addition, the students’ voices were recorded as they read dialogue readings both before and after the evaluation, in order to examine the degree of improvement of the segmental and suprasegmental features. Finally, the students were asked to describe certain pictures before and after the course to see if there were any beneficial effects regarding their pronunciation of English.The findings showed that communication-oriented activities led to significant improvements, but only on the consonant recognition tests, not on the tests with vowels. The results from the reading aloud tests also demonstrated that the communication-oriented activities helped students to improve suprasegmental elements, such as stress, intonation, and rhythm. Moreover, there was a significant improvement in the speaking ability of the students, from which I can infer that communication-oriented activities have a positive effect on improving the students’ suprasegmentals, as well as their speaking ability.


2021 ◽  
Vol 6 (1) ◽  
Author(s):  
Gabriela Diamante

Introduction: In the present work, a review is conducted on hearing loss, its origin, the consequences of the lack of a timely treatment, the possibilities and benefits provided by the access to a cochlear implant and the corresponding auditory rehabilitation. Materials and Methods: The methods used in the present work were Ling Test, Vowel Recognition, Consonant Recognition, Two-syllable Word Test, and Open-set Sentence Recognition. Results: By means of a descriptive study of a population that received a cochlear implant and after one year of use, we can observe high rates of auditory discrimination in all the variables presented and their levels of requirement and complexity, showing a close relationship between an appropriate and timely approach and a successful auditory rehabilitation. Conclusion: In all of the cases in this sample, the cochlear implant led to an improved quality of life for both the patients and their family members, allowing children to develop the language and improve their performance in school activities, and adults to appropriately reintegrate into society.


2020 ◽  
Vol 29 (4) ◽  
pp. 916-929
Author(s):  
Yihsin Tai ◽  
Fatima T. Husain

Purpose Difficulties in speech-in-noise understanding are often reported in individuals with tinnitus. Building on our previous findings that speech-in-noise performance is correlated with subjective loudness of tinnitus, this study aimed to investigate the effect of tinnitus pitch on consonant recognition in noise. Method Pure-tone audiometry and the Quick Speech-in-Noise Test were conducted on 66 participants categorized into four groups by their hearing sensitivity and self-report of tinnitus. Consonant recognition scores at various frequency ranges were obtained at the 5 dB SNR condition of the Quick Speech-in-Noise Test. Participants with tinnitus also completed a tinnitus pitch-matching procedure. Correlation analyses were conducted between tinnitus pitch and the frequency of the worst consonant recognition, and the error rates based on word and sentence position were compared. Results Regardless of hearing sensitivity, tinnitus pitch did not correlate with the frequency of the worst consonant recognition. Sentence-initial word recognition was affected by hearing loss, whereas sentence-final word recognition was not affected by hearing loss or tinnitus. In contrast to individuals with normal hearing, participants with hearing loss varied in full-sentence recognition, with those reporting tinnitus exhibiting significantly higher error rates. Conclusions The findings suggest that the effect of tinnitus on consonant recognition in noise may involve higher level functions more than perceptual characteristics of tinnitus. Furthermore, for individuals with speech-in-noise concerns, clinical evaluation should address both hearing sensitivity and the presence of tinnitus. Future speech-in-noise studies should incorporate cognitive tests and, possibly, brain imaging to parse out the contribution of cognitive factors, such as cognitive control, in speech-in-noise in tinnitus.


2020 ◽  
Vol 147 (5) ◽  
pp. 3712-3727
Author(s):  
Joshua G. W. Bernstein ◽  
Jonathan H. Venezia ◽  
Ken W. Grant

2019 ◽  
Vol 128 (6_suppl) ◽  
pp. 139S-145S
Author(s):  
Yang-Soo Yoon ◽  
Britteny Riley ◽  
Henna Patel ◽  
Amanda Frost ◽  
Paul Fillmore ◽  
...  

Objectives: The present study investigated the effects of 3-dimensional deep search (3DDS) signal processing on the enhancement of consonant perception in bimodal and normal hearing listeners. Methods: Using an articulation-index gram and 3DDS signal processing, consonant segments that greatly affected performance were identified and intensified with a 6-dB gain. Then consonant recognition was measured unilaterally and bilaterally before and after 3DDS processing both in quiet and noise. Results: The 3DDS signal processing provided a benefit to both groups, with greater benefit occurring in noise than quiet. The benefit rendered by 3DDS was the greatest in binaural listening condition. Ability to integrate acoustic features across ears was also enhanced with 3DDS processing. In listeners with normal hearing, manner and place of articulation were improved in binaural listening condition. In bimodal listeners, voicing and manner and place of articulation were also improved in bimodal and hearing aid ear–alone conditions. Conclusions: Consonant recognition was improved with 3DDS in both groups. This observed benefit suggests 3DDS can be used as an auditory training tool for improved integration and for bimodal users who receive little or no benefit from their current bimodal hearing.


Sign in / Sign up

Export Citation Format

Share Document