Recognition of vocal emotion in noise-vocoded speech by normal hearing and cochlear implant listeners

2016 ◽  
Vol 140 (4) ◽  
pp. 3271-3271 ◽  
Author(s):  
Zhi Zhu ◽  
Ryota Miyauchi ◽  
Yukiko Araki ◽  
Masashi Unoki
Author(s):  
Martin Chavant ◽  
Alexis Hervais-Adelman ◽  
Olivier Macherey

Purpose An increasing number of individuals with residual or even normal contralateral hearing are being considered for cochlear implantation. It remains unknown whether the presence of contralateral hearing is beneficial or detrimental to their perceptual learning of cochlear implant (CI)–processed speech. The aim of this experiment was to provide a first insight into this question using acoustic simulations of CI processing. Method Sixty normal-hearing listeners took part in an auditory perceptual learning experiment. Each subject was randomly assigned to one of three groups of 20 referred to as NORMAL, LOWPASS, and NOTHING. The experiment consisted of two test phases separated by a training phase. In the test phases, all subjects were tested on recognition of monosyllabic words passed through a six-channel “PSHC” vocoder presented to a single ear. In the training phase, which consisted of listening to a 25-min audio book, all subjects were also presented with the same vocoded speech in one ear but the signal they received in their other ear differed across groups. The NORMAL group was presented with the unprocessed speech signal, the LOWPASS group with a low-pass filtered version of the speech signal, and the NOTHING group with no sound at all. Results The improvement in speech scores following training was significantly smaller for the NORMAL than for the LOWPASS and NOTHING groups. Conclusions This study suggests that the presentation of normal speech in the contralateral ear reduces or slows down perceptual learning of vocoded speech but that an unintelligible low-pass filtered contralateral signal does not have this effect. Potential implications for the rehabilitation of CI patients with partial or full contralateral hearing are discussed.


2010 ◽  
Vol 10 ◽  
pp. 329-339 ◽  
Author(s):  
Torsten Rahne ◽  
Michael Ziese ◽  
Dorothea Rostalski ◽  
Roland Mühler

This paper describes a logatome discrimination test for the assessment of speech perception in cochlear implant users (CI users), based on a multilingual speech database, the Oldenburg Logatome Corpus, which was originally recorded for the comparison of human and automated speech recognition. The logatome discrimination task is based on the presentation of 100 logatome pairs (i.e., nonsense syllables) with balanced representations of alternating “vowel-replacement” and “consonant-replacement” paradigms in order to assess phoneme confusions. Thirteen adult normal hearing listeners and eight adult CI users, including both good and poor performers, were included in the study and completed the test after their speech intelligibility abilities were evaluated with an established sentence test in noise. Furthermore, the discrimination abilities were measured electrophysiologically by recording the mismatch negativity (MMN) as a component of auditory event-related potentials. The results show a clear MMN response only for normal hearing listeners and CI users with good performance, correlating with their logatome discrimination abilities. Higher discrimination scores for vowel-replacement paradigms than for the consonant-replacement paradigms were found. We conclude that the logatome discrimination test is well suited to monitor the speech perception skills of CI users. Due to the large number of available spoken logatome items, the Oldenburg Logatome Corpus appears to provide a useful and powerful basis for further development of speech perception tests for CI users.


2015 ◽  
Vol 54 (06) ◽  
pp. 500-504 ◽  
Author(s):  
A. G. Maglione ◽  
A. Scorpecci ◽  
P. Malerba ◽  
P. Marsella ◽  
S. Giannantonio ◽  
...  

SummaryObjectives: The aim of the present study is to investigate the variations of the electroencephalographic (EEG) alpha rhythm in order to measure the appreciation of bilateral and unilateral young cochlear implant users during the observation of a musical cartoon. The cartoon has been modified for the generation of three experimental conditions: one with the original audio, another one with a distorted sound and, finally, a mute version.Methods: The EEG data have been recorded during the observation of the cartoons in the three experimental conditions. The frontal alpha EEG imbalance has been calculated as a measure of motivation and pleasantness to be compared across experimental populations and conditions.Results: The EEG frontal imbalance of the alpha rhythm showed significant variations during the perception of the different cartoons. In particular, the pattern of activation of normal-hearing children is very similar to the one elicited by the bilateral implanted patients. On the other hand, results related to the unilateral subjects do not present significant variations of the imbalance index across the three cartoons.Conclusion: The presented results suggest that the unilateral patients could not appreciate the difference in the audio format as well as bilaterally implanted and normal hearing subjects. The frontal alpha EEG imbalance is a useful tool to detect the differences in the appreciation of audiovisual stimuli in cochlear implant patients.


2015 ◽  
Vol 43 (2) ◽  
pp. 310-337 ◽  
Author(s):  
MARCEL R. GIEZEN ◽  
PAOLA ESCUDERO ◽  
ANNE E. BAKER

AbstractThis study investigates the role of acoustic salience and hearing impairment in learning phonologically minimal pairs. Picture-matching and object-matching tasks were used to investigate the learning of consonant and vowel minimal pairs in five- to six-year-old deaf children with a cochlear implant (CI), and children of the same age with normal hearing (NH). In both tasks, the CI children showed clear difficulties with learning minimal pairs. The NH children also showed some difficulties, however, particularly in the picture-matching task. Vowel minimal pairs were learned more successfully than consonant minimal pairs, particularly in the object-matching task. These results suggest that the ability to encode phonetic detail in novel words is not fully developed at age six and is affected by task demands and acoustic salience. CI children experience persistent difficulties with accurately mapping sound contrasts to novel meanings, but seem to benefit from the relative acoustic salience of vowel sounds.


Sign in / Sign up

Export Citation Format

Share Document