degraded speech
Recently Published Documents


TOTAL DOCUMENTS

186
(FIVE YEARS 55)

H-INDEX

25
(FIVE YEARS 3)

NeuroSci ◽  
2022 ◽  
Vol 3 (1) ◽  
pp. 52-62
Author(s):  
Mira White ◽  
Fauve Duquette-Laplante ◽  
Benoît Jutras ◽  
Caryn Bursch ◽  
Amineh Koravand

Purpose: The main purpose of this retrospective study was to identify auditory dysfunctions related to traumatic brain injury (TBI) in individuals evaluated in an Audiology clinic. Method: Peripheral and central auditory evaluations were performed from March 2014 to June 2018 in 26 patients (14 males) with TBI. The age of the participants ranged from 9 to 59 years old (34.24 ± 15.21). Six participants had blast-related TBI and 20 had blunt force TBI. Sixteen experienced a single TBI event whereas ten experienced several. Correlation analyses were performed to verify the relationship, if any, between the number of auditory tests failed and the number, type, and severity of TBIs. Result: All participants failed at least one auditory test. Nearly 60% had abnormal results on degraded speech tests (compressed and echoed, filtered or in background noise) and 25% had a high frequency hearing loss. There was no statistically significant correlation between the number of auditory tests failed and the number, type, and severity of TBIs. Conclusion: Results indicated negative and heterogenous effects of TBI on peripheral and central auditory function and highlighted the need for a more extensive auditory assessment in individuals with TBI.


2021 ◽  
Author(s):  
Anna Uta Rysop ◽  
Lea-Maria Schmitt ◽  
Jonas Obleser ◽  
Gesa Hartwigsen

AbstractSpeech comprehension is often challenged by increased background noise, but can be facilitated via the semantic context of a sentence. This predictability gain relies on an interplay of language-specific semantic and domain-general brain regions. However, age-related differences in the interactions within and between semantic and domain-general networks remain poorly understood. Here we investigated commonalities and differences in degraded speech processing in healthy young and old participants. Participants performed a sentence repetition task while listening to sentences with high and low predictable endings and varying intelligibility. Stimulus intelligibility was adjusted to individual hearing abilities. Older adults showed an undiminished behavioural predictability gain. Likewise, both groups recruited a similar set of semantic and cingulo-opercular brain regions. However, we observed age-related differences in effective connectivity for high predictable speech of increasing intelligibility. Young adults exhibited stronger coupling within the cingulo-opercular network and between a cingulo-opercular and a posterior temporal semantic node. Moreover, these interactions were excitatory in young adults but inhibitory in old adults. Finally, the degree of the inhibitory influence between cingulo-opercular regions was predictive of the behavioural sensitivity towards changes in intelligibility for high predictable sentences in older adults only. Our results demonstrate that the predictability gain is relatively preserved in older adults when stimulus intelligibility is individually adjusted. While young and old participants recruit similar brain regions, differences manifest in network dynamics. Together, these results suggest that ageing affects the network configuration rather than regional activity during successful speech comprehension under challenging listening conditions.


Author(s):  
Jiaqiang Zhu ◽  
Xiaoxiang Chen ◽  
Fei Chen ◽  
Seth Wiener

Purpose: Individuals with congenital amusia exhibit degraded speech perception. This study examined whether adult Chinese Mandarin listeners with amusia were still able to extract the statistical regularities of Mandarin speech sounds, despite their degraded speech perception. Method: Using the gating paradigm with monosyllabic syllable–tone words, we tested 19 Mandarin-speaking amusics and 19 musically intact controls. Listeners heard increasingly longer fragments of the acoustic signal across eight duration-blocked gates. The stimuli varied in syllable token frequency and syllable–tone co-occurrence probability. The correct syllable–tone word, correct syllable-only, correct tone-only, and correct syllable–incorrect tone responses were compared respectively between the two groups using mixed-effects models. Results: Amusics were less accurate than controls in terms of the correct word, correct syllable-only, and correct tone-only responses. Amusics, however, showed consistent patterns of top-down processing, as indicated by more accurate responses to high-frequency syllables, high-probability tones, and tone errors all in manners similar to those of the control listeners. Conclusions: Amusics are able to learn syllable and tone statistical regularities from the language input. This extends previous work by showing that amusics can track phonological segment and pitch cues despite their degraded speech perception. The observed speech deficits in amusics are therefore not due to an abnormal statistical learning mechanism. These results support rehabilitation programs aimed at improving amusics' sensitivity to pitch.


2021 ◽  
Vol 12 ◽  
Author(s):  
Kendra Gimhani Kandana Arachchige ◽  
Wivine Blekic ◽  
Isabelle Simoes Loureiro ◽  
Laurent Lefebvre

Numerous studies have explored the benefit of iconic gestures in speech comprehension. However, only few studies have investigated how visual attention was allocated to these gestures in the context of clear versus degraded speech and the way information is extracted for enhancing comprehension. This study aimed to explore the effect of iconic gestures on comprehension and whether fixating the gesture is required for information extraction. Four types of gestures (i.e., semantically and syntactically incongruent iconic gestures, meaningless configurations, and congruent iconic gestures) were presented in a sentence context in three different listening conditions (i.e., clear, partly degraded or fully degraded speech). Using eye tracking technology, participants’ gaze was recorded, while they watched video clips after which they were invited to answer simple comprehension questions. Results first showed that different types of gestures differently attract attention and that the more speech was degraded, the less participants would pay attention to gestures. Furthermore, semantically incongruent gestures appeared to particularly impair comprehension although not being fixated while congruent gestures appeared to improve comprehension despite also not being fixated. These results suggest that covert attention is sufficient to convey information that will be processed by the listener.


Author(s):  
Faizah Mushtaq ◽  
Ian M. Wiggins ◽  
Pádraig T. Kitterick ◽  
Carly A. Anderson ◽  
Douglas E. H. Hartley

AbstractWhilst functional neuroimaging has been used to investigate cortical processing of degraded speech in adults, much less is known about how these signals are processed in children. An enhanced understanding of cortical correlates of poor speech perception in children would be highly valuable to oral communication applications, including hearing devices. We utilised vocoded speech stimuli to investigate brain responses to degraded speech in 29 normally hearing children aged 6–12 years. Intelligibility of the speech stimuli was altered in two ways by (i) reducing the number of spectral channels and (ii) reducing the amplitude modulation depth of the signal. A total of five different noise-vocoded conditions (with zero, partial or high intelligibility) were presented in an event-related format whilst participants underwent functional near-infrared spectroscopy (fNIRS) neuroimaging. Participants completed a word recognition task during imaging, as well as a separate behavioural speech perception assessment. fNIRS recordings revealed statistically significant sensitivity to stimulus intelligibility across several brain regions. More intelligible stimuli elicited stronger responses in temporal regions, predominantly within the left hemisphere, while right inferior parietal regions showed an opposite, negative relationship. Although there was some evidence that partially intelligible stimuli elicited the strongest responses in the left inferior frontal cortex, a region previous studies have suggested is associated with effortful listening in adults, this effect did not reach statistical significance. These results further our understanding of cortical mechanisms underlying successful speech perception in children. Furthermore, fNIRS holds promise as a clinical technique to help assess speech intelligibility in paediatric populations.


2021 ◽  
Vol 12 ◽  
Author(s):  
Pratik Bhandari ◽  
Vera Demberg ◽  
Jutta Kray

Previous studies have shown that at moderate levels of spectral degradation, semantic predictability facilitates language comprehension. It is argued that when speech is degraded, listeners have narrowed expectations about the sentence endings; i.e., semantic prediction may be limited to only most highly predictable sentence completions. The main objectives of this study were to (i) examine whether listeners form narrowed expectations or whether they form predictions across a wide range of probable sentence endings, (ii) assess whether the facilitatory effect of semantic predictability is modulated by perceptual adaptation to degraded speech, and (iii) use and establish a sensitive metric for the measurement of language comprehension. For this, we created 360 German Subject-Verb-Object sentences that varied in semantic predictability of a sentence-final target word in a graded manner (high, medium, and low) and levels of spectral degradation (1, 4, 6, and 8 channels noise-vocoding). These sentences were presented auditorily to two groups: One group (n =48) performed a listening task in an unpredictable channel context in which the degraded speech levels were randomized, while the other group (n =50) performed the task in a predictable channel context in which the degraded speech levels were blocked. The results showed that at 4 channels noise-vocoding, response accuracy was higher in high-predictability sentences than in the medium-predictability sentences, which in turn was higher than in the low-predictability sentences. This suggests that, in contrast to the narrowed expectations view, comprehension of moderately degraded speech, ranging from low- to high- including medium-predictability sentences, is facilitated in a graded manner; listeners probabilistically preactivate upcoming words from a wide range of semantic space, not limiting only to highly probable sentence endings. Additionally, in both channel contexts, we did not observe learning effects; i.e., response accuracy did not increase over the course of experiment, and response accuracy was higher in the predictable than in the unpredictable channel context. We speculate from these observations that when there is no trial-by-trial variation of the levels of speech degradation, listeners adapt to speech quality at a long timescale; however, when there is a trial-by-trial variation of the high-level semantic feature (e.g., sentence predictability), listeners do not adapt to low-level perceptual property (e.g., speech quality) at a short timescale.


Author(s):  
Briony Banks ◽  
Emma Gowen ◽  
Kevin J. Munro ◽  
Patti Adank

Purpose Visual cues from a speaker's face may benefit perceptual adaptation to degraded speech, but current evidence is limited. We aimed to replicate results from previous studies to establish the extent to which visual speech cues can lead to greater adaptation over time, extending existing results to a real-time adaptation paradigm (i.e., without a separate training period). A second aim was to investigate whether eye gaze patterns toward the speaker's mouth were related to better perception, hypothesizing that listeners who looked more at the speaker's mouth would show greater adaptation. Method A group of listeners ( n = 30) was presented with 90 noise-vocoded sentences in audiovisual format, whereas a control group ( n = 29) was presented with the audio signal only. Recognition accuracy was measured throughout and eye tracking was used to measure fixations toward the speaker's eyes and mouth in the audiovisual group. Results Previous studies were partially replicated: The audiovisual group had better recognition throughout and adapted slightly more rapidly, but both groups showed an equal amount of improvement overall. Longer fixations on the speaker's mouth in the audiovisual group were related to better overall accuracy. An exploratory analysis further demonstrated that the duration of fixations to the speaker's mouth decreased over time. Conclusions The results suggest that visual cues may not benefit adaptation to degraded speech as much as previously thought. Longer fixations on a speaker's mouth may play a role in successfully decoding visual speech cues; however, this will need to be confirmed in future research to fully understand how patterns of eye gaze are related to audiovisual speech recognition. All materials, data, and code are available at https://osf.io/2wqkf/ .


2021 ◽  
Author(s):  
Jianwei Zhang ◽  
Suren Jayasuriya ◽  
Visar Berisha

2021 ◽  
Author(s):  
Camryn Terblanche ◽  
Philip Harrison ◽  
Amelia J. Gully

Sign in / Sign up

Export Citation Format

Share Document