scholarly journals Influence of face surgical and N95 face masks on speech perception and listening effort in noise

Author(s):  
Torsten Rahne ◽  
Laura Fröhlich ◽  
Stefan Plontke ◽  
Luise Wagner

Abstract Daily life conversation relies on speech perception in quiet and noise. Due to the COVID-19 pandemic, the usage of face masks has become mandatory in many situations. Acoustic attenuation of sound pressure by the mask tissue reduces speech perception ability especially in noisy situations. Mask also impede the process of speech information from sight of the moving mouth (lip reading). In a prospective exploratory experimental study including 17 normal hearing participants, speech perception in noise and listening effort were measured with and without a surgical mask or an N95 mask between the speech signal and the listener. Additionally, the influence of the masks on the acoustic frequency spectrum was measured. Averaged over all noise signals, speech perception threshold in noise was significantly reduced by 1.6 dB (95% CI[1.0, 2.1]) by a surgical mask and by 2.7 dB (95% CI[2.2, 3.2]) using an N95 mask. The 50%-listening effort signal-to-noise ratio was not significantly increased by 0.58 dB (95% CI[0.4, 1.5]) by a surgical mask but significantly increased by 2.2 dB (95% CI[1.2, 3.1]) using an N95 mask. In acoustic measures the amplitudes were reduced by the mask tissue by up to 8 dB at frequencies above 1 kHz while no reduction was observed below 1 kHz. We conclude that face masks reduce speech perception and increase listening effort in different noise signals. Together with an additional interference by missing lip reading the compound effect of face masks would have a relevant impact on daily life communication already in normal hearing subjects.

PLoS ONE ◽  
2021 ◽  
Vol 16 (7) ◽  
pp. e0253874
Author(s):  
Torsten Rahne ◽  
Laura Fröhlich ◽  
Stefan Plontke ◽  
Luise Wagner

Daily-life conversation relies on speech perception in quiet and noise. Because of the COVID-19 pandemic, face masks have become mandatory in many situations. Acoustic attenuation of sound pressure by the mask tissue reduces speech perception ability, especially in noisy situations. Masks also can impede the process of speech comprehension by concealing the movements of the mouth, interfering with lip reading. In this prospective observational, cross-sectional study including 17 participants with normal hearing, we measured the influence of acoustic attenuation caused by medical face masks (mouth and nose protection) according to EN 14683 and of N95 masks according to EN 1149 (EN 14683) on the speech recognition threshold and listening effort in various types of background noise. Averaged over all noise signals, a surgical mask significantly reduced the speech perception threshold in noise was by 1.6 dB (95% confidence interval [CI], 1.0, 2.1) and an N95 mask reduced it significantly by 2.7 dB (95% CI, 2.2, 3.2). Use of a surgical mask did not significantly increase the 50% listening effort signal-to-noise ratio (increase of 0.58 dB; 95% CI, 0.4, 1.5), but use of an N95 mask did so significantly, by 2.2 dB (95% CI, 1.2, 3.1). In acoustic measures, mask tissue reduced amplitudes by up to 8 dB at frequencies above 1 kHz, whereas no reduction was observed below 1 kHz. We conclude that face masks reduce speech perception and increase listening effort in different noise signals. Together with additional interference because of impeded lip reading, the compound effect of face masks could have a relevant impact on daily life communication even in those with normal hearing.


2015 ◽  
Vol 26 (06) ◽  
pp. 572-581 ◽  
Author(s):  
Stanley Sheft ◽  
Min-Yu Cheng ◽  
Valeriy Shafiro

Background: Past work has shown that low-rate frequency modulation (FM) may help preserve signal coherence, aid segmentation at word and syllable boundaries, and benefit speech intelligibility in the presence of a masker. Purpose: This study evaluated whether difficulties in speech perception by cochlear implant (CI) users relate to a deficit in the ability to discriminate among stochastic low-rate patterns of FM. Research Design: This is a correlational study assessing the association between the ability to discriminate stochastic patterns of low-rate FM and the intelligibility of speech in noise. Study Sample: Thirteen postlingually deafened adult CI users participated in this study. Data Collection and Analysis: Using modulators derived from 5-Hz lowpass noise applied to a 1-kHz carrier, thresholds were measured in terms of frequency excursion both in quiet and with a speech-babble masker present, stimulus duration, and signal-to-noise ratio in the presence of a speech-babble masker. Speech perception ability was assessed in the presence of the same speech-babble masker. Relationships were evaluated with Pearson product–moment correlation analysis with correction for family-wise error, and commonality analysis to determine the unique and common contributions across psychoacoustic variables to the association with speech ability. Results: Significant correlations were obtained between masked speech intelligibility and three metrics of FM discrimination involving either signal-to-noise ratio or stimulus duration, with shared variance among the three measures accounting for much of the effect. Compared to past results from young normal-hearing adults and older adults with either normal hearing or a mild-to-moderate hearing loss, mean FM discrimination thresholds obtained from CI users were higher in all conditions. Conclusions: The ability to process the pattern of frequency excursions of stochastic FM may, in part, have a common basis with speech perception in noise. Discrimination of differences in the temporally distributed place coding of the stimulus could serve as this common basis for CI users.


QJM ◽  
2020 ◽  
Vol 113 (Supplement_1) ◽  
Author(s):  
A M Saad ◽  
M A Hegazi ◽  
M S Khodeir

Abstract Background Lip-reading is considered an important skill which varies considerably among normal hearing and hearing impaired (HI) children. It helps HI children to perceive speech, acquire spoken language and acquire phonological awareness. Speech perception is considered to be a multisensory process that involves attention to auditory signals as well as visual articulatory movements. Integration of auditory and visual signals occurs naturally and automatically in normal individuals across all ages. Many researches suggested that normal hearing children use audition as the primary sensory modality for speech perception, whereas HI children use lip-reading cues as the primary sensory modality for speech perception. Aim of the Work The aim of this study is to compare the lip-reading ability between normal and HI children. Participants and methods This is a comparative descriptive case control study. It was applied on 60 hearing impaired children (cases) and 60 normal hearing children (controls) of the same age and gender. The age range was (3-8 years). The Egyptian Arabic Lip-reading Test was applied to all children. Results There was statistically significant difference between the total mean scores of the EALRT between normal and HI children. Conclusion The results of the study proved that normal children are better lip-readers than HI children of the matched age range.


2020 ◽  
Vol 63 (5) ◽  
pp. 1595-1607
Author(s):  
Julia Campbell ◽  
Mashhood Nielsen ◽  
Alison LaBrec ◽  
Connor Bean

Purpose Speech perception in noise (SPiN) varies widely in individuals with normal hearing, which may be attributed to factors that are not reflected in the audiogram, such as inhibition. However, inhibition is involved at both sensory and cognitive stages of auditory perception, and while inhibition at the cognitive level has been shown to be a significant factor in SPiN processes, it is unknown whether sensory inhibition may also contribute to SPiN variability. Therefore, the goal of this study was to evaluate sensory inhibition in adults with normal hearing and mild SPiN impairment. Method Cortical auditory evoked potentials (CAEPs) were recorded in 49 adults via high-density electroencephalography using an auditory gating paradigm. Participants were categorized according to a median signal-to-noise ratio (SNR) loss of 1.5 dB: typical SNR loss ≤ 1.5 dB ( n = 32), mild SNR loss > 1.5 dB ( n = 17). CAEP gating responses were compared and correlated with SNR loss and extended high-frequency thresholds. Current density reconstructions were performed to qualitatively observe underlying cortical inhibitory networks in each group. Results In comparison to adults with typical SPiN ability, adults with mild SPiN impairment showed an absence of the gating response. A CAEP gating component (P2) reflected decreased sensory inhibition and correlated with increased SNR loss. Extended high-frequency thresholds were also found to correlate with SNR loss, but not gating function. An atypical cortical inhibitory network was observed in the mild SNR loss group, with reduced frontal and absent prefrontal activation. Conclusion Sensory inhibition appears to be atypical and related to SPiN deficits in adults with mild impairment. In addition, cortical inhibitory networks appear to be incomplete, with a possible compensatory parietal network. Further research is needed to delineate between types or levels of central inhibitory mechanisms and their contribution to SPiN processes.


Revista CEFAC ◽  
2018 ◽  
Vol 20 (4) ◽  
pp. 541-549
Author(s):  
Laura Mochiatti Guijo ◽  
Ana Cláudia Vieira Cardoso

ABSTRACT Purpose: to review the scientific literature and present existing instruments and methods for the objective assessment of the listening effort in normal hearing individuals worldwide. Methods: a literature integrative review whose purpose was to gather and summarize the scientific knowledge regarding the objective methods theme for measuring the listening effort, developed through the search of articles in specialized national and international journals, in the English and Portuguese languages, available in the databases: PUBMED, Cochrane Library, LILACS and SCIELO. Results: 18 articles which used physiological methods to measure the listening effort in individuals with normal hearing were reviewed. The main findings described in those articles refer to the author(s) and purpose(s) of the research, country where the research was conducted, casuistry, physiological method used and results. Conclusion: there is no consensus among the researchers about the best physiological method to measure this parameter, that is, this effort in the speech perception tasks, although the level of skin conductance is considered the most accurate measure to date.


Author(s):  
Yones Lotfi ◽  
Jamileh Chupani ◽  
Mohanna Javanbakht ◽  
Enayatollah Bakhshi

Background and Aim: In most everyday sett­ings, speech is heard in the presence of com­peting sounds and speech perception in noise is affected by various factors, including cognitive factors. In this regard, bilingualism is a pheno­menon that changes cognitive and behavioral processes as well as the nervous system. This study aimed to evaluate speech perception in noise and compare differences in Kurd-Persian bilinguals versus Persian monolinguals. Methods: This descriptive-analytic study was performed on 92 students with normal hearing, 46 of whom were bilingual Kurd-Persian with a mean (SD) age of 22.73 (1.92) years, and 46 other Persian monolinguals with a mean (SD) age of 22.71 (2.28) years. They were examined by consonant-vowel in noise (CV in noise) test and quick speech in noise (Q-SIN) test. The obtained data were analyzed by SPSS 21. Results: The comparison of the results showed differences in both tests between bilingual and monolingual subjects. In both groups, the reduc­tion of signal-to-noise ratio led to lower scores, but decrease in CV in noise test in bilinguals was less than monolinguals (p < 0.001) and in the Q-SIN test, the drop in bilinguals’ score was  more than monolinguals (p = 0.002). Conclusion: Kurd-Persian bilinguals had a bet­ter performance in CV in noise test but had a worse performance in Q-SIN test than Persian monolinguals.


2020 ◽  
Author(s):  
Matthew Winn ◽  
Katherine H. Teece

Speech perception and listening effort are complicated and interrelated concepts. One might assume that intelligibility performance (percent correct) is a proxy for listening effort, but there are some reasons to challenge whether that is actually true. Correct responses in speech perception tests could reflect effortful mental processing, and a completely wrong answer could evoke very little effort, especially if the misperception itself is linguistically well-formed and sensible. This paper presents evidence that listening effort is not a function of the proportion of words correct, but is rather driven by the types of errors, position of errors within a sentence, and the need to resolve ambiguity, reflecting how easily the listener can make sense of a perception. We offer a taxonomy of error types that is both intuitive and also consistent with data from two experiments measuring listening effort with careful controls to either elicit specific kinds of mistakes or to track specific mistakes retrospectively. Participants included individuals with normal hearing or with cochlear implants. In two experiments of sentence repetition, listening effort – indexed by changes in pupil size – was found to scale with the amount of perceptual restoration needed (phoneme versus whole word), and also scale with the sensibility of responses, but not with the number of intelligibility errors. Although mental corrective action and number of mistakes can scale together in many experiments, it is possible to dissociate them in order to advance toward a more explanatory (rather than correlational) account of listening effort.


2008 ◽  
Vol 17 (6) ◽  
pp. 405-409 ◽  
Author(s):  
Lawrence D. Rosenblum

Speech perception is inherently multimodal. Visual speech (lip-reading) information is used by all perceivers and readily integrates with auditory speech. Imaging research suggests that the brain treats auditory and visual speech similarly. These findings have led some researchers to consider that speech perception works by extracting amodal information that takes the same form across modalities. From this perspective, speech integration is a property of the input information itself. Amodal speech information could explain the reported automaticity, immediacy, and completeness of audiovisual speech integration. However, recent findings suggest that speech integration can be influenced by higher cognitive properties such as lexical status and semantic context. Proponents of amodal accounts will need to explain these results.


2017 ◽  
Vol 60 (5) ◽  
pp. 1236-1245 ◽  
Author(s):  
Rebecca E. Millman ◽  
Sven L. Mattys

Purpose Background noise can interfere with our ability to understand speech. Working memory capacity (WMC) has been shown to contribute to the perception of speech in modulated noise maskers. WMC has been assessed with a variety of auditory and visual tests, often pertaining to different components of working memory. This study assessed the relationship between speech perception in modulated maskers and components of auditory verbal working memory (AVWM) over a range of signal-to-noise ratios. Method Speech perception in noise and AVWM were measured in 30 listeners (age range 31–67 years) with normal hearing. AVWM was estimated using forward digit recall, backward digit recall, and nonword repetition. Results After controlling for the effects of age and average pure-tone hearing threshold, speech perception in modulated maskers was related to individual differences in the phonological component of working memory (as assessed by nonword repetition) but only in the least favorable signal-to-noise ratio. The executive component of working memory (as assessed by backward digit) was not predictive of speech perception in any conditions. Conclusions AVWM is predictive of the ability to benefit from temporal dips in modulated maskers: Listeners with greater phonological WMC are better able to correctly identify sentences in modulated noise backgrounds.


eLife ◽  
2020 ◽  
Vol 9 ◽  
Author(s):  
Aravindakshan Parthasarathy ◽  
Kenneth E Hancock ◽  
Kara Bennett ◽  
Victor DeGruttola ◽  
Daniel B Polley

In social settings, speech waveforms from nearby speakers mix together in our ear canals. Normally, the brain unmixes the attended speech stream from the chorus of background speakers using a combination of fast temporal processing and cognitive active listening mechanisms. Of >100,000 patient records,~10% of adults visited our clinic because of reduced hearing, only to learn that their hearing was clinically normal and should not cause communication difficulties. We found that multi-talker speech intelligibility thresholds varied widely in normal hearing adults, but could be predicted from neural phase-locking to frequency modulation (FM) cues measured with ear canal EEG recordings. Combining neural temporal fine structure processing, pupil-indexed listening effort, and behavioral FM thresholds accounted for 78% of the variability in multi-talker speech intelligibility. The disordered bottom-up and top-down markers of poor multi-talker speech perception identified here could inform the design of next-generation clinical tests for hidden hearing disorders.


Sign in / Sign up

Export Citation Format

Share Document