sound source segregation
Recently Published Documents


TOTAL DOCUMENTS

23
(FIVE YEARS 3)

H-INDEX

5
(FIVE YEARS 1)

2021 ◽  
Vol 15 ◽  
Author(s):  
Christiane R. Neubert ◽  
Alexander P. Förstel ◽  
Stefan Debener ◽  
Alexandra Bendixen

When multiple sound sources are present at the same time, auditory perception is often challenged with disentangling the resulting mixture and focusing attention on the target source. It has been repeatedly demonstrated that background (distractor) sound sources are easier to ignore when their spectrotemporal signature is predictable. Prior evidence suggests that this ability to exploit predictability for foreground-background segregation degrades with age. On a theoretical level, this has been related with an impairment in elderly adults’ capabilities to detect certain types of sensory deviance in unattended sound sequences. Yet the link between those two capacities, deviance detection and predictability-based sound source segregation, has not been empirically demonstrated. Here we report on a combined behavioral-EEG study investigating the ability of elderly listeners (60–75 years of age) to use predictability as a cue for sound source segregation, as well as their sensory deviance detection capacities. Listeners performed a detection task on a target stream that can only be solved when a concurrent distractor stream is successfully ignored. We contrast two conditions whose distractor streams differ in their predictability. The ability to benefit from predictability was operationalized as performance difference between the two conditions. Results show that elderly listeners can use predictability for sound source segregation at group level, yet with a high degree of inter-individual variation in this ability. In a further, passive-listening control condition, we measured correlates of deviance detection in the event-related brain potential (ERP) elicited by occasional deviations from the same spectrotemporal pattern as used for the predictable distractor sequence during the behavioral task. ERP results confirmed neural signatures of deviance detection in terms of mismatch negativity (MMN) at group level. Correlation analyses at single-subject level provide no evidence for the hypothesis that deviance detection ability (measured by MMN amplitude) is related to the ability to benefit from predictability for sound source segregation. These results are discussed in the frameworks of sensory deviance detection and predictive coding.


Author(s):  
Guus Christian van Bentum ◽  
Marc Mathijs van Wanrooij ◽  
A. John Van Opstal

To program a goal-directed response in the presence of acoustic reflections, the audio-motor system should suppress the detection of time-delayed sources. We examined the effects of spatial separation and inter-stimulus delay on the ability of human listeners to localize a pair of broadband sounds in the horizontal plane. Participants indicated how many sounds were heard and where these were perceived by making one or two head-orienting localization responses. Results suggest that perceptual fusion of the two sounds depends on delay and spatial separation. Leading and lagging stimuli in close spatial proximity required longer stimulus delays to be perceptually separated than those further apart. Whenever participants heard one sound, their localization responses for synchronous sounds were oriented to a weighted average of both source locations. For short delays, responses were directed towards the leading stimulus location. Increasing spatial separation enhanced this effect. For longer delays, responses were again directed towards a weighted average. When participants perceived two sounds, the first and the second response were directed to either of the leading and lagging source locations. Perceived locations were interchanged often in their temporal order (in ~40% of trials). We show that the percept of two sounds occurring requires sufficient spatiotemporal separation, after which localization can be performed with high accuracy. We propose that the percept of temporal order of two concurrent sounds results from a different process than localization, and discuss how dynamic lateral excitatory-inhibitory interactions within a spatial sensorimotor map could explain the findings.


Brain ◽  
2020 ◽  
Vol 143 (9) ◽  
pp. 2689-2695 ◽  
Author(s):  
Chris J D Hardy ◽  
Keir X X Yong ◽  
Johanna C Goll ◽  
Sebastian J Crutch ◽  
Jason D Warren

Abstract Although posterior cortical atrophy is often regarded as the canonical ‘visual dementia’, auditory symptoms may also be salient in this disorder. Patients often report particular difficulty hearing in busy environments; however, the core cognitive process—parsing of the auditory environment (‘auditory scene analysis’)—has been poorly characterized. In this cross-sectional study, we used customized perceptual tasks to assess two generic cognitive operations underpinning auditory scene analysis—sound source segregation and sound event grouping—in a cohort of 21 patients with posterior cortical atrophy, referenced to 15 healthy age-matched individuals and 21 patients with typical Alzheimer’s disease. After adjusting for peripheral hearing function and performance on control tasks assessing perceptual and executive response demands, patients with posterior cortical atrophy performed significantly worse on both auditory scene analysis tasks relative to healthy controls and patients with typical Alzheimer’s disease (all P < 0.05). Our findings provide further evidence of central auditory dysfunction in posterior cortical atrophy, with implications for our pathophysiological understanding of Alzheimer syndromes as well as clinical diagnosis and management.


2018 ◽  
Vol 8 (2) ◽  
Author(s):  
Erik Schebsdat ◽  
Manuel C. Kohl ◽  
Farah I. Corona-Strauss ◽  
Harald Seidler ◽  
Daniel J. Strauss

The importance of binaural cues in auditory stream formation and sound source segregation is widely accepted. When treating one ear with a cochlear implant (CI) the peripheral auditory system gets partially replaced and processing delays get added potentially, thus important interaural time differences get altered. However, these effects are not fully understood, leaving a lack of systematic binaural fitting strategies with respect to an optimal binaural fusion. To get new insights into such alterations, we suggest a novel method of free-field auditory brainstem evoked responses (ABRs) analysis in CI users. This method does not bypass the technically induced intrinsic delays of the sound processor while leaving the whole electrode array active, thus the most natural way of stimulation is provided. We compared the ABRs collected of 12 CI users and 12 normal hearing listeners using two different stimuli (chirp, click) at four different intensities each. We analyzed the ABRs using the average of 2000 trials as well as a single trial analysis and found consistent results in the ABRs’ amplitudes and latencies, as well as in single trial relationships between both groups. This method provides a new perspective into the natural CI users’ ABRs and can be useful in future research regarding binaural interaction and fusion.


2018 ◽  
Vol 61 (11) ◽  
pp. 2804-2813
Author(s):  
Frédéric Apoux ◽  
Brittney L. Carter ◽  
Eric W. Healy

Purpose The goal of this study was to examine the role of carrier cues in sound source segregation and the possibility to enhance the intelligibility of 2 sentences presented simultaneously. Dual-carrier (DC) processing (Apoux, Youngdahl, Yoho, & Healy, 2015) was used to introduce synthetic carrier cues in vocoded speech. Method Listeners with normal hearing heard sentences processed either with a DC or with a traditional single-carrier (SC) vocoder. One group was asked to repeat both sentences in a sentence pair (Experiment 1). The other group was asked to repeat only 1 sentence of the pair and was provided additional segregation cues involving onset asynchrony (Experiment 2). Results Both experiments showed that not only is the “target” sentence more intelligible in DC compared with SC, but the “background” sentence intelligibility is equally enhanced. The participants did not benefit from the additional segregation cues. Conclusions The data showed a clear benefit of using a distinct carrier to convey each sentence (i.e., DC processing). Accordingly, the poor speech intelligibility in noise typically observed with SC-vocoded speech may be partly attributed to the envelope of independent sound sources sharing the same carrier. Moreover, this work suggests that noise reduction may not be the only viable option to improve speech intelligibility in noise for users of cochlear implants. Alternative approaches aimed at enhancing sound source segregation such as DC processing may help to improve speech intelligibility while preserving and enhancing the background.


eLife ◽  
2017 ◽  
Vol 6 ◽  
Author(s):  
Norman Lee ◽  
Andrew C Mason

Spatial release from masking (SRM) occurs when spatial separation between a signal and masker decreases masked thresholds. The mechanically-coupled ears of Ormia ochracea are specialized for hyperacute directional hearing, but the possible role of SRM, or whether such specializations exhibit limitations for sound source segregation, is unknown. We recorded phonotaxis to a cricket song masked by band-limited noise. With a masker, response thresholds increased and localization was diverted away from the signal and masker. Increased separation from 6° to 90° did not decrease response thresholds or improve localization accuracy, thus SRM does not operate in this range of spatial separations. Tympanal vibrations and auditory nerve responses reveal that localization errors were consistent with changes in peripheral coding of signal location and flies localized towards the ear with better signal detection. Our results demonstrate that, in a mechanically coupled auditory system, specialization for directional hearing does not contribute to source segregation.


2015 ◽  
Vol 27 (2) ◽  
pp. 200-207
Author(s):  
Yeng Weng Leong ◽  
◽  
Hiroaki Seki ◽  
Yoshitsugu Kamiya ◽  
Masatoshi Hikizu ◽  
...  

<div class=""abs_img""> <img src=""[disp_template_path]/JRM/abst-image/00270002/10.jpg"" width=""300"" /> Mobile localization: grouping</div> Mobile devices are caught in an inverse relationship between mobility and ease of use. This paper presents incremental technology related to a previously proposed method of mobile yet easy-to-use input using triboacoustic signals generated from the action of a user tracing shapes on surfaces in environment. Mobile devices must function in various environments, thus requiring that they be immune to noise interference. We propose an improvement in accuracy and automated multiple sound source segregation supported by experiments evidencing the proposal’ s effectiveness, results of which show the proposal’ s accuracy and capability have merit and should be pursued further. </span>


Sign in / Sign up

Export Citation Format

Share Document