sound segregation
Recently Published Documents


TOTAL DOCUMENTS

51
(FIVE YEARS 3)

H-INDEX

15
(FIVE YEARS 0)

2020 ◽  
Author(s):  
Kenny F Chou ◽  
Virginia Best ◽  
H Steven Colburn ◽  
Kamal Sen

AbstractListening in an acoustically cluttered scene remains a difficult task for both machines and hearing-impaired listeners. Normal-hearing listeners accomplish this task with relative ease by segregating the scene into its constituent sound sources, then selecting and attending to a target source. An assistive listening device that mimics the biological mechanisms underlying this behavior may provide an effective solution for those with difficulty listening in acoustically cluttered environments (e.g., a cocktail party). Here, we present a binaural sound segregation algorithm based on a hierarchical network model of the auditory system. In the algorithm, binaural sound inputs first drive populations of neurons tuned to specific spatial locations and frequencies. Lateral inhibition then sharpens the spatial response of the neurons. Finally, the spiking response of neurons in the output layer are then reconstructed into audible waveforms via a novel reconstruction method. We evaluate the performance of the algorithm with psychoacoustic measures of normal-hearing listeners. This two-microphone algorithm is shown to provide listeners with perceptual benefit similar to that of a 16-microphone acoustic beamformer in a difficult listening task. Unlike deep-learning approaches, the proposed algorithm is biologically interpretable and does not need to be trained on large datasets. This study presents a biologically based algorithm for sound source segregation as well as a method to reconstruct highly intelligible audio signals from spiking models.Author SummaryAnimal and humans can navigate complex auditory environments with relative ease, attending to certain sounds while suppressing others. Normally, various sounds originate from various spatial locations. This paper presents an algorithmic model to perform sound segregation based on how animals make use of this spatial information at various stages of the auditory pathway. We showed that the performance of this two-microphone algorithm provides as much benefit to normal-hearing listeners a multi-microphone algorithm. Unlike mathematical and machine-learning approaches, our model is fully interpretable and does not require training with large datasets. Such an approach may benefit the design of machine hearing algorithms. To interpret the spike-trains generated in the model, we designed a method to recover sounds from model spikes with high intelligibility. This method can be applied to spiking neural networks for audio-related applications, or to interpret each node within a spiking model of the auditory cortex.


2019 ◽  
Vol 146 (4) ◽  
pp. 2917-2917
Author(s):  
Kenny Chou ◽  
H. Steven Colburn ◽  
Kamal Sen
Keyword(s):  

2018 ◽  
Author(s):  
Aarti Sehdev ◽  
Yunusa G. Mohammed ◽  
Tilman Triphan ◽  
Paul Szyszka

SUMMARYOdorants of behaviorally relevant objects (e.g., food sources) intermingle with those from other sources. Therefore, to sniff out whether an odor source is good or bad – without actually visiting it – animals first need to segregate the odorants from different sources. To do so, animals could use temporal cues, since odorants from one source exhibit correlated fluctuations, while odorants from different sources are less correlated. However, it remains unclear whether animals can rely solely on temporal cues for odor source segregation. Here we show that 1) flies can use a few milliseconds differences in odorant arrival to segregate a target odorant from a binary mixture, 2) segregation does not improve when the target odorant arrives first, and 3) segregation works for odorants with innate, as well as learned valences. These properties of odor segregation parallel those of concurrent sound segregation and figure-ground segregation by onset asynchrony in humans.


2018 ◽  
Vol 367 ◽  
pp. 161-168
Author(s):  
Sara M.K. Madsen ◽  
Torsten Dau ◽  
Brian C.J. Moore
Keyword(s):  

2017 ◽  
Vol 372 (1714) ◽  
pp. 20160114 ◽  
Author(s):  
Anahita H. Mehta ◽  
Nori Jacoby ◽  
Ifat Yasin ◽  
Andrew J. Oxenham ◽  
Shihab A. Shamma

This study investigates the neural correlates and processes underlying the ambiguous percept produced by a stimulus similar to Deutsch's ‘octave illusion’, in which each ear is presented with a sequence of alternating pure tones of low and high frequencies. The same sequence is presented to each ear, but in opposite phase, such that the left and right ears receive a high–low–high … and a low–high–low … pattern, respectively. Listeners generally report hearing the illusion of an alternating pattern of low and high tones, with all the low tones lateralized to one side and all the high tones lateralized to the other side. The current explanation of the illusion is that it reflects an illusory feature conjunction of pitch and perceived location. Using psychophysics and electroencephalogram measures, we test this and an alternative hypothesis involving synchronous and sequential stream segregation, and investigate potential neural correlates of the illusion. We find that the illusion of alternating tones arises from the synchronous tone pairs across ears rather than sequential tones in one ear, suggesting that the illusion involves a misattribution of time across perceptual streams, rather than a misattribution of location within a stream. The results provide new insights into the mechanisms of binaural streaming and synchronous sound segregation. This article is part of the themed issue ‘Auditory and visual scene analysis’.


2016 ◽  
Vol 42 (3) ◽  
pp. 386-400 ◽  
Author(s):  
Keiko Masutomi ◽  
Nicolas Barascud ◽  
Makio Kashino ◽  
Josh H. McDermott ◽  
Maria Chait
Keyword(s):  

2016 ◽  
Vol 170 (1) ◽  
pp. 95-101
Author(s):  
Erin M. Ramage ◽  
Nedka Klimas ◽  
Sally J. Vogel ◽  
Mary Vertinski ◽  
Breanne D. Yerkes ◽  
...  

2015 ◽  
Vol 27 (11) ◽  
pp. 2186-2196 ◽  
Author(s):  
Ada W. S. Leung ◽  
Pierre Jolicoeur ◽  
Claude Alain

Detecting a brief silent interval (i.e., a gap) is more difficult when listeners perceive two concurrent sounds rather than one in a sound containing a mistuned harmonic in otherwise in-tune harmonics. This impairment in gap detection may reflect the interaction of low-level encoding or the division of attention between two sound objects, both of which could interfere with signal detection. To distinguish between these two alternatives, we compared ERPs during active and passive listening with complex harmonic tones that could include a gap, a mistuned harmonic, both features, or neither. During active listening, participants indicated whether they heard a gap irrespective of mistuning. During passive listening, participants watched a subtitled muted movie of their choice while the same sounds were presented. Gap detection was impaired when the complex sounds included a mistuned harmonic that popped out as a separate object. The ERP analysis revealed an early gap-related activity that was little affected by mistuning during the active or passive listening condition. However, during active listening, there was a marked decrease in the late positive wave that was thought to index attention and response-related processes. These results suggest that the limitation in detecting the gap is related to attentional processing, possibly divided attention induced by the concurrent sound objects, rather than deficits in preattentional sensory encoding.


Sign in / Sign up

Export Citation Format

Share Document