Auditory scene analysis by songbirds: Stream segregation of birdsong by European starlings (Sturnus vulgaris).

1997 ◽  
Vol 111 (1) ◽  
pp. 3-13 ◽  
Author(s):  
Stewart H. Hulse ◽  
Scott A. MacDougall-Shackleton ◽  
Amy B. Wisniewski
1998 ◽  
Vol 103 (6) ◽  
pp. 3581-3587 ◽  
Author(s):  
Scott A. MacDougall-Shackleton ◽  
Stewart H. Hulse ◽  
Timothy Q. Gentner ◽  
Wesley White

2012 ◽  
Vol 107 (9) ◽  
pp. 2366-2382 ◽  
Author(s):  
Yonatan I. Fishman ◽  
Christophe Micheyl ◽  
Mitchell Steinschneider

The ability to detect and track relevant acoustic signals embedded in a background of other sounds is crucial for hearing in complex acoustic environments. This ability is exemplified by a perceptual phenomenon known as “rhythmic masking release” (RMR). To demonstrate RMR, a sequence of tones forming a target rhythm is intermingled with physically identical “Distracter” sounds that perceptually mask the rhythm. The rhythm can be “released from masking” by adding “Flanker” tones in adjacent frequency channels that are synchronous with the Distracters. RMR represents a special case of auditory stream segregation, whereby the target rhythm is perceptually segregated from the background of Distracters when they are accompanied by the synchronous Flankers. The neural basis of RMR is unknown. Previous studies suggest the involvement of primary auditory cortex (A1) in the perceptual organization of sound patterns. Here, we recorded neural responses to RMR sequences in A1 of awake monkeys in order to identify neural correlates and potential mechanisms of RMR. We also tested whether two current models of stream segregation, when applied to these responses, could account for the perceptual organization of RMR sequences. Results suggest a key role for suppression of Distracter-evoked responses by the simultaneous Flankers in the perceptual restoration of the target rhythm in RMR. Furthermore, predictions of stream segregation models paralleled the psychoacoustics of RMR in humans. These findings reinforce the view that preattentive or “primitive” aspects of auditory scene analysis may be explained by relatively basic neural mechanisms at the cortical level.


2017 ◽  
Vol 60 (10) ◽  
pp. 2989-3000 ◽  
Author(s):  
Elyse S. Sussman

Purpose This review article provides a new perspective on the role of attention in auditory scene analysis. Method A framework for understanding how attention interacts with stimulus-driven processes to facilitate task goals is presented. Previously reported data obtained through behavioral and electrophysiological measures in adults with normal hearing are summarized to demonstrate attention effects on auditory perception—from passive processes that organize unattended input to attention effects that act at different levels of the system. Data will show that attention can sharpen stream organization toward behavioral goals, identify auditory events obscured by noise, and limit passive processing capacity. Conclusions A model of attention is provided that illustrates how the auditory system performs multilevel analyses that involve interactions between stimulus-driven input and top-down processes. Overall, these studies show that (a) stream segregation occurs automatically and sets the basis for auditory event formation; (b) attention interacts with automatic processing to facilitate task goals; and (c) information about unattended sounds is not lost when selecting one organization over another. Our results support a neural model that allows multiple sound organizations to be held in memory and accessed simultaneously through a balance of automatic and task-specific processes, allowing flexibility for navigating noisy environments with competing sound sources. Presentation Video http://cred.pubs.asha.org/article.aspx?articleid=2601618


2009 ◽  
Vol 101 (6) ◽  
pp. 3212-3225 ◽  
Author(s):  
Naoya Itatani ◽  
Georg M. Klump

Streaming in auditory scene analysis refers to the perceptual grouping of multiple interleaved sounds having similar characteristics while sounds with different characteristics are segregated. In human perception, auditory streaming occurs on the basis of temporal features of sounds such as the rate of amplitude modulation. We present results from multiunit recordings in the auditory forebrain of awake European starlings ( Sturnus vulgaris) on the representation of sinusoidally amplitude modulated (SAM) tones to investigate the effect of temporal envelope structure on neural stream segregation. Different types of rate modulation transfer functions in response to SAM tones were observed. The strongest responses were found for modulation frequencies (fmod) <160 Hz. The streaming stimulus consisted of sequences of alternating SAM tones with the same carrier frequency but differing in fmod (ABA-ABA-ABA-…). A signals had a modulation frequency evoking a large excitation, whereas the fmod of B signals was ≤4 octaves higher. Synchrony of B signal responses to the modulation decreased as fmod increased. Spike rate in response to B signals dropped as fmod increased. Faster signal repetition resulted in fewer spikes, suggesting the contribution of forward suppression to the response that may be due to both signals having similar spectral energy and that is not related to the temporal pattern of modulation. These two effects are additive and may provide the basis for a more separated representation of A and B signals by two populations of neurons that can be viewed as a neuronal correlate of segregated streams.


2014 ◽  
Vol 78 (3) ◽  
pp. 361-378 ◽  
Author(s):  
Mona Isabel Spielmann ◽  
Erich Schröger ◽  
Sonja A. Kotz ◽  
Alexandra Bendixen

Sign in / Sign up

Export Citation Format

Share Document