Specializations of the Auditory System for the Analysis of Natural Sounds

Author(s):  
Israel Nelken ◽  
Yaron Rotman ◽  
Omer Bar Yosef
2002 ◽  
Vol 88 (5) ◽  
pp. 2684-2699 ◽  
Author(s):  
Dennis L. Barbour ◽  
Xiaoqin Wang

Natural sounds often contain energy over a broad spectral range and consequently overlap in frequency when they occur simultaneously; however, such sounds under normal circumstances can be distinguished perceptually (e.g., the cocktail party effect). Sound components arising from different sources have distinct (i.e., incoherent) modulations, and incoherence appears to be one important cue used by the auditory system to segregate sounds into separately perceived acoustic objects. Here we show that, in the primary auditory cortex of awake marmoset monkeys, many neurons responsive to amplitude- or frequency-modulated tones at a particular carrier frequency [the characteristic frequency (CF)] also demonstrate sensitivity to the relative modulation phase between two otherwise identically modulated tones: one at CF and one at a different carrier frequency. Changes in relative modulation phase reflect alterations in temporal coherence between the two tones, and the most common neuronal response was found to be a maximum of suppression for the coherent condition. Coherence sensitivity was generally found in a narrow frequency range in the inhibitory portions of the frequency response areas (FRA), indicating that only some off-CF neuronal inputs into these cortical neurons interact with on-CF inputs on the same time scales. Over the population of neurons studied, carrier frequencies showing coherence sensitivity were found to coincide with the carrier frequencies of inhibition, implying that inhibitory inputs create the effect. The lack of strong coherence-induced facilitation also supports this interpretation. Coherence sensitivity was found to be greatest for modulation frequencies of 16–128 Hz, which is higher than the phase-locking capability of most cortical neurons, implying that subcortical neurons could play a role in the phenomenon. Collectively, these results reveal that auditory cortical neurons receive some off-CF inputs temporally matched and some temporally unmatched to the on-CF input(s) and respond in a fashion that could be utilized by the auditory system to segregate natural sounds containing similar spectral components (such as vocalizations from multiple conspecifics) based on stimulus coherence.


Author(s):  
Israel Nelken

Understanding the principles by which sensory systems represent natural stimuli is one of the holy grails of neuroscience. In the auditory system, the study of the coding of natural sounds has a particular prominence. Indeed, the relationships between neural responses to simple stimuli (usually pure tone bursts)—often used to characterize auditory neurons—and complex sounds (in particular natural sounds) may be complex. Many different classes of natural sounds have been used to study the auditory system. Sound families that researchers have used to good effect in this endeavor include human speech, species-specific vocalizations, an “acoustic biotope” selected in one way or another, and sets of artificial sounds that mimic important features of natural sounds. Peripheral and brainstem representations of natural sounds are relatively well understood. The properties of the peripheral auditory system play a dominant role, and further processing occurs mostly within the frequency channels determined by these properties. At the level of the inferior colliculus, the highest brainstem station, representational complexity increases substantially due to the convergence of multiple processing streams. Undoubtedly, the most explored part of the auditory system, in term of responses to natural sounds, is the primary auditory cortex. In spite of over 50 years of research, there is still no commonly accepted view of the nature of the population code for natural sounds in the auditory cortex. Neurons in the auditory cortex are believed by some to be primarily linear spectro-temporal filters, by others to respond to conjunctions of important sound features, or even to encode perceptual concepts such as “auditory objects.” Whatever the exact mechanism is, many studies consistently report a substantial increase in the variability of the response patterns of cortical neurons to natural sounds. The generation of such variation may be the main contribution of auditory cortex to the coding of natural sounds.


2019 ◽  
Vol 30 (4) ◽  
pp. 2600-2614 ◽  
Author(s):  
Xiangbin Teng ◽  
David Poeppel

Abstract Natural sounds contain acoustic dynamics ranging from tens to hundreds of milliseconds. How does the human auditory system encode acoustic information over wide-ranging timescales to achieve sound recognition? Previous work (Teng et al. 2017) demonstrated a temporal coding preference for the theta and gamma ranges, but it remains unclear how acoustic dynamics between these two ranges are coded. Here, we generated artificial sounds with temporal structures over timescales from ~200 to ~30 ms and investigated temporal coding on different timescales. Participants discriminated sounds with temporal structures at different timescales while undergoing magnetoencephalography recording. Although considerable intertrial phase coherence can be induced by acoustic dynamics of all the timescales, classification analyses reveal that the acoustic information of all timescales is preferentially differentiated through the theta and gamma bands, but not through the alpha and beta bands; stimulus reconstruction shows that the acoustic dynamics in the theta and gamma ranges are preferentially coded. We demonstrate that the theta and gamma bands show the generality of temporal coding with comparable capacity. Our findings provide a novel perspective—acoustic information of all timescales is discretised into two discrete temporal chunks for further perceptual analysis.


2019 ◽  
Author(s):  
Xiangbin Teng ◽  
David Poeppel

AbstractNatural sounds have broadband modulation spectra and contain acoustic dynamics ranging from tens to hundreds of milliseconds. How does the human auditory system encode acoustic information over wide-ranging timescales to achieve sound recognition? Previous work (Teng et al., 2017) demonstrated a temporal coding preference in the auditory system for the theta (4 – 7 Hz) and gamma (30 – 45 Hz) ranges, but it remains unclear how acoustic dynamics between these two ranges is encoded. Here we generated artificial sounds with temporal structures over timescales from ~200 ms to ~30 ms and investigated temporal coding on different timescales in the human auditory cortex. Participants discriminated sounds with temporal structures at different timescales while undergoing magnetoencephalography (MEG) recording. The data show robust neural entrainment in the theta and the gamma bands, but not in the alpha and beta bands. Classification analyses as well as stimulus reconstruction reveal that the acoustic information of all timescales can be differentiated through the theta and gamma bands, but the acoustic dynamics in the theta and gamma ranges are preferentially encoded. We replicate earlier findings of multi-time scale processing and further demonstrate that the theta and gamma bands show generality of temporal coding across all timescales with comparable capacity. The results support the hypothesis that the human auditory cortex primarily encodes auditory information employing neural processes within two discrete temporal regimes.SignificanceNatural sounds contain rich acoustic dynamics over wide-ranging timescales, but perceptually relevant regularities often occupy specific temporal ranges. For instance, speech carries phonemic information on a shorter timescale than syllabic information at ~ 200 ms. How does the brain efficiently ‘sample’ continuous acoustic input to perceive temporally structured sounds? We presented sounds with temporal structures at different timescales and measured cortical entrainment using magnetoencephalography. We found, unexpectedly, that the human auditory system preserves high temporal coding precision on two non-overlapping timescales, the slower (theta) and faster (gamma) bands, to track acoustic dynamics over all timescales. The results suggest that the acoustic environment which we experience as seamless and continuous is segregated by discontinuous neural processing, or ‘sampled.’


2020 ◽  
Author(s):  
X. Zhai ◽  
F. Khatami ◽  
M. Sadeghi ◽  
F. He ◽  
H.L. Read ◽  
...  

ABSTRACTThe perception of sound textures, a class of natural sounds defined by statistical sound structure such as fire wind, and rain, has been proposed to arise through the integration of time-averaged summary statistics. Where and how the auditory system might encode these summary statistics to create internal representations of these stationary sounds, however, is unknown. Here, using natural textures and synthetic variants with reduced statistics, we show that summary statistics modulate the correlations between frequency organized neuron ensembles in the awake rabbit inferior colliculus. These neural ensemble correlation statistics capture high-order sound structure and allow for accurate neural decoding in a single trial recognition task with evidence accumulation times approaching 1 s. In contrast, the average activity across the neural ensemble (neural spectrum) provides a fast (tens of ms) and salient signal that contributes primarily to texture discrimination. Intriguingly, perceptual studies in human listeners reveals analogous trends: the sound spectrum is integrated quickly and serves as salient discrimination cue while high-order sound statistics are integrated slowly and contribute substantially more towards recognition. The findings suggest statistical sound cues such as the sound spectrum and correlation structure are represented by distinct response statistics in auditory midbrain ensembles, and that these neural response statistics may have dissociable roles and time scales for the recognition and discrimination of natural sounds.SIGNIFICANCE STATEMENTBeing able to recognize and discriminate natural sounds, such as from a running stream, a crowd clapping, or ruffling leaves is a critical task of the normal functioning auditory system. Humans can easily perform such tasks, yet they can be particularly difficult for the hearing impaired and they challenge our most sophisticated computer algorithms. This difficulty is attributed to the complex physical structure of such natural sounds and the fact they are not unique: they vary randomly in a statistically defined manner from one excerpt to the other. Here we provide the first evidence, to our knowledge, that the central auditory system is able to encode and utilize statistical sound cues for natural sound recognition and discrimination behaviors.


Author(s):  
Nuriye Yıldırım Gökay ◽  
Bülent Gündüz ◽  
Fatih Söke ◽  
Recep Karamert

Purpose The effects of neurological diseases on the auditory system have been a notable issue for investigators because the auditory pathway is closely associated with neural systems. The purposes of this study are to evaluate the efferent auditory system function and hearing quality in Parkinson's disease (PD) and to compare the findings with age-matched individuals without PD to present a perspective on aging. Method The study included 35 individuals with PD (mean age of 48.50 ± 8.00 years) and 35 normal-hearing peers (mean age of 49 ± 10 years). The following tests were administered for all participants: the first section of the Speech, Spatial and Qualities of Hearing Scale; pure-tone audiometry, speech audiometry, tympanometry, and acoustic reflexes; and distortion product otoacoustic emissions (DPOAEs) and contralateral suppression of DPOAEs. SPSS Version 25 was used for statistical analyses, and values of p < .05 were considered statistically significant. Results There were no statistically significant differences in the pure-tone audiometry thresholds and DPOAE responses between the individuals with PD and their normal-hearing peers ( p = .732). However, statistically significant differences were found between the groups in suppression levels of DPOAEs and hearing quality ( p < .05). In addition, a statistically significant and positive correlation was found between the amount of suppression at some frequencies and the Speech, Spatial and Qualities of Hearing Scale scores. Conclusions This study indicates that medial olivocochlear efferent system function and the hearing quality of individuals with PD were affected adversely due to the results of PD pathophysiology on the hearing system. For optimal intervention and follow-up, tasks related to hearing quality in daily life can also be added to therapies for PD.


2005 ◽  
Vol 32 (Spring) ◽  
pp. 5-10 ◽  
Author(s):  
Elizabeth Hester
Keyword(s):  

Sign in / Sign up

Export Citation Format

Share Document