scholarly journals Envelope reconstruction of speech and music highlights stronger tracking of speech at low frequencies

2021 ◽  
Vol 17 (9) ◽  
pp. e1009358
Author(s):  
Nathaniel J. Zuk ◽  
Jeremy W. Murphy ◽  
Richard B. Reilly ◽  
Edmund C. Lalor

The human brain tracks amplitude fluctuations of both speech and music, which reflects acoustic processing in addition to the encoding of higher-order features and one’s cognitive state. Comparing neural tracking of speech and music envelopes can elucidate stimulus-general mechanisms, but direct comparisons are confounded by differences in their envelope spectra. Here, we use a novel method of frequency-constrained reconstruction of stimulus envelopes using EEG recorded during passive listening. We expected to see music reconstruction match speech in a narrow range of frequencies, but instead we found that speech was reconstructed better than music for all frequencies we examined. Additionally, models trained on all stimulus types performed as well or better than the stimulus-specific models at higher modulation frequencies, suggesting a common neural mechanism for tracking speech and music. However, speech envelope tracking at low frequencies, below 1 Hz, was associated with increased weighting over parietal channels, which was not present for the other stimuli. Our results highlight the importance of low-frequency speech tracking and suggest an origin from speech-specific processing in the brain.

2021 ◽  
Author(s):  
Nathaniel J Zuk ◽  
Jeremy W Murphy ◽  
Richard B Reilly ◽  
Edmund C Lalor

AbstractThe human brain tracks amplitude fluctuations of both speech and music, which reflects acoustic processing in addition to the processing of higher-order features and one’s cognitive state. Comparing neural tracking of speech and music envelopes can elucidate stimulus-general mechanisms, but direct comparisons are confounded by differences in their envelope spectra. Here, we use a novel method of frequency-constrained reconstruction of stimulus envelopes using EEG recorded during passive listening. We expected to see music reconstruction match speech in a narrow range of frequencies, but instead we found that speech was reconstructed better than music for all frequencies we examined. Additionally, speech envelope tracking at low frequencies, below 1 Hz, was uniquely associated with increased weighting over parietal channels. Our results highlight the importance of low-frequency speech tracking and its origin from speech-specific processing in the brain.


2020 ◽  
Vol 10 (1) ◽  
Author(s):  
Maya Inbar ◽  
Eitan Grossman ◽  
Ayelet N. Landau

Abstract Studies of speech processing investigate the relationship between temporal structure in speech stimuli and neural activity. Despite clear evidence that the brain tracks speech at low frequencies (~ 1 Hz), it is not well understood what linguistic information gives rise to this rhythm. In this study, we harness linguistic theory to draw attention to Intonation Units (IUs), a fundamental prosodic unit of human language, and characterize their temporal structure as captured in the speech envelope, an acoustic representation relevant to the neural processing of speech. IUs are defined by a specific pattern of syllable delivery, together with resets in pitch and articulatory force. Linguistic studies of spontaneous speech indicate that this prosodic segmentation paces new information in language use across diverse languages. Therefore, IUs provide a universal structural cue for the cognitive dynamics of speech production and comprehension. We study the relation between IUs and periodicities in the speech envelope, applying methods from investigations of neural synchronization. Our sample includes recordings from every-day speech contexts of over 100 speakers and six languages. We find that sequences of IUs form a consistent low-frequency rhythm and constitute a significant periodic cue within the speech envelope. Our findings allow to predict that IUs are utilized by the neural system when tracking speech. The methods we introduce here facilitate testing this prediction in the future (i.e., with physiological data).


ACTA IMEKO ◽  
2020 ◽  
Vol 9 (5) ◽  
pp. 374
Author(s):  
Henrik Ingerslev ◽  
Soren Andresen ◽  
Jacob Holm Winther

The demand from industry to produce accurate acceleration measurements down to ever lower frequencies and with ever lower noise is increasing. Different vibration transducers are used today for many different purposes within this area, like detection and warning for earthquakes, detection of nuclear testing, and monitoring of the environment. Accelerometers for such purposes must be calibrated in order to yield trustworthy results and provide traceability to the SI-system accordingly. For these calibrations to be feasible, suitable ultra low-noise accelerometers and/or signal processing functions are needed. <br />Here we present two digital signal processing (DSP) functions designed to measure ultra low-noise acceleration in calibration systems. The DSP functions use dual channel signal analysis on signals from two accelerometers measuring the same stimuli and use the coherence between the two signals to reduce noise. Simulations show that the two DSP functions are estimating calibration signals better than the standard analysis. <br />The results presented here are intended to be used in key comparison studies of accelerometer calibration systems, and may help extend current general low frequency range from e.g. 100 mHz down to ultra-low frequencies of around 10mHz, possibly using somewhat same instrumentation.


2019 ◽  
Author(s):  
Maya Inbar ◽  
Eitan Grossman ◽  
Ayelet N. Landau

AbstractStudies of speech processing investigate the relationship between temporal structure in speech stimuli and neural activity. Despite clear evidence that the brain tracks speech at low frequencies (~1 Hz), it is not well understood what linguistic information gives rise to this rhythm. Here, we harness linguistic theory to draw attention to Intonation Units (IUs), a fundamental prosodic unit of human language, and characterize their temporal structure as captured in the speech envelope, an acoustic representation relevant to the neural processing of speech.IUs are defined by a specific pattern of syllable delivery, together with resets in pitch and articulatory force. Linguistic studies of spontaneous speech indicate that this prosodic segmentation paces new information in language use across diverse languages. Therefore, IUs provide a universal structural cue for the cognitive dynamics of speech production and comprehension.We study the relation between IUs and periodicities in the speech envelope, applying methods from investigations of neural synchronization. Our sample includes recordings from every-day speech contexts of over 100 speakers and six languages. We find that sequences of IUs form a consistent low-frequency rhythm and constitute a significant periodic cue within the speech envelope. Our findings allow to predict that IUs are utilized by the neural system when tracking speech, and the methods we introduce facilitate testing this prediction given physiological data.


2020 ◽  
Vol 635 ◽  
pp. A150 ◽  
Author(s):  
F. de Gasperin ◽  
J. Vink ◽  
J. P. McKean ◽  
A. Asgekar ◽  
I. Avruch ◽  
...  

Context. The four persistent radio sources in the northern sky with the highest flux density at metre wavelengths are Cassiopeia A, Cygnus A, Taurus A, and Virgo A; collectively they are called the A-team. Their flux densities at ultra-low frequencies (< 100 MHz) can reach several thousands of janskys, and they often contaminate observations of the low-frequency sky by interfering with image processing. Furthermore, these sources are foreground objects for all-sky observations hampering the study of faint signals, such as the cosmological 21 cm line from the epoch of reionisation. Aims. We aim to produce robust models for the surface brightness emission as a function of frequency for the A-team sources at ultra-low frequencies. These models are needed for the calibration and imaging of wide-area surveys of the sky with low-frequency interferometers. This requires obtaining images at an angular resolution better than 15″ with a high dynamic range and good image fidelity. Methods. We observed the A-team with the Low Frequency Array (LOFAR) at frequencies between 30 MHz and 77 MHz using the Low Band Antenna system. We reduced the datasets and obtained an image for each A-team source. Results. The paper presents the best models to date for the sources Cassiopeia A, Cygnus A, Taurus A, and Virgo A between 30 MHz and 77 MHz. We were able to obtain the aimed resolution and dynamic range in all cases. Owing to its compactness and complexity, observations with the long baselines of the International LOFAR Telescope will be required to improve the source model for Cygnus A further.


2017 ◽  
Vol 284 (1864) ◽  
pp. 20171670 ◽  
Author(s):  
Molly C. Womack ◽  
Jakob Christensen-Dalsgaard ◽  
Luis A. Coloma ◽  
Juan C. Chaparro ◽  
Kim L. Hoke

Sensory losses or reductions are frequently attributed to relaxed selection. However, anuran species have lost tympanic middle ears many times, despite anurans' use of acoustic communication and the benefit of middle ears for hearing airborne sound. Here we determine whether pre-existing alternative sensory pathways enable anurans lacking tympanic middle ears (termed earless anurans) to hear airborne sound as well as eared species or to better sense vibrations in the environment. We used auditory brainstem recordings to compare hearing and vibrational sensitivity among 10 species (six eared, four earless) within the Neotropical true toad family (Bufonidae). We found that species lacking middle ears are less sensitive to high-frequency sounds, however, low-frequency hearing and vibrational sensitivity are equivalent between eared and earless species. Furthermore, extratympanic hearing sensitivity varies among earless species, highlighting potential species differences in extratympanic hearing mechanisms. We argue that ancestral bufonids may have sufficient extratympanic hearing and vibrational sensitivity such that earless lineages tolerated the loss of high frequency hearing sensitivity by adopting species-specific behavioural strategies to detect conspecifics, predators and prey.


Geophysics ◽  
1992 ◽  
Vol 57 (6) ◽  
pp. 854-859 ◽  
Author(s):  
Xiao Ming Tang

A new technique for measuring elastic wave attenuation in the frequency range of 10–150 kHz consists of measuring low‐frequency waveforms using two cylindrical bars of the same material but of different lengths. The attenuation is obtained through two steps. In the first, the waveform measured within the shorter bar is propagated to the length of the longer bar, and the distortion of the waveform due to the dispersion effect of the cylindrical waveguide is compensated. The second step is the inversion for the attenuation or Q of the bar material by minimizing the difference between the waveform propagated from the shorter bar and the waveform measured within the longer bar. The waveform inversion is performed in the time domain, and the waveforms can be appropriately truncated to avoid multiple reflections due to the finite size of the (shorter) sample, allowing attenuation to be measured at long wavelengths or low frequencies. The frequency range in which this technique operates fills the gap between the resonant bar measurement (∼10 kHz) and ultrasonic measurement (∼100–1000 kHz). By using the technique, attenuation values in a PVC (a highly attenuative) material and in Sierra White granite were measured in the frequency range of 40–140 kHz. The obtained attenuation values for the two materials are found to be reliable and consistent.


2019 ◽  
Vol 219 (2) ◽  
pp. 975-994 ◽  
Author(s):  
Gabriel Gribler ◽  
T Dylan Mikesell

SUMMARY Estimating shear wave velocity with depth from Rayleigh-wave dispersion data is limited by the accuracy of fundamental and higher mode identification and characterization. In many cases, the fundamental mode signal propagates exclusively in retrograde motion, while higher modes propagate in prograde motion. It has previously been shown that differences in particle motion can be identified with multicomponent recordings and used to separate prograde from retrograde signals. Here we explore the domain of existence of prograde motion of the fundamental mode, arising from a combination of two conditions: (1) a shallow, high-impedance contrast and (2) a high Poisson ratio material. We present solutions to isolate fundamental and higher mode signals using multicomponent recordings. Previously, a time-domain polarity mute was used with limited success due to the overlap in the time domain of fundamental and higher mode signals at low frequencies. We present several new approaches to overcome this low-frequency obstacle, all of which utilize the different particle motions of retrograde and prograde signals. First, the Hilbert transform is used to phase shift one component by 90° prior to summation or subtraction of the other component. This enhances either retrograde or prograde motion and can increase the mode amplitude. Secondly, we present a new time–frequency domain polarity mute to separate retrograde and prograde signals. We demonstrate these methods with synthetic and field data to highlight the improvements to dispersion images and the resulting dispersion curve extraction.


1990 ◽  
Vol 69 (2) ◽  
pp. 630-639 ◽  
Author(s):  
M. Modarreszadeh ◽  
E. N. Bruce ◽  
B. Gothe

We analyzed breath-to-breath inspiratory time (TI), expiratory time (TE), inspiratory volume (VI), and minute ventilation (Vm) from 11 normal subjects during stage 2 sleep. The analysis consisted of 1) fitting first- and second-order autoregressive models (AR1 and AR2) and 2) obtaining the power spectra of the data by fast-Fourier transform. For the AR2 model, the only coefficients that were statistically different from zero were the average alpha 1 (a1) for TI, VI, and Vm (a1 = 0.19, 0.29, and 0.15, respectively). However, the power spectra of all parameters often exhibited peaks at low frequency (less than 0.2 cycles/breath) and/or at high frequency (greater than 0.2 cycles/breath), indicative of periodic oscillations. After accounting for the corrupting effects of added oscillations on the a1 estimates, we conclude that 1) breath-to-breath fluctuations of VI, and to a lesser extent TI and Vm, exhibit a first-order autoregressive structure such that fluctuations of each breath are positively correlated with those of immediately preceding breaths and 2) the correlated components of variability in TE are mostly due to discrete high- and/or low-frequency oscillations with no underlying autoregressive structure. We propose that the autoregressive structure of VI, TI, and Vm during spontaneous breathing in stage 2 sleep may reflect either a central neural mechanism or the effects of noise in respiratory chemical feedback loops; the presence of low-frequency oscillations, seen more often in Vm, suggests possible instability in the chemical feedback loops. Mechanisms of high-frequency periodicities, seen more often in TE, are unknown.


Sign in / Sign up

Export Citation Format

Share Document