harmonic complex
Recently Published Documents


TOTAL DOCUMENTS

85
(FIVE YEARS 15)

H-INDEX

15
(FIVE YEARS 2)

Author(s):  
Joseph D Wagner ◽  
Alice Gelman ◽  
Kenneth E. Hancock ◽  
Yoojin Chung ◽  
Bertrand Delgutte

The pitch of harmonic complex tones (HCT) common in speech, music and animal vocalizations plays a key role in the perceptual organization of sound. Unraveling the neural mechanisms of pitch perception requires animal models but little is known about complex pitch perception by animals, and some species appear to use different pitch mechanisms than humans. Here, we tested rabbits' ability to discriminate the fundamental frequency (F0) of HCTs with missing fundamentals using a behavioral paradigm inspired by foraging behavior in which rabbits learned to harness a spatial gradient in F0 to find the location of a virtual target within a room for a food reward. Rabbits were initially trained to discriminate HCTs with F0s in the range 400-800 Hz and with harmonics covering a wide frequency range (800-16,000 Hz), and then tested with stimuli differing either in spectral composition to test the role of harmonic resolvability (Experiment 1), or in F0 range (Experiment 2), or both F0 and spectral content (Experiment 3). Together, these experiments show that rabbits can discriminate HCTs over a wide F0 range (200-1600 Hz) encompassing the range of conspecific vocalizations, and can use either the spectral pattern of harmonics resolved by the cochlea for higher F0s or temporal envelope cues resulting from interaction between unresolved harmonics for lower F0s. The qualitative similarity of these results to human performance supports using rabbits as an animal model for studies of pitch mechanisms providing species differences in cochlear frequency selectivity and F0 range of vocalizations are taken into account.


2021 ◽  
Vol 11 (12) ◽  
pp. 1592
Author(s):  
Devin Inabinet ◽  
Jan De La Cruz ◽  
Justin Cha ◽  
Kevin Ng ◽  
Gabriella Musacchia

The perception of harmonic complexes provides important information for musical and vocal communication. Numerous studies have shown that musical training and expertise are associated with better processing of harmonic complexes, however, it is unclear whether the perceptual improvement associated with musical training is universal to different pitch models. The current study addresses this issue by measuring discrimination thresholds of musicians (n = 20) and non-musicians (n = 18) to diotic (same sound to both ears) and dichotic (different sounds to each ear) sounds of four stimulus types: (1) pure sinusoidal tones, PT; (2) four-harmonic complex tones, CT; (3) iterated rippled noise, IRN; and (4) interaurally correlated broadband noise, called the “Huggins” or “dichotic” pitch, DP. Frequency difference limens (DLF) for each stimulus type were obtained via a three-alternative-forced-choice adaptive task requiring selection of the interval with the highest pitch, yielding the smallest perceptible fundamental frequency (F0) distance (in Hz) between two sounds. Music skill was measured by an online test of musical pitch, melody and timing maintained by the International Laboratory for Brain Music and Sound Research. Musicianship, length of music experience and self-evaluation of musical skill were assessed by questionnaire. Results showed musicians had smaller DLFs in all four conditions with the largest group difference in the dichotic condition. DLF thresholds were related to both subjective and objective musical ability. In addition, subjective self-report of musical ability was shown to be a significant variable in group classification. Taken together, the results suggest that music-related plasticity benefits multiple mechanisms of pitch encoding and that self-evaluation of musicality can be reliably associated with objective measures of perception.


2021 ◽  
Author(s):  
Emma Holmes

Pitch discrimination is better for complex tones than for pure tones, but how more subtle differences in timbre affect pitch discrimination is not fully understood. This study compared pitch discrimination thresholds of flat-spectrum harmonic complex tones with those of natural sounds played by musical instruments of three different timbres (violin, trumpet, and flute). To investigate whether natural familiarity with sounds of particular timbres affects pitch discrimination thresholds, this study recruited musicians who were trained on one of the three instruments. We found that flautists and trumpeters could discriminate smaller differences in pitch for artificial flat-spectrum tones, despite their unfamiliar timbre, than for sounds played by musical instruments, which are regularly heard in everyday life (particularly by musicians who play those instruments). Furthermore, thresholds were no better for the instrument a musician was trained to play than for other instruments, suggesting that even extensive experience listening to and producing sounds of particular timbres does not reliably improve pitch discrimination thresholds for those timbres. The results show that timbre familiarity provides minimal improvements to auditory acuity, and physical acoustics (i.e., the presence of equal-amplitude harmonics) determine pitch-discrimination thresholds more than does experience with natural sounds and timbre-specific training.


Symmetry ◽  
2021 ◽  
Vol 13 (9) ◽  
pp. 1748
Author(s):  
Dawei Shen ◽  
Claude Alain ◽  
Bernhard Ross

The presence of binaural low-level background noise has been shown to enhance the transient evoked N1 response at about 100 ms after sound onset. This increase in N1 amplitude is thought to reflect noise-mediated efferent feedback facilitation from the auditory cortex to lower auditory centers. To test this hypothesis, we recorded auditory-evoked fields using magnetoencephalography while participants were presented with binaural harmonic complex tones embedded in binaural or monaural background noise at signal-to-noise ratios of 25 dB (low noise) or 5 dB (higher noise). Half of the stimuli contained a gap in the middle of the sound. The source activities were measured in bilateral auditory cortices. The onset and gap N1 response increased with low binaural noise, but high binaural and low monaural noise did not affect the N1 amplitudes. P1 and P2 onset and gap responses were consistently attenuated by background noise, and noise level and binaural/monaural presentation showed distinct effects. Moreover, the evoked gamma synchronization was also reduced by background noise, and it showed a lateralized reduction for monaural noise. The effects of noise on the N1 amplitude follow a bell-shaped characteristic that could reflect an optimal representation of acoustic information for transient events embedded in noise.


PLoS ONE ◽  
2021 ◽  
Vol 16 (4) ◽  
pp. e0249654
Author(s):  
Sara M. K. Madsen ◽  
Torsten Dau ◽  
Andrew J. Oxenham

Differences in fundamental frequency (F0) or pitch between competing voices facilitate our ability to segregate a target voice from interferers, thereby enhancing speech intelligibility. Although lower-numbered harmonics elicit a stronger and more accurate pitch sensation than higher-numbered harmonics, it is unclear whether the stronger pitch leads to an increased benefit of pitch differences when segregating competing talkers. To answer this question, sentence recognition was tested in young normal-hearing listeners in the presence of a single competing talker. The stimuli were presented in a broadband condition or were highpass or lowpass filtered to manipulate the pitch accuracy of the voicing, while maintaining roughly equal speech intelligibility in the highpass and lowpass regions. Performance was measured with average F0 differences (ΔF0) between the target and single-talker masker of 0, 2, and 4 semitones. Pitch discrimination abilities were also measured to confirm that the lowpass-filtered stimuli elicited greater pitch accuracy than the highpass-filtered stimuli. No interaction was found between filter type and ΔF0 in the sentence recognition task, suggesting little or no effect of harmonic rank or pitch accuracy on the ability to use F0 to segregate natural voices, even when the average ΔF0 is relatively small. The results suggest that listeners are able to obtain some benefit of pitch differences between competing voices, even when pitch salience and accuracy is low. The accuracy with which we are able to discriminate the pitch of a harmonic complex tone depends on the F0 and the harmonic numbers present. For F0s in the average range of speech (100–200 Hz), pitch discrimination is best (implying accurate F0 coding) when harmonics below about the 10th are present [6–10]. When these lower-numbered harmonics are present, pitch discrimination is also independent of the phase relationships between the harmonics, suggesting that these harmonics are spectrally resolved to some extent. In contrast, when only harmonics above the 10th are present in this range of F0s, pitch discrimination is poorer and is affected by the phase relationships between harmonics, suggesting that interactions occur between these spectrally unresolved harmonics [6–10]. Psychoacoustic studies of sound segregation have often been carried out with interleaved sequences of tones. Some of these studies have investigated segregation based on differences in pitch accuracy and have varied the accuracy by systematically varying whether resolved or only unresolved harmonics are present. Previous studies have found that stream segregation can occur with alternating sequences of tones, even if the tones consist only of unresolved harmonics [11–14]. However, the question of whether streaming is greater with resolved than unresolved harmonics has received mixed answers. In cases where the listeners’ task was to segregate the streams, some studies have shown little difference in streaming between conditions containing resolved or only unresolved harmonics [11, 15], whereas another study using a similar approach found significantly greater stream segregation when resolved harmonics were present than when only unresolved harmonics were present [12]. However, in situations where the task was either neutral or encouraged listeners to integrate the sequences into a single stream, the results have been consistent across studies in showing greater segregation for complex tones containing resolved harmonics than for tones containing only unresolved harmonics [13, 14]. These findings support the idea that pitch accuracy can affect our ability to segregate sounds. Less is known about the role of low-numbered harmonics in the context of segregating competing speech. Bird and Darwin [2] showed that lower harmonics dominate performance in a speech-segregation task based on F0 differences, but they did not test any conditions containing only high-numbered harmonics. Oxenham and Simonson [16] explored the effect of harmonic rank on speech intelligibility by comparing conditions where the target and single-talker masker had been lowpass (LP) or highpass (HP) filtered to either retain (LP-filtered) or remove (HP-filtered) the spectrally resolved components from the target and masker [16]. The LP and HP cutoff frequencies were selected to produce roughly equal performance in noise for both conditions. Surprisingly, performance in the LP and HP conditions improved by similar amounts when the noise masker was replaced by a single-talker masker with a different average F0, suggesting no clear benefit of having resolved harmonic components in the speech. However, that study only used relatively large values of average ΔF0 that according to recent F0 estimates were approximately 4 and 8 semitones (ST). Moreover, this study did not parametrically vary the ΔF0 between the target and masker. It may be that pitch accuracy is only relevant for more challenging conditions, i.e. for conditions with smaller average values of ΔF0. Thus, it remains unclear whether the effect of ΔF0 on performance is affected by the presence or absence of low-numbered, spectrally resolved harmonics. The aim of the present study was to determine whether there is an effect of spectral region, and hence pitch coding accuracy, on the ability of listeners to use average F0 differences between a target and an interfering talker to understand natural speech.


2021 ◽  
Vol 11 (1) ◽  
pp. 38-46
Author(s):  
Fan-Yin Cheng ◽  
Craig A. Champlin

Temporal acuity is the ability to differentiate between sounds based on fluctuations in the waveform envelope. The proximity of successive sounds and background noise diminishes the ability to track rapid changes between consecutive sounds. We determined whether a physiological correlate of temporal acuity is also affected by these factors. We recorded the auditory brainstem response (ABR) from human listeners using a harmonic complex (S1) followed by a brief tone burst (S2) with the latter serving as the evoking signal. The duration and depth of the silent gap between S1 and S2 were manipulated, and the peak latency and amplitude of wave V were measured. The latency of the responses decreased significantly as the duration or depth of the gap increased. The amplitude of the responses was not affected by the duration or depth of the gap. These findings suggest that changing the physical parameters of the gap affects the auditory system’s ability to encode successive sounds.


2021 ◽  
Vol 33 (1) ◽  
pp. 8-27
Author(s):  
Mylène Barbaroux ◽  
Arnaud Norena ◽  
Maud Rasamimanana ◽  
Eric Castet ◽  
Mireille Besson

Musical expertise has been shown to positively influence high-level speech abilities such as novel word learning. This study addresses the question whether low-level enhanced perceptual skills causally drives successful novel word learning. We used a longitudinal approach with psychoacoustic procedures to train 2 groups of nonmusicians either on pitch discrimination or on intensity discrimination, using harmonic complex sounds. After short (approximately 3 hr) psychoacoustic training, discrimination thresholds were lower on the specific feature (pitch or intensity) that was trained. Moreover, compared to the intensity group, participants trained on pitch were faster to categorize words varying in pitch. Finally, although the N400 components in both the word learning phase and in the semantic task were larger in the pitch group than in the intensity group, no between-group differences were found at the behavioral level in the semantic task. Thus, these results provide mixed evidence that enhanced perception of relevant features through a few hours of acoustic training with harmonic sounds causally impacts the categorization of speech sounds as well as novel word learning. These results are discussed within the framework of near and far transfer effects from music training to speech processing.


Mathematics ◽  
2020 ◽  
Vol 8 (11) ◽  
pp. 2041
Author(s):  
Georgia Irina Oros

The theory of differential subordinations has been extended from the analytic functions to the harmonic complex-valued functions in 2015. In a recent paper published in 2019, the authors have considered the dual problem of the differential subordination for the harmonic complex-valued functions and have defined the differential superordination for harmonic complex-valued functions. Finding the best subordinant of a differential superordination is among the main purposes in this research subject. In this article, conditions for a harmonic complex-valued function p to be the best subordinant of a differential superordination for harmonic complex-valued functions are given. Examples are also provided to show how the theoretical findings can be used and also to prove the connection with the results obtained in 2015.


2019 ◽  
Vol 122 (6) ◽  
pp. 2468-2485 ◽  
Author(s):  
Yaqing Su ◽  
Bertrand Delgutte

Harmonic complex tones (HCTs) found in speech, music, and animal vocalizations evoke strong pitch percepts at their fundamental frequencies. The strongest pitches are produced by HCTs that contain harmonics resolved by cochlear frequency analysis, but HCTs containing solely unresolved harmonics also evoke a weaker pitch at their envelope repetition rate (ERR). In the auditory periphery, neurons phase lock to the stimulus envelope, but this temporal representation of ERR degrades and gives way to rate codes along the ascending auditory pathway. To assess the role of the inferior colliculus (IC) in such transformations, we recorded IC neuron responses to HCT and sinusoidally modulated broadband noise (SAMN) with varying ERR from unanesthetized rabbits. Different interharmonic phase relationships of HCT were used to manipulate the temporal envelope without changing the power spectrum. Many IC neurons demonstrated band-pass rate tuning to ERR between 60 and 1,600 Hz for HCT and between 40 and 500 Hz for SAMN. The tuning was not related to the pure-tone best frequency of neurons but was dependent on the shape of the stimulus envelope, indicating a temporal rather than spectral origin. A phenomenological model suggests that the tuning may arise from peripheral temporal response patterns via synaptic inhibition. We also characterized temporal coding to ERR. Some IC neurons could phase lock to the stimulus envelope up to 900 Hz for either HCT or SAMN, but phase locking was weaker with SAMN. Together, the rate code and the temporal code represent a wide range of ERR, providing strong cues for the pitch of unresolved harmonics. NEW & NOTEWORTHY Envelope repetition rate (ERR) provides crucial cues for pitch perception of frequency components that are not individually resolved by the cochlea, but the neural representation of ERR for stimuli containing many harmonics is poorly characterized. Here we show that the pitch of stimuli with unresolved harmonics is represented by both a rate code and a temporal code for ERR in auditory midbrain neurons and propose possible underlying neural mechanisms with a computational model.


2019 ◽  
Vol 64 (4) ◽  
pp. 487-496 ◽  
Author(s):  
Georgia Irina Oros ◽  
◽  
Gheorghe Oros ◽  
◽  

Sign in / Sign up

Export Citation Format

Share Document