acoustic correlates
Recently Published Documents


TOTAL DOCUMENTS

409
(FIVE YEARS 75)

H-INDEX

34
(FIVE YEARS 3)

PLoS ONE ◽  
2022 ◽  
Vol 17 (1) ◽  
pp. e0261151
Author(s):  
Jonna K. Vuoskoski ◽  
Janis H. Zickfeld ◽  
Vinoo Alluri ◽  
Vishnu Moorthigari ◽  
Beate Seibt

The experience often described as feeling moved, understood chiefly as a social-relational emotion with social bonding functions, has gained significant research interest in recent years. Although listening to music often evokes what people describe as feeling moved, very little is known about the appraisals or musical features contributing to the experience. In the present study, we investigated experiences of feeling moved in response to music using a continuous rating paradigm. A total of 415 US participants completed an online experiment where they listened to seven moving musical excerpts and rated their experience while listening. Each excerpt was randomly coupled with one of seven rating scales (perceived sadness, perceived joy, feeling moved or touched, sense of connection, perceived beauty, warmth [in the chest], or chills) for each participant. The results revealed that musically evoked experiences of feeling moved are associated with a similar pattern of appraisals, physiological sensations, and trait correlations as feeling moved by videos depicting social scenarios (found in previous studies). Feeling moved or touched by both sadly and joyfully moving music was associated with experiencing a sense of connection and perceiving joy in the music, while perceived sadness was associated with feeling moved or touched only in the case of sadly moving music. Acoustic features related to arousal contributed to feeling moved only in the case of joyfully moving music. Finally, trait empathic concern was positively associated with feeling moved or touched by music. These findings support the role of social cognitive and empathic processes in music listening, and highlight the social-relational aspects of feeling moved or touched by music.


Author(s):  
Chieh Kao ◽  
Maria D. Sera ◽  
Yang Zhang

Purpose: The aim of this study was to investigate infants' listening preference for emotional prosodies in spoken words and identify their acoustic correlates. Method: Forty-six 3- to-12-month-old infants ( M age = 7.6 months) completed a central fixation (or look-to-listen) paradigm in which four emotional prosodies (happy, sad, angry, and neutral) were presented. Infants' looking time to the string of words was recorded as a proxy of their listening attention. Five acoustic variables—mean fundamental frequency (F0), word duration, intensity variation, harmonics-to-noise ratio (HNR), and spectral centroid—were also analyzed to account for infants' attentiveness to each emotion. Results: Infants generally preferred affective over neutral prosody, with more listening attention to the happy and sad voices. Happy sounds with breathy voice quality (low HNR) and less brightness (low spectral centroid) maintained infants' attention more. Sad speech with shorter word duration (i.e., faster speech rate), less breathiness, and more brightness gained infants' attention more than happy speech did. Infants listened less to angry than to happy and sad prosodies, and none of the acoustic variables were associated with infants' listening interests in angry voices. Neutral words with a lower F0 attracted infants' attention more than those with a higher F0. Neither age nor sex effects were observed. Conclusions: This study provides evidence for infants' sensitivity to the prosodic patterns for the basic emotion categories in spoken words and how the acoustic properties of emotional speech may guide their attention. The results point to the need to study the interplay between early socioaffective and language development.


Author(s):  
Sabrina Bendjaballah ◽  
David Le Gac

This article seeks to determine the acoustic correlates of gemination in Standard Somali (Afroasiatic, Cushitic), in particular whether closure duration is the primary acoustic correlate distinguishing singleton and geminate stops, with immediate consequences for the analysis of word-initial strengthening. We provide an acoustic analysis of word-initial and word-internal voiced singletons as well as of their geminate counterparts on the basis of a production experiment conducted with four native speakers. Three temporal and four non-temporal acoustic properties of /b d ɡ/ and /bb dd ɡɡ/ are examined and systematically compared (closure duration, release burst duration, vowel duration; and closure amplitude, release amplitude, presence of a release burst, (de)voicing). We argue that the opposition between singleton and geminate voiced stops is primarily realized as the manner contrast approximant [β̞ ð̞ ɣ̞] vs. stop [b d ɡ]. Word-initially, Somali exhibits various peculiarities that are reminiscent of the cross-linguistically attested phenomenon of domain-initial strengthening. This article provides the first study of this phenomenon in Somali. We establish that word-initial /b d ɡ/ and word-medial /bb dd ɡɡ/ share the same closure duration, release burst duration, and vowel duration within the Prosodic Word. They also have a similar closure amplitude, and voicing properties. Moreover, the acoustic properties of word-initial /b d ɡ/ remain constant, and do not depend on their position in the prosodic hierarchy. On the basis of these results, the article also aims at providing new insights in the phonological representation of Somali geminates and word boundaries, and thus contributes to the understanding of word-initial strengthening in Somali.


2021 ◽  
Vol 2 (4) ◽  
pp. 45-79
Author(s):  
Margarita Ivanova

Vowel harmony based on the advanced tongue root (±ATR) feature is a well-known attribute of the languages of the Macro-Sudan belt. Igbo (< Benue-Congo), one of these languages, on basis of which first instrumental studies of ±ATR articulation were conducted, has an asymmetric vowel system /i, ɪ, e, a, ɔ, o, ʊ, u/, so that root ±ATR harmony exists within three pairs and a single vowel /e/. This paper describes an acoustic realization of ±ATR in Igbo. According to the hypothesis that the value of the first formant in combination with the distribution of energy over the spectrum combine into a single perceptual cue that allows to distinguish vowels opposed by ±ATR we investigate (along with the description of the formant space) three spectral parameters that were shown to be acoustic correlates of ±ATR in other Macro-Sudan languages. The results of the instrumental analysis indicate that acoustic correlates of ±ATR harmony in Igbo are the value of the first formant, its bandwidth and normalized center of gravity. In addition, all parameters relevant for the ±ATR contrast in our data divided /ɔ/ utterances in different roots into two groups with contrasting values of the feature.


Author(s):  
Karel Kleisner ◽  
Juan David Leongómez ◽  
Katarzyna Pisanski ◽  
Vojtěch Fiala ◽  
Clément Cornec ◽  
...  

The human voice carries information about a vocalizer's physical strength that listeners can perceive and that may influence mate choice and intrasexual competition. Yet, reliable acoustic correlates of strength in human speech remain unclear. Compared to speech, aggressive nonverbal vocalizations (roars) may function to maximize perceived strength, suggesting that their acoustic structure has been selected to communicate formidability, similar to the vocal threat displays of other animals. Here, we test this prediction in two non-WEIRD African samples: an urban community of Cameroonians and rural nomadic Hadza hunter–gatherers in the Tanzanian bushlands. Participants produced standardized speech and volitional roars and provided handgrip strength measures. Using acoustic analysis and information-theoretic multi-model inference and averaging techniques, we show that strength can be measured from both speech and roars, and as predicted, strength is more reliably gauged from roars than vowels, words or greetings. The acoustic structure of roars explains 40–70% of the variance in actual strength within adults of either sex. However, strength is predicted by multiple acoustic parameters whose combinations vary by sex, sample and vocal type. Thus, while roars may maximally signal strength, more research is needed to uncover consistent and likely interacting acoustic correlates of strength in the human voice. This article is part of the theme issue ‘Voice modulation: from origin and mechanism to social impact (Part I)’.


2021 ◽  
Author(s):  
Joshua Penney ◽  
Andy Gibson ◽  
Felicity Cox ◽  
Michael Proctor ◽  
Anita Szakay

2021 ◽  
pp. 002383092110377
Author(s):  
Hyoju Kim ◽  
Allard Jongman

This exploratory study investigates the acoustic correlates of the Korean three-way laryngeal stop distinction in Gyeongsang long-term (LT) transplants who were born in the Gyeongsang region but moved to Seoul to pursue higher education. Acoustic data were collected from eight LT transplants, five short-term (ST) transplants, and 11 Seoul speakers to examine whether exposure to Seoul Korean (SK) affects Gyeongsang speakers’ cue-weighting in distinguishing stops in production. LT transplants produced stimuli in both Gyeongsang and Seoul dialects. A cue-weighting model based on the acoustic data reveals that voice onset time (VOT) is less important to distinguish lenis from aspirated stops for Seoul speakers and for LT transplants’ SK, as compared to ST transplants and LT transplants’ Gyeongsang Korean (GK). In addition, fundamental frequency (F0) is more important for the lenis–aspirated distinction for Seoul speakers and LT transplants’ SK, as compared to ST and LT transplants’ GK, showing that LT transplants rely less on VOT and more on F0 to distinguish lenis from aspirated stops compared to ST transplants. LT transplants’ SK reveals that they rely more on VOT and less on F0 compared to SK speakers. The cue-weighting model of the LT transplants provide empirical evidence that a series of sound changes in GK is due to inter-dialect contact.


Languages ◽  
2021 ◽  
Vol 6 (3) ◽  
pp. 139
Author(s):  
Qandeel Hussain

Shina is an endangered Indo-Aryan (Dardic) language spoken in Gilgit, Northern Pakistan. The present study investigates the acoustic correlates of Shina’s three-way stop laryngeal contrast across five places of articulation. A wide range of acoustic correlates were measured including fundamental frequency (F0), spectral tilt (H1*-H2*, H1*-A1*, H1*-A2*, and H1*-A3*), and cepstral peak prominence (CPP). Voiceless aspirated stops were characterized by higher fundamental frequency, spectral tilt, and cepstral peak prominence, compared to voiceless unaspirated and voiced unaspirated stops. These results suggest that Shina is among those languages which have a raising effect of aspiration on the pitch and spectral tilt onsets of the following vowels. Positive correlations among fundamental frequency, spectral tilt, and cepstral peak prominence were observed. The findings of this study will contribute to the phonetic documentation of endangered Dardic languages.


2021 ◽  
Author(s):  
Hans Rutger Bosker ◽  
Marieke Hoetjes ◽  
Wim Pouw ◽  
Lieke van Maastricht

The prosody of a second language (L2) is notoriously difficult to acquire. It requires the mastery of a range of nested multimodal systems, including articulatory but also gestural signals, as hand gestures are produced in close synchrony with spoken prosody. It remains unclear how easily the articulatory and gestural systems acquire new prosodic patterns in the L2 and how the two systems interact, especially when L1 patterns interfere. This interdisciplinary pre-registered study investigates how Dutch learners of Spanish produce multimodal lexical stress in Spanish-Dutch cognates (e.g., Spanish profeSOR vs. Dutch proFESsor). Acoustic analyses assess whether gesturing helps L2 speakers to place stress on the correct syllable; and whether gesturing boosts the acoustic correlates of stress through biomechanic coupling. Moreover, motion-tracking and time-series analyses test whether gesture-prosody synchrony is enhanced for stress-matching vs. stress-mismatching cognate pairs, perhaps revealing that gestural timing is biased in the L1 (or L2) direction (e.g., Spanish profeSOR with the gesture biased towards Dutch stressed syllable -fes). Thus, we will uncover how speakers deal with manual, articulatory, and cognitive constraints that need to be brought in harmony for efficient speech production, bearing implications for theories on gesture-speech interaction and multimodal L2 acquisition.


Sign in / Sign up

Export Citation Format

Share Document