acoustic cues
Recently Published Documents


TOTAL DOCUMENTS

547
(FIVE YEARS 115)

H-INDEX

46
(FIVE YEARS 2)

Author(s):  
Youssef Elfahm ◽  
Nesrine Abajaddi ◽  
Badia Mounir ◽  
Laila Elmaazouzi ◽  
Ilham Mounir ◽  
...  

<span>Many technology systems have used voice recognition applications to transcribe a speaker’s speech into text that can be used by these systems. One of the most complex tasks in speech identification is to know, which acoustic cues will be used to classify sounds. This study presents an approach for characterizing Arabic fricative consonants in two groups (sibilant and non-sibilant). From an acoustic point of view, our approach is based on the analysis of the energy distribution, in frequency bands, in a syllable of the consonant-vowel type. From a practical point of view, our technique has been implemented, in the MATLAB software, and tested on a corpus built in our laboratory. The results obtained show that the percentage energy distribution in a speech signal is a very powerful parameter in the classification of Arabic fricatives. We obtained an accuracy of 92% for non-sibilant consonants /f, χ, ɣ, ʕ, ћ, and h/, 84% for sibilants /s, sҁ, z, Ӡ and ∫/, and 89% for the whole classification rate. In comparison to other algorithms based on neural networks and support vector machines (SVM), our classification system was able to provide a higher classification rate.</span>


2021 ◽  
Vol 12 (1) ◽  
pp. 348
Author(s):  
Vincent Martin ◽  
Isabelle Viaud-Delmon ◽  
Olivier Warusfel

Audio-only augmented reality consists of enhancing a real environment with virtual sound events. A seamless integration of the virtual events within the environment requires processing them with artificial spatialization and reverberation effects that simulate the acoustic properties of the room. However, in augmented reality, the visual and acoustic environment of the listener may not be fully mastered. This study aims to gain some insight into the acoustic cues (intensity and reverberation) that are used by the listener to form an auditory distance judgment, and to observe if these strategies can be influenced by the listener’s environment. To do so, we present a perceptual evaluation of two distance-rendering models informed by a measured Spatial Room Impulse Response. The choice of the rendering methods was made to design stimuli categories in which the availability and reproduction quality of acoustic cues are different. The proposed models have been evaluated in an online experiment gathering 108 participants who were asked to provide judgments of auditory distance about a stationary source. To evaluate the importance of environmental cues, participants had to describe the environment in which they were running the experiment, and more specifically the volume of the room and the distance to the wall they were facing. It could be shown that these context cues had a limited, but significant, influence on the perceived auditory distance.


Author(s):  
Viviane Déprez ◽  
Jeremy Daniel Yeaton

While it has long been assumed that prosody can help resolve syntactic and semantic ambiguities, empirical evidence has shown that the mapping between prosody and meaning is complex (Hirschberg & Avesani, 2000; Jackendoff, 1972). This paper investigates the prosody of ambiguous French sentences with multiple potentially negative terms that allow two semantically very distinct interpretations—a single negation reading involving so-called negative concord (NC), and a double negative reading (DN) with a positive meaning reflecting a strictly compositional interpretation—with the goal to further research on the role of prosody in ambiguities by examining whether intonation can be recruited by speakers to signal distinct interpretations of these sentences to hearers. Twenty native speakers produced transitive sentences with potentially negative terms embedded in contexts designed to elicit single-negation or double-negation readings. Analysis regarding the F0 and the duration of the utterances revealed distinct prosodic profiles for the two readings, confirming previous evidence that speakers can produce characteristic acoustic cues to signal intended distinctive meanings (Kraljic & Brennan, 2005; Syrett, Simon, & Nisula, 2014). Our results reveal that NC readings feature a focused subject and a deaccented object, in contrast to DN readings where both the subject and the object were independently focused. They do not relate DN to contradiction but link negative meaning with focus on French negative concord items (NCI). The paper discusses the implications of these findings for theoretical approaches to NC and outlines further questions for the syntax-prosody interface of these constructions.


Author(s):  
Luodi Yu ◽  
Jiajing Zeng ◽  
Suiping Wang ◽  
Yang Zhang

Purpose This study aimed to examine whether abstract knowledge of word-level linguistic prosody is independent of or integrated with phonetic knowledge. Method Event-related potential (ERP) responses were measured from 18 adult listeners while they listened to native and nonnative word-level prosody in speech and in nonspeech. The prosodic phonology (speech) conditions included disyllabic pseudowords spoken in Chinese and in English matched for syllabic structure, duration, and intensity. The prosodic acoustic (nonspeech) conditions were hummed versions of the speech stimuli, which eliminated the phonetic content while preserving the acoustic prosodic features. Results We observed language-specific effects on the ERP that native stimuli elicited larger late negative response (LNR) amplitude than nonnative stimuli in the prosodic phonology conditions. However, no such effect was observed in the phoneme-free prosodic acoustic control conditions. Conclusions The results support the integration view that word-level linguistic prosody likely relies on the phonetic content where the acoustic cues embedded in. It remains to be examined whether the LNR may serve as a neural signature for language-specific processing of prosodic phonology beyond auditory processing of the critical acoustic cues at the suprasyllabic level.


Author(s):  
Laurence Bruggeman ◽  
Julien Millasseau ◽  
Ivan Yuen ◽  
Katherine Demuth

Purpose Children with hearing loss (HL), including those with hearing aids (HAs) and cochlear implants (CIs), often have difficulties contrasting words like “ b each ” versus “ p each ” and “ do g ” versus “ do ck ” due to challenges producing systematic voicing contrasts. Even when acoustic contrasts are present, these may not be perceived as such by others. This can cause miscommunication, leading to poor self-esteem and social isolation. Acoustic evidence is therefore needed to determine if these children have established distinct voicing categories before entering school and if misperceptions are due to a lack of phonological representations or due to a still-maturing implementation system. The findings should help inform more effective early intervention. Method Participants included 14 children with HL (eight HA users, five CI users, and one bimodal) and 20 with normal hearing, all English-speaking preschoolers. In an elicited imitation task, they produced consonant–vowel–consonant minimal pair words that contrasted voicing in word-initial (onset) or word-final (coda) position at all three places of articulation (PoAs). Results Overall, children with HL showed acoustically distinct voicing categories for both onsets and codas at all three PoAs. Contrasts were less systematic for codas than for onsets, as also confirmed by adults' perceptual ratings. Conclusions Preschoolers with HL produce acoustic differences for voiced versus voiceless onsets and codas, indicating distinct phonological representations for both. Nonetheless, codas were less accurately perceived by adult raters, especially when produced by CI users. This suggests a protracted development of the phonetic implementation of codas, where CI users, in particular, may benefit from targeted intervention.


2021 ◽  
pp. 002383092110494 ◽  
Author(s):  
William Choi

Musical experience facilitates speech perception. French musicians, to whom stress is foreign, have been found to perceive English stress more accurately than French non-musicians. This study investigated whether this musical advantage also applies to native listeners. English musicians and non-musicians completed an English stress discrimination task and two control tasks. With age, non-verbal intelligence and short-term memory controlled, the musicians exhibited a perceptual advantage relative to the non-musicians. This perceptual advantage was equally potent to both trochaic and iambic stress patterns. In terms of perceptual strategy, the two groups showed differential use of acoustic cues for iambic but not trochaic stress. Collectively, the results could be taken to suggest that musical experience enhances stress discrimination even among native listeners. Remarkably, this musical advantage is highly consistent and does not particularly favour either stress pattern. For iambic stress, the musical advantage appears to stem from the differential use of acoustic cues by musicians. For trochaic stress, the musical advantage may be rooted in enhanced durational sensitivity.


2021 ◽  
Vol 150 (4) ◽  
pp. A70-A71
Author(s):  
Kevin Lilley ◽  
Rebecca Scarborough ◽  
Georgia Zellou
Keyword(s):  

2021 ◽  
Vol 150 (4) ◽  
pp. A71-A71
Author(s):  
Ian Chan ◽  
Alec DeCaprio ◽  
Javier Arango ◽  
Luca De Nardis ◽  
Stefanie Shattuck-Hufnagel ◽  
...  
Keyword(s):  

2021 ◽  
Vol 150 (4) ◽  
pp. A356-A356
Author(s):  
Sagnik Anupam ◽  
Jeung-Yoon Choi ◽  
Stefanie Shattuck-Hufnagel

Sign in / Sign up

Export Citation Format

Share Document