speech rate
Recently Published Documents


TOTAL DOCUMENTS

867
(FIVE YEARS 234)

H-INDEX

46
(FIVE YEARS 6)

2022 ◽  
pp. 002383092110648
Author(s):  
Malte Belz ◽  
Oksana Rasskazova ◽  
Jelena Krivokapić ◽  
Christine Mooshammer

Phrase-final lengthening affects the segments preceding a prosodic boundary. This prosodic variation is generally assumed to be independent of the phonemic identity. We refer to this as the ‘uniform lengthening hypothesis’ (ULH). However, in German, lax vowels do not undergo lengthening for word stress or shortening for increased speech rate, indicating that temporal properties might interact with phonemic identity. We test the ULH by comparing the effect of the boundary on acoustic and kinematic measures for tense and lax vowels and several coda consonants. We further examine if the boundary effect decreases with distance from the boundary. Ten native speakers of German were recorded by means of electromagnetic articulography (EMA) while reading sentences that contained six minimal pairs varying in vowel tenseness and boundary type. In line with the ULH, the results show that the acoustic durations of lax vowels are lengthened phrase-finally, similarly to tense vowels. We find that acoustic lengthening is stronger the closer the segments are to the boundary. Articulatory parameters of the closing movements toward the post-vocalic consonants are affected by both phrasal position and identity of the preceding vowel. The results are discussed with regard to the interaction between prosodic structure and vowel tenseness.


Author(s):  
Chieh Kao ◽  
Maria D. Sera ◽  
Yang Zhang

Purpose: The aim of this study was to investigate infants' listening preference for emotional prosodies in spoken words and identify their acoustic correlates. Method: Forty-six 3- to-12-month-old infants ( M age = 7.6 months) completed a central fixation (or look-to-listen) paradigm in which four emotional prosodies (happy, sad, angry, and neutral) were presented. Infants' looking time to the string of words was recorded as a proxy of their listening attention. Five acoustic variables—mean fundamental frequency (F0), word duration, intensity variation, harmonics-to-noise ratio (HNR), and spectral centroid—were also analyzed to account for infants' attentiveness to each emotion. Results: Infants generally preferred affective over neutral prosody, with more listening attention to the happy and sad voices. Happy sounds with breathy voice quality (low HNR) and less brightness (low spectral centroid) maintained infants' attention more. Sad speech with shorter word duration (i.e., faster speech rate), less breathiness, and more brightness gained infants' attention more than happy speech did. Infants listened less to angry than to happy and sad prosodies, and none of the acoustic variables were associated with infants' listening interests in angry voices. Neutral words with a lower F0 attracted infants' attention more than those with a higher F0. Neither age nor sex effects were observed. Conclusions: This study provides evidence for infants' sensitivity to the prosodic patterns for the basic emotion categories in spoken words and how the acoustic properties of emotional speech may guide their attention. The results point to the need to study the interplay between early socioaffective and language development.


2021 ◽  
Vol 15 (3) ◽  
pp. 141-150
Author(s):  
Zahra Sadat Ghoreishi ◽  
◽  
Mojtaba Azimian ◽  
Javad Alaghband Rad ◽  
Anahita Khorrami Banaraki ◽  
...  

Background and Objectives: There are several tests to determine the type and severity of aphasia, but they take a long time to administer when assessing aphasic patients. In recent years, the analysis of spontaneous speech has gained great attention because it is important to diagnose and follow post-treatment improvement in aphasic patients. This study was done to assess some parameters of connected speech in aphasic patients. In addition, the correlation between connected speech parameters and the type and severity of aphasia was measured. Methods: We compared the connected speech parameters of 27 aphasics (10 fluent and 17 non-fluent), Persian speakers, compared with the control group. There were two groups matched by age, sex, and education. Nest’s bird story pictures were used to elicit a speech sample. In the next step, the connected speech was analyzed to define speech parameters, including speech rate, mean length of utterance, number of utterances, total words, content words, function words, nouns, and verbs. Moreover, the severity of aphasia was measured using a Persian Western Aphasia Battery (WAB). Results: The findings showed significant differences between groups in all parameters of the connected speech (P<0.01). The correlation coefficient between speech parameters and severity of aphasia demonstrated that all parameters were highly correlated (r >0.71) with the severity of aphasia (P<0.01) except for speech rate and the number of function words. There were some typical differences between linguistics grammatical and pragmatical characters of different types of aphasia. Conclusion: Connected speech is one of the most sensitive parts of language in all types of aphasic patients. There are some clinical signs for differential diagnosis of aphasia based on speech measures. According to the findings, the type and severity of aphasia and connected speech were highly correlated. Thus, the use of the connected speech analysis is necessary as an assessment tool for the diagnosis of aphasia.


2021 ◽  
pp. 1-9
Author(s):  
Jess C.S. Chan ◽  
Julie C. Stout ◽  
Christopher A. Shirbin ◽  
Adam P. Vogel

Background: Subtle progressive changes in speech motor function and cognition begin prior to diagnosis of Huntington’s disease (HD). Objective: To determine the nature of listener-rated speech differences in premanifest and early-stage HD (i.e., PreHD and EarlyHD), compared to neurologically healthy controls. Methods: We administered a speech battery to 60 adults (16 people with PreHD, 14 with EarlyHD, and 30 neurologically healthy controls), and conducted a cognitive test of processing speed/visual attention, the Symbol Digit Modalities Test (SDMT) on participants with HD. Voice recordings were rated by expert listeners and analyzed for acoustic and perceptual speech features. Results: Listeners perceived subtle differences in the speech of PreHD compared to controls, including abnormal pitch level and speech rate, reduced loudness and loudness inflection, altered voice quality, hypernasality, imprecise articulation, and reduced naturalness of speech. Listeners detected abnormal speech rate in PreHD compared to healthy speakers on a reading task, which correlated with slower speech rate from acoustic analysis and a lower cognitive performance score. In early-stage HD, continuous speech was characterized by longer pauses, a higher proportion of silence, and slower rate. Conclusion: Differences in speech and voice acoustic features are detectable in PreHD by expert listeners and align with some acoustically-derived objective speech measures. Slower speech rate in PreHD suggests altered oral motor control and/or subtle cognitive deficits that begin prior to diagnosis. Speakers with EarlyHD exhibited more silences compared to the PreHD and control groups, raising the likelihood of a link between speech and cognition that is not yet well characterized in HD.


Loquens ◽  
2021 ◽  
Vol 7 (2) ◽  
pp. e075
Author(s):  
Enrique Santamaría Busto

The aim of this study is to identify which linguistic variables are most strongly related to comprehensibility, accentedness and fluency in Spanish as L2, and how the results may also vary according to the type of listener (expert or non-expert) and to the speakers’ level of ability in these three dimensions. To this end, 40 native English speakers of Spanish orally described picture narratives that were subsequently rated by four groups of raters using 9-point Likert scales. The first two groups were composed of 109 novice and 42 expert raters assessing comprehensibility, accentedness and fluency. The last two groups consisted of 35 phoneticians and 35 linguistic raters responsible for analyzing and scoring a total of 14 speech measures (7 per group) targeting pronunciation, disfluencies, lexis, grammar and discourse. The results reveal that comprehensibility is associated with a wide range of elements, while accentedness is determined especially by segmentals, and fluency by speech rate. Additionally, the impact of linguistic parameters varies according to the speakers’ level of performance, and there are differences between novice and expert raters, with the former being more affected by pronunciation and the experts by lexis and grammar.


Author(s):  
Stephanie A. Borrie ◽  
Camille J. Wynn ◽  
Visar Berisha ◽  
Tyson S. Barrett

Purpose: We proposed and tested a causal instantiation of the World Health Organization's International Classification of Functioning, Disability and Health (ICF) framework, linking acoustics, intelligibility, and communicative participation in the context of dysarthria. Method: Speech samples and communicative participation scores were collected from individuals with dysarthria ( n = 32). Speech was analyzed for two acoustic metrics (i.e., articulatory precision and speech rate), and an objective measure of intelligibility was generated from listener transcripts. Mediation analysis was used to evaluate pathways of effect between acoustics, intelligibility, and communicative participation. Results: We observed a strong relationship between articulatory precision and intelligibility and a moderate relationship between intelligibility and communicative participation. Collectively, data supported a significant relationship between articulatory precision and communicative participation, which was almost entirely mediated through intelligibility. These relationships were not significant when speech rate was specified as the acoustic variable of interest. Conclusion: The statistical corroboration of our causal instantiation of the ICF framework with articulatory acoustics affords important support toward the development of a comprehensive causal framework to understand and, ultimately, address restricted communicative participation in dysarthria.


Author(s):  
Mathilde Hutin ◽  
Adèle Jatteau ◽  
Ioana Vasilescu ◽  
Lori Lamel ◽  
Yaru Wu ◽  
...  

What is commonly considered as an epenthetic vowel can actually refer to at least two different realities: phonological epenthesis or phonetic excrescence. French schwa, noted [ә], is a vowel alternating with zero and limited to unstressed syllables that can appear word-internally or word-finally. This paper presents an extensive description of the distribution of word-final schwa in Standard French in order to shed light on its nature: is it an intrusive vowel or a full epenthetic vowel? To that extent, three large corpora of French containing more than 110 hours of speech were used to establish the presence of word-final schwa as a function of sociolinguistics, orthography, phonotactics and phonetics. Our conclusions are that word-final schwa is impacted by speech style, gender, orthography, phonotactics (i.e., the number of adjacent consonants and their sonority profile), and the phonological properties of the codas. However, speech rate does not impact word-final schwa realization. The specific results lead us to suggest that word-final schwa in Standard French shares similarities with intrusive vowels but ultimately behaves like a legit epenthetic vowel.


2021 ◽  
pp. 009365022110593
Author(s):  
Emma Rodero ◽  
Lucía Cores-Sarría

Studies in different languages have identified a broadcast speaking style, a particular manner that broadcasters have of reading news. This speaking style is characterized by an emphatic intonation with a fast speech rate easily recognizable by listeners. Some authors have stated that messages in this style are not positively perceived by listeners, as it is repetitive and regular, but there is no empirical data to support this conclusion, nor has the style been analyzed with physiological measures. The physiological approach has some advantages, such as a more objective assessment and real-time evaluation. Therefore, this study aims to analyze the effectiveness, adequacy, and physiological response of this broadcast style compared to a narrative pattern. We combined self-report with physiological measures. Fifty-six participants listened to six news pieces in both styles and with two voices, male and female. They had to rate the effectiveness and adequacy of the news while we measured their physiological responses (heart rate and electrodermal activity). The results showed that news conveyed through the broadcast style elicited less cognitive resource allocation and emotional arousal than the narrative pattern, but there were no significant differences in self-report evaluations.


Author(s):  
Sanne Koops ◽  
Sanne G. Brederoo ◽  
Janna N. de Boer ◽  
Femke G. Nadema ◽  
Alban E. Voppel ◽  
...  

Background: Depression is a debilitating disorder that at present lacks a reliable biomarker to aid in diagnosis and early detection. Recent advances in computational analytic approaches have opened up new avenues in developing such a biomarker by taking advantage of the wealth of information that can be extracted from a person’s speech. Objective: The current review provides an overview of the latest findings in the rapidly evolving field of computational language analysis for the detection of depression. We cover a wide range of both acoustic and content-related linguistic features, data types (i.e., spoken and written language), and data sources (i.e., lab settings, social media, and smartphone-based). We put special focus on the current methodological advances with regard to feature extraction and computational modeling techniques. Furthermore, we pay attention to potential hurdles in the implementation of automatic speech analysis. Conclusion: Depressive speech is characterized by several anomalies, such as lower speech rate, less pitch variability and more self-referential speech. With current computational modeling techniques, such features can be used to detect depression with an accuracy of up to 91%. The performance of the models is optimized when machine learning techniques are implemented that suit the type and amount of data. Recent studies now work towards further optimization and generalizability of the computational language models to detect depression. Finally, privacy and ethical issues are of paramount importance to be addressed when automatic speech analysis techniques are further implemented in, for example, smartphones. Altogether, computational speech analysis is well underway towards becoming an effective diagnostic aid for depression.


Author(s):  
Katharina Lehner ◽  
Wolfram Ziegler ◽  

Purpose: Despite extensive research into communication-related parameters in dysarthria, such as intelligibility, naturalness, and perceived listener effort, the existing evidence has not been translated into a clinically applicable, comprehensive, and valid diagnostic tool so far. This study addresses Communication-Related Parameters in Speech Disorders (KommPaS), a new web-based diagnostic instrument for measuring indices of communication limitation in individuals with dysarthria through online crowdsourcing. More specifically, it answers questions about the construct validity of KommPaS. In the first part, the interrelationship of the KommPaS variables intelligibility, naturalness, perceived listener effort, and speech rate were explored in order to draw a comprehensive picture of a patient's limitations and avoid the collection of redundant information. Second, the influences of motor speech symptoms on the KommPaS variables were studied in order to delineate the structural relationships between two complementary diagnostic perspectives. Method: One hundred persons with dysarthria of different etiologies and varying degrees of severity were examined with KommPaS to obtain layperson-based data on communication-level parameters, and with the Bogenhausen Dysarthria Scale (BoDyS) to obtain expert-based, function-level data on dysarthria symptoms. The internal structure of the KommPaS variables and their dependence on the BoDyS variables were analyzed using structural equation modeling. Results: Despite a high multicollinearity, all KommPaS variables were shown to provide complementary diagnostic information and their mutual interconnections were delineated in a path graph model. Regarding the influence of the BoDyS scales on the KommPaS variables, separate linear regression models revealed plausible predictor sets. A complete path model of KommPaS and BoDyS variables was developed to map the complex interplay between variables at the functional and the communication levels of dysarthria assessment. Conclusion: In validating a new clinical tool for the diagnostics of communication limitations in dysarthria, this study is the first to draw a comprehensive picture of how auditory-perceptual characteristics of dysarthria interact at the levels of expert-based functional and layperson-based communicative assessments.


Sign in / Sign up

Export Citation Format

Share Document