scholarly journals GMM FOR EMOTION RECOGNITION OF VIETNAMESE

2018 ◽  
Vol 33 (3) ◽  
pp. 229-246
Author(s):  
Đào Thị Lệ Thủy ◽  
Trinh Van Loan ◽  
Nguyen Hong Quang

This paper presents the results of GMM-based recognition for four basic emotions of Vietnamese such as neutral, sadness, anger and happiness. The characteristic parameters of these emotions are extracted from speech signals and divided into different parameter sets for experiments. The experiments are carried out according to speaker-dependent or speaker-independent and content-dependent or content-independent recognitions. The results showed that the recognition scores are rather high with the case for which there is a full combination of parameters as MFCC and its first and second derivatives, fundamental frequency, energy, formants and its correspondent bandwidths, spectral characteristics and F0 variants. In average, the speaker-dependent and content-dependent recognition scrore is 89.21%. Next, the average score is 82.27% for the speaker-dependent and content-independent recognition. For the speaker-independent and content-dependent recognition, the average score is 70.35%. The average score is 66.99% for speaker-independent and content-independent recognition. Information on F0 has significantly increased the score of recognition

1990 ◽  
Vol 112 (1) ◽  
pp. 83-87 ◽  
Author(s):  
R. H. Fries ◽  
B. M. Coffey

Solution of rail vehicle dynamics models by means of numerical simulation has become more prevalent and more sophisticated in recent years. At the same time, analysts and designers are increasingly interested in the response of vehicles to random rail irregularities. The work described in this paper provides a convenient method to generate random vertical and crosslevel irregularities when their time histories are required as inputs to a numerical simulation. The solution begins with mathematical models of vertical and crosslevel power spectral densities (PSDs) representing PSDs of track classes 4, 5, and 6. The method implements state-space models of shape filters whose frequency response magnitude squared matches the desired PSDs. The shape filters give time histories possessing the proper spectral content when driven by white noise inputs. The state equations are solved directly under the assumption that the white noise inputs are constant between time steps. Thus, the state transition matrix and the forcing matrix are obtained in closed form. Some simulations require not only vertical and crosslevel alignments, but also the first and occasionally the second derivatives of these signals. To accommodate these requirements, the first and second derivatives of the signals are also generated. The responses of the random vertical and crosslevel generators depend upon vehicle speed, sample interval, and track class. They possess the desired PSDs over wide ranges of speed and sample interval. The paper includes a comparison between synthetic and measured spectral characteristics of class 4 track. The agreement is very good.


2021 ◽  
Author(s):  
Talieh Seyed Tabtabae

Automatic Emotion Recognition (AER) is an emerging research area in the Human-Computer Interaction (HCI) field. As Computers are becoming more and more popular every day, the study of interaction between humans (users) and computers is catching more attention. In order to have a more natural and friendly interface between humans and computers, it would be beneficial to give computers the ability to recognize situations the same way a human does. Equipped with an emotion recognition system, computers will be able to recognize their users' emotional state and show the appropriate reaction to that. In today's HCI systems, machines can recognize the speaker and also content of the speech, using speech recognition and speaker identification techniques. If machines are equipped with emotion recognition techniques, they can also know "how it is said" to react more appropriately, and make the interaction more natural. One of the most important human communication channels is the auditory channel which carries speech and vocal intonation. In fact people can perceive each other's emotional state by the way they talk. Therefore in this work the speech signals are analyzed in order to set up an automatic system which recognizes the human emotional state. Six discrete emotional states have been considered and categorized in this research: anger, happiness, fear, surprise, sadness, and disgust. A set of novel spectral features are proposed in this contribution. Two approaches are applied and the results are compared. In the first approach, all the acoustic features are extracted from consequent frames along the speech signals. The statistical values of features are considered to constitute the features vectors. Suport Vector Machine (SVM), which is a relatively new approach in the field of machine learning is used to classify the emotional states. In the second approach, spectral features are extracted from non-overlapping logarithmically-spaced frequency sub-bands. In order to make use of all the extracted information, sequence discriminant SVMs are adopted. The empirical results show that the employed techniques are very promising.


Revista CEFAC ◽  
2019 ◽  
Vol 21 (6) ◽  
Author(s):  
Flávia Viegas ◽  
Danieli Viegas ◽  
Glaucio Serra Guimarães ◽  
Margareth Maria Gomes de Souza ◽  
Ronir Raggio Luiz ◽  
...  

ABSTRACT Purpose: to compare the measurements of fundamental frequency (F0) and frequency of the first two formants (F1 and F2) of the seven oral vowels of the Brazilian Portuguese in two speech tasks, in adults without voice and speech disorders. Methods: eighty participants in the age range 18 and 40 years, paired by gender, were selected after orofacial, orthodontic and auditory-perceptual assessments of voice and speech. The speech signals were obtained from carrier phrases and sustained vowels and the values of the F0 and frequencies of F1 and F2 were estimated. The differences were verified through the t Test, and the effect size was calculated. Results: differences were found in the F0 measurements between the two speech tasks, in two vowels in males, and in five vowels, in females. In the F1 frequencies, differences were noted in six vowels, in men, and in two, in women. In the F2 frequencies, there was a difference in four vowels, in men, and three, in women. Conclusion: based on the differences found, it is concluded that the speech task for evaluation of fundamental frequency and formants’ frequencies, in the Brazilian Portuguese, can show distinct results in both glottal and supraglottal measures in the production of different oral vowels of this language. Thus, it is suggested that clinicians and researchers consider both forms of emission for a more accurate interpretation of the implications of these data in the evaluation of oral communication and therapeutic conducts.


2015 ◽  
Vol 14 ◽  
pp. 57-76
Author(s):  
Hasrul Mohd Nazid ◽  
Hariharan Muthusamy ◽  
Vikneswaran Vijean ◽  
Sazali Yaacob

2017 ◽  
Vol 41 (S1) ◽  
pp. S190-S190
Author(s):  
V.P. Bozikas ◽  
S. Tsotsi ◽  
A. Dardagani ◽  
E. Dandi ◽  
E.I. Nazlidou ◽  
...  

Deficits in emotion perception in patients with first episode of psychosis have been reported by many researchers. Till now, training programs have focused mainly in patients with schizophrenia and not in first psychotic episode (FEP) patients. We used a new intervention for facial affect recognition in a group of 35 FEP patients (26 male). The emotion recognition intervention included coloured pictures of individuals expressing six basic emotions (happiness, sadness, anger, disgust, surprise, fear) and a neutral emotion. The patients were trained to detect changes in facial features, according to the emotion displayed. A comprehensive battery of neuropsychological tests was also administered, measuring attention, memory, working memory, visuospatial ability and executive function by using specific tests of the Cambridge Neuropsychological Test Automated Battery (CANTAB). We tried to explore whether cognitive performance can explain the difference noted between the original assessment of emotion recognition and the post-intervention assessment. According to our data, overall cognitive performance did not correlate with post-intervention change in emotion recognition. Specific cognitive domains did not correlate with this change, either. According the above mentioned results, no significant correlation between neuropsychological performance and post-intervention improvement in emotion recognition was noted. This finding may suggest that interventions for emotion recognition may target specific processes that underlie emotion perception and their effect can be independent of general cognitive function.Disclosure of interestThe authors have not supplied their declaration of competing interest.


Author(s):  
Riadh Ouerchefani ◽  
Naoufel Ouerchefani ◽  
Mohamed Riadh Ben Rejeb ◽  
Didier Le Gall

Abstract Objective Patients with prefrontal cortex damage often transgress social rules and show lower accuracy in identifying and explaining inappropriate social behavior. The objective of this study was to examine the relationship between the ability to perceive other unintentional transgressions of social norms and both decision making and emotion recognition as these abilities are critical for appropriate social behavior. Method We examined a group of patients with focal prefrontal cortex damage (N = 28) and a group of matched control participants (N = 28) for their abilities to detect unintentional transgression of social norms using the “Faux-Pas” task of theory of mind, to make advantageous decisions on the Iowa gambling task, and to recognize basic emotions on the Ekman facial affect test. Results The group of patients with frontal lobe damage was impaired in all of these tasks compared with control participants. Moreover, all the “Faux-Pas”, Iowa gambling, and emotion recognition tasks were significantly associated and predicted by executive measures of inhibition, flexibility, or planning. However, only measures from the Iowa gambling task were associated and predicted performance on the “Faux-Pas” task. These tasks were not associated with performance in recognition of basic emotions. These findings suggest that theory of mind, executive functions, and decision-making abilities act in an interdependent way for appropriate social behavior. However, theory of mind and emotion recognition seem to have distinct but additive effects upon social behavior. Results from VLSM analysis also corroborate these data by showing a partially overlapped prefrontal circuitry underlying these cognitive domains.


Sign in / Sign up

Export Citation Format

Share Document