Emotion detection with hybrid voice quality and prosodic features using Neural Network

Author(s):  
Inshirah Idris ◽  
Md Sah Hj Salam
Author(s):  
Chunyan Ji ◽  
Thosini Bamunu Mudiyanselage ◽  
Yutong Gao ◽  
Yi Pan

AbstractThis paper reviews recent research works in infant cry signal analysis and classification tasks. A broad range of literatures are reviewed mainly from the aspects of data acquisition, cross domain signal processing techniques, and machine learning classification methods. We introduce pre-processing approaches and describe a diversity of features such as MFCC, spectrogram, and fundamental frequency, etc. Both acoustic features and prosodic features extracted from different domains can discriminate frame-based signals from one another and can be used to train machine learning classifiers. Together with traditional machine learning classifiers such as KNN, SVM, and GMM, newly developed neural network architectures such as CNN and RNN are applied in infant cry research. We present some significant experimental results on pathological cry identification, cry reason classification, and cry sound detection with some typical databases. This survey systematically studies the previous research in all relevant areas of infant cry and provides an insight on the current cutting-edge works in infant cry signal analysis and classification. We also propose future research directions in data processing, feature extraction, and neural network classification fields to better understand, interpret, and process infant cry signals.


Facial emotion analysis is the basic idea to train the system to understand the different facial expressions of human beings. The Facial expressions are recorded by the use of camera which is attached to user device. Additionally this project will be helpful for the online marketing of the products as it can detect the facial expressions and sentiment of the person. It is the study of people sentiment, opinions and emotions. Sentiment analysis is the method by which information is taken from the facial expressions of people in regard to different situations. The main aim is to read the facial expressions of the human beings using a good resolution camera so that the machine can identify the human sentiments. Convolutional neural network is used as an existing system which is unsupervised neural network to replace that with a supervised mechanism which is called supervised neural network. It can be used in gaming sector, unlock smart phones, automated facial language translation etc.


2021 ◽  
Author(s):  
Shi Feng ◽  
Jia Wei ◽  
Daling Wang ◽  
Xiaocui Yang ◽  
Zhenfei Yang ◽  
...  

2016 ◽  
Vol 59 (2) ◽  
pp. 216-229 ◽  
Author(s):  
Theresa Schölderle ◽  
Anja Staiger ◽  
Renée Lampe ◽  
Katrin Strecker ◽  
Wolfram Ziegler

Purpose Although dysarthria affects the large majority of individuals with cerebral palsy (CP) and can substantially complicate everyday communication, previous research has provided an incomplete picture of its clinical features. We aimed to comprehensively describe characteristics of dysarthria in adults with CP and to elucidate the impact of dysarthric symptoms on parameters relevant for communication. Method Forty-two adults with CP underwent speech assessment by means of standardized auditory rating scales. Listening experiments were conducted to obtain communication-related parameters—that is, intelligibility and naturalness—as well as age and gender estimates. Results The majority of adults with CP showed moderate to severe dysarthria with symptoms on all dimensions of speech, most prominently voice quality, respiration, and prosody. Regression analyses revealed that articulatory, respiratory, and prosodic features were the strongest predictors of intelligibility and naturalness of speech. Listeners' estimates of the speakers' age and gender were predominantly determined by voice parameters. Conclusion This study provides an overview on the clinical presentation of dysarthria in a convenience sample of adults with CP. The complexity of the functional impairment described and the consequences on the individuals' communication call for a stronger consideration of dysarthria in CP both in clinical care and in research.


In this paper we will identify a cry signals of infants and the explanation behind the screams below 0-6 months of segment age. Detection of baby cry signals is essential for the pre-processing of various applications involving crial analysis for baby caregivers, such as emotion detection. Since cry signals hold baby well-being information and can be understood to an extent by experienced parents and experts. We train and validate the neural network architecture for baby cry detection and also test the fastAI with the neural network. Trained neural networks will provide a model and this model can predict the reason behind the cry sound. Only the cry sounds are recognized, and alert the user automatically. Created a web application by responding and detecting different emotions including hunger, tired, discomfort, bellypain.


Sign in / Sign up

Export Citation Format

Share Document