scholarly journals Real Time Facial Expression Recognition System Based on Deep Learning

2019 ◽  
Vol 8 (2S11) ◽  
pp. 4047-4051

The automatic detection of facial expressions is an active research topic, since its wide fields of applications in human-computer interaction, games, security or education. However, the latest studies have been made in controlled laboratory environments, which is not according to real world scenarios. For that reason, a real time Facial Expression Recognition System (FERS) is proposed in this paper, in which a deep learning approach is applied to enhance the detection of six basic emotions: happiness, sadness, anger, disgust, fear and surprise in a real-time video streaming. This system is composed of three main components: face detection, face preparation and face expression classification. The results of proposed FERS achieve a 65% of accuracy, trained over 35558 face images..

2008 ◽  
Vol 381-382 ◽  
pp. 375-378
Author(s):  
K.T. Song ◽  
M.J. Han ◽  
F.Y. Chang ◽  
S.H. Chang

The capability of recognizing human facial expression plays an important role in advanced human-robot interaction development. Through recognizing facial expressions, a robot can interact with a user in a more natural and friendly manner. In this paper, we proposed a facial expression recognition system based on an embedded image processing platform to classify different facial expressions on-line in real time. A low-cost embedded vision system has been designed and realized for robotic applications using a CMOS image sensor and digital signal processor (DSP). The current design acquires thirty 640x480 image frames per second (30 fps). The proposed emotion recognition algorithm has been successfully implemented on the real-time vision system. Experimental results on a pet robot show that the robot can interact with a person in a responding manner. The developed image processing platform is effective for accelerating the recognition speed to 25 recognitions per second with an average on-line recognition rate of 74.4% for five facial expressions.


2013 ◽  
pp. 1434-1460
Author(s):  
Ong Chin Ann ◽  
Marlene Valerie Lu ◽  
Lau Bee Theng

The main purpose of this research is to enhance the communication of the disabled community. The authors of this chapter propose an enhanced interpersonal-human interaction for people with special needs, especially those with physical and communication disabilities. The proposed model comprises of automated real time behaviour monitoring, designed and implemented with the ubiquitous and affordable concept in mind to suit the underprivileged. In this chapter, the authors present the prototype which encapsulates an automated facial expression recognition system for monitoring the disabled, equipped with a feature to send Short Messaging System (SMS) for notification purposes. The authors adapted the Viola-Jones face detection algorithm at the face detection stage and implemented template matching technique for the expression classification and recognition stage. They tested their model with a few users and achieved satisfactory results. The enhanced real time behaviour monitoring system is an assistive tool to improve the quality of life for the disabled by assisting them anytime and anywhere when needed. They can do their own tasks more independently without constantly being monitored physically or accompanied by their care takers, teachers, or even parents. The rest of this chapter is organized as follows. The background of the facial expression recognition system is reviewed in Section 2. Section 3 is the description and explanations of the conceptual model of facial expression recognition. Evaluation of the proposed system is in Section 4. Results and findings on the testing are laid out in Section 5, and the final section concludes the chapter.


2017 ◽  
Vol 63 ◽  
pp. 114-125 ◽  
Author(s):  
Md. Zia Uddin ◽  
Mohammed Mehedi Hassan ◽  
Ahmad Almogren ◽  
Mansour Zuair ◽  
Giancarlo Fortino ◽  
...  

Sign in / Sign up

Export Citation Format

Share Document