Impostures of Talking Face Systems Using Automatic Face Animation

Author(s):  
Florian Verdet ◽  
Jean Hennebert
Keyword(s):  
Author(s):  
Walid Karam ◽  
Chafic Mokbel ◽  
Hanna Greige ◽  
Guido Aversano ◽  
Catherine Pelachaud ◽  
...  
Keyword(s):  

2021 ◽  
Author(s):  
Sandika Biswas ◽  
Sanjana Sinha ◽  
Dipanjan Das ◽  
Brojeshwar Bhowmick

2021 ◽  
Vol 11 (15) ◽  
pp. 6975
Author(s):  
Tao Zhang ◽  
Lun He ◽  
Xudong Li ◽  
Guoqing Feng

Lipreading aims to recognize sentences being spoken by a talking face. In recent years, the lipreading method has achieved a high level of accuracy on large datasets and made breakthrough progress. However, lipreading is still far from being solved, and existing methods tend to have high error rates on the wild data and have the defects of disappearing training gradient and slow convergence. To overcome these problems, we proposed an efficient end-to-end sentence-level lipreading model, using an encoder based on a 3D convolutional network, ResNet50, Temporal Convolutional Network (TCN), and a CTC objective function as the decoder. More importantly, the proposed architecture incorporates TCN as a feature learner to decode feature. It can partly eliminate the defects of RNN (LSTM, GRU) gradient disappearance and insufficient performance, and this yields notable performance improvement as well as faster convergence. Experiments show that the training and convergence speed are 50% faster than the state-of-the-art method, and improved accuracy by 2.4% on the GRID dataset.


2015 ◽  
Vol 26 (4) ◽  
pp. 490-498 ◽  
Author(s):  
Ferran Pons ◽  
Laura Bosch ◽  
David J. Lewkowicz

Author(s):  
Takacs Gyorgy ◽  
Tihanyi Attila ◽  
Bardi Tamas ◽  
Feldhoffer Gergely
Keyword(s):  

2006 ◽  
Vol 2 (2-3) ◽  
pp. 95-110 ◽  
Author(s):  
Tomislav Kosutic ◽  
Miran Mosmondor ◽  
Ivan Andrisek ◽  
Mario Weber ◽  
Maja Matijasevic ◽  
...  

With evolution in computer and mobile networking technologies comes the challenge of offering novel and complex multimedia applications and end-user services in heterogeneous environments for both developers and service providers. This paper describes one novel service, called LiveMail that explores the potential of existing face animation technologies for innovative and attractive services intended for the mobile market. This prototype service allows mobile subscribers to communicate using personalized 3D face models created from images taken by their phone cameras. The user can take a snapshot of someone's face – a friend, famous person, themselves, even a pet – using the mobile phone's camera. After a quick manipulation on the phone, a 3D model of that face is created and can be animated simply by typing in some text. Speech and appropriate animation of the face are created automatically by speech synthesis. Furthermore, these highly personalized animations can be sent to others as real 3D animated messages or as short videos in MMS. The clients were implemented on different platforms, and different network and face animation techniques, and connected into one complex system. This paper presents the architecture and experience gained in building such a system.


Author(s):  
Sefik Emre Eskimez ◽  
Ross K. Maddox ◽  
Chenliang Xu ◽  
Zhiyao Duan
Keyword(s):  

Sign in / Sign up

Export Citation Format

Share Document