Real-time Sign Language Recognition with Guided Deep Convolutional Neural Networks

Author(s):  
Zhengzhe Liu ◽  
Fuyang Huang ◽  
Gladys Wai Lan Tang ◽  
Felix Yim Binh Sze ◽  
Jing Qin ◽  
...  
Author(s):  
F. M. Javed Mehedi Shamrat ◽  
Sovon Chakraborty ◽  
Md. Masum Billah ◽  
Moumita Kabir ◽  
Nazmus Shakib Shadin ◽  
...  

<p>The amount of deaf and mute individuals on the earth is rising at an alarmingrate. Bangladesh has about 2.6 million people who are unable to interact with the community using language. Hearing-impaired citizens in Bangladesh use Bangladeshi sign language (BSL) as a means of communication. In this article,we propose a new method for Bengali sign language recognition based on deep convolutional neural networks. Our framework employs convolutional neural networks (CNN) to learn from the images in our dataset and interpret hand signs from input images. Checking their collections of ten indications (we usedten sets of images with 31 distinct signs) for a total of 310 images. The proposed system takes snap shots from a video by using a webcam with applying a computer vision-based approach. After that, it compares those photos to a previously trained dataset generated with CNN and displays the Bengali numbers (০-৯). After estimating the model on our dataset, weobtained an overall accuracy of 99.8%. We want to streng then things as far aswe can to make silent contact with the majority of society as simple asprobable.</p>


2021 ◽  
Vol 5 (2 (113)) ◽  
pp. 44-54
Author(s):  
Chingiz Kenshimov ◽  
Samat Mukhanov ◽  
Timur Merembayev ◽  
Didar Yedilkhan

For people with disabilities, sign language is the most important means of communication. Therefore, more and more authors of various papers and scientists around the world are proposing solutions to use intelligent hand gesture recognition systems. Such a system is aimed not only for those who wish to understand a sign language, but also speak using gesture recognition software. In this paper, a new benchmark dataset for Kazakh fingerspelling, able to train deep neural networks, is introduced. The dataset contains more than 10122 gesture samples for 42 alphabets. The alphabet has its own peculiarities as some characters are shown in motion, which may influence sign recognition. Research and analysis of convolutional neural networks, comparison, testing, results and analysis of LeNet, AlexNet, ResNet and EffectiveNet – EfficientNetB7 methods are described in the paper. EffectiveNet architecture is state-of-the-art (SOTA) and is supposed to be a new one compared to other architectures under consideration. On this dataset, we showed that the LeNet and EffectiveNet networks outperform other competing algorithms. Moreover, EffectiveNet can achieve state-of-the-art performance on nother hand gesture datasets. The architecture and operation principle of these algorithms reflect the effectiveness of their application in sign language recognition. The evaluation of the CNN model score is conducted by using the accuracy and penalty matrix. During training epochs, LeNet and EffectiveNet showed better results: accuracy and loss function had similar and close trends. The results of EffectiveNet were explained by the tools of the SHapley Additive exPlanations (SHAP) framework. SHAP explored the model to detect complex relationships between features in the images. Focusing on the SHAP tool may help to further improve the accuracy of the model


Sign in / Sign up

Export Citation Format

Share Document