Deep Learning Model for Face Recognition in Unconstrained Environment

2019 ◽  
Vol 16 (10) ◽  
pp. 4309-4312
Author(s):  
Rajeshwar Moghekar ◽  
Sachin Ahuja

Face recognition from videos is challenging problem as the face image captured has variations in terms of pose, Occlusion, blur and resolution. It has many applications including security monitoring and authentication. A subset of Indian Movies Face database (IMFDB) which has collection of face images retrieved from movie/video of actors which vary in terms of blur, pose, noise and illumination is used in our work. Our work focuses on the use of pre-trained deep learning models and applies transfer learning to the features extracted from the CNN layers. Later we compare it Fine tuned model. The results show that the accuracy is 99.89 using CNN as feature extractor and 96.3 when we fine tune the VGG-Face. The Fine tuned network of VGG-Face learnt more generic features when compared with its counterpart transfer learning. When applied on VGG16 transfer learning achieved 93.9.

2021 ◽  
Vol 2021 ◽  
pp. 1-12
Author(s):  
Zhixue Liang

In the contactless delivery scenario, the self-pickup cabinet is an important terminal delivery device, and face recognition is one of the efficient ways to achieve contactless access express delivery. In order to effectively recognize face images under unrestricted environments, an unrestricted face recognition algorithm based on transfer learning is proposed in this study. First, the region extraction network of the faster RCNN algorithm is improved to improve the recognition speed of the algorithm. Then, the first transfer learning is applied between the large ImageNet dataset and the face image dataset under restricted conditions. The second transfer learning is applied between face image under restricted conditions and unrestricted face image datasets. Finally, the unrestricted face image is processed by the image enhancement algorithm to increase its similarity with the restricted face image, so that the second transfer learning can be carried out effectively. Experimental results show that the proposed algorithm has better recognition rate and recognition speed on the CASIA-WebFace dataset, FLW dataset, and MegaFace dataset.


Forests ◽  
2021 ◽  
Vol 12 (2) ◽  
pp. 212
Author(s):  
Mingyu Gao ◽  
Dawei Qi ◽  
Hongbo Mu ◽  
Jianfeng Chen

In recent years, due to the shortage of timber resources, it has become necessary to reduce the excessive consumption of forest resources. Non-destructive testing technology can quickly find wood defects and effectively improve wood utilization. Deep learning has achieved significant results as one of the most commonly used methods in the detection of wood knots. However, compared with convolutional neural networks in other fields, the depth of deep learning models for the detection of wood knots is still very shallow. This is because the number of samples marked in the wood detection is too small, which limits the accuracy of the final prediction of the results. In this paper, ResNet-34 is combined with transfer learning, and a new TL-ResNet34 deep learning model with 35 convolution depths is proposed to detect wood knot defects. Among them, ResNet-34 is used as a feature extractor for wood knot defects. At the same time, a new method TL-ResNet34 is proposed, which combines ResNet-34 with transfer learning. After that, the wood knot defect dataset was applied to TL-ResNet34 for testing. The results show that the detection accuracy of the dataset trained by TL-ResNet34 is significantly higher than that of other methods. This shows that the final prediction accuracy of the detection of wood knot defects can be improved by TL-ResNet34.


Face recognition is used to biometric authentication method to analyze the face extract and photographs useful to reputation formation from them, which can be usually called as a characteristic vector this is used to differentiate the organic features. In this paper to detect the suspect by extracting facial features from the captured image of the suspect from CCTV and match it with the pictures stored in the database and also to achieve an accuracy rate of 100 %, negligible loss using deep learning technique. For extracting the facial features, we are using deep learning model known as Convolutional Neural Network (CNN). It is one of the best models to extract features with the highest accuracy rate .


2019 ◽  
Vol 70 (2) ◽  
pp. 113-121
Author(s):  
Guang Yi Chen ◽  
Tien D. Bui ◽  
Adam Krzyzak

Abstract In this article, we develop a new algorithm for illumination invariant face recognition. We first transform the face images to the logarithm domain, which makes the dark regions brighter. We then use dual-tree complex wavelet transform to generate face images that are approximately invariant to illumination changes and use collaborative representation-based classifier to classify the unknown faces to one known class. We set the approximation sub-band and the highest two DTCWT coefficient sub-bands to zero values before the inverse DTCWT transform is performed. Experimental results demonstrate that our proposed method improves upon a few existing methods under both the noise-free and noisy environments for the Extended Yale Face Database B and the CMU-PIE face database.


Deep learning has attracted several researchers in the field of computer vision due to its ability to perform face and object recognition tasks with high accuracy than the traditional shallow learning systems. The convolutional layers present in the deep learning systems help to successfully capture the distinctive features of the face. For biometric authentication, face recognition (FR) has been preferred due to its passive nature. Processing face images are accompanied by a series of complexities, like variation of pose, light, face expression, and make up. Although all aspects are important, the one that impacts the most face-related computer vision applications is pose. In face recognition, it has been long desired to have a method capable of bringing faces to the same pose, usually a frontal view, in order to ease recognition. Synthesizing different views of a face is still a great challenge, mostly because in nonfrontal face images there are loss of information when one side of the face occludes the other. Most solutions for FR fail to perform well in cases involving extreme pose variations as in such scenarios, the convolutional layers of the deep models are unable to find discriminative parts of the face for extracting information. Most of the architectures proposed earlier deal with the scenarios where the face images used for training as well as testing the deep learning models are frontal and nearfrontal. On the contrary, here a limited number of face images at different poses is used to train the model, where a number of separate generator models learn to map a single face image at any arbitrary pose to specific poses and the discriminator performs the task of face recognition along with discriminating a synthetic face from a realworld sample. To this end, this paper proposes a representation learning by rotating the face. Here an encoderdecoder structure of the generator enables to learn a representation that is both generative and discriminative, which can be used for face image synthesis and pose-invariant face recognition. This representation is explicitly disentangled from other face variations such as pose, through the pose code provided to the decoder and pose estimation in the discriminator.


Author(s):  
Zhongkui Fan ◽  
Ye-Peng Guan

Deep learning has achieved a great success in face recognition (FR), however, little work has been done to apply deep learning for face photo-sketch recognition. This paper proposes an adaptive scale local binary pattern extraction method for optical face features. The extracted features are classified by Gaussian process. The most authoritative optical face test set LFW is used to train the trained model. Test, the test accuracy is 98.7%. Finally, the face features extracted by this method and the face features extracted from the convolutional neural network method are adapted to sketch faces through transfer learning, and the results of the adaptation are compared and analyzed. Finally, the paper tested the open-source sketch face data set CUHK Face Sketch database(CUFS) using the multimedia experiment of the Chinese University of Hong Kong. The test result was 97.4%. The result was compared with the test results of traditional sketch face recognition methods. It was found that the method recognized High efficiency, it is worth promoting.


Face recognition plays a vital role in security purpose. In recent years, the researchers have focused on the pose illumination, face recognition, etc,. The traditional methods of face recognition focus on Open CV’s fisher faces which results in analyzing the face expressions and attributes. Deep learning method used in this proposed system is Convolutional Neural Network (CNN). Proposed work includes the following modules: [1] Face Detection [2] Gender Recognition [3] Age Prediction. Thus the results obtained from this work prove that real time age and gender detection using CNN provides better accuracy results compared to other existing approaches.


2021 ◽  
Vol 10 (3) ◽  
pp. 137
Author(s):  
Youngok Kang ◽  
Nahye Cho ◽  
Jiyoung Yoon ◽  
Soyeon Park ◽  
Jiyeon Kim

Recently, as computer vision and image processing technologies have rapidly advanced in the artificial intelligence (AI) field, deep learning technologies have been applied in the field of urban and regional study through transfer learning. In the tourism field, studies are emerging to analyze the tourists’ urban image by identifying the visual content of photos. However, previous studies have limitations in properly reflecting unique landscape, cultural characteristics, and traditional elements of the region that are prominent in tourism. With the purpose of going beyond these limitations of previous studies, we crawled 168,216 Flickr photos, created 75 scenes and 13 categories as a tourist’ photo classification by analyzing the characteristics of photos posted by tourists and developed a deep learning model by continuously re-training the Inception-v3 model. The final model shows high accuracy of 85.77% for the Top 1 and 95.69% for the Top 5. The final model was applied to the entire dataset to analyze the regions of attraction and the tourists’ urban image in Seoul. We found that tourists feel attracted to Seoul where the modern features such as skyscrapers and uniquely designed architectures and traditional features such as palaces and cultural elements are mixed together in the city. This work demonstrates a tourist photo classification suitable for local characteristics and the process of re-training a deep learning model to effectively classify a large volume of tourists’ photos.


2021 ◽  
pp. 1-11
Author(s):  
Suphawimon Phawinee ◽  
Jing-Fang Cai ◽  
Zhe-Yu Guo ◽  
Hao-Ze Zheng ◽  
Guan-Chen Chen

Internet of Things is considerably increasing the levels of convenience at homes. The smart door lock is an entry product for smart homes. This work used Raspberry Pi, because of its low cost, as the main control board to apply face recognition technology to a door lock. The installation of the control sensing module with the GPIO expansion function of Raspberry Pi also improved the antitheft mechanism of the door lock. For ease of use, a mobile application (hereafter, app) was developed for users to upload their face images for processing. The app sends the images to Firebase and then the program downloads the images and captures the face as a training set. The face detection system was designed on the basis of machine learning and equipped with a Haar built-in OpenCV graphics recognition program. The system used four training methods: convolutional neural network, VGG-16, VGG-19, and ResNet50. After the training process, the program could recognize the user’s face to open the door lock. A prototype was constructed that could control the door lock and the antitheft system and stream real-time images from the camera to the app.


2021 ◽  
Vol 27 ◽  
Author(s):  
Qi Zhou ◽  
Wenjie Zhu ◽  
Fuchen Li ◽  
Mingqing Yuan ◽  
Linfeng Zheng ◽  
...  

Objective: To verify the ability of the deep learning model in identifying five subtypes and normal images in noncontrast enhancement CT of intracranial hemorrhage. Method: A total of 351 patients (39 patients in the normal group, 312 patients in the intracranial hemorrhage group) performed with intracranial hemorrhage noncontrast enhanced CT were selected, with 2768 images in total (514 images for the normal group, 398 images for the epidural hemorrhage group, 501 images for the subdural hemorrhage group, 497 images for the intraventricular hemorrhage group, 415 images for the cerebral parenchymal hemorrhage group, and 443 images for the subarachnoid hemorrhage group). Based on the diagnostic reports of two radiologists with more than 10 years of experience, the ResNet-18 and DenseNet-121 deep learning models were selected. Transfer learning was used. 80% of the data was used for training models, 10% was used for validating model performance against overfitting, and the last 10% was used for the final evaluation of the model. Assessment indicators included accuracy, sensitivity, specificity, and AUC values. Results: The overall accuracy of ResNet-18 and DenseNet-121 models were 89.64% and 82.5%, respectively. The sensitivity and specificity of identifying five subtypes and normal images were above 0.80. The sensitivity of DenseNet-121 model to recognize intraventricular hemorrhage and cerebral parenchymal hemorrhage was lower than 0.80, 0.73, and 0.76 respectively. The AUC values of the two deep learning models were above 0.9. Conclusion: The deep learning model can accurately identify the five subtypes of intracranial hemorrhage and normal images, and it can be used as a new tool for clinical diagnosis in the future.


Sign in / Sign up

Export Citation Format

Share Document