scholarly journals The structure of the local detector of the reprint model of the object in the image

Author(s):  
A. A. Kulikov

Currently, methods for recognizing objects in images work poorly and use intellectually unsatisfactory methods. The existing identification systems and methods do not completely solve the problem of identification, namely, identification in difficult conditions: interference, lighting, various changes on the face, etc. To solve these problems, a local detector for a reprint model of an object in an image was developed and described. A transforming autocoder (TA), a model of a neural network, was developed for the local detector. This neural network model is a subspecies of the general class of neural networks of reduced dimension. The local detector is able, in addition to determining the modified object, to determine the original shape of the object as well. A special feature of TA is the representation of image sections in a compact form and the evaluation of the parameters of the affine transformation. The transforming autocoder is a heterogeneous network (HS) consisting of a set of networks of smaller dimension. These networks are called capsules. Artificial neural networks should use local capsules that perform some rather complex internal calculations on their inputs, and then encapsulate the results of these calculations in a small vector of highly informative outputs. Each capsule learns to recognize an implicitly defined visual object in a limited area of viewing conditions and deformations. It outputs both the probability that the object is present in its limited area and a set of “instance parameters” that can include the exact pose, lighting, and deformation of the visual object relative to an implicitly defined canonical version of this object. The main advantage of capsules that output instance parameters is a simple way to recognize entire objects by recognizing their parts. The capsule can learn to display the pose of its visual object in a vector that is linearly related to the “natural” representations of the pose that are used in computer graphics. There is a simple and highly selective test for whether visual objects represented by two active capsules A and B have the correct spatial relationships for activating a higher-level capsule C. The transforming autoencoder solves the problem of identifying facial images in conditions of interference (noise), changes in illumination and angle.

2009 ◽  
Vol 8 (3) ◽  
pp. 887-897
Author(s):  
Vishal Paika ◽  
Er. Pankaj Bhambri

The face is the feature which distinguishes a person. Facial appearance is vital for human recognition. It has certain features like forehead, skin, eyes, ears, nose, cheeks, mouth, lip, teeth etc which helps us, humans, to recognize a particular face from millions of faces even after a large span of time and despite large changes in their appearance due to ageing, expression, viewing conditions and distractions such as disfigurement of face, scars, beard or hair style. A face is not merely a set of facial features but is rather but is rather something meaningful in its form.In this paper, depending on the various facial features, a system is designed to recognize them. To reveal the outline of the face, eyes, ears, nose, teeth etc different edge detection techniques have been used. These features are extracted in the term of distance between important feature points. The feature set obtained is then normalized and are feed to artificial neural networks so as to train them for reorganization of facial images.


Biology ◽  
2021 ◽  
Vol 10 (3) ◽  
pp. 182
Author(s):  
Rodrigo Dalvit Carvalho da Silva ◽  
Thomas Richard Jenkyn ◽  
Victor Alexander Carranza

In reconstructive craniofacial surgery, the bilateral symmetry of the midplane of the facial skeleton plays an important role in surgical planning. Surgeons can take advantage of the intact side of the face as a template for the malformed side by accurately locating the midplane to assist in the preparation of the surgical procedure. However, despite its importance, the location of the midline is still a subjective procedure. The aim of this study was to present a 3D technique using a convolutional neural network and geometric moments to automatically calculate the craniofacial midline symmetry of the facial skeleton from CT scans. To perform this task, a total of 195 skull images were assessed to validate the proposed technique. In the symmetry planes, the technique was found to be reliable and provided good accuracy. However, further investigations to improve the results of asymmetric images may be carried out.


2020 ◽  
Vol 4 (4(73)) ◽  
pp. 15-24
Author(s):  
S.N. Boranbayev ◽  
M.S. Amirtayev

The purpose of this article is to summarize the knowledge gained in the development and implementation of a neural network for facial recognition. Neural networks are used to solve complex tasks that require analytical calculations similar to what the human brain does. Machine learning algorithms are the foundation of a neural network. As input, the algorithm receives an image with people's faces, then searches for faces in this image using HOG (Histogram of oriented gradients). The result is images with explicit face structures. To determine unique facial features, the Face landmark algorithm is used, which finds 68 special points on the face. These points can be used to centerthe eyes and mouth for more accurate encoding. To get an accurate “face map” consisting of 128 dimensions, you need to use image encoding. Using the obtained data, the convolutional neural network can determine people's faces using the SVM linear classifier algorithm.


2018 ◽  
Vol 7 (2.13) ◽  
pp. 402
Author(s):  
Y Yusmartato ◽  
Zulkarnain Lubis ◽  
Solly Arza ◽  
Zulfadli Pelawi ◽  
A Armansah ◽  
...  

Lockers are one of the facilities that people use to store stuff. Artificial neural networks are computational systems where architecture and operations are inspired by the knowledge of biological neurons in the brain, which is one of the artificial representations of the human brain that always tries to stimulate the learning process of the human brain. One of the utilization of artificial neural network is for pattern recognition. The face of a person must be different but sometimes has a shape similar to the face of others, because the facial pattern is a good pattern to try to be recognized by using artificial neural networks. Pattern recognition on artificial neural network can be done by back propagation method. Back propagation method consists of input layer, hidden layer and output layer.  


2020 ◽  
Vol 10 (24) ◽  
pp. 8940
Author(s):  
Wanshun Gao ◽  
Xi Zhao ◽  
Jianhua Zou

Face recognition under drastic pose drops rapidly due to the limited samples during the model training. In this paper, we propose a pose-autoaugment face recognition framework (PAFR) based on the training of a Convolutional Neural Network (CNN) with multi-view face augmentation. The proposed framework consists of three parts: face augmentation, CNN training, and face matching. The face augmentation part is composed of pose autoaugment and background appending for increasing the pose variations of each subject. In the second part, we train a CNN model with the generated facial images to enhance the pose-invariant feature extraction. In the third part, we concatenate the feature vectors of each face and its horizontally flipped face from the trained CNN model to obtain a robust feature. The correlation score between the two faces is computed by the cosine similarity of their robust features. Comparable experiments are demonstrated on Bosphorus and CASIA-3D databases.


Sensors ◽  
2020 ◽  
Vol 20 (8) ◽  
pp. 2393 ◽  
Author(s):  
Daniel Octavian Melinte ◽  
Luige Vladareanu

The interaction between humans and an NAO robot using deep convolutional neural networks (CNN) is presented in this paper based on an innovative end-to-end pipeline method that applies two optimized CNNs, one for face recognition (FR) and another one for the facial expression recognition (FER) in order to obtain real-time inference speed for the entire process. Two different models for FR are considered, one known to be very accurate, but has low inference speed (faster region-based convolutional neural network), and one that is not as accurate but has high inference speed (single shot detector convolutional neural network). For emotion recognition transfer learning and fine-tuning of three CNN models (VGG, Inception V3 and ResNet) has been used. The overall results show that single shot detector convolutional neural network (SSD CNN) and faster region-based convolutional neural network (Faster R-CNN) models for face detection share almost the same accuracy: 97.8% for Faster R-CNN on PASCAL visual object classes (PASCAL VOCs) evaluation metrics and 97.42% for SSD Inception. In terms of FER, ResNet obtained the highest training accuracy (90.14%), while the visual geometry group (VGG) network had 87% accuracy and Inception V3 reached 81%. The results show improvements over 10% when using two serialized CNN, instead of using only the FER CNN, while the recent optimization model, called rectified adaptive moment optimization (RAdam), lead to a better generalization and accuracy improvement of 3%-4% on each emotion recognition CNN.


2021 ◽  
Vol 38 (4) ◽  
pp. 1007-1012
Author(s):  
Shakiba Ahmadimehr ◽  
Mohammad Karimi Moridani

This paper aims to explore the essence of facial attractiveness from the viewpoint of geometric features toward the classification and identification of attractive and unattractive individuals. We present a simple but useful feature extraction for facial beauty classification. Evaluation of facial attractiveness was performed with different combinations of geometric facial features using the deep learning method. In this method, we focus on the geometry of a face and use actual faces for our analysis. The proposed method has been tested on, image database containing 60 images of men's faces (attractive or unattractive) ranging from 20-50 years old. The images are taken from both frontal and lateral position. In the next step, principle components analysis (PCA) was applied to feature a reduction of beauty, and finally, the neural network was used for judging whether the obtained analysis of various faces is attractive or not. The results show that one of the indexes in identifying facial attractiveness base of science, is the values of the geometric features in the face, changing facial parameters can change the face from unattractive to attractive and vice versa. The experimental results are based on 60 facial images, high accuracy of 88%, and Sensitivity of 92% is obtained for 2-level classification (attractive or not).


2018 ◽  
Author(s):  
Werton P. De Araujo ◽  
Thelmo P. De Araujo ◽  
Gustavo A. L. De Campos

Eye detection is a preprocessing step in many methods using facial images. Some algorithms to detect eyes are based on the characteristics of the gradient flow in the iris-sclera boundary. These algorithms are usually applied to the whole face and a posterior heuristic is used to remove false positives. In this paper, we reverse that approach by using a Convolutional Neural Network (CNN) to solve a regression problem and give a coarse estimate of the eye regions, and only then do we apply the gradient-based algorithms. The CNN was combined with two gradient-based algorithms and the results were evaluated regarding their accuracy and processing time, showing the applicability of both methods for eye localization.


2019 ◽  
Vol 8 (3) ◽  
pp. 1932-1938

In this work, deep learning methods are used to classify the facial images. ORL Database is used for the purpose of training the models and for testing. Three kinds of models are developed and their performances are measured. Convolutional Neural Networks (CNN), Convolutional Neural Network Based Inception Model with single training image per class (CNN-INC) and Convolutional Neural Network Based Inception Model with several training images per class (CNN-INC-MEAN) are developed. The ORL database has ten facial images for each person. Five images are used for training purpose and remaining 5 images are used for testing. The five images for the training are chosen randomly so that two sets of training and testing data is generated. The models are trained and tested on the two sets that are drawn from the same population. The results are presented for accuracy of face recognition


2011 ◽  
Vol 204-210 ◽  
pp. 216-219
Author(s):  
Hong Zhang

It's well known that the technology of human face recognition has become a hot topicin pattern recognition field. Though a lot of progress has been made by many researchersthese years, many key problems still have to be solved in order to popularize the application of face recognition because of the complexity of face recognition. The background, development and main methods of face recognition are introducedfirstly in this paper, then a face recognition method which is based on wavelet transform,KL transform and BP neural networks is used in the paper.Here the face feature extraction includes wavelet transform and KL transform.Moreover,the recognition classifier is BP neural networks.The simulation testing in the paper holds good recognition rate.


Sign in / Sign up

Export Citation Format

Share Document