scholarly journals Efficient Facial Landmark Localization Based on Binarized Neural Networks

Electronics ◽  
2020 ◽  
Vol 9 (8) ◽  
pp. 1236
Author(s):  
Hanlin Chen ◽  
Xudong Zhang ◽  
Teli Ma ◽  
Haosong Yue ◽  
Xin Wang ◽  
...  

Facial landmark localization is a significant yet challenging computer vision task, whose accuracy has been remarkably improved due to the successful application of deep Convolutional Neural Networks (CNNs). However, CNNs require huge storage and computation overhead, thus impeding their deployment on computationally limited platforms. In this paper, to the best of our knowledge, it is the first time that an efficient facial landmark localization is implemented via binarized CNNs. We introduce a new network architecture to calculate the binarized models, referred to as Amplitude Convolutional Networks (ACNs), based on the proposed asynchronous back propagation algorithm. We can efficiently recover the full-precision filters only using a single factor in an end-to-end manner, and the efficiency of CNNs for facial landmark localization is further improved by the extremely compressed 1-bit ACNs. Our ACNs reduce the storage space of convolutional filters by a factor of 32 compared with the full-precision models on dataset LFW+Webface, CelebA, BioID and 300W, while achieving a comparable performance to the full-precision facial landmark localization algorithms.

2019 ◽  
Vol 1 (11) ◽  
Author(s):  
Chollette C. Olisah ◽  
Lyndon Smith

Abstract Deep convolutional neural networks have achieved huge successes in application domains like object and face recognition. The performance gain is attributed to different facets of the network architecture such as: depth of the convolutional layers, activation function, pooling, batch normalization, forward and back propagation and many more. However, very little emphasis is made on the preprocessor’s module of the network. Therefore, in this paper, the network’s preprocessing module is varied across different preprocessing approaches while keeping constant other facets of the deep network architecture, to investigate the contribution preprocessing makes to the network. Commonly used preprocessors are the data augmentation and normalization and are termed conventional preprocessors. Others are termed the unconventional preprocessors, they are: color space converters; grey-level resolution preprocessors; full-based and plane-based image quantization, Gaussian blur, illumination normalization and insensitive feature preprocessors. To achieve fixed network parameters, CNNs with transfer learning is employed. The aim is to transfer knowledge from the high-level feature vectors of the Inception-V3 network to offline preprocessed LFW target data; and features is trained using the SoftMax classifier for face identification. The experiments show that the discriminative capability of the deep networks can be improved by preprocessing RGB data with some of the unconventional preprocessors before feeding it to the CNNs. However, for best performance, the right setup of preprocessed data with augmentation and/or normalization is required. Summarily, preprocessing data before it is fed to the deep network is found to increase the homogeneity of neighborhood pixels even at reduced bit depth which serves for better storage efficiency.


2021 ◽  
pp. 10-17
Author(s):  
S. S. Yudachev ◽  
N. A. Gordienko ◽  
F. M. Bosy

The article describes an algorithm for the synthesis of neural networks for controlling the gyrostabilizer. The neural network acts as an observer of the state vector. The role of such an observer is to provide feedback to the gyrostabilizer, which is illustrated in the article. Gyrostabilizer is a gyroscopic device designed to stabilize individual objects or devices, as well as to determine the angular deviations of objects. Gyrostabilizer systems will be more widely used, as they provide an effective means of motion control with a number of significant advantages for various designs. The article deals in detail with the issue of specific stage features of classical algorithms: selecting the network architecture, training the neural network, and verifying the results of feedback control. In recent years, neural networks have become an increasingly powerful tool in scientific computing. The universal approximation theorem states that a neural network can be constructed to approximate any given continuous function with the required accuracy. The back propagation algorithm also allows effectively optimizing the parameters when training a neural network. Due to the use of graphics processors, it is possible to perform efficient calculations for scientific and engineering tasks. The article presents the optimal configuration of the neural network, such as the depth of memory, the number of layers and neurons in these layers, as well as the functions of the activation layer. In addition, it provides data on dynamic systems to improve neural network training. An optimal training scheme is also provided.


2021 ◽  
Vol 12 (1) ◽  
Author(s):  
Florian Stelzer ◽  
André Röhm ◽  
Raul Vicente ◽  
Ingo Fischer ◽  
Serhiy Yanchuk

AbstractDeep neural networks are among the most widely applied machine learning tools showing outstanding performance in a broad range of tasks. We present a method for folding a deep neural network of arbitrary size into a single neuron with multiple time-delayed feedback loops. This single-neuron deep neural network comprises only a single nonlinearity and appropriately adjusted modulations of the feedback signals. The network states emerge in time as a temporal unfolding of the neuron’s dynamics. By adjusting the feedback-modulation within the loops, we adapt the network’s connection weights. These connection weights are determined via a back-propagation algorithm, where both the delay-induced and local network connections must be taken into account. Our approach can fully represent standard Deep Neural Networks (DNN), encompasses sparse DNNs, and extends the DNN concept toward dynamical systems implementations. The new method, which we call Folded-in-time DNN (Fit-DNN), exhibits promising performance in a set of benchmark tasks.


Sensors ◽  
2021 ◽  
Vol 21 (8) ◽  
pp. 2852
Author(s):  
Parvathaneni Naga Srinivasu ◽  
Jalluri Gnana SivaSai ◽  
Muhammad Fazal Ijaz ◽  
Akash Kumar Bhoi ◽  
Wonjoon Kim ◽  
...  

Deep learning models are efficient in learning the features that assist in understanding complex patterns precisely. This study proposed a computerized process of classifying skin disease through deep learning based MobileNet V2 and Long Short Term Memory (LSTM). The MobileNet V2 model proved to be efficient with a better accuracy that can work on lightweight computational devices. The proposed model is efficient in maintaining stateful information for precise predictions. A grey-level co-occurrence matrix is used for assessing the progress of diseased growth. The performance has been compared against other state-of-the-art models such as Fine-Tuned Neural Networks (FTNN), Convolutional Neural Network (CNN), Very Deep Convolutional Networks for Large-Scale Image Recognition developed by Visual Geometry Group (VGG), and convolutional neural network architecture that expanded with few changes. The HAM10000 dataset is used and the proposed method has outperformed other methods with more than 85% accuracy. Its robustness in recognizing the affected region much faster with almost 2× lesser computations than the conventional MobileNet model results in minimal computational efforts. Furthermore, a mobile application is designed for instant and proper action. It helps the patient and dermatologists identify the type of disease from the affected region’s image at the initial stage of the skin disease. These findings suggest that the proposed system can help general practitioners efficiently and effectively diagnose skin conditions, thereby reducing further complications and morbidity.


2021 ◽  
Author(s):  
Mateus Alexandre da Silva ◽  
Marina Neves Merlo ◽  
Michael Silveira Thebaldi ◽  
Danton Diego Ferreira ◽  
Felipe Schwerz ◽  
...  

Abstract Predicting rainfall can prevent and mitigate damages caused by its deficit or excess, besides providing necessary tools for adequate planning for the use of water. This research aimed to predict the monthly rainfall, one month in advance, in four municipalities in the metropolitan region of Belo Horizonte, using artificial neural networks (ANN) trained with different climate variables, and to indicate the suitability of such variables as inputs to these models. The models were developed through the MATLAB® software version R2011a, using the NNTOOL toolbox. The ANN’s were trained by the multilayer perceptron architecture and the Feedforward and Back propagation algorithm, using two combinations of input data were used, with 2 and 6 variables, and one combination of input data with 3 of the 6 variables most correlated to observed rainfall from 1970 to 1999, to predict the rainfall from 2000 to 2009. The most correlated variables to the rainfall of the following month are the sequential number corresponding to the month, total rainfall and average compensated temperature, and the best performance was obtained with these variables. Furthermore, it was concluded that the performance of the models was satisfactory; however, they presented limitations for predicting months with high rainfall.


2013 ◽  
Vol 14 (6) ◽  
pp. 431-439 ◽  
Author(s):  
Issam Hanafi ◽  
Francisco Mata Cabrera ◽  
Abdellatif Khamlichi ◽  
Ignacio Garrido ◽  
José Tejero Manzanares

Sign in / Sign up

Export Citation Format

Share Document