MAPPING 2D TO 3D FORENSIC FACIAL RECOGNITION VIA BIO- INSPIRED ACTIVE APPEARANCE MODEL

2015 ◽  
Vol 78 (2-2) ◽  
Author(s):  
Siti Zaharah Abd. Rahman ◽  
Siti Norul Huda Sheikh Abdullah ◽  
Lim Eng Hao ◽  
Mohammed Hasan Abdulameer ◽  
Nazri Ahmad Zamani ◽  
...  

This research done is to solve the problems faced by digital forensic analysts in identifying a suspect captured on their CCTV. Identifying the suspect through the CCTV video footage is a very challenging task for them as it involves tedious rounds of processes to match the facial information in the video footage to a set of suspect’s images. The biggest problem faced by digital forensic analysis is modeling 2D model extracted from CCTV video as the model does not provide enough information to carry out the identification process. Problems occur when a suspect in the video is not facing the camera, the image extracted is the side image of the suspect and it is difficult to make a matching with portrait image in the database. There are also many factors that contribute to the process of extracting facial information from a video to be difficult, such as low-quality video. Through 2D to 3D image model mapping, any partial face information that is incomplete can be matched more efficiently with 3D data by rotating it to matched position. The first methodology in this research is data collection; any data obtained through video recorder. Then, the video will be converted into an image. Images are used to develop the Active Appearance Model (the 2D face model is AAM) 2D and AAM 3D. AAM is used as an input for learning and testing process involving three classifiers, which are Random Forest, Support Vector Machine (SVM), and Neural Networks classifier. The experimental results show that the 3D model is more suitable for use in face recognition as the percentage of the recognition is higher compared with the 2D model.

2014 ◽  
Vol 2014 ◽  
pp. 1-11 ◽  
Author(s):  
Ji-Xiang Du ◽  
Xing Wu ◽  
Chuan-Min Zhai

Face verification in the presence of age progression is an important problem that has not been widely addressed. In this paper, we propose to use the active appearance model (AAM) and gradient orientation pyramid (GOP) feature representation for this problem. First, we use the AAM on the dataset and generate the AAM images; we then get the representation of gradient orientation on a hierarchical model, which is the appearance of GOP. When combined with a support vector machine (SVM), experimental results show that our approach has excellent performance on two public domain face aging datasets: FGNET and MORPH. Second, we compare the performance of the proposed methods with a number of related face verification methods; the results show that the new approach is more robust and performs better.


2016 ◽  
Vol 16 (04) ◽  
pp. 1650019 ◽  
Author(s):  
Flávio Altinier Maximiano da Silva ◽  
Helio Pedrini

One of the most effective ways of expressing emotion is through facial expressions. This work proposes and discusses a geometrical descriptor based on the calculation of distances from coordinates of facial fiducial points, which are used as features for training support vector machines (SVM) to classify emotions. Three data sets are studied and six basic emotions are considered in our experiments. In comparison to other approaches available in the literature, the results obtained with our geometrical descriptor demonstrated to be very competitive, achieving high classification F-score rates. Additionally, we evaluate whether the combination of our geometrical descriptor with an appearance feature, the Gabor filter, allows emotions to be even more distinguishable for the classifier. The result is positive for two out of three data sets. Finally, to simulate in-the-wild scenarios, an active appearance model (AAM) is trained to position the fiducial points on the correct facial locations, instead of using the ones provided by the data sets. As the fitting error is considered acceptable, the former experiments are also conducted with the new data generated by the AAM. The results show a small drop on the F-score values when compared to the data originally provided by the data sets,but are still satisfactory.


Data ◽  
2021 ◽  
Vol 6 (8) ◽  
pp. 87
Author(s):  
Sara Ferreira ◽  
Mário Antunes ◽  
Manuel E. Correia

Deepfake and manipulated digital photos and videos are being increasingly used in a myriad of cybercrimes. Ransomware, the dissemination of fake news, and digital kidnapping-related crimes are the most recurrent, in which tampered multimedia content has been the primordial disseminating vehicle. Digital forensic analysis tools are being widely used by criminal investigations to automate the identification of digital evidence in seized electronic equipment. The number of files to be processed and the complexity of the crimes under analysis have highlighted the need to employ efficient digital forensics techniques grounded on state-of-the-art technologies. Machine Learning (ML) researchers have been challenged to apply techniques and methods to improve the automatic detection of manipulated multimedia content. However, the implementation of such methods have not yet been massively incorporated into digital forensic tools, mostly due to the lack of realistic and well-structured datasets of photos and videos. The diversity and richness of the datasets are crucial to benchmark the ML models and to evaluate their appropriateness to be applied in real-world digital forensics applications. An example is the development of third-party modules for the widely used Autopsy digital forensic application. This paper presents a dataset obtained by extracting a set of simple features from genuine and manipulated photos and videos, which are part of state-of-the-art existing datasets. The resulting dataset is balanced, and each entry comprises a label and a vector of numeric values corresponding to the features extracted through a Discrete Fourier Transform (DFT). The dataset is available in a GitHub repository, and the total amount of photos and video frames is 40,588 and 12,400, respectively. The dataset was validated and benchmarked with deep learning Convolutional Neural Networks (CNN) and Support Vector Machines (SVM) methods; however, a plethora of other existing ones can be applied. Generically, the results show a better F1-score for CNN when comparing with SVM, both for photos and videos processing. CNN achieved an F1-score of 0.9968 and 0.8415 for photos and videos, respectively. Regarding SVM, the results obtained with 5-fold cross-validation are 0.9953 and 0.7955, respectively, for photos and videos processing. A set of methods written in Python is available for the researchers, namely to preprocess and extract the features from the original photos and videos files and to build the training and testing sets. Additional methods are also available to convert the original PKL files into CSV and TXT, which gives more flexibility for the ML researchers to use the dataset on existing ML frameworks and tools.


Sign in / Sign up

Export Citation Format

Share Document