scholarly journals Design and Implementation of Multilayer Security for ATM Machines

Author(s):  
Anjalin Joy ◽  
◽  
Caren Babu ◽  

Automated teller machine (ATM) nowadays are a favourite spot for attackers as they are available everywhere and are much easier to rob. Generally, ATM attacks can be either physical ATM attacks or ATM-related fraud attacks. In this paper the idea of an ATM system with multilayer security is proposed with the help of internet of things (IoT), fingerprint identification and face recognition to increase the security of ATM. The physical ATM counter attacks can be identified by using specific sensors to detect changes in vibration and temperature in the ATM counter. To prevent ATM related fraud attacks the proposed system has additional security features like fingerprint identification and face recognition along with ATM number verification. The convolutional neural network (CNN) and machine learning based face recognition is used in this work which is quite reliable. Failures in any of the above steps cancel the transactions and so the proposed system provides multi layer security which makes it impossible for the attackers to break the ATM security. The proposed system will help to increase the security of the ATM and provide safe and secure ATM transactions.

Author(s):  
Abhay Patil

Abstract: Animal intervention is significant intimidation to the potency of the crops, which influences food security and decreases the value to the farmers. This suggested model displays the growth of the Internet of Things and Machine learning technique-based resolutions to surmount this obstacle. Raspberry Pi commands the machine algorithm, which is interfaced with the ESP8266 Wireless Fidelity module, Pi-Camera, Speaker/Buzzer, and LED. Machine learning algorithms similar to Regionbased Convolutional Neural Network and Single Shot Detection technology represents an essential function to identify the target in the pictures and classify the creatures. The experimentation exhibits that the Single Shot Detection algorithm exceeds than Region-based Convolutional Neural Network algorithm. Ultimately, the Twilio API interfaced software decimates the data to the farmers to take conclusive work in their farm territory. Keywords: Region-Based Convolutional Neural Network (R-CNN), Tensor Flow, Raspberry Pi, Internet of Things (IoT), Single Shot Detector (SSD)


Recognition holds great significance to give biometric authentications that are utilized in various applications particularly in attendance and security. A gathered database of the subjects is converted applying image processing methods to make this task. This paper suggests a cascade object detector based face detection and convolutional neural network alexnet based face recognition that can recognize the faces. The techniques used for face recognition are machine learning-based methods because of their great precision as associated with different methods. Face detection is the initial level before face recognition that is done utilizing a cascade object detector classifier. Face recognition is performed utilizing Deep Learning's sub-field that is Convolutional Neural Network (CNN). It is a multi-layer network which is used to train the network, to perform a particular task using classification. Check learning of a trained CNN model that is AlexNet is used for face recognition.


Author(s):  
Satoru Tsuiki ◽  
Takuya Nagaoka ◽  
Tatsuya Fukuda ◽  
Yuki Sakamoto ◽  
Fernanda R. Almeida ◽  
...  

Abstract Purpose In 2-dimensional lateral cephalometric radiographs, patients with severe obstructive sleep apnea (OSA) exhibit a more crowded oropharynx in comparison with non-OSA. We tested the hypothesis that machine learning, an application of artificial intelligence (AI), could be used to detect patients with severe OSA based on 2-dimensional images. Methods A deep convolutional neural network was developed (n = 1258; 90%) and tested (n = 131; 10%) using data from 1389 (100%) lateral cephalometric radiographs obtained from individuals diagnosed with severe OSA (n = 867; apnea hypopnea index > 30 events/h sleep) or non-OSA (n = 522; apnea hypopnea index < 5 events/h sleep) at a single center for sleep disorders. Three kinds of data sets were prepared by changing the area of interest using a single image: the original image without any modification (full image), an image containing a facial profile, upper airway, and craniofacial soft/hard tissues (main region), and an image containing part of the occipital region (head only). A radiologist also performed a conventional manual cephalometric analysis of the full image for comparison. Results The sensitivity/specificity was 0.87/0.82 for full image, 0.88/0.75 for main region, 0.71/0.63 for head only, and 0.54/0.80 for the manual analysis. The area under the receiver-operating characteristic curve was the highest for main region 0.92, for full image 0.89, for head only 0.70, and for manual cephalometric analysis 0.75. Conclusions A deep convolutional neural network identified individuals with severe OSA with high accuracy. Future research on this concept using AI and images can be further encouraged when discussing triage of OSA.


2021 ◽  
Vol 4 (1) ◽  
Author(s):  
Peter M. Maloca ◽  
Philipp L. Müller ◽  
Aaron Y. Lee ◽  
Adnan Tufail ◽  
Konstantinos Balaskas ◽  
...  

AbstractMachine learning has greatly facilitated the analysis of medical data, while the internal operations usually remain intransparent. To better comprehend these opaque procedures, a convolutional neural network for optical coherence tomography image segmentation was enhanced with a Traceable Relevance Explainability (T-REX) technique. The proposed application was based on three components: ground truth generation by multiple graders, calculation of Hamming distances among graders and the machine learning algorithm, as well as a smart data visualization (‘neural recording’). An overall average variability of 1.75% between the human graders and the algorithm was found, slightly minor to 2.02% among human graders. The ambiguity in ground truth had noteworthy impact on machine learning results, which could be visualized. The convolutional neural network balanced between graders and allowed for modifiable predictions dependent on the compartment. Using the proposed T-REX setup, machine learning processes could be rendered more transparent and understandable, possibly leading to optimized applications.


Sensors ◽  
2019 ◽  
Vol 19 (1) ◽  
pp. 210 ◽  
Author(s):  
Zied Tayeb ◽  
Juri Fedjaev ◽  
Nejla Ghaboosi ◽  
Christoph Richter ◽  
Lukas Everding ◽  
...  

Non-invasive, electroencephalography (EEG)-based brain-computer interfaces (BCIs) on motor imagery movements translate the subject’s motor intention into control signals through classifying the EEG patterns caused by different imagination tasks, e.g., hand movements. This type of BCI has been widely studied and used as an alternative mode of communication and environmental control for disabled patients, such as those suffering from a brainstem stroke or a spinal cord injury (SCI). Notwithstanding the success of traditional machine learning methods in classifying EEG signals, these methods still rely on hand-crafted features. The extraction of such features is a difficult task due to the high non-stationarity of EEG signals, which is a major cause by the stagnating progress in classification performance. Remarkable advances in deep learning methods allow end-to-end learning without any feature engineering, which could benefit BCI motor imagery applications. We developed three deep learning models: (1) A long short-term memory (LSTM); (2) a spectrogram-based convolutional neural network model (CNN); and (3) a recurrent convolutional neural network (RCNN), for decoding motor imagery movements directly from raw EEG signals without (any manual) feature engineering. Results were evaluated on our own publicly available, EEG data collected from 20 subjects and on an existing dataset known as 2b EEG dataset from “BCI Competition IV”. Overall, better classification performance was achieved with deep learning models compared to state-of-the art machine learning techniques, which could chart a route ahead for developing new robust techniques for EEG signal decoding. We underpin this point by demonstrating the successful real-time control of a robotic arm using our CNN based BCI.


Author(s):  
E. Yu. Shchetinin

The recognition of human emotions is one of the most relevant and dynamically developing areas of modern speech technologies, and the recognition of emotions in speech (RER) is the most demanded part of them. In this paper, we propose a computer model of emotion recognition based on an ensemble of bidirectional recurrent neural network with LSTM memory cell and deep convolutional neural network ResNet18. In this paper, computer studies of the RAVDESS database containing emotional speech of a person are carried out. RAVDESS-a data set containing 7356 files. Entries contain the following emotions: 0 – neutral, 1 – calm, 2 – happiness, 3 – sadness, 4 – anger, 5 – fear, 6 – disgust, 7 – surprise. In total, the database contains 16 classes (8 emotions divided into male and female) for a total of 1440 samples (speech only). To train machine learning algorithms and deep neural networks to recognize emotions, existing audio recordings must be pre-processed in such a way as to extract the main characteristic features of certain emotions. This was done using Mel-frequency cepstral coefficients, chroma coefficients, as well as the characteristics of the frequency spectrum of audio recordings. In this paper, computer studies of various models of neural networks for emotion recognition are carried out on the example of the data described above. In addition, machine learning algorithms were used for comparative analysis. Thus, the following models were trained during the experiments: logistic regression (LR), classifier based on the support vector machine (SVM), decision tree (DT), random forest (RF), gradient boosting over trees – XGBoost, convolutional neural network CNN, recurrent neural network RNN (ResNet18), as well as an ensemble of convolutional and recurrent networks Stacked CNN-RNN. The results show that neural networks showed much higher accuracy in recognizing and classifying emotions than the machine learning algorithms used. Of the three neural network models presented, the CNN + BLSTM ensemble showed higher accuracy.


Sign in / Sign up

Export Citation Format

Share Document