scholarly journals Learning, compression, and leakage: Minimising classification error via meta-universal compression principles

Author(s):  
Fernando E. Rosas ◽  
Pedro A. M. Mediano ◽  
Michael Gastpar
Cybersecurity ◽  
2021 ◽  
Vol 4 (1) ◽  
Author(s):  
Shushan Arakelyan ◽  
Sima Arasteh ◽  
Christophe Hauser ◽  
Erik Kline ◽  
Aram Galstyan

AbstractTackling binary program analysis problems has traditionally implied manually defining rules and heuristics, a tedious and time consuming task for human analysts. In order to improve automation and scalability, we propose an alternative direction based on distributed representations of binary programs with applicability to a number of downstream tasks. We introduce Bin2vec, a new approach leveraging Graph Convolutional Networks (GCN) along with computational program graphs in order to learn a high dimensional representation of binary executable programs. We demonstrate the versatility of this approach by using our representations to solve two semantically different binary analysis tasks – functional algorithm classification and vulnerability discovery. We compare the proposed approach to our own strong baseline as well as published results, and demonstrate improvement over state-of-the-art methods for both tasks. We evaluated Bin2vec on 49191 binaries for the functional algorithm classification task, and on 30 different CWE-IDs including at least 100 CVE entries each for the vulnerability discovery task. We set a new state-of-the-art result by reducing the classification error by 40% compared to the source-code based inst2vec approach, while working on binary code. For almost every vulnerability class in our dataset, our prediction accuracy is over 80% (and over 90% in multiple classes).


Author(s):  
Leijin Long ◽  
Feng He ◽  
Hongjiang Liu

AbstractIn order to monitor the high-level landslides frequently occurring in Jinsha River area of Southwest China, and protect the lives and property safety of people in mountainous areas, the data of satellite remote sensing images are combined with various factors inducing landslides and transformed into landslide influence factors, which provides data basis for the establishment of landslide detection model. Then, based on the deep belief networks (DBN) and convolutional neural network (CNN) algorithm, two landslide detection models DBN and convolutional neural-deep belief network (CDN) are established to monitor the high-level landslide in Jinsha River. The influence of the model parameters on the landslide detection results is analyzed, and the accuracy of DBN and CDN models in dealing with actual landslide problems is compared. The results show that when the number of neurons in the DBN is 100, the overall error is the minimum, and when the number of learning layers is 3, the classification error is the minimum. The detection accuracy of DBN and CDN is 97.56% and 97.63%, respectively, which indicates that both DBN and CDN models are feasible in dealing with landslides from remote sensing images. This exploration provides a reference for the study of high-level landslide disasters in Jinsha River.


2021 ◽  
Vol 11 (9) ◽  
pp. 4292
Author(s):  
Mónica Y. Moreno-Revelo ◽  
Lorena Guachi-Guachi ◽  
Juan Bernardo Gómez-Mendoza ◽  
Javier Revelo-Fuelagán ◽  
Diego H. Peluffo-Ordóñez

Automatic crop identification and monitoring is a key element in enhancing food production processes as well as diminishing the related environmental impact. Although several efficient deep learning techniques have emerged in the field of multispectral imagery analysis, the crop classification problem still needs more accurate solutions. This work introduces a competitive methodology for crop classification from multispectral satellite imagery mainly using an enhanced 2D convolutional neural network (2D-CNN) designed at a smaller-scale architecture, as well as a novel post-processing step. The proposed methodology contains four steps: image stacking, patch extraction, classification model design (based on a 2D-CNN architecture), and post-processing. First, the images are stacked to increase the number of features. Second, the input images are split into patches and fed into the 2D-CNN model. Then, the 2D-CNN model is constructed within a small-scale framework, and properly trained to recognize 10 different types of crops. Finally, a post-processing step is performed in order to reduce the classification error caused by lower-spatial-resolution images. Experiments were carried over the so-named Campo Verde database, which consists of a set of satellite images captured by Landsat and Sentinel satellites from the municipality of Campo Verde, Brazil. In contrast to the maximum accuracy values reached by remarkable works reported in the literature (amounting to an overall accuracy of about 81%, a f1 score of 75.89%, and average accuracy of 73.35%), the proposed methodology achieves a competitive overall accuracy of 81.20%, a f1 score of 75.89%, and an average accuracy of 88.72% when classifying 10 different crops, while ensuring an adequate trade-off between the number of multiply-accumulate operations (MACs) and accuracy. Furthermore, given its ability to effectively classify patches from two image sequences, this methodology may result appealing for other real-world applications, such as the classification of urban materials.


2011 ◽  
Vol 22 (8) ◽  
pp. 1334-1339 ◽  
Author(s):  
T. Windeatt ◽  
Cemre Zor
Keyword(s):  

1971 ◽  
Vol C-20 (12) ◽  
pp. 1521-1527 ◽  
Author(s):  
K. Fukunaga ◽  
D.L. Kessell
Keyword(s):  

Author(s):  
Zbigniew Omiotek

The purpose of the study was to construct an efficient classifier that, along with a given reduced set of discriminant features, could be used as a part of the computer system in automatic identification and classification of ultrasound images of the thyroid gland, which is aimed to detect cases affected by Hashimoto’s thyroiditis. A total of 10 supervised learning techniques and a majority vote for the combined classifier were used. Two models were proposed as a result of the classifier’s construction. The first one is based on the K-nearest neighbours method (for K = 7). It uses three discriminant features and affords sensitivity equal to 88.1%, specificity of 66.7% and classification error at a level of 21.8%. The second model is a combined classifier, which was constructed using three-component classifiers. They are based on the K-nearest neighbours method (for K = 7), linear discriminant analysis and a boosting algorithm. The combined classifier is based on 48 discriminant features. It allows to achieve the classification sensitivity equal to 88.1%, specificity of 69.4% and classification error at a level of 20.5%. The combined classifier allows to improve the classification quality compared to the single model. The models, built as a part of the automatic computer system, may support the physician, especially in first-contact hospitals, in diagnosis of cases that are difficult to recognise based on ultrasound images. The high sensitivity of constructed classification models indicates high detection accuracy of the sick cases, and this is beneficial to the patients from a medical point of view.


2006 ◽  
Vol 32 (3) ◽  
pp. 417-438 ◽  
Author(s):  
Diane Litman ◽  
Julia Hirschberg ◽  
Marc Swerts

This article focuses on the analysis and prediction of corrections, defined as turns where a user tries to correct a prior error made by a spoken dialogue system. We describe our labeling procedure of various corrections types and statistical analyses of their features in a corpus collected from a train information spoken dialogue system. We then present results of machine-learning experiments designed to identify user corrections of speech recognition errors. We investigate the predictive power of features automatically computable from the prosody of the turn, the speech recognition process, experimental conditions, and the dialogue history. Our best-performing features reduce classification error from baselines of 25.70–28.99% to 15.72%.


2021 ◽  
Vol 17 (1) ◽  
pp. 53-67
Author(s):  
Rajneesh Rani ◽  
Harpreet Singh

In this busy world, biometric authentication methods are serving as fast authentication means. But with growing dependencies on these systems, attackers have tried to exploit these systems through various attacks; thus, there is a strong need to protect authentication systems. Many software and hardware methods have been proposed in the past to make existing authentication systems more robust. Liveness detection/presentation attack detection is one such method that provides protection against malicious agents by detecting fake samples of biometric traits. This paper has worked on fingerprint liveness detection/presentation attack detection using transfer learning for which the authors have used a pre-trained NASNetMobile model. The experiments are performed on publicly available liveness datasets LivDet 2011 and LivDet 2013 and have obtained good results as compared to state of art techniques in terms of ACE(average classification error).


Sign in / Sign up

Export Citation Format

Share Document