scholarly journals Malicious Javascript Detection based on Clustering Techniques

2021 ◽  
Vol 13 (6) ◽  
pp. 11-21
Author(s):  
Nguyen Hong Son ◽  
Ha Thanh Dung

Malicious JavaScript code is still a problem for website and web users. The complication and equivocation of this code make the detection which is based on signatures of antivirus programs becomes ineffective. So far, the alternative methods using machine learning have achieved encouraging results, and have detected malicious JavaScript code with high accuracy. However, according to the supervised learning method, the models, which are introduced, depend on the number of labeled symbols and require significant computational resources to activate. The rapid growth of malicious JavaScript is a real challenge to the solutions based on supervised learning due to the lacking of experience in detecting new forms of malicious JavaScript code. In this paper, we deal with the challenge by the method of detecting malicious JavaScript based on clustering techniques. The known symbols that will be analyzed, the characteristics which are extracted, and a detection processing technique applied on output clusters are included in the model. This method is not computationally complicated, as well as the typical case experiments gave positive results; specifically, it has detected new forms of malicious JavaScript code.

1998 ◽  
Vol 26 (5) ◽  
pp. 679-708 ◽  
Author(s):  
Horst Spielmann ◽  
Michael Balls ◽  
Jack Dupuis ◽  
Wolfgang J. W. Pape ◽  
Odile de Silva ◽  
...  

In 1996, the Scientific Committee on Cosmetology of DGXXIV of the European Commission asked the European Centre for the Validation of Alternative Methods to test eight UV filter chemicals from the 1995 edition of Annex VII of Directive 76/768/EEC in a blind trial in the in vitro 3T3 cell neutral red uptake phototoxicity (3T3 NRU PT) test, which had been scientifically validated between 1992 and 1996. Since all the UV filter chemicals on the positive list of EU Directive 76/768/EEC have been shown not to be phototoxic in vivo in humans under use conditions, only negative effects would be expected in the 3T3 NRU PT test. To balance the number of positive and negative chemicals, ten phototoxic and ten non-phototoxic chemicals were tested under blind conditions in four laboratories. Moreover, to assess the optimum concentration range for testing, information was provided on appropriate solvents and on the solubility of the coded chemicals. In this study, the phototoxic potential of test chemicals was evaluated in a prediction model in which either the Photoirritation Factor (PIF) or the Mean Photo Effect (MPE) were determined. The results obtained with both PIF and MPE were highly reproducible in the four laboratories, and the correlation between in vitro and in vivo data was almost perfect. All the phototoxic test chemicals provided a positive result at concentrations of 1μg/ml, while nine of the ten non-phototoxic chemicals gave clear negative results, even at the highest test concentrations. One of the UV filter chemicals gave positive results in three of the four laboratories only at concentrations greater than 100μg/ml; the other laboratory correctly identified all 20 of the test chemicals. An analysis of the impact that exposure concentrations had on the performance of the test revealed that the optimum concentration range in the 3T3 NRU PT test for determining the phototoxic potential of chemicals is between 0.1μg/ml and 10μg/ml, and that false positive results can be obtained at concentrations greater than 100μg/ml. Therefore, the positive results obtained with some of the UV filter chemicals only at concentrations greater than 100μg/ml do not indicate a phototoxic potential in vivo. When this information was taken into account during calculation of the overall predictivity of the 3T3 NRU PT test in the present study, an almost perfect correlation of in vitro versus in vivo results was obtained (between 95% and 100%), when either PIF or MPE were used to predict the phototoxic potential. The management team and participants therefore conclude that the 3T3 NRU PT test is a valid test for correctly assessing the phototoxic potential of UV filter chemicals, if the defined concentration limits are taken into account.


Author(s):  
Mamta Bisht ◽  
Richa Gupta

Script recognition is the first necessary preliminary step for text recognition. In the deep learning era, for this task two essential requirements are the availability of a large labeled dataset for training and computational resources to train models. But if we have limitations on these requirements then we need to think of alternative methods. This provides an impetus to explore the field of transfer learning, in which the previously trained model knowledge established in the benchmark dataset can be reused in another smaller dataset for another task, thus saving computational power as it requires to train only less number of parameters from the total parameters in the model. Here we study two pre-trained models and fine-tune them for script classification tasks. Firstly, the VGG-16 pre-trained model is fine-tuned for publically available CVSI-15 and MLe2e datasets for script recognition. Secondly, a well-performed model on Devanagari handwritten characters dataset has been adopted and fine-tuned for the Kaggle Devanagari numeral dataset for numeral recognition. The performance of proposed fine-tune models is related to the nature of the target dataset as similar or dissimilar from the original dataset and it has been analyzed with widely used optimizers.


2021 ◽  
Vol 12 (1) ◽  
pp. 29-52
Author(s):  
Raja Guru R. ◽  
Naresh Kumar P.

Unmanned aerial vehicles (UAV) play a significant role in finding victims affected in the post-disaster zone, where a man cannot risk his life under a critical condition of the disaster environment. The proposed design incorporates autonomous vision-based navigation through the disaster environment based on general graph theory with dynamic changes on the length between two or multiple nodes, where a node is a pathway. Camera fixed on it continuously captures the surrounding footage, processing it frame by frame on-site using image processing technique based on a SOC. Identifies victims in the zone and the pathways available for traversal. UAV uses an ultrasonic rangefinder to avoid collision with obstacles. The system alerts the rescue team if any victim detected and transmits the frames using CRN to the off-site console. UAV learns navigation policy that achieves high accuracy in real-time environments; communication using CRN is uninterrupted and useful during such emergencies.


Author(s):  
Baoying Wang ◽  
Imad Rahal ◽  
Richard Leipold

Data clustering is a discovery process that partitions a data set into groups (clusters) such that data points within the same group have high similarity while being very dissimilar to points in other groups (Han & Kamber, 2001). The ultimate goal of data clustering is to discover natural groupings in a set of patterns, points, or objects without prior knowledge of any class labels. In fact, in the machine-learning literature, data clustering is typically regarded as a form of unsupervised learning as opposed to supervised learning. In unsupervised learning or clustering, there is no training function as in supervised learning. There are many applications for data clustering including, but not limited to, pattern recognition, data analysis, data compression, image processing, understanding genomic data, and market-basket research.


Author(s):  
Patricia Rich ◽  
Mark Blokpoel ◽  
Ronald de Haan ◽  
Maria Otworowska ◽  
Marieke Sweers ◽  
...  

Abstract Many compelling examples have recently been provided in which people can achieve impressive epistemic success, e.g. draw highly accurate inferences, by using simple heuristics and very little information. This is possible by taking advantage of the features of the environment. The examples suggest an easy and appealing naturalization of rationality: on the one hand, people clearly can apply simple heuristics, and on the other hand, they intuitively ought do so when this brings them high accuracy at little cost.. The ‘ought-can’ principle is satisfied, and rationality is meaningfully normative. We show, however, that this naturalization program is endangered by a computational wrinkle in the adaptation process taken to be responsible for this heuristics-based (‘ecological’) rationality: for the adaptation process to guarantee even minimal rationality, it requires astronomical computational resources, making the problem intractable. We consider various plausible auxiliary assumptions in attempt to remove this obstacle, and show that they do not succeed; intractability is a robust property of adaptation. We discuss the implications of our findings for the project of naturalizing rationality.


2019 ◽  
Vol 487 (2) ◽  
pp. 2701-2717 ◽  
Author(s):  
Ross O’Connell ◽  
Daniel J Eisenstein

Abstract Covariance matrix estimation is a persistent challenge for cosmology. We focus on a class of model covariance matrices that can be generated with high accuracy and precision, using a tiny fraction of the computational resources that would be required to achieve comparably precise covariance matrices using mock catalogues. In previous work, the free parameters in these models were determined using sample covariance matrices computed using a large number of mocks, but we demonstrate that those parameters can be estimated consistently and with good precision by applying jackknife methods to a single survey volume. This enables model covariance matrices that are calibrated from data alone, with no reference to mocks.


2020 ◽  
Vol 1 (2) ◽  
pp. 1-4
Author(s):  
Priyam Guha ◽  
Abhishek Mukherjee ◽  
Abhishek Verma

This research paper deals with using supervised machine learning algorithms to detect authenticity of bank notes. In this research we were successful in achieving very high accuracy (of the order of 99%) by applying some data preprocessing tricks and then running the processed data on supervised learning algorithms like SVM, Decision Trees, Logistic Regression, KNN. We then proceed to analyze the misclassified points. We examine the confusion matrix to find out which algorithms had more number of false positives and which algorithm had more number of False negatives. This research paper deals with using supervised machine learning algorithms to detect authenticity of bank notes. In this research we were successful in achieving very high accuracy (of the order of 99%) by applying some data preprocessing tricks and then running the processed data on supervised learning algorithms like SVM, Decision Trees, Logistic Regression, KNN. We then proceed to analyze the misclassified points. We examine the confusion matrix to find out which algorithms had more number of false positives and which algorithm had more number of False negatives.


2014 ◽  
Vol 490-491 ◽  
pp. 1391-1398
Author(s):  
Xiao Ping Li

This article proposes the necessity and feasibility of the use of Data Mining and Knowledge Discovery in CBR reasoning. This paper focuses on the method of empowering feature items based on least squares method parameter identification, and achieve the method of Similarity case retrieval on this basis, the object is the typical case database of railway rescue. The simulation results show that: the least square method can effectively make estimation and identification of the feature parameters, and can continuously correct on-line. High accuracy and fast convergence characteristics of the assigned parameters show that the algorithm has a certain application value.


AI Magazine ◽  
2020 ◽  
Vol 41 (3) ◽  
pp. 45-62
Author(s):  
Shivali Agarwal ◽  
Jayachandu Bandlamudi ◽  
Atri Mandal ◽  
Anupama Ray ◽  
Giriprasad Sridhara

In this article, we present an end-to-end automated helpdesk email ticket assignment system driven by high accuracy, coverage, business continuity, scalability, and optimal usage of computational resources. The primary objective of the system is to determine the problem mentioned in an incoming email ticket and then automatically dispatch it to an appropriate resolver group with high accuracy. While meeting this objective, it should also meet the objective of being able to operate at desired accuracy levels in the face of changing business needs by automatically adapting to the changes. The proposed system uses a system of classifiers with separate strategies for handling frequent and sparse resolver groups augmented with a semiautomatic rule engine and retraining strategies to ensure that it is accurate, robust, and adaptive to changing business needs. Our system has been deployed in the production of six major service providers in diverse service domains and currently assigns 100,000 emails per month, on an average, with an accuracy close to ninety percent and covering at least ninety percent of email tickets. This translates to achieving human-level accuracy and results in a net savings of more than 50,000 man-hours of effort per annum. To date, our deployed system has already served more than two million tickets in production.


2020 ◽  
Vol 14 (3) ◽  
pp. 329-341
Author(s):  
Yiming Lin ◽  
Daokun Jiang ◽  
Roberto Yus ◽  
Georgios Bouloukakis ◽  
Andrew Chio ◽  
...  

This paper explores the data cleaning challenges that arise in using WiFi connectivity data to locate users to semantic indoor locations such as buildings, regions, rooms. WiFi connectivity data consists of sporadic connections between devices and nearby WiFi access points (APs), each of which may cover a relatively large area within a building. Our system, entitled semantic LOCATion cleanER (LOCATER), postulates semantic localization as a series of data cleaning tasks - first, it treats the problem of determining the AP to which a device is connected between any two of its connection events as a missing value detection and repair problem. It then associates the device with the semantic subregion (e.g., a conference room in the region) by postulating it as a location disambiguation problem. LOCATER uses a bootstrapping semi-supervised learning method for coarse localization and a probabilistic method to achieve finer localization. The paper shows that LOCATER can achieve significantly high accuracy at both the coarse and fine levels.


Sign in / Sign up

Export Citation Format

Share Document