scholarly journals Improving Salience Retention and  Identification in the Automated Filtering  of Event Log Messages

2021 ◽  
Author(s):  
◽  
Paul Radford

<p>Event log messages are currently the only genuine interface through which computer systems administrators can effectively monitor their systems and assemble a mental perception of system state. The popularisation of the Internet and the accompanying meteoric growth of business-critical systems has resulted in an overwhelming volume of event log messages, channeled through mechanisms whose designers could not have envisaged the scale of the problem. Messages regarding intrusion detection, hardware status, operating system status changes, database tablespaces, and so on, are being produced at the rate of many gigabytes per day for a significant computing environment. Filtering technologies have not been able to keep up. Most messages go unnoticed; no  filtering whatsoever is performed on them, at least in part due to the difficulty of implementing and maintaining an effective filtering solution. The most commonly-deployed  filtering alternatives rely on regular expressions to match pre-defi ned strings, with 100% accuracy, which can then become ineffective as the code base for the software producing the messages 'drifts' away from those strings. The exactness requirement means all possible failure scenarios must be accurately anticipated and their events catered for with regular expressions, in order to make full use of this technique. Alternatives to regular expressions remain largely academic. Data mining, automated corpus construction, and neural networks, to name the highest-profi le ones, only produce probabilistic results and are either difficult or impossible to alter in any deterministic way. Policies are therefore not supported under these alternatives. This thesis explores a new architecture which utilises rich metadata in order to avoid the burden of message interpretation. The metadata itself is based on an intention to improve end-to-end communication and reduce ambiguity. A simple yet effective filtering scheme is also presented which fi lters log messages through a short and easily-customisable set of rules. With such an architecture, it is envisaged that systems administrators could signi ficantly improve their awareness of their systems while avoiding many of the false-positives and -negatives which plague today's fi ltering solutions.</p>

2021 ◽  
Author(s):  
◽  
Paul Radford

<p>Event log messages are currently the only genuine interface through which computer systems administrators can effectively monitor their systems and assemble a mental perception of system state. The popularisation of the Internet and the accompanying meteoric growth of business-critical systems has resulted in an overwhelming volume of event log messages, channeled through mechanisms whose designers could not have envisaged the scale of the problem. Messages regarding intrusion detection, hardware status, operating system status changes, database tablespaces, and so on, are being produced at the rate of many gigabytes per day for a significant computing environment. Filtering technologies have not been able to keep up. Most messages go unnoticed; no  filtering whatsoever is performed on them, at least in part due to the difficulty of implementing and maintaining an effective filtering solution. The most commonly-deployed  filtering alternatives rely on regular expressions to match pre-defi ned strings, with 100% accuracy, which can then become ineffective as the code base for the software producing the messages 'drifts' away from those strings. The exactness requirement means all possible failure scenarios must be accurately anticipated and their events catered for with regular expressions, in order to make full use of this technique. Alternatives to regular expressions remain largely academic. Data mining, automated corpus construction, and neural networks, to name the highest-profi le ones, only produce probabilistic results and are either difficult or impossible to alter in any deterministic way. Policies are therefore not supported under these alternatives. This thesis explores a new architecture which utilises rich metadata in order to avoid the burden of message interpretation. The metadata itself is based on an intention to improve end-to-end communication and reduce ambiguity. A simple yet effective filtering scheme is also presented which fi lters log messages through a short and easily-customisable set of rules. With such an architecture, it is envisaged that systems administrators could signi ficantly improve their awareness of their systems while avoiding many of the false-positives and -negatives which plague today's fi ltering solutions.</p>


Author(s):  
Shereen Yousef Mohamed ◽  
◽  
Mohamed Hamed N. Taha ◽  
Hesham N. Elmahdy ◽  
Hany Harb ◽  
...  

Cloud computing refers to the services and applications that are accessible throughout the world from data centers. All services and applications are available online. Virtual machine migration is an important part of virtualization which is considered as essential part in cloud computing environment. Virtual Machine Migration means transferring a running Virtual Machine with all its applications and the operating system state as it is to target destination machine where it continues to run as if nothing happened. It makes balancing between servers. This improves the performance by redistributing the workload among available servers. There are many algorithms of load balancing classified into two types: static load balancing algorithms and dynamic load balancing algorithms. This paper presents the algorithm (Balanced Throttled Load Balancing Algorithm- BTLB). It compares the results of the BTLB with round robin algorithm, AMLB algorithm and throttled load balancing algorithm. The results of these four algorithms would be presented in this paper. The proposed algorithm shows the improvement in response time (75 µs). Cloud analyst simulator is used to evaluate the results. BTLB was developed and tested using Java.


2017 ◽  
Vol 2017 (3) ◽  
pp. 21-38 ◽  
Author(s):  
Hung Dang ◽  
Tien Tuan Anh Dinh ◽  
Ee-Chien Chang ◽  
Beng Chin Ooi

Abstract We consider privacy-preserving computation of big data using trusted computing primitives with limited private memory. Simply ensuring that the data remains encrypted outside the trusted computing environment is insufficient to preserve data privacy, for data movement observed during computation could leak information. While it is possible to thwart such leakage using generic solution such as ORAM [42], designing efficient privacy-preserving algorithms is challenging. Besides computation efficiency, it is critical to keep trusted code bases lean, for large ones are unwieldy to vet and verify. In this paper, we advocate a simple approach wherein many basic algorithms (e.g., sorting) can be made privacy-preserving by adding a step that securely scrambles the data before feeding it to the original algorithms. We call this approach Scramble-then-Compute (StC), and give a sufficient condition whereby existing external memory algorithms can be made privacy-preserving via StC. This approach facilitates code-reuse, and its simplicity contributes to a smaller trusted code base. It is also general, allowing algorithm designers to leverage an extensive body of known efficient algorithms for better performance. Our experiments show that StC could offer up to 4.1× speedups over known, application-specific alternatives.


Corpora ◽  
2020 ◽  
Vol 15 (2) ◽  
pp. 125-140
Author(s):  
Yukiko Ohashi ◽  
Noriaki Katagiri ◽  
Katsutoshi Oka ◽  
Michiko Hanada

This paper reports on two research results: ( 1) designing an English for Specific Purposes (esp) corpus architecture complete with annotations structured by regular expressions; and ( 2) a case study to test the design to cater for creating a specific vocabulary list using the compiled corpus. The first half of this study involved designing a precisely structured esp corpus from 190 veterinary medical charts with a hierarchy of the data. The data hierarchy in the corpus consists of document types, outline elements and inline elements, such as species and breed. Perl scripts extracted the data attached to veterinary-specific categories, and the extraction led to creating wordlists. The second part of the research tested the corpus mode, creating a list of commonly observed lexical items in veterinary medicine. The coverage rate of the wordlists by General Service List (gsl) and Academic Word List (awl) was tested, with the result that 66.4 percent of all lexical items appeared in gsl and awl, whereas 33.7 percent appeared in none of those lists. The corpus compilation procedures as well as the annotation scheme introduced in this study enable the compilation of specific corpora with explicit annotations, allowing teachers to have access to data required for creating esp classroom materials.


2019 ◽  
Vol 2 (3) ◽  
pp. 216-229
Author(s):  
Vasily Larshin ◽  
Natalia Lishchenko

2019 ◽  
Vol 13 ◽  
pp. 57-79
Author(s):  
Tetsuro KAKESHITA ◽  
Mika OHTSUKI

We conducted the first national survey of computing education at Japanese universities in 2016. In this paper, we report the survey result of the computing education at non-IT departments and faculties whose major subject is not computing. The survey covers various aspects of computing education including program organization, quality and quantity of educational achievement, students, teaching staff and computing environment. We collected 994 answers through the survey. At least 87,000 non-ICT students are taking computing education in Japan. Although computing education is carried out at every major academic discipline, teaching effort greatly differs depending on the academic discipline. We also find shortage of teaching staff for computing education. The analysis result will be an essential input to develop reasonable curriculum guidelines and accreditation criteria to improve computing education at non-IT departments.


Sign in / Sign up

Export Citation Format

Share Document