Building and Evaluating Adaptive Web Pages

2013 ◽  
Vol 12 (01) ◽  
pp. 1350006 ◽  
Author(s):  
M. Koutb ◽  
H. Kelash ◽  
M. Abu Al-Khair

Each year the number of consumers and the variety of their interests increase. As a result, providers are seeking ways to infer the customer's interests and to adapt their websites to make the content of interest more easily accessible. Assume that past navigation behaviour as an indicator of the user's interests. Then, the records of this behaviour, kept in the web-server logs, can be mined to extract the user's interests. On this principal, recommendations can be generated, to help old and new website's visitors to find the information about their interest faster.

2012 ◽  
Vol 3 (1) ◽  
pp. 30
Author(s):  
Mona M. Abu Al-Khair ◽  
M. Koutb ◽  
H. Kelash

Each year the number of consumers and the variety of their interests increase. As a result, providers are seeking ways to infer the customer's interests and to adapt their websites to make the content of interest more easily accessible. Assume that past navigation behavior as an indicator of the user's interests. Then, the records of this behavior, kept in the web-server logs, can be mined to extract the user's interests. On this principal, recommendations can be generated, to help old and new website's visitors to find the information about their interest faster.


2021 ◽  
Author(s):  
Ramon Abilio ◽  
Cristiano Garcia ◽  
Victor Fernandes

Browsing on Internet is part of the world population’s daily routine. The number of web pages is increasing and so is the amount of published content (news, tutorials, images, videos) provided by them. Search engines use web robots to index web contents and to offer better results to their users. However, web robots have also been used for exploiting vulnerabilities in web pages. Thus, monitoring and detecting web robots’ accesses is important in order to keep the web server as safe as possible. Data Mining methods have been applied to web server logs (used as data source) in order to detect web robots. Then, the main objective of this work was to observe evidences of definition or use of web robots detection by analyzing web server-side logs using Data Mining methods. Thus, we conducted a systematic Literature mapping, analyzing papers published between 2013 and 2020. In the systematic mapping, we analyzed 34 studies and they allowed us to better understand the area of web robots detection, mapping what is being done, the data used to perform web robots detection, the tools, and algorithms used in the Literature. From those studies, we extracted 33 machine learning algorithms, 64 features, and 13 tools. This study is helpful for researchers to find machine learning algorithms, features, and tools to detect web robots by analyzing web server logs.


Author(s):  
Yijun Gao

This study analyzed the Web server logs from the People's Daily Online and revealed some interesting findings: Pageview numbers of the mportant news in editors’ mind on the most obvious sections of the homepage, are not significantly different than those of the "common" news put on the less obvious sections.Cette étude a porté sur l'analyse des fichiers de journalisation de serveurs Web du Quotidien du Peuple en ligne et a révélé quelques données intéressantes : le nombre de pages vues pour les dépêches jugées importantes par la rédaction et placées en évidence de la page d'accueil n'est pas significativement différent du nombre de pages vues pour les dépêches plus « courantes » placés moins en évidence. 


2012 ◽  
Vol 37 (3) ◽  
pp. 1-5 ◽  
Author(s):  
K. Sudheer Reddy ◽  
G. Partha Saradhi Varma ◽  
I. Ramesh Babu

2021 ◽  
Vol ahead-of-print (ahead-of-print) ◽  
Author(s):  
Husna Sarirah Husin ◽  
James Thom ◽  
Xiuzhen Zhang

Purpose The purpose of the study is to use web serer logs in analyzing the changes of user behavior in reading online news, in terms of desktop and mobile users. Advances in mobile technology and social media have paved the way for online news consumption to evolve. There is an absence of research into the changes of user behavior in terms of desktop versus mobile users, particularly by analyzing the server logs. Design/methodology/approach In this paper, the authors investigate the evolution of user behavior using logs from the Malaysian newspaper Berita Harian Online in April 2012 and April 2017. Web usage mining techniques were used for pre-processing the logs and identifying user sessions. A Markov model is used to analyze navigation flows, and association rule mining is used to analyze user behavior within sessions. Findings It was found that page accesses have increased tremendously, particularly from Android phones, and about half of the requests in 2017 are referred from Facebook. Navigation flow between the main page, articles and section pages has changed from 2012 to 2017; while most users started navigation with the main page in 2012, readers often started with an article in 2017. Based on association rules, National and Sports are the most frequent section pages in 2012 and 2017 for desktop and mobile. However, based on the lift and conviction, these two sections are not read together in the same session as frequently as might be expected. Other less popular items have higher probability of being read together in a session. Research limitations/implications The localized data set is from Berita Harian Online; although unique to this particular newspaper, the findings and the methodology for investigating user behavior can be applied to other online news. On another note, the data set could be extended to be more than a month. Although initially data for the year 2012 was collected, unfortunately only the data for April 2012 is complete. Other months have missing days. Therefore, to make an impartial comparison for the evolution of user behavior in five years, the Web server logs for April 2017 were used. Originality/value The user behavior in 2012 and 2017 was compared using association rules and Markov flow. Different from existing studies analyzing online newspaper Web server logs, this paper uniquely investigates changes in user behavior as a result of mobile phones becoming a mainstream technology for accessing the Web.


2016 ◽  
Vol 1 (1) ◽  
pp. 001
Author(s):  
Harry Setya Hadi

String searching is a common process in the processes that made the computer because the text is the main form of data storage. Boyer-Moore is the search string from right to left is considered the most efficient methods in practice, and matching string from the specified direction specifically an algorithm that has the best results theoretically. A system that is connected to a computer network that literally pick a web server that is accessed by multiple users in different parts of both good and bad aim. Any activity performed by the user, will be stored in Web server logs. With a log report contained in the web server can help a web server administrator to search the web request error. Web server log is a record of the activities of a web site that contains the data associated with the IP address, time of access, the page is opened, activities, and access methods. The amount of data contained in the resulting log is a log shed useful information.


2018 ◽  
Vol 7 (3.6) ◽  
pp. 106
Author(s):  
B J. Santhosh Kumar ◽  
Kankanala Pujitha

Application uses URL as contribution for Web Application Vulnerabilities recognition. if the length of URL is too long then it will consume more time to scan the URL (Ain Zubaidah et.al 2014).Existing system can notice the web pages but not overall web application. This application will test for URL of any length using String matching algorithm. To avoid XSS and CSRF and detect attacks that try to sidestep program upheld arrangements by white list and DOM sandboxing techniques (Elias Athanasopoulos et.al.2012). The web application incorporates a rundown of cryptographic hashes of legitimate (trusted) client side contents. In the event that there is a cryptographic hash for the content in the white list. On the off chance that the hash is discovered the content is viewed as trusted or not trusted. This application makes utilization of SHA-1 for making a message process. The web server stores reliable scripts inside div or span HTML components that are attribute as reliable. DOM sandboxing helps in identifying the script or code. Partitioning Program Symbols into Code and Non-code. This helps to identify any hidden code in trusted tag, which bypass web server. Scanning the website for detecting the injection locations and injecting the mischievous XSS assault vectors in such infusion focuses and check for these assaults in the helpless web application( Shashank Gupta et.al 2015).The proposed application improve the false negative rate.  


Sign in / Sign up

Export Citation Format

Share Document