Finding Suitable Membership Functions for Mining Fuzzy Association Rules in Web Data Using Learning Automata

Author(s):  
Zohreh Anari ◽  
Abdolreza Hatamlou ◽  
Babak Anari

Transactions in web data are huge amounts of data, often consisting of fuzzy and quantitative values. Mining fuzzy association rules can help discover interesting relationships between web data. The quality of these rules depends on membership functions, and thus, it is essential to find the suitable number and position of membership functions. The time spent by users on each web page, which shows their level of interest in those web pages, can be considered as a trapezoidal membership function (TMF). In this paper, the optimization problem was finding the appropriate number and position of TMFs for each web page. To solve this optimization problem, a learning automata-based algorithm was proposed to optimize the number and position of TMFs (LA-ONPTMF). Experiments conducted on two real datasets confirmed that the proposed algorithm enhances the efficiency of mining fuzzy association rules by extracting the optimized TMFs.

Author(s):  
Satinder Kaur ◽  
Sunil Gupta

Inform plays a very important role in life and nowadays, the world largely depends on the World Wide Web to obtain any information. Web comprises of a lot of websites of every discipline, whereas websites consists of web pages which are interlinked with each other with the help of hyperlinks. The success of a website largely depends on the design aspects of the web pages. Researchers have done a lot of work to appraise the web pages quantitatively. Keeping in mind the importance of the design aspects of a web page, this paper aims at the design of an automated evaluation tool which evaluate the aspects for any web page. The tool takes the HTML code of the web page as input, and then it extracts and checks the HTML tags for the uniformity. The tool comprises of normalized modules which quantify the measures of design aspects. For realization, the tool has been applied on four web pages of distinct sites and design aspects have been reported for comparison. The tool will have various advantages for web developers who can predict the design quality of web pages and enhance it before and after implementation of website without user interaction.


2015 ◽  
Vol 14 (06) ◽  
pp. 1215-1242 ◽  
Author(s):  
Chun-Hao Chen ◽  
Tzung-Pei Hong ◽  
Yeong-Chyi Lee ◽  
Vincent S. Tseng

Since transactions may contain quantitative values, many approaches have been proposed to derive membership functions for mining fuzzy association rules using genetic algorithms (GAs), a process known as genetic-fuzzy data mining. However, existing approaches assume that the number of linguistic terms is predefined. Thus, this study proposes a genetic-fuzzy mining approach for extracting an appropriate number of linguistic terms and their membership functions used in fuzzy data mining for the given items. The proposed algorithm adjusts membership functions using GAs and then uses them to fuzzify the quantitative transactions. Each individual in the population represents a possible set of membership functions for the items and is divided into two parts, control genes (CGs) and parametric genes (PGs). CGs are encoded into binary strings and used to determine whether membership functions are active. Each set of membership functions for an item is encoded as PGs with real-number schema. In addition, seven fitness functions are proposed, each of which is used to evaluate the goodness of the obtained membership functions and used as the evolutionary criteria in GA. After the GA process terminates, a better set of association rules with a suitable set of membership functions is obtained. Experiments are made to show the effectiveness of the proposed approach.


2015 ◽  
Vol 295 ◽  
pp. 358-378 ◽  
Author(s):  
Ana María Palacios ◽  
José Luis Palacios ◽  
Luciano Sánchez ◽  
Jesús Alcalá-Fdez

Author(s):  
CHUN-HAO CHEN ◽  
TZUNG-PEI HONG ◽  
YEONG-CHYI LEE

Data mining is most commonly used in attempts to induce association rules from transaction data. Since transactions in real-world applications usually consist of quantitative values, many fuzzy association-rule mining approaches have been proposed on single- or multiple-concept levels. However, the given membership functions may have a critical influence on the final mining results. In this paper, we propose a multiple-level genetic-fuzzy mining algorithm for mining membership functions and fuzzy association rules using multiple-concept levels. It first encodes the membership functions of each item class (category) into a chromosome according to the given taxonomy. The fitness value of each individual is then evaluated by the summation of large 1-itemsets of each item in different concept levels and the suitability of membership functions in the chromosome. After the GA process terminates, a better set of multiple-level fuzzy association rules can then be expected with a more suitable set of membership functions. Experimental results on a simulation dataset also show the effectiveness of the algorithm.


Author(s):  
Ily Amalina Ahmad Sabri ◽  
Mustafa Man

<p>Web data extraction is the process of extracting user required information from web page. The information consists of semi-structured data not in structured format. The extraction data involves the web documents in html format. Nowadays, most people uses web data extractors because the extraction involve large information which makes the process of manual information extraction takes time and complicated. We present in this paper WEIDJ approach to extract images from the web, whose goal is to harvest images as object from template-based html pages. The WEIDJ (Web Extraction Image using DOM (Document Object Model) and JSON (JavaScript Object Notation)) applies DOM theory in order to build the structure and JSON as environment of programming. The extraction process leverages both the input of web address and the structure of extraction. Then, WEIDJ splits DOM tree into small subtrees and applies searching algorithm by visual blocks for each web page to find images. Our approach focus on three level of extraction; single web page, multiple web page and the whole web page. Extensive experiments on several biodiversity web pages has been done to show the comparison time performance between image extraction using DOM, JSON and WEIDJ for single web page. The experimental results advocate via our model, WEIDJ image extraction can be done fast and effectively.</p>


2020 ◽  
Vol 9 (1) ◽  
pp. 1937-1942

Right now, the Internet is the most common medium people use to obtain access to knowledge. One's requirements vary from a single word sense to specialized papers and self-education. Revisiting previously visited web pages is one of the most common and unpleasant activities a person does. In fact, a significant primary feedback technique is included in tailoring the individual's memory of vigor and re-examination patterns. Our successful control of environment and object recollections, including decay and reinforcing techniques, may imitate device retention and retention mechanisms. This proposed solution takes into consideration the normal human retrieval method of utilizing episodic and textual memory signals for fast retrieval and includes a different site revisitation mechanism known as the "Web Page Preview" via meaning and information gained through user discovery and web page visits. The underlying methods for the collection, preservation, degradation and usage of meaning and quality of consumer memories for page re-discovery are addressed. Our 3 month user study says that of the time, location, and activity context factors in Web Page Prep, activity is the best recall guide, and context + content based re-discovery provides the highest results relative to context and content individual rediscovery. This concept, if completed, can be used to improve web surfing functionality by providing users a variety of apps with a rather simple UI.


2004 ◽  
Vol 4 (1) ◽  
Author(s):  
David Carabantes Alarcón ◽  
Carmen García Carrión ◽  
Juan Vicente Beneit Montesinos

La calidad en Internet tiene un gran valor, y más aún cuando se trata de una página web sobre salud como es un recurso sobre drogodependencias. El presente artículo recoge los estimadores y sistemas más destacados sobre calidad web para el desarrollo de un sistema específico de valoración de la calidad de recursos web sobre drogodependencias. Se ha realizado una prueba de viabilidad mediante el análisis de las principales páginas web sobre este tema (n=60), recogiendo la valoración, desde el punto de vista del usuario, de la calidad de los recursos. Se han detectado aspectos de mejora en cuanto a la exactitud y fiabilidad de la información, autoría, y desarrollo de descripciones y valoraciones de los enlaces externos. AbstractThe quality in Internet has a great value, and still more when is a web page on health like a resource of drug dependence. This paper contains the estimators and systems on quality in the web for the development of a specific system to value the quality of a web site about drug dependence. A test of viability by means of the analysis of the main web pages has been made on this subject, gathering the valuation from the point of view of the user of the quality of the resources. Aspects of improvement as the exactitude and reliability of the information, responsibility, and development of descriptions and valuations of the external links have been detected.


2011 ◽  
Vol 3 (4) ◽  
pp. 62-70 ◽  
Author(s):  
Stephen O’Neill ◽  
Kevin Curran

Search engine optimization (SEO) is the process of improving the visibility, volume and quality of traffic to website or a web page in search engines via the natural search results. SEO can also target other areas of a search, including image search and local search. SEO is one of many different strategies used for marketing a website but SEO has been proven the most effective. An Internet marketing campaign may drive organic search results to websites or web pages but can be involved with paid advertising on search engines. All search engines have a unique way of ranking the importance of a website. Some search engines focus on the content while others review Meta tags to identify who and what a web site’s business is. Most engines use a combination of Meta tags, content, link popularity, click popularity and longevity to determine a sites ranking. To make it even more complicated, they change their ranking policies frequently. This paper provides an overview of search engine optimisation strategies and pitfalls.


Sign in / Sign up

Export Citation Format

Share Document