web graph
Recently Published Documents


TOTAL DOCUMENTS

142
(FIVE YEARS 4)

H-INDEX

17
(FIVE YEARS 0)

Author(s):  
Khoi Duy Vo ◽  
Sergej Zerr ◽  
Xiaofei Zhu ◽  
Wolfgang Nejdl
Keyword(s):  


PLoS ONE ◽  
2021 ◽  
Vol 16 (8) ◽  
pp. e0255908
Author(s):  
Xiaorong Guo ◽  
Benhua Zhao ◽  
Tianmu Chen ◽  
Bin Hao ◽  
Tao Yang ◽  
...  

This study aimed to investigate the spatial distribution and patterns of multimorbidity among the elderly in China. Data on the occurrence of 14 chronic diseases were collected for 9710 elderly participants in the 2015 waves of the China Health and Retirement Longitudinal Study (CHARLS). Web graph, Apriori algorithm, age-adjusted Charlson comorbidity index (AAC), and Spatial autocorrelation were used to perform the multimorbidity analysis. The multimorbidity prevalence rate was estimated as 49.64% in the elderly in China. Three major multimorbidity patterns were identified: [Asthma/Chronic lungs diseases]: (Support (S) = 6.17%, Confidence (C) = 63.77%, Lift (L) = 5.15); [Asthma, Arthritis, or rheumatism/ Chronic lungs diseases]: (S = 3.12%, C = 64.03%, L = 5.17); [Dyslipidemia, Hypertension, Arthritis or rheumatism/Heart attack]: (S = 3.96%, C = 51.56, L = 2.69). Results of the AAC analysis showed that the more chronic diseases an elderly has, the lower is the 10-year survival rate (P < 0.001). Global spatial autocorrelation showed a positive spatial correlation distribution for the prevalence of the third multimorbidity pattern in China (P = 0.032). The status of chronic diseases and multimorbidity among the elderly with a spatial correlation is a significant health issue in China.



2021 ◽  
Vol 3 (4 (111)) ◽  
pp. 24-31
Author(s):  
Natalia Guk ◽  
Olga Verba ◽  
Vladyslav Yevlakov

A recommendation system has been built for a web resource’s users that applies statistics about user activities to provide recommendations. The purpose of the system operation is to provide recommendations in the form of an orderly sequence of HTML pages of the resource suggested for the user. The ranking procedure uses statistical information about user transitions between web resource pages. The web resource model is represented in the form of a web graph; the user behavior model is shown as a graph of transitions between resource pages. The web graph is represented by an adjacency matrix; for the transition graph, a weighted matrix of probabilities of transitions between the vertices of the graph has been constructed. It was taken into consideration that user transitions between pages of a web resource may involve entering a URL in the address bar of a browser or by clicking on a link in the current page. The user’s transition between vertices in a finite graph according to probabilities determined by the weight of the graph’s edges is represented by a homogeneous Markov chain and is considered a process of random walk on the graph with the possibility of moving to a random vertex. Random Walk with Restarts was used to rank web resource pages for a particular user. Numerical analysis has been performed for an actual online store website. The initial data on user sessions are divided into training and test samples. According to the training sample, a weighted matrix of the probability of user transitions between web resource pages was constructed. To assess the quality of the built recommendation system, the accuracy, completeness, and Half-life Utility metrics were used. On the elements of the test sample, the accuracy value of 65‒68 % was obtained, the optimal number of elements in the recommendation list was determined. The influence of model parameters on the quality of recommendation systems was investigated.



Author(s):  
Vadim Zverovich

This chapter gives a brief overview of selected applications of graph theory, many of which gave rise to the development of graph theory itself. A range of such applications extends from puzzles and games to serious scientific and real-life problems, thus illustrating the diversity of applications. The first section is devoted to the six earliest applications of graph theory. The next section introduces so-called scale-free networks, which include the web graph, social and biological networks. The last section describes a number of graph-theoretic algorithms, which can be used to tackle a number of interesting applications and problems of graph theory.



2020 ◽  
Author(s):  
Michelangelo Ceci ◽  
Pasqua Fabiana Lanotte

Abstract A sitemap represents an explicit specification of the design concept and knowledge organization of a website and is therefore considered as the website’s basic ontology. It not only presents the main usage flows for users, but also hierarchically organizes concepts of the website. Typically, sitemaps are defined by webmasters in the very early stages of the website design. However, during their life websites significantly change their structure, their content and their possible navigation paths. Even if this is not the case, webmasters can fail to either define sitemaps that reflect the actual website content or, vice versa, to define the actual organization of pages and links which do not reflect the intended organization of the content coded in the sitemaps. In this paper we propose an approach which automatically generates sitemaps. Contrary to other approaches proposed in the literature, which mainly generate sitemaps from the textual content of the pages, in this work sitemaps are generated by analyzing the Web graph of a website. This allows us to: i) automatically generate a sitemap on the basis of possible navigation paths, ii) compare the generated sitemaps with either the sitemap provided by the Web designer or with the intended sitemap of the website and, consequently, iii) plan possible website re-organization. The solution we propose is based on closed frequent sequence extraction and only concentrates on hyperlinks organized in “Web lists”, which are logical lists embedded in the pages. These “Web lists” are typically used for supporting users in Web site navigation and they include menus, navbars and content tables. Experiments performed on three real datasets show that the extracted sitemaps are much more similar to those defined by website curators than those obtained by competitor algorithms.



2020 ◽  
Vol 23 (3) ◽  
pp. 526-542
Author(s):  
Andrey Anatolievich Pechnikov

The web graph is the most popular model of real Web fragments used in Web science. The study of communities in the web graph contributes to a better understanding of the organization of the fragment of the Web and the processes occurring in it. It is proposed to allocate a communication graph in a web graph containing only those vertices (and arcs between them) that have counter arcs, and in it to investigate the problem of splitting into communities. By analogy with social studies, connections realized through edges in a communication graph are proposed to be called "strong" and all others "weak". Thematic communities with meaningful interpretations are built on strong connections. At the same time, weak links facilitate communication between sites that do not have common features in the field of activity, geography, subordination, etc., and basically preserve the coherence of the fragments of the Web even in the absence of strong links. Experiments conducted for a fragment of the scientific and educational Web of Russia show the possibility of meaningful interpretation of the results and the prospects of such an approach.





Author(s):  
N. A. Guk ◽  
S. V. Dykhanov ◽  
S. F. Syryk

A technique for analyzing the structure of a website based on data on hypertext links between pages is proposed. An analysis method based on the topology of links between pages was selected. The mathematical model of the website in the form of a web graph is developed. Structural relationships between pages are represented by binary values in the graph adjacency matrix. The problem of clustering is formulated. To analyze the structure of the web graph the clustering method k-means is used. A metric for determining the distance between cluster elements has been introduced. Assessment of the complexity of the algorithm is performed. Website pages correspond to hierarchical units of the structure. The structure distinguishes between pages of categories and subcategories of goods, pages of goods, and thematic articles. Types of site pages are selected as clusters. Typical pages for each cluster are selected as centroids. An iterative algorithm for constructing a web graph has been developed. The queue is selected as the data structure for storing local information when crawling pages. Testing of the proposed approach is carried out on the example of an existing online store. A division of the site pages into clusters was obtained as a result of the analysis. A division is corresponded to hierarchical elements of the structure: product categories, subcategories, product pages. The type of pages that are poorly identified by the algorithm is revealed. Using the results of clustering, you can improve the site structure during reengineering. Application of the developed methodology will improve the indexing of the site in the search engine.



2019 ◽  
Vol 9 (4) ◽  
pp. 36-49
Author(s):  
Vasantha Thangasamy

Information available on the internet is wide, diverse, and dynamic. Since an enormous amount of information is available online, finding similarity between webpages using efficient hyperlink analysis is a challenging task. In this article, the researcher proposes an improved PageSim algorithm which measurse the importance of a webpage based on the PageRank values of connected webpage. Therefore, the proposed algorithm uses heterogeneous propagation of the PageRank score, based on the prestige measure of each webpage. The existing and the improved PageSim algorithms are implemented with a sample web graph. Real time Citation Networks, namely the ZEWAIL Citation Network and the DBLP Citation Network are used to test and compare the existing and improved PageSim algorithms. By using this proposed algorithm, it has been found that a similarity score between two different webpages significantly increases based on common information features and significantly decreases based on distinct factors.



Sign in / Sign up

Export Citation Format

Share Document