SWAMI

Author(s):  
Mark Kilfoil ◽  
Ali Ghorbani

The rapid growth of the World Wide Web has complicated the process of Web browsing by providing an overwhelming wealth of choices for the end user. To alleviate this burden, intelligent tools can do much of the drudge-work of looking ahead, searching and performing a preliminary evaluation of the end pages on the user’s behalf, anticipating the user’s needs and providing the user with more information with which to make fewer, more informed decisions. However, to accomplish this task, the tools need some form of representation of the interests of the user. This article describes the SWAMI system: SWAMI stands for Searching the Web with Agents having Mobility and Intelligence. SWAMI is a prototype that uses a multi-agent system to represent the interests of a user dynamically, and take advantage of the active nature of agents to provide a platform for look-ahead evaluation, page searching, and link swapping. The collection of agents is organized hierarchically according to the apparent interests of the user, which are discovered on-the-fly through multi-stage clustering. Results from initial testing show that such a system is able to follow the multiple changing interests of a user accurately, and that it is capable of acting fruitfully on these interests to provide a user with useful navigational suggestions.

Author(s):  
Mark Kilfoil ◽  
Ali Ghorbani

The rapid growth of the World Wide Web has complicated the process of Web browsing by providing an overwhelming wealth of choices for the end user. To alleviate this burden, intelligent tools can do much of the drudge-work of looking ahead, searching and performing a preliminary evaluation of the end pages on the user’s behalf, anticipating the user’s needs and providing the user with more information with which to make fewer, more informed decisions. However, to accomplish this task, the tools need some form of representation of the interests of the user. This article describes the SWAMI system: SWAMI stands for Searching the Web with Agents having Mobility and Intelligence. SWAMI is a prototype that uses a multi-agent system to represent the interests of a user dynamically, and take advantage of the active nature of agents to provide a platform for look-ahead evaluation, page searching, and link swapping. The collection of agents is organized hierarchically according to the apparent interests of the user, which are discovered on-the-fly through multistage clustering. Results from initial testing show that such a system is able to follow the multiple changing interests of a user accurately, and that it is capable of acting fruitfully on these interests to provide a user with useful navigational suggestions.


Author(s):  
R. Ryan Nelson ◽  
Peter Todd

Beginning in the early 1980s, end-user computing (EUC) began to permeate organizations following the advent of the personal computer and a host of applications directed at the non-IS professional. Along with EUC came a whole new set of organizational opportunities and risks. Ten years later, the World Wide Web has opened the door to a yet more powerful set of EUC applications capable of reaching well beyond the boundaries of the organization. Indeed, Web technology permits end users to design applications that are immediately accessible by unlimited numbers of people from anywhere in the world. As a result, EUC using Web technology has introduced a whole new set of opportunities and risks for organizations. The purpose of this research is to examine what strategies organizations are using in their attempt to maximize the benefits of the Web for end users while mitigating the inherent risks. To this end, individuals from 12 major organizations were surveyed via the Web. The results indicate that while organizations seem to be doing an adequate job of establishing roles and standards, mechanisms for resource allocation, development management, and maintenance appear to be lacking. In fact, most firms seem to be relying on a monopolist control strategy at this point in time. While such a strategy may be the best approach given the relative infancy of Web technology, it could prove to be an unstable strategy in the long run given the reach, range and flexibility of access that Web technology provides. Organizations are encouraged to take a proactive, formal posture toward EUC development on the Web.


Author(s):  
Anthony D. Andre

This paper provides an overview of the various human factors and ergonomics (HF/E) resources on the World Wide Web (WWW). A list of the most popular and useful HF/E sites will be provided, along with several critical guidelines relevant to using the WWW. The reader will gain a clear understanding of how to find HF/E information on the Web and how to successfully use the Web towards various HF/E professional consulting activities. Finally, we consider the ergonomic implications of surfing the Web.


2016 ◽  
Vol 28 (2) ◽  
pp. 241-251 ◽  
Author(s):  
Luciane Lena Pessanha Monteiro ◽  
Mark Douglas de Azevedo Jacyntho

The study addresses the use of the Semantic Web and Linked Data principles proposed by the World Wide Web Consortium for the development of Web application for semantic management of scanned documents. The main goal is to record scanned documents describing them in a way the machine is able to understand and process them, filtering content and assisting us in searching for such documents when a decision-making process is in course. To this end, machine-understandable metadata, created through the use of reference Linked Data ontologies, are associated to documents, creating a knowledge base. To further enrich the process, (semi)automatic mashup of these metadata with data from the new Web of Linked Data is carried out, considerably increasing the scope of the knowledge base and enabling to extract new data related to the content of stored documents from the Web and combine them, without the user making any effort or perceiving the complexity of the whole process.


2018 ◽  
Vol 31 (5) ◽  
pp. 154-182
Author(s):  
Cadence Kinsey

This article analyses Camille Henrot’s 2013 film Grosse Fatigue in relation to the histories of hypermedia and modes of interaction with the World Wide Web. It considers the development of non-hierarchical systems for the organisation of information, and uses Grosse Fatigue to draw comparisons between the Web, the natural history museum and the archive. At stake in focusing on the way in which information is organised through hypermedia is the question of subjectivity, and this article argues that such systems are made ‘user-friendly’ by appearing to accommodate intuitive processes of information retrieval, reflecting the subject back to itself as autonomous. This produces an ideology of individualism which belies the forms of heteronomy that in fact shape and structure access to information online in significant ways. At the heart of this argument is an attention to the visual, and the significance of art as an immanent mode of analysis. Through the themes of transparency and opacity, and order and chaos, the article thus proposes a defining dynamic between autonomy and automation as a model for understanding the contemporary subject.


2017 ◽  
Vol 4 (1) ◽  
pp. 95-110 ◽  
Author(s):  
Deepika Punj ◽  
Ashutosh Dixit

In order to manage the vast information available on web, crawler plays a significant role. The working of crawler should be optimized to get maximum and unique information from the World Wide Web. In this paper, architecture of migrating crawler is proposed which is based on URL ordering, URL scheduling and document redundancy elimination mechanism. The proposed ordering technique is based on URL structure, which plays a crucial role in utilizing the web efficiently. Scheduling ensures that URLs should go to optimum agent for downloading. To ensure this, characteristics of both agents and URLs are taken into consideration for scheduling. Duplicate documents are also removed to make the database unique. To reduce matching time, document matching is made on the basis of their Meta information only. The agents of proposed migrating crawler work more efficiently than traditional single crawler by providing ordering and scheduling of URLs.


2021 ◽  
Author(s):  
Michael Dick

Since it was first formally proposed in 1990 (and since the first website was launched in 1991), the World Wide Web has evolved from a collection of linked hypertext documents residing on the Internet, to a "meta-medium" featuring platforms that older media have leveraged to reach their publics through alternative means. However, this pathway towards the modernization of the Web has not been entirely linear, nor will it proceed as such. Accordingly, this paper problematizes the notion of "progress" as it relates to the online realm by illuminating two distinct perspectives on the realized and proposed evolution of the Web, both of which can be grounded in the broader debate concerning technological determinism versus the social construction of technology: on the one hand, the centralized and ontology-driven shift from a human-centred "Web of Documents" to a machine-understandable "Web of Data" or "Semantic Web", which is supported by the Web's inventor, Tim Berners-Lee, and the organization he heads, the World Wide Web Consortium (W3C); on the other, the decentralized and folksonomy-driven mechanisms through which individuals and collectives exert control over the online environment (e.g. through the social networking applications that have come to characterize the contemporary period of "Web 2.0"). Methodologically, the above is accomplished through a sustained exploration of theory derived from communication and cultural studies, which discursively weaves these two viewpoints together with a technical history of recent W3C projects. As a case study, it is asserted that the forward slashes contained in a Uniform Resource Identifier (URI) were a social construct that was eventually rendered extraneous by the end-user community. By focusing On the context of the technology itself, it is anticipated that this paper will contribute to the broader debate concerning the future of the Web and its need to move beyond a determinant "modernization paradigm" or over-arching ontology, as well as advance the potential connections that can be cultivated with cognate disciplines.


Author(s):  
Punam Bedi ◽  
Neha Gupta ◽  
Vinita Jindal

The World Wide Web is a part of the Internet that provides data dissemination facility to people. The contents of the Web are crawled and indexed by search engines so that they can be retrieved, ranked, and displayed as a result of users' search queries. These contents that can be easily retrieved using Web browsers and search engines comprise the Surface Web. All information that cannot be crawled by search engines' crawlers falls under Deep Web. Deep Web content never appears in the results displayed by search engines. Though this part of the Web remains hidden, it can be reached using targeted search over normal Web browsers. Unlike Deep Web, there exists a portion of the World Wide Web that cannot be accessed without special software. This is known as the Dark Web. This chapter describes how the Dark Web differs from the Deep Web and elaborates on the commonly used software to enter the Dark Web. It highlights the illegitimate and legitimate sides of the Dark Web and specifies the role played by cryptocurrencies in the expansion of Dark Web's user base.


Web Services ◽  
2019 ◽  
pp. 1068-1076
Author(s):  
Vudattu Kiran Kumar

The World Wide Web (WWW) is global information medium, where users can read and write using computers over internet. Web is one of the services available on internet. The Web was created in 1989 by Sir Tim Berners-Lee. Since then a great refinement has done in the web usage and development of its applications. Semantic Web Technologies enable machines to interpret data published in a machine-interpretable form on the web. Semantic web is not a separate web it is an extension to the current web with additional semantics. Semantic technologies play a crucial role to provide data understandable to machines. To achieve machine understandable, we should add semantics to existing websites. With additional semantics, we can achieve next level web where knowledge repositories are available for better understanding of web data. This facilitates better search, accurate filtering and intelligent retrieval of data. This paper discusses about the Semantic Web and languages involved in describing documents in machine understandable format.


Author(s):  
K.G. Srinivasa ◽  
Anil Kumar Muppalla ◽  
Varun A. Bharghava ◽  
M. Amulya

In this paper, the authors discuss the MapReduce implementation of crawler, indexer and ranking algorithms in search engines. The proposed algorithms are used in search engines to retrieve results from the World Wide Web. A crawler and an indexer in a MapReduce environment are used to improve the speed of crawling and indexing. The proposed ranking algorithm is an iterative method that makes use of the link structure of the Web and is developed using MapReduce framework to improve the speed of convergence of ranking the WebPages. Categorization is used to retrieve and order the results according to the user choice to personalize the search. A new score is introduced in this paper that is associated with each WebPage and is calculated using user’s query and number of occurrences of the terms in the query in the document corpus. The experiments are conducted on Web graph datasets and the results are compared with the serial versions of crawler, indexer and ranking algorithms.


Sign in / Sign up

Export Citation Format

Share Document