scholarly journals Improved Query Processing in Web Search Engines Using Grey Wolf Algorithm

2018 ◽  
Vol 7 (2.24) ◽  
pp. 353
Author(s):  
Nishant Pal ◽  
Akshat Chawla ◽  
A Meena Priyadharsini

In Information systems working at a large scale where retrieval of information is an essential operation for example search engines etc. The users are not only concerned with the quality of results but also the time they consume for querying the data. These aspects lead to a natural tradeoff in which the approaches that lead to an increase in data have a similar larger response time and vice-versa. Hence, as the requirement for faster search query processing time along with efficient results is increasing, we need to identify other ways for increasing efficiency. This work proposes an application of the meta-heuristic algorithm called Grey Wolf Optimization (GWO) algorithm to improve Query Processing Time in Search Engines. The GWO algorithm is an alter ego of the way in which the grey wolves are organised and their hunting techniques. There are four categories of  grey wolves in a single pack of grey wolves which are alpha, beta, delta, and omega respectively. They are used to work in a simulating hierarchy. These help achieve better search results at decrease query response timings.

Abakós ◽  
2012 ◽  
Vol 1 (1) ◽  
pp. 28-49
Author(s):  
Kaio Wagner ◽  
Edleno Silva de Moura ◽  
David Fernandes ◽  
Marco Cristo ◽  
Altigran Soares da Silva

Previous work in literature has indicated that template of web pages represent noisy information in web collections, and advocate that the simple removal of template result in improvements in quality of results provided by Web search systems. In this paper, we study the impact of template removal in two distinct scenarios: large scale web search collections, which consist of several distinct websites, and intrasite web collections, involving searches inside of web sites.  Our  work  is the  first  in literature to  study the  impact of template removal  to  search systems in large  scale  Web  collections. The study was carried out using an automatic template detection method previously proposed by us. As contributions, we present statistics about the application of this automatic template detection method to the well known GOV2 reference collection, a large scale Web collection. We also present experiments comparing the amount of template detected by our automatic method to the ones obtained when humans select templates. And finally, experiments which indicate that, in both experimented scenarios, template removal does not improve the quality of results provided by search systems, but can play the role of an effective loss compression method by reducing the size of their indexes.


2021 ◽  
pp. 089443932110068
Author(s):  
Aleksandra Urman ◽  
Mykola Makhortykh ◽  
Roberto Ulloa

We examine how six search engines filter and rank information in relation to the queries on the U.S. 2020 presidential primary elections under the default—that is nonpersonalized—conditions. For that, we utilize an algorithmic auditing methodology that uses virtual agents to conduct large-scale analysis of algorithmic information curation in a controlled environment. Specifically, we look at the text search results for “us elections,” “donald trump,” “joe biden,” “bernie sanders” queries on Google, Baidu, Bing, DuckDuckGo, Yahoo, and Yandex, during the 2020 primaries. Our findings indicate substantial differences in the search results between search engines and multiple discrepancies within the results generated for different agents using the same search engine. It highlights that whether users see certain information is decided by chance due to the inherent randomization of search results. We also find that some search engines prioritize different categories of information sources with respect to specific candidates. These observations demonstrate that algorithmic curation of political information can create information inequalities between the search engine users even under nonpersonalized conditions. Such inequalities are particularly troubling considering that search results are highly trusted by the public and can shift the opinions of undecided voters as demonstrated by previous research.


2011 ◽  
pp. 762-784 ◽  
Author(s):  
Le-Shin Wu ◽  
Ruj Akavipat ◽  
Ana Gabriela Maguitman ◽  
Filippo Menczer

This chapter proposed a collaborative peer network application called 6Search (6S) to address the scalability limitations of centralized search engines. Each peer crawls the Web in a focused way, guided by its user’s information context. Through this approach, better (distributed) coverage can be achieved. Each peer also acts as a search “servent” (server + client) by submitting and responding to queries to/from its neighbors. This search process has no centralized bottleneck. Peers depend on a local adaptive routing algorithm to dynamically change the topology of the peer network and search for the best neighbors to answer their queries. We present and evaluate learning techniques to improve local query routing. We validate prototypes of the 6S network via simulations with model users based on actual Web crawls. We find that the network topology rapidly converges from a random network to a small world network, with clusters emerging from user communities with shared interests. We finally compare the quality of the results with those obtained by centralized search engines such as Google.


Author(s):  
Le-Shin Wu ◽  
Ruj Akavipat ◽  
Ana Gabriela Maguitman ◽  
Filippo Menczer

This chapter proposed a collaborative peer network application called 6Search (6S) to address the scalability limitations of centralized search engines. Each peer crawls the Web in a focused way, guided by its user’s information context. Through this approach, better (distributed) coverage can be achieved. Each peer also acts as a search “servent” (server + client) by submitting and responding to queries to/from its neighbors. This search process has no centralized bottleneck. Peers depend on a local adaptive routing algorithm to dynamically change the topology of the peer network and search for the best neighbors to answer their queries. We present and evaluate learning techniques to improve local query routing. We validate prototypes of the 6S network via simulations with model users based on actual Web crawls. We find that the network topology rapidly converges from a random network to a small world network, with clusters emerging from user communities with shared interests. We finally compare the quality of the results with those obtained by centralized search engines such as Google.


Author(s):  
Denis Shestakov

Finding information on the Web using a web search engine is one of the primary activities of today’s web users. For a majority of users results returned by conventional search engines are an essentially complete set of links to all pages on the Web relevant to their queries. However, currentday searchers do not crawl and index a significant portion of the Web and, hence, web users relying on search engines only are unable to discover and access a large amount of information from the nonindexable part of the Web. Specifically, dynamic pages generated based on parameters provided by a user via web search forms are not indexed by search engines and cannot be found in searchers’ results. Such search interfaces provide web users with an online access to myriads of databases on the Web. In order to obtain some information from a web database of interest, a user issues his/her query by specifying query terms in a search form and receives the query results, a set of dynamic pages which embed required information from a database. At the same time, issuing a query via an arbitrary search interface is an extremely complex task for any kind of automatic agents including web crawlers, which, at least up to the present day, do not even attempt to pass through web forms on a large scale.


Author(s):  
Martin Feuz ◽  
Matthew Fuller ◽  
Felix Stalder

Web search engines have become indispensable tools for finding information online effectively. As the range of information, context and users of Internet searches has grown, the relationship between the search query, search interest and user has become more tenuous. Not all users are seeking the same information, even if they use the same query term. Thus, the quality of search results has, at least potentially, been decreasing. Search engines have begun to respond to this problem by trying to personalise search in order to deliver more relevant results to the users. A query is now evaluated in the context of a user’s search history and other data compiled into a personal profile and associated with statistical groups. This, at least, is the promise stated by the search engines themselves. This paper tries to assess the current reality of the personalisation of search results. We analyse the mechanisms of personalisation in the case of Google web search by empirically testing three commonly held assumptions about what personalisation does. To do this, we developed new digital methods which are explained here. The findings suggest that Google personal search does not fully provide the much-touted benefits for its search users. More likely, it seems to serve the interest of advertisers in providing more relevant audiences to them.


Sign in / Sign up

Export Citation Format

Share Document