Visual comparison of search results: A censorship case study

First Monday ◽  
2008 ◽  
Author(s):  
Mark Meiss ◽  
Filippo Menczer

Understanding the qualitative differences between the sets of results from different search engines can be a difficult task. How many links must you follow from each list before you can reach a conclusion? We describe a user interface that allows users to quickly identify the most significant differences in content between two lists of Web pages. We have implemented this interface in CenSEARCHip, a system for comparing the effects of censorship policies on search engines.

Author(s):  
Natasha Tusikov

This chapter explains how the transnational regime uses search engines (especially Google) and domain name registrars (specifically GoDaddy) to throttle access to infringing sites. It traces efforts by the U.S. and U.K. governments, along with rights holders, to pressure Google and GoDaddy into adopting the non-binding agreements. It then presents two case studies. The first discusses search engines’ regulation of search results linking to infringing sites and a non-binding agreement struck among search engines (Google, Yahoo, and Microsoft) at the behest of the U.K. government. The second case study examines GoDaddy’s efforts to disable so-called illegal online pharmacies that operate in violation of U.S. federal and state laws. The chapter concludes that Internet firms’ practice of using chokepoints to dissuade access to targeted websites is highly problematic as legitimate websites are mistakenly targeted and sanctioned. Automated enforcement programs exacerbate this problem as they significantly increase the scale and speed of rights holders’ enforcement efforts without a corresponding increase in oversight.


2002 ◽  
Vol 63 (4) ◽  
pp. 354-365 ◽  
Author(s):  
Susan Augustine ◽  
Courtney Greene

Have Internet search engines influenced the way students search library Web pages? The results of this usability study reveal that students consistently and frequently use the library Web site’s internal search engine to find information rather than navigating through pages. If students are searching rather than navigating, library Web page designers must make metadata and powerful search engines priorities. The study also shows that students have difficulty interpreting library terminology, experience confusion discerning difference amongst library resources, and prefer to seek human assistance when encountering problems online. These findings imply that library Web sites have not alleviated some of the basic and long-range problems that have challenged librarians in the past.


Author(s):  
José Antonio Robles-Flores ◽  
Gregory Schymik ◽  
Julie Smith-David ◽  
Robert St. Louis

Web search engines typically retrieve a large number of web pages and overload business analysts with irrelevant information. One approach that has been proposed for overcoming some of these problems is automated Question Answering (QA). This paper describes a case study that was designed to determine the efficacy of QA systems for generating answers to original, fusion, list questions (questions that have not previously been asked and answered, questions for which the answer cannot be found on a single web site, and questions for which the answer is a list of items). Results indicate that QA algorithms are not very good at producing complete answer lists and that searchers are not very good at constructing answer lists from snippets. These findings indicate a need for QA research to focus on crowd sourcing answer lists and improving output format.


Author(s):  
Sunny Sharma ◽  
Vijay Rana

: The Existing studies have already revealed that the information on the web is increasing rapidly. Ambiguous queries and user’s ability to express their intention through queries have been one of the key challenges in retrieving the accurate search results from the search engine. This paper in response explored different methodologies proposed during 2005-2019 by the eminent researchers for recommending better search results. Some of these methodologies are based on the users’ geographical location while others rely on re- rank the web results and refinement of user’s query. Fellow researchers can use this literature, to define the fundamental literature for their own work. Further a brief case study of major search engines like Google, Yahoo, Bing etc. along with the techniques used by these search engines for personalization are also discussed. Finally, the paper discusses some current issues and challenges related to the personalization which further lays the future research directions.


2011 ◽  
Vol 3 (4) ◽  
pp. 62-70 ◽  
Author(s):  
Stephen O’Neill ◽  
Kevin Curran

Search engine optimization (SEO) is the process of improving the visibility, volume and quality of traffic to website or a web page in search engines via the natural search results. SEO can also target other areas of a search, including image search and local search. SEO is one of many different strategies used for marketing a website but SEO has been proven the most effective. An Internet marketing campaign may drive organic search results to websites or web pages but can be involved with paid advertising on search engines. All search engines have a unique way of ranking the importance of a website. Some search engines focus on the content while others review Meta tags to identify who and what a web site’s business is. Most engines use a combination of Meta tags, content, link popularity, click popularity and longevity to determine a sites ranking. To make it even more complicated, they change their ranking policies frequently. This paper provides an overview of search engine optimisation strategies and pitfalls.


2011 ◽  
Vol 2 (1) ◽  
pp. 46-63
Author(s):  
José Antonio Robles-Flores ◽  
Gregory Schymik ◽  
Julie Smith-David ◽  
Robert St. Louis

Web search engines typically retrieve a large number of web pages and overload business analysts with irrelevant information. One approach that has been proposed for overcoming some of these problems is automated Question Answering (QA). This paper describes a case study that was designed to determine the efficacy of QA systems for generating answers to original, fusion, list questions (questions that have not previously been asked and answered, questions for which the answer cannot be found on a single web site, and questions for which the answer is a list of items). Results indicate that QA algorithms are not very good at producing complete answer lists and that searchers are not very good at constructing answer lists from snippets. These findings indicate a need for QA research to focus on crowd sourcing answer lists and improving output format.


2012 ◽  
Vol 38 (3) ◽  
pp. 284-296 ◽  
Author(s):  
Mohammed Al-Kabi ◽  
Heider Wahsheh ◽  
Izzat Alsmadi ◽  
Emad Al-Shawakfa ◽  
Abdullah Wahbeh ◽  
...  

Search engines are important outlets for information query and retrieval. They have to deal with the continual increase of information available on the web, and provide users with convenient access to such huge amounts of information. Furthermore, with this huge amount of information, a more complex challenge that continuously gets more and more difficult to illuminate is the spam in web pages. For several reasons, web spammers try to intrude in the search results and inject artificially biased results in favour of their websites or pages. Spam pages are added to the internet on a daily basis, thus making it difficult for search engines to keep up with the fast-growing and dynamic nature of the web, especially since spammers tend to add more keywords to their websites to deceive the search engines and increase the rank of their pages. In this research, we have investigated four different classification algorithms (naïve Bayes, decision tree, SVM and K-NN) to detect Arabic web spam pages, based on content. The three groups of datasets used, with 1%, 15% and 50% spam contents, were collected using a crawler that was customized for this study. Spam pages were classified manually. Different tests and comparisons have revealed that the Decision Tree was the best classifier for this purpose.


2018 ◽  
Vol 41 (1) ◽  
pp. 21-37 ◽  
Author(s):  
Andrei Zavadski ◽  
Florian Toepfl

In the digital memories literature, the practice of searching for information – one of the most frequent online activities worldwide – has received comparatively little attention. To fill the gap, this exploratory study asks how search engines affect the representations of the past that they produce in query results. Designed as a single revelatory case study, with a focus on Russia, this article delineates a typology of four types of memory events based on four types of websites dominating search results. For each type of event, we discuss recurring locations and mechanisms of power struggles over competing memory narratives. We conclude that within Russia’s authoritarian context, the mnemonic practice of Internet searching tends to reproduce and reinforce the dominant narratives supported by the ruling elites. Search engine companies are thus only one of several powerful institutions that constitute the social framework within which querying the Internet is pursued as a mnemonic practice. Others include mass media, non-governmental organisations (NGOs), and online encyclopaedias.


Author(s):  
Stephen O’Neill ◽  
Kevin Curran

Search engine optimization (SEO) is the process of improving the visibility, volume and quality of traffic to website or a web page in search engines via the natural search results. SEO can also target other areas of a search, including image search and local search. SEO is one of many different strategies used for marketing a website but SEO has been proven the most effective. An Internet marketing campaign may drive organic search results to websites or web pages but can be involved with paid advertising on search engines. All search engines have a unique way of ranking the importance of a website. Some search engines focus on the content while others review Meta tags to identify who and what a web site’s business is. Most engines use a combination of Meta tags, content, link popularity, click popularity and longevity to determine a sites ranking. To make it even more complicated, they change their ranking policies frequently. This paper provides an overview of search engine optimisation strategies and pitfalls.


Author(s):  
Sakher Khalil Alqaaidi

Search engines save copies of crawled web pages to provide instant search results, saved pages may become old and un-updated as original pages change providing new information and new links, and most of websites don’t submit these new changes to search engines so search engines don’t depend mainly on websites techniques of submitting changes. Keeping pages fresh and updated in search engine is important for giving real page ranks and for providing real time information. Techniques were invented to improve the page update process by search engine. In this paper the author combines two of good known techniques and implements the new one via experiments that improve better results in different experiment cases.


Sign in / Sign up

Export Citation Format

Share Document