Website removal from search engines due to copyright violation

2019 ◽  
Vol 71 (1) ◽  
pp. 54-71 ◽  
Author(s):  
Artur Strzelecki

Purpose The purpose of this paper is to clarify how many removal requests are made, how often, and who makes these requests, as well as which websites are reported to search engines so they can be removed from the search results. Design/methodology/approach Undertakes a deep analysis of more than 3.2bn removed pages from Google’s search results requested by reporting organizations from 2011 to 2018 and over 460m removed pages from Bing’s search results requested by reporting organizations from 2015 to 2017. The paper focuses on pages that belong to the .pl country coded top-level domain (ccTLD). Findings Although the number of requests to remove data from search results has been growing year on year, fewer URLs have been reported in recent years. Some of the requests are, however, unjustified and are rejected by teams representing the search engines. In terms of reporting copyright violations, one company in particular stands out (AudioLock.Net), accounting for 28.1 percent of all reports sent to Google (the top ten companies combined were responsible for 61.3 percent of the total number of reports). Research limitations/implications As not every request can be published, the study is based only what is publicly available. Also, the data assigned to Poland is only based on the ccTLD domain name (.pl); other domain extensions for Polish internet users were not considered. Originality/value This is first global analysis of data from transparency reports published by search engine companies as prior research has been based on specific notices.

2016 ◽  
Vol 34 (2) ◽  
pp. 197-206 ◽  
Author(s):  
Sungin Lee ◽  
Wonhong Jang ◽  
Eunsol Lee ◽  
Sam G. Oh

Purpose – The purpose of this paper is to examine the effect of, and identify core techniques of, search engine optimization (SEO) techniques applied to the web (http://lg-sl.net) and mobile (http//m.lg-sl.net) Science Land content and services at LG Sangnam Library in Korea. Design/methodology/approach – In accordance with three major SEO guidelines, ten SEO techniques were identified and applied, and their implications were extracted on three areas: improved search engine accessibility, increased relevance between site content and search engine keywords, and improved site credibility. The effects were quantitatively analyzed in terms of registered search engine keywords and influx of visits via search engines. Findings – This study shows that SEO techniques help increase the exposure of the library services and the number of visitors through search engines. Practical implications – SEO techniques have been applied to a few non-Korean information service organizations, but it is not a well-accepted practice in Korean libraries. And the dominant search engines in Korea have published their own SEO guidelines. Prior to this study, no significant endeavors have been undertaken in the context of Korean library services that have adopted SEO techniques to boost exposure of library services and increase user traffics. Originality/value – This is the first published study that has applied optimized SEO techniques to Korean web and mobile library services, in order to demonstrate the usefulness of the techniques for maximized exposure of library content.


2019 ◽  
Author(s):  
Muhammad Ilham Verardi Pradana

Thanks to the existence of Search engines, all of informations and datas could be easily found in the internet, one of the search engine that users use the most is Google. Google still be the most popular search engine to provide any informations available on the internet. The search result that Google provide, doesn't always give the result we wanted. Google just displayed the results based on the keyword we type. So sometimes, they show us the negative contents on the internet, such as pornography, pornsites, and many more that seems to be related to the keyword, whether the title or the other that makes the result going that way. In this paper, we will implement the "DNS SEHAT" to pass along client's request queries so the Google search engine on the client's side will provide more relevant search results without any negative contents.


2016 ◽  
Vol 68 (5) ◽  
pp. 566-588 ◽  
Author(s):  
Isto Huvila

Purpose The purpose of this paper is to discuss the affective premises and economics of the influence of search engines on knowing and informing in the contemporary society. Design/methodology/approach A conceptual discussion of the affective premises and framings of the capitalist economics of knowing is presented. Findings The main proposition of this text is that the exploitation of affects is entwined in the competing market and emancipatory discourses and counter-discourses both as intentional interventions, and perhaps even more significantly, as unintentional influences that shape the ways of knowing in the peripheries of the regime that shape cultural constellations of their own. Affective capitalism bounds and frames our ways of knowing in ways that are difficult to anticipate and read even from the context of the regime itself. Originality/value In the relatively extensive discussion on the role of affects in the contemporary capitalism, influence of affects on knowing and their relation to search engine use has received little explicit attention so far.


2019 ◽  
Vol 37 (2) ◽  
pp. 312-328 ◽  
Author(s):  
Yiming Zhao ◽  
Jin Zhang ◽  
Xue Xia ◽  
Taowen Le

Purpose The purpose of this paper is to evaluate Google question-answering (QA) quality. Design/methodology/approach Given the large variety and complexity of Google answer boxes in search result pages, existing evaluation criteria for both search engines and QA systems seemed unsuitable. This study developed an evaluation criteria system for the evaluation of Google QA quality by coding and analyzing search results of questions from a representative question set. The study then evaluated Google’s overall QA quality as well as QA quality across four target types and across six question types, using the newly developed criteria system. ANOVA and Tukey tests were used to compare QA quality among different target types and question types. Findings It was found that Google provided significantly higher-quality answers to person-related questions than to thing-related, event-related and organization-related questions. Google also provided significantly higher-quality answers to where- questions than to who-, what- and how-questions. The more specific a question is, the higher the QA quality would be. Research limitations/implications Suggestions for both search engine users and designers are presented to help enhance user experience and QA quality. Originality/value Particularly suitable for search engine QA quality analysis, the newly developed evaluation criteria system expanded and enriched assessment metrics of both search engines and QA systems.


2016 ◽  
Vol 34 (3) ◽  
pp. 457-470 ◽  
Author(s):  
Zahid Ashraf Wani ◽  
Adil Ahmad Sofi

Purpose This paper aims to gauge the visibility of open content available in different formats of select open courseware (OCW) repositories through prominent search engines. Design/methodology/approach Open content in three formats (pdf, audio and video) from four OCW repositories listed in the OCW consortium under the science and technology subject heading were searched through seven select search engines. Findings None of the selected OCW repositories are fully visible on the selected search engines. Visibility of OCW content varied from one search engine to the other and was affected by the format in which it is available. Google is the best search engine for retrieving OCW content, whereas OCWfinder – a specialized search engine for retrieving OCW – has performed dismally. Research limitations/implications The study demonstrates the need for enhancing the visibility of open content through using search engine optimization techniques. Originality/value The study intends to supply findings that could be used by stakeholders to improve the visibility of OCW repositories. It is an attempt to draw a comparison between search engines for their ability to index different formats of OCW in the selected repositories. Findings can be used by information professionals to brush their information hunting skills.


2020 ◽  
Vol ahead-of-print (ahead-of-print) ◽  
Author(s):  
Sebastian Schultheiß ◽  
Dirk Lewandowski

PurposeIn commercial web search engine results rankings, four stakeholder groups are involved: search engine providers, users, content providers and search engine optimizers. Search engine optimization (SEO) is a multi-billion-dollar industry and responsible for making content visible through search engines. Despite this importance, little is known about its role in the interaction of the stakeholder groups.Design/methodology/approachWe conducted expert interviews with 15 German search engine optimizers and content providers, the latter represented by content managers and online journalists. The interviewees were asked about their perspectives on SEO and how they assess the views of users about SEO.FindingsSEO was considered necessary for content providers to ensure visibility, which is why dependencies between both stakeholder groups have evolved. Despite its importance, SEO was seen as largely unknown to users. Therefore, it is assumed that users cannot realistically assess the impact SEO has and that user opinions about SEO depend heavily on their knowledge of the topic.Originality/valueThis study investigated search engine optimization from the perspective of those involved in the optimization business: content providers, online journalists and search engine optimization professionals. The study therefore contributes to a more nuanced view on and a deeper understanding of the SEO domain.


2019 ◽  
Vol 54 (4) ◽  
pp. 548-558
Author(s):  
Laura Sinay ◽  
Maria Cristina Fogliatti de Sinay ◽  
Rodney William (Bill) Carter ◽  
Aurea Martins

Purpose The purpose of this paper is to critically analyze the influence of the algorithm used on scholarly search engines (Garfield’s algorithm) and propose metrics to improve it so that science could be based on a more democratic way. Design/methodology/approach This paper used a snow-ball approach to collect data that allowed identifying the history and the logic behind the Garfield’s algorithm. It follows on excerpting the foundation of existing algorithm and databases of major scholarly search engine. It concluded proposing new metrics so as to surpass restraints and to democratize the scientific discourse. Findings This paper finds that the studied algorithm currently biases the scientific discourse toward a narrow perspective, while it should take into consideration several researchers’ characteristics. It proposes the substitution of the h-index by the number of times the scholar’s most cited work has been cited. Finally, it proposes that works in languages different than English should be included. Research limitations/implications The broad comprehension of any phenomena should be based on multiple perspectives; therefore, the inclusion of diverse metrics will extend the scientific discourse. Practical implications The improvement of the existing algorithm will increase the chances of contact among different cultures, which stimulate rapid progress on the development of knowledge. Originality/value The value of this paper resides in demonstrating that the algorithm used in scholarly search engines biases the development of science. If updated as proposed here, science will be unbiased and bias aware.


2014 ◽  
Vol 30 (4) ◽  
pp. 206-211
Author(s):  
Robert Fox

Purpose – This column aims to look at various aspects of search engine optimization (SEO) and the potential risks and rewards from exposing library-related content using techniques such as microdata and descriptive frameworks such as that outlined on schema.org. Design/methodology/approach – Regular column. Findings – This column explores concepts related to SEO and is speculative in nature. Originality/value – Academic libraries can greatly benefit from exploring the potential benefits of using SEO and microdata/microformats. The landscape for SEO has changed dramatically over the past decade, and the benefit to libraries who have in once sense seen themselves as competitors with the major search engines is significant.


2021 ◽  
pp. 089443932110068
Author(s):  
Aleksandra Urman ◽  
Mykola Makhortykh ◽  
Roberto Ulloa

We examine how six search engines filter and rank information in relation to the queries on the U.S. 2020 presidential primary elections under the default—that is nonpersonalized—conditions. For that, we utilize an algorithmic auditing methodology that uses virtual agents to conduct large-scale analysis of algorithmic information curation in a controlled environment. Specifically, we look at the text search results for “us elections,” “donald trump,” “joe biden,” “bernie sanders” queries on Google, Baidu, Bing, DuckDuckGo, Yahoo, and Yandex, during the 2020 primaries. Our findings indicate substantial differences in the search results between search engines and multiple discrepancies within the results generated for different agents using the same search engine. It highlights that whether users see certain information is decided by chance due to the inherent randomization of search results. We also find that some search engines prioritize different categories of information sources with respect to specific candidates. These observations demonstrate that algorithmic curation of political information can create information inequalities between the search engine users even under nonpersonalized conditions. Such inequalities are particularly troubling considering that search results are highly trusted by the public and can shift the opinions of undecided voters as demonstrated by previous research.


Author(s):  
Novario Jaya Perdana

The accuracy of search result using search engine depends on the keywords that are used. Lack of the information provided on the keywords can lead to reduced accuracy of the search result. This means searching information on the internet is a hard work. In this research, a software has been built to create document keywords sequences. The software uses Google Latent Semantic Distance which can extract relevant information from the document. The information is expressed in the form of specific words sequences which could be used as keyword recommendations in search engines. The result shows that the implementation of the method for creating document keyword recommendation achieved high accuracy and could finds the most relevant information in the top search results.


Sign in / Sign up

Export Citation Format

Share Document