scholarly journals IMPLEMENTASI ALGORITMA GOOGLE LATENT SEMANTIC DISTANCE UNTUK EKSTRAKSI RANGKAIAN KATA KUNCI ARTIKEL JURNAL ILMIAH

Author(s):  
Novario Jaya Perdana

The accuracy of search result using search engine depends on the keywords that are used. Lack of the information provided on the keywords can lead to reduced accuracy of the search result. This means searching information on the internet is a hard work. In this research, a software has been built to create document keywords sequences. The software uses Google Latent Semantic Distance which can extract relevant information from the document. The information is expressed in the form of specific words sequences which could be used as keyword recommendations in search engines. The result shows that the implementation of the method for creating document keyword recommendation achieved high accuracy and could finds the most relevant information in the top search results.

2019 ◽  
Author(s):  
Muhammad Ilham Verardi Pradana

Thanks to the existence of Search engines, all of informations and datas could be easily found in the internet, one of the search engine that users use the most is Google. Google still be the most popular search engine to provide any informations available on the internet. The search result that Google provide, doesn't always give the result we wanted. Google just displayed the results based on the keyword we type. So sometimes, they show us the negative contents on the internet, such as pornography, pornsites, and many more that seems to be related to the keyword, whether the title or the other that makes the result going that way. In this paper, we will implement the "DNS SEHAT" to pass along client's request queries so the Google search engine on the client's side will provide more relevant search results without any negative contents.


Author(s):  
Chandran M ◽  
Ramani A. V

<p>The research work is about to test the quality of the website and to improve the quality by analyzing the hit counts, impressions, clicks, count through rates and average positions. This is accomplished using WRPA and SEO technique. The quality of the website mainly lies on the keywords which are present in it. The keywords can be of a search query which is typed by the users in the search engines and based on these keywords, the websites are displayed in the search results. This research work concentrates on bringing the particular websites to the first of the search result in the search engine. The website chosen for research is SRKV. The research work is carried out by creating an index array of Meta tags. This array will hold all the Meta tags. All the search keywords for the website from the users are stored in another array. The index array is matched and compared with the search keywords array. From this, hit to count is calculated for the analysis. Now the calculated hit count and the searched keywords will be analyzed to improve the performance of the website. The matched special keywords from the above comparison are included in the Meta tag to improve the performance of the website. Again all the Meta tags and newly specified keywords in the index array are matched with the SEO keywords. If this matches, then the matched keyword will be stored for improving the quality of the website. Metrics such as impressions, clicks, CTR, average positions are also measured along with the hit counts. The research is carried out under different types of browsers and different types of platforms. Queries about the website from different countries are also measured. In conclusion, if the number of the clicks for the website is more than the average number of clicks, then the quality of the website is good. This research helps in improvising the keywords using WRPA and SEO and thereby improves the quality of the website easily.</p>


2018 ◽  
Vol 1 (1) ◽  
pp. 15-22
Author(s):  
Diki Arisandi ◽  
Sukri Sukri ◽  
Salamun Salamun ◽  
Roni Salambue

Access to information from the internet is become the thing that is needed by almost all society, and also the student of SMK N II Taluk Kuantan. In this community service, the material that has been given is about how the search results obtained to be more effective by using power searching. Power searching delivered to students of SMK N II Taluk Kuantan is by inserting mathematical symbols on the keywords entered into the search engine on the internet. In addition, the material also discussed about more specific search by combining mathematical symbols, host names and file types to search. The method In this community service was giving the material and demo about how the implementation of power searching on search engines. After this activity is implemented, the community service team evaluates the material that has been given before. The result was the students in SMK N II Taluk Kuantan can implement power searching well.\  


2019 ◽  
Vol 2019 ◽  
Author(s):  
Jan Rensinghoff ◽  
Florian Marius Farke ◽  
Markus Dürmuth ◽  
Tobias Gostomzyk

The new European right to be forgotten (Art. 17 of the European General Data Protection Regulation (GDPR) grants EU citizens the right to demand the erasure of their personal data from anyone who processes their personal data. To enforce this right to erasure may be a problem for many of those data processors. On the one hand, they need to examine any claim to remove search results. On the other hand, they have to balance conflicting rights in order to prevent over-blocking and the accusation of censorship. The paper examines the criteria which are potentially involved in the decision-making process of search engines when it comes to the right to erasure. We present an approach helping search engine operators and individuals to assess and decide whether search results may have to be deleted or not. Our goal is to make this process more transparent and consistent, providing more legal certainty for both the search engine operator and the person concerned by the search result in question. As a result, we develop a model to estimate the chances of success to delete a particular search result for a given person. This is a work in progress.


Author(s):  
Ajitkumar Kadam

Basically, the search engine is used to search for any information on the Internet. The purpose of any website or webpage is to appear on top of the search engine when any user searches for relevant information in the content on their website. There are several factors that affect website ratings and make it down in the search crawler's crop list and some websites do not even come in the top search engine list despite the best content. The art of improving website visibility is known as search engine optimization. Most of times, main focus is set on making the website as user-friendly as possible, stable, fast and secure. But when creating the website, we need to take into consideration an essential factor, which is making the world aware of their website and their content. In this paper, we will survey how various techniques optimize the search result in search engine.


Author(s):  
Vijay Kasi ◽  
Radhika Jain

In the context of the Internet, a search engine can be defined as a software program designed to help one access information, documents, and other content on the World Wide Web. The adoption and growth of the Internet in the last decade has been unprecedented. The World Wide Web has always been applauded for its simplicity and ease of use. This is evident looking at the extent of the knowledge one requires to build a Web page. The flexible nature of the Internet has enabled the rapid growth and adoption of it, making it hard to search for relevant information on the Web. The number of Web pages has been increasing at an astronomical pace, from around 2 million registered domains in 1995 to 233 million registered domains in 2004 (Consortium, 2004). The Internet, considered a distributed database of information, has the CRUD (create, retrieve, update, and delete) rule applied to it. While the Internet has been effective at creating, updating, and deleting content, it has considerably lacked in enabling the retrieval of relevant information. After all, there is no point in having a Web page that has little or no visibility on the Web. Since the 1990s when the first search program was released, we have come a long way in terms of searching for information. Although we are currently witnessing a tremendous growth in search engine technology, the growth of the Internet has overtaken it, leading to a state in which the existing search engine technology is falling short. When we apply the metrics of relevance, rigor, efficiency, and effectiveness to the search domain, it becomes very clear that we have progressed on the rigor and efficiency metrics by utilizing abundant computing power to produce faster searches with a lot of information. Rigor and efficiency are evident in the large number of indexed pages by the leading search engines (Barroso, Dean, & Holzle, 2003). However, more research needs to be done to address the relevance and effectiveness metrics. Users typically type in two to three keywords when searching, only to end up with a search result having thousands of Web pages! This has made it increasingly hard to effectively find any useful, relevant information. Search engines face a number of challenges today requiring them to perform rigorous searches with relevant results efficiently so that they are effective. These challenges include the following (“Search Engines,” 2004). 1. The Web is growing at a much faster rate than any present search engine technology can index. 2. Web pages are updated frequently, forcing search engines to revisit them periodically. 3. Dynamically generated Web sites may be slow or difficult to index, or may result in excessive results from a single Web site. 4. Many dynamically generated Web sites are not able to be indexed by search engines. 5. The commercial interests of a search engine can interfere with the order of relevant results the search engine shows. 6. Content that is behind a firewall or that is password protected is not accessible to search engines (such as those found in several digital libraries).1 7. Some Web sites have started using tricks such as spamdexing and cloaking to manipulate search engines to display them as the top results for a set of keywords. This can make the search results polluted, with more relevant links being pushed down in the result list. This is a result of the popularity of Web searches and the business potential search engines can generate today. 8. Search engines index all the content of the Web without any bounds on the sensitivity of information. This has raised a few security and privacy flags. With the above background and challenges in mind, we lay out the article as follows. In the next section, we begin with a discussion of search engine evolution. To facilitate the examination and discussion of the search engine development’s progress, we break down this discussion into the three generations of search engines. Figure 1 depicts this evolution pictorially and highlights the need for better search engine technologies. Next, we present a brief discussion on the contemporary state of search engine technology and various types of content searches available today. With this background, the next section documents various concerns about existing search engines setting the stage for better search engine technology. These concerns include information overload, relevance, representation, and categorization. Finally, we briefly address the research efforts under way to alleviate these concerns and then present our conclusion.


2020 ◽  
Author(s):  
Hong Xiong

The automatic search performance of search engines has become an essential part of measuring the difference in user experience. An efficient automatic search system can significantly improve the performance of search engines and increase user traffic. Hadoop has strong data integration and analysis capabilities, while R has excellent statistical capabilities in linear regression. This article will propose a linear regression based on Hadoop and R to quantify the efficiency of the automatic retrieval system. We use R's functional properties to transform the user's search results upon linear correlations. In this way, the final output results have multiple display forms instead of web page preview interfaces. This article provides feasible solutions to the drawbacks of current search engine algorithms lacking once or twice search accuracies and multiple types of search results. We can conduct personalized regression analysis for user’s needs with public datasets and optimize resources integration for most relevant information.


2021 ◽  
pp. 089443932110068
Author(s):  
Aleksandra Urman ◽  
Mykola Makhortykh ◽  
Roberto Ulloa

We examine how six search engines filter and rank information in relation to the queries on the U.S. 2020 presidential primary elections under the default—that is nonpersonalized—conditions. For that, we utilize an algorithmic auditing methodology that uses virtual agents to conduct large-scale analysis of algorithmic information curation in a controlled environment. Specifically, we look at the text search results for “us elections,” “donald trump,” “joe biden,” “bernie sanders” queries on Google, Baidu, Bing, DuckDuckGo, Yahoo, and Yandex, during the 2020 primaries. Our findings indicate substantial differences in the search results between search engines and multiple discrepancies within the results generated for different agents using the same search engine. It highlights that whether users see certain information is decided by chance due to the inherent randomization of search results. We also find that some search engines prioritize different categories of information sources with respect to specific candidates. These observations demonstrate that algorithmic curation of political information can create information inequalities between the search engine users even under nonpersonalized conditions. Such inequalities are particularly troubling considering that search results are highly trusted by the public and can shift the opinions of undecided voters as demonstrated by previous research.


2021 ◽  
Vol 11 (15) ◽  
pp. 7063
Author(s):  
Esmaeel Rezaee ◽  
Ali Mohammad Saghiri ◽  
Agostino Forestiero

With the increasing growth of different types of data, search engines have become an essential tool on the Internet. Every day, billions of queries are run through few search engines with several privacy violations and monopoly problems. The blockchain, as a trending technology applied in various fields, including banking, IoT, education, etc., can be a beneficial alternative. Blockchain-based search engines, unlike monopolistic ones, do not have centralized controls. With a blockchain-based search system, no company can lay claims to user’s data or access search history and other related information. All these data will be encrypted and stored on a blockchain. Valuing users’ searches and paying them in return is another advantage of a blockchain-based search engine. Additionally, in smart environments, as a trending research field, blockchain-based search engines can provide context-aware and privacy-preserved search results. According to our research, few efforts have been made to develop blockchain use, which include studies generally in the early stages and few white papers. To the best of our knowledge, no research article has been published in this regard thus far. In this paper, a survey on blockchain-based search engines is provided. Additionally, we state that the blockchain is an essential paradigm for the search ecosystem by describing the advantages.


2019 ◽  
Vol 71 (1) ◽  
pp. 54-71 ◽  
Author(s):  
Artur Strzelecki

Purpose The purpose of this paper is to clarify how many removal requests are made, how often, and who makes these requests, as well as which websites are reported to search engines so they can be removed from the search results. Design/methodology/approach Undertakes a deep analysis of more than 3.2bn removed pages from Google’s search results requested by reporting organizations from 2011 to 2018 and over 460m removed pages from Bing’s search results requested by reporting organizations from 2015 to 2017. The paper focuses on pages that belong to the .pl country coded top-level domain (ccTLD). Findings Although the number of requests to remove data from search results has been growing year on year, fewer URLs have been reported in recent years. Some of the requests are, however, unjustified and are rejected by teams representing the search engines. In terms of reporting copyright violations, one company in particular stands out (AudioLock.Net), accounting for 28.1 percent of all reports sent to Google (the top ten companies combined were responsible for 61.3 percent of the total number of reports). Research limitations/implications As not every request can be published, the study is based only what is publicly available. Also, the data assigned to Poland is only based on the ccTLD domain name (.pl); other domain extensions for Polish internet users were not considered. Originality/value This is first global analysis of data from transparency reports published by search engine companies as prior research has been based on specific notices.


Sign in / Sign up

Export Citation Format

Share Document