Website Quality Evaluation Based on Search Engine Queries using Web Rank Position Algorithm (WRPA)

Author(s):  
Chandran M ◽  
Ramani A. V

<p>The research work is about to test the quality of the website and to improve the quality by analyzing the hit counts, impressions, clicks, count through rates and average positions. This is accomplished using WRPA and SEO technique. The quality of the website mainly lies on the keywords which are present in it. The keywords can be of a search query which is typed by the users in the search engines and based on these keywords, the websites are displayed in the search results. This research work concentrates on bringing the particular websites to the first of the search result in the search engine. The website chosen for research is SRKV. The research work is carried out by creating an index array of Meta tags. This array will hold all the Meta tags. All the search keywords for the website from the users are stored in another array. The index array is matched and compared with the search keywords array. From this, hit to count is calculated for the analysis. Now the calculated hit count and the searched keywords will be analyzed to improve the performance of the website. The matched special keywords from the above comparison are included in the Meta tag to improve the performance of the website. Again all the Meta tags and newly specified keywords in the index array are matched with the SEO keywords. If this matches, then the matched keyword will be stored for improving the quality of the website. Metrics such as impressions, clicks, CTR, average positions are also measured along with the hit counts. The research is carried out under different types of browsers and different types of platforms. Queries about the website from different countries are also measured. In conclusion, if the number of the clicks for the website is more than the average number of clicks, then the quality of the website is good. This research helps in improvising the keywords using WRPA and SEO and thereby improves the quality of the website easily.</p>

Author(s):  
Novario Jaya Perdana

The accuracy of search result using search engine depends on the keywords that are used. Lack of the information provided on the keywords can lead to reduced accuracy of the search result. This means searching information on the internet is a hard work. In this research, a software has been built to create document keywords sequences. The software uses Google Latent Semantic Distance which can extract relevant information from the document. The information is expressed in the form of specific words sequences which could be used as keyword recommendations in search engines. The result shows that the implementation of the method for creating document keyword recommendation achieved high accuracy and could finds the most relevant information in the top search results.


Author(s):  
Xiannong Meng ◽  
Song Xing

This chapter reports the results of a project attempting to assess the performance of a few major search engines from various perspectives. The search engines involved in the study include the Microsoft Search Engine (MSE) when it was in its beta test stage, AllTheWeb, and Yahoo. In a few comparisons, other search engines such as Google, Vivisimo are also included. The study collects statistics such as the average user response time, average process time for a query reported by MSE, as well as the number of pages relevant to a query reported by all search engines involved. The project also studies the quality of search results generated by MSE and other search engines using RankPower as the metric. We found MSE performs well in speed and diversity of the query results, while weaker in other statistics, compared to some other leading search engines. The contribution of this chapter is to review the performance evaluation techniques for search engines and use different measures to assess and compare the quality of different search engines, especially MSE.


2019 ◽  
Author(s):  
Muhammad Ilham Verardi Pradana

Thanks to the existence of Search engines, all of informations and datas could be easily found in the internet, one of the search engine that users use the most is Google. Google still be the most popular search engine to provide any informations available on the internet. The search result that Google provide, doesn't always give the result we wanted. Google just displayed the results based on the keyword we type. So sometimes, they show us the negative contents on the internet, such as pornography, pornsites, and many more that seems to be related to the keyword, whether the title or the other that makes the result going that way. In this paper, we will implement the "DNS SEHAT" to pass along client's request queries so the Google search engine on the client's side will provide more relevant search results without any negative contents.


2021 ◽  
Author(s):  
Kwok-Pun Chan

Meta search engines allow multiple engine searches to minimize biased information and improve the quality of the results it generates. However, existing meta engine applications contain many foreign language results, and only run on Windows platform. The meta search engine we develop will resolve these problems. Our search engine will run on both Windows and Linus platforms, and has some desirable properties: 1) users can shorten the search waiting time if one of the search engines is down 2) users can sort the result titles in an alphabetic or relevancy order. Current meta search websites only allow users to sort results by relevancy. Our search engine allows users to do an alphabetical search from the previous relevancy search result, so that the users can identify the required title within a shorter time frame.


2011 ◽  
Vol 3 (4) ◽  
pp. 62-70 ◽  
Author(s):  
Stephen O’Neill ◽  
Kevin Curran

Search engine optimization (SEO) is the process of improving the visibility, volume and quality of traffic to website or a web page in search engines via the natural search results. SEO can also target other areas of a search, including image search and local search. SEO is one of many different strategies used for marketing a website but SEO has been proven the most effective. An Internet marketing campaign may drive organic search results to websites or web pages but can be involved with paid advertising on search engines. All search engines have a unique way of ranking the importance of a website. Some search engines focus on the content while others review Meta tags to identify who and what a web site’s business is. Most engines use a combination of Meta tags, content, link popularity, click popularity and longevity to determine a sites ranking. To make it even more complicated, they change their ranking policies frequently. This paper provides an overview of search engine optimisation strategies and pitfalls.


2019 ◽  
Vol 2019 ◽  
Author(s):  
Jan Rensinghoff ◽  
Florian Marius Farke ◽  
Markus Dürmuth ◽  
Tobias Gostomzyk

The new European right to be forgotten (Art. 17 of the European General Data Protection Regulation (GDPR) grants EU citizens the right to demand the erasure of their personal data from anyone who processes their personal data. To enforce this right to erasure may be a problem for many of those data processors. On the one hand, they need to examine any claim to remove search results. On the other hand, they have to balance conflicting rights in order to prevent over-blocking and the accusation of censorship. The paper examines the criteria which are potentially involved in the decision-making process of search engines when it comes to the right to erasure. We present an approach helping search engine operators and individuals to assess and decide whether search results may have to be deleted or not. Our goal is to make this process more transparent and consistent, providing more legal certainty for both the search engine operator and the person concerned by the search result in question. As a result, we develop a model to estimate the chances of success to delete a particular search result for a given person. This is a work in progress.


Author(s):  
Stephen O’Neill ◽  
Kevin Curran

Search engine optimization (SEO) is the process of improving the visibility, volume and quality of traffic to website or a web page in search engines via the natural search results. SEO can also target other areas of a search, including image search and local search. SEO is one of many different strategies used for marketing a website but SEO has been proven the most effective. An Internet marketing campaign may drive organic search results to websites or web pages but can be involved with paid advertising on search engines. All search engines have a unique way of ranking the importance of a website. Some search engines focus on the content while others review Meta tags to identify who and what a web site’s business is. Most engines use a combination of Meta tags, content, link popularity, click popularity and longevity to determine a sites ranking. To make it even more complicated, they change their ranking policies frequently. This paper provides an overview of search engine optimisation strategies and pitfalls.


2021 ◽  
Author(s):  
Kwok-Pun Chan

Meta search engines allow multiple engine searches to minimize biased information and improve the quality of the results it generates. However, existing meta engine applications contain many foreign language results, and only run on Windows platform. The meta search engine we develop will resolve these problems. Our search engine will run on both Windows and Linus platforms, and has some desirable properties: 1) users can shorten the search waiting time if one of the search engines is down 2) users can sort the result titles in an alphabetic or relevancy order. Current meta search websites only allow users to sort results by relevancy. Our search engine allows users to do an alphabetical search from the previous relevancy search result, so that the users can identify the required title within a shorter time frame.


2021 ◽  
Vol 5 (2) ◽  
Author(s):  
Hannah C Cai ◽  
Leanne E King ◽  
Johanna T Dwyer

ABSTRACT We assessed the quality of online health and nutrition information using a Google™ search on “supplements for cancer”. Search results were scored using the Health Information Quality Index (HIQI), a quality-rating tool consisting of 12 objective criteria related to website domain, lack of commercial aspects, and authoritative nature of the health and nutrition information provided. Possible scores ranged from 0 (lowest) to 12 (“perfect” or highest quality). After eliminating irrelevant results, the remaining 160 search results had median and mean scores of 8. One-quarter of the results were of high quality (score of 10–12). There was no correlation between high-quality scores and early appearance in the sequence of search results, where results are presumably more visible. Also, 496 advertisements, over twice the number of search results, appeared. We conclude that the Google™ search engine may have shortcomings when used to obtain information on dietary supplements and cancer.


2021 ◽  
pp. 089443932110068
Author(s):  
Aleksandra Urman ◽  
Mykola Makhortykh ◽  
Roberto Ulloa

We examine how six search engines filter and rank information in relation to the queries on the U.S. 2020 presidential primary elections under the default—that is nonpersonalized—conditions. For that, we utilize an algorithmic auditing methodology that uses virtual agents to conduct large-scale analysis of algorithmic information curation in a controlled environment. Specifically, we look at the text search results for “us elections,” “donald trump,” “joe biden,” “bernie sanders” queries on Google, Baidu, Bing, DuckDuckGo, Yahoo, and Yandex, during the 2020 primaries. Our findings indicate substantial differences in the search results between search engines and multiple discrepancies within the results generated for different agents using the same search engine. It highlights that whether users see certain information is decided by chance due to the inherent randomization of search results. We also find that some search engines prioritize different categories of information sources with respect to specific candidates. These observations demonstrate that algorithmic curation of political information can create information inequalities between the search engine users even under nonpersonalized conditions. Such inequalities are particularly troubling considering that search results are highly trusted by the public and can shift the opinions of undecided voters as demonstrated by previous research.


Sign in / Sign up

Export Citation Format

Share Document