User and Context-Aware Quality Filters Based on Web Metadata Retrieval

Author(s):  
Ricardo Barros ◽  
Geraldo Xexéo ◽  
Wallace A. Pinheiro ◽  
Jano de Souza

Due to the amount of information on the Web being so large and being of varying levels of quality, it is becoming increasingly difficult to find precisely what is required on the Web, particularly if the information consumer does not have precise knowledge of his or her information needs. On the Web, while searching for information, users can find data that is old, imprecise, invalid, intentionally wrong, or biased, due to this large amount of available data and comparative ease of access. In this environment users constantly receive useless, outdated, or false data, which they have no means to assess. This chapter addresses the issues regarding the large amount and low quality of Web information by proposing a methodology that adopts user and context-aware quality filters based on Web metadata retrieval. This starts with an initial evaluation and adjusts it to consider context characteristics and user perspectives to obtain aggregated evaluation values.

2017 ◽  
Author(s):  
Kouichi Tanabe ◽  
Kaho Fujiwara ◽  
Hana Ogura ◽  
Hatsuna Yasuda ◽  
Nobuyuki Goto ◽  
...  

BACKGROUND Patients and their families are able to obtain information about palliative care from websites easily nowadays. However, there are concerns on the accuracy of information on the Web and how up to date it is. OBJECTIVE The objective of this study was to elucidate problematic points of medical information about palliative care obtained from websites, and to compare the quality of the information between Japanese and US websites. METHODS We searched Google Japan and Google USA for websites relating to palliative care. We then evaluated the top 50 websites from each search using the DISCERN and LIDA instruments. RESULTS We found that Japanese websites were given a lower evaluation of reliability than US websites. In 3 LIDA instrument subcategories—engagability (P<.001), currency (P=.001), and content production procedure (P<.001)—US websites scored significantly higher and had large effect sizes. CONCLUSIONS Our results suggest that Japanese websites have problems with the frequency with which they are updated, their update procedures and policies, and the scrutiny process the evidence must undergo. Additionally, there was a weak association between search ranking and reliability, and simultaneously we found that reliability could not be assessed by search ranking alone.


2016 ◽  
Vol 2 (1) ◽  
pp. 97-110
Author(s):  
Ulis Bella ◽  
Ruli Supriati

Technological current developments more rapidly, many users to improve of information technology the way of delivering information in various ways, one of them with the website. in technology advances especially in the field of information and communication has encouraged globalization in the office and the more it needs new information, where the information is easy in the can by using web technology, namely the Internet, where information needs to be better in terms of speed, capacity and sophistication in providing information. Many people have realized that the web is not just a mere web page, but also can interact with visitors. The web has become a user interface for database applications, and Web information systems have become scattered, nowadays almost anyone can do accessing information over the web. As well as the Tangerang District Government will improve system performance through running. Likewise in the warehouse that is not computerized Web -based, but still using Microsoft excel, so that the warehouse often experience difficulties and delays in processing the data and create reports . With the web-based goods inventory system, is expected to help users interested in parts warehouse Revenue Office Tangerang District Government, and effectiveness in many ways.


Author(s):  
Abdelkrim Bouramoul

Users of Web search engines are generally confronted to numerous responses that are rarely structured, making it difficult to analyze the available results. Indeed, the linear results displayed through lists ordered according to a relevance criterion, although still widely used, seem often limitless. A solution to this problem is to improve the interfaces for better visualization of large number of results. In this paper, we propose modeling and implementation of a tool for graphical visualization and manipulation of results returned by search engines. The goal is to facilitate the analysis, the interpretation and the supervision of users' information needs. The architecture of the ‘Gravisor' tool is based on Multi-Agent paradigm. It is composed of four agents working in full cooperation and coordination. We hope that besides the web information retrieval field, the three graphical visualization modes offered by the ‘Gravisor' tool will be a promising alternative for better information visualization in other areas.


Author(s):  
Nicolas Guelfi ◽  
Cédric Pruski ◽  
Chantal Reynaud

The evolution of Web information is of utmost importance in the design of good Web Information Systems applications. New emerging paradigms, like the Semantic Web, use ontologies for describing metadata and are defined, in part, to aid in Web evolution. In this chapter, we survey techniques for ontology evolution. After identifying the different kinds of evolution with which the Web is confronted, we detail the various existing languages and techniques devoted to Web data evolution, with particular attention to Semantic Web concepts, and how these languages and techniques can be adapted to evolving data in order to improve the quality of Web Information Systems applications.


Author(s):  
Hiroaki Yamane ◽  
◽  
Masafumi Hagiwara

This paper proposes a tag line generating systemusing information extracted from the web. Tag lines sometimes attract attention even when they consist of indirect word group of the target. We use web information to extract hidden data and use several tag line corpora to collect a large number of tag lines. First, knowledge related to the input is obtained from the web. Then, the proposed system selects suitable words according to the theme. Also, model tag lines are selected from the corpora using the knowledge. By inserting nouns, verbs and adjectives into model tag lines’ structure, candidate sentences are generated. These tag line candidates are selected by the suitability as a sentence using a text N-gram corpus. The subjective experiment measures the quality of system-generated tag lines and some of them are quite comparable to human-made ones.


2021 ◽  
Vol 13 (4) ◽  
pp. 1-35
Author(s):  
Gabriel Amaral ◽  
Alessandro Piscopo ◽  
Lucie-aimée Kaffee ◽  
Odinaldo Rodrigues ◽  
Elena Simperl

Wikidata is one of the most important sources of structured data on the web, built by a worldwide community of volunteers. As a secondary source, its contents must be backed by credible references; this is particularly important, as Wikidata explicitly encourages editors to add claims for which there is no broad consensus, as long as they are corroborated by references. Nevertheless, despite this essential link between content and references, Wikidata's ability to systematically assess and assure the quality of its references remains limited. To this end, we carry out a mixed-methods study to determine the relevance, ease of access, and authoritativeness of Wikidata references, at scale and in different languages, using online crowdsourcing, descriptive statistics, and machine learning. Building on previous work of ours, we run a series of microtasks experiments to evaluate a large corpus of references, sampled from Wikidata triples with labels in several languages. We use a consolidated, curated version of the crowdsourced assessments to train several machine learning models to scale up the analysis to the whole of Wikidata. The findings help us ascertain the quality of references in Wikidata and identify common challenges in defining and capturing the quality of user-generated multilingual structured data on the web. We also discuss ongoing editorial practices, which could encourage the use of higher-quality references in a more immediate way. All data and code used in the study are available on GitHub for feedback and further improvement and deployment by the research community.


2013 ◽  
Vol 9 (1) ◽  
Author(s):  
Laerte Pereira da Silva Júnior

Resumo O portal do Centro de Ciências Humanas, Letras e Artes (CCHLA) compõe o conjunto de informações web indicadas pela Carta de Serviços ao Cidadão da Universidade Federal da Paraíba (UFPB). Esta pesquisa procura colaborar para o incremento da qualidade do uso dessa fonte de informação por meio dos Estudos de Usuários, no campo da Ciência da Informação, com base na Engenharia da Usabilidade, mais precisamente, no conceito de usabilidade, definido por Jakob Nielsen, associado aos atributos: facilidade de aprendizado, eficiência de uso, facilidade de memorização, incidência de erros e satisfação subjetiva.Palavras-chave Estudos de Usuários, Ciência da Informação, Usabilidade, Engenharia de Usabilidade, Portal do CCHLA.Abstract The homepage of the Centre for Humanities, Arts and Letters of the Federal University of Paraíba is listed among the web information sources indicated by the Univerity’s Carta de Serviços ao Cidadão. The present study intends to collaborate towards increasing the quality of this source of information through the Users Study in Information Science based on Usability Engineering, mainly the concept of usability adopted by Jakob Nielsen, associated to the atributes of learning facility, use efficiency, memorizing facility, error occurrences and subject satisfaction.Keywords Users Studies, Information Science, Usability, Usability Engineering, CCHLA Homepage.


2014 ◽  
Vol 08 (04) ◽  
pp. 389-413
Author(s):  
Moritz von Hoffen ◽  
Abdulbaki Uzun

The amount of data within the Linking Open Data (LOD) Cloud is steadily increasing and resembles a rich source of information. Since Context-aware Services (CAS) are based on the correlation of heterogeneous data sources for deriving the contextual situation of a target, it makes sense to leverage that enormous amount of data already present in the LOD Cloud to enhance the quality of these services. Within this work, the applicability of the LOD Cloud as a context provider for enriching CAS is investigated. For this purpose, a deep analysis according to the discoverability and availability of datasets is performed. Furthermore, in order to ease the process of finding a dataset that matches the information needs of a CAS developer, techniques for retrieving contents of LOD datasets are discussed and different approaches to condense the dataset to its most important concepts are shown. Finally, a Context Data Lookup Service is introduced that enables context data discovery within the LOD Cloud and its applicability is highlighted based on an example.


2011 ◽  
pp. 218-252 ◽  
Author(s):  
Guillaume Cabanac ◽  
Max Chevalier ◽  
Claude Chrisment ◽  
Christine Julien ◽  
Chantal Soulé-Dupuy ◽  
...  

Nowadays, the Web has become the most queried information source. To solve their information needs, individuals can use different types of tools or services like a search engine, for instance. Due to the high amount of information and the diversity of human factors, searching for information requires patience, perseverance, and sometimes luck. To help individuals during this task, search assistants feature adaptive techniques aiming at personalizing retrieved information. Moreover, thanks to the “new Web” (the Web 2.0), personal search assistants are evolving, using social techniques (social networks, sharing-based methods). Let us enter into the Social Web, where everyone collaborates with others in providing their experience, their expertise. This chapter introduces search assistants and underlines their evolution toward Social Information Search Assistants.


Author(s):  
Deni Darmawan ◽  
Edi Suryadi ◽  
Dinn Wahyudin

This study focuses on smart digital for mobile communication through television streaming. It makes use of the television station located in Universitas Pendidikan Indonesia (UPI) which requires further development. In 2017, research was carried out to develop web streaming and Mobile Audience through Smart Digital Mobile.  However, this study uses a research and development approach comprising of 10 main steps some of which are broadcast programs needed in education and local wisdom. The UPI TV Station gained then complete trust from all segments of viewers from the internal academic community. The content test and ease of access are expected to get an overview of the information needs of campus television by the wider community. The results of this project are as follows : (a) Syllabus and SAP Tools in Education Communication courses; (b) Evaluation of the sustainability programs of TVUPI broadcast based on the web streaming; and (c) Mobile Smart Digital Development in expanding the reach of TVUPI viewers based on web streaming.


Sign in / Sign up

Export Citation Format

Share Document