Towards An Objective Assessment Framework for Linked Data Quality

2016 ◽  
Vol 12 (3) ◽  
pp. 111-133 ◽  
Author(s):  
Ahmad Assaf ◽  
Aline Senart ◽  
Raphaël Troncy

Ensuring data quality in Linked Open Data is a complex process as it consists of structured information supported by models, ontologies and vocabularies and contains queryable endpoints and links. In this paper, the authors first propose an objective assessment framework for Linked Data quality. The authors build upon previous efforts that have identified potential quality issues but focus only on objective quality indicators that can measured regardless on the underlying use case. Secondly, the authors present an extensible quality measurement tool that helps on one hand data owners to rate the quality of their datasets, and on the other hand data consumers to choose their data sources from a ranked set. The authors evaluate this tool by measuring the quality of the LOD cloud. The results demonstrate that the general state of the datasets needs attention as they mostly have low completeness, provenance, licensing and comprehensibility quality scores.

Author(s):  
Ahmad Assaf ◽  
Aline Senart ◽  
Raphaël Troncy

Ensuring data quality in Linked Open Data is a complex process as it consists of structured information supported by models, ontologies and vocabularies and contains queryable endpoints and links. In this paper, the authors first propose an objective assessment framework for Linked Data quality. The authors build upon previous efforts that have identified potential quality issues but focus only on objective quality indicators that can measured regardless on the underlying use case. Secondly, the authors present an extensible quality measurement tool that helps on one hand data owners to rate the quality of their datasets, and on the other hand data consumers to choose their data sources from a ranked set. The authors evaluate this tool by measuring the quality of the LOD cloud. The results demonstrate that the general state of the datasets needs attention as they mostly have low completeness, provenance, licensing and comprehensibility quality scores.


Author(s):  
Syed Mustafa Ali ◽  
Farah Naureen ◽  
Arif Noor ◽  
Maged Kamel N. Boulos ◽  
Javariya Aamir ◽  
...  

Background Increasingly, healthcare organizations are using technology for the efficient management of data. The aim of this study was to compare the data quality of digital records with the quality of the corresponding paper-based records by using data quality assessment framework. Methodology We conducted a desk review of paper-based and digital records over the study duration from April 2016 to July 2016 at six enrolled TB clinics. We input all data fields of the patient treatment (TB01) card into a spreadsheet-based template to undertake a field-to-field comparison of the shared fields between TB01 and digital data. Findings A total of 117 TB01 cards were prepared at six enrolled sites, whereas just 50% of the records (n=59; 59 out of 117 TB01 cards) were digitized. There were 1,239 comparable data fields, out of which 65% (n=803) were correctly matched between paper based and digital records. However, 35% of the data fields (n=436) had anomalies, either in paper-based records or in digital records. 1.9 data quality issues were calculated per digital patient record, whereas it was 2.1 issues per record for paper-based record. Based on the analysis of valid data quality issues, it was found that there were more data quality issues in paper-based records (n=123) than in digital records (n=110). Conclusion There were fewer data quality issues in digital records as compared to the corresponding paper-based records. Greater use of mobile data capture and continued use of the data quality assessment framework can deliver more meaningful information for decision making.


2019 ◽  
Vol 29 ◽  
pp. e10194
Author(s):  
Camila Andrea Herrera-Melo ◽  
Juan Sebastián González Sanabria

The provision of portals that serve as a source of access and availability of public domain data is part of the adoption of public policies that some government entities have implemented in response to the establishment of an open, transparent, multidirectional, collaborative and focused on citizen participation government, both in monitoring and in making public decisions. However, the publication of this data must meet certain characteristics to be considered open and of quality. For this reason, studies arise that focus on the approach of methodologies and indicators that measure the quality of the portals and their data. For the aim of this paper, the search of referential sources of the last six years regarding the evaluation of data quality and open data portals in Spain, Brazil, Costa Rica, Taiwan and the European Union was carried out with the objective of gathering the necessary inputs for the approach of the methodology presented in the document.


2005 ◽  
Vol 24 (S1) ◽  
pp. 153-170 ◽  
Author(s):  
Leslie L. Roos ◽  
Sumit Gupta ◽  
Ruth-Ann Soodeen ◽  
Laurel Jebamani

ABSTRACTThis review evaluates the quality of available administrative data in the Canadian provinces, emphasizing the information needed to create integrated systems. We explicitly compare approaches to quality measurement, indicating where record linkage can and cannot substitute for more expensive record re-abstraction. Forty-nine original studies evaluating Canadian administrative data (registries, hospital abstracts, physician claims, and prescription drugs) are summarized in a structured manner. Registries, hospital abstracts, and physician files appear to be generally of satisfactory quality, though much work remains to be done. Data quality did not vary systematically among provinces. Primary data collection to check place of residence and longitudinal follow-up in provincial registries is needed. Promising initial checks of pharmaceutical data should be expanded. Because record linkage studies were “conservative” in reporting reliability, the reduction of time-consuming record re-abstraction appears feasible in many cases. Finally, expanding the scope of administrative data to study health, as well as health care, seems possible for some chronic conditions. The research potential of the information-rich environments being created highlights the importance of data quality.


2000 ◽  
Vol 13 (4) ◽  
pp. 205-215 ◽  
Author(s):  
A. Curry ◽  
S. Stark

This paper illustrates the relevance of using the Servqual instrument as a service quality measurement tool in the healthcare context. The expectations and perceptions of nursing home residents and their relatives are compared and contrasted to determine the priorities and satisfaction levels of both groups with respect to the services provided. Some interesting differences emerged between the two respondent groups, which gave more accurate and relevant pointers as to potential areas of future improvement. Because a number of nursing homes were surveyed, it was also possible to ascertain which homes might serve as useful examples of the best practice and which were in need of improvement action. In frail, elderly respondent groups it is not especially easy to elicit views as to quality of service. Servqual, carefully and sympathetically designed and applied, proved appropriate for this task.


Author(s):  
Abdulbaki Uzun ◽  
Eric Neidhardt ◽  
Axel Küpper

Mobile network operators maintain data about their mobile network topology, which is mainly used for network provisioning and planning purposes restricting its full business potential. Utilizing this data in combination with the extensive pool of semantically modeled data in the Linking Open Data Cloud, innovative applications can be realized that would establish network operators as service providers and enablers in the highly competitive services market. In this article, the authors introduce the OpenMobileNetwork (available at http://www.openmobilenetwork.org/) as an open solution for providing approximated network topology data based on the principles of Linked Data along with a business concept for network operators to exploit their valuable asset. Since the quality of the estimated network topology is crucial when providing services on top of it, the authors further analyze and evaluate state-of-the-art approaches for estimating base station positions out of crowdsourced data and discuss the results in comparison to real base station locations.


Semantic Web ◽  
2017 ◽  
Vol 9 (1) ◽  
pp. 77-129 ◽  
Author(s):  
Michael Färber ◽  
Frederic Bartscherer ◽  
Carsten Menne ◽  
Achim Rettinger
Keyword(s):  

Author(s):  
Sarah Rees ◽  
Arfon Rees

ABSTRACTObjectivesThe SAIL databank brings together a range of datasets gathered primarily for administrative rather than research processes. These datasets contain information regarding different aspects of an individual’s contact with services which when combined form a detailed health record for individuals living (or deceased) in Wales. Understanding the quality of data in SAIL supports the research process by providing a level of assurance about the robustness of data, identifying and describing where there may be sources of potential bias due to invalid, incomplete, inconsistent or inaccurate data and therefore helping to increase the accuracy of research using these data. Designing processes to investigate and report on data quality within and between multiple datasets can be a time-consuming task to undertake; it requires a high degree of effort to ensure it is genuinely meaningful and useful to SAIL users and may require a range of different approaches. ApproachData quality tests for each dataset were written, considering a range of data quality dimensions including validity, consistency, accuracy and completeness. Tests were designed to capture not just the quality of data within each dataset, but also to assess consistency of data items between datasets. SQL scripts were written to test each of these aspects: in order to minimise repetition, automated processes were implemented where appropriate. Batch automation was used to called SQL stored procedures, which utilise metadata to generate dynamic SQL. The metadata (created as part of the data quality process) describes each dataset and the measurement parameters used to assess each field within the dataset. However automation on its own is insufficient and data quality process outputs require scrutiny and oversight to ensure they are actually capturing what they set out to do. SAIL users were consulted on the development of the data quality reports to ensure usability and appropriateness to support data utilisation for research. ResultsThe data quality reporting process is beneficial to the SAIL databank as it provides additional information to support the research process and in some cases may act as a diagnostic tool, detecting problems with data which can then be rectified. ConclusionThe development of data quality processes in SAIL is ongoing, and changes or developments in each dataset lead to new requirements for data quality measurement and reporting. A vital component of the process is the production of output that is genuinely meaningful and useful.


2019 ◽  
Vol 1 ◽  
pp. ed1
Author(s):  
Shaun Yon-Seng Khoo

Almost every open access neuroscience journal is pay-to-publish. This leaves neuroscientists with a choice of submitting to journals that not all of our colleagues can legitimately access and choosing to pay large sums of money to publish open access. Neuroanatomy and Behaviour is a new platinum open access journal published by a non-profit association of scientists. Since we do not charge fees, we will focus entirely on the quality of submitted articles and encourage the adoption of reproducibility-enhancing practices, like open data, preregistration, and data quality checks. We hope that our colleagues will join us in this endeavour so that we can support good neuroscience no matter where it comes from.


2019 ◽  
Vol 1 (2) ◽  
pp. 21-39
Author(s):  
Dewa Nyoman Benni Kusyana ◽  
Komang Ary Pratiwi

Service quality measurement techniques along with dimensions of service quality have become a major issue in the marketing literature over the past few decades. The increasing importance of quality services, both practitioners and scholars began to focus on service delivery. Measurement of service quality is now seen as an important measurement tool for companies to understand the needs and desires of customers by analyzing customer satisfaction and experience with the services provided. Although until now there has been no agreement on a particular model that should be used to measure service quality, the fact is there are several effective models that can be used by both practitioners and scholars that have been produced by researchers. Until now, researchers believe and agree that service quality is multilevel or multidimensional. This study focuses on various types of service quality models that existed until year of 2000. The methodology used is conducting a review of service quality literature. The limitation of this literature review is that the existing service quality model was developed before 2000 because after 2000 the development of quality services was focused on the quality of electronic services.


Sign in / Sign up

Export Citation Format

Share Document