scholarly journals Ontologies and Data Management: A Brief Survey

2020 ◽  
Vol 34 (3) ◽  
pp. 329-353 ◽  
Author(s):  
Thomas Schneider ◽  
Mantas Šimkus

Abstract Information systems have to deal with an increasing amount of data that is heterogeneous, unstructured, or incomplete. In order to align and complete data, systems may rely on taxonomies and background knowledge that are provided in the form of an ontology. This survey gives an overview of research work on the use of ontologies for accessing incomplete and/or heterogeneous data.

Author(s):  
Adolphe Ayissi Eteme ◽  
Justin Moskolai Ngossaha

The use of information technology in council management has resulted in the generation of a large amount of data through various autonomous urban bodies. The relevant bodies barely or never reuse such locally-generated data. This may be due particularly to managers', policy makers' and users' lack of awareness of existing information. The Platform for the Integration and Interoperability of the Yaounde Urban Information Systems (YUSIIP) project seeks to reduce this deficit by establishing a federated operational platform of heterogeneous and distributed data systems based on a distributed data repository. The position developed in this paper is that Master Data Management (MDM) will contribute to achieving this objective in a context marked by the dispersion and duplication of data and diversity of information systems.


2018 ◽  
pp. 154-178
Author(s):  
Adolphe Ayissi Eteme ◽  
Justin Moskolai Ngossaha

The use of information technology in council management has resulted in the generation of a large amount of data through various autonomous urban bodies. The relevant bodies barely or never reuse such locally-generated data. This may be due particularly to managers', policy makers' and users' lack of awareness of existing information. The Platform for the Integration and Interoperability of the Yaounde Urban Information Systems (YUSIIP) project seeks to reduce this deficit by establishing a federated operational platform of heterogeneous and distributed data systems based on a distributed data repository. The position developed in this paper is that Master Data Management (MDM) will contribute to achieving this objective in a context marked by the dispersion and duplication of data and diversity of information systems.


Author(s):  
Е.Е. ДЕВЯТКИН ◽  
М.В. ИВАНКОВИЧ ◽  
М.Н. КУПИН

Представлены результаты исследовательской работы по унификации форматов данных,которые требуются при реализации аналитической составляющей в структуре информационных систем, предназначенных для решения экспертно-прогностических задач в сфере связи и вещания. По результатам анализа существующих в отрасли основных автоматизированных систем, их функциональных возможностей обоснована необходимость введения в действующие и перспективные системы аналитической составляющей и унификации форматов данных как непременного условия их взаимодействия. Рассматривается набор унифицированных форматов данных, разработанный для перспективного макета аналитической системы развития связи и вещания. The results of research work on unifying the data formats that are required for the implementation of the analytical component in the structure of information systems intended for the solution of expert predictive problems in the sphere of communications and broadcasting are presented. Based on the results of the analysis of the main automated systems and their functional capabilities the necessity of introducing an analytical component into existing and future systems as well as unifying data formats as an indispensable condition for their interaction is substantiated. A set of unified data formats developed for a promising layout of an analytical system for the development of communications and broadcasting is considered.


AI Magazine ◽  
2015 ◽  
Vol 36 (1) ◽  
pp. 75-86 ◽  
Author(s):  
Jennifer Sleeman ◽  
Tim Finin ◽  
Anupam Joshi

We describe an approach for identifying fine-grained entity types in heterogeneous data graphs that is effective for unstructured data or when the underlying ontologies or semantic schemas are unknown. Identifying fine-grained entity types, rather than a few high-level types, supports coreference resolution in heterogeneous graphs by reducing the number of possible coreference relations that must be considered. Big data problems that involve integrating data from multiple sources can benefit from our approach when the datas ontologies are unknown, inaccessible or semantically trivial. For such cases, we use supervised machine learning to map entity attributes and relations to a known set of attributes and relations from appropriate background knowledge bases to predict instance entity types. We evaluated this approach in experiments on data from DBpedia, Freebase, and Arnetminer using DBpedia as the background knowledge base.


1999 ◽  
Vol 33 (3) ◽  
pp. 55-66 ◽  
Author(s):  
L. Charles Sun

An interactive data access and retrieval system, developed at the U.S. National Oceanographic Data Genter (NODG) and available at <ext-link ext-link-type="uri" href="http://www.node.noaa.gov">http://www.node.noaa.gov</ext-link>, is presented in this paper. The purposes of this paper are: (1) to illustrate the procedures of quality control and loading oceanographic data into the NODG ocean databases and (2) to describe the development of a system to manage, visualize, and disseminate the NODG data holdings over the Internet. The objective of the system is to provide ease of access to data that will be required for data assimilation models. With advances in scientific understanding of the ocean dynamics, data assimilation models require the synthesis of data from a variety of resources. Modern intelligent data systems usually involve integrating distributed heterogeneous data and information sources. As the repository for oceanographic data, NOAA’s National Oceanographic Data Genter (NODG) is in a unique position to develop such a data system. In support of the data assimilation needs, NODG has developed a system to facilitate browsing of the oceanographic environmental data and information that is available on-line at NODG. Users may select oceanographic data based on geographic areas, time periods and measured parameters. Once the selection is complete, users may produce a station location plot, produce plots of the parameters or retrieve the data.


2007 ◽  
Vol 40 (4) ◽  
pp. 2070 ◽  
Author(s):  
S. Vassilopoulou ◽  
K. Chousianitis ◽  
V. Sakkas ◽  
B. Damiata ◽  
E. Lagios

The present study is concerned with the management of multi-thematic geo-data of Cephallonia Island, related to crustal deformation. A large amount of heterogeneous data (vector, raster, ascii files) involving geology, tectonics, topography, geomorphology and DGPS measurements was compiled. Crustal deformation was studied using GPS network consisting of '23 stations. This was installed and measured in October 2001 and re-measured during September 2003 following the Lefkas earthquake of August 2003 (Mw=6.2), and also in July 2006. With proper spatial analysis, a large number of thematic and synthetic layers and maps were produced. Simultaneously, a GIS Data base was organized in order to make an easy extraction of conclusions in specific questions.


Organon ◽  
2011 ◽  
Vol 25 (50) ◽  
Author(s):  
Anna Maria Becker Maciel

This paper reports on the quest for the interface between Terminology, Informaticsand the Termisul Project. Three landmarks highlight the account, DOS, Windows, and Internet.They characterize the steps of a team of linguists towards the computerized era. As far as thetechnology advanced, their perception of digital resources changed from being mereoperational help in data management to representing sophisticated tools assisting research andeventually to the demand of cooperative work with computer scientists. The difficulties of theearly stages encountered in the path tread are remembered. The meaningful results of theefforts taken are described. The challenges of the ongoing research work are outlined.


Author(s):  
Amanda Leanne Butler ◽  
Mark Smith ◽  
Wayne Jones ◽  
Carol E Adair ◽  
Simone Vigod ◽  
...  

BackgroundCanada has a publicly-funded universal healthcare system with information systems managed by 13 different provinces and territories. This context creates inconsistencies in data collection and challenges for research or surveillance conducted at the national or multi-jurisdictional level. ObjectiveUsing a recent Canadian research project as a case study, we document the strengths and challenges of using administrative health data in a multi-jurisdictional context. We discuss the implications of using different health information systems and the solutions we adopted to deal with variations. Our goal is to contribute to better understanding of these challenges and the development of a more integrated and harmonized approach to conducting multi-jurisdictional research using administrative data. Context and ModelUsing data from five separate provincial healthcare data systems, we sought to create and report on a set of provincially-comparable mental health and addiction services performance indicators. In this paper, we document the research process, challenges, and solutions. Finally, we conclude by making recommendations for investment in national infrastructure that could help cut costs, broaden scope, and increase use of administrative health data that exists in Canada. ConclusionCanada has an incredible wealth of administrative data that resides in 13 territorial and provincial government systems. Navigating access and improving comparability across these systems has been an ongoing challenge for the past 20 years, but progress is being made. We believe that with some investment, a more harmonized and integrated information network could be developed that supports a broad range of surveillance and research activities with strong policy and program implications.


Sign in / Sign up

Export Citation Format

Share Document