scholarly journals The knowledge graph as the default data model for learning on heterogeneous knowledge

Data Science ◽  
2017 ◽  
Vol 1 (1-2) ◽  
pp. 39-57 ◽  
Author(s):  
Xander Wilcke ◽  
Peter Bloem ◽  
Victor de Boer
2020 ◽  
Vol 34 (10) ◽  
pp. 13953-13954
Author(s):  
Xu Wang ◽  
Shuai Zhao ◽  
Bo Cheng ◽  
Jiale Han ◽  
Yingting Li ◽  
...  

Multi-hop question answering models based on knowledge graph have been extensively studied. Most existing models predict a single answer with the highest probability by ranking candidate answers. However, they are stuck in predicting all the right answers caused by the ranking method. In this paper, we propose a novel model that converts the ranking of candidate answers into individual predictions for each candidate, named heterogeneous knowledge graph based multi-hop and multi-answer model (HGMAN). HGMAN is capable of capturing more informative representations for relations assisted by our heterogeneous graph, which consists of multiple entity nodes and relation nodes. We rely on graph convolutional network for multi-hop reasoning and then binary classification for each node to get multiple answers. Experimental results on MetaQA dataset show the performance of our proposed model over all baselines.


Life ◽  
2021 ◽  
Vol 11 (1) ◽  
pp. 42
Author(s):  
Charlotte A. Nelson ◽  
Ana Uriarte Acuna ◽  
Amber M. Paul ◽  
Ryan T. Scott ◽  
Atul J. Butte ◽  
...  

There has long been an interest in understanding how the hazards from spaceflight may trigger or exacerbate human diseases. With the goal of advancing our knowledge on physiological changes during space travel, NASA GeneLab provides an open-source repository of multi-omics data from real and simulated spaceflight studies. Alone, this data enables identification of biological changes during spaceflight, but cannot infer how that may impact an astronaut at the phenotypic level. To bridge this gap, Scalable Precision Medicine Oriented Knowledge Engine (SPOKE), a heterogeneous knowledge graph connecting biological and clinical data from over 30 databases, was used in combination with GeneLab transcriptomic data from six studies. This integration identified critical symptoms and physiological changes incurred during spaceflight.


2021 ◽  
Vol 16 (1) ◽  
Author(s):  
Qian Zhu ◽  
Ðắc-Trung Nguyễn ◽  
Timothy Sheils ◽  
Gioconda Alyea ◽  
Eric Sid ◽  
...  

Abstract Background Limited knowledge and unclear underlying biology of many rare diseases pose significant challenges to patients, clinicians, and scientists. To address these challenges, there is an urgent need to inspire and encourage scientists to propose and pursue innovative research studies that aim to uncover the genetic and molecular causes of more rare diseases and ultimately to identify effective therapeutic solutions. A clear understanding of current research efforts, knowledge/research gaps, and funding patterns as scientific evidence is crucial to systematically accelerate the pace of research discovery in rare diseases, which is an overarching goal of this study. Methods To semantically represent NIH funding data for rare diseases and advance its use of effectively promoting rare disease research, we identified NIH funded projects for rare diseases by mapping GARD diseases to the project based on project titles; subsequently we presented and managed those identified projects in a knowledge graph using Neo4j software, hosted at NCATS, based on a pre-defined data model that captures semantics among the data. With this developed knowledge graph, we were able to perform several case studies to demonstrate scientific evidence generation for supporting rare disease research discovery. Results Of 5001 rare diseases belonging to 32 distinct disease categories, we identified 1294 diseases that are mapped to 45,647 distinct, NIH-funded projects obtained from the NIH ExPORTER by implementing semantic annotation of project titles. To capture semantic relationships presenting amongst mapped research funding data, we defined a data model comprised of seven primary classes and corresponding object and data properties. A Neo4j knowledge graph based on this predefined data model has been developed, and we performed multiple case studies over this knowledge graph to demonstrate its use in directing and promoting rare disease research. Conclusion We developed an integrative knowledge graph with rare disease funding data and demonstrated its use as a source from where we can effectively identify and generate scientific evidence to support rare disease research. With the success of this preliminary study, we plan to implement advanced computational approaches for analyzing more funding related data, e.g., project abstracts and PubMed article abstracts, and linking to other types of biomedical data to perform more sophisticated research gap analysis and identify opportunities for future research in rare diseases.


Author(s):  
Yuting Wu ◽  
Xiao Liu ◽  
Yansong Feng ◽  
Zheng Wang ◽  
Rui Yan ◽  
...  

Entity alignment is the task of linking entities with the same real-world identity from different knowledge graphs (KGs), which has been recently dominated by embedding-based methods. Such approaches work by learning KG representations so that entity alignment can be performed by measuring the similarities between entity embeddings. While promising, prior works in the field often fail to properly capture complex relation information that commonly exists in multi-relational KGs, leaving much room for improvement. In this paper, we propose a novel Relation-aware Dual-Graph Convolutional Network (RDGCN) to incorporate relation information via attentive interactions between the knowledge graph and its dual relation counterpart, and further capture neighboring structures to learn better entity representations. Experiments on three real-world cross-lingual datasets show that our approach delivers better and more robust results over the state-of-the-art alignment methods by learning better KG representations.


Semantic Web ◽  
2021 ◽  
pp. 1-23
Author(s):  
Steven J. Baskauf ◽  
Jessica K. Baskauf

The W3C Generating RDF from Tabular Data on the Web Recommendation provides a mechanism for mapping CSV-formatted data to any RDF graph model. Since the Wikibase data model used by Wikidata can be expressed as RDF, this Recommendation can be used to document tabular snapshots of parts of the Wikidata knowledge graph in a simple form that is easy for humans and applications to read. Those snapshots can be used to document how subgraphs of Wikidata have changed over time and can be compared with the current state of Wikidata using its Query Service to detect vandalism and value added through community contributions.


2021 ◽  
Vol 2 (3) ◽  
pp. 336-347
Author(s):  
Ariam Rivas ◽  
Irlan Grangel-Gonzalez ◽  
Diego Collarana ◽  
Jens Lehmann ◽  
Maria-esther Vidal

Industry 4.0 (I4.0) standards and standardization frameworks provide a unified way to describe smart factories. Standards specify the main components, systems, and processes inside a smart factory and the interaction among all of them. Furthermore, standardization frameworks classify standards according to their functions into layers and dimensions. Albeit informative, frameworks can categorize similar standards differently. As a result, interoperability conflicts are generated whenever smart factories are described with miss-classified standards. Approaches like ontologies and knowledge graphs enable the integration of standards and frameworks in a structured way. They also encode the meaning of the standards, known relations among them, as well as their classification according to existing frameworks. This structured modeling of the I4.0 landscape using a graph data model provides the basis for graph-based analytical methods to uncover alignments among standards. This paper contributes to analyzing the relatedness among standards and frameworks; it presents an unsupervised approach for discovering links among standards. The proposed method resorts to knowledge graph embeddings to determine relatedness among standards-based on similarity metrics. The proposed method is agnostic to the technique followed to create the embeddings and to the similarity measure. Building on the similarity values, community detection algorithms can automatically create communities of highly similar standards. Our approach follows the homophily principle, and assumes that related standards are together in a community. Thus, alignments across standards are predicted and interoperability issues across them are solved. We empirically evaluate our approach on a knowledge graph of 249 I4.0 standards using the Trans$^*$ family of embedding models for knowledge graph entities. Our results are promising and suggest that relations among standards can be detected accurately.


2008 ◽  
Author(s):  
Pedro J. M. Passos ◽  
Duarte Araujo ◽  
Keith Davids ◽  
Ana Diniz ◽  
Luis Gouveia ◽  
...  

Sign in / Sign up

Export Citation Format

Share Document