scholarly journals Inspecting the concept knowledge graph encoded by modern language models

Author(s):  
Carlos Aspillaga ◽  
Marcelo Mendoza ◽  
Alvaro Soto
Author(s):  
Bosung Kim ◽  
Taesuk Hong ◽  
Youngjoong Ko ◽  
Jungyun Seo

2021 ◽  
Vol 72 ◽  
pp. 1343-1384
Author(s):  
Vassilina Nikoulina ◽  
Maxat Tezekbayev ◽  
Nuradil Kozhakhmet ◽  
Madina Babazhanova ◽  
Matthias Gallé ◽  
...  

There is an ongoing debate in the NLP community whether modern language models contain linguistic knowledge, recovered through so-called probes. In this paper, we study whether linguistic knowledge is a necessary condition for the good performance of modern language models, which we call the rediscovery hypothesis. In the first place, we show that language models that are significantly compressed but perform well on their pretraining objectives retain good scores when probed for linguistic structures. This result supports the rediscovery hypothesis and leads to the second contribution of our paper: an information-theoretic framework that relates language modeling objectives with linguistic information. This framework also provides a metric to measure the impact of linguistic information on the word prediction task. We reinforce our analytical results with various experiments, both on synthetic and on real NLP tasks in English.


2021 ◽  
Vol 21 (S9) ◽  
Author(s):  
Yinyu Lan ◽  
Shizhu He ◽  
Kang Liu ◽  
Xiangrong Zeng ◽  
Shengping Liu ◽  
...  

Abstract Background Knowledge graphs (KGs), especially medical knowledge graphs, are often significantly incomplete, so it necessitating a demand for medical knowledge graph completion (MedKGC). MedKGC can find new facts based on the existed knowledge in the KGs. The path-based knowledge reasoning algorithm is one of the most important approaches to this task. This type of method has received great attention in recent years because of its high performance and interpretability. In fact, traditional methods such as path ranking algorithm take the paths between an entity pair as atomic features. However, the medical KGs are very sparse, which makes it difficult to model effective semantic representation for extremely sparse path features. The sparsity in the medical KGs is mainly reflected in the long-tailed distribution of entities and paths. Previous methods merely consider the context structure in the paths of knowledge graph and ignore the textual semantics of the symbols in the path. Therefore, their performance cannot be further improved due to the two aspects of entity sparseness and path sparseness. Methods To address the above issues, this paper proposes two novel path-based reasoning methods to solve the sparsity issues of entity and path respectively, which adopts the textual semantic information of entities and paths for MedKGC. By using the pre-trained model BERT, combining the textual semantic representations of the entities and the relationships, we model the task of symbolic reasoning in the medical KG as a numerical computing issue in textual semantic representation. Results Experiments results on the publicly authoritative Chinese symptom knowledge graph demonstrated that the proposed method is significantly better than the state-of-the-art path-based knowledge graph reasoning methods, and the average performance is improved by 5.83% for all relations. Conclusions In this paper, we propose two new knowledge graph reasoning algorithms, which adopt textual semantic information of entities and paths and can effectively alleviate the sparsity problem of entities and paths in the MedKGC. As far as we know, it is the first method to use pre-trained language models and text path representations for medical knowledge reasoning. Our method can complete the impaired symptom knowledge graph in an interpretable way, and it outperforms the state-of-the-art path-based reasoning methods.


2020 ◽  
Vol 34 (05) ◽  
pp. 7911-7918
Author(s):  
Hiroaki Hayashi ◽  
Zecong Hu ◽  
Chenyan Xiong ◽  
Graham Neubig

In this paper, we propose Latent Relation Language Models (LRLMs), a class of language models that parameterizes the joint distribution over the words in a document and the entities that occur therein via knowledge graph relations. This model has a number of attractive properties: it not only improves language modeling performance, but is also able to annotate the posterior probability of entity spans for a given text through relations. Experiments demonstrate empirical improvements over both word-based language models and a previous approach that incorporates knowledge graph information. Qualitative analysis further demonstrates the proposed model's ability to learn to predict appropriate relations in context. 1


2020 ◽  
Author(s):  
Charlotte Caucheteux ◽  
Jean-Rémi King

AbstractDeep learning has recently allowed substantial progress in language tasks such as translation and completion. Do such models process language similarly to humans, and is this similarity driven by systematic structural, functional and learning principles? To address these issues, we tested whether the activations of 7,400 artificial neural networks trained on image, word and sentence processing linearly map onto the hierarchy of human brain responses elicited during a reading task, using source-localized magneto-encephalography (MEG) recordings of one hundred and four subjects. Our results confirm that visual, word and language models sequentially correlate with distinct areas of the left-lateralized cortical hierarchy of reading. However, only specific subsets of these models converge towards brain-like representations during their training. Specifically, when the algorithms are trained on language modeling, their middle layers become increasingly similar to the late responses of the language network in the brain. By contrast, input and output word embedding layers often diverge away from brain activity during training. These differences are primarily rooted in the sustained and bilateral responses of the temporal and frontal cortices. Together, these results suggest that the compositional - but not the lexical - representations of modern language models converge to a brain-like solution.


2020 ◽  
Author(s):  
Stéphane Aroca-Ouellette ◽  
Frank Rudzicz

2020 ◽  
Vol 142 (10) ◽  
Author(s):  
Xinyu Li ◽  
Chun-Hsien Chen ◽  
Pai Zheng ◽  
Zuoxu Wang ◽  
Zuhua Jiang ◽  
...  

Abstract In order to meet user expectations and to optimize user experience with a higher degree of flexibility and sustainability, the Smart product–service system (Smart PSS), as a novel value proposition paradigm considering both online and offline smartness, was proposed. However, conventional manners for developing PSS require many professional consultations and still cannot meet with the new features of Smart PSS, such as user context-awareness and ever-evolving knowledge management. Therefore, aiming to assist Smart PSS development cost-effectively, this paper adopted the knowledge graph (KG) technique and concept–knowledge (C-K) model to propose an evolutionary design approach. Two knowledge graphs are firstly established with open-source knowledge, prototype specifications, and user-generated textual data. Then, triggered by personalized requirements, four KG-aided C-K operators are conducted based on graph-based query patterns and computational linguistics algorithms, thus generating innovative solutions for evolving Smart PSS. To validate the performance of the proposed approach, a case study of a smart nursing bed fulfilling multiple personalized requirements is conducted, and the evaluation result of its knowledge evolution is acceptable. It hopes that this work can offer insightful guidance to industrial organizations in their development of Smart PSS.


Author(s):  
Junyi Li ◽  
Tianyi Tang ◽  
Wayne Xin Zhao ◽  
Zhicheng Wei ◽  
Nicholas Jing Yuan ◽  
...  

2021 ◽  
Author(s):  
Melania Nițu ◽  
◽  
Mihai Dascălu ◽  
Gabriel Guțu-Robu ◽  
Maria-Iuliana Dascălu ◽  
...  

Author(s):  
Alexander Meduna ◽  
Ondřej Soukup

Sign in / Sign up

Export Citation Format

Share Document