A Relation Pattern-Driven Probability Model for Related Entity Retrieval

2012 ◽  
Vol 3 (1) ◽  
pp. 64-77
Author(s):  
Peng Jiang ◽  
Qing Yang ◽  
Chunxia Zhang ◽  
Zhendong Niu ◽  
Hongping Fu

As the Web is becoming the largest knowledge repository which contains various entities and their relations, the task of related entity retrieval excites interest in the field of information retrieval. This challenging task is introduced in TREC 2009 Entity Track. In this task, given an entity and the type of the target entity, a retrieval system is required to return a ranked list of related entities extracted from a given large corpus. It means that entity ranking goes beyond entity relevance and integrates the judgment of relation into the evaluation of the retrieved entities. This paper proposes a probability model using relation patterns to address the task of related entity retrieval. This model takes into account both relevance and relation between entities. The authors focus on using relation patterns to measure the level of relations matching between entities, and then to estimate the probability of occurrence of relation between two entities. In addition, the authors represent entity by its context language model and measure the relevance between two entities by a language model. Experimental results on TREC Entity Track dataset show that the proposed model significantly improves retrieval performances over baseline. The comparison with other approaches also reveals the effectiveness of the model.

2020 ◽  
Vol 4 (3) ◽  
pp. 551-557
Author(s):  
Muhammad zaky ramadhan ◽  
Kemas Muslim Lhaksmana

Hadith has several levels of authenticity, among which are weak (dhaif), and fabricated (maudhu) hadith that may not originate from the prophet Muhammad PBUH, and thus should not be considered in concluding an Islamic law (sharia). However, many such hadiths have been commonly confused as authentic hadiths among ordinary Muslims. To easily distinguish such hadiths, this paper proposes a method to check the authenticity of a hadith by comparing them with a collection of fabricated hadiths in Indonesian. The proposed method applies the vector space model and also performs spelling correction using symspell to check whether the use of spelling check can improve the accuracy of hadith retrieval, because it has never been done in previous works and typos are common on Indonesian-translated hadiths on the Web and social media raw text. The experiment result shows that the use of spell checking improves the mean average precision and recall to become 81% (from 73%) and 89% (from 80%), respectively. Therefore, the improvement in accuracy by implementing spelling correction make the hadith retrieval system more feasible and encouraged to be implemented in future works because it can correct typos that are common in the raw text on the Internet.


2021 ◽  
Vol 15 (6) ◽  
pp. 1-22
Author(s):  
Yashen Wang ◽  
Huanhuan Zhang ◽  
Zhirun Liu ◽  
Qiang Zhou

For guiding natural language generation, many semantic-driven methods have been proposed. While clearly improving the performance of the end-to-end training task, these existing semantic-driven methods still have clear limitations: for example, (i) they only utilize shallow semantic signals (e.g., from topic models) with only a single stochastic hidden layer in their data generation process, which suffer easily from noise (especially adapted for short-text etc.) and lack of interpretation; (ii) they ignore the sentence order and document context, as they treat each document as a bag of sentences, and fail to capture the long-distance dependencies and global semantic meaning of a document. To overcome these problems, we propose a novel semantic-driven language modeling framework, which is a method to learn a Hierarchical Language Model and a Recurrent Conceptualization-enhanced Gamma Belief Network, simultaneously. For scalable inference, we develop the auto-encoding Variational Recurrent Inference, allowing efficient end-to-end training and simultaneously capturing global semantics from a text corpus. Especially, this article introduces concept information derived from high-quality lexical knowledge graph Probase, which leverages strong interpretability and anti-nose capability for the proposed model. Moreover, the proposed model captures not only intra-sentence word dependencies, but also temporal transitions between sentences and inter-sentence concept dependence. Experiments conducted on several NLP tasks validate the superiority of the proposed approach, which could effectively infer meaningful hierarchical concept structure of document and hierarchical multi-scale structures of sequences, even compared with latest state-of-the-art Transformer-based models.


2018 ◽  
Vol 33 (4) ◽  
Author(s):  
Murari Kumar ◽  
Samir Farooqi ◽  
K. K. Chaturvedi ◽  
Chandan Kumar Deb ◽  
Pankaj Das

Bibliographic data contains necessary information about literature to help users to recognize and retrieve that resource. These data are used quantitatively by a “Bibliometrician” for analysis and dissemination purpose but with the increasing rate of literature publication in open access journals such as Nucleic Acids Research (NAR), Springer, Oxford Journals etc., it has become difficult to retrieve structured bibliographic information in desired format. A digital bibliographic database contains necessary and structured information about published literature. Bibliographic records of different articles are scattered and resides on different web pages. This thesis presents the retrieval system for bibliographic data of NAR at a single place. For this purpose, parser agents have been developed which access the web pages of NAR and parse the scattered bibliographic data and finally store it into a local bibliographic database. Based on the bibliographic database, “three-tier architecture” has been utilized to display the bibliographic information in systematized format. Using this system, it would be possible to build the network between different authors and affiliations and also other analytical reports can be generated.


Author(s):  
Mohammad Amin Hariri-Ardebili

Risk analysis of concrete dams and quantification of the failure probability are important tasks in dam safety assessment. The conditional probability of demand and capacity is usually estimated by numerical simulation and Monte Carlo technique. However, the estimated failure probability (or the reliability index) is dam-dependent which makes its application limited to some case studies. This article proposes an analytical failure model for generic gravity dam classes which is optimized based on large number of nonlinear finite element analyses. A hybrid parametric–probabilistic–statistical approach is used to estimate the failure probability as a function of dam size, material distributional models and external hydrological hazard. The proposed model can be used for preliminary design and evaluation of two-dimensional gravity dam models.


Author(s):  
Ricardo Baeza-Yates ◽  
Roi Blanco ◽  
Malú Castellanos

Web search has become a ubiquitous commodity for Internet users. This fact puts a large number of documents with plenty of text content at our fingertips. To make good use of this data, we need to mine web text. This triggers the two problems covered here: sentiment analysis and entity retrieval in the context of the Web. The first problem answers the question of what people think about a given product or a topic, in particular sentiment analysis in social media. The second problem addresses the issue of solving certain enquiries precisely by returning a particular object: for instance, where the next concert of my favourite band will be or who the best cooks are in a particular region. Where to find these objects and how to retrieve, rank, and display them are tasks related to the entity retrieval problem.


Author(s):  
Sazid Zaman Khan ◽  
Alan Colman ◽  
Iqbal H. Sarker

A large number of smart devices (things) are being deployed with the swift development of Internet of Things (IOT). These devices, owned by different organizations, have a wide variety of services to offer over the web. During a natural disaster or emergency (i.e., a situation), for example, relevant IOT services can be found and put to use. However, appropriate service matching methods are required to find the relevant services. Organizations that manage situation responses and organizations that provide IOT services are likely to be independent of each other, and therefore it is difficult for them to adopt a common ontological model to facilitate the service matching. Moreover, there exists a large conceptual gap between the domain of discourse for situations and the domain of discourse for services, which cannot be adequately bridged by existing techniques. In this paper, we address these issues and propose a new method, WikiServe, to identify IOT services that are functionally relevant to a given situation. Using concepts (terms) from situation and service descriptions, WikiServe employs Wikipedia as a knowledge source to bridge the conceptual gap between situation and service descriptions and match functionally relevant IOT services for a situation. It uses situation terms to retrieve situation related articles from Wikipedia. Then it creates a ranked list of services for the situation using the weighted occurrences of service terms in weighted situation articles. WikiServe performs better than a commonly used baseline method in terms of Precision, Recall and F measure for service matching.


2021 ◽  
Vol 15 (04) ◽  
pp. 487-510
Author(s):  
Prakhar Mishra ◽  
Chaitali Diwan ◽  
Srinath Srinivasa ◽  
G. Srinivasaraghavan

To create curiosity and interest for a topic in online learning is a challenging task. A good preview that outlines the contents of a learning pathway could help learners know the topic and get interested in it. Towards this end, we propose a hierarchical title generation approach to generate semantically relevant titles for the learning resources in a learning pathway and a title for the pathway itself. Our approach to Automatic Title Generation for a given text is based on pre-trained Transformer Language Model GPT-2. A pool of candidate titles are generated and an appropriate title is selected among them which is then refined or de-noised to get the final title. The model is trained on research paper abstracts from arXiv and evaluated on three different test sets. We show that it generates semantically and syntactically relevant titles as reflected in ROUGE, BLEU scores and human evaluations. We propose an optional abstractive Summarizer module based on pre-trained Transformer model T5 to shorten medium length documents. This module is also trained and evaluated on research papers from arXiv dataset. Finally, we show that the proposed model of hierarchical title generation for learning pathways has promising results.


Sign in / Sign up

Export Citation Format

Share Document