scholarly journals Pretrained Transformers for Text Ranking: BERT and Beyond

Author(s):  
Andrew Yates ◽  
Rodrigo Nogueira ◽  
Jimmy Lin
Keyword(s):  
Author(s):  
Tsutomu MATSUNAGA ◽  
Shogo SHINKAI ◽  
Takashi SUENAGA

2020 ◽  
Vol 8 ◽  
pp. 759-775
Author(s):  
Edwin Simpson ◽  
Yang Gao ◽  
Iryna Gurevych

For many NLP applications, such as question answering and summarization, the goal is to select the best solution from a large space of candidates to meet a particular user’s needs. To address the lack of user or task-specific training data, we propose an interactive text ranking approach that actively selects pairs of candidates, from which the user selects the best. Unlike previous strategies, which attempt to learn a ranking across the whole candidate space, our method uses Bayesian optimization to focus the user’s labeling effort on high quality candidates and integrate prior knowledge to cope better with small data scenarios. We apply our method to community question answering (cQA) and extractive multidocument summarization, finding that it significantly outperforms existing interactive approaches. We also show that the ranking function learned by our method is an effective reward function for reinforcement learning, which improves the state of the art for interactive summarization.


Author(s):  
Abdelghani Bellaachia ◽  
Mohammed Al-Dhelaan

Random walks on graphs have been extensively used for a variety of graph-based problems such as ranking vertices, predicting links, recommendations, and clustering. However, many complex problems mandate a high-order graph representation to accurately capture the relationship structure inherent in them. Hypergraphs are particularly useful for such models due to the density of information stored in their structure. In this paper, we propose a novel extension to defining random walks on hypergraphs. Our proposed approach combines the weights of destination vertices and hyperedges in a probabilistic manner to accurately capture transition probabilities. We study and analyze our generalized form of random walks suitable for the structure of hypergraphs. We show the effectiveness of our model by conducting a text ranking experiment on a real world data set with a 9% to 33% improvement in precision and a range of 7% to 50% improvement in Bpref over other random walk approaches.


Quantity of data produced per day is around 2.5 quintillion bytes. Right now, no one has the time to pursue each and everything. With the growth of technology and digital media, people are becoming very lazy; they are looking for everything more smartly. If they want to read any article or newspaper, they cannot go through every line that has been given. To overcome this problem, an automatic text summarizer using Machine Learning (ML) and Natural Language Processing (NLP) with the python programming language has been introduced. This automatic text summarizer will generate a concise and meaningful summary of the text from resources like textbooks, articles, messages by using a text ranking algorithm. The input text that is given will be split into sentences; these sentences are again converted into vectors. These vectors are represented as a similarity matrix and based on these similarities; matrices sentence rankings will be given. The higher ranked sentences will be the final summary of the given input text.


2021 ◽  
Vol 28 (3) ◽  
pp. 292-311
Author(s):  
Vitaly I. Yuferev ◽  
Nikolai A. Razin

It is known that in the tasks of natural language processing, the representation of texts by vectors of fixed length using word-embedding models makes sense in cases where the vectorized texts are short.The longer the texts being compared, the worse the approach works. This situation is due to the fact that when using word-embedding models, information is lost when converting the vector representations of the words that make up the text into a vector representation of the entire text, which usually has the same dimension as the vector of a single word.This paper proposes an alternative way for using pre-trained word-embedding models for text vectorization. The essence of the proposed method consists in combining semantically similar elements of the dictionary of the existing text corpus by clustering their (dictionary elements) embeddings, as a result of which a new dictionary is formed with a size smaller than the original one, each element of which corresponds to one cluster. The original corpus of texts is reformulated in terms of this new dictionary, after which vectorization is performed on the reformulated texts using one of the dictionary approaches (TF-IDF was used in the work). The resulting vector representation of the text can be additionally enriched using the vectors of words of the original dictionary obtained by decreasing the dimension of their embeddings for each cluster.A series of experiments to determine the optimal parameters of the method is described in the paper, the proposed approach is compared with other methods of text vectorization for the text ranking problem – averaging word embeddings with TF-IDF weighting and without weighting, as well as vectorization based on TF-IDF coefficients.


2020 ◽  
Vol 9 (05) ◽  
pp. 25039-25046 ◽  
Author(s):  
Rahul C Kore ◽  
Prachi Ray ◽  
Priyanka Lade ◽  
Amit Nerurkar

Reading legal documents are tedious and sometimes it requires domain knowledge related to that document. It is hard to read the full legal document without missing the key important sentences. With increasing number of legal documents it would be convenient to get the essential information from the document without having to go through the whole document. The purpose of this study is to understand a large legal document within a short duration of time. Summarization gives flexibility and convenience to the reader. Using vector representation of words, text ranking algorithms, similarity techniques, this study gives a way to produce the highest ranked sentences. Summarization produces the result in such a way that it covers the most vital information of the document in a concise manner. The paper proposes how the different natural language processing concepts can be used to produce the desired result and give readers the relief from going through the whole complex document. This study definitively presents the steps that are required to achieve the aim and elaborates all the algorithms used at each and every step in the process.


Sign in / Sign up

Export Citation Format

Share Document