A New Deep Neural Network Based Learning to Rank Method for Information Retrieval

Author(s):  
Mingsheng Fu ◽  
Hong Qu ◽  
Fan Li ◽  
Yanjun Liu
Author(s):  
Dr. V. Suma

The recent technology development fascinates the people towards information and its services. Managing the personal and pubic data is a perennial research topic among researchers. In particular retrieval of information gains more attention as it is important similar to data storing. Clustering based, similarity based, graph based information retrieval systems are evolved to reduce the issues in conventional information retrieval systems. Learning based information retrieval is the present trend and in particular deep neural network is widely adopted due to its retrieval performance. However, the similarity between the information has uncertainties due to its measuring procedures. Considering these issues also to improve the retrieval performance, a hybrid deep fuzzy hashing algorithm is introduced in this research work. Hashing efficiently retrieves the information based on mapping the similar information as correlated binary codes and this underlying information is trained using deep neural network and fuzzy logic to retrieve the necessary information from distributed cloud. Experimental results prove that the proposed model attains better retrieval accuracy and accuracy compared to conventional models such as support vector machine and deep neural network.


2021 ◽  
Vol 11 (5) ◽  
pp. 7598-7604
Author(s):  
H. V. T. Chi ◽  
D. L. Anh ◽  
N. L. Thanh ◽  
D. Dinh

Paraphrase identification is a crucial task in natural language understanding, especially in cross-language information retrieval. Nowadays, Multi-Task Deep Neural Network (MT-DNN) has become a state-of-the-art method that brings outstanding results in paraphrase identification [1]. In this paper, our proposed method based on MT-DNN [2] to detect similarities between English and Vietnamese sentences, is proposed. We changed the shared layers of the original MT-DNN from original the BERT [3] to other pre-trained multi-language models such as M-BERT [3] or XLM-R [4] so that our model could work on cross-language (in our case, English and Vietnamese) information retrieval. We also added some tasks as improvements to gain better results. As a result, we gained 2.3% and 2.5% increase in evaluated accuracy and F1. The proposed method was also implemented on other language pairs such as English – German and English – French. With those implementations, we got a 1.0%/0.7% improvement for English – German and a 0.7%/0.5% increase for English – French.


2019 ◽  
Vol 8 (1) ◽  
pp. 32-35
Author(s):  
A. Uma Maheswari ◽  
N. Revathy

Semantic drift is a common problem in iterative information extraction. Unsupervised bagging and incorporated distributional similarity is used to reduce the difficulty of semantic drift in iterative bootstrapping algorithms, particularly when extracting large semantic lexicons. In this research work, a method to minimize semantic drift by identifying the (Drifting Points) DPs and removing the effect introduced by the DPs is proposed. Previous methods for identifying drifting errors can be roughly divided into two categories: (1) multi-class based, and (2) single-class based, according to the settings of Information Extraction systems that adopt them. Compared to previous approaches which usually incur substantial loss in recall, DP-based cleaning method can effectively clean a large proportion of semantic drift errors while keeping a high recall.


Author(s):  
David T. Wang ◽  
Brady Williamson ◽  
Thomas Eluvathingal ◽  
Bruce Mahoney ◽  
Jennifer Scheler

Sign in / Sign up

Export Citation Format

Share Document