scholarly journals A Practical q -Gram Index for Text Retrieval Allowing Errors

2018 ◽  
Vol 1 (2) ◽  
Author(s):  
Gonzalo Navarro ◽  
Ricardo Baeza-Yates

We propose an indexing technique for approximate text searching, which is practical and powerful, and especially optimized for natural language text. Unlike other indices of this kind, it is able to retrieve any string that approximately matches the search pattern, not only words. Every text substring of a fixed length q is stored in the index, together with pointers to all the text positions where it appears. The search pattern is partitioned into pieces which are searched in the index, and all their occurrences in the text are verified for a complete match. To reduce space requirements, pointers to blocks instead of exact positions can be used, which increases querying costs. We design an algorithm to optimize the pattern partition into pieces so that the total number of verifications is minimized. This is especially well suited for natural language texts, and allows to know in advance the expected cost of the search and the expected relevance of the query to the user. We show experimentally the building time, space requirements and querying time of our index, finding that it is a practical alternative for text retrieval. The retrieval times are reduced from 10% to 60% of the best on-line algorithm. 

Author(s):  
P. Monisha ◽  
R. Rubanya ◽  
N. Malarvizhi

The overwhelming majority of existing approaches to opinion feature extraction trust mining patterns for one review corpus, ignoring the nontrivial disparities in word spacing characteristics of opinion options across completely different corpora. During this research a unique technique to spot opinion options from on-line reviews by exploiting the distinction in opinion feature statistics across two corpora, one domain-specific corpus (i.e., the given review corpus) and one domain-independent corpus (i.e., the contrasting corpus). The tendency to capture this inequality called domain relevance (DR), characterizes the relevancy of a term to a text assortment. The tendency to extract an inventory of candidate opinion options from the domain review corpus by shaping a group of grammar dependence rules. for every extracted candidate feature, to have a tendency to estimate its intrinsic-domain relevancy (IDR) and extrinsic-domain relevance(EDR) scores on the domain-dependent and domain-independent corpora, severally. Natural language processing (NLP) refers to computer systems that analyze, attempt understand, or produce one or more human languages, such as English, Japanese, Italian, or Russian. Process information contained in natural language text. The input might be text, spoken language, or keyboard input. The field of NLP is primarily concerned with getting computers to perform useful and interesting tasks with human languages. The field of NLP is secondarily concerned with helping us come to a better understanding of human language


Author(s):  
Matheus C. Pavan ◽  
Vitor G. Santos ◽  
Alex G. J. Lan ◽  
Joao Martins ◽  
Wesley Ramos Santos ◽  
...  

2012 ◽  
Vol 30 (1) ◽  
pp. 1-34 ◽  
Author(s):  
Antonio Fariña ◽  
Nieves R. Brisaboa ◽  
Gonzalo Navarro ◽  
Francisco Claude ◽  
Ángeles S. Places ◽  
...  

Author(s):  
S.G. Antonov

In the article discuss the application aspects of wordforms of natural language text for decision the mistakes correction problem. Discuss the merits and demerits of two known approaches for decision – deterministic and based on probabilities/ Construction principles of natural language corpus described, wich apply in probability approach. Declare conclusion about necessity of complex using these approaches in dependence on properties of texts.


2022 ◽  
Vol 40 (1) ◽  
pp. 1-43
Author(s):  
Ruqing Zhang ◽  
Jiafeng Guo ◽  
Lu Chen ◽  
Yixing Fan ◽  
Xueqi Cheng

Question generation is an important yet challenging problem in Artificial Intelligence (AI), which aims to generate natural and relevant questions from various input formats, e.g., natural language text, structure database, knowledge base, and image. In this article, we focus on question generation from natural language text, which has received tremendous interest in recent years due to the widespread applications such as data augmentation for question answering systems. During the past decades, many different question generation models have been proposed, from traditional rule-based methods to advanced neural network-based methods. Since there have been a large variety of research works proposed, we believe it is the right time to summarize the current status, learn from existing methodologies, and gain some insights for future development. In contrast to existing reviews, in this survey, we try to provide a more comprehensive taxonomy of question generation tasks from three different perspectives, i.e., the types of the input context text, the target answer, and the generated question. We take a deep look into existing models from different dimensions to analyze their underlying ideas, major design principles, and training strategies We compare these models through benchmark tasks to obtain an empirical understanding of the existing techniques. Moreover, we discuss what is missing in the current literature and what are the promising and desired future directions.


Sign in / Sign up

Export Citation Format

Share Document