A Principle-Based System for Syntactic Analysis

Author(s):  
Matthew W. Crocker

Traditional approaches to natural language processing (NLP) can be considered construction-based. That is to say, they employ surface oriented, language specific rules, whether in the form of an Augmented Transition Network (ATN), logic grammar or some other grammar/parsing formalism. The problems of such approaches have always been apparent; they involve large sets of rules, often ad hoc, and their adequacy with respect to the grammar of the language is difficult to ensure.

2018 ◽  
pp. 35-38
Author(s):  
O. Hyryn

The article deals with natural language processing, namely that of an English sentence. The article describes the problems, which might arise during the process and which are connected with graphic, semantic, and syntactic ambiguity. The article provides the description of how the problems had been solved before the automatic syntactic analysis was applied and the way, such analysis methods could be helpful in developing new analysis algorithms. The analysis focuses on the issues, blocking the basis for the natural language processing — parsing — the process of sentence analysis according to their structure, content and meaning, which aims to analyze the grammatical structure of the sentence, the division of sentences into constituent components and defining links between them.


2021 ◽  
Author(s):  
Carolinne Roque e Faria ◽  
Cinthyan Renata Sachs Camerlengo de Barb

Technology is becoming expressively popular among agribusiness producers and is progressing in all agricultural area. One of the difficulties in this context is to handle data in natural language to solve problems in the field of agriculture. In order to build up dialogs and provide rich researchers, the present work uses Natural Language Processing (NLP) techniques to develop an automatic and effective computer system to interact with the user and assist in the identification of pests and diseases in the soybean farming, stored in a database repository to provide accurate diagnoses to simplify the work of the agricultural professional and also for those who deal with a lot of information in this area. Information on 108 pests and 19 diseases that damage Brazilian soybean was collected from Brazilian bibliographic manuals with the purpose to optimize the data and improve production, using the spaCy library for syntactic analysis of NLP, which allowed the pre-process the texts, recognize the named entities, calculate the similarity between the words, verify dependency parsing and also provided the support for the development requirements of the CAROLINA tool (Robotized Agronomic Conversation in Natural Language) using the language belonging to the agricultural area.


2020 ◽  
pp. 41-45
Author(s):  
O. Hyryn

The article proceeds from the intended use of parsing for the purposes of automatic information search, question answering, logical conclusions, authorship verification, text authenticity verification, grammar check, natural language synthesis and other related tasks, such as ungrammatical speech analysis, morphological class definition, anaphora resolution etc. The study covers natural language processing challenges, namely of an English sentence. The article describes formal and linguistic problems, which might arise during the process and which are connected with graphic, semantic, and syntactic ambiguity. The article provides the description of how the problems had been solved before the automatic syntactic analysis was applied and the way, such analysis methods could be helpful in developing new analysis algorithms today. The analysis focuses on the issues, blocking the basis for the natural language processing — parsing — the process of sentence analysis according to their structure, content and meaning, which aims to examine the grammatical structure of the sentence, the division of sentences into constituent components and defining links between them. The analysis identifies a number of linguistic issues that will contribute to the development of an improved model of automatic syntactic analysis: lexical and grammatical synonymy and homonymy, hypo- and hyperonymy, lexical and semantic fields, anaphora resolution, ellipsis, inversion etc. The scope of natural language processing reveals obvious directions for the improvement of parsing models. The improvement will consequently expand the scope and improve the results in areas that already employ automatic parsing. Indispensable achievements in vocabulary and morphology processing shall not be neglected while improving automatic syntactic analysis mechanisms for natural languages.


2021 ◽  
Author(s):  
Minoru Yoshida ◽  
Kenji Kita

Both words and numerals are tokens found in almost all documents but they have different properties. However, relatively little attention has been paid in numerals found in texts and many systems treated the numbers found in the document in ad-hoc ways, such as regarded them as mere strings in the same way as words, normalized them to zeros, or simply ignored them. Recent growth of natural language processing (NLP) research areas has change this situations and more and more attentions have been paid to the numeracy in documents. In this survey, we provide a quick overview of the history and recent advances of the research of mining such relations between numerals and words found in text data.


Author(s):  
JungHo Jeon ◽  
Xin Xu ◽  
Yuxi Zhang ◽  
Liu Yang ◽  
Hubo Cai

Construction inspection is an essential component of the quality assurance programs of state transportation agencies (STAs), and the guidelines for this process reside in lengthy textual specifications. In the current practice, engineers and inspectors must manually go through these documents to plan, conduct, and document their inspections, which is time-consuming, very subjective, inconsistent, and prone to error. A promising alternative to this manual process is the application of natural language processing (NLP) techniques (e.g., text parsing, sentence classification, and syntactic analysis) to automatically extract construction inspection requirements from textual documents and present them as straightforward check questions. This paper introduces an NLP-based method that: 1) extracts individual sentences from the construction specification; 2) preprocesses the resulting sentences; 3) applies Word2Vec and GloVe algorithms to extract vector features; 4) uses a convolutional neural network (CNN) and recurrent neural network to classify sentences; and 5) converts the requirement sentences into check questions via syntactic analysis. The overall methodology was assessed using the Indiana Department of Transportation (DOT) specification as a test case. Our results revealed that the CNN + GloVe combination led to the highest accuracy, at 91.9%, and the lowest loss, at 11.7%. To further validate its use across STAs nationwide, we applied it to the construction specification of the South Carolina DOT as a test case, and our average accuracy was 92.6%.


Computers ◽  
2019 ◽  
Vol 8 (1) ◽  
pp. 22
Author(s):  
Frederik Bäumer ◽  
Joschka Kersting ◽  
Michaela Geierhos

The vision of On-the-Fly (OTF) Computing is to compose and provide software services ad hoc, based on requirement descriptions in natural language. Since non-technical users write their software requirements themselves and in unrestricted natural language, deficits occur such as inaccuracy and incompleteness. These deficits are usually met by natural language processing methods, which have to face special challenges in OTF Computing because maximum automation is the goal. In this paper, we present current automatic approaches for solving inaccuracies and incompletenesses in natural language requirement descriptions and elaborate open challenges. In particular, we will discuss the necessity of domain-specific resources and show why, despite far-reaching automation, an intelligent and guided integration of end users into the compensation process is required. In this context, we present our idea of a chat bot that integrates users into the compensation process depending on the given circumstances.


2011 ◽  
Vol 6 ◽  
Author(s):  
Eva Hajičová

We want to demonstrate on some selected linguistic issues that classical structural and functional linguistics even with its seemingly traditional approaches has something to offer to a formal description of language and its applications in natural language processing and to illustrate by a brief reference to Functional Generative Grammar (on the theoretical side of CL) and Prague Dependency Treebank (on the applicational side) a possible interaction between linguistics and CL.


2019 ◽  
Vol 87 (2) ◽  
pp. 21-23
Author(s):  
Rohin Attrey ◽  
Alexander Levit

The healthcare industry generates data at a rapid rate, with no signs of slowing down. A large portion of this information takes the form of unstructured narrative text, making it difficult for computer systems to analyze the data in a usable format. However, automated analysis of this information could be incredibly useful in daily practice. This could be accomplished with natural language processing, an area of artificial intelligence and computational linguistics that is used to analyze and process large sets of unstructured data, namely spoken or written communication. Natural language processing has already been implemented in many sectors, and the industry is projected to be worth US$16 billion by 2021. Natural language processing could take unstructured patient data and interpret meaning from the text, allowing that information to inform healthcare delivery. Natural language processing can also enable intelligent chatbots, interacting and providing medical support to patients. It has the potential to aid physicians by efficiently summarizing patient charts and predicting patient outcomes. In hospitals, it has the ability to analyze patient satisfaction and facilitate quality improvement. Despite current technical limitations, natural language processing is a rapidly developing technology that promises to improve the quality and efficiency of healthcare delivery.


Author(s):  
Dastan Hussen Maulud ◽  
Siddeeq Y. Ameen ◽  
Naaman Omar ◽  
Shakir Fattah Kak ◽  
Zryan Najat Rashid ◽  
...  

With the exponential growth of the information on the Internet, there is a high demand for making this information readable and processable by machines. For this purpose, there is a need for the Natural Language Processing (NLP) pipeline. Natural language analysis is a tool used by computers to grasp, perceive, and control human language. This paper discusses various techniques addressed by different researchers on NLP and compares their performance. The comparison among the reviewed researches illustrated that good accuracy levels haved been achieved. Adding to that, the researches that depended on the Sentiment Analysis and ontology methods achieved small prediction error. The syntactic analysis or parsing or syntax analysis is the third stage of the NLP as a conclusion to use NLP technology. This step aims to accurately mean or, from the text, you may state a dictionary meaning. Syntax analysis analyzes the meaning of the text in comparison with the formal grammatical rules.


Sign in / Sign up

Export Citation Format

Share Document