scholarly journals A computational linguistic approach to natural language processing with applications to garden path sentences analysis

Author(s):  
DU Jia-li ◽  
YU Ping-fang
Author(s):  
Alex Lash ◽  
Kevin Murray ◽  
Gregory Mocko

In the design process, the requirements serve as the benchmark for the entire product. Therefore, the quality of requirement statements is essential to the success of a design. Because of their ergonomic-nature, most requirements are written in natural language (NL). However, writing requirements in natural language presents many issues such as ambiguity, specification issues, and incompleteness. Therefore, identifying issues in requirements involves analyzing these NL statements. This paper presents a linguistic approach to requirement analysis, which utilizes grammatical elements of requirements statements to identify requirement statement issues. These issues are organized by the entity—word, sentence, or document—that they affect. The field of natural language processing (NLP) provides a core set of tools that can aid with this linguistic analysis and provide a method to create a requirement analysis support tool. NLP addresses requirements on processing levels: lexical, syntactic, semantic, and pragmatic. While processing on the lexical and syntactic level are well-defined, mining semantic and pragmatic data is performed in a number of different methods. This paper provides an overview of these current requirement analysis methods in light of the presented linguistic approach. This overview will be used to identify areas for further research and development. Finally, a prototype requirement analysis support tool will be presented. This tool seeks to demonstrate how the semantic processing level can begin to be addressed in requirement analysis. The tool will analyze a sample set of requirements from a family of military tactical vehicles (FMTV) requirements document. It implements NLP tools to semantically compare requirements statements based upon their grammatical subject.


2021 ◽  
Vol 10 (4) ◽  
pp. 2130-2136
Author(s):  
Ryan Adipradana ◽  
Bagas Pradipabista Nayoga ◽  
Ryan Suryadi ◽  
Derwin Suhartono

Misinformation has become an innocuous yet potentially harmful problem ever since the development of internet. Numbers of efforts are done to prevent the consumption of misinformation, including the use of artificial intelligence (AI), mainly natural language processing (NLP). Unfortunately, most of natural language processing use English as its linguistic approach since English is a high resource language. On the contrary, Indonesia language is considered a low resource language thus the amount of effort to diminish consumption of misinformation is low compared to English-based natural language processing. This experiment is intended to compare fastText and GloVe embeddings for four deep neural networks (DNN) models: long short-term memory (LSTM), bidirectional long short-term memory (BI-LSTM), gated recurrent unit (GRU) and bidirectional gated recurrent unit (BI-GRU) in terms of metrics score when classifying news between three classes: fake, valid, and satire. The latter results show that fastText embedding is better than GloVe embedding in supervised text classification, along with BI-GRU + fastText yielding the best result.


2020 ◽  
pp. 3-17
Author(s):  
Peter Nabende

Natural Language Processing for under-resourced languages is now a mainstream research area. However, there are limited studies on Natural Language Processing applications for many indigenous East African languages. As a contribution to covering the current gap of knowledge, this paper focuses on evaluating the application of well-established machine translation methods for one heavily under-resourced indigenous East African language called Lumasaaba. Specifically, we review the most common machine translation methods in the context of Lumasaaba including both rule-based and data-driven methods. Then we apply a state of the art data-driven machine translation method to learn models for automating translation between Lumasaaba and English using a very limited data set of parallel sentences. Automatic evaluation results show that a transformer-based Neural Machine Translation model architecture leads to consistently better BLEU scores than the recurrent neural network-based models. Moreover, the automatically generated translations can be comprehended to a reasonable extent and are usually associated with the source language input.


Diabetes ◽  
2019 ◽  
Vol 68 (Supplement 1) ◽  
pp. 1243-P
Author(s):  
JIANMIN WU ◽  
FRITHA J. MORRISON ◽  
ZHENXIANG ZHAO ◽  
XUANYAO HE ◽  
MARIA SHUBINA ◽  
...  

Author(s):  
Pamela Rogalski ◽  
Eric Mikulin ◽  
Deborah Tihanyi

In 2018, we overheard many CEEA-AGEC members stating that they have "found their people"; this led us to wonder what makes this evolving community unique. Using cultural historical activity theory to view the proceedings of CEEA-ACEG 2004-2018 in comparison with the geographically and intellectually adjacent ASEE, we used both machine-driven (Natural Language Processing, NLP) and human-driven (literature review of the proceedings) methods. Here, we hoped to build on surveys—most recently by Nelson and Brennan (2018)—to understand, beyond what members say about themselves, what makes the CEEA-AGEC community distinct, where it has come from, and where it is going. Engaging in the two methods of data collection quickly diverted our focus from an analysis of the data themselves to the characteristics of the data in terms of cultural historical activity theory. Our preliminary findings point to some unique characteristics of machine- and human-driven results, with the former, as might be expected, focusing on the micro-level (words and language patterns) and the latter on the macro-level (ideas and concepts). NLP generated data within the realms of "community" and "division of labour" while the review of proceedings centred on "subject" and "object"; both found "instruments," although NLP with greater granularity. With this new understanding of the relative strengths of each method, we have a revised framework for addressing our original question.  


Sign in / Sign up

Export Citation Format

Share Document