scholarly journals Basic challenges in natural language processing systems

2020 ◽  
pp. 41-45
Author(s):  
O. Hyryn

The article proceeds from the intended use of parsing for the purposes of automatic information search, question answering, logical conclusions, authorship verification, text authenticity verification, grammar check, natural language synthesis and other related tasks, such as ungrammatical speech analysis, morphological class definition, anaphora resolution etc. The study covers natural language processing challenges, namely of an English sentence. The article describes formal and linguistic problems, which might arise during the process and which are connected with graphic, semantic, and syntactic ambiguity. The article provides the description of how the problems had been solved before the automatic syntactic analysis was applied and the way, such analysis methods could be helpful in developing new analysis algorithms today. The analysis focuses on the issues, blocking the basis for the natural language processing — parsing — the process of sentence analysis according to their structure, content and meaning, which aims to examine the grammatical structure of the sentence, the division of sentences into constituent components and defining links between them. The analysis identifies a number of linguistic issues that will contribute to the development of an improved model of automatic syntactic analysis: lexical and grammatical synonymy and homonymy, hypo- and hyperonymy, lexical and semantic fields, anaphora resolution, ellipsis, inversion etc. The scope of natural language processing reveals obvious directions for the improvement of parsing models. The improvement will consequently expand the scope and improve the results in areas that already employ automatic parsing. Indispensable achievements in vocabulary and morphology processing shall not be neglected while improving automatic syntactic analysis mechanisms for natural languages.

2018 ◽  
pp. 35-38
Author(s):  
O. Hyryn

The article deals with natural language processing, namely that of an English sentence. The article describes the problems, which might arise during the process and which are connected with graphic, semantic, and syntactic ambiguity. The article provides the description of how the problems had been solved before the automatic syntactic analysis was applied and the way, such analysis methods could be helpful in developing new analysis algorithms. The analysis focuses on the issues, blocking the basis for the natural language processing — parsing — the process of sentence analysis according to their structure, content and meaning, which aims to analyze the grammatical structure of the sentence, the division of sentences into constituent components and defining links between them.


Poetics ◽  
1990 ◽  
Vol 19 (1-2) ◽  
pp. 99-120
Author(s):  
Stefan Wermter ◽  
Wendy G. Lehnert

2021 ◽  
Author(s):  
Carolinne Roque e Faria ◽  
Cinthyan Renata Sachs Camerlengo de Barb

Technology is becoming expressively popular among agribusiness producers and is progressing in all agricultural area. One of the difficulties in this context is to handle data in natural language to solve problems in the field of agriculture. In order to build up dialogs and provide rich researchers, the present work uses Natural Language Processing (NLP) techniques to develop an automatic and effective computer system to interact with the user and assist in the identification of pests and diseases in the soybean farming, stored in a database repository to provide accurate diagnoses to simplify the work of the agricultural professional and also for those who deal with a lot of information in this area. Information on 108 pests and 19 diseases that damage Brazilian soybean was collected from Brazilian bibliographic manuals with the purpose to optimize the data and improve production, using the spaCy library for syntactic analysis of NLP, which allowed the pre-process the texts, recognize the named entities, calculate the similarity between the words, verify dependency parsing and also provided the support for the development requirements of the CAROLINA tool (Robotized Agronomic Conversation in Natural Language) using the language belonging to the agricultural area.


2020 ◽  
Vol 34 (05) ◽  
pp. 8504-8511
Author(s):  
Arindam Mitra ◽  
Ishan Shrivastava ◽  
Chitta Baral

Natural Language Inference (NLI) plays an important role in many natural language processing tasks such as question answering. However, existing NLI modules that are trained on existing NLI datasets have several drawbacks. For example, they do not capture the notion of entity and role well and often end up making mistakes such as “Peter signed a deal” can be inferred from “John signed a deal”. As part of this work, we have developed two datasets that help mitigate such issues and make the systems better at understanding the notion of “entities” and “roles”. After training the existing models on the new dataset we observe that the existing models do not perform well on one of the new benchmark. We then propose a modification to the “word-to-word” attention function which has been uniformly reused across several popular NLI architectures. The resulting models perform as well as their unmodified counterparts on the existing benchmarks and perform significantly well on the new benchmarks that emphasize “roles” and “entities”.


2021 ◽  
Vol 47 (05) ◽  
Author(s):  
NGUYỄN CHÍ HIẾU

Knowledge Graphs are applied in many fields such as search engines, semantic analysis, and question answering in recent years. However, there are many obstacles for building knowledge graphs as methodologies, data and tools. This paper introduces a novel methodology to build knowledge graph from heterogeneous documents.  We use the methodologies of Natural Language Processing and deep learning to build this graph. The knowledge graph can use in Question answering systems and Information retrieval especially in Computing domain


Author(s):  
Saravanakumar Kandasamy ◽  
Aswani Kumar Cherukuri

Semantic similarity quantification between concepts is one of the inevitable parts in domains like Natural Language Processing, Information Retrieval, Question Answering, etc. to understand the text and their relationships better. Last few decades, many measures have been proposed by incorporating various corpus-based and knowledge-based resources. WordNet and Wikipedia are two of the Knowledge-based resources. The contribution of WordNet in the above said domain is enormous due to its richness in defining a word and all of its relationship with others. In this paper, we proposed an approach to quantify the similarity between concepts that exploits the synsets and the gloss definitions of different concepts using WordNet. Our method considers the gloss definitions, contextual words that are helping in defining a word, synsets of contextual word and the confidence of occurrence of a word in other word’s definition for calculating the similarity. The evaluation based on different gold standard benchmark datasets shows the efficiency of our system in comparison with other existing taxonomical and definitional measures.


2019 ◽  
Vol 2 (1) ◽  
pp. 53-64
Author(s):  
Herwin H Herwin

STMIK Amik Riau memiliki portal pada website http://www.sar.ac.id difungsikan sebagai media penyebaran informasi bagi sivitas akademika dan stakeholder. Rerata pengunjung setiap hari dalam 3 bulan terakhir adalah 150 kunjungan, namun terjadi peningkatan pada saat penerimaan mahasiswa di setiap tahun akademik. Hal ini mengindikasikan terjadinya peningkatan minat masyarakat untuk mengetahui informasi STMIK Amik Riau. Sayangnya, sampai saat ini pemanfaatan portal web site masih satu arah, dari STMIK Amik Riau ke stakeholder dan masyarakat, tidak terjadi sebaliknya. Komunikasi stakeholder dengan PT sehubungan dengan muatan yang ada di dalam portal menggunakan media sosial dan tidak terintegrasi dengan web.  Begitu juga dengan masukan, koreksi, tanggapan, maupun komunikasi lain menggunakan media sosial.  Sampai saat ini, masyarakat yang mengunjungi portal website baik masyarakat luas, maupun stakeholder tidak dapat dideteksi waktu berkunjung sehingga tidak dapat disapa dengan filosofi “3S”, padahal masyarakat luas yang telah berkunjung merupakan pasar potensial untuk di edukasi. Masyarakat yang berkunjung ke portal website, dengan sopan di sapa oleh sistem, kemudian dilanjutkan dengan komunikasi langsung, tersedia mesin yang siap memberikan salam  dan melayani setiap pertanyaan yang diajukan oleh pengunjung. Penelitian ini bertujuan membuat chatbot yang mampu berkomunikasi dengan pengunjung website.  Chatbot  yang telah dibuat diberi nama STMIK Amik Riau Intelligence Virtual Information disingkat SILVI.  Chatbot dibuat berdasarkan Question Answering Systems (QAS), bekerja dengan algoritma kemiripan antara dua teks. Penelitian ini menghasilkan aplikasi yang siap digunakan, diberi nama SILVI, mampu berkomunikasi dengan pengunjung website. Chatbot mengoptimalkan komunikasi seolah tidak menyadari, tetap menganggap lawan bicara adalah pegawai yang tepat dalam tugas pokok dan fungsi.  


Sign in / Sign up

Export Citation Format

Share Document