scholarly journals Identifikasi Chatbot dalam Meningkatkan Pelayanan Online Menggunakan Metode Natural Language Processing

2021 ◽  
pp. 142-147
Author(s):  
M Muliyono ◽  
S Sumijan

Chatbot is a software with artificial intelligence that can imitate human conversations through text messages or voice messages. This chatbot can convey information, according to the knowledge that has been given previously. Helping the limitations of the academic section in answering questions posed by students. The method in this study was sourced from a questionnaire distributed to students at the Muhammadiyah University of West Sumatra. Based on the analysis of the questionnaire, there are 40 questions that are often asked by students to the academic section. Then it is processed using Natural Language Processing (NLP). Natural Language Processing is a branch of science from artificial intelligence that is able to study communication between humans and computers through natural language. The processing stage is to identify the intent, process the input and display the results according to the input. The results of the test using a questionnaire addressed to 227 students got a score of 3,55 with a very good predicate. Then do the test using 40 question and answer data. So, obtained 37 appropriate answers and 3 answers that are not in accordance with the percentage of answer accuracy generated from the chatbot is 92.5 percent. The results of this test have been able to respond to the questions asked by students. This chatbot can make it easier for students to get information with a very good level of accuracy

Author(s):  
Jesús Fernández-Avelino ◽  
Giner Alor-Hernández ◽  
Mario Andrés Paredes-Valverde ◽  
Laura Nely Sánchez-Morales

A chatbot is a software agent that mimics human conversation using artificial intelligence technologies. Chatbots help to accomplish tasks ranging from answering questions, playing music, to managing smart home devices. The adoption of this kind of agent is increasing since people are discovering the benefits of them, such as saving time and money, higher customer satisfaction, customer base growing, among others. However, developing a chatbot is a challenging task that requires addressing several issues such as pattern matching, natural language understanding, and natural language processing, as well as to design a knowledge base that encapsulates the intelligence of the system. This chapter describes the design and implementation of a text/speech chatbot for supporting health self-management. This chatbot is currently based on Spanish. The main goal of this chapter is to clearly describe the main components and phases of the chatbot development process, the methods, and tools used for this purpose, as well as to describe and discuss our findings from the practice side of things.


2021 ◽  
pp. 1-13
Author(s):  
Lamiae Benhayoun ◽  
Daniel Lang

BACKGROUND: The renewed advent of Artificial Intelligence (AI) is inducing profound changes in the classic categories of technology professions and is creating the need for new specific skills. OBJECTIVE: Identify the gaps in terms of skills between academic training on AI in French engineering and Business Schools, and the requirements of the labour market. METHOD: Extraction of AI training contents from the schools’ websites and scraping of a job advertisements’ website. Then, analysis based on a text mining approach with a Python code for Natural Language Processing. RESULTS: Categorization of occupations related to AI. Characterization of three classes of skills for the AI market: Technical, Soft and Interdisciplinary. Skills’ gaps concern some professional certifications and the mastery of specific tools, research abilities, and awareness of ethical and regulatory dimensions of AI. CONCLUSIONS: A deep analysis using algorithms for Natural Language Processing. Results that provide a better understanding of the AI capability components at the individual and the organizational levels. A study that can help shape educational programs to respond to the AI market requirements.


2019 ◽  
Vol 8 (4) ◽  
pp. 10289-10293

Sentiment Analysis is a tool used for determining the Polarity or Emotion of a Sentence. It is a field of Natural Language Processing which focuses on the study of opinions. In this study, the researchers solved one key challenge in Sentiment Analysis, which is to consider the Ending Punctuation Marks present in a sentence. Ending punctuation marks plays a significant role in Emotion Recognition and Intensity Level Recognition. The research made used of tweets expressing opinions about Philippine President Rodrigo Duterte. These downloaded tweets served as the inputs. It was initially subjected to pre-processing stage to be able to prepare the sentences for processing. A Language Model was created to serve as the classifier for determining the scores of the tweets. The scores give the polarity of the sentence. Accuracy is very important in sentiment analysis. To increase the chance of correctly identifying the polarity of the tweets, the input undergone Intensity Level Recognition which determines the intensifiers and negations within the sentences. The system was evaluated with overall performance of 80.27%.


Author(s):  
Evrenii Polyakov ◽  
Leonid Voskov ◽  
Pavel Abramov ◽  
Sergey Polyakov

Introduction: Sentiment analysis is a complex problem whose solution essentially depends on the context, field of study andamount of text data. Analysis of publications shows that the authors often do not use the full range of possible data transformationsand their combinations. Only a part of the transformations is used, limiting the ways to develop high-quality classification models.Purpose: Developing and exploring a generalized approach to building a model, which consists in sequentially passing throughthe stages of exploratory data analysis, obtaining a basic solution, vectorization, preprocessing, hyperparameter optimization, andmodeling. Results: Comparative experiments conducted using a generalized approach for classical machine learning and deeplearning algorithms in order to solve the problem of sentiment analysis of short text messages in natural language processinghave demonstrated that the classification quality grows from one stage to another. For classical algorithms, such an increasein quality was insignificant, but for deep learning, it was 8% on average at each stage. Additional studies have shown that theuse of automatic machine learning which uses classical classification algorithms is comparable in quality to manual modeldevelopment; however, it takes much longer. The use of transfer learning has a small but positive effect on the classificationquality. Practical relevance: The proposed sequential approach can significantly improve the quality of models under developmentin natural language processing problems.


Author(s):  
Seonho Kim ◽  
Jungjoon Kim ◽  
Hong-Woo Chun

Interest in research involving health-medical information analysis based on artificial intelligence, especially for deep learning techniques, has recently been increasing. Most of the research in this field has been focused on searching for new knowledge for predicting and diagnosing disease by revealing the relation between disease and various information features of data. These features are extracted by analyzing various clinical pathology data, such as EHR (electronic health records), and academic literature using the techniques of data analysis, natural language processing, etc. However, still needed are more research and interest in applying the latest advanced artificial intelligence-based data analysis technique to bio-signal data, which are continuous physiological records, such as EEG (electroencephalography) and ECG (electrocardiogram). Unlike the other types of data, applying deep learning to bio-signal data, which is in the form of time series of real numbers, has many issues that need to be resolved in preprocessing, learning, and analysis. Such issues include leaving feature selection, learning parts that are black boxes, difficulties in recognizing and identifying effective features, high computational complexities, etc. In this paper, to solve these issues, we provide an encoding-based Wave2vec time series classifier model, which combines signal-processing and deep learning-based natural language processing techniques. To demonstrate its advantages, we provide the results of three experiments conducted with EEG data of the University of California Irvine, which are a real-world benchmark bio-signal dataset. After converting the bio-signals (in the form of waves), which are a real number time series, into a sequence of symbols or a sequence of wavelet patterns that are converted into symbols, through encoding, the proposed model vectorizes the symbols by learning the sequence using deep learning-based natural language processing. The models of each class can be constructed through learning from the vectorized wavelet patterns and training data. The implemented models can be used for prediction and diagnosis of diseases by classifying the new data. The proposed method enhanced data readability and intuition of feature selection and learning processes by converting the time series of real number data into sequences of symbols. In addition, it facilitates intuitive and easy recognition, and identification of influential patterns. Furthermore, real-time large-capacity data analysis is facilitated, which is essential in the development of real-time analysis diagnosis systems, by drastically reducing the complexity of calculation without deterioration of analysis performance by data simplification through the encoding process.


Author(s):  
Katie Miller

The challenge presented is an age when some decisions are made by humans, some are made by AI, and some are made by a combination of AI and humans. For the person refused housing, a phone service, or employment, the experience is the same, but the ability to understand what has happened and obtain a remedy may be very different if the discrimination is attributable to or contributed by an AI system. If we are to preserve the policy intentions of our discrimination, equal opportunity, and human rights laws, we need to understand how discrimination arises in AI systems; how design in AI systems can mitigate such discrimination; and whether our existing laws are adequate to address discrimination in AI. This chapter endeavours to provide this understanding. In doing so, it focuses on narrow but advanced forms of artificial intelligence, such as natural language processing, facial recognition, and cognitive neural networks.


Sign in / Sign up

Export Citation Format

Share Document