scholarly journals EMOTION DETECTION AND SEMANTICS OF EMOTIVES: DISTRESS AND ANGER IN ANNOTATED TEXT DATASET

2021 ◽  
Vol 26 (2) ◽  
pp. 78-89
Author(s):  
Anastasia V. Kolmogorova ◽  

The article explores the ways of making emotional lexemes semantic description consistent with interpretative intuition of the ordinary language speaker. The research novelty is determined by the fact that it is based on the data retrieved from the emotional assessment of 3920 internet-texts in Russian made by informants via using a specially designed computer interface. When applied this interface, we can aggregate the weight of 8 emotions (distress, enjoyment, anger, surprise, shame, excitement, disgust, fear) in text. Thus, the data we have used for this publication includes two sets of 150 internet-texts assessed by 2000 informants with the highest score of emotions of distress or anger. The scope of the study covers the semantics of two mentioned above lexemes (grust’ and gnev) analyzed through the prism of collective introspection of informants. The article purpose is to discuss the case when a semantic description of emotives is given by an expert, which largely uses “the best texts” of corresponding emotions, according to the collective opinion of informants. Our methods include psycholinguistic experiment, corpus and semantic analysis. The research led us to three main conclusions. Firstly, the semantic descriptions of emotives grust’ and gnev obtained in proposed way represent prototypical scenarios of living an emotion in social context and take into account not only the introspective sensations of an expert-linguist, but the interpretative strategies of language users. Secondly, such semantic explanation provides us with keys for explaining, why machine learning technologies are better at detecting anger than sadness in text. Finally, it creates a precedent in using new technologies for making an ecological semantic description of emotive vocabulary. The research results can find application in emotiology, lexicographic practice and didactics.

Author(s):  
Francis J Alexander ◽  
James Ang ◽  
Jenna A Bilbrey ◽  
Jan Balewski ◽  
Tiernan Casey ◽  
...  

Rapid growth in data, computational methods, and computing power is driving a remarkable revolution in what variously is termed machine learning (ML), statistical learning, computational learning, and artificial intelligence. In addition to highly visible successes in machine-based natural language translation, playing the game Go, and self-driving cars, these new technologies also have profound implications for computational and experimental science and engineering, as well as for the exascale computing systems that the Department of Energy (DOE) is developing to support those disciplines. Not only do these learning technologies open up exciting opportunities for scientific discovery on exascale systems, they also appear poised to have important implications for the design and use of exascale computers themselves, including high-performance computing (HPC) for ML and ML for HPC. The overarching goal of the ExaLearn co-design project is to provide exascale ML software for use by Exascale Computing Project (ECP) applications, other ECP co-design centers, and DOE experimental facilities and leadership class computing facilities.


Author(s):  
Prarthana Dutta ◽  
Naresh Babu Muppalaneni ◽  
Ripon Patgiri

The world has been evolving with new technologies and advances day-by-day. With the advent of various learning technologies in every field, the research community is able to provide solution in every aspect of life with the applications of Artificial Intelligence, Machine Learning, Deep Learning, Computer Vision, etc. However, with such high achievements, it is found to lag behind the ability to provide explanation against its prediction. The current situation is such that these modern technologies are able to predict and decide upon various cases more accurately and speedily than a human, but failed to provide an answer when the question of why to trust its prediction is put forward. In order to attain a deeper understanding into this rising trend, we explore a very recent and talked-about novel contribution which provides rich insight on a prediction being made -- ``Explainability.'' The main premise of this survey is to provide an overview for researches explored in the domain and obtain an idea of the current scenario along with the advancements published to-date in this field. This survey is intended to provide a comprehensive background of the broad spectrum of Explainability.


2017 ◽  
Vol 3 (10) ◽  
Author(s):  
Dr. A. Joycilin Shermila

Times have changed and teachers have evolved. New technologies have opened up the classroom to the outside world. Teachers who were seen with textbooks and blackboard are now using varied technological tools to empower learners to publish works and engage learners with live audience in real contexts. In this digital era an ever-expanding array of powerful software has been made available. The flipped classroom is a shift from passive to active learning to focus on higher order thinking skills such as analysis, synthesis and evaluation. This model of teaching combines pedagogy and learning technologies. Significant learning happens through facilitating active learning through engaged learners. In this approach learning materials are provided through text, video, audio and multimedia. Students take responsibility of their learning. They work together applying course concepts with guidance from the instructor. This increased interaction helps to create a learning community that encourages them to build knowledge inside and outside the classroom.


2021 ◽  
Vol 186 (Supplement_1) ◽  
pp. 445-451
Author(s):  
Yifei Sun ◽  
Navid Rashedi ◽  
Vikrant Vaze ◽  
Parikshit Shah ◽  
Ryan Halter ◽  
...  

ABSTRACT Introduction Early prediction of the acute hypotensive episode (AHE) in critically ill patients has the potential to improve outcomes. In this study, we apply different machine learning algorithms to the MIMIC III Physionet dataset, containing more than 60,000 real-world intensive care unit records, to test commonly used machine learning technologies and compare their performances. Materials and Methods Five classification methods including K-nearest neighbor, logistic regression, support vector machine, random forest, and a deep learning method called long short-term memory are applied to predict an AHE 30 minutes in advance. An analysis comparing model performance when including versus excluding invasive features was conducted. To further study the pattern of the underlying mean arterial pressure (MAP), we apply a regression method to predict the continuous MAP values using linear regression over the next 60 minutes. Results Support vector machine yields the best performance in terms of recall (84%). Including the invasive features in the classification improves the performance significantly with both recall and precision increasing by more than 20 percentage points. We were able to predict the MAP with a root mean square error (a frequently used measure of the differences between the predicted values and the observed values) of 10 mmHg 60 minutes in the future. After converting continuous MAP predictions into AHE binary predictions, we achieve a 91% recall and 68% precision. In addition to predicting AHE, the MAP predictions provide clinically useful information regarding the timing and severity of the AHE occurrence. Conclusion We were able to predict AHE with precision and recall above 80% 30 minutes in advance with the large real-world dataset. The prediction of regression model can provide a more fine-grained, interpretable signal to practitioners. Model performance is improved by the inclusion of invasive features in predicting AHE, when compared to predicting the AHE based on only the available, restricted set of noninvasive technologies. This demonstrates the importance of exploring more noninvasive technologies for AHE prediction.


Author(s):  
Petar Radanliev ◽  
David De Roure ◽  
Kevin Page ◽  
Max Van Kleek ◽  
Omar Santos ◽  
...  

AbstractMultiple governmental agencies and private organisations have made commitments for the colonisation of Mars. Such colonisation requires complex systems and infrastructure that could be very costly to repair or replace in cases of cyber-attacks. This paper surveys deep learning algorithms, IoT cyber security and risk models, and established mathematical formulas to identify the best approach for developing a dynamic and self-adapting system for predictive cyber risk analytics supported with Artificial Intelligence and Machine Learning and real-time intelligence in edge computing. The paper presents a new mathematical approach for integrating concepts for cognition engine design, edge computing and Artificial Intelligence and Machine Learning to automate anomaly detection. This engine instigates a step change by applying Artificial Intelligence and Machine Learning embedded at the edge of IoT networks, to deliver safe and functional real-time intelligence for predictive cyber risk analytics. This will enhance capacities for risk analytics and assists in the creation of a comprehensive and systematic understanding of the opportunities and threats that arise when edge computing nodes are deployed, and when Artificial Intelligence and Machine Learning technologies are migrated to the periphery of the internet and into local IoT networks.


Energies ◽  
2021 ◽  
Vol 14 (4) ◽  
pp. 1055
Author(s):  
Qian Sun ◽  
William Ampomah ◽  
Junyu You ◽  
Martha Cather ◽  
Robert Balch

Machine-learning technologies have exhibited robust competences in solving many petroleum engineering problems. The accurate predictivity and fast computational speed enable a large volume of time-consuming engineering processes such as history-matching and field development optimization. The Southwest Regional Partnership on Carbon Sequestration (SWP) project desires rigorous history-matching and multi-objective optimization processes, which fits the superiorities of the machine-learning approaches. Although the machine-learning proxy models are trained and validated before imposing to solve practical problems, the error margin would essentially introduce uncertainties to the results. In this paper, a hybrid numerical machine-learning workflow solving various optimization problems is presented. By coupling the expert machine-learning proxies with a global optimizer, the workflow successfully solves the history-matching and CO2 water alternative gas (WAG) design problem with low computational overheads. The history-matching work considers the heterogeneities of multiphase relative characteristics, and the CO2-WAG injection design takes multiple techno-economic objective functions into accounts. This work trained an expert response surface, a support vector machine, and a multi-layer neural network as proxy models to effectively learn the high-dimensional nonlinear data structure. The proposed workflow suggests revisiting the high-fidelity numerical simulator for validation purposes. The experience gained from this work would provide valuable guiding insights to similar CO2 enhanced oil recovery (EOR) projects.


Sign in / Sign up

Export Citation Format

Share Document