Natural Language Processing and Futures Studies

2019 ◽  
Vol 12 (2) ◽  
pp. 181-197 ◽  
Author(s):  
Walter Kehl ◽  
Mike Jackson ◽  
Alessandro Fergnani

Because the input for Futures Studies is to a very high degree formulated as written words and texts, methods which automate the processing of texts can substantially help Futures Studies. At Shaping Tomorrow, we have developed a software system using Natural Language Processing (NLP), a subfield of Artificial Intelligence, which automatically analyzes publicly available texts and extracts future-relevant data from theses texts. This process can be used to study the futures. This article discusses this software system, explains how it works with a detailed example, and shows real-life applications and visualizations of the resulting data. The current state of this method is just the first step; a number of technological improvements and their possible benefits are explained. The implications of using this software system for the field of Futures Studies are mostly positive, but there are also a number of caveats.

10.29007/pc58 ◽  
2018 ◽  
Author(s):  
Julia Lavid ◽  
Marta Carretero ◽  
Juan Rafael Zamorano

In this paper we set forth an annotation model for dynamic modality in English and Spanish, given its relevance not only for contrastive linguistic purposes, but also for its impact on practical annotation tasks in the Natural Language Processing (NLP) community. An annotation scheme is proposed, which captures both the functional-semantic meanings and the language-specific realisations of dynamic meanings in both languages. The scheme is validated through a reliability study performed on a randomly selected set of one hundred and twenty sentences from the MULTINOT corpus, resulting in a high degree of inter-annotator agreement. We discuss our main findings and give attention to the difficult cases as they are currently being used to develop detailed guidelines for the large-scale annotation of dynamic modality in English and Spanish.


2021 ◽  
Author(s):  
Raphael Souza de Oliveira ◽  
Erick Giovani Sperandio Nascimento

The Brazilian legal system postulates the expeditious resolution of judicial proceedings. However, legal courts are working under budgetary constraints and with reduced staff. As a way to face these restrictions, artificial intelligence (AI) has been tackling many complex problems in natural language processing (NLP). This work aims to detect the degree of similarity between judicial documents that can be achieved in the inference group using unsupervised learning, by applying three NLP techniques, namely term frequency-inverse document frequency (TF-IDF), Word2Vec CBoW, and Word2Vec Skip-gram, the last two being specialized with a Brazilian language corpus. We developed a template for grouping lawsuits, which is calculated based on the cosine distance between the elements of the group to its centroid. The Ordinary Appeal was chosen as a reference file since it triggers legal proceedings to follow to the higher court and because of the existence of a relevant contingent of lawsuits awaiting judgment. After the data-processing steps, documents had their content transformed into a vector representation, using the three NLP techniques. We notice that specialized word-embedding models—like Word2Vec—present better performance, making it possible to advance in the current state of the art in the area of NLP applied to the legal sector.


1990 ◽  
Vol 5 (4) ◽  
pp. 225-249 ◽  
Author(s):  
Ann Copestake ◽  
Karen Sparck Jones

AbstractThis paper reviews the current state of the art in natural language access to databases. This has been a long-standing area of work in natural language processing. But though some commercial systems are now available, providing front ends has proved much harder than was expected, and the necessary limitations on front ends have to be recognized. The paper discusses the issues, both general to language and task-specific, involved in front end design, and the way these have been addressed, concentrating on the work of the last decade. The focus is on the central process of translating a natural language question into a database query, but other supporting functions are also covered. The points are illustrated by the use of a single example application. The paper concludes with an evaluation of the current state, indicating that future progress will depend on the one hand on general advances in natural language processing, and on the other on expanding the capabilities of traditional databases.


Author(s):  
Oana Frunza ◽  
Diana Inkpen

This book chapter presents several natural language processing (NLP) and machine learning (ML) techniques that can help achieve a better medical practice by means of extracting relevant medical information from the wealth of textual data. The chapter describes three major tasks: building intelligent tools that can help in the clinical decision making, tools that can automatically identify relevant medical information from the life-science literature, and tools that can extract semantic relations between medical concepts. Besides introducing and describing these tasks, methodological settings accompanied by representative results obtained on real-life data sets are presented.


2020 ◽  
Vol 4 (Supplement_1) ◽  
Author(s):  
Yasmeen Almog ◽  
Angshu Rai ◽  
Anirban Mishra ◽  
Amanda Moulaison ◽  
Ross Powell ◽  
...  

Abstract Fragility fractures due to osteoporosis are common and are associated with significant clinical, personal, and economic burden. Even after a fragility fracture, osteoporosis remains widely underdiagnosed and undertreated. Common fracture risk assessment tools, such as FRAX1 and Garvan,2 confer risk over the long term but do not provide short-term risk estimates necessary to identify very high-risk patients likely to fracture in the next 1–2 years. Furthermore, these tools utilize cross-sectional data representing a subset of all available clinical risk factors for risk prediction. Thus, these methods are generalized across patient populations and may not fully utilize patient histories commonly found in electronic health records (EHRs) that contain temporal information for thousands of unique features. The Optum® de-identified EHR dataset (2007–2018) provides an opportunity to use historical medical data to generate short-term, personalized fracture risk predictions for individual patients. We used the Optum® dataset to develop Crystal Bone, a method that applies machine learning techniques commonly used in natural language processing to the temporal nature of patient histories in order to predict fracture risk over a 1- to 2-year timeframe. Specifically, we repurposed deep-learning models typically applied to language-based prediction tasks in which the goal is to learn the meanings of words and sentences to classify them. Crystal Bone uses context-based embedding techniques to learn an equivalent “semantic” meaning of various medical events. Similar to how language models predict the next word in a given sentence or the topic of an overall document, Crystal Bone can predict that a patient’s future trajectory may contain a fracture or that the “signature” of the patient’s overall journey is similar to that of a typical fracture patient. We applied Crystal Bone to two datasets, one enriched for fracture patients and one representative of a typical hospital system. In both datasets, when predicting likelihood of fracture in the next 1–2 years, Crystal Bone had an area under the receiver operating characteristic (AUROC) score ranging from 72% to 83% on a test (hold-out) dataset. These results suggest performance similar to that of FRAX and Garvan, which have 10-year fracture risk prediction AUROC scores of 64.4% +/- 3.7%.3 Our results suggest that it is possible to use each patient’s unique medical history as it changes over time to predict patients at risk for fracture in 1–2 years. Furthermore, it is theoretically possible to integrate a model like Crystal Bone directly into an EHR system, enabling “hands-off” fracture risk prediction, which could lead to improved identification of patients at very high risk for fracture. 1Kanis JA, Osteoporos Int. 2012;23:2239–56. 2Rubin KH, J Bone Miner Res. 2013;28:1701–17. 3Leslie WD, Osteoporos Int. 2014;25:1–21.


Author(s):  
Michael Caballero

Question Answering (QA) is a subfield of Natural Language Processing (NLP) and computer science focused on building systems that automatically answer questions from humans in natural language. This survey summarizes the history and current state of the field and is intended as an introductory overview of QA systems. After discussing QA history, this paper summarizes the different approaches to the architecture of QA systems -- whether they are closed or open-domain and whether they are text-based, knowledge-based, or hybrid systems. Lastly, some common datasets in this field are introduced and different evaluation metrics are discussed.


2021 ◽  
Vol 1 (3) ◽  
Author(s):  
García Navarro C

The purpose of this research is to learn how engagement has been measured so far and what new techniques will be used to measure it in the future. To this end, firstly, there is a review of all the current research in engagement has been conducted, in addition to a review of the current traditional techniques used to measure it. Secondly, the concept of Artificial Intelligence has been analyzed and how one of its most common techniques (Natural Language Processing) is starting to be used as a new technique to measure engagement. Once the traditional and new techniques had been presented, a theoretical differentiation was made between them in order to test the benefits of the latter. The main conclusions were that Artificial Intelligence is increasing its fields of action, specifically in the psychology of organizations. In this field, the new techniques allow companies to save time in the administration and the conduction of surveys. Moreover, the data reported by AI is less biased than the one that comes from surveys, since the data is collected directly and these techniques do not bias the employee when answering the items. As a final conclusion, it is proposed that a study be carried out to compare the results of both techniques in real-life companies.


Sign in / Sign up

Export Citation Format

Share Document