scholarly journals How to combine text-mining methods to validate induced Verb-Object relations?

2014 ◽  
Vol 11 (1) ◽  
pp. 133-155 ◽  
Author(s):  
Nicolas Béchet ◽  
Jacques Chauché ◽  
Violaine Prince ◽  
Mathieu Roche

This paper describes methods using Natural Language Processing approaches to extract and validate induced syntactic relations (here restricted to the Verb-Object relation). These methods use a syntactic parser and a semantic closeness measure to extract such relations. Then, their validation is based on two different techniques: A Web Validation system on one part, then a Semantic-Vectorbased approach, and finally different combinations of both techniques in order to rank induced Verb-Object relations. The Semantic Vector approach is a Roget-based method which computes a syntactic relation as a vector. Web Validation uses a search engine to determine the relevance of a syntactic relation according to its popularity. An experimental protocol is set up to judge automatically the relevance of the sorted induced relations. We finally apply our approach on a French corpus of news by using ROC Curves to evaluate the results.

2019 ◽  
Vol 10 (1) ◽  
pp. 51-67
Author(s):  
Marta Iwaszuk

Thesis: Aim of the paper is to present Melanie Klein and Charles S. Peirce concept of symbol in order to combine them into scheme that presents conscious and unconscious aspect of thinking through symbolic signs (signs based on convention). Presented concepts: Paper presents concept of symbol in psychoanalytical and semiotic perspective. Psychoanalytical view is based on interpretation of symbol according to object relation paradigm proposed by Klein. There are two reasons for selecting her theory for the model: it is most closely bound with interdependency between communication and thinking plus her concept of proper symbol fulfills definition of symbolic sign in Peirce theory, due to deployment of matter of absence in substitution process. Peirce theory however is selected to present semiotic perspective not only for its good linkage to Klein’s “proper symbol” but also for its accurate understating of object representation in quasi- mind through Representamen and as a result recognition of symbol embedment in code through unlimited semiosis. Chosen concepts are consolidated into psycho-semiotic model of thinking which recognizes symbol to be co-created by unique internal world of unconscious phantasy with simultaneous employment of semiotic devices oriented to external, group order perspective. Results and conclusions: Proposed psycho-semiotic model of thinking enhances psychoanalytic view, based on drive for object, by recognizing communication means required for meaningful relation and with that for thinking itself. As a result conceptualizing thinking processes is enriched with semiotic discoveries such as mechanics and structure of Representamen and Interpretant, along with indispensable code rules, with unlimited semiosis at its core. In turn psychoanalytical view adds to semiotic perspective sensitivity to individual potential and constrains when code is in use and with that raises precision of exploration in the field. Contribution to the field: Proposed model enriches theory of thinking based on object relations with semiotic sign theory, which being focused on communication serves as a frame for establishing object relations and their conceptualization. In turn employing psychoanalytic perspective into semiotic field brings back code theory to actual code usage, and by that expands it to various unconscious forces, which ultimately determine Interpretant


Author(s):  
Zdenek Struska ◽  
Jirí Vanícek ◽  
Martin Závodný

The area of applications development for government purposes can be characterized to be task specific. In this context, development projects are usually more complex and there are some differences in comparison with commercial projects. The mission of the proposed chapter is an explanation of methods of project complexity evaluation based on analogy, crisp and fuzzy expert estimation and measure models. The selected methods for aggregation of expert’s estimations are also presented. Further the chapter introduces selected methods designed for complexity estimation. All the introduced methods are widely known except one that was designed by the lead author of the chapter. The method is called BORM Points and is developed for an IS project designed in BORM method (Business Object Relation Modeling). Each method is introduced first, then its step-by-step computation procedure is described and finally suggestion of software, which is supported method computation procedure. The results of the methods are in non-dimensional numbers and it is necessary to set up the relationship between complexity and effort, and introduces COCOMO model and its variants. Efforts are given about the implementation of this form of estimation approach in the area of ICT governance, especially at the grass roots e-governance.


1988 ◽  
Vol 11 (4) ◽  
pp. 235-242 ◽  
Author(s):  
C. Lamberti ◽  
E. Sarti ◽  
A. Santoro ◽  
M. Spongano ◽  
P. Zucchelli ◽  
...  

A mathematical model of hydroelectrolyte exchanges and arterial pressure regulation in the human body during dialysis has been set up. It is conceived as a tool for a new dialysis unit which will be able to “interpret” the signals supplied by suitable instruments connected to the patient and modify the machine set-points in real time in order to obtain clinical results defined by the physician. The main aim is the prevention of hypotensive episodes during treatment. An experimental protocol has been developed for parameter estimation of each patient during a single dialysis. Clinical tests illustrated the model's ability to fit the patient's state during dialysis. This is the first step in the more general task of validation of the model, necessary for the achievement of a closed-loop dialysis unit.


2019 ◽  
Vol 141 (9) ◽  
Author(s):  
Tianjun Hou ◽  
Bernard Yannou ◽  
Yann Leroy ◽  
Emilie Poirson

Customers post online reviews at any time. With the timestamp of online reviews, they can be regarded as a flow of information. With this characteristic, designers can capture the changes in customer feedback to help set up product improvement strategies. Here, we propose an approach for capturing changes in user expectation on product affordances based on the online reviews for two generations of products. First, the approach uses a rule-based natural language processing method to automatically identify and structure product affordances from review text. Then, inspired by the Kano model which classifies preferences of product attributes in five categories, conjoint analysis is used to quantitatively categorize the structured affordances. Finally, changes in user expectation can be found by applying the conjoint analysis on the online reviews posted for two successive generations of products. A case study based on the online reviews of Kindle e-readers downloaded from amazon.com shows that designers can use our proposed approach to evaluate their product improvement strategies for previous products and develop new product improvement strategies for future products.


Author(s):  
John Archibald

The distinction between representations and processes is central to most models of the cognitive science of language. Linguistic theory informs the types of representations assumed, and these representations are what are taken to be the targets of second language acquisition. Epistemologically, this is often taken to be knowledge, or knowledge-that. Techniques such as Grammaticality Judgment tasks are paradigmatic as we seek to gain insight into what a learner’s grammar looks like. Learners behave as if certain phonological, morphological, or syntactic strings (which may or may not be target-like) were well-formed. It is the task of the researcher to understand the nature of the knowledge that governs those well-formedness beliefs. Traditional accounts of processing, on the other hand, look to the real-time use of language, either in production or perception, and invoke discussions of skill or knowledge-how. A range of experimental psycholinguistic techniques have been used to assess these skills: self-paced reading, eye-tracking, ERPs, priming, lexical decision, AXB discrimination, and the like. Such online measures can show us how we “do” language when it comes to activities such as production or comprehension. There has long been a connection between linguistic theory and theories of processing as evidenced by the work of Berwick (The Grammatical Basis of Linguistic Performance). The task of the parser is to assign abstract structure to a phonological, morphological, or syntactic string; structure that does not come directly labeled in the acoustic input. Such processing studies as the Garden Path phenomenon have revealed that grammaticality and processability are distinct constructs. In some models, however, the distinction between grammar and processing is less distinct. Phillips says that “parsing is grammar,” while O’Grady builds an emergentist theory with no grammar, only processing. Bayesian models of acquisition, and indeed of knowledge, assume that the grammars we set up are governed by a principle of entropy, which governs other aspects of human behavior; knowledge and skill are combined. Exemplar models view the processing of the input as a storing of all phonetic detail that is in the environment, not storing abstract categories; the categories emerge via a process of comparing exemplars. Linguistic theory helps us to understand the processing of input to acquire new L2 representations, and the access of those representations in real time.


2021 ◽  
Author(s):  
K. Segaert ◽  
C. Poulisse ◽  
R. Markiewicz ◽  
L. Wheeldon ◽  
D. Marchment ◽  
...  

AbstractMild cognitive impairment (MCI) is the term used to identify those individuals with subjective and objective cognitive decline but with preserved activities of daily living and an absence of dementia. While MCI can impact functioning in different cognitive domains, most notably episodic memory, relatively little is known about the comprehension of language in MCI. In this study we used around-the-ear electrodes (cEEGrids) to identify impairments during language comprehension in MCI patients. In a group of 23 MCI patients and 23 age-matched controls, language comprehension was tested in a two-word phrase paradigm. We examined the oscillatory changes following word onset as a function of lexical retrieval (e.g. swrfeq versus swift) and semantic binding (e.g. horse preceded by swift versus preceded by swrfeq). Electrophysiological signatures (as measured by the cEEGrids) were significantly different between MCI patients and controls. In controls lexical retrieval was associated with a rebound in the alpha/beta range and semantic binding was associated with a post-word alpha/beta suppression. In contrast, both the lexical retrieval and semantic binding signatures were absent in the MCI group. The signatures observed using cEEGrids in controls were comparable to those signatures obtained with a full-cap EEG set-up. Importantly, our findings suggest that MCI patients have impaired electrophysiological signatures for comprehending single-words and multi-word phrases. Moreover, cEEGrids set-ups provide a non-invasive and sensitive clinical tool for detecting early impairments in language comprehension in MCI.


2021 ◽  
Author(s):  
Insook Cho ◽  
Minyoung Lee ◽  
Yeonjin Kim

Patient safety is a fundamental aspect of the quality of healthcare and there is a growing interest in improving safety among healthcare stakeholders in many countries. The Korean government recognized that patient safety is a threat to society following several serious adverse events, and so the Ministry of Health and Welfare of the Korean government set up the Patient Safety Act in January 2015. This study analyzed text data on patient safety collected from web-based, user-generated documents related to the legislation to see if they accurately represent the specific concerns of various healthcare stakeholders. We adopted the unsupervised natural language processing method of probabilistic topic modeling and also Latent Dirichlet Allocation. The results showed that text data are useful for inferring the latent concerns of healthcare consumers, providers, government bodies, and researchers as well as changes therein over time.


Robotica ◽  
2018 ◽  
Vol 37 (2) ◽  
pp. 246-263 ◽  
Author(s):  
Hachem A. Lamti ◽  
Mohamed Moncef Ben Khelifa ◽  
Vincent Hugel

SUMMARYThe goal of this paper is to present a new hybrid system based on the fusion of gaze data and Steady State Visual Evoked Potentials (SSVEP) not only to command a powered wheelchair, but also to account for users distraction levels (concentrated or distracted). For this purpose, a multi-layer perception neural network was set up in order to combine relevant gazing and blinking features from gaze sequence and brainwave features from occipital and parietal brain regions. The motivation behind this work is the shortages raised from the individual use of gaze-based and SSVEP-based wheelchair command techniques. The proposed framework is based on three main modules: a gaze module to select command and activate the flashing stimuli. An SSVEP module to validate the selected command. In parallel, a distraction level module estimates the intention of the user by mean of behavioral entropy and validates/inhibits the command accordingly. An experimental protocol was set up and the prototype was tested on five paraplegic subjects and compared with standard SSVEP and gaze-based systems. The results showed that the new framework performed better than conventional gaze-based and SSVEP-based systems. Navigation performance was assessed based on navigation time and obstacles collisions.


2008 ◽  
Vol 2008 ◽  
pp. 1-9 ◽  
Author(s):  
Mi-Young Kim

Interactions between proteins and genes are considered essential in the description of biomolecular phenomena, and networks of interactions are applied in a system's biology approach. Recently, many studies have sought to extract information from biomolecular text using natural language processing technology. Previous studies have asserted that linguistic information is useful for improving the detection of gene interactions. In particular, syntactic relations among linguistic information are good for detecting gene interactions. However, previous systems give a reasonably good precision but poor recall. To improve recall without sacrificing precision, this paper proposes a three-phase method for detecting gene interactions based on syntactic relations. In the first phase, we retrieve syntactic encapsulation categories for each candidate agent and target. In the second phase, we construct a verb list that indicates the nature of the interaction between pairs of genes. In the last phase, we determine direction rules to detect which of two genes is the agent or target. Even without biomolecular knowledge, our method performs reasonably well using a small training dataset. While the first phase contributes to improve recall, the second and third phases contribute to improve precision. In the experimental results using ICML 05 Workshop on Learning Language in Logic (LLL05) data, our proposed method gave an F-measure of 67.2% for the test data, significantly outperforming previous methods. We also describe the contribution of each phase to the performance.


2020 ◽  
Vol 7 (Supplement_1) ◽  
pp. S92-S92
Author(s):  
Valeria Fabre ◽  
Valeria Fabre ◽  
George Jones ◽  
Joe Amoah ◽  
Eili Klein ◽  
...  

Abstract Background Syndrome-based antibiotic stewardship can be limited by difficulty in finding cases for evaluation. We developed an electronic extraction algorithm to prospectively identify CAP patients. Methods We included non-oncology patients ≥18 years old admitted to The Johns Hopkins Hospital from 12/2018 to 3/2019 who 1) received common CAP antibiotics for ≥48 hours after admission and 2) had a bacterial urinary antigen and chest imaging ordered within 48 hours of admission that was not for assessment of endotracheal tube or central line placement. Charts of patients meeting these criteria were reviewed by 2 authors to identify true cases of CAP based on IDSA guidelines. Cases identified in 12/2018 (n=111) were used to explore potential indicators of CAP, and cases identified 1–3/2019 (n=173) were used to evaluate combinations of indicators that could identify patients treated for CAP who did have CAP (true CAP) and did not have CAP (false CAP). This cohort was divided into a training and a validation set (2/3 and 1/3, respectively). Potential indicators included vitals signs, laboratory data and free text extracted via natural language processing (NLP). Predictive performance of composite indicators for true CAP were assessed using receiver-operating characteristics (ROC) curves. The Hosmer-Lemeshow goodness fit test was used to test model fit and the Akaike Information Criteria was used to determine model selection. Results True CAP was observed in 41% (71/173) of cases and 14 potential individual indicators were identified (Table). These were combined to make 45 potential composite indicators. ROC curves for selected composite indicators are shown in the Figure. Models without use of NLP-derived variables had poor discriminative ability. The best model included fever, hypoxemia, leukocytosis, and “consolidation” on imaging with a sensitivity and positive predictive value 78.7% and specificity and negative predictive value of 85.7%. Table. Indicators evaluated to identify patients with CAP Figure. ROC curves for composite indicators Conclusion Patients with CAP can be identified using electronic data but use of NLP-derived radiographic criteria is required. These data can be linked with data on antibiotic use and duration to develop reports for clinicians regarding appropriate CAP diagnosis and treatment. Disclosures All Authors: No reported disclosures


Sign in / Sign up

Export Citation Format

Share Document