A closed-domain question answering framework using reliable resources to assist students

2018 ◽  
Vol 24 (5) ◽  
pp. 725-762 ◽  
Author(s):  
CANER DERİCİ ◽  
YİĞİT AYDIN ◽  
ÇİĞDEM YENİALACA ◽  
NİHAL YAĞMUR AYDIN ◽  
GÜNİZİ KARTAL ◽  
...  

AbstractThis paper describes a question answering framework that can answer student questions given in natural language. We suggest a methodology that makes use of reliable resources only, provides the answer in the form of a multi-document summary for both factoid and open-ended questions, and produces an answer also from foreign resources by translating into the native language. The resources are compiled using a question database in the selected domains based on reliability and coverage metrics. A question is parsed using a dependency parser, important parts are extracted by rule-based and statistical methods, the question is converted into a representation, and a query is built. Documents relevant to the query are retrieved from the set of resources. The documents are summarized and the answers to the question together with other relevant information about the topic of the question are shown to the user. A summary answer from the foreign resources is also built by the translation of the input question and the retrieved documents. The proposed approach was applied to the Turkish language and it was tested with several experiments and a pilot study. The experiments have shown that the summaries returned include the answer for about 50–60 percent of the questions. The data bank built for factoid and open-ended questions in the two domains covered is made publicly available.

Author(s):  
Bettina von Helversen ◽  
Stefan M. Herzog ◽  
Jörg Rieskamp

Judging other people is a common and important task. Every day professionals make decisions that affect the lives of other people when they diagnose medical conditions, grant parole, or hire new employees. To prevent discrimination, professional standards require that decision makers render accurate and unbiased judgments solely based on relevant information. Facial similarity to previously encountered persons can be a potential source of bias. Psychological research suggests that people only rely on similarity-based judgment strategies if the provided information does not allow them to make accurate rule-based judgments. Our study shows, however, that facial similarity to previously encountered persons influences judgment even in situations in which relevant information is available for making accurate rule-based judgments and where similarity is irrelevant for the task and relying on similarity is detrimental. In two experiments in an employment context we show that applicants who looked similar to high-performing former employees were judged as more suitable than applicants who looked similar to low-performing former employees. This similarity effect was found despite the fact that the participants used the relevant résumé information about the applicants by following a rule-based judgment strategy. These findings suggest that similarity-based and rule-based processes simultaneously underlie human judgment.


2019 ◽  
Vol 15 (3) ◽  
pp. 79-100 ◽  
Author(s):  
Watanee Jearanaiwongkul ◽  
Frederic Andres ◽  
Chutiporn Anutariya

Nowadays, farmers can search for treatments for their plants using search engines and applications. Most existing works are developed in the form of rule-based question answering platforms. However, an observation could be incorrectly given by the farmer. This work recommends that diseases and treatments must be considered from a set of related observations. Thus, we develop a theoretical framework for systems to manage a farmer's observation data. We investigate and formalize desirable characteristics of such systems. The observation data is attached with a geolocation in which related contextual data is found. The framework is formalized based on algebra, in which required types and functions are identified. Its key characteristics are described by: (1) the defined type called warncons for representing observation data; (2) the similarity function for warncons; and (3) the warncons composition function for composing similar warncons. Finally, we show that the framework helps observation data to become richer and improve advice-finding.


2016 ◽  
Vol 49 (1) ◽  
pp. 302-310 ◽  
Author(s):  
Michael Kachala ◽  
John Westbrook ◽  
Dmitri Svergun

Recent advances in small-angle scattering (SAS) experimental facilities and data analysis methods have prompted a dramatic increase in the number of users and of projects conducted, causing an upsurge in the number of objects studied, experimental data available and structural models generated. To organize the data and models and make them accessible to the community, the Task Forces on SAS and hybrid methods for the International Union of Crystallography and the Worldwide Protein Data Bank envisage developing a federated approach to SAS data and model archiving. Within the framework of this approach, the existing databases may exchange information and provide independent but synchronized entries to users. At present, ways of exchanging information between the various SAS databases are not established, leading to possible duplication and incompatibility of entries, and limiting the opportunities for data-driven research for SAS users. In this work, a solution is developed to resolve these issues and provide a universal exchange format for the community, based on the use of the widely adopted crystallographic information framework (CIF). The previous version of the sasCIF format, implemented as an extension of the core CIF dictionary, has been available since 2000 to facilitate SAS data exchange between laboratories. The sasCIF format has now been extended to describe comprehensively the necessary experimental information, results and models, including relevant metadata for SAS data analysis and for deposition into a database. Processing tools for these files (sasCIFtools) have been developed, and these are available both as standalone open-source programs and integrated into the SAS Biological Data Bank, allowing the export and import of data entries as sasCIF files. Software modules to save the relevant information directly from beamline data-processing pipelines in sasCIF format are also developed. This update of sasCIF and the relevant tools are an important step in the standardization of the way SAS data are presented and exchanged, to make the results easily accessible to users and to promote further the application of SAS in the structural biology community.


2021 ◽  
Vol 183 (23) ◽  
pp. 1-5
Author(s):  
Haniel G. Cavalcante ◽  
Jéferson N. Soares ◽  
José E.B. Maia

2019 ◽  
Vol 184 (18) ◽  
pp. 556-556
Author(s):  
Carla Correia-Gomes ◽  
Madeleine Kate Henry ◽  
Susanna Williamson ◽  
Richard M Irvine ◽  
George J Gunn ◽  
...  

Traditional indicator-based livestock surveillance has been focused on case definitions, definitive diagnoses and laboratory confirmation. The use of syndromic disease surveillance would increase the population base from which animal health data are captured and facilitate earlier detection of new and re-emerging threats to animal health. Veterinary practitioners could potentially play a vital role in such activities. In a pilot study, specialist private veterinary practitioners (PVP) working in the English pig industry were asked to collect and transfer background data and disease incident reports for pig farms visited during the study period. Baseline data from 110 pig farms were received, along with 68 disease incident reports. Reports took an average of approximately 25 minutes to complete. Feedback from the PVPs indicated that they saw value in syndromic surveillance. Maintenance of anonymity in the outputs would be essential, as would timely access for the PVPs to relevant information on syndromic trends. Further guidance and standardisation would also be required. Syndromic surveillance by PVPs is possible for the pig industry. It has potential to fill current gaps in the collection of animal health data, as long as the engagement and participation of data providers can be obtained and maintained.


2015 ◽  
Vol 6 (4) ◽  
pp. 35-49 ◽  
Author(s):  
Laurent Issertial ◽  
Hiroshi Tsuji

This paper proposes a system called CFP Manager specialized on IT field and designed to ease the process of searching conference suitable to one's need. At present, the handling of CFP faces two problems: for emails, the huge quantity of CFP received can be easily skimmed through. For websites, the reviewing of some of the main CFP aggregators available online points out the lack of usable criteria. This system proposes to answer to these problems via its architecture consisting of three components: firstly an Information Extraction module extracting relevant information (as date, location, etc...) from CFP using rule based text mining algorithm. The second component enriches the now extracted data with external one from ontology models. Finally the last one displays the said data and allows the end user to perform complex queries on the CFP dataset and thus allow him to only access to CFP suitable for him. In order to validate the authors' proposal, they eventually process the well-known precision / recall metric on our information extraction component with an average of 0.95 for precision and 0.91 for recall on three different 100 CFP dataset. This paper finally discusses the validity of our approach by confronting our system for different queries with two systems already available online (WikiCFP and IEEE Conference Search) and basic text searching approach standing for searching in an email box. On a 100 CFP dataset with the wide variety of usable data and the possibility to perform complex queries we surpass basic text searching method and WikiCFP by not returning the false positive usually returned by them and find a result close to the IEEE system.


1970 ◽  
Vol 46 (1) ◽  
pp. 39-43
Author(s):  
W. Pleines ◽  
L. Letourneau

Forest surveys based on permanent plots possess peculiarities unusual in other computer applications. The enormous amount of information (30,000 plots, 1,000,000 trees), must be checked, corrected. Relevant information must be selected from the data bank for statistical computations. Because information for decision-making changes, the computer programs must be flexible.This article explains how this was done. In a temporary phase, all card data were "converted" to standard codes and format and written on magnetic tapes. In the file maintenance phase, the data bank is checked and corrected. Volumes are computed, plots checked by accumulating tree data, etc. The file creation phase builds a unit record from plot and tree information. Stratification data can also be merged on the new file. The reporting phase consists of modified versions of programs of the Northeastern Forest Experiment Station in the U.S. (Wilson and Peters, 1967). TABLE computes statistics by strata and condenses them in matrices, OUTPUT prints them in desired form.This computer system is a harmonious combination of special and general purpose programs. CIP experiences in developing these programs may help other foresters hence more exchange of information about data processing is desired.


Sign in / Sign up

Export Citation Format

Share Document