scholarly journals SKIMMR: Facilitating knowledge discovery in life sciences by machine-aided skim reading

Author(s):  
Vit Novacek ◽  
Gully APC Burns

Background: Unlike full reading, 'skim-reading' involves the process of looking quickly over information in an attempt to cover more material whilst still being able to retain a superficial view of the underlying content. Within this work, we specifically emulate this natural human activity by providing a dynamic graph-based view of entities automatically extracted from text. For the extraction, we use shallow parsing, co-occurrence analysis and semantic similarity computation techniques. Our main motivation is to assist biomedical researchers and clinicians in coping with increasingly large amounts of potentially relevant articles that are being published ongoingly in life sciences. Methods: To construct the high-level network overview of articles, we extract weighted binary statements from the text. We consider two types of these statements, co-occurrence and similarity, both organised in the same distributional representation (i.e., in a vector-space model). For the co-occurrence weights, we use point-wise mutual information that indicates the degree of non-random association between two co-occurring entities. For computing the similarity statement weights, we use cosine distance based on the relevant co-occurrence vectors. These statements are used to build fuzzy indices of terms, statements and provenance article identifiers, which support fuzzy querying and subsequent result ranking. These indexing and querying processes are then used to construct a graph-based interface for searching and browsing entity networks extracted from articles, as well as articles relevant to the networks being browsed. Last but not least, we describe a methodology for automated experimental evaluation of the presented approach. The method uses formal comparison of the graphs generated by our tool to relevant gold standards based on manually curated PubMed, TREC challenge and MeSH data. Results: We provide a web-based prototype (called `SKIMMR') that generates a network of inter-related entities from a set of documents which a user may explore through our interface. When a particular area of the entity network looks interesting to a user, the tool displays the documents that are the most relevant to those entities of interest currently shown in the network. We present this as a methodology for browsing a collection of research articles. To illustrate the practical applicability of SKIMMR, we present examples of its use in the domains of Spinal Muscular Atrophy and Parkinson's Disease. Finally, we report on the results of experimental evaluation using the two domains and one additional dataset based on the TREC challenge. The results show that the presented method for machine-aided skim reading outperforms tools like PubMed regarding focused browsing and informativeness of the browsing context.

2014 ◽  
Author(s):  
Vit Novacek ◽  
Gully APC Burns

Background: Unlike full reading, 'skim-reading' involves the process of looking quickly over information in an attempt to cover more material whilst still being able to retain a superficial view of the underlying content. Within this work, we specifically emulate this natural human activity by providing a dynamic graph-based view of entities automatically extracted from text. For the extraction, we use shallow parsing, co-occurrence analysis and semantic similarity computation techniques. Our main motivation is to assist biomedical researchers and clinicians in coping with increasingly large amounts of potentially relevant articles that are being published ongoingly in life sciences. Methods: To construct the high-level network overview of articles, we extract weighted binary statements from the text. We consider two types of these statements, co-occurrence and similarity, both organised in the same distributional representation (i.e., in a vector-space model). For the co-occurrence weights, we use point-wise mutual information that indicates the degree of non-random association between two co-occurring entities. For computing the similarity statement weights, we use cosine distance based on the relevant co-occurrence vectors. These statements are used to build fuzzy indices of terms, statements and provenance article identifiers, which support fuzzy querying and subsequent result ranking. These indexing and querying processes are then used to construct a graph-based interface for searching and browsing entity networks extracted from articles, as well as articles relevant to the networks being browsed. Last but not least, we describe a methodology for automated experimental evaluation of the presented approach. The method uses formal comparison of the graphs generated by our tool to relevant gold standards based on manually curated PubMed, TREC challenge and MeSH data. Results: We provide a web-based prototype (called `SKIMMR') that generates a network of inter-related entities from a set of documents which a user may explore through our interface. When a particular area of the entity network looks interesting to a user, the tool displays the documents that are the most relevant to those entities of interest currently shown in the network. We present this as a methodology for browsing a collection of research articles. To illustrate the practical applicability of SKIMMR, we present examples of its use in the domains of Spinal Muscular Atrophy and Parkinson's Disease. Finally, we report on the results of experimental evaluation using the two domains and one additional dataset based on the TREC challenge. The results show that the presented method for machine-aided skim reading outperforms tools like PubMed regarding focused browsing and informativeness of the browsing context.


2014 ◽  
Author(s):  
Vit Novacek ◽  
Gully APC Burns

Background: Unlike full reading, 'skim-reading' involves the process of looking quickly over information in an attempt to cover more material whilst still being able to retain a superficial view of the underlying content. Within this work, we specifically emulate this natural human activity by providing a dynamic graph-based view of entities automatically extracted from text. For the extraction, we use shallow parsing, co-occurrence analysis and semantic similarity computation techniques. Our main motivation is to assist biomedical researchers and clinicians in coping with increasingly large amounts of potentially relevant articles in life sciences. Methods: To construct the high-level network overview of articles, we extract weighted binary statements from the text. We consider two types of these statements, co-occurrence and similarity, both organised in the same distributional representation (i.e., in a vector-space model). For the co-occurrence weights, we use point-wise mutual information that indicates the degree of non-random association between two co-occurring entities. For computing the similarity statement weights, we use cosine distance based on the relevant co-occurrence vectors. These statements are used to build fuzzy indices of terms, statements and provenance article identifiers, which support fuzzy querying and subsequent result ranking. These indexing and querying processes are then used top construct a graph-based interface for searching and browsing entity networks extracted from articles, as well as articles relevant to the networks being browsed. Results: We provide a web-based prototype (called `SKIMMR') that generates a network of inter-related entities from a set of documents which a user may explore through our interface. When a particular area of the entity network looks interesting to a user, the tool displays the documents that are most relevant to those entities of interest currently shown in the network. We present this as a methodology for browsing a collection of research articles. To illustrate the practical applicability of SKIMMR, we present examples of its use in the domains of Spinal Muscular Atrophy and Parkinson's Disease. Last but not least, we describe a methodology for automated experimental evaluation of SKIMMR instances. The method uses formal comparison of the graphs generated by our tool to relevant gold standards based on manually curated PubMed, TREC challenge and MeSH data. The results of experiments performed on three different instances of SKIMMR show that the presented method for machine-aided skim reading outperforms state of the art tools like PubMed regarding focused browsing and informativeness of the browsing context. Conclusions: In preliminary trials, users find new, interesting and non-trivial facts with SKIMMR. Our evaluation showed a high potential of the presented work for facilitating knowledge discovery in life sciences.


2014 ◽  
Author(s):  
Vit Novacek ◽  
Gully APC Burns

Background: Unlike full reading, 'skim-reading' involves the process of looking quickly over information in an attempt to cover more material whilst still being able to retain a superficial view of the underlying content. Within this work, we specifically emulate this natural human activity by providing a dynamic graph-based view of entities automatically extracted from text. For the extraction, we use shallow parsing, co-occurrence analysis and semantic similarity computation techniques. Our main motivation is to assist biomedical researchers and clinicians in coping with increasingly large amounts of potentially relevant articles that are being published ongoingly in life sciences. Methods: To construct the high-level network overview of articles, we extract weighted binary statements from the text. We consider two types of these statements, co-occurrence and similarity, both organised in the same distributional representation (i.e., in a vector-space model). For the co-occurrence weights, we use point-wise mutual information that indicates the degree of non-random association between two co-occurring entities. For computing the similarity statement weights, we use cosine distance based on the relevant co-occurrence vectors. These statements are used to build fuzzy indices of terms, statements and provenance article identifiers, which support fuzzy querying and subsequent result ranking. These indexing and querying processes are then used top construct a graph-based interface for searching and browsing entity networks extracted from articles, as well as articles relevant to the networks being browsed. Results: We provide a prototype (called SKIMMR) that generates a network of inter-related entities from a set of documents which users may explore through our interface. When a particular area of the entity network looks interesting to a user, the tool displays the documents that are most relevant entities currently shown in the network. We present this as a methodology for browsing a collection of research articles. To illustrate the practical applicability of SKIMMR, we present examples of its use in the domains of Spinal Muscular Atrophy and Parkinson's Disease. Finally, we describe a methodology for automated experimental evaluation of SKIMMR instances. The method uses formal comparison of the graphs generated by our tool to relevant gold standards based on manually curated PubMed, TREC challenge and MeSH data. The results of experiments performed on three different instances of SKIMMR show that the presented method for machine-aided skim reading outperforms state of the art tools like PubMed regarding focused browsing and informativeness of the browsing context. Conclusions: In preliminary trials, sample users find new, interesting and non-trivial facts with the tool. Our evaluation showed a high potential of the presented work for facilitating knowledge discovery in life sciences.


2005 ◽  
Vol 68 (1) ◽  
pp. 36-43 ◽  
Author(s):  
Gayle Vogt ◽  
Catherine Atwong ◽  
Jean Fuller

Student Assessment of Learning Gains (SALGains) is a Web-based instrument for measuring student perception of their learning in a variety of courses. The authors adapted this instrument to measure students’ achieved proficiency in analyzing cases in an advanced business communication class. The instrument showed that students did achieve a high level of proficiency and that they did so equally in both traditional and online classes.


Sensors ◽  
2021 ◽  
Vol 21 (12) ◽  
pp. 4045
Author(s):  
Alessandro Sassu ◽  
Jose Francisco Saenz-Cogollo ◽  
Maurizio Agelli

Edge computing is the best approach for meeting the exponential demand and the real-time requirements of many video analytics applications. Since most of the recent advances regarding the extraction of information from images and video rely on computation heavy deep learning algorithms, there is a growing need for solutions that allow the deployment and use of new models on scalable and flexible edge architectures. In this work, we present Deep-Framework, a novel open source framework for developing edge-oriented real-time video analytics applications based on deep learning. Deep-Framework has a scalable multi-stream architecture based on Docker and abstracts away from the user the complexity of cluster configuration, orchestration of services, and GPU resources allocation. It provides Python interfaces for integrating deep learning models developed with the most popular frameworks and also provides high-level APIs based on standard HTTP and WebRTC interfaces for consuming the extracted video data on clients running on browsers or any other web-based platform.


2020 ◽  
Vol 15 (1) ◽  
pp. 711-720
Author(s):  
Janetta Niemann ◽  
Justyna Szwarc ◽  
Jan Bocianowski ◽  
Dorota Weigt ◽  
Marek Mrówczyński

AbstractRapeseed (Brassica napus) can be attacked by a wide range of pests, for example, cabbage root fly (Delia radicum) and cabbage aphid (Brevicoryne brassicae). One of the best methods of pest management is breeding for insect resistance in rapeseed. Wild genotypes of Brassicaceae and rapeseed cultivars can be used as a source of resistance. In 2017, 2018, and 2019, field trials were performed to assess the level of resistance to D. radicum and B. brassicae within 53 registered rapeseed cultivars and 31 interspecific hybrid combinations originating from the resources of the Department of Genetics and Plant Breeding of Poznań University of Life Sciences (PULS). The level of resistance varied among genotypes and years. Only one hybrid combination and two B. napus cultivars maintained high level of resistance in all tested years, i.e., B. napus cv. Jet Neuf × B. carinata – PI 649096, Galileus, and Markolo. The results of this research indicate that resistance to insects is present in Brassicaceae family and can be transferred to rapeseed cultivars. The importance of continuous improvement of rapeseed pest resistance and the search for new sources of resistance is discussed; furthermore, plans for future investigations are presented.


Vaccines ◽  
2022 ◽  
Vol 10 (1) ◽  
pp. 90
Author(s):  
Yi Kong ◽  
Hao Jiang ◽  
Zhisheng Liu ◽  
Yi Guo ◽  
Dehua Hu

Objective: To investigate the uptake and vaccination willingness of the COVID-19 vaccine among Chinese residents and analyze the difference and factors that impact vaccination. Methods: The snowball sampling method was used to distribute online questionnaires. Relevant sociodemographic data along with the circumstances of COVID-19 vaccination were collected from the respondents. The χ2 test, independent samples t test and binary logistic regression analysis were used to analyze the data. Results: Among 786 respondents, 84.22% had been vaccinated. Over 80% of the vaccinated population have completed all the injections because of supporting the national vaccination policies of China, while the unvaccinated population (23.91%) is mainly due to personal health status. Meanwhile, statistical analysis revealed that the main predictors of not being vaccinated were younger age (3 to 18 years old), personal health status, and lower vaccinated proportion of family members and close friends (p < 0.05). Conclusions: There was a high level of uptake of the COVID-19 vaccine in China, and people who have not been vaccinated generally had a low willingness to vaccinate in the future. Based on our results, it suggested the next work to expand the coverage of the COVID-19 vaccination should be concentrated on targeted publicity and education for people who have not been vaccinated.


Author(s):  
Krishna N. Jha ◽  
Andrea Morris ◽  
Ed Mytych ◽  
Judith Spering

Abstract Designing aircraft parts requires extensive coordination among multiple distributed design groups. Achieving such a coordination is time-consuming and expensive, but the cost of ignoring or minimizing it is much higher in terms of delayed and inferior quality products. We have built a multi-agent-based system to provide the desired coordination among the design groups, the legacy applications, and other resources during the preliminary design (PD) process. A variety of agents are used to model the various design and control functionalities. The agent-representation includes a formal representation of the task-structures. A web-based user-interface provides high-level interface to the users. The agents collaborate to achieve the design goals.


Author(s):  
Lichia Yiu ◽  
Raymond Saner

Since the 1990s, more and more corporate learning has been moved online to allow for flexibility, just-in-time learning, and cost saving in delivering training. This trend has been evolved along with the introduction of Web-based applications for HRM purposes, known as electronic Human Resource Management (e-HRM). By 2005, 39.67% of the corporate learning, among the ASTD (American Society for Training and Development) benchmarking forum companies, was delivered online in comparison to 10.5% in 2001. E-learning has now reached “a high level of (technical) sophistication, both in terms of instructional development and the effective management of resources” in companies with high performance learning function (ASTD, 2006, p.4). The cost per unit, reported by ASTD in its 2006 State of Industry Report, has been declining since 2000 despite the higher training hours received per employee thanks to the use of technology based training delivery and its scalability. However, the overall quality of e-learning either public available in the market or implemented at the workplace remains unstable.


Semantic Web ◽  
2020 ◽  
Vol 12 (1) ◽  
pp. 143-161 ◽  
Author(s):  
Mads Holten Rasmussen ◽  
Maxime Lefrançois ◽  
Georg Ferdinand Schneider ◽  
Pieter Pauwels

Actors in the Architecture, Engineering, Construction, Owner and Operation (AECOO) industry traditionally exchange building models as files. The Building Information Modelling (BIM) methodology advocates the seamless exchange of all information between related stakeholders using digital technologies. The ultimate evolution of the methodology, BIM Maturity Level 3, envisions interoperable, distributed, web-based, interdisciplinary information exchange among stakeholders across the life-cycle of buildings. The World Wide Web Consortium Linked Building Data Community Group (W3C LBD-CG) hypothesises that the Linked Data models and best practices can be leveraged to achieve this vision in modern web-based applications. In this paper, we introduce the Building Topology Ontology (BOT) as a core vocabulary to this approach. It provides a high-level description of the topology of buildings including storeys and spaces, the building elements they contain, and their web-friendly 3D models. We describe how existing applications produce and consume datasets combining BOT with other ontologies that describe product catalogues, sensor observations, or Internet of Things (IoT) devices effectively implementing BIM Maturity Level 3. We evaluate our approach by exporting and querying three real-life large building models.


Sign in / Sign up

Export Citation Format

Share Document