scholarly journals Photon-HDF5: open data format and computational tools for timestamp-based single-molecule experiments

Author(s):  
Antonino Ingargiola ◽  
Ted Laurence ◽  
Robert Boutelle ◽  
Shimon Weiss ◽  
Xavier Michalet
2021 ◽  
Author(s):  
Nadia M Huisjes ◽  
Thomas M Retzer ◽  
Matthias J Scherr ◽  
Rohit Agarwal ◽  
Barbara Safaric ◽  
...  

The rapid development of new imaging approaches is generating larger and more complex datasets revealing the time evolution of individual cells and biomolecules. Single-molecule techniques, in particular, provide access to rare intermediates in complex, multistage molecular pathways, but few standards exist for processing these information-rich datasets, posing challenges for wider dissemination. Here, we present Mars, an open-source platform for storage and processing of image-derived properties of biomolecules. Mars provides Fiji/ImageJ2 commands written in Java for common single-molecule analysis tasks using a Molecule Archive architecture that is easily adapted to complex, multistep analysis workflows. Three diverse workflows involving molecule tracking, multichannel fluorescence imaging, and force spectroscopy, demonstrate the range of analysis applications. A comprehensive graphical user interface written in JavaFX enhances biomolecule feature exploration by providing charting, tagging, region highlighting, scriptable dashboards, and interactive image views. The interoperability of ImageJ2 ensures Molecule Archives can easily be opened in multiple environments, including those written in Python using PyImageJ, for interactive scripting and visualization. Mars provides a flexible solution for reproducible analysis of image-derived properties facilitating the discovery and quantitative classification of new biological phenomena with an open data format accessible to everyone.


2016 ◽  
Vol 110 (3) ◽  
pp. 634a ◽  
Author(s):  
Antonino Ingargiola ◽  
Ted Laurence ◽  
Robert Boutelle ◽  
Shimon Weiss ◽  
Xavier Michalet

Author(s):  
Aatif Ahmad Khan ◽  
Sanjay Kumar Malik

Semantic Search refers to set of approaches dealing with usage of Semantic Web technologies for information retrieval in order to make the process machine understandable and fetch precise results. Knowledge Bases (KB) act as the backbone for semantic search approaches to provide machine interpretable information for query processing and retrieval of results. These KB include Resource Description Framework (RDF) datasets and populated ontologies. In this paper, an assessment of the largest cross-domain KB is presented that are exploited in large scale semantic search and are freely available on Linked Open Data Cloud. Analysis of these datasets is a prerequisite for modeling effective semantic search approaches because of their suitability for particular applications. Only the large scale, cross-domain datasets are considered, which are having sizes more than 10 million RDF triples. Survey of sizes of the datasets in triples count has been depicted along with triples data format(s) supported by them, which is quite significant to develop effective semantic search models.


2020 ◽  
Vol 21 (1) ◽  
Author(s):  
Stefania Alexandra Iakab ◽  
Lluc Sementé ◽  
María García-Altares ◽  
Xavier Correig ◽  
Pere Ràfols

Abstract Background Multimodal imaging that combines mass spectrometry imaging (MSI) with Raman imaging is a rapidly developing multidisciplinary analytical method used by a growing number of research groups. Computational tools that can visualize and aid the analysis of datasets by both techniques are in demand. Results Raman2imzML was developed as an open-source converter that transforms Raman imaging data into imzML, a standardized common data format created and adopted by the mass spectrometry community. We successfully converted Raman datasets to imzML and visualized Raman images using open-source software designed for MSI applications. Conclusion Raman2imzML enables both MSI and Raman images to be visualized using the same file format and the same software for a straightforward exploratory imaging analysis.


2017 ◽  
Vol 6 (3) ◽  
pp. 44-56 ◽  
Author(s):  
Pierre Benning

The industry foundation classes (IFCs) data model is a neutral and open data format defined by an international standard (ISO 16739), which allows the description of a construction as a collection of standard objects. These objects are quite well defined for describing a building, but their use is still far from being adapted (and then adopted) for infrastructure. The article presents a new methodology to enrich the IFC model for an infrastructure, in particular, for the scope of bridges, based on a system approach. The first step is to identify all the absent concepts and classes in the current IFC definition, procedural geometry, coordinate systems, etc., and then proposes “bridge oriented” new entities in order to enrich the current IFC model. The next IFC development phases, dedicated to other infrastructure domains, will be based on this experienced methodology.


2017 ◽  
Vol 136 (1) ◽  
pp. 75-92
Author(s):  
Torsten Strasser ◽  
Tobias Peters ◽  
Herbert Jägle ◽  
Eberhart Zrenner

2021 ◽  
Vol 22 (1) ◽  
Author(s):  
Ankeeta Shah ◽  
Briana E. Mittleman ◽  
Yoav Gilad ◽  
Yang I. Li

Abstract Background Alternative cleavage and polyadenylation (APA), an RNA processing event, occurs in over 70% of human protein-coding genes. APA results in mRNA transcripts with distinct 3′ ends. Most APA occurs within 3′ UTRs, which harbor regulatory elements that can impact mRNA stability, translation, and localization. Results APA can be profiled using a number of established computational tools that infer polyadenylation sites from standard, short-read RNA-seq datasets. Here, we benchmarked a number of such tools—TAPAS, QAPA, DaPars2, GETUTR, and APATrap— against 3′-Seq, a specialized RNA-seq protocol that enriches for reads at the 3′ ends of genes, and Iso-Seq, a Pacific Biosciences (PacBio) single-molecule full-length RNA-seq method in their ability to identify polyadenylation sites and quantify polyadenylation site usage. We demonstrate that 3′-Seq and Iso-Seq are able to identify and quantify the usage of polyadenylation sites more reliably than computational tools that take short-read RNA-seq as input. However, we find that running one such tool, QAPA, with a set of polyadenylation site annotations derived from small quantities of 3′-Seq or Iso-Seq can reliably quantify variation in APA across conditions, such asacross genotypes, as demonstrated by the successful mapping of alternative polyadenylation quantitative trait loci (apaQTL). Conclusions We envisage that our analyses will shed light on the advantages of studying APA with more specialized sequencing protocols, such as 3′-Seq or Iso-Seq, and the limitations of studying APA with short-read RNA-seq. We provide a computational pipeline to aid in the identification of polyadenylation sites and quantification of polyadenylation site usages using Iso-Seq data as input.


2021 ◽  
Vol 24 (67) ◽  
pp. 121-128
Author(s):  
Gerardo Ernesto Rolong Agudelo ◽  
Carlos Enrique Montenegro Marin ◽  
Paulo Alonso Gaona-Garcia

In the world and some countries like Colombia, the number of missing person is a phenome very worrying and growing, every year, thousands of people are reported missing all over the world, the fact that this keeps happening might indicate that there are still analyses that have not been done and tools that have not been considered in order to find patterns in the information of missing person. The present article presents a study of the way informatics and computational tools can be used to help find missing person and what patterns can be found in missing person datasets using as a study case open data about missing person in Colombia in 2017. The goal of this study is to review how computational tools like data mining and image analysis can be used to help find missing person and draw patterns in the available information about missing person. For this, first it will be review of the state of art of image analysis in real world applications was made in order to explore the possibilities when studying the photos of missing person, then a data mining process with data of missing person in Colombia was conducted to produce a set of decision rules that can explain the cause of the disappearance, as a result is generated decision rules algorithm suggest links between socioeconomic stratification, age, gender and specific locations of Colombia and the missing person reports. In conclusion, this work reviews what information about missing person is available publicly and what analysis can me made with them, showing that data mining and face recognition can be useful tools to extract patterns and identify patterns in missing person data.


Author(s):  
Nam Gyu Kang ◽  
Kyung Tae Lim ◽  
Woo Seung Jo ◽  
Jeong Oh Kim ◽  
Tae Hyun Kim ◽  
...  

2016 ◽  
Author(s):  
Andrea Motti ◽  
Norman Natali

Detailed information about geology, hydrogeology and seismic hazard issues for Umbria region are contained in a spatial database available as open data format (shapefile or KMZ) and distributed under the regional open data portal called Open Data Umbria (http://dati.umbria.it) where 297 datasets have been produced by Umbria Region until now and most of them are made by Geological Survey. Geological Survey of Regione Umbria carried out a 20 years program to produce 276 geological maps at 1:10.000 reference scale with an accurate geological model of the regional surface and providing millions of geological data. The key word is the characteristic index of the single geologic unit. Characteristic index , shown in percentage, calculates the ratio between the surface of the geologic units compared to their thickness. Thickness value for each geologic unit is intended to be based on rank level and calculated as weighted average of the thickness for each geologic unit.


Sign in / Sign up

Export Citation Format

Share Document