scholarly journals Adaptive multiorientation resolution analysis of complex filamentous network images

2020 ◽  
Vol 36 (20) ◽  
pp. 5093-5103 ◽  
Author(s):  
Mark Kittisopikul ◽  
Amir Vahabikashi ◽  
Takeshi Shimi ◽  
Robert D Goldman ◽  
Khuloud Jaqaman

Abstract Motivation Microscopy images of cytoskeletal, nucleoskeletal and other structures contain complex junctions of overlapping filaments with arbitrary geometry. Yet, state-of-the-art algorithms generally perform single orientation analysis to segment these structures, resulting in gaps near junctions, or assume particular junction geometries to detect them. Results We developed a fully automated image analysis approach to address the challenge of determining the number of orientations and their values at each point in space to detect both lines and their junctions. Our approach does not assume any fixed number of orientations or any particular geometry in the case of multiple coincident orientations. It is based on analytically resolving coincident orientations revealed by steerable ridge filtering in an adaptive manner that balances orientation resolution and spatial localization. Combining this multiorientation resolution information with a generalization of the concept of non-maximum suppression allowed us to then identify the centers of lines and their junctions in an image. We validated our approach using a wide array of synthetic junctions and by comparison to manual segmentation. We also applied it to light microscopy images of cytoskeletal and nucleoskeletal networks. Availability and implementation https://github.com/mkitti/AdaptiveResolutionOrientationSpace. Supplementary information Supplementary information is available at Bioinformatics online.

2019 ◽  
Author(s):  
Mark Kittisopikul ◽  
Amir Vahabikashi ◽  
Takeshi Shimi ◽  
Robert D. Goldman ◽  
Khuloud Jaqaman

AbstractMicroscopy images of cytoskeletal, nucleoskeletal, and other filamentous structures contain complex junctions where multiple filaments with distinct orientations overlap, yet state-of-the-art software generally uses single orientation analysis to segment these structures. We describe an image analysis approach to simultaneously segment both filamentous structures and their intersections in microscopy images, based on analytically resolving coincident multiple orientations upon image filtering in a manner that balances orientation resolution and spatial localization.


2020 ◽  
Vol 36 (17) ◽  
pp. 4668-4670
Author(s):  
Tim Scherr ◽  
Karolin Streule ◽  
Andreas Bartschat ◽  
Moritz Böhland ◽  
Johannes Stegmaier ◽  
...  

Abstract Motivation An automated counting of beads is required for many high-throughput experiments such as studying mimicked bacterial invasion processes. However, state-of-the-art algorithms under- or overestimate the number of beads in low-resolution images. In addition, expert knowledge is needed to adjust parameters. Results In combination with our image labeling tool, BeadNet enables biologists to easily annotate and process their data reducing the expertise required in many existing image analysis pipelines. BeadNet outperforms state-of-the-art-algorithms in terms of missing, added and total amount of beads. Availability and implementation BeadNet (software, code and dataset) is available at https://bitbucket.org/t_scherr/beadnet. The image labeling tool is available at https://bitbucket.org/abartschat/imagelabelingtool. Supplementary information Supplementary data are available at Bioinformatics online.


Author(s):  
Matteo Chiara ◽  
Federico Zambelli ◽  
Marco Antonio Tangaro ◽  
Pietro Mandreoli ◽  
David S Horner ◽  
...  

Abstract Summary While over 200 000 genomic sequences are currently available through dedicated repositories, ad hoc methods for the functional annotation of SARS-CoV-2 genomes do not harness all currently available resources for the annotation of functionally relevant genomic sites. Here, we present CorGAT, a novel tool for the functional annotation of SARS-CoV-2 genomic variants. By comparisons with other state of the art methods we demonstrate that, by providing a more comprehensive and rich annotation, our method can facilitate the identification of evolutionary patterns in the genome of SARS-CoV-2. Availabilityand implementation Galaxy   http://corgat.cloud.ba.infn.it/galaxy; software: https://github.com/matteo14c/CorGAT/tree/Revision_V1; docker: https://hub.docker.com/r/laniakeacloud/galaxy_corgat. Supplementary information Supplementary data are available at Bioinformatics online.


Algorithms ◽  
2021 ◽  
Vol 14 (5) ◽  
pp. 146
Author(s):  
Aleksei Vakhnin ◽  
Evgenii Sopov

Modern real-valued optimization problems are complex and high-dimensional, and they are known as “large-scale global optimization (LSGO)” problems. Classic evolutionary algorithms (EAs) perform poorly on this class of problems because of the curse of dimensionality. Cooperative Coevolution (CC) is a high-performed framework for performing the decomposition of large-scale problems into smaller and easier subproblems by grouping objective variables. The efficiency of CC strongly depends on the size of groups and the grouping approach. In this study, an improved CC (iCC) approach for solving LSGO problems has been proposed and investigated. iCC changes the number of variables in subcomponents dynamically during the optimization process. The SHADE algorithm is used as a subcomponent optimizer. We have investigated the performance of iCC-SHADE and CC-SHADE on fifteen problems from the LSGO CEC’13 benchmark set provided by the IEEE Congress of Evolutionary Computation. The results of numerical experiments have shown that iCC-SHADE outperforms, on average, CC-SHADE with a fixed number of subcomponents. Also, we have compared iCC-SHADE with some state-of-the-art LSGO metaheuristics. The experimental results have shown that the proposed algorithm is competitive with other efficient metaheuristics.


Author(s):  
J.M. Murray ◽  
P. Pfeffer ◽  
R. Seifert ◽  
A. Hermann ◽  
J. Handke ◽  
...  

Objective: Manual plaque segmentation in microscopy images is a time-consuming process in atherosclerosis research and potentially subject to unacceptable user-to-user variability and observer bias. We address this by releasing Vesseg a tool that includes state-of-the-art deep learning models for atherosclerotic plaque segmentation. Approach and Results: Vesseg is a containerized, extensible, open-source, and user-oriented tool. It includes 2 models, trained and tested on 1089 hematoxylin-eosin stained mouse model atherosclerotic brachiocephalic artery sections. The models were compared to 3 human raters. Vesseg can be accessed at https://vesseg .online or downloaded. The models show mean Soerensen-Dice scores of 0.91±0.15 for plaque and 0.97±0.08 for lumen pixels. The mean accuracy is 0.98±0.05. Vesseg is already in active use, generating time savings of >10 minutes per slide. Conclusions: Vesseg brings state-of-the-art deep learning methods to atherosclerosis research, providing drastic time savings, while allowing for continuous improvement of models and the underlying pipeline.


2020 ◽  
Vol 36 (10) ◽  
pp. 3011-3017 ◽  
Author(s):  
Olga Mineeva ◽  
Mateo Rojas-Carulla ◽  
Ruth E Ley ◽  
Bernhard Schölkopf ◽  
Nicholas D Youngblut

Abstract Motivation Methodological advances in metagenome assembly are rapidly increasing in the number of published metagenome assemblies. However, identifying misassemblies is challenging due to a lack of closely related reference genomes that can act as pseudo ground truth. Existing reference-free methods are no longer maintained, can make strong assumptions that may not hold across a diversity of research projects, and have not been validated on large-scale metagenome assemblies. Results We present DeepMAsED, a deep learning approach for identifying misassembled contigs without the need for reference genomes. Moreover, we provide an in silico pipeline for generating large-scale, realistic metagenome assemblies for comprehensive model training and testing. DeepMAsED accuracy substantially exceeds the state-of-the-art when applied to large and complex metagenome assemblies. Our model estimates a 1% contig misassembly rate in two recent large-scale metagenome assembly publications. Conclusions DeepMAsED accurately identifies misassemblies in metagenome-assembled contigs from a broad diversity of bacteria and archaea without the need for reference genomes or strong modeling assumptions. Running DeepMAsED is straight-forward, as well as is model re-training with our dataset generation pipeline. Therefore, DeepMAsED is a flexible misassembly classifier that can be applied to a wide range of metagenome assembly projects. Availability and implementation DeepMAsED is available from GitHub at https://github.com/leylabmpi/DeepMAsED. Supplementary information Supplementary data are available at Bioinformatics online.


Author(s):  
Gabriel Landini ◽  
Giovanni Martinelli ◽  
Filippo Piccinini

Abstract Motivation Microscopy images of stained cells and tissues play a central role in most biomedical experiments and routine histopathology. Storing colour histological images digitally opens the possibility to process numerically colour distribution and intensity to extract quantitative data. Among those numerical procedures are colour deconvolution, which enable decomposing an RGB image into channels representing the optical absorbance and transmittance of the dyes when their RGB representation is known. Consequently, a range of new applications become possible for morphological and histochemical segmentation, automated marker localization and image enhancement. Availability and implementation Colour deconvolution is presented here in two open-source forms: a MATLAB program/function and an ImageJ plugin written in Java. Both versions run in Windows, Macintosh and UNIX-based systems under the respective platforms. Source code and further documentation are available at: https://blog.bham.ac.uk/intellimic/g-landini-software/colour-deconvolution-2/. Supplementary information Supplementary data are available at Bioinformatics online.


Author(s):  
Kexin Huang ◽  
Tianfan Fu ◽  
Lucas M Glass ◽  
Marinka Zitnik ◽  
Cao Xiao ◽  
...  

Abstract Summary Accurate prediction of drug–target interactions (DTI) is crucial for drug discovery. Recently, deep learning (DL) models for show promising performance for DTI prediction. However, these models can be difficult to use for both computer scientists entering the biomedical field and bioinformaticians with limited DL experience. We present DeepPurpose, a comprehensive and easy-to-use DL library for DTI prediction. DeepPurpose supports training of customized DTI prediction models by implementing 15 compound and protein encoders and over 50 neural architectures, along with providing many other useful features. We demonstrate state-of-the-art performance of DeepPurpose on several benchmark datasets. Availability and implementation https://github.com/kexinhuang12345/DeepPurpose. Contact [email protected] Supplementary information Supplementary data are available at Bioinformatics online.


2020 ◽  
Vol 2020 ◽  
pp. 1-11
Author(s):  
Weijia Wu ◽  
Jici Xing ◽  
Cheng Yang ◽  
Yuxing Wang ◽  
Hong Zhou

The performance of text detection is crucial for the subsequent recognition task. Currently, the accuracy of the text detector still needs further improvement, particularly those with irregular shapes in a complex environment. We propose a pixel-wise method based on instance segmentation for scene text detection. Specifically, a text instance is split into five components: a Text Skeleton and four Directional Pixel Regions, then restoring itself based on these elements and receiving supplementary information from other areas when one fails. Besides, a Confidence Scoring Mechanism is designed to filter characters similar to text instances. Experiments on several challenging benchmarks demonstrate that our method achieves state-of-the-art results in scene text detection with an F-measure of 84.6% on Total-Text and 86.3% on CTW1500.


2019 ◽  
Vol 35 (21) ◽  
pp. 4525-4527 ◽  
Author(s):  
Alex X Lu ◽  
Taraneh Zarin ◽  
Ian S Hsu ◽  
Alan M Moses

Abstract Summary We introduce YeastSpotter, a web application for the segmentation of yeast microscopy images into single cells. YeastSpotter is user-friendly and generalizable, reducing the computational expertise required for this critical preprocessing step in many image analysis pipelines. Availability and implementation YeastSpotter is available at http://yeastspotter.csb.utoronto.ca/. Code is available at https://github.com/alexxijielu/yeast_segmentation. Supplementary information Supplementary data are available at Bioinformatics online.


Sign in / Sign up

Export Citation Format

Share Document