scholarly journals Decomposing Cell Identity for Transfer Learning across Cellular Measurements, Platforms, Tissues, and Species

Cell Systems ◽  
2019 ◽  
Vol 8 (5) ◽  
pp. 395-411.e8 ◽  
Author(s):  
Genevieve L. Stein-O’Brien ◽  
Brian S. Clark ◽  
Thomas Sherman ◽  
Cristina Zibetti ◽  
Qiwen Hu ◽  
...  
Cell Systems ◽  
2021 ◽  
Vol 12 (2) ◽  
pp. 203
Author(s):  
Genevieve L. Stein-O’Brien ◽  
Brian S. Clark ◽  
Thomas Sherman ◽  
Cristina Zibetti ◽  
Qiwen Hu ◽  
...  

2018 ◽  
Author(s):  
Genevieve L. Stein-O’Brien ◽  
Brian S. Clark ◽  
Thomas Sherman ◽  
Cristina Zibetti ◽  
Qiwen Hu ◽  
...  

ABSTRACTNew approaches are urgently needed to glean biological insights from the vast amounts of single cell RNA sequencing (scRNA-Seq) data now being generated. To this end, we propose that cell identity should map to a reduced set of factors which will describe both exclusive and shared biology of individual cells, and that the dimensions which contain these factors reflect biologically meaningful relationships across different platforms, tissues and species. To find a robust set of dependent factors in large-scale scRNA- Seq data, we developed a Bayesian non-negative matrix factorization (NMF) algorithm, scCoGAPS. Application of scCoGAPS to scRNA-Seq data obtained over the course of mouse retinal development identified gene expression signatures for factors associated with specific cell types and continuous biological processes. To test whether these signatures are shared across diverse cellular contexts, we developed projectR to map biologically disparate datasets into the factors learned by scCoGAPS. Because projecting these dimensions preserve relative distances between samples, biologically meaningful relationships/factors will stratify new data consistent with their underlying processes, allowing labels or information from one dataset to be used for annotation of the other—a machine learning concept called transfer learning. Using projectR, data from multiple datasets was used to annotate latent spaces and reveal novel parallels between developmental programs in other tissues, species and cellular assays. Using this approach we are able to transfer cell type and state designations across datasets to rapidly annotate cellular features in a new dataset without a priori knowledge of their type, identify a species-specific signature of microglial cells, and identify a previously undescribed subpopulation of neurosecretory cells within the lung. Together, these algorithms define biologically meaningful dimensions of cellular identity, state, and trajectories that persist across technologies, molecular features, and species.GRAPHICAL ABSTRACT


Diabetes ◽  
2019 ◽  
Vol 68 (Supplement 1) ◽  
pp. 2184-P
Author(s):  
NORIKO KODANI ◽  
MASAKI KOBAYASHI ◽  
OSAMU KIKUCHI ◽  
TADAHIRO KITAMURA ◽  
HIROSHI ITOH ◽  
...  

2019 ◽  
Author(s):  
Qi Yuan ◽  
Alejandro Santana-Bonilla ◽  
Martijn Zwijnenburg ◽  
Kim Jelfs

<p>The chemical space for novel electronic donor-acceptor oligomers with targeted properties was explored using deep generative models and transfer learning. A General Recurrent Neural Network model was trained from the ChEMBL database to generate chemically valid SMILES strings. The parameters of the General Recurrent Neural Network were fine-tuned via transfer learning using the electronic donor-acceptor database from the Computational Material Repository to generate novel donor-acceptor oligomers. Six different transfer learning models were developed with different subsets of the donor-acceptor database as training sets. We concluded that electronic properties such as HOMO-LUMO gaps and dipole moments of the training sets can be learned using the SMILES representation with deep generative models, and that the chemical space of the training sets can be efficiently explored. This approach identified approximately 1700 new molecules that have promising electronic properties (HOMO-LUMO gap <2 eV and dipole moment <2 Debye), 6-times more than in the original database. Amongst the molecular transformations, the deep generative model has learned how to produce novel molecules by trading off between selected atomic substitutions (such as halogenation or methylation) and molecular features such as the spatial extension of the oligomer. The method can be extended as a plausible source of new chemical combinations to effectively explore the chemical space for targeted properties.</p>


2014 ◽  
Author(s):  
Hiroshi Kanayama ◽  
Youngja Park ◽  
Yuta Tsuboi ◽  
Dongmook Yi
Keyword(s):  

Sign in / Sign up

Export Citation Format

Share Document