scholarly journals Deep neural networks identify sequence context features predictive of transcription factor binding

2021 ◽  
Vol 3 (2) ◽  
pp. 172-180
Author(s):  
An Zheng ◽  
Michael Lamkin ◽  
Hanqing Zhao ◽  
Cynthia Wu ◽  
Hao Su ◽  
...  
2020 ◽  
Vol 15 ◽  
Author(s):  
Kanu Geete ◽  
Manish Pandey

Aims: Robust and more accurate method for identifying transcription factor binding sites (TFBS) for gene expression. Background: Deep neural networks (DNNs) have shown promising growth in solving complex machine learning problems. Conventional techniques are comfortably replaced by DNNs in computer vision, signal processing, healthcare and genomics. Understanding DNA sequences is always a crucial task in healthcare and regulatory genomics. For DNA motif prediction, choosing the right dataset with sufficient number of input sequences is crucial in order to design an effective model. Objective: Designing a new algorithm which works on different dataset while an improved performance for TFBS prediction. Methods: With the help of Layerwise Relevance Propagation, the proposed algorithm identifies the invariant features with adaptive noise patterns. Results: The performance is compared by calculating various metrics on standard as well as recent methods and significant improvement is noted. Conclusion: By identifying the invariant and robust features in the DNA sequences,the classification performance can be increased.


2018 ◽  
Author(s):  
Christopher F. Blum ◽  
Markus Kollmann

AbstractMotivationNucleic acids and proteins often have localized sequence motifs that enable highly specific interactions. Due to the biological relevance of sequence motifs, numerous inference methods have been developed. Recently, convolutional neural networks (CNNs) achieved state of the art performance because they can approximate complex motif distributions. These methods were able to learn transcription factor binding sites from ChIP-seq data and to make accurate predictions. However, CNNs learn filters that are difficult to interpret, and networks trained on small data sets often do not generalize optimally to new sequences.ResultsHere we present circular filters, a novel convolutional architecture, that contains all circularly shifted variants of the same filter. We motivate circular filters by the observation that CNNs frequently learn filters that correspond to shifted and truncated variants of the true motif. Circular filters enable learning of non-truncated motifs and allow easy interpretation of the learned filters. We show that circular filters improve motif inference performance over a wide range of hyperparameters. Furthermore, we show that CNNs with circular filters perform better at inferring transcription factor binding motifs from ChIP-seq data than conventional [email protected]


Sign in / Sign up

Export Citation Format

Share Document