A new method for exponential synchronization of memristive recurrent neural networks

2018 ◽  
Vol 466 ◽  
pp. 152-169 ◽  
Author(s):  
Ruimei Zhang ◽  
Ju H. Park ◽  
Deqiang Zeng ◽  
Yajuan Liu ◽  
Shouming Zhong
2011 ◽  
Vol 74 (17) ◽  
pp. 3043-3050 ◽  
Author(s):  
Ailong Wu ◽  
Zhigang Zeng ◽  
Xusheng Zhu ◽  
Jine Zhang

10.29007/j5hd ◽  
2020 ◽  
Author(s):  
Bartosz Piotrowski ◽  
Josef Urban

In this work we develop a new learning-based method for selecting facts (premises) when proving new goals over large formal libraries. Unlike previous methods that choose sets of facts independently of each other by their rank, the new method uses the notion of state that is updated each time a choice of a fact is made. Our stateful architecture is based on recurrent neural networks which have been recently very successful in stateful tasks such as language translation. The new method is combined with data augmentation techniques, evaluated in several ways on a standard large-theory benchmark and compared to state-of-the-art premise approach based on gradient boosted trees. It is shown to perform significantly better and to solve many new problems.


Author(s):  
Shin-ichiro Hashimukai ◽  
Chikahiro Araki ◽  
Mikio Mori ◽  
Shuji Taniguchi ◽  
Shozo Kato ◽  
...  

Sign in / Sign up

Export Citation Format

Share Document