scholarly journals FIXAR: A Fixed-Point Deep Reinforcement Learning Platform with Quantization-Aware Training and Adaptive Parallelism

Author(s):  
Je Yang ◽  
Seongmin Hong ◽  
Joo-Young Kim
2021 ◽  
Author(s):  
Jean Paul Sebastian Piest ◽  
Maria-Eugenia Iacob ◽  
Marten van Sinderen ◽  
Martijn Gemmink ◽  
Bjorn Goossens

2020 ◽  
Vol 15 (2) ◽  
pp. 197-204
Author(s):  
Taewon Kim ◽  
◽  
Yeseong Park ◽  
Jong Bok Kim ◽  
Youngbin Park ◽  
...  

Author(s):  
Matteo Hessel ◽  
Hubert Soyer ◽  
Lasse Espeholt ◽  
Wojciech Czarnecki ◽  
Simon Schmitt ◽  
...  

The reinforcement learning (RL) community has made great strides in designing algorithms capable of exceeding human performance on specific tasks. These algorithms are mostly trained one task at the time, each new task requiring to train a brand new agent instance. This means the learning algorithm is general, but each solution is not; each agent can only solve the one task it was trained on. In this work, we study the problem of learning to master not one but multiple sequentialdecision tasks at once. A general issue in multi-task learning is that a balance must be found between the needs of multiple tasks competing for the limited resources of a single learning system. Many learning algorithms can get distracted by certain tasks in the set of tasks to solve. Such tasks appear more salient to the learning process, for instance because of the density or magnitude of the in-task rewards. This causes the algorithm to focus on those salient tasks at the expense of generality. We propose to automatically adapt the contribution of each task to the agent’s updates, so that all tasks have a similar impact on the learning dynamics. This resulted in state of the art performance on learning to play all games in a set of 57 diverse Atari games. Excitingly, our method learned a single trained policy - with a single set of weights - that exceeds median human performance. To our knowledge, this was the first time a single agent surpassed human-level performance on this multi-task domain. The same approach also demonstrated state of the art performance on a set of 30 tasks in the 3D reinforcement learning platform DeepMind Lab.


2021 ◽  
Vol 2113 (1) ◽  
pp. 012086
Author(s):  
Zhongxuan Cai ◽  
Zhen Liang ◽  
Jing Ren

Abstract Deep reinforcement learning (DRL) has greatly improved the intelligence of AI in recent years and the community has proposed several common software to facilitate the development of DRL. However, in robotics the utility of common DRL software is limited and the development is time-consuming due to the complexity of various robot software. In this paper, we propose a software engineering approach leveraging modularity to facilitate robot DRL development. The platform decouples learning environment into task, simulator and hierarchical robot modules, which in turn enables diverse environment generation using existing modules as building blocks, regardless of the underlying robot software details. Experimental results show that our platform provides composable environment building, introduces high module reuse and efficiently facilitates robot DRL.


Decision ◽  
2016 ◽  
Vol 3 (2) ◽  
pp. 115-131 ◽  
Author(s):  
Helen Steingroever ◽  
Ruud Wetzels ◽  
Eric-Jan Wagenmakers

Sign in / Sign up

Export Citation Format

Share Document