Comparison between model-based and non-isothermal model-free computational procedures for prediction of conversion-time curves of calcium carbonate decomposition

2019 ◽  
Vol 679 ◽  
pp. 178322 ◽  
Author(s):  
P. Budrugeac
2020 ◽  
Vol 43 ◽  
Author(s):  
Peter Dayan

Abstract Bayesian decision theory provides a simple formal elucidation of some of the ways that representation and representational abstraction are involved with, and exploit, both prediction and its rather distant cousin, predictive coding. Both model-free and model-based methods are involved.


2019 ◽  
Author(s):  
Leor M Hackel ◽  
Jeffrey Jordan Berg ◽  
Björn Lindström ◽  
David Amodio

Do habits play a role in our social impressions? To investigate the contribution of habits to the formation of social attitudes, we examined the roles of model-free and model-based reinforcement learning in social interactions—computations linked in past work to habit and planning, respectively. Participants in this study learned about novel individuals in a sequential reinforcement learning paradigm, choosing financial advisors who led them to high- or low-paying stocks. Results indicated that participants relied on both model-based and model-free learning, such that each independently predicted choice during the learning task and self-reported liking in a post-task assessment. Specifically, participants liked advisors who could provide large future rewards as well as advisors who had provided them with large rewards in the past. Moreover, participants varied in their use of model-based and model-free learning strategies, and this individual difference influenced the way in which learning related to self-reported attitudes: among participants who relied more on model-free learning, model-free social learning related more to post-task attitudes. We discuss implications for attitudes, trait impressions, and social behavior, as well as the role of habits in a memory systems model of social cognition.


2021 ◽  
Vol 11 (1) ◽  
Author(s):  
Lieneke K. Janssen ◽  
Florian P. Mahner ◽  
Florian Schlagenhauf ◽  
Lorenz Deserno ◽  
Annette Horstmann

An amendment to this paper has been published and can be accessed via a link at the top of the paper.


Author(s):  
Javier Loranca ◽  
Jonathan Carlos Mayo Maldonado ◽  
Gerardo Escobar ◽  
Carlos Villarreal-Hernandez ◽  
Thabiso Maupong ◽  
...  

2021 ◽  
Vol 54 (5) ◽  
pp. 19-24
Author(s):  
Tyler Westenbroek ◽  
Ayush Agrawal ◽  
Fernando Castañeda ◽  
S Shankar Sastry ◽  
Koushil Sreenath

2022 ◽  
pp. 1-12
Author(s):  
Shuailong Li ◽  
Wei Zhang ◽  
Huiwen Zhang ◽  
Xin Zhang ◽  
Yuquan Leng

Model-free reinforcement learning methods have successfully been applied to practical applications such as decision-making problems in Atari games. However, these methods have inherent shortcomings, such as a high variance and low sample efficiency. To improve the policy performance and sample efficiency of model-free reinforcement learning, we propose proximal policy optimization with model-based methods (PPOMM), a fusion method of both model-based and model-free reinforcement learning. PPOMM not only considers the information of past experience but also the prediction information of the future state. PPOMM adds the information of the next state to the objective function of the proximal policy optimization (PPO) algorithm through a model-based method. This method uses two components to optimize the policy: the error of PPO and the error of model-based reinforcement learning. We use the latter to optimize a latent transition model and predict the information of the next state. For most games, this method outperforms the state-of-the-art PPO algorithm when we evaluate across 49 Atari games in the Arcade Learning Environment (ALE). The experimental results show that PPOMM performs better or the same as the original algorithm in 33 games.


Sign in / Sign up

Export Citation Format

Share Document