scholarly journals Timing and Parameter Optimization for One-time Motion Problem Based on Reinforcement Learning

2020 ◽  
Vol 5 (1) ◽  
pp. 10
Author(s):  
Boxuan Fan ◽  
Guiming Chen ◽  
Hongtao Lin
2021 ◽  
Vol 2021 ◽  
pp. 1-9
Author(s):  
Yi Zou ◽  
Jijuan Zhong ◽  
Zhihao Jiang ◽  
Hong Zhang ◽  
Xuyu Pu

Agents face challenges to achieve adaptability and stability when interacting with dynamic counterparts in a complex multiagent system (MAS). To strike a balance between these two goals, this paper proposes a learning algorithm for heterogeneous agents with bounded rationality. It integrates reinforcement learning as well as fictitious play to evaluate the historical information and adopt mechanisms in evolutionary game to adapt to uncertainty, which is referred to as experience weighted learning (EWL) in this paper. We have conducted multiagent simulations to test the performance of EWL in various games. The results demonstrate that the average payoff of EWL exceeds that of the baseline in all 4 games. In addition, we find that most of the EWL agents converge to pure strategy and become stable finally. Furthermore, we test the impact of 2 import parameters, respectively. The results show that the performance of EWL is quite stable and there is a potential to improve its performance by parameter optimization.


2020 ◽  
Vol 17 (2) ◽  
pp. 172988142091149
Author(s):  
Kai Zhang ◽  
Sterling McLeod ◽  
Minwoo Lee ◽  
Jing Xiao

This article introduces a continuous reinforcement learning framework to enable online adaptation of multi-objective optimization functions for guiding a mobile robot to move in changing dynamic environments. The robot with this framework can continuously learn from multiple or changing environments where it encounters different numbers of obstacles moving in unknown ways at different times. Using both planned trajectories from a real-time motion planner and already executed trajectories as feedback observations, our reinforcement learning agent enables the robot to adapt motion behaviors to environmental changes. The agent contains a Q network connected to a long short-term memory network. The proposed framework is tested in both simulations and real robot experiments over various, dynamically varied task environments. The results show the efficacy of online continuous reinforcement learning for quick adaption to different, unknown, and dynamic environments.


Sign in / Sign up

Export Citation Format

Share Document