SARSA-based reinforcement learning for motion planning in serial manipulators

Author(s):  
Ignazio Aleo ◽  
Paolo Arena ◽  
Luca Patane
2021 ◽  
Author(s):  
Qiang Li ◽  
Jun Nie ◽  
Haixia Wang ◽  
Xiao Lu ◽  
Shibin Song

2014 ◽  
Vol 7 ◽  
Author(s):  
Mikhail Frank ◽  
Jürgen Leitner ◽  
Marijn Stollenga ◽  
Alexander Förster ◽  
Jürgen Schmidhuber

2021 ◽  
pp. 318-329
Author(s):  
Nikodem Pankiewicz ◽  
Tomasz Wrona ◽  
Wojciech Turlej ◽  
Mateusz Orłowski

Sensors ◽  
2020 ◽  
Vol 20 (7) ◽  
pp. 1890 ◽  
Author(s):  
Zijian Hu ◽  
Kaifang Wan ◽  
Xiaoguang Gao ◽  
Yiwei Zhai ◽  
Qianglong Wang

Autonomous motion planning (AMP) of unmanned aerial vehicles (UAVs) is aimed at enabling a UAV to safely fly to the target without human intervention. Recently, several emerging deep reinforcement learning (DRL) methods have been employed to address the AMP problem in some simplified environments, and these methods have yielded good results. This paper proposes a multiple experience pools (MEPs) framework leveraging human expert experiences for DRL to speed up the learning process. Based on the deep deterministic policy gradient (DDPG) algorithm, a MEP–DDPG algorithm was designed using model predictive control and simulated annealing to generate expert experiences. On applying this algorithm to a complex unknown simulation environment constructed based on the parameters of the real UAV, the training experiment results showed that the novel DRL algorithm resulted in a performance improvement exceeding 20% as compared with the state-of-the-art DDPG. The results of the experimental testing indicate that UAVs trained using MEP–DDPG can stably complete a variety of tasks in complex, unknown environments.


Procedia CIRP ◽  
2017 ◽  
Vol 63 ◽  
pp. 107-112 ◽  
Author(s):  
Richard Meyes ◽  
Hasan Tercan ◽  
Simon Roggendorf ◽  
Thomas Thiele ◽  
Christian Büscher ◽  
...  

Robotica ◽  
2011 ◽  
Vol 30 (2) ◽  
pp. 159-170 ◽  
Author(s):  
M. Gómez ◽  
R. V. González ◽  
T. Martínez-Marín ◽  
D. Meziat ◽  
S. Sánchez

SUMMARYThe aim of this work has been the implementation and testing in real conditions of a new algorithm based on the cell-mapping techniques and reinforcement learning methods to obtain the optimal motion planning of a vehicle considering kinematics, dynamics and obstacle constraints. The algorithm is an extension of the control adjoining cell mapping technique for learning the dynamics of the vehicle instead of using its analytical state equations. It uses a transformation of cell-to-cell mapping in order to reduce the time spent during the learning stage. Real experimental results are reported to show the satisfactory performance of the algorithm.


Sign in / Sign up

Export Citation Format

Share Document