Motion capture and reinforcement learning of dynamically stable humanoid movement primitives

Author(s):  
Rok Vuga ◽  
Matjaz Ogrinc ◽  
Andrej Gams ◽  
Tadej Petric ◽  
Norikazu Sugimoto ◽  
...  
2019 ◽  
Vol 38 (14) ◽  
pp. 1560-1580 ◽  
Author(s):  
Carlos Celemin ◽  
Guilherme Maeda ◽  
Javier Ruiz-del-Solar ◽  
Jan Peters ◽  
Jens Kober

Robot learning problems are limited by physical constraints, which make learning successful policies for complex motor skills on real systems unfeasible. Some reinforcement learning methods, like Policy Search, offer stable convergence toward locally optimal solutions, whereas interactive machine learning or learning-from-demonstration methods allow fast transfer of human knowledge to the agents. However, most methods require expert demonstrations. In this work, we propose the use of human corrective advice in the actions domain for learning motor trajectories. Additionally, we combine this human feedback with reward functions in a Policy Search learning scheme. The use of both sources of information speeds up the learning process, since the intuitive knowledge of the human teacher can be easily transferred to the agent, while the Policy Search method with the cost/reward function take over for supervising the process and reducing the influence of occasional wrong human corrections. This interactive approach has been validated for learning movement primitives with simulated arms with several degrees of freedom in reaching via-point movements, and also using real robots in such tasks as “writing characters” and the ball-in-a-cup game. Compared with standard reinforcement learning without human advice, the results show that the proposed method not only converges to higher rewards when learning movement primitives, but also that the learning is sped up by a factor of 4–40 times, depending on the task.


2020 ◽  
Vol 5 (4) ◽  
pp. 6678-6685
Author(s):  
Rahul Tallamraju ◽  
Nitin Saini ◽  
Elia Bonetto ◽  
Michael Pabst ◽  
Yu Tang Liu ◽  
...  

2018 ◽  
Vol 23 (1) ◽  
pp. 121-131 ◽  
Author(s):  
Zhijun Li ◽  
Ting Zhao ◽  
Fei Chen ◽  
Yingbai Hu ◽  
Chun-Yi Su ◽  
...  

Sensors ◽  
2021 ◽  
Vol 21 (13) ◽  
pp. 4560
Author(s):  
Chen-Huan Pi ◽  
Yi-Wei Dai ◽  
Kai-Chun Hu ◽  
Stone Cheng

This paper proposes a multipurpose reinforcement learning based low-level multirotor unmanned aerial vehicles control structure constructed using neural networks with model-free training. Other low-level reinforcement learning controllers developed in studies have only been applicable to a model-specific and physical-parameter-specific multirotor, and time-consuming training is required when switching to a different vehicle. We use a 6-degree-of-freedom dynamic model combining acceleration-based control from the policy neural network to overcome these problems. The UAV automatically learns the maneuver by an end-to-end neural network from fusion states to acceleration command. The state estimation is performed using the data from on-board sensors and motion capture. The motion capture system provides spatial position information and a multisensory fusion framework fuses the measurement from the onboard inertia measurement units for compensating the time delay and low update frequency of the capture system. Without requiring expert demonstration, the trained control policy implemented using an improved algorithm can be applied to various multirotors with the output directly mapped to actuators. The algorithm’s ability to control multirotors in the hovering and the tracking task is evaluated. Through simulation and actual experiments, we demonstrate the flight control with a quadrotor and hexrotor by using the trained policy. With the same policy, we verify that we can stabilize the quadrotor and hexrotor in the air under random initial states.


2021 ◽  
Author(s):  
Clinton Mo ◽  
Kun Hu ◽  
Shaohui Mei ◽  
Zebin Chen ◽  
Zhiyong Wang

Sign in / Sign up

Export Citation Format

Share Document