scholarly journals Robust Feedback Motion Policy Design Using Reinforcement Learning on a 3D Digit Bipedal Robot

Author(s):  
Guillermo A. Castillo ◽  
Bowen Weng ◽  
Wei Zhang ◽  
Ayonga Hereid
2019 ◽  
pp. 027836491985944 ◽  
Author(s):  
David Surovik ◽  
Kun Wang ◽  
Massimo Vespignani ◽  
Jonathan Bruce ◽  
Kostas E Bekris

Tensegrity robots, which are prototypical examples of hybrid soft–rigid robots, exhibit dynamical properties that provide ruggedness and adaptability. They also bring about, however, major challenges for locomotion control. Owing to high dimensionality and the complex evolution of contact states, data-driven approaches are appropriate for producing viable feedback policies for tensegrities. Guided policy search (GPS), a sample-efficient hybrid framework for optimization and reinforcement learning, has previously been applied to generate periodic, axis-constrained locomotion by an icosahedral tensegrity on flat ground. Varying environments and tasks, however, create a need for more adaptive and general locomotion control that actively utilizes an expanded space of robot states. This implies significantly higher needs in terms of sample data and setup effort. This work mitigates such requirements by proposing a new GPS -based reinforcement learning pipeline, which exploits the vehicle’s high degree of symmetry and appropriately learns contextual behaviors that are sustainable without periodicity. Newly achieved capabilities include axially unconstrained rolling, rough terrain traversal, and rough incline ascent. These tasks are evaluated for a small variety of key model parameters in simulation and tested on the NASA hardware prototype, SUPERball. Results confirm the utility of symmetry exploitation and the adaptability of the vehicle. They also shed light on numerous strengths and limitations of the GPS framework for policy design and transfer to real hybrid soft–rigid robots.


2021 ◽  
pp. 103891
Author(s):  
Christos Kouppas ◽  
Mohamad Saada ◽  
Qinggang Meng ◽  
Mark King ◽  
Dennis Majoe

Author(s):  
Ziyao Zhang ◽  
Liang Ma ◽  
Konstantinos Poularakis ◽  
Kin K. Leung ◽  
Jeremy Tucker ◽  
...  

2019 ◽  
Author(s):  
Christos Kouppas ◽  
Qinggang Meng ◽  
Mark King ◽  
Dennis Majoe

2021 ◽  
Author(s):  
Stephan Zheng ◽  
Alexander Trott ◽  
Sunil Srinivasa ◽  
David C. Parkes ◽  
Richard Socher

Robotica ◽  
2004 ◽  
Vol 22 (1) ◽  
pp. 29-39 ◽  
Author(s):  
Chee-Meng Chew ◽  
Gill A. Pratt

This paper presents two frontal plane algorithms for 3D dynamic bipedal walking. One of which is based on the notion of symmetry and the other uses reinforcement learning algorithm to learn the lateral foot placement. The algorithms are combined with a sagittal plane algorithm and successfully applied to a simulated 3D bipedal robot to achieve level ground walking. The simulation results showed that the choice of the local control law for the stance-ankle roll joint could significantly affect the performance of the frontal plane algorithms.


Sign in / Sign up

Export Citation Format

Share Document