Locomotion Control Framework for Snake-like Robot using Deep Reinforcement Learning

Author(s):  
Obe Olumide O ◽  
Ayogu Thomas O
2020 ◽  
Vol 35 (6) ◽  
pp. 4644-4654 ◽  
Author(s):  
Shengyi Wang ◽  
Jiajun Duan ◽  
Di Shi ◽  
Chunlei Xu ◽  
Haifeng Li ◽  
...  

2019 ◽  
pp. 027836491985944 ◽  
Author(s):  
David Surovik ◽  
Kun Wang ◽  
Massimo Vespignani ◽  
Jonathan Bruce ◽  
Kostas E Bekris

Tensegrity robots, which are prototypical examples of hybrid soft–rigid robots, exhibit dynamical properties that provide ruggedness and adaptability. They also bring about, however, major challenges for locomotion control. Owing to high dimensionality and the complex evolution of contact states, data-driven approaches are appropriate for producing viable feedback policies for tensegrities. Guided policy search (GPS), a sample-efficient hybrid framework for optimization and reinforcement learning, has previously been applied to generate periodic, axis-constrained locomotion by an icosahedral tensegrity on flat ground. Varying environments and tasks, however, create a need for more adaptive and general locomotion control that actively utilizes an expanded space of robot states. This implies significantly higher needs in terms of sample data and setup effort. This work mitigates such requirements by proposing a new GPS -based reinforcement learning pipeline, which exploits the vehicle’s high degree of symmetry and appropriately learns contextual behaviors that are sustainable without periodicity. Newly achieved capabilities include axially unconstrained rolling, rough terrain traversal, and rough incline ascent. These tasks are evaluated for a small variety of key model parameters in simulation and tested on the NASA hardware prototype, SUPERball. Results confirm the utility of symmetry exploitation and the adaptability of the vehicle. They also shed light on numerous strengths and limitations of the GPS framework for policy design and transfer to real hybrid soft–rigid robots.


Author(s):  
Francisco Hugo Costa Neto ◽  
Daniel Costa Araujo ◽  
Mateus Pontes Mota ◽  
Tarcisio Maciel ◽  
Andre L. F. De Almeida

2020 ◽  
Vol 5 (47) ◽  
pp. eabc5986 ◽  
Author(s):  
Joonho Lee ◽  
Jemin Hwangbo ◽  
Lorenz Wellhausen ◽  
Vladlen Koltun ◽  
Marco Hutter

Legged locomotion can extend the operational domain of robots to some of the most challenging environments on Earth. However, conventional controllers for legged locomotion are based on elaborate state machines that explicitly trigger the execution of motion primitives and reflexes. These designs have increased in complexity but fallen short of the generality and robustness of animal locomotion. Here, we present a robust controller for blind quadrupedal locomotion in challenging natural environments. Our approach incorporates proprioceptive feedback in locomotion control and demonstrates zero-shot generalization from simulation to natural environments. The controller is trained by reinforcement learning in simulation. The controller is driven by a neural network policy that acts on a stream of proprioceptive signals. The controller retains its robustness under conditions that were never encountered during training: deformable terrains such as mud and snow, dynamic footholds such as rubble, and overground impediments such as thick vegetation and gushing water. The presented work indicates that robust locomotion in natural environments can be achieved by training in simple domains.


Sign in / Sign up

Export Citation Format

Share Document