scholarly journals Autonomous helicopter control using reinforcement learning policy search methods

Author(s):  
J.A. Bagnell ◽  
J.G. Schneider
2021 ◽  
Author(s):  
N. Snehal ◽  
W. Pooja ◽  
K. Sonam ◽  
S. R. Wagh ◽  
N. M. Singh

2019 ◽  
Vol 92 (2) ◽  
pp. 172-179
Author(s):  
Firat Sal

Purpose The purpose of this study is to examine the effect of passive and active morphing of blade root chord length and blade taper on the control effort of the flight control system (FCS) of a helicopter. Design/methodology/approach Physics-based helicopter models, which are functions of passive and active morphing, are created and applied in helicopter FCS design to determine the control effort. Findings Helicopters, having both passively and actively morphing blade root chord length and blade taper, experience less control effort than the ones having either only passively morphing blade root chord length or only blade taper or only actively morphing blade root chord length and blade taper. Practical implications Both passively and actively morphing blade root chord length and blade taper can be implemented for more economical autonomous helicopter flights. Originality/value Main novelty of our article is simultaneous application of passive and active morphing ideas on helicopter root chord length and blade taper. It is also proved in this study that using both passive and active morphing ideas on helicopter blade root chord and blade taper causes much less energy consumption than using either only passive morphing idea on helicopter blade root chord and blade taper or only active morphing idea on helicopter blade root chord and blade taper. This also reduces fuel consumption and also makes environment cleaner.


Author(s):  
Abhinav Verma

We study the problem of generating interpretable and verifiable policies for Reinforcement Learning (RL). Unlike the popular Deep Reinforcement Learning (DRL) paradigm, in which the policy is represented by a neural network, the aim of this work is to find policies that can be represented in highlevel programming languages. Such programmatic policies have several benefits, including being more easily interpreted than neural networks, and being amenable to verification by scalable symbolic methods. The generation methods for programmatic policies also provide a mechanism for systematically using domain knowledge for guiding the policy search. The interpretability and verifiability of these policies provides the opportunity to deploy RL based solutions in safety critical environments. This thesis draws on, and extends, work from both the machine learning and formal methods communities.


2011 ◽  
Vol 23 (11) ◽  
pp. 2798-2832 ◽  
Author(s):  
Hirotaka Hachiya ◽  
Jan Peters ◽  
Masashi Sugiyama

Direct policy search is a promising reinforcement learning framework, in particular for controlling continuous, high-dimensional systems. Policy search often requires a large number of samples for obtaining a stable policy update estimator, and this is prohibitive when the sampling cost is expensive. In this letter, we extend an expectation-maximization-based policy search method so that previously collected samples can be efficiently reused. The usefulness of the proposed method, reward-weighted regression with sample reuse (R[Formula: see text]), is demonstrated through robot learning experiments. (This letter is an extended version of our earlier conference paper: Hachiya, Peters, & Sugiyama, 2009 .)


Sign in / Sign up

Export Citation Format

Share Document