scholarly journals Reinforcement Learning from Imperfect Demonstrations under Soft Expert Guidance

2020 ◽  
Vol 34 (04) ◽  
pp. 5109-5116
Author(s):  
Mingxuan Jing ◽  
Xiaojian Ma ◽  
Wenbing Huang ◽  
Fuchun Sun ◽  
Chao Yang ◽  
...  

In this paper, we study Reinforcement Learning from Demonstrations (RLfD) that improves the exploration efficiency of Reinforcement Learning (RL) by providing expert demonstrations. Most of existing RLfD methods require demonstrations to be perfect and sufficient, which yet is unrealistic to meet in practice. To work on imperfect demonstrations, we first define an imperfect expert setting for RLfD in a formal way, and then point out that previous methods suffer from two issues in terms of optimality and convergence, respectively. Upon the theoretical findings we have derived, we tackle these two issues by regarding the expert guidance as a soft constraint on regulating the policy exploration of the agent, which eventually leads to a constrained optimization problem. We further demonstrate that such problem is able to be addressed efficiently by performing a local linear search on its dual form. Considerable empirical evaluations on a comprehensive collection of benchmarks indicate our method attains consistent improvement over other RLfD counterparts.

Author(s):  
Gabriele Eichfelder ◽  
Kathrin Klamroth ◽  
Julia Niebling

AbstractA major difficulty in optimization with nonconvex constraints is to find feasible solutions. As simple examples show, the $$\alpha $$ α BB-algorithm for single-objective optimization may fail to compute feasible solutions even though this algorithm is a popular method in global optimization. In this work, we introduce a filtering approach motivated by a multiobjective reformulation of the constrained optimization problem. Moreover, the multiobjective reformulation enables to identify the trade-off between constraint satisfaction and objective value which is also reflected in the quality guarantee. Numerical tests validate that we indeed can find feasible and often optimal solutions where the classical single-objective $$\alpha $$ α BB method fails, i.e., it terminates without ever finding a feasible solution.


Energies ◽  
2018 ◽  
Vol 11 (8) ◽  
pp. 2107 ◽  
Author(s):  
Min-Rong Chen ◽  
Huan Wang ◽  
Guo-Qiang Zeng ◽  
Yu-Xing Dai ◽  
Da-Qiang Bi

The optimal P-Q control issue of the active and reactive power for a microgrid in the grid-connected mode has attracted increasing interests recently. In this paper, an optimal active and reactive power control is developed for a three-phase grid-connected inverter in a microgrid by using an adaptive population-based extremal optimization algorithm (APEO). Firstly, the optimal P-Q control issue of grid-connected inverters in a microgrid is formulated as a constrained optimization problem, where six parameters of three decoupled PI controllers are real-coded as the decision variables, and the integral time absolute error (ITAE) between the output and referenced active power and the ITAE between the output and referenced reactive power are weighted as the objective function. Then, an effective and efficient APEO algorithm with an adaptive mutation operation is proposed for solving this constrained optimization problem. The simulation and experiments for a 3kW three-phase grid-connected inverter under both nominal and variable reference active power values have shown that the proposed APEO-based P-Q control method outperforms the traditional Z-N empirical method, the adaptive genetic algorithm-based, and particle swarm optimization-based P-Q control methods.


Sign in / Sign up

Export Citation Format

Share Document