Model-Based Usage Control Policy Derivation

Author(s):  
Prachi Kumari ◽  
Alexander Pretschner
2021 ◽  
Author(s):  
Xinglong Zhang ◽  
Yaoqian Peng ◽  
Biao Luo ◽  
Wei Pan ◽  
Xin Xu ◽  
...  

<div>Recently, barrier function-based safe reinforcement learning (RL) with the actor-critic structure for continuous control tasks has received increasing attention. It is still challenging to learn a near-optimal control policy with safety and convergence guarantees. Also, few works have addressed the safe RL algorithm design under time-varying safety constraints. This paper proposes a model-based safe RL algorithm for optimal control of nonlinear systems with time-varying state and control constraints. In the proposed approach, we construct a novel barrier-based control policy structure that can guarantee control safety. A multi-step policy evaluation mechanism is proposed to predict the policy's safety risk under time-varying safety constraints and guide the policy to update safely. Theoretical results on stability and robustness are proven. Also, the convergence of the actor-critic learning algorithm is analyzed. The performance of the proposed algorithm outperforms several state-of-the-art RL algorithms in the simulated Safety Gym environment. Furthermore, the approach is applied to the integrated path following and collision avoidance problem for two real-world intelligent vehicles. A differential-drive vehicle and an Ackermann-drive one are used to verify the offline deployment performance and the online learning performance, respectively. Our approach shows an impressive sim-to-real transfer capability and a satisfactory online control performance in the experiment.</div>


2020 ◽  
Vol 253 ◽  
pp. 109751 ◽  
Author(s):  
Qianru Zhang ◽  
Peifeng Tong ◽  
Maodian Liu ◽  
Huiming Lin ◽  
Xiao Yun ◽  
...  

Author(s):  
Fabio Martinelli ◽  
Ilaria Matteucci ◽  
Paolo Mori ◽  
Andrea Saracino
Keyword(s):  

Sign in / Sign up

Export Citation Format

Share Document