Discrete-Time Optimal Control via Local Policy Iteration Adaptive Dynamic Programming

2017 ◽  
Vol 47 (10) ◽  
pp. 3367-3379 ◽  
Author(s):  
Qinglai Wei ◽  
Derong Liu ◽  
Qiao Lin ◽  
Ruizhuo Song
Complexity ◽  
2021 ◽  
Vol 2021 ◽  
pp. 1-10
Author(s):  
Feng Jiang ◽  
Kai Zhang ◽  
Jinjing Hu ◽  
Shunjiang Wang

Adaptive dynamic programming (ADP), which belongs to the field of computational intelligence, is a powerful tool to address optimal control problems. To overcome the bottleneck of solving Hamilton–Jacobi–Bellman equations, several state-of-the-art ADP approaches are reviewed in this paper. First, two model-based offline iterative ADP methods including policy iteration (PI) and value iteration (VI) are given, and their respective advantages and shortcomings are discussed in detail. Second, the multistep heuristic dynamic programming (HDP) method is introduced, which avoids the requirement of initial admissible control and achieves fast convergence. This method successfully utilizes the advantages of PI and VI and overcomes their drawbacks at the same time. Finally, the discrete-time optimal control strategy is tested on a power system.


Sign in / Sign up

Export Citation Format

Share Document