Discrete-Time Optimal Control of Nonlinear Systems via Value Iteration-Based $$ Q $$ -Learning

Author(s):  
Qinglai Wei ◽  
Ruizhuo Song ◽  
Benkai Li ◽  
Xiaofeng Lin
Complexity ◽  
2021 ◽  
Vol 2021 ◽  
pp. 1-10
Author(s):  
Feng Jiang ◽  
Kai Zhang ◽  
Jinjing Hu ◽  
Shunjiang Wang

Adaptive dynamic programming (ADP), which belongs to the field of computational intelligence, is a powerful tool to address optimal control problems. To overcome the bottleneck of solving Hamilton–Jacobi–Bellman equations, several state-of-the-art ADP approaches are reviewed in this paper. First, two model-based offline iterative ADP methods including policy iteration (PI) and value iteration (VI) are given, and their respective advantages and shortcomings are discussed in detail. Second, the multistep heuristic dynamic programming (HDP) method is introduced, which avoids the requirement of initial admissible control and achieves fast convergence. This method successfully utilizes the advantages of PI and VI and overcomes their drawbacks at the same time. Finally, the discrete-time optimal control strategy is tested on a power system.


1982 ◽  
Vol 104 (2) ◽  
pp. 151-157 ◽  
Author(s):  
M. J. Grimble ◽  
J. Fotakis

The deterministic discrete-time optimal control problem for a finite optimization interval is considered. A solution is obtained in the z-domain by embedding the problem within a equivalent infinite time problem. The optimal controller is time-invariant and may be easily implemented. The controller is related to the solution of the usual infinite time optimal control problem due to Wiener. This new controller should be of value in self-tuning control laws where a finite interval controller is particularly important.


Sign in / Sign up

Export Citation Format

Share Document