Model-Free Dual Heuristic Dynamic Programming

2015 ◽  
Vol 26 (8) ◽  
pp. 1834-1839 ◽  
Author(s):  
Zhen Ni ◽  
Haibo He ◽  
Xiangnan Zhong ◽  
Danil V. Prokhorov
Author(s):  
Hossein Nejatbakhsh Esfahani ◽  
Rafal Szlapczynski

AbstractThis paper proposes a hybrid robust-adaptive learning-based control scheme based on Approximate Dynamic Programming (ADP) for the tracking control of autonomous ship maneuvering. We adopt a Time-Delay Control (TDC) approach, which is known as a simple, practical, model free and roughly robust strategy, combined with an Actor-Critic Approximate Dynamic Programming (ACADP) algorithm as an adaptive part in the proposed hybrid control algorithm. Based on this integration, Actor-Critic Time-Delay Control (AC-TDC) is proposed. It offers a high-performance robust-adaptive control approach for path following of autonomous ships under deterministic and stochastic disturbances induced by the winds, waves, and ocean currents. Computer simulations have been conducted under two different conditions in terms of the deterministic and stochastic disturbances and all simulation results indicate an acceptable performance in tracking of paths for the proposed control algorithm in comparison with the conventional TDC approach.


2016 ◽  
Vol 817 ◽  
pp. 150-161 ◽  
Author(s):  
Marcin Szuster ◽  
Piotr Gierlak

The article focuses on the implementation of the globalized dual-heuristic dynamic programming algorithm in the discrete tracking control system of the three degrees of freedom robotic manipulator. The globalized dual-heuristic dynamic programming algorithm is included in the approximate dynamic programming algorithms family, that bases on the Bellman’s dynamic programming idea. These algorithms generally consist of the actor and the critic structures realized in a form of artificial neural networks. Moreover, the control system includes the PD controller, the supervisory term and an additional control signal. The structure of the supervisory term derives from the stability analysis, which was realized using the Lyapunov stability theorem. The control system works on-line and the neural networks’ weight adaptation process is realized in every iteration step. A series of computer simulations was realized in Matlab/Simulink software to confirm performance of the control system.


2021 ◽  
Author(s):  
Yunfan Su

Vehicular ad hoc network (VANET) is a promising technique that improves traffic safety and transportation efficiency and provides a comfortable driving experience. However, due to the rapid growth of applications that demand channel resources, efficient channel allocation schemes are required to utilize the performance of the vehicular networks. In this thesis, two Reinforcement learning (RL)-based channel allocation methods are proposed for a cognitive enabled VANET environment to maximize a long-term average system reward. First, we present a model-based dynamic programming method, which requires the calculations of the transition probabilities and time intervals between decision epochs. After obtaining the transition probabilities and time intervals, a relative value iteration (RVI) algorithm is used to find the asymptotically optimal policy. Then, we propose a model-free reinforcement learning method, in which we employ an agent to interact with the environment iteratively and learn from the feedback to approximate the optimal policy. Simulation results show that our reinforcement learning method can acquire a similar performance to that of the dynamic programming while both outperform the greedy method.


Sign in / Sign up

Export Citation Format

Share Document