Deep reinforcement learning based missile guidance law design for maneuvering target interception

Author(s):  
Mingjian Du ◽  
Chi Peng ◽  
Jianjun Ma
2004 ◽  
Vol 37 (6) ◽  
pp. 599-604 ◽  
Author(s):  
Naoto Dohi ◽  
Yoriaki Baba ◽  
Hiroyuki Takano

Author(s):  
Weifan Li ◽  
Yuanheng Zhu ◽  
Dongbin Zhao

AbstractIn missile guidance, pursuit performance is seriously degraded due to the uncertainty and randomness in target maneuverability, detection delay, and environmental noise. In many methods, accurately estimating the acceleration of the target or the time-to-go is needed to intercept the maneuvering target, which is hard in an environment with uncertainty. In this paper, we propose an assisted deep reinforcement learning (ARL) algorithm to optimize the neural network-based missile guidance controller for head-on interception. Based on the relative velocity, distance, and angle, ARL can control the missile to intercept the maneuvering target and achieve large terminal intercept angle. To reduce the influence of environmental uncertainty, ARL predicts the target’s acceleration as an auxiliary supervised task. The supervised learning task improves the ability of the agent to extract information from observations. To exploit the agent’s good trajectories, ARL presents the Gaussian self-imitation learning to make the mean of action distribution approach the agent’s good actions. Compared with vanilla self-imitation learning, Gaussian self-imitation learning improves the exploration in continuous control. Simulation results validate that ARL outperforms traditional methods and proximal policy optimization algorithm with higher hit rate and larger terminal intercept angle in the simulation environment with noise, delay, and maneuverable target.


2020 ◽  
Vol 10 (18) ◽  
pp. 6567
Author(s):  
Daseon Hong ◽  
Minjeong Kim ◽  
Sungsu Park

Reinforcement learning is generating considerable interest in terms of building guidance law and solving optimization problems that were previously difficult to solve. Since reinforcement learning-based guidance laws often show better robustness than a previously optimized algorithm, several studies have been carried out on the subject. This paper presents a new approach to training missile guidance law by reinforcement learning and introducing some notable characteristics. The novel missile guidance law shows better robustness to the controller-model compared to the proportional navigation guidance. The neural network in this paper has identical inputs with proportional navigation guidance, which makes the comparison fair, distinguishing it from other research. The proposed guidance law will be compared to the proportional navigation guidance, which is widely known as quasi-optimal of missile guidance law. Our work aims to find effective missile training methods through reinforcement learning, and how better the new method is. Additionally, with the derived policy, we contemplated which is better, and in which circumstances it is better. A novel methodology for the training will be proposed first, and the performance comparison results will be continued therefrom.


Author(s):  
Jun-Yong Lee ◽  
Hyeong-Guen Kim ◽  
H Jin Kim

This article proposes an impact-time-control guidance law that can keep a non-maneuvering moving target in the seeker’s field of view (FOV). For a moving target, the missile calculates a predicted intercept point (PIP), designates the PIP as a new virtual stationary target, and flies to the PIP at the desired impact time. The main contribution of the article is that the guidance law is designed to always lock onto the moving target by adjusting the guidance gain. The guidance law for the purpose is based on the backstepping control technique and designed to regulate the defined impact time error. In this procedure, the desired look angle, which is a virtual control, is designed not to violate the FOV limit, and the actual look angle of the missile is kept within the FOV by tracking the desired look angle. To validate the performance of the guidance law, numerical simulation is conducted with different impact times. The result shows that the proposed guidance law intercepts the moving target at the desired impact time maintaining the target lock-on condition.


Author(s):  
Feng Tyan ◽  
Jeng Fu Shen

Sign in / Sign up

Export Citation Format

Share Document