Reinforcement Learning-Based 3-D Sliding Mode Interception Guidance via Proximal Policy Optimization

Jianguo Guo, Mengxuan Li, Zongyi Guo, Zhiyong She

Research output: Contribution to journalArticlepeer-review

3 Scopus citations

Abstract

This article proposes a novel 3-D sliding mode interception guidance law for maneuvering targets, which explores the potential of reinforcement learning (RL) techniques to enhance guidance accuracy and reduce chattering. The guidance problem of intercepting maneuvering targets is abstracted into a Markov decision process whose reward function is established to estimate the off-target amount and line-of-sight angular rate chattering. Importantly, a design framework of reward function suitable for general guidance problems based on RL can be proposed. Then, the proximal policy optimization algorithm with a satisfactory training performance is introduced to learn an action policy which represents the observed engagements states to sliding mode interception guidance. Finally, numerical simulations and comparisons are conducted to demonstrate the effectiveness of the proposed guidance law.

Original languageEnglish
Pages (from-to)423-430
Number of pages8
JournalIEEE Journal on Miniaturization for Air and Space Systems
Volume4
Issue number4
DOIs
StatePublished - 1 Dec 2023

Keywords

  • 3-D
  • guidance law
  • proximal policy optimization (PPO)
  • reinforcement learning (RL)
  • sliding mode control

Fingerprint

Dive into the research topics of 'Reinforcement Learning-Based 3-D Sliding Mode Interception Guidance via Proximal Policy Optimization'. Together they form a unique fingerprint.

Cite this