Robust motion control for UAV in dynamic uncertain environments using deep reinforcement learning

Kaifang Wan, Xiaoguang Gao, Zijian Hu, Gaofeng Wu

Research output: Contribution to journalArticlepeer-review

80 Scopus citations

Abstract

In this paper, a novel deep reinforcement learning (DRL) method, and robust deep deterministic policy gradient (Robust-DDPG), is proposed for developing a controller that allows robust flying of an unmanned aerial vehicle (UAV) in dynamic uncertain environments. This technique is applicable in many fields, such as penetration and remote surveillance. The learning-based controller is constructed with an actor-critic framework, and can perform a dual-channel continuous control (roll and speed) of the UAV. To overcome the fragility and volatility of original DDPG, three critical learning tricks are introduced in Robust-DDPG: (1) Delayed-learning trick, providing stable learnings, while facing dynamic environments; (2) adversarial attack trick, improving policy's adaptability to uncertain environments; (3) mixed exploration trick, enabling faster convergence of the model. The training experiments show great improvement in its convergence speed, convergence effect, and stability. The exploiting experiments demonstrate high efficiency in providing the UAV a shorter and smoother path. While, the generalization experiments verify its better adaptability to complicated, dynamic and uncertain environments, comparing to Deep Q Network (DQN) and DDPG algorithms.

Original languageEnglish
Article number640
JournalRemote Sensing
Volume12
Issue number4
DOIs
StatePublished - 1 Feb 2020

Keywords

  • Adversarial attack
  • Deep reinforcement learning
  • Delayed learning
  • Mixed exploration
  • Robust motion control
  • UAV

Fingerprint

Dive into the research topics of 'Robust motion control for UAV in dynamic uncertain environments using deep reinforcement learning'. Together they form a unique fingerprint.

Cite this