End-to-End Autonomous Exploration with Deep Reinforcement Learning and Intrinsic Motivation

Comput Intell Neurosci. 2021 Dec 16:2021:9945044. doi: 10.1155/2021/9945044. eCollection 2021.

Abstract

Developing artificial intelligence (AI) agents is challenging for efficient exploration in visually rich and complex environments. In this study, we formulate the exploration question as a reinforcement learning problem and rely on intrinsic motivation to guide exploration behavior. Such intrinsic motivation is driven by curiosity and is calculated based on episode memory. To distribute the intrinsic motivation, we use a count-based method and temporal distance to generate it synchronously. We tested our approach in 3D maze-like environments and validated its performance in exploration tasks through extensive experiments. The experimental results show that our agent can learn exploration ability from raw sensory input and accomplish autonomous exploration across different mazes. In addition, the learned policy is not biased by stochastic objects. We also analyze the effects of different training methods and driving forces on exploration policy.

MeSH terms

  • Artificial Intelligence*
  • Exploratory Behavior
  • Learning
  • Motivation*
  • Reinforcement, Psychology