Recurrent Neural Networks With External Addressable Long-Term and Working Memory for Learning Long-Term Dependences

IEEE Trans Neural Netw Learn Syst. 2020 Mar;31(3):813-826. doi: 10.1109/TNNLS.2019.2910302. Epub 2019 May 3.

Abstract

Learning long-term dependences (LTDs) with recurrent neural networks (RNNs) is challenging due to their limited internal memories. In this paper, we propose a new external memory architecture for RNNs called an external addressable long-term and working memory (EALWM)-augmented RNN. This architecture has two distinct advantages over existing neural external memory architectures, namely the division of the external memory into two parts-long-term memory and working memory-with both addressable and the capability to learn LTDs without suffering from vanishing gradients with necessary assumptions. The experimental results on algorithm learning, language modeling, and question answering demonstrate that the proposed neural memory architecture is promising for practical applications.

Publication types

  • Research Support, Non-U.S. Gov't

MeSH terms

  • Databases, Factual*
  • Memory, Long-Term* / physiology
  • Memory, Short-Term* / physiology
  • Neural Networks, Computer*