Enhancing Chinese Character Representation With Lattice-Aligned Attention

IEEE Trans Neural Netw Learn Syst. 2023 Jul;34(7):3727-3736. doi: 10.1109/TNNLS.2021.3114378. Epub 2023 Jul 6.

Abstract

Word-character lattice models have been proved to be effective for some Chinese natural language processing (NLP) tasks, in which word boundary information is fused into character sequences. However, due to the inherently unidirectional sequential nature, prior approaches have only learned sequential interactions of character-word instances but fail to capture fine-grained correlations in word-character spaces. In this article, we propose a lattice-aligned attention network (LAN) that aims to model dense interactions over word-character lattice structure for enhancing character representations. By carefully combining cross-lattice module, gated word-character semantic fusion unit, and self-lattice attention module, the network can explicitly capture fine-grained correlations across different spaces (e.g., word-to-character and character-to-character), thus significantly improving model performance. Experimental results on three Chinese NLP benchmark tasks demonstrate that LAN obtains state-of-the-art results compared to several competitive approaches.

MeSH terms

  • Language*
  • Natural Language Processing
  • Neural Networks, Computer*
  • Semantics