Hierarchical Attention Neural Network for Event Types to Improve Event Detection

Sensors (Basel). 2022 May 31;22(11):4202. doi: 10.3390/s22114202.

Abstract

Event detection is an important task in the field of natural language processing, which aims to detect trigger words in a sentence and classify them into specific event types. Event detection tasks suffer from data sparsity and event instances imbalance problems in small-scale datasets. For this reason, the correlation information of event types can be used to alleviate the above problems. In this paper, we design a Hierarchical Attention Neural Network for Event Types (HANN-ET). Specifically, we select Long Short-Term Memory (LSTM) as the semantic encoder and utilize dynamic multi-pooling and the Graph Attention Network (GAT) to enrich the sentence feature. Meanwhile, we build several upper-level event type modules and employ a weighted attention aggregation mechanism to integrate these modules to obtain the correlation event type information. Each upper-level module is completed by a Neural Module Network (NMNs), event types within the same upper-level module can share information, and an attention aggregation mechanism can provide effective bias scores for the trigger word classifier. We conduct extensive experiments on the ACE2005 and the MAVEN datasets, and the results show that our approach outperforms previous state-of-the-art methods and achieves the competitive F1 scores of 78.9% on the ACE2005 dataset and 68.8% on the MAVEN dataset.

Keywords: LSTM; attention aggregation mechanism; event detection; neural module network.

MeSH terms

  • Language
  • Memory, Long-Term
  • Natural Language Processing*
  • Neural Networks, Computer*
  • Semantics

Grants and funding

This research was funded by the Shanghai Science Committee, grant number 22N51900200.