Fast Haar Transforms for Graph Neural Networks

Neural Netw. 2020 Aug:128:188-198. doi: 10.1016/j.neunet.2020.04.028. Epub 2020 May 4.

Abstract

Graph Neural Networks (GNNs) have become a topic of intense research recently due to their powerful capability in high-dimensional classification and regression tasks for graph-structured data. However, as GNNs typically define the graph convolution by the orthonormal basis for the graph Laplacian, they suffer from high computational cost when the graph size is large. This paper introduces a Haar basis, which is a sparse and localized orthonormal system for a coarse-grained chain on the graph. The graph convolution under Haar basis, called Haar convolution, can be defined accordingly for GNNs. The sparsity and locality of the Haar basis allow Fast Haar Transforms (FHTs) on the graph, by which one then achieves a fast evaluation of Haar convolution between graph data and filters. We conduct experiments on GNNs equipped with Haar convolution, which demonstrates state-of-the-art results on graph-based regression and node classification tasks.

Keywords: Fast Haar Transforms; Geometric deep learning; Graph Laplacian; Graph Neural Networks; Graph convolution; Haar basis.

MeSH terms

  • Neural Networks, Computer*
  • Time Factors