HookNet: Multi-resolution convolutional neural networks for semantic segmentation in histopathology whole-slide images

Med Image Anal. 2021 Feb:68:101890. doi: 10.1016/j.media.2020.101890. Epub 2020 Oct 29.

Abstract

We propose HookNet, a semantic segmentation model for histopathology whole-slide images, which combines context and details via multiple branches of encoder-decoder convolutional neural networks. Concentric patches at multiple resolutions with different fields of view, feed different branches of HookNet, and intermediate representations are combined via a hooking mechanism. We describe a framework to design and train HookNet for achieving high-resolution semantic segmentation and introduce constraints to guarantee pixel-wise alignment in feature maps during hooking. We show the advantages of using HookNet in two histopathology image segmentation tasks where tissue type prediction accuracy strongly depends on contextual information, namely (1) multi-class tissue segmentation in breast cancer and, (2) segmentation of tertiary lymphoid structures and germinal centers in lung cancer. We show the superiority of HookNet when compared with single-resolution U-Net models working at different resolutions as well as with a recently published multi-resolution model for histopathology image segmentation. We have made HookNet publicly available by releasing the source code1 as well as in the form of web-based applications2,3 based on the grand-challenge.org platform.

Keywords: Computational pathology; Deep learning; Multi-resolution; Semantic segmentation.

Publication types

  • Research Support, Non-U.S. Gov't

MeSH terms

  • Breast
  • Humans
  • Image Processing, Computer-Assisted*
  • Neural Networks, Computer
  • Semantics*
  • Software