Real-time instance segmentation of surgical instruments using attention and multi-scale feature fusion

Med Image Anal. 2022 Oct:81:102569. doi: 10.1016/j.media.2022.102569. Epub 2022 Aug 6.

Abstract

Precise instrument segmentation aids surgeons to navigate the body more easily and increases patient safety. While accurate tracking of surgical instruments in real-time plays a crucial role in minimally invasive computer-assisted surgeries, it is a challenging task to achieve, mainly due to: (1) a complex surgical environment, and (2) model design trade-off in terms of both optimal accuracy and speed. Deep learning gives us the opportunity to learn complex environment from large surgery scene environments and placements of these instruments in real world scenarios. The Robust Medical Instrument Segmentation 2019 challenge (ROBUST-MIS) provides more than 10,000 frames with surgical tools in different clinical settings. In this paper, we propose a light-weight single stage instance segmentation model complemented with a convolutional block attention module for achieving both faster and accurate inference. We further improve accuracy through data augmentation and optimal anchor localization strategies. To our knowledge, this is the first work that explicitly focuses on both real-time performance and improved accuracy. Our approach out-performed top team performances in the most recent edition of ROBUST-MIS challenge with over 44% improvement on area-based multi-instance dice metric MI_DSC and 39% on distance-based multi-instance normalized surface dice MI_NSD. We also demonstrate real-time performance (>60 frames-per-second) with different but competitive variants of our final approach.

Keywords: Attention; Deep learning; MIS instance segmentation; Multi-scale feature fusion; Real-time; Single-stage.

MeSH terms

  • Attention
  • Humans
  • Image Processing, Computer-Assisted
  • Minimally Invasive Surgical Procedures
  • Surgery, Computer-Assisted*
  • Surgical Instruments*