Enhancing medical image segmentation with a multi-transformer U-Net

PeerJ. 2024 Feb 29:12:e17005. doi: 10.7717/peerj.17005. eCollection 2024.

Abstract

Various segmentation networks based on Swin Transformer have shown promise in medical segmentation tasks. Nonetheless, challenges such as lower accuracy and slower training convergence have persisted. To tackle these issues, we introduce a novel approach that combines the Swin Transformer and Deformable Transformer to enhance overall model performance. We leverage the Swin Transformer's window attention mechanism to capture local feature information and employ the Deformable Transformer to adjust sampling positions dynamically, accelerating model convergence and aligning it more closely with object shapes and sizes. By amalgamating both Transformer modules and incorporating additional skip connections to minimize information loss, our proposed model excels at rapidly and accurately segmenting CT or X-ray lung images. Experimental results demonstrate the remarkable, showcasing the significant prowess of our model. It surpasses the performance of the standalone Swin Transformer's Swin Unet and converges more rapidly under identical conditions, yielding accuracy improvements of 0.7% (resulting in 88.18%) and 2.7% (resulting in 98.01%) on the COVID-19 CT scan lesion segmentation dataset and Chest X-ray Masks and Labels dataset, respectively. This advancement has the potential to aid medical practitioners in early diagnosis and treatment decision-making.

Keywords: CT or X-ray lung images; Medical image segmentation; Multi-transformer; Unet.

MeSH terms

  • COVID-19* / diagnostic imaging
  • Electric Power Supplies
  • Health Personnel
  • Humans
  • Pemoline
  • Thorax

Substances

  • Pemoline

Grants and funding

The authors received no funding for this work.