M4FNet: Multimodal medical image fusion network via multi-receptive-field and multi-scale feature integration

Comput Biol Med. 2023 Jun:159:106923. doi: 10.1016/j.compbiomed.2023.106923. Epub 2023 Apr 14.

Abstract

The main purpose of multimodal medical image fusion is to aggregate the significant information from different modalities and obtain an informative image, which provides comprehensive content and may help to boost other image processing tasks. Many existing methods based on deep learning neglect the extraction and retention of multi-scale features of medical images and the construction of long-distance relationships between depth feature blocks. Therefore, a robust multimodal medical image fusion network via the multi-receptive-field and multi-scale feature (M4FNet) is proposed to achieve the purpose of preserving detailed textures and highlighting the structural characteristics. Specifically, the dual-branch dense hybrid dilated convolution blocks (DHDCB) is proposed to extract the depth features from multi-modalities by expanding the receptive field of the convolution kernel as well as reusing features, and establish long-range dependencies. In order to make full use of the semantic features of the source images, the depth features are decomposed into multi-scale domain by combining the 2-D scale function and wavelet function. Subsequently, the down-sampling depth features are fused by the proposed attention-aware fusion strategy and inversed to the feature space with equal size of source images. Ultimately, the fusion result is reconstructed by a deconvolution block. To force the fusion network balancing information preservation, a local standard deviation-driven structural similarity is proposed as the loss function. Extensive experiments prove that the performance of the proposed fusion network outperforms six state-of-the-art methods, which SD, MI, QABF and QEP are about 12.8%, 4.1%, 8.5% and 9.7% gains, respectively.

Keywords: Long-range dependencies; Medical image fusion; Multi-scale features; Unsupervised learning.

Publication types

  • Research Support, Non-U.S. Gov't

MeSH terms

  • Algorithms*
  • Image Processing, Computer-Assisted*
  • Semantics