Multi-Modal Fake News Detection via Bridging the Gap between Modals

Entropy (Basel). 2023 Apr 4;25(4):614. doi: 10.3390/e25040614.

Abstract

Multi-modal fake news detection aims to identify fake information through text and corresponding images. The current methods purely combine images and text scenarios by a vanilla attention module but there exists a semantic gap between different scenarios. To address this issue, we introduce an image caption-based method to enhance the model's ability to capture semantic information from images. Formally, we integrate image description information into the text to bridge the semantic gap between text and images. Moreover, to optimize image utilization and enhance the semantic interaction between images and text, we combine global and object features from the images for the final representation. Finally, we leverage a transformer to fuse the above multi-modal content. We carried out extensive experiments on two publicly available datasets, and the results show that our proposed method significantly improves performance compared to other existing methods.

Keywords: caption-based; fake news detection; multi-modal; transformer.

Grants and funding

This work was supported by the Research Foundation of Yunnan Province No. 202002AD080001, 202001BB050043, and 2019FA044, the National Natural Science Foundation of China under Grants No. 62162065, and the Provincial Foundation for Leaders of Disciplines in Science and Technology No. 2019HB121.