Fast discrete cross-modal hashing with semantic consistency

Neural Netw. 2020 May:125:142-152. doi: 10.1016/j.neunet.2020.01.035. Epub 2020 Feb 11.

Abstract

Supervised cross-modal hashing has attracted widespread concentrations for large-scale retrieval task due to its promising retrieval performance. However, most existing works suffer from some of following issues. Firstly, most of them only leverage the pair-wise similarity matrix to learn hash codes, which may result in class information loss. Secondly, the pair-wise similarity matrix generally lead to high computing complexity and memory cost. Thirdly, most of them relax the discrete constraints during optimization, which generally results in large cumulative quantization error and consequent inferior hash codes. To address above problems, we present a Fast Discrete Cross-modal Hashing method in this paper, FDCH for short. Specifically, it firstly leverages both class labels and the pair-wise similarity matrix to learn a sharing Hamming space where the semantic consistency can be better preserved. Then we propose an asymmetric hash codes learning model to avoid the challenging issue of symmetric matrix factorization. Finally, an effective and efficient discrete optimal scheme is designed to generate discrete hash codes directly, and the computing complexity and memory cost caused by the pair-wise similarity matrix are reduced from O(n2) to O(n), where n denotes the size of training set. Extensive experiments conducted on three real world datasets highlight the superiority of FDCH compared with several cross-modal hashing methods and demonstrate its effectiveness and efficiency.

Keywords: Cross-modal retrieval; Discrete optimization; Hashing; Semantic consistency.

MeSH terms

  • Algorithms*
  • Deep Learning / trends
  • Humans
  • Pattern Recognition, Automated / methods*
  • Pattern Recognition, Automated / trends
  • Semantics*
  • Time Factors