The search functionality is under construction.

IEICE TRANSACTIONS on Information

Modality-Fused Graph Network for Cross-Modal Retrieval

Fei WU, Shuaishuai LI, Guangchuan PENG, Yongheng MA, Xiao-Yuan JING

  • Full Text Views

    1

  • Cite this

Summary :

Cross-modal hashing technology has attracted much attention for its favorable retrieval performance and low storage cost. However, for existing cross-modal hashing methods, the heterogeneity of data across modalities is still a challenge and how to fully explore and utilize the intra-modality features has not been well studied. In this paper, we propose a novel cross-modal hashing approach called Modality-fused Graph Network (MFGN). The network architecture consists of a text channel and an image channel that are used to learn modality-specific features, and a modality fusion channel that uses the graph network to learn the modality-shared representations to reduce the heterogeneity across modalities. In addition, an integration module is introduced for the image and text channels to fully explore intra-modality features. Experiments on two widely used datasets show that our approach achieves better results than the state-of-the-art cross-modal hashing methods.

Publication
IEICE TRANSACTIONS on Information Vol.E106-D No.5 pp.1094-1097
Publication Date
2023/05/01
Publicized
2023/02/09
Online ISSN
1745-1361
DOI
10.1587/transinf.2022EDL8069
Type of Manuscript
LETTER
Category
Pattern Recognition

Authors

Fei WU
  Nanjing University of Posts and Telecommunications
Shuaishuai LI
  Nanjing University of Posts and Telecommunications
Guangchuan PENG
  Nanjing University of Posts and Telecommunications
Yongheng MA
  Nanjing University of Posts and Telecommunications
Xiao-Yuan JING
  Wuhan University

Keyword