超图
计算机科学
散列函数
模式识别(心理学)
人工智能
相似性(几何)
图形
图像检索
图嵌入
聚类分析
情态动词
数据挖掘
嵌入
理论计算机科学
图像(数学)
数学
离散数学
计算机安全
化学
高分子化学
作者
Fangming Zhong,Chia-Yih Chu,Zijie Zhu,Zhikui Chen
标识
DOI:10.1145/3581783.3612116
摘要
Unsupervised cross-modal hashing retrieval across image and text modality is a challenging task because of the suboptimality of similarity guidance, i.e., the joint similarity matrix constructed by existing methods does not possess clear enough guiding significance. How to construct more robust similarity matrix is the key to solve this problem. The unsupervised cross-modal retrieval methods based on graph have a good performance in mining semantic information of input samples, but the graph hashing based on traditional affinity graph cannot capture the high-order semantic information of input samples effectively. In order to overcome the aforementioned limitations, this paper presents a novel hypergraph-based approach for unsupervised cross-modal retrieval that differs from previous works in two significant ways. Firstly, to address the ubiquitous redundant information present in current methods, this paper introduces a robust similarity matrix constructing method. Secondly, we propose a novel hypergraph enhanced module that produces embedding vectors by hypergraph convolution and attention mechanism for input data, capturing important high-order semantics. Our approach is evaluated on the NUS-WIDE and MIRFlickr datasets, and yields state-of-the-art performance for unsupervised cross-modal retrieval.
科研通智能强力驱动
Strongly Powered by AbleSci AI