汉明空间
计算机科学
散列函数
二进制代码
特征哈希
通用哈希
动态完美哈希
人工智能
理论计算机科学
判别式
模式识别(心理学)
机器学习
二进制数
哈希表
汉明码
双重哈希
算法
数学
区块代码
计算机安全
算术
解码方法
作者
Wen Gu,Xiaoyan Gu,Jingzi Gu,Bo Li,Zhi Xiong,Weiping Wang
标识
DOI:10.1145/3323873.3325045
摘要
Cross-modal hashing has attracted considerable attention for large-scale multimodal retrieval task. A majority of hashing methods have been proposed for cross-modal retrieval. However, these methods inadequately focus on feature learning process and cannot fully preserve higher-ranking correlation of various item pairs as well as the multi-label semantics of each item, so that the quality of binary codes may be downgraded. To tackle these problems, in this paper, we propose a novel deep cross-modal hashing method, called Adversary Guided Asymmetric Hashing (AGAH). Specifically, it employs an adversarial learning guided multi-label attention module to enhance the feature learning part which can learn discriminative feature representations and keep the cross-modal invariability. Furthermore, in order to generate hash codes which can fully preserve the multi-label semantics of all items, we propose an asymmetric hashing method which utilizes a multi-label binary code map that can equip the hash codes with multi-label semantic information. In addition, to ensure higher-ranking correlation of all similar item pairs than those of dissimilar ones, we adopt a new triplet-margin constraint and a cosine quantization technique for Hamming space similarity preservation. Extensive empirical studies show that AGAH outperforms several state-of-the-art methods for cross-modal retrieval.
科研通智能强力驱动
Strongly Powered by AbleSci AI