计算机科学
模式
量化(信号处理)
人工智能
散列函数
机器学习
选择(遗传算法)
情态动词
信息丢失
模式识别(心理学)
数据挖掘
计算机视觉
社会学
计算机安全
化学
高分子化学
社会科学
作者
Erkun Yang,Dongren Yao,Tongliang Liu,Cheng Deng
标识
DOI:10.1109/cvpr52688.2022.00740
摘要
Deep cross-modal hashing has become an essential tool for supervised multimodal search. These models tend to be optimized with large, curated multimodal datasets, where most labels have been manually verified. Unfortunately, in many scenarios, such accurate labeling may not be avail-able. In contrast, datasets with low-quality annotations may be acquired, which inevitably introduce numerous mis-takes or label noise and therefore degrade the search per-formance. To address the challenge, we present a general robust cross-modal hashing framework to correlate distinct modalities and combat noisy labels simultaneously. More specifically, we propose a proxy-based contrastive (PC) loss to mitigate the gap between different modalities and train networks for different modalities Jointly with small-loss samples that are selected with the PC loss and a mu-tual quantization loss. The small-loss sample selection from such Joint loss can help choose confident examples to guide the model training, and the mutual quantization loss can maximize the agreement between different modalities and is beneficial to improve the effectiveness of sample selection. Experiments on three widely-used multimodal datasets show that our method significantly outperforms existing state-of-the-arts.
科研通智能强力驱动
Strongly Powered by AbleSci AI