计算机科学
子空间拓扑
人工智能
判别式
特征选择
模式识别(心理学)
机器学习
正规化(语言学)
特征(语言学)
图形
相似性(几何)
理论计算机科学
哲学
语言学
图像(数学)
作者
Kaiye Wang,Ran He,Liang Wang,Wei Wang,Tieniu Tan
标识
DOI:10.1109/tpami.2015.2505311
摘要
Cross-modal retrieval has recently drawn much attention due to the widespread existence of multimodal data. It takes one type of data as the query to retrieve relevant data objects of another type, and generally involves two basic problems: the measure of relevance and coupled feature selection. Most previous methods just focus on solving the first problem. In this paper, we aim to deal with both problems in a novel joint learning framework. To address the first problem, we learn projection matrices to map multimodal data into a common subspace, in which the similarity between different modalities of data can be measured. In the learning procedure, the l21-norm penalties are imposed on the projection matrices separately to solve the second problem, which selects relevant and discriminative features from different feature spaces simultaneously. A multimodal graph regularization term is further imposed on the projected data,which preserves the inter-modality and intra-modality similarity relationships.An iterative algorithm is presented to solve the proposed joint learning problem, along with its convergence analysis. Experimental results on cross-modal retrieval tasks demonstrate that the proposed method outperforms the state-of-the-art subspace approaches.
科研通智能强力驱动
Strongly Powered by AbleSci AI