判别式
特征(语言学)
计算机科学
特征学习
嵌入
加权
利用
特征提取
一致性(知识库)
人工智能
机器学习
编码(集合论)
模式识别(心理学)
正规化(语言学)
子空间拓扑
源代码
水准点(测量)
频域
数据挖掘
特征向量
排名(信息检索)
领域(数学分析)
降维
学习排名
先验与后验
矩阵分解
灵活性(工程)
模态(人机交互)
秩(图论)
作者
Hongyang Gu,Xiaogang Yang,Ruitao Lu,Lei Pu,Siming Han,Ming Wu
标识
DOI:10.1109/tcsvt.2025.3612751
摘要
Visible-Infrared Person Re-identification (VI-ReID) is critical for round-the-clock surveillance systems yet is hindered by significant modality discrepancies. Existing methods often fail to fully exploit frequency domain information, focusing predominantly on spatial domain feature learning or limited frequency decompositions. To address this, we propose the Multi-Frequency Embedding Network (MFENet), a feature-level method that operates in the frequency domain through multi-frequency decomposition to learn discriminative and modality-invariant features. Specifically, the HiLo-Frequency Modulation (HiLo-FM) module efficiently extracts low-frequency features via frequency-domain filtering and high-frequency details through lightweight multiscale convolutions, followed by attention-based fusion. The Frequency-Aware Diversity Enhancer (FADE) module further enriches feature discriminability by weighting multi-frequency components and learning diverse features through multi-branch architectures. To further enhance the performance of our method, we introduce two innovative loss functions. The Cross-Modality Soft Retrieval (CMSR) loss prioritizes cross-modality consistency over intra-modality similarity, while the Cross-Modality Ranking Regularization (CMRR) loss enhances feature diversity through differentiable rank correlation optimization. Extensive experiments demonstrate the state-of-the-art performance of our method, achieving 61.06% Rank-1 and 67.75% mAP in the challenging IR to VIS mode on the largest VI-ReID benchmark LLCM, surpassing existing methods by significant margins without resorting to reranking or additional labeled data. Code is available at https://github.com/GuHY777/MFENet-VIReID.
科研通智能强力驱动
Strongly Powered by AbleSci AI