判别式
计算机科学
人工智能
特征(语言学)
光学(聚焦)
模态(人机交互)
模式识别(心理学)
滤波器(信号处理)
特征提取
特征学习
任务(项目管理)
小波
计算机视觉
透视图(图形)
匹配(统计)
特征向量
深度学习
语音识别
小波
任务分析
亮度
可视化
隐藏字幕
小波变换
频带
Gabor滤波器
作者
Ting Yu,Deqiang Cheng,He Jiang,Libin Chen,Jiansheng Qian,Qiqi Kou,Guangtao Zhai
标识
DOI:10.1109/tce.2026.3659243
摘要
Visible-infrared person re-identification (VI-ReID) is a challenging person retrieval task under two different modality cameras, due to the large gap between visible and infrared images and intra-class variations within the same individuals. Existing works mainly focus on extracting discriminative spatial features to achieve matching purposes, while overlooking the difference between the low-frequency and high-frequency components in different modalities. To solve this problem, we propose a Wavelet-based Frequency Feature Learning network called WFLNet, which tackles the VI-ReID task from a frequency perspective by filtering out the irrelevant modality-specific components and leveraging the frequency-domain cues to learn more comprehensive and discriminative features. Firstly, we introduce a Processed Low-Frequency Mask (PLFM) to filter out the irrelevant low-frequency components, guiding the model to focus more on modality-invariant features and suppress the negative impact of modality differences caused by the brightness and color information in VIS and thermal radiation information in IR. Secondly, a Multi-scale Wavelet Query Attention module (MWQA) is designed to extract the modality-invariant feature by integrating low- and high-frequency features in cross-scales to enhance global structure and detail information. Finally, the dual frequency loss is proposed to reduce the modality gap by pulling one modality feature closer to another modality center. Extensive experiments on the SYSU-MM01, RegDB, and LLCM datasets demonstrate that the proposed WFLNet outperforms most existing works.
科研通智能强力驱动
Strongly Powered by AbleSci AI