Learning Discriminative Cross-Modality Features for RGB-D Saliency Detection

RGB颜色模型 人工智能 判别式 模式识别(心理学) 计算机科学 相关性 模态(人机交互) 特征(语言学) 计算机视觉 串联(数学) 像素 光学(聚焦) 分割 数学 组合数学 语言学 光学 物理 哲学 几何学
作者
Fengyun Wang,Jinshan Pan,Sugang Xu,Jinhui Tang
出处
期刊:IEEE transactions on image processing [Institute of Electrical and Electronics Engineers]
卷期号:31: 1285-1297 被引量:42
标识
DOI:10.1109/tip.2022.3140606
摘要

How to explore useful information from depth is the key success of the RGB-D saliency detection methods. While the RGB and depth images are from different domains, a modality gap will lead to unsatisfactory results for simple feature concatenation. Towards better performance, most methods focus on bridging this gap and designing different cross-modal fusion modules for features, while ignoring explicitly extracting some useful consistent information from them. To overcome this problem, we develop a simple yet effective RGB-D saliency detection method by learning discriminative cross-modality features based on the deep neural network. The proposed method first learns modality-specific features for RGB and depth inputs. And then we separately calculate the correlations of every pixel-pair in a cross-modality consistent way, i.e., the distribution ranges are consistent for the correlations calculated based on features extracted from RGB (RGB correlation) or depth inputs (depth correlation). From different perspectives, color or spatial, the RGB and depth correlations end up at the same point to depict how tightly each pixel-pair is related. Secondly, to complemently gather RGB and depth information, we propose a novel correlation-fusion to fuse RGB and depth correlations, resulting in a cross-modality correlation. Finally, the features are refined with both long-range cross-modality correlations and local depth correlations to predict salient maps. In which, the long-range cross-modality correlation provides context information for accurate localization, and the local depth correlation keeps good subtle structures for fine segmentation. In addition, a lightweight DepthNet is designed for efficient depth feature extraction. We solve the proposed network in an end-to-end manner. Both quantitative and qualitative experimental results demonstrate the proposed algorithm achieves favorable performance against state-of-the-art methods.
最长约 10秒,即可获得该文献文件

科研通智能强力驱动
Strongly Powered by AbleSci AI
更新
大幅提高文件上传限制,最高150M (2024-4-1)

科研通是完全免费的文献互助平台,具备全网最快的应助速度,最高的求助完成率。 对每一个文献求助,科研通都将尽心尽力,给求助人一个满意的交代。
实时播报
1秒前
1秒前
2秒前
平淡秋白发布了新的文献求助10
2秒前
2秒前
Mayinhere发布了新的文献求助10
3秒前
常常完成签到 ,获得积分10
4秒前
4秒前
4秒前
zhangyumin发布了新的文献求助10
5秒前
Greed应助D&L采纳,获得10
6秒前
peter发布了新的文献求助10
6秒前
李木子发布了新的文献求助10
7秒前
胖胖龙发布了新的文献求助30
7秒前
7秒前
阿宋发布了新的文献求助30
8秒前
xiongyh10发布了新的文献求助10
8秒前
8秒前
9秒前
cctv18应助callous采纳,获得30
10秒前
zhenzheng完成签到 ,获得积分10
10秒前
小石头完成签到 ,获得积分10
10秒前
SaSa完成签到,获得积分10
12秒前
Jasper应助顾长生采纳,获得10
13秒前
科研通AI2S应助李木子采纳,获得10
13秒前
九州发布了新的文献求助10
13秒前
13秒前
14秒前
RealPeace应助傻鱼吃猫采纳,获得10
14秒前
chu发布了新的文献求助10
15秒前
xingyi完成签到 ,获得积分10
15秒前
白冷之发布了新的文献求助10
15秒前
犹豫的笑旋完成签到,获得积分10
16秒前
强健的心锁完成签到,获得积分10
16秒前
Mayinhere完成签到,获得积分10
16秒前
陈昕炜发布了新的文献求助10
16秒前
whh发布了新的文献求助10
16秒前
19秒前
Ziwei发布了新的文献求助10
19秒前
纷纷故事完成签到 ,获得积分10
19秒前
高分求助中
The three stars each : the Astrolabes and related texts 1070
Manual of Clinical Microbiology, 4 Volume Set (ASM Books) 13th Edition 1000
Sport in der Antike 800
Aspect and Predication: The Semantics of Argument Structure 666
De arte gymnastica. The art of gymnastics 600
少脉山油柑叶的化学成分研究 530
Sport in der Antike Hardcover – March 1, 2015 500
热门求助领域 (近24小时)
化学 材料科学 医学 生物 有机化学 工程类 生物化学 纳米技术 物理 内科学 计算机科学 化学工程 复合材料 遗传学 基因 物理化学 催化作用 电极 光电子学 量子力学
热门帖子
关注 科研通微信公众号,转发送积分 2409311
求助须知:如何正确求助?哪些是违规求助? 2105252
关于积分的说明 5316657
捐赠科研通 1832725
什么是DOI,文献DOI怎么找? 913204
版权声明 560754
科研通“疑难数据库(出版商)”最低求助积分说明 488289