Learning Discriminative Cross-Modality Features for RGB-D Saliency Detection

RGB颜色模型 人工智能 判别式 模式识别(心理学) 计算机科学 相关性 模态(人机交互) 特征(语言学) 计算机视觉 串联(数学) 像素 光学(聚焦) 分割 数学 组合数学 语言学 光学 物理 哲学 几何学
作者
Fengyun Wang,Jinshan Pan,Shoukun Xu,Jinhui Tang
出处
期刊:IEEE transactions on image processing [Institute of Electrical and Electronics Engineers]
卷期号:31: 1285-1297 被引量:88
标识
DOI:10.1109/tip.2022.3140606
摘要

How to explore useful information from depth is the key success of the RGB-D saliency detection methods. While the RGB and depth images are from different domains, a modality gap will lead to unsatisfactory results for simple feature concatenation. Towards better performance, most methods focus on bridging this gap and designing different cross-modal fusion modules for features, while ignoring explicitly extracting some useful consistent information from them. To overcome this problem, we develop a simple yet effective RGB-D saliency detection method by learning discriminative cross-modality features based on the deep neural network. The proposed method first learns modality-specific features for RGB and depth inputs. And then we separately calculate the correlations of every pixel-pair in a cross-modality consistent way, i.e., the distribution ranges are consistent for the correlations calculated based on features extracted from RGB (RGB correlation) or depth inputs (depth correlation). From different perspectives, color or spatial, the RGB and depth correlations end up at the same point to depict how tightly each pixel-pair is related. Secondly, to complemently gather RGB and depth information, we propose a novel correlation-fusion to fuse RGB and depth correlations, resulting in a cross-modality correlation. Finally, the features are refined with both long-range cross-modality correlations and local depth correlations to predict salient maps. In which, the long-range cross-modality correlation provides context information for accurate localization, and the local depth correlation keeps good subtle structures for fine segmentation. In addition, a lightweight DepthNet is designed for efficient depth feature extraction. We solve the proposed network in an end-to-end manner. Both quantitative and qualitative experimental results demonstrate the proposed algorithm achieves favorable performance against state-of-the-art methods.
最长约 10秒,即可获得该文献文件

科研通智能强力驱动
Strongly Powered by AbleSci AI
科研通是完全免费的文献互助平台,具备全网最快的应助速度,最高的求助完成率。 对每一个文献求助,科研通都将尽心尽力,给求助人一个满意的交代。
实时播报
刚刚
1秒前
斯文败类应助bian采纳,获得10
1秒前
1秒前
简约发布了新的文献求助10
1秒前
黄队的橄榄完成签到,获得积分10
1秒前
路痴发布了新的文献求助10
1秒前
丘比特应助TaoJ采纳,获得10
1秒前
tcwei发布了新的文献求助10
2秒前
jiayoujijin完成签到 ,获得积分10
3秒前
Sleven完成签到,获得积分10
4秒前
小小孙发布了新的文献求助10
5秒前
5秒前
冷艳宛亦发布了新的文献求助10
5秒前
ddd杜完成签到,获得积分10
5秒前
ZZZZ发布了新的文献求助10
5秒前
6秒前
正在载入中完成签到,获得积分10
6秒前
7秒前
7秒前
顺利鸡完成签到,获得积分10
9秒前
CC完成签到,获得积分10
9秒前
chenchen完成签到,获得积分10
9秒前
NexusExplorer应助科研通管家采纳,获得10
9秒前
9秒前
NexusExplorer应助科研通管家采纳,获得10
9秒前
wanci应助科研通管家采纳,获得10
9秒前
9秒前
酷波er应助科研通管家采纳,获得10
9秒前
wanci应助科研通管家采纳,获得10
9秒前
深情安青应助科研通管家采纳,获得10
9秒前
酷波er应助科研通管家采纳,获得10
9秒前
10秒前
深情安青应助科研通管家采纳,获得10
10秒前
10秒前
10秒前
乐乐应助科研通管家采纳,获得10
10秒前
10秒前
10秒前
乐乐应助科研通管家采纳,获得10
10秒前
高分求助中
(应助此贴封号)【重要!!请各用户(尤其是新用户)详细阅读】【科研通的精品贴汇总】 10000
Kinesiophobia : a new view of chronic pain behavior 3000
Les Mantodea de guyane 2500
Molecular Biology of Cancer: Mechanisms, Targets, and Therapeutics 2000
Standard: In-Space Storable Fluid Transfer for Prepared Spacecraft (AIAA S-157-2024) 1000
Signals, Systems, and Signal Processing 510
Discrete-Time Signals and Systems 510
热门求助领域 (近24小时)
化学 材料科学 生物 医学 工程类 计算机科学 有机化学 物理 生物化学 纳米技术 复合材料 内科学 化学工程 人工智能 催化作用 遗传学 数学 基因 量子力学 物理化学
热门帖子
关注 科研通微信公众号,转发送积分 5948826
求助须知:如何正确求助?哪些是违规求助? 7118154
关于积分的说明 15913428
捐赠科研通 5081759
什么是DOI,文献DOI怎么找? 2732213
邀请新用户注册赠送积分活动 1692603
关于科研通互助平台的介绍 1615456