计算机科学
伪装
人工智能
RGB颜色模型
边距(机器学习)
保险丝(电气)
水准点(测量)
计算机视觉
特征(语言学)
目标检测
深度学习
对象(语法)
模式识别(心理学)
机器学习
地质学
工程类
哲学
电气工程
语言学
大地测量学
作者
Qingwei Wang,Jinyu Yang,Xiaosheng Yu,F. Wang,Peng Chen,Feng Zheng
标识
DOI:10.1145/3581783.3611874
摘要
Camouflaged Object Detection (COD) aims to identify and segment objects that blend into their surroundings. Since the color and texture of the camouflaged objects are extremely similar to the surrounding environment, it is super challenging for vision models to precisely detect them. Inspired by research on biology and evolution, we introduce depth information as an additional cue to help break camouflage, which can provide spatial information and texture-free separation for foreground and background. To dig clues of camouflaged objects in both RGB and depth modalities, we innovatively propose Depth-aided Camouflaged Object Detection (DaCOD), which involves two key components. We firstly propose the Multi-modal Collaborative Learning (MCL) module, which aims to collaboratively learning deep features from both RGB and depth channels via a hybrid backbone. Then, we propose a novel Cross-modal Asymmetric Fusion (CAF) strategy, which asymmetrically fuse RGB and depth information for complementary depth feature enhancement to produce accurate predictions. We conducted numerous experiments of the proposed DaCOD on three widely-used challenging COD benchmark datasets, in which DaCOD outperforms the current state-of-the-arts by a large margin. All resources are available at https://github.com/qingwei-wang/DaCOD.
科研通智能强力驱动
Strongly Powered by AbleSci AI