计算机科学
单眼
人工智能
水准点(测量)
卷积神经网络
编码(集合论)
特征(语言学)
深度学习
特征提取
计算
卷积(计算机科学)
推论
模式识别(心理学)
人工神经网络
算法
语言学
哲学
大地测量学
集合(抽象数据类型)
程序设计语言
地理
作者
Boya Wang,Shuo Wang,Dong Ye,Ziwen Dou
标识
DOI:10.1109/icassp48485.2024.10447168
摘要
With the frequent use of self-supervised monocular depth estimation in robotics and autonomous driving, the model's efficiency is becoming increasingly important. Most current approaches apply much larger and more complex networks to improve the precision of depth estimation. Some researchers incorporated Transformer into self-supervised monocular depth estimation to achieve better performance. However, this method leads to high parameters and high computation. We present a fully convolutional depth estimation network using contextual feature fusion. Compared to UNet++ and HRNet, we use high-resolution and low-resolution features to reserve information on small targets and fast-moving objects instead of long-range fusion. We further promote depth estimation results employing lightweight channel attention based on convolution in the decoder stage. Our method reduces the parameters without sacrificing accuracy. Experiments on the KITTI benchmark show that our method can get better results than many large models, such as Monodepth2, with only 30% parameters. The source code is available at https://github.com/boyagesmile/DNA-Depth.
科研通智能强力驱动
Strongly Powered by AbleSci AI