姿势
人工智能
激光雷达
计算机科学
计算机视觉
单眼
深度图
三维姿态估计
帧(网络)
估计
关节式人体姿态估计
深度知觉
图像(数学)
遥感
地理
工程类
电信
系统工程
神经科学
感知
生物
作者
Wenhua Wu,Guangming Wang,Jiquan Zhong,Hesheng Wang,Zhe Liu
标识
DOI:10.1109/icra48891.2023.10160391
摘要
Depth estimation is one of the most important tasks in scene understanding. In the existing joint self-supervised learning approaches of depth-pose estimation, depth estimation and pose estimation networks are independent of each other. They only use the adjacent image frames for pose estimation and lack the use of the estimated geometric information. To enhance the depth-pose association, we propose a monocular multi-frame unsupervised depth estimation framework, named PLPE-Depth. There are a depth estimation network and two pose estimation networks with image input and pseudo-LiDAR input. The main idea of our approach is to use the pseudo-LiDAR reconstructed from the depth map to estimate the pose of adjacent frames. We propose depth re-estimation with a better pose between the image pose and the pseudo-LiDAR pose to improve the accuracy of estimation. Besides, we improve the reconstruction loss and design a pseudo-LiDAR pose enhancement loss to facilitate the joint learning. Our approach enhances the use of the estimated depth information and strengthens the coupling between depth estimation and pose estimation. Experiments on the KITTI dataset show that our depth estimation achieves state-of-the-art performance at low resolution. Our source codes will be released on https://github.com/IRMVLabIPLPE-Depth.
科研通智能强力驱动
Strongly Powered by AbleSci AI