姿势
计算机科学
过度拟合
人工智能
接头(建筑物)
模棱两可
一般化
机器学习
单眼
三维姿态估计
计算机视觉
模式识别(心理学)
人工神经网络
数学
工程类
建筑工程
数学分析
程序设计语言
作者
Chuchu Han,Xin Yu,Changxin Gao,Nong Sang,Yi Yang
标识
DOI:10.1016/j.patcog.2022.108934
摘要
In monocular image scenes, 3D human pose estimation exhibits inherent ambiguity due to the loss of depth information and occlusions. Simply regressing body joints with high uncertainties will lead to model overfitting and poor generalization. In this paper, we propose an uncertainty-based framework to jointly learn 3D human poses and the uncertainty of each joint. Our proposed joint estimation framework aims to mitigate the adverse effects of training samples with high uncertainties and facilitate the training procedure. To be specific, we model each body joint as a Laplace distribution for uncertainty representation. Since visual joints often exhibit low uncertainties while occluded ones have high uncertainties, we develop an adaptive scaling factor, named the uncertainty-aware scaling factor, to ease the network optimization in accordance with the joint uncertainties. By doing so, our network is able to converge faster and significantly reduce the adverse effects caused by those ambiguous joints. Furthermore, we present an uncertainty-aware graph convolutional network by exploiting the learned joint uncertainties and the relationships among joints to refine the initial joint localization. Extensive experiments on single-person (Human3.6M) and multi-person (MuCo-3DHP & MuPoTS-3D) 3D human pose estimation datasets demonstrate the effectiveness of our method.
科研通智能强力驱动
Strongly Powered by AbleSci AI