Region Pixel Voting Network (RPVNet) for 6D Pose Estimation from Monocular Image

人工智能 计算机科学 计算机视觉 像素 兰萨克 姿势 条件随机场 对象(语法) 模式识别(心理学) 投票 卷积神经网络 图像(数学) 政治 政治学 法学
作者
Feng Xiong,Chengju Liu,Qijun Chen
出处
期刊:Applied sciences [MDPI AG]
卷期号:11 (2): 743-743 被引量:3
标识
DOI:10.3390/app11020743
摘要

Recent studies have shown that deep learning achieves superior results in the task of estimating 6D-pose of target object from an image. End-to-end techniques use deep networks to predict pose directly from image, avoiding the limitations of handcraft features, but rely on training dataset to deal with occlusion. Two-stage algorithms alleviate this problem by finding keypoints in the image and then solving the Perspective-n-Point (PnP) problem to avoid directly fitting the transformation from image space to 6D-pose space. This paper proposes a novel two-stage method using only local features for pixel voting, called Region Pixel Voting Network (RPVNet). Front-end network detects target object and predicts its direction maps, from which the keypoints are recovered by pixel voting using Random Sample Consensus (RANSAC). The backbone, object detection network and mask prediction network of RPVNet are designed based on Mask R-CNN. Direction map is a vector field with the direction of each point pointing to its source keypoint. It is shown that predicting an object’s keypoints is related to its own pixels and independent of other pixels, which means the influence of occlusion decreases in the object’s region. Based on this phenomenon, in RPVNet, local features instead of the whole features, i.e., the output of the backbone, are used by a well-designed Convolutional Neural Networks (CNN) to compute direction maps. The local features are extracted from the whole features through RoIAlign, based on the region provided by detection network. Experiments on LINEMOD dataset show that RPVNet’s average accuracy (86.1%) is almost equal to state-of-the-art (86.4%) when no occlusion occurs. Meanwhile, results on Occlusion LINEMOD dataset show that RPVNet outperforms state-of-the-art (43.7% vs. 40.8%) and is more accurate for small object in occluded scenes.
最长约 10秒,即可获得该文献文件

科研通智能强力驱动
Strongly Powered by AbleSci AI
更新
大幅提高文件上传限制,最高150M (2024-4-1)

科研通是完全免费的文献互助平台,具备全网最快的应助速度,最高的求助完成率。 对每一个文献求助,科研通都将尽心尽力,给求助人一个满意的交代。
实时播报
shinysparrow应助Dobby采纳,获得10
刚刚
shinysparrow应助Dobby采纳,获得10
刚刚
shinysparrow应助Dobby采纳,获得10
刚刚
wanci应助Dobby采纳,获得10
刚刚
CC完成签到,获得积分10
1秒前
2秒前
平常善若发布了新的文献求助10
3秒前
fei发布了新的文献求助10
3秒前
4秒前
科研薯条发布了新的文献求助20
4秒前
4秒前
bingchem发布了新的文献求助10
5秒前
英俊的铭应助CC采纳,获得10
6秒前
7秒前
希望天下0贩的0应助152455采纳,获得10
7秒前
8秒前
8秒前
斯文败类应助99668采纳,获得10
9秒前
可爱的函函应助1205114938采纳,获得10
9秒前
iVANPENNY应助廖元枫采纳,获得10
11秒前
共享精神应助wujiwuhui采纳,获得10
12秒前
稍稍发布了新的文献求助10
13秒前
英姑应助能干的梦岚采纳,获得10
13秒前
偷书贼发布了新的文献求助10
13秒前
老王发布了新的文献求助10
14秒前
莫友安完成签到,获得积分10
14秒前
田様应助小金采纳,获得10
14秒前
小潘完成签到,获得积分10
15秒前
15秒前
林香香完成签到,获得积分10
15秒前
李爱国应助Rachel采纳,获得10
17秒前
zsh完成签到,获得积分10
17秒前
如意道消完成签到,获得积分10
17秒前
大方访文完成签到,获得积分10
18秒前
桐桐应助昵称采纳,获得10
19秒前
迷路的小狗完成签到,获得积分10
19秒前
JamesPei应助Yamila采纳,获得10
19秒前
20秒前
CipherSage应助半夏采纳,获得10
21秒前
稍稍完成签到,获得积分10
21秒前
高分求助中
Thermodynamic data for steelmaking 3000
Teaching Social and Emotional Learning in Physical Education 900
Lexique et typologie des poteries: pour la normalisation de la description des poteries (Full Book) 400
Cardiology: Board and Certification Review 300
Transformerboard III 300
Degenerative Cervical Myelopathy From Basic Science to Clinical Practice 200
Institution Building, Organisational Restructuring and Everyday Negotiations in Uganda's Roads Sector 200
热门求助领域 (近24小时)
化学 材料科学 医学 生物 有机化学 工程类 生物化学 纳米技术 物理 内科学 计算机科学 化学工程 复合材料 遗传学 基因 物理化学 催化作用 电极 光电子学 量子力学
热门帖子
关注 科研通微信公众号,转发送积分 2356165
求助须知:如何正确求助?哪些是违规求助? 2062895
关于积分的说明 5148244
捐赠科研通 1792640
什么是DOI,文献DOI怎么找? 895413
版权声明 557448
科研通“疑难数据库(出版商)”最低求助积分说明 477988