计算机科学
人工智能
对偶(语法数字)
翻译(生物学)
图像(数学)
对象(语法)
计算机视觉
极限(数学)
方案(数学)
对抗制
模式识别(心理学)
机器学习
数学
艺术
数学分析
文学类
信使核糖核酸
化学
基因
生物化学
作者
Zimeng Zhao,Binghui Zuo,Zhiyu Long,Yangang Wang
标识
DOI:10.1109/cvpr52729.2023.01167
摘要
Enormous hand images with reliable annotations are collected through marker-based MoCap. Unfortunately, degradations caused by markers limit their application in hand appearance reconstruction. A clear appearance recovery insight is an image-to-image translation trained with unpaired data. However, most frameworks fail because there exists structure inconsistency from a degraded hand to a bare one. The core of our approach is to first disentangle the bare hand structure from those degraded images and then wrap the appearance to this structure with a dual adversarial discrimination (DAD) scheme. Both modules take full advantage of the semi-supervised learning paradigm: The structure disentanglement benefits from the modeling ability of ViT, and the translator is enhanced by the dual discrimination on both translation processes and translation results. Comprehensive evaluations have been conducted to prove that our framework can robustly recover photo-realistic hand appearance from diverse marker-contained and even object-occluded datasets. It provides a novel avenue to acquire bare hand appearance data for other down-stream learning problems.
科研通智能强力驱动
Strongly Powered by AbleSci AI