论文标题

MVP-Human数据集用于3D人类头像重建

MVP-Human Dataset for 3D Human Avatar Reconstruction from Unconstrained Frames

论文作者

Zhu, Xiangyu, Liao, Tingting, Lyu, Jiangjing, Yan, Xiang, Wang, Yunfeng, Guo, Kan, Cao, Qiong, Li, Stan Z., Lei, Zhen

论文摘要

在本文中,我们考虑了一个新的问题,即从多个无约束的框架中重建3D人类化身,而与摄像机校准,捕获空间和受约束作用的假设无关。该问题应通过一个将多个无约束图像作为输入的框架解决,并在规范空间中生成形状,并在一个馈送方面完成。为此,我们介绍了野生(ARWILD)中的3D化身重建,该重建首先以多层次的方式重建隐式皮肤领域,通过该图像来自多个图像的图像特征并集成以估算代表穿衣形状的Pixel-Synigned Intagit函数。为了启用新框架的培训和测试,我们贡献了一个大规模的数据集MVP-Human(多视图和多姿势3D人),其中包含400名受试者,每个受试者都有15个主题,每个姿势中有15个扫描,每个姿势的8视图图像,提供6,000 3D Scans和48,000张图像。总体而言,从特定的网络体系结构和不同数据中受益,受过训练的模型可从无约束的框架中实现3D头像重建,并实现最新的性能。

In this paper, we consider a novel problem of reconstructing a 3D human avatar from multiple unconstrained frames, independent of assumptions on camera calibration, capture space, and constrained actions. The problem should be addressed by a framework that takes multiple unconstrained images as inputs, and generates a shape-with-skinning avatar in the canonical space, finished in one feed-forward pass. To this end, we present 3D Avatar Reconstruction in the wild (ARwild), which first reconstructs the implicit skinning fields in a multi-level manner, by which the image features from multiple images are aligned and integrated to estimate a pixel-aligned implicit function that represents the clothed shape. To enable the training and testing of the new framework, we contribute a large-scale dataset, MVP-Human (Multi-View and multi-Pose 3D Human), which contains 400 subjects, each of which has 15 scans in different poses and 8-view images for each pose, providing 6,000 3D scans and 48,000 images in total. Overall, benefits from the specific network architecture and the diverse data, the trained model enables 3D avatar reconstruction from unconstrained frames and achieves state-of-the-art performance.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源