论文标题
从单眼视频中捕获和动画的身体和衣服
Capturing and Animation of Body and Clothing from Monocular Video
论文作者
论文摘要
尽管最近的工作显示了从单个图像,视频或一组3D扫描中提取衣服的3D人体化身的进展,但仍然存在一些限制。大多数方法都使用整体代表来共同对身体和衣服进行建模,这意味着无法在虚拟试验之类的应用中分离服装和身体。其他方法分别对车身和衣服进行了建模,但需要从3D/4D扫描仪或物理模拟的大量3D衣服的人网络中进行训练。我们的见解是,身体和衣服有不同的建模要求。尽管身体由基于网格的参数3D模型很好地表示,但隐式表示和神经辐射场更适合捕获服装中存在的大量形状和外观。在这种见解的基础上,我们提出了围巾(分段服装的头像辐射场),这是一种混合模型,将基于网格的身体与神经辐射场结合在一起。将网格与容量渲染结合在一起,并结合使用可区分的栅格器,使我们能够直接从单眼视频中优化围巾,而无需任何3D监督。混合建模使围巾(i)通过改变身体姿势(包括手铰接和面部表情),(ii)合成avatar新颖的视图,以及(iii)在虚拟试用应用中的头像之间转移服装。我们证明,围巾比现有方法更高的视觉质量重建衣服,衣服随着身体姿势和身体形状变化而变形,并且可以在不同受试者的化身之间成功地转移衣服。代码和型号可在https://github.com/yadiraf/scarf上找到。
While recent work has shown progress on extracting clothed 3D human avatars from a single image, video, or a set of 3D scans, several limitations remain. Most methods use a holistic representation to jointly model the body and clothing, which means that the clothing and body cannot be separated for applications like virtual try-on. Other methods separately model the body and clothing, but they require training from a large set of 3D clothed human meshes obtained from 3D/4D scanners or physics simulations. Our insight is that the body and clothing have different modeling requirements. While the body is well represented by a mesh-based parametric 3D model, implicit representations and neural radiance fields are better suited to capturing the large variety in shape and appearance present in clothing. Building on this insight, we propose SCARF (Segmented Clothed Avatar Radiance Field), a hybrid model combining a mesh-based body with a neural radiance field. Integrating the mesh into the volumetric rendering in combination with a differentiable rasterizer enables us to optimize SCARF directly from monocular videos, without any 3D supervision. The hybrid modeling enables SCARF to (i) animate the clothed body avatar by changing body poses (including hand articulation and facial expressions), (ii) synthesize novel views of the avatar, and (iii) transfer clothing between avatars in virtual try-on applications. We demonstrate that SCARF reconstructs clothing with higher visual quality than existing methods, that the clothing deforms with changing body pose and body shape, and that clothing can be successfully transferred between avatars of different subjects. The code and models are available at https://github.com/YadiraF/SCARF.