论文标题
图像动画的薄板样条运动模型
Thin-Plate Spline Motion Model for Image Animation
论文作者
论文摘要
图像动画根据驱动视频将寿命带入源图像中的静态对象。最近的工作试图通过无监督的方法对任意对象进行运动转移,而无需使用先验知识。但是,对于当前的无监督方法,当源中的对象和驾驶图像之间存在较大的姿势间隙时,这仍然是一个重大挑战。在本文中,提出了一个新的端到端无监督运动转移框架来克服此类问题。首先,我们提出了薄板样条运动估计,以产生更灵活的光流,该光流将源图像的特征图延迟到驱动图像的特征域。其次,为了更现实地恢复缺失的区域,我们利用多分辨率的闭塞面膜来实现更有效的特征融合。最后,设计了其他辅助损失功能,以确保网络模块中有明确的劳动力划分,从而鼓励网络生成高质量的图像。我们的方法可以使各种对象动画,包括说话的面孔,人体和像素动画。实验表明,我们的方法在大多数基准测试基准上的性能要比最先进的状态更好,具有可见的姿势相关指标。
Image animation brings life to the static object in the source image according to the driving video. Recent works attempt to perform motion transfer on arbitrary objects through unsupervised methods without using a priori knowledge. However, it remains a significant challenge for current unsupervised methods when there is a large pose gap between the objects in the source and driving images. In this paper, a new end-to-end unsupervised motion transfer framework is proposed to overcome such issue. Firstly, we propose thin-plate spline motion estimation to produce a more flexible optical flow, which warps the feature maps of the source image to the feature domain of the driving image. Secondly, in order to restore the missing regions more realistically, we leverage multi-resolution occlusion masks to achieve more effective feature fusion. Finally, additional auxiliary loss functions are designed to ensure that there is a clear division of labor in the network modules, encouraging the network to generate high-quality images. Our method can animate a variety of objects, including talking faces, human bodies, and pixel animations. Experiments demonstrate that our method performs better on most benchmarks than the state of the art with visible improvements in pose-related metrics.