论文标题
MetaporTrait:具有快速个性化适应性的具有身份的交谈型校长
MetaPortrait: Identity-Preserving Talking Head Generation with Fast Personalized Adaptation
论文作者
论文摘要
在这项工作中,我们提出了一个具有ID ID的谈话头部生成框架,该框架在两个方面都采用了以前的方法。首先,与从稀疏流量中插值相反,我们声称密集的地标对于实现准确的几何感知流场至关重要。其次,受面部交换方法的启发,我们在合成过程中自适应地融合了源身份,因此网络可以更好地保留图像肖像的关键特征。尽管所提出的模型超过了既定基准的先前一代忠诚度,但要进一步使会说话的头部生成有资格获得真实用法,通常需要个性化的微调。但是,这个过程在计算上是对标准用户无法承受的要求。为了解决这个问题,我们建议使用元学习方法提出快速适应模型。学到的模型可以适应高质量的个性化模型,以至30秒钟。最后但并非最不重要的一点是,提出了一个空间增强模块,以改善细节,同时确保时间连贯性。广泛的实验证明了我们的方法在单次和个性化的环境中都比艺术状态具有重要的优势。
In this work, we propose an ID-preserving talking head generation framework, which advances previous methods in two aspects. First, as opposed to interpolating from sparse flow, we claim that dense landmarks are crucial to achieving accurate geometry-aware flow fields. Second, inspired by face-swapping methods, we adaptively fuse the source identity during synthesis, so that the network better preserves the key characteristics of the image portrait. Although the proposed model surpasses prior generation fidelity on established benchmarks, to further make the talking head generation qualified for real usage, personalized fine-tuning is usually needed. However, this process is rather computationally demanding that is unaffordable to standard users. To solve this, we propose a fast adaptation model using a meta-learning approach. The learned model can be adapted to a high-quality personalized model as fast as 30 seconds. Last but not the least, a spatial-temporal enhancement module is proposed to improve the fine details while ensuring temporal coherency. Extensive experiments prove the significant superiority of our approach over the state of the arts in both one-shot and personalized settings.