论文标题
域增强了通过对比度学习增强的任意图像样式转移
Domain Enhanced Arbitrary Image Style Transfer via Contrastive Learning
论文作者
论文摘要
在这项工作中,我们使用新型样式特征表示方法解决了任意图像样式转移的具有挑战性的问题。合适的样式表示形式,作为图像样式任务中的关键组成部分,对于获得令人满意的结果至关重要。现有的基于神经网络的方法通过二阶统计数据(例如内容特征的革兰氏矩阵)获得了合理的结果。但是,它们不利用足够的样式信息,这会导致诸如局部扭曲和风格不一致之类的人工制品。为了解决这些问题,我们建议通过分析多种样式之间的相似性和差异并考虑样式分布,直接从图像功能而不是其二阶统计数据中学习样式表示形式。具体而言,我们提出对比度的任意风格转移(CAST),这是一种通过对比学习的新样式表示学习和样式转移方法。我们的框架包括三个关键组件,即用于样式代码编码的多层样式投影仪,用于有效学习样式分布的域增强模块以及用于图像样式传输的生成网络。我们全面进行定性和定量评估,以证明与通过最新方法获得的方法相比,我们的方法取得了明显更好的结果。代码和型号可从https://github.com/zyxelsa/cast_pytorch获得
In this work, we tackle the challenging problem of arbitrary image style transfer using a novel style feature representation learning method. A suitable style representation, as a key component in image stylization tasks, is essential to achieve satisfactory results. Existing deep neural network based approaches achieve reasonable results with the guidance from second-order statistics such as Gram matrix of content features. However, they do not leverage sufficient style information, which results in artifacts such as local distortions and style inconsistency. To address these issues, we propose to learn style representation directly from image features instead of their second-order statistics, by analyzing the similarities and differences between multiple styles and considering the style distribution. Specifically, we present Contrastive Arbitrary Style Transfer (CAST), which is a new style representation learning and style transfer method via contrastive learning. Our framework consists of three key components, i.e., a multi-layer style projector for style code encoding, a domain enhancement module for effective learning of style distribution, and a generative network for image style transfer. We conduct qualitative and quantitative evaluations comprehensively to demonstrate that our approach achieves significantly better results compared to those obtained via state-of-the-art methods. Code and models are available at https://github.com/zyxElsa/CAST_pytorch