论文标题
熊猫:无监督的零件和外观在甘斯的特征图中
PandA: Unsupervised Learning of Parts and Appearances in the Feature Maps of GANs
论文作者
论文摘要
了解生成对抗网络(GAN)的最新进展导致了视觉编辑和合成任务的显着进步,并利用了嵌入在预训练的GAN的潜在空间中的丰富语义。但是,现有方法通常是针对特定的GAN体系结构量身定制的,并且仅限于发现不促进局部控制的全球语义方向,或者需要通过手动提供的区域或分段口罩进行某种形式的监督。从这个角度来看,我们提出了一种建筑不足的方法,该方法共同发现代表空间部分及其外观的因素,以一种完全无监督的方式。这些因素是通过在特征图上应用半不合格的张量分解来获得的,这又可以通过像素级控件进行上下文感知的本地图像编辑。此外,我们表明,发现的外观因子对应于无需使用任何标签的显着图。对广泛的GAN体系结构和数据集进行的实验表明,与最新的状态相比,我们的方法在训练时间方面更有效,最重要的是,提供了更准确的局部控制。我们的代码可在以下网址找到:https://github.com/james oldfield/panda。
Recent advances in the understanding of Generative Adversarial Networks (GANs) have led to remarkable progress in visual editing and synthesis tasks, capitalizing on the rich semantics that are embedded in the latent spaces of pre-trained GANs. However, existing methods are often tailored to specific GAN architectures and are limited to either discovering global semantic directions that do not facilitate localized control, or require some form of supervision through manually provided regions or segmentation masks. In this light, we present an architecture-agnostic approach that jointly discovers factors representing spatial parts and their appearances in an entirely unsupervised fashion. These factors are obtained by applying a semi-nonnegative tensor factorization on the feature maps, which in turn enables context-aware local image editing with pixel-level control. In addition, we show that the discovered appearance factors correspond to saliency maps that localize concepts of interest, without using any labels. Experiments on a wide range of GAN architectures and datasets show that, in comparison to the state of the art, our method is far more efficient in terms of training time and, most importantly, provides much more accurate localized control. Our code is available at: https://github.com/james-oldfield/PandA.