论文标题

关于二元性差距,作为监测gan训练的措施

On Duality Gap as a Measure for Monitoring GAN Training

论文作者

Sidheekh, Sahil, Aimen, Aroof, Madan, Vineet, Krishnan, Narayanan C.

论文摘要

生成对抗网络(GAN)是学习复杂数据分布的最流行的深度学习模型之一。但是,众所周知,培训gan是一项具有挑战性的任务。这通常归因于训练进度与发电机损失的轨迹之间缺乏相关性以及对GAN主观评估的需求。最近提出的受游戏理论启发的措施 - 双重性差距,旨在弥合这一差距。但是,正如我们所证明的那样,由于其估计过程构成的局限性,二元性差距的能力仍然受到限制。本文介绍了对这一限制的理论理解,并提出了二元性差距的更可靠的估计过程。我们方法的症结在于,本地扰动可以有效地帮助零和游戏中的代理,从而有效地逃脱了非纳什鞍点。通过跨GAN模型和数据集进行的详尽实验,我们确定了方法在捕获GAN训练进度的效力,而对计算复杂性的增长最小。此外,我们表明,我们的估计能够识别模型收敛/差异的能力,是一种潜在的性能度量,可用于调整GAN的超参数。

Generative adversarial network (GAN) is among the most popular deep learning models for learning complex data distributions. However, training a GAN is known to be a challenging task. This is often attributed to the lack of correlation between the training progress and the trajectory of the generator and discriminator losses and the need for the GAN's subjective evaluation. A recently proposed measure inspired by game theory - the duality gap, aims to bridge this gap. However, as we demonstrate, the duality gap's capability remains constrained due to limitations posed by its estimation process. This paper presents a theoretical understanding of this limitation and proposes a more dependable estimation process for the duality gap. At the crux of our approach is the idea that local perturbations can help agents in a zero-sum game escape non-Nash saddle points efficiently. Through exhaustive experimentation across GAN models and datasets, we establish the efficacy of our approach in capturing the GAN training progress with minimal increase to the computational complexity. Further, we show that our estimate, with its ability to identify model convergence/divergence, is a potential performance measure that can be used to tune the hyperparameters of a GAN.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源