论文标题
从人工数据集的gan培训中学到的教训
Lessons Learned from the Training of GANs on Artificial Datasets
论文作者
论文摘要
近年来,生成的对抗网络(GAN)在综合逼真的图像方面取得了长足的进步。但是,它们通常在图像数据集上进行培训,该数据集的样本太少或属于不同数据分布的类太多。因此,甘恩容易不适合或过度拟合,从而使它们的分析变得困难和受到限制。因此,为了对gan进行彻底的研究,同时避免了数据集引入的不必要的干扰,我们在人工数据集中训练它们,那里有很多样本,而真实的数据分布是简单,高维的,并且具有结构化的歧管。此外,生成器的设计使得存在最佳参数集。从经验上讲,我们发现在各种距离度量下,发电机无法通过GAN训练程序学习此类参数。我们还发现,与在模型复杂性足够高时增加网络深度或宽度相比,gan的训练混合物会导致更大的性能增长。我们的实验结果表明,发电机的混合物可以在无监督的设置中自动发现不同的模式或不同类别,我们将其归因于生成器和歧视器之间的分布和歧视任务。作为我们结论对现实数据集的普遍性的一个例子,我们在CIFAR-10数据集上训练GAN的混合物,而我们的方法在流行的指标(即Inception Sercep(IS)和Fréchet-Inception距离(FID)方面,我们的方法显着优于最先进的指标。
Generative Adversarial Networks (GANs) have made great progress in synthesizing realistic images in recent years. However, they are often trained on image datasets with either too few samples or too many classes belonging to different data distributions. Consequently, GANs are prone to underfitting or overfitting, making the analysis of them difficult and constrained. Therefore, in order to conduct a thorough study on GANs while obviating unnecessary interferences introduced by the datasets, we train them on artificial datasets where there are infinitely many samples and the real data distributions are simple, high-dimensional and have structured manifolds. Moreover, the generators are designed such that optimal sets of parameters exist. Empirically, we find that under various distance measures, the generator fails to learn such parameters with the GAN training procedure. We also find that training mixtures of GANs leads to more performance gain compared to increasing the network depth or width when the model complexity is high enough. Our experimental results demonstrate that a mixture of generators can discover different modes or different classes automatically in an unsupervised setting, which we attribute to the distribution of the generation and discrimination tasks across multiple generators and discriminators. As an example of the generalizability of our conclusions to realistic datasets, we train a mixture of GANs on the CIFAR-10 dataset and our method significantly outperforms the state-of-the-art in terms of popular metrics, i.e., Inception Score (IS) and Fréchet Inception Distance (FID).