论文标题
广义对抗性的推论
Generalized Adversarially Learned Inference
论文作者
论文摘要
允许在训练gan的同时有效推断潜在向量可以大大提高其在各种下游任务中的适用性。最近的方法,例如Ali和Bigan框架,通过对抗训练图像发生器以及编码器,以匹配图像和潜在向量对的两个联合分布,从而开发了gan中潜在变量的推理方法。我们概括了这些方法,以基于对所需解决方案的先前或学习的知识来结合有关重建,自学和其他形式的监督的多层反馈。我们通过修改歧视者的目标来正确识别由图像,潜在向量和通过辅助任务产生的其他变量组成的任意数量的随机变量的两个以上的联合分布,例如重建和内置或作为成分,或作为适当训练的模型的输出。我们为生成器编码对设计了一个不饱和的最大化目标,并证明所得的对抗游戏对应于同时匹配所有分布的全局最佳距离。在我们提出的框架内,我们引入了一组新型技术,用于根据属性为模型提供自我监督的反馈,例如贴片级对应关系和重建的循环一致性。通过全面的实验,我们证明了针对各种任务的拟议方法的功效,可伸缩性和灵活性。
Allowing effective inference of latent vectors while training GANs can greatly increase their applicability in various downstream tasks. Recent approaches, such as ALI and BiGAN frameworks, develop methods of inference of latent variables in GANs by adversarially training an image generator along with an encoder to match two joint distributions of image and latent vector pairs. We generalize these approaches to incorporate multiple layers of feedback on reconstructions, self-supervision, and other forms of supervision based on prior or learned knowledge about the desired solutions. We achieve this by modifying the discriminator's objective to correctly identify more than two joint distributions of tuples of an arbitrary number of random variables consisting of images, latent vectors, and other variables generated through auxiliary tasks, such as reconstruction and inpainting or as outputs of suitable pre-trained models. We design a non-saturating maximization objective for the generator-encoder pair and prove that the resulting adversarial game corresponds to a global optimum that simultaneously matches all the distributions. Within our proposed framework, we introduce a novel set of techniques for providing self-supervised feedback to the model based on properties, such as patch-level correspondence and cycle consistency of reconstructions. Through comprehensive experiments, we demonstrate the efficacy, scalability, and flexibility of the proposed approach for a variety of tasks.