Loading [MathJax]/extensions/MathMenu.js
Neural Characteristic Function Learning for Conditional Image Generation | IEEE Conference Publication | IEEE Xplore

Neural Characteristic Function Learning for Conditional Image Generation


Abstract:

The emergence of conditional generative adversarial networks (cGANs) has revolutionised the way we approach and control the generation, by means of adversarially learning...Show More

Abstract:

The emergence of conditional generative adversarial networks (cGANs) has revolutionised the way we approach and control the generation, by means of adversarially learning joint distributions of data and auxiliary information. Despite the success, cGANs have been consistently put under scrutiny due to their ill-posed discrepancy measure between distributions, leading to mode collapse and instability problems in training. To address this issue, we propose a novel conditional characteristic function generative adversarial network (CCF-GAN) to reduce the discrepancy by the characteristic functions (CFs), which is able to learn accurate distance measure of joint distributions under theoretical soundness. More specifically, the difference between CFs is first proved to be complete and optimisation-friendly, for measuring the discrepancy of two joint distributions. To relieve the problem of curse of dimensionality in calculating CF difference, we propose to employ the neural network, namely neural CF (NCF), to efficiently minimise an upper bound of the difference. Based on the NCF, we establish the CCF-GAN framework to explicitly decompose CFs of joint distributions, which allows for learning the data distribution and auxiliary information with classified importance. The experimental results on synthetic and real-world datasets verify the superior performances of our CCF-GAN, on both the generation quality and stability.
Date of Conference: 01-06 October 2023
Date Added to IEEE Xplore: 15 January 2024
ISBN Information:

ISSN Information:

Conference Location: Paris, France
School of Electronic and Information Engineering, Beihang University, Beijing, China
School of Electronic and Information Engineering, Beihang University, Beijing, China
School of Electronic and Information Engineering, Beihang University, Beijing, China
School of Electronic and Information Engineering, Beihang University, Beijing, China
School of Cyber Science and Technology, Beihang University, Beijing, China
School of Electronic and Information Engineering, Beihang University, Beijing, China

1. Introduction

Generative adversarial network (GAN) has been the workhorse in deep generative models since its birth for image generation [16], and its popularity arises from the capability of generating clear and realistic images from merely small dimensions. Despite success, the original architecture of GAN only allows for randomly generating images from Gaussian noise, and an important variant of GANs aims to control the generation by pre-defined auxiliary information (e.g., the class labels or texts), constituting the conditional GAN (cGAN). Taking advantages of the auxiliary information, cGANs have been proved to be capable of enhancing the realistic image generation that is conditioned on extra semantic cues [42], [32], [33]. Therefore, the past few years have witnessed the extensive applications of cGANs, including class-conditioned generation [31], [37], style transfer [55], text-to-image translation [42], [51], to name but a few.

School of Electronic and Information Engineering, Beihang University, Beijing, China
School of Electronic and Information Engineering, Beihang University, Beijing, China
School of Electronic and Information Engineering, Beihang University, Beijing, China
School of Electronic and Information Engineering, Beihang University, Beijing, China
School of Cyber Science and Technology, Beihang University, Beijing, China
School of Electronic and Information Engineering, Beihang University, Beijing, China

Contact IEEE to Subscribe

References

References is not available for this document.