Loading web-font TeX/Math/Italic
On the Effectiveness of Least Squares Generative Adversarial Networks | IEEE Journals & Magazine | IEEE Xplore

On the Effectiveness of Least Squares Generative Adversarial Networks


Abstract:

Unsupervised learning with generative adversarial networks (GANs) has proven to be hugely successful. Regular GANs hypothesize the discriminator as a classifier with the ...Show More

Abstract:

Unsupervised learning with generative adversarial networks (GANs) has proven to be hugely successful. Regular GANs hypothesize the discriminator as a classifier with the sigmoid cross entropy loss function. However, we found that this loss function may lead to the vanishing gradients problem during the learning process. To overcome such a problem, we propose in this paper the Least Squares Generative Adversarial Networks (LSGANs) which adopt the least squares loss for both the discriminator and the generator. We show that minimizing the objective function of LSGAN yields minimizing the Pearson \chi ^2 divergence. We also show that the derived objective function that yields minimizing the Pearson \chi ^2 divergence performs better than the classical one of using least squares for classification. There are two benefits of LSGANs over regular GANs. First, LSGANs are able to generate higher quality images than regular GANs. Second, LSGANs perform more stably during the learning process. For evaluating the image quality, we conduct both qualitative and quantitative experiments, and the experimental results show that LSGANs can generate higher quality images than regular GANs. Furthermore, we evaluate the stability of LSGANs in two groups. One is to compare between LSGANs and regular GANs without gradient penalty. We conduct three experiments, including Gaussian mixture distribution, difficult architectures, and a newly proposed method — datasets with small variability, to illustrate the stability of LSGANs. The other one is to compare between LSGANs with gradient penalty (LSGANs-GP) and WGANs with gradient penalty (WGANs-GP). The experimental results show that LSGANs-GP succeed in training for all the difficult architectures used in WGANs-GP, including 101-layer ResNet.
Published in: IEEE Transactions on Pattern Analysis and Machine Intelligence ( Volume: 41, Issue: 12, 01 December 2019)
Page(s): 2947 - 2960
Date of Publication: 24 September 2018

ISSN Information:

PubMed ID: 30273144

Funding Agency:

References is not available for this document.

1 Introduction

Deep learning has launched a profound reformation and even been applied to many real-world tasks, such as image classification [1], object detection [2], and segmentation [3]. These tasks fall into the scope of supervised learning, which means that a lot of labeled data is provided for the learning processes. Compared with supervised learning, however, unsupervised learning (such as generative models) obtains limited impact from deep learning. Although some deep generative models, e.g., RBM [4], DBM [5], and VAE [6], have been proposed, these models all face the difficulties of intractable functions (e.g., intractable partition function) or intractable inference, which in turn restricts the effectiveness of these models.

Select All
1.
K. He, X. Zhang, S. Ren and J. Sun, "Deep residual learning for image recognition", Proc. IEEE Conf. Comput. Vis. Pattern Recognit., pp. 770-778, 2016.
2.
S. Ren, K. He, R. Girshick and J. Sun, "Faster r-CNN: Towards real-time object detection with region proposal networks", Proc. Advances Neural Inf. Process. Syst., vol. 28, pp. 91-99, 2015.
3.
J. Long, E. Shelhamer and T. Darrell, "Fully convolutional networks for semantic segmentation", Proc. IEEE Conf. Comput. Vis. Pattern Recognit., pp. 3431-3440, 2015.
4.
G. Hinton and R. Salakhutdinov, "Reducing the dimensionality of data with neural networks", Sci., vol. 313, no. 5786, pp. 504-507, 2006.
5.
R. Salakhutdinov and G. Hinton, "Deep Boltzmann machines", Proc. Int. Conf. Artif. Intell. Statist., vol. 5, pp. 448-455, 2009.
6.
D. P. Kingma and M. Welling, "Auto-encoding variational bayes", CoRR, 2013.
7.
I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, et al., "Generative adversarial nets", Proc. Advances Neural Inf. Process. Syst., pp. 2672-2680, 2014.
8.
I. Goodfellow, Y. Bengio and A. Courville, Deep Learning, Cambridge, MA, USA:MIT Press, 2016.
9.
A. Nguyen, J. Yosinski, Y. Bengio, A. Dosovitskiy and J. Clune, "Plug play generative networks: Conditional iterative generation of images in latent space", Proc. Comput. Vis. Pattern Recognit., pp. 4467-4477, 2017.
10.
X. Chen, Y. Duan, R. Houthooft, J. Schulman, I. Sutskever and P. Abbeel, "Infogan: Interpretable representation learning by information maximizing generative adversarial nets", Proc. Advances Neural Inf. Process. Syst., pp. 2172-2180, 2016.
11.
C. Ledig, L. Theis, F. Huszar, J. Caballero, A. Cunningham, A. Acosta, et al., "Photo-realistic single image super-resolution using a generative adversarial network", Proc. Comput. Vis. Pattern Recognit., pp. 4681-4690, 2017.
12.
T. Salimans, I. Goodfellow, W. Zaremba, V. Cheung, A. Radford, X. Chen, et al., "Improved techniques for training gans", Proc. Advances Neural Inf. Process. Syst., pp. 2226-2234, 2016.
13.
A. Radford, L. Metz and S. Chintala, "Unsupervised representation learning with deep convolutional generative adversarial networks", CoRR.
14.
L. Metz, B. Poole, D. Pfau and J. Sohl-Dickstein, "Unrolled generative adversarial networks", CoRR, [online] Available: http://arxiv.org/abs/1611.02163.
15.
M. Arjovsky, S. Chintala and L. Bottou, "Wasserstein gan", Proc. 33rd Int. Conf. Mach. Learn., pp. 214-223, 2017.
16.
G.-J. Qi, "Loss-sensitive generative adversarial networks on lipschitz densities", CoRR, [online] Available: http://arxiv.org/abs/1701.06264.
17.
C. M. Bishop, Pattern Recognition and Machine Learning, Berlin, Germany:Springer, 2006.
18.
N. Kodali, J. D. Abernethy, J. Hays and Z. Kira, "On convergence and stability of gans", CoRR, [online] Available: http://arxiv.org/abs/1705.07215.
19.
I. Gulrajani, F. Ahmed, M. Arjovsky, V. Dumoulin and A. Courville, "Improved training of wasserstein gans", Adv. Neural Inf. Process. Syst. (NIPS), pp. 5767-5777, 2017.
20.
X. Mao, Q. Li, H. Xie, R. Y. Lau, Z. Wang and S. P. Smolley, "Least squares generative adversarial networks", Proc. Int. Conf. Comput. Vis., pp. 2813-2821, 2017.
21.
G. W. Taylor, R. Fergus, Y. LeCun and C. Bregler, "Convolutional learning of spatio-temporal features", Proc. Eur. Conf. Comput. Vis., pp. 140-153, 2010.
22.
G. E. Hinton and R. R. Salakhutdinov, "Replicated softmax: An undirected topic model", Proc. Advances Neural Inf. Process. Syst., pp. 1607-1614, 2009.
23.
G. E. Hinton, S. Osindero and Y.-W. Teh, "A fast learning algorithm for deep belief nets", Neural Comput., vol. 18, no. 7, pp. 1527-1554, Jul. 2006.
24.
Y. Ganin, E. Ustinova, H. Ajakan, P. Germain, H. Larochelle, F. Laviolette, et al., "Domain-adversarial training of neural networks", J. Mach. Learn. Res., vol. 17, no. 1, pp. 2096-2030, 2016.
25.
S. Reed, Z. Akata, X. Yan, L. Logeswaran, B. Schiele and H. Lee, "Generative adversarial text-to-image synthesis", Proc. 33rd Int. Conf. Mach. Learn., pp. 1060-1069, 2016.
26.
P. Isola, J.-Y. Zhu, T. Zhou and A. A. Efros, "Image-to-image translation with conditional adversarial networks", Comput. Vis. Pattern Recognit., pp. 5967-5976, 2017.
27.
M. Mirza and S. Osindero, "Conditional Generative Adversarial Nets", CoRR, [online] Available: http://arxiv.org/abs/1411.1784.
28.
J. Donahue, P. Krähenbühl and T. Darrell, "Adversarial feature learning", CoRR, [online] Available: http://arxiv.org/abs/1605.09782.
29.
V. Dumoulin, I. Belghazi, B. Poole, O. Mastropietro, A. Lamb, M. Arjovsky, et al., "Adversarially learned inference", CoRR, [online] Available: http://arxiv.org/abs/1606.00704.
30.
C. Li, H. Liu, C. Chen, Y. Pu, L. Chen, R. Henao, et al., "Alice: Towards understanding adversarial learning for joint distribution matching", Adv. Neural Inf. Process. Syst. (NIPS), pp. 5495-5503, 2017.

Contact IEEE to Subscribe

References

References is not available for this document.