1. Introduction
Generative Adversarial Networks (GANs) are capable of generating high-resolution, photorealistic images [24], [25], [26]. However, these GANs are often confined to two dimensions because of a lack of photorealistic 3D training data; therefore, they cannot support tasks such as synthesizing multiple views of a single object. 3D-aware image synthesis offers to learn neural scene representations unsupervised from 2D images. The learned representations can be used to render view-consistent images from new camera poses [41], [54], [18].