
1
A Survey on Deep Generative 3D-aware Image Synthesis
WEIHAO XIA and JING-HAO XUE, University College London, UK
Recent years have seen remarkable progress in deep learning powered visual content creation. This includes
deep generative 3D-aware image synthesis, which produces high-delity images in a 3D-consistent manner
while simultaneously capturing compact surfaces of objects from pure image collections without the need
for any 3D supervision, thus bridging the gap between 2D imagery and 3D reality. The eld of computer
vision has been recently captivated by the task of deep generative 3D-aware image synthesis, with hundreds
of papers appearing in top-tier journals and conferences over the past few years (mainly the past two years),
but there lacks a comprehensive survey of this remarkable and swift progress. Our survey aims to introduce
new researchers to this topic, provide a useful reference for related works, and stimulate future research
directions through our discussion section. Apart from the presented papers, we aim to constantly update the
latest relevant papers along with corresponding implementations at https://weihaox.github.io/3D-aware-Gen.
CCS Concepts: •General and reference
→
Surveys and overviews;•Computing methodologies
→
Machine learning;Computer vision;Image manipulation.
Additional Key Words and Phrases: 3D-aware image synthesis, deep generative models, implicit neural
representation, generative adversarial network, diusion probabilistic models
ACM Reference Format:
Weihao Xia and Jing-Hao Xue. 2023. A Survey on Deep Generative 3D-aware Image Synthesis. ACM Comput.
Surv. 1, 1, Article 1 (January 2023), 34 pages. https://doi.org/10.1145/3626193
1 INTRODUCTION
A tremendous amount of progress has been made in deep neural networks that lead to photorealistic
image synthesis. Despite achieving compelling results, most approaches focus on two-dimensional
(2D) images, overlooking the three-dimensional (3D) nature of the physical world. The lack of 3D
structure, therefore, inevitably limits some of their practical applications. Recent studies have thus
proposed generative models that are 3D-aware. That is, they incorporate 3D information into the
generative models to enhance control (especially in terms of multiconsistency) over the generated
images. Examples depicted in Fig. 1 elucidate that the objective is to produce high-quality renderings
which maintain consistency across various views. In contrast to the 2D generative models, the
recently developed 3D-aware generative models [
13
,
33
] bridge the gap between 2D images and 3D
physical world. The physical world surrounding us is intrinsically three-dimensional and images
depict reality under certain conditions of geometry, material, and illumination, making it natural
to model the image generation process in 3D spaces. As shown in Fig. 2, classical rendering (a)
renders images at certain camera positions given human-designed or scanned 3D shape models;
inverse rendering (b) recovers the underlying intrinsic properties of the 3D physical world from
2D images; 2D image generation (c) is mostly driven by generative models, which have achieved
impressive results in photorealistic image synthesis; and 3D-aware generative models (d) oers the
Authors’ address: Weihao Xia, weihao.xia.21@ucl.ac.uk; Jing-Hao Xue, jinghao.xue@ucl.ac.uk, Department of Statistical
Science, University College London, London, UK.
Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee
provided that copies are not made or distributed for prot or commercial advantage and that copies bear this notice and
the full citation on the rst page. Copyrights for third-party components of this work must be honored. For all other uses,
contact the owner/author(s).
©2023 Copyright held by the owner/author(s).
0360-0300/2023/1-ART1
https://doi.org/10.1145/3626193
ACM Comput. Surv., Vol. 1, No. 1, Article 1. Publication date: January 2023.
arXiv:2210.14267v3 [cs.CV] 2 Oct 2023