
3D Brain and Heart Volume Generative Models: A Survey
YANBIN LIU,Harry Perkins Institute of Medical Research, Department of Computer Science and Software
Engineering, The University of Western Australia, Australia
GIRISH DWIVEDI
∗
,Harry Perkins Institute of Medical Research, The University of Western Australia,
Fiona Stanley Hospital, Australia
FARID BOUSSAID, Department of Electrical, Electronic and Computer Engineering, The University of
Western Australia, Australia
MOHAMMED BENNAMOUN, Department of Computer Science and Software Engineering, The
University of Western Australia, Australia
Generative models such as generative adversarial networks and autoencoders have gained a great deal
of attention in the medical eld due to their excellent data generation capability. This paper provides a
comprehensive survey of generative models for three-dimensional (3D) volumes, focusing on the brain and
heart. A new and elaborate taxonomy of unconditional and conditional generative models is proposed to cover
diverse medical tasks for the brain and heart: unconditional synthesis, classication, conditional synthesis,
segmentation, denoising, detection, and registration. We provide relevant background, examine each task and
also suggest potential future directions. A list of the latest publications will be updated on GitHub to keep up
with the rapid inux of papers at https://github.com/csyanbin/3D-Medical-Generative-Survey.
CCS Concepts: •Computing methodologies →Computer vision;3D imaging.
Additional Key Words and Phrases: generative models, three-dimensional, medical images, brain and heart
1 INTRODUCTION
A wide range of research elds has embraced deep learning (DL) in recent years, including image
processing [
52
,
63
,
64
,
97
], speech recognition [
57
,
68
,
128
], natural language processing [
22
,
30
,
90
,
190
,
203
], and robotics [
149
,
189
]. Thus, the medical imaging community has put in signicant
eorts to take advantage of deep learning advances, and medical imaging research has made
signicant progress with respect to a variety of applications including classication [
14
,
98
,
153
,
154
,
245
], segmentation [
46
,
121
,
137
,
195
,
195
], registration [
160
,
255
], detection [
150
,
151
,
197
],
denoising [
161
,
213
,
214
,
216
,
252
], and synthesis [
34
,
74
,
78
,
111
,
119
], as well as with various
imaging modalities, including Computed Tomography (CT) [
107
,
217
], ultrasound [
117
], Magnetic
Resonance Imaging (MRI) [3,123], and Positron Emission Tomography (PET) [163].
A large number of annotated training images, obtained with the aid of crowd-sourcing annotation
platforms like Amazon Mechanical Turk [
144
], were required for deep learning to be successful in
natural image processing. However, the complexity of collection procedures, the lack of experts,
privacy concerns, and the mandatory requirement of consent from patients make the annotation
process a major bottleneck in medical imaging. In order to mitigate this issue, deep generative
models (e.g., generative adversarial networks (GANs) [
55
] and variational autoencoder (VAE) [
92
])
have been introduced to medical imaging. In these generative models, the original data distribution
∗This work was supported by MRFF Frontier Health and Medical Research - RFRHPI000147.
Authors’ addresses: Yanbin Liu, csyanbin@gmail.com, Harry Perkins Institute of Medical Research, Department of Computer
Science and Software Engineering, The University of Western Australia, Canberra, ACT, Australia, 2601; Girish Dwivedi,
Harry Perkins Institute of Medical Research, The University of Western Australia, Fiona Stanley Hospital, Perth, WA,
Australia, girish.dwivedi@perkins.uwa.edu.au; Farid Boussaid, Department of Electrical, Electronic and Computer Engi-
neering, The University of Western Australia, Perth, WA, Australia, farid.boussaid@uwa.edu.au; Mohammed Bennamoun,
Department of Computer Science and Software Engineering, The University of Western Australia, Perth, WA, Australia,
mohammed.bennamoun@uwa.edu.au.
arXiv:2210.05952v2 [eess.IV] 6 Dec 2023