the point-based module to effectively capture the geometry
information for generating the compact neuron skeletoniza-
tion.
2. Related Works
Traditional neuron reconstruction algorithms consist of
three main steps: pre-processing the raw 3D microscopy
image stacks, initializing the tree-like neuronal graph map,
and then pruning the reconstruction map until the com-
pact result is obtained. APP [35] and APP2 [51] cover
all the potential neuron signals on the raw image input for
the initial reconstruction map and remove the surplus neu-
ron branches for a compact structure at the pruning step.
Like the APP family, FMST [52] applies the fast march-
ing algorithm with edge weights to initialize neuron traces
and prunes them based on the coverage ratio of two inter-
sected neuron nodes. NeuTube [16] implements free editing
functions and the multi-branch tracing algorithm from seed
source points. Reversely, Rivulet [54] and Rivulet2 [29]
capture the neuron traces from the furthest branch termini
to the seed point. LCMBoost [20] and SmartTracing [8]
incorporate the deep learning-based modules into an auto-
matic tracing algorithm without human intervention.
With the emergence of 3D U-Net [13] showing great suc-
cess in medical image segmentation tasks, learning-based
segmentation prior to applying the tracing algorithm is able
to highlight the neuron signal and enhance the input neuron
image quality. Some advanced deep learning techniques are
applied to improve the image segmentation performance,
such as inception learning module [26], multi-scale ker-
nel fusion [46], atrous convolution [9], and Atrous Spa-
tial Pyramid Pooling (ASPP) [10, 25]. [43, 41] incorporate
graph reasoning module to the multi-scale encoder-decoder
network for eliminating the semantic gap of image feature
learning. For computational saving and faster inference,
[47] proposes a light-weighted student inference model
guided by the more complex teacher model via knowledge
distillation. To handle the small-size neuron dataset, [44]
improves neuron image segmentation performance through
the VCV-RL module extracting intra- and inter-volume vox-
els of same semantics into the latent space. [39] builds a
GAN-based framework to synthesize neuron training im-
ages from the manually annotated skeletons.
As the deep learning advances in medical image anal-
ysis, researchers have raised the interests to analyze 3D
medical images by applying deep learning techniques. Al-
though the existing works process medical images in voxel-
wise representation, an increasing number of researchers
are studying the 3D structures with the insight of point
clouds. They leverage the 3D point cloud representation to
learn more discriminative object features for different med-
ical image tasks [53], such as cardiac mesh reconstruction
[11], volumetric segmentation [23, 2], and vessel centerline
extraction [22]. For example, [23, 22, 2] use the character-
ization of point clouds to learn the global context feature
for enhancing the CNN-based image segmentation perfor-
mance. Also, [1] and [4] take into account the anatomical
properties of streamline and mesh structure in the form of
point cloud representation.
The great success of introducing point cloud concepts
into the domain of medical image analysis and the fact that
existing tracing methods have not considered the usage of
the point cloud encourage us to address the challenging neu-
ron reconstruction task from a novel perspective. We aim to
improve 3D neuron reconstruction performance through the
powerful geometric and topological representation of point
clouds. Therefore, we shift one of the most challenging
medical image tasks to the scope of point clouds.
3. Method
We propose a novel pipeline, PointNeuron, to perform
3D neuron morphology reconstruction in the point-based
manner. Given the voxel-wise neuron image input, we ini-
tially convert it to a point cloud in Section 3.1. Then we
forward the neuron point cloud into the Skeleton Predic-
tion module to generate a series of neuron skeletal points in
Section 3.2. After that, we design the Connectivity Predic-
tion module to link these skeletal points through analyzing
the node relationships of graph data structure in Section 3.3.
Lastly, we present the specific training losses in Section 3.4.
Our pipeline is shown in Figure 3.
3.1. Voxel-to-Point Transformation
Given a raw volumetric neuron image of size RH×W×D,
a thresholding value θis pre-defined to segment the neuron
structure and remove a majority of noises. Every voxel with
an intensity larger than θis positioned and transformed to
a point. To handle large amount of points, we split all the
neuron points into Kpatches. Hence, the neuron point in-
put can be represented as P=K× {pi: [xi;Ii]}Np
i=1 where
Npis the number of points per patch with the Cartesian co-
ordinate xi∈R3and the intensity Ii∈R.
3.2. Neuron Skeleton Prediction
In this module, we extract Nsskeletal points from the
neuronal point cloud input to constitute a neuron skele-
ton with the point-wise F-dimensional geometric features.
There are three primary steps: learning the deep geometric
features of neuron points through a graph-based encoder,
generating the center proposals at local regions, and pro-
ducing the compact neuron skeleton.
Point cloud geometry learning. Since the point clouds
representing neuron structures are uneven and unordered
in coordinate space, they cannot be simply processed by
a regular grided convolution kernel like typical pixel- or