
2
TABLE I: Summary of datasets used for LiDAR-based point cloud object detection
Dataset Year Real/Simulated Frames BBoxes Classes Corruptions Corruption
Severities
Robustness
Metric
KITTI [6] 2012 real 15K 200K 8 cutout, noise 2 -
NuScenes [7] 2019 real 400K 1.4M 23 rain, sun, clouds, cutout, various vehicle types, noise 2 -
Waymo [8] 2019 real 200K 12M 4 rain, fog, cutout, dust, various vehicle types, noise 2 -
Boreas [15] 2022 real 7.1K 320K 4 snow, rain, sun, clouds, cutout, noise 2 -
STF [10] 2020 real 13.5K 100K 4 fog, rain, snow, cutout, noise 3 -
CADC [14] 2020 real 7K 334K 10 snow, bright light, cutout, noise 5 -
ModelNet40-C [11] 2022 real+simulated 185K - 40
occlusion, LiDAR, local_density_inc/dec, cutout,
uniform, Gaussian, impulse, upsampling, background,
rotation, shear, FFD, RBF, inv_RBF
6X
ModelNet-C [18] 2022 real+simulated 185K - 40 scale, rotate, jitter, drop_global/local, add_global/local 6X
Argoverse [19] 2019 real 468K 993K 15 rain, cutout, dust, noise 2 -
Lyft Level 5 [20] 2020 real 30K 1.3M 9 rain, cutout, noise 2 -
Ours 2022 real+simulated 1.1M 15M 8
Scene: rain, snow, fog, uniform_rad, gaussian_rad,
impulse_rad, upsample, background, cutout, beam_del,
local_dec/inc, layer_del; Object: uniform, gaussian,
impulse, upsample, cutout, local_dec/inc, shear, scale,
rotation, FFD, translation
6X
raw 3D scenes (i.e., clean point clouds) from [
6
] and build
large-scale corrupted datasets by adding
25
corruptions with
6
different severity levels to each clean point cloud. Finally, we
obtain a total of 1,122,150 examples covering 7,481 scenes,
25 common corruption types, and 6 severity levels. Compared
with real-world data benchmark (see Table I), the proposed
benchmark synthesized more examples for benchmarking
robustness. Compared with other synthesized benchmark (see
Table I), our benchmark provides more types of corruption
patterns to specifically support benchmarking object detection.
Note that, we conduct extensive experiments to quantitatively
validate the effectiveness of simulation methods by evaluating
the naturalness of synthesized data.
With such a novel benchmark, we investigate the robustness
of current point cloud detectors by conducting extensive
empirical studies on 8 existing detectors, covering 3 different
representations and 2 different proposal architectures. In
particular, we study the following four research questions to
identify the challenges and potential opportunities for building
robust point cloud detectors:
•How do the common corruption patterns affect the point
cloud detector’s performance?
Given overall common
corruptions, an accuracy drop of
11.01%
(on average) on all
detectors anticipates a noticeable accuracy drop of detectors
against diverse corruption patterns.
•How does the design of a point cloud detector affect its
robustness against corruption patterns?
Compared with
two-stage detectors, one-stage detectors perform more robust
against a majority of corruptions. Compared with point-based
detectors, voxel-involving detectors perform more robust
against the most of corruptions.
•What kind of detection bugs exist in point cloud detec-
tors against common corruption patterns?
Followed by
the decrease in the rate of true detection, common corruptions
widely trigger a number of false detections on all point cloud
detectors.
•How do the robustness enhancement techniques im-
prove point cloud detectors against common corruption
patterns?
Even with the help of data augmentation and
denoising, common corruptions still cause a severe accuracy
drop of over 10% on detection.
In summary, this work makes the following contributions:
•
We design physical-aware simulation methods covering
25 common corruptions related to natural weather, noise
disturbance, density change, and object transformations at
the object and scene level.
•
We create the first robustness benchmark of point cloud
detection against common corruptions.
•
Based on the benchmark, we conduct extensive empirical
studies to evaluate the robustness of 8 existing detectors
to reveal the vulnerabilities of the detectors under common
corruptions.
•
We study the existing data augmentation (DA) method and
denoising method’s performance on robustness enhancement
for point cloud detection and further discuss their limitations.
II. RELATED WORK
A. LiDAR Perception
LiDAR perception is sensitive to both internal and external
factors that could result in different corruptions. Adversarial
weather [
9
] (e.g., snow, rain, and fog) can dim or even block
transmissions of lasers by dense liquid or solid droplets. Regard-
ing noise characteristics of point clouds, strong illumination
[
22
] affects the signal transmission by lowering Signal-to-Noise
Ratio (SNR), increasing the noise level of LiDAR ranging [
23
].
Besides, the intrinsically inaccurately ranging and the sensor
vibration [
24
], [
25
] potentially trigger noisy observations during
LiDAR scanning. Environmental floating particles (e.g., dust
[
26
]) could perturb point cloud with the background noise.
Density distribution of LiDAR-based point clouds can also
easily affect autonomous driving. For instance, common object-
object occlusions block LiDAR scanning on objects in the
scene [
13
]. Besides, the dark-color cover and rough surface