2
additional time at inference stage. Thus it may retard some
real-time SOD models.
In order to handle the aforementioned limitations, this paper
proposes a learnable noise against adversarial attacks of SOD
networks. The learnable noise consists of a shallow noise and
a noise estimation. Different from SWS that introduces noise
directly to the input image, the shallow noise inserts a noisy
layer between stem and stage 1 of the backbone. It introduces
noise in feature-level, thereby the noise is learnable and able to
balance between learning clean images and adversarial images.
Inspired from the image denoising method [14], we propose a
lightweight noise estimation component to refine the feature of
adversarial images. Our shallow noise and noise estimation are
embedded in the encoder and decoder respectively, allowing
parallel computation. Furthermore, the noise estimation only
affects one channel of the decoder. Consequently, our defense
method introduces much less extra time and performs better
than ROSA, see Fig. 1(b) and (c).
Our main contributions can be summarized as three-folds:
1) We launch adversarial attacks on both state-of-the-art
RGB and RGB-D SOD models successfully. Experi-
mental results verify that a wide range of existing SOD
models are sensitive to adversarial perturbations.
2) We propose a simple but efficient learnable noise (LeNo)
which hardly modifies the original SOD network struc-
ture. It consists of a plug-and-play shallow noise and
noise estimation. It is parallel computing and hardly
influences the inference speed.
3) With the deeply-supervised noise-decoupled training
scheme, the proposed defense method promotes adver-
sarial robustness of extensive RGB and RGB-D SOD
networks. The experimental results show that our pro-
posed defense method outperforms previous works not
only on adversarial images but also clean images.
II. RELATED WORKS
A. Salient Object Detection
An impressive mechanism of human vision system is the
internal process that quickly scans the global image to obtain
region of interest. In the field of computer vision, this task
is referred to as Salient Object Detection. It plays a key role
in a range of real-world applications, such as medical image
segmentation [8], [39], camouflaged object detection [7], etc.
Although significant progress has been made in the past several
years [23], [35], [44], there is still room for improvement
when faced with challenging factors, such as complicated
backgrounds or varying lighting conditions in the scenes. One
way to overcome such challenges is to employ depth maps,
which provides complementary spatial information and have
become easier to capture due to the ready availability of depth
sensors. Recently, RGB-D based salient object detection has
gained increasing attention, and various methods have been
developed [3], [9]. Early RGB-D based salient object detection
models tended to extract handcrafted features and then fused
the RGB image and depth map. Despite the effectiveness
of traditional methods using handcrafted features, their low-
level features tend to limit generalization ability, and they
lack the high-level reasoning required for complex scenes. To
address these limitations, several deep learning based RGB-
D salient object detection methods [10] have been developed,
with improved performance.
B. Adversarial Attacks
Existing adversarial attacks consist of several groups,
one-step gradient-based methods; iterative methods [6];
optimization-based methods [42]; and generative networks
[34], [46] based methods.
1) FGSM: Fast Gradient Sign Method (FGSM) [13] is an
efficient single-step adversarial attack method. Given vector-
ized input xand corresponding target label y, FGSM alters
each element of xalong the direction of its gradient w.r.t the
inference loss ∂L/∂x. The generation of adversarial example
ˆx(i.e., perturbed input) can be described as:
ˆx=x+·sgn (∇xL(g(x;θ), y)) ,(1)
where is the perturbation constraint that determines the attack
strength. g(x;θ)computes the output of DNN paramterized by
θ.sgn (·)is the sign function.
2) PGD: Projected Gradient Descent (PGD) [29] is a multi-
step variant of FGSM, which is one of the strongest L∞
adversarial example generation algorithm. With ˆxt=1 =xas
the initialization, the iterative update of perturbed data ˆxin
iteration t can be expressed as:
ˆxt= ΠP(x)(ˆxt−1+a·sgn (∇xL(g(ˆxt−1;θ), y))) ,(2)
where P(x)is the projection space which is bounded by x±,
and ais the step size. [29] also propose that PGD is a universal
adversary among all the first-order adversaries.
3) ROSA: ROSA [21] is an iterative gradient-based
pipeline. It is the first adversarial attack on the state-of-the-
art salient object detection models. They try to make the
predictions of all pixels in xgo wrong. In each iteration t,
supposing that adversarial sample ˆxfrom previous time step
or initialization is prepared, the adversarial sample is updated
as:
ˆx0=x, ˆxt+1 = ˆxt+pt,(3)
p0
t=X
i∈St
[∇ˆxtgi,1−yi(ˆxt;θ)− ∇ˆxtgi,yi(ˆxt;θ)] .(4)
Here, ptdenotes the adversarial perturbation computed at t-th
step, it is obtained by normalization as α·p0
t/kp0
tk∞where α
is a fixed step length, i denotes one pixel in x,Stdenotes the
set of pixels that gcan still classify correctly and yidenotes
two categories: salient and nonsalient.
C. Defenses Against Adversarial Attacks
Many researchers resort to randomization schemes [5], [27]
for mitigating the effects of adversarial perturbations in the
input/feature domain. The intuition behind this type of defense
is that DNNs are always robust to random perturbations.
A randomization based defense attempts to randomize the
adversarial effects into random effects, which are not a concern
for most DNNs. Some of them also add noise to the network
as we do, but their noise is random and not learnable, like
ROSA.
Previous works in feature denoising [25] attempts to al-
leviate the effects of adversarial perturbations on high-level