2
•We analyze the image super-resolution task from the
optimization perspective and develop an ISTA block for
image super-resolution.
•We develop the multi-scale exploration and multi-scale
attention mechanism in the ISTA block, which improves
the network representation and boosts the performance.
•Experimental results show the proposed network
achieves competitive or better performance than other
optimization-based works with much fewer parameters
and lower computation complexity.
II. RELATED WORKS
A. Deep Learning for Image Super-Resolution
Deep learning has demonstrated its amazing performance
on various computer vision tasks. There are numerous con-
volutional neural networks (CNNs) specially designed for
image super-resolution (SR). SRCNN [10] is the first CNN-
based image SR method composed of three convolution layers,
which follows a sparse-coding manner. After SRCNN, deeper
and wider networks has proposed to improve the restoration
performance. FSRCNN [25] increases the network depth and
decreases the input resolution, which makes the method faster
and more effective. VDSR [11] develops a very deep network
with residual connection to restore the high-resolution (HR)
images. EDSR [12] then utilizes the residual blocks in the
network and improves the network capacity. ESPCN [26]
provides a different upsampling strategy to restore the HR
images, which is more effective than the deconvolution op-
eration. Recently, researchers concentrate more on effective
block design for better restoration performance. RDN [13]
combines the residual connection [27] and densely connec-
tion [28], and develops a residual dense block for image SR.
After that, the researchers introduce the residual-in-residual
design with channel attention [29] for image SR and build an
effective network termed RCAN [14]. RFANet [30] expands
the residual connection and aggregates the residual features for
better information transmission. IMDN [15] and RFDN [16]
build the lightweight networks with the help of an information
distillation mechanism. SHSR [31] and MSRN [32] utilize hi-
erarchical exploration to further investigate the image features.
These works usually concentrate on the effective block designs
but neglect to analyze the image SR from the optimization
perspective.
B. Optimization-Inspired Image Super-Resolution
There are also optimization-inspired networks for inter-
pretable image SR. ADMM-Net [23] provides a good example
of dealing with the image restoration problem by the optimiza-
tion strategy and develops a CNN-based denoiser for plug-
and-play restoration. IRCNN [18] then analyzes the image
restoration with the help of the half-quadratic splitting (HQS)
strategy and recovers the image with a CNN-based denoiser
prior. After IRCNN, there are numerous HQS-based methods
for effective image SR. DPSR [20] proposes a different
observation model for image SR and uses kernel estimation
and CNN denoiser for plug-and-play image SR. USRNet [21]
develops an end-to-end network for different image SR tasks.
ISRN [19] devises an effective network for image SR under the
guidance of HQS and maximum likelihood estimation (MLE).
HSRNet [33] also investigates the HQS strategy and develops
a network for aliasing suppression image SR. However, these
works just take the solution as a whole and calculate it directly
by CNNs, without giving an explicit optimization step for each
iteration.
III. METHODOLOGY
In this section, we first analyze the image super-resolution
(SR) from the optimization perspective and propose an itera-
tive solution with the help of ISTA. Then, we introduce the
designed end-to-end network ISTAR. After that, we discuss
the design of the ISTA block. Finally, the network settings are
described in detail.
A. ISTA for Image Super-Resolution
Given an low-resolution (LR) image ILR, the task of image
SR is to find a corresponding image ISR, satisfying
ISR = arg min
ILR ||DISR −ILR||2
`+λ||ILR||1,(1)
where Dis the down-sampling matrix, and λis a weighting
factor. The prior term λ||ILR||1is utilized to introduce the
sparsity of the natural image.
To solve this function, we use ISTA to convert it into an
iterative manner. Then, the solution is
ISR
k+1 =Tλαk(ISR
k−αkDT(DISR
k−ILR)),(2)
where αkis the weighting factor for the k-th iteration and
T(·)is the soft-thresholding operation.
It can be found that the right hand side of Equation 2 has
two independent variables ISR
kand ILR. To make it clear for
understanding, we re-write Equation 2 as
ISR
k+1 =Tλαk((E−αkDTD)ISR
k−αkDTILR),(3)
where Eis the identity matrix.
In Equation 3, we can find that DTILR is shared for every
iteration. In this point of view, we can calculate this term
before the ISTA optimization, and regard it as an invariant to
speed up the optimization.
B. Network Design
Figure 2 shows the entire network design of our ISTAR.
Firstly, the input image ILR is converted into the feature space
by one convolutional layer as
ˆ
ILR =Conv(ILR).(4)
Then, two convolutional layers and one ReLU activation are
utilized to calculate the DTILR for ISTA steps, as shown in
the figure. There are Ksteps for ISTA optimization. For the
k-th step, there is
ˆ
ISR
k=IST ABlock(DTILR,ˆ
ISR
k−1),(5)
where IST ABlock(·)is the designed ISTA block.