
of [
72
] consider minimizing the linear least-squares objective function over a convex set. They show
that a reliable estimation of the signal can be obtained by such a simple method despite the unknown
nonlinear link function.
As mentioned earlier, MPR with sparse priors has been studied in [
64
,
98
]. The work [
64
] implements
a two-step procedure based on convex optimization, with the first step being a spectral initialization
method. More specifically, a semidefinite program is solved to produce an initial vector, and then an
`1
regularized least square is solved to obtain a refined estimator. Such a procedure suffers from high
computational costs. A more efficient two-step approach, which is a simple variant of the thresholded
Wirtinger flow method [
10
], is proposed in [
98
]. In the first step, identical to that in [
10
], the initial
vector is calculated by a thresholded spectral method that first estimates the support of the sparse
signal by thresholding and then performs the classic power method over the submatrix corresponding
to the estimated support. In the second step, a thresholded gradient descent algorithm is employed.
Both approaches in [
64
,
98
] can attain the optimal statistical rate of order
p(slog n)/m
, provided
that the sensing vector is standard Gaussian and the number of samples
m= Ωs2log n
. In addition,
the second step of the approach proposed in [98] is shown to achieve a linear convergence rate.
PR and SIM with generative priors
: PR with generative priors has been studied in [
31
,
37
,
38
,
80
,
3
,
47
]. More specifically, an approximate message passing algorithm is proposed in [
3
]. The authors
of [
31
,
80
] minimize the objective over the latent space in
Rk
using gradient descent, where
k
is
the latent dimension of the generative prior. Corresponding algorithms may easily get stuck in local
minima since the explorable solution space is limited. Recovery guarantees for projected gradient
descent algorithms over the ambient space in
Rn
for noiseless PR with pre-trained or untrained neural
network priors have been proposed in [
37
,
38
]. No initialization methods have been proposed in these
works, making the assumption on the initial vector therein a bit stringent. On the other hand, the
authors of [
47
] propose a spectral initialization method for PR with generative priors and provide
recovery guarantees with respect to globally optimal solutions of a corresponding optimization
problem. The optimization problem is non-convex, and a projected power method is proposed in [
50
]
to approximately find an optimal solution.
Generative model based SIMs have been studied in [
51
,
49
,
46
,
94
]. The authors of [
51
,
49
,
46
]
provide optimal sample complexity upper bounds under the assumption of Gaussian sensing vectors.
But their results rely on the assumption
(4)
, which is not satisfied by widely adopted PR models. The
SIM studied in [
94
] encompasses certain PR models as special cases, and the sensing vector can
be non-Gaussian. However, the nonlinear link function
f
is assumed to be differentiable, making
it not directly applicable to the PR model in
(2)
. Moreover, it is worth mentioning that in both
works [
94
,
51
], the recovery guarantees are with respect to globally optimal solutions of typically
highly non-convex optimization problems. Attaining these optimal solutions is practically difficult.
1.2 Contributions
Throughout this paper, we assume that the signal
x
lies in the range of an
L
-Lipschitz continuous
generative model with bounded k-dimensional inputs. Our main contributions are as follows:
•
We propose a two-step approach for MPR with generative priors. In particular, in the first
step, we make use of the projected power method proposed in [
50
] to obtain a good initial
vector for the iterative algorithm used in the second step.
•
We show that under appropriate initialization, both steps attain a statistical rate of order
p(klog L)·(log m)/m
, and the second step achieves a linear convergence rate. The
initialization condition for the first step is mild in the sense that it allows the inner product
between the starting point and the signal
x
to be a sufficiently small positive constant. In
contrary, the initialization condition for the second step is more restrictive, making the
first step necessary. Notably, unlike for the works on MPR with sparse priors [
64
,
98
] that
require the sub-optimal
Os2log n
sample complexity, the sample complexity requirement
for our recovery guarantees is
O((klog L)·(log m))
, which is naturally conjectured to be
near-optimal [52, 42].
•
We perform numerical experiments on image datasets to corroborate our theoretical results.
In particular, for the noisy magnitude-only measurement model
(2)
, we observe that our
approach gives reconstructions that are competitive with those of the alternating phase
projected gradient descent (APPGD) algorithm proposed in [
37
], which is the corresponding
3