2
long vector and thus a huge matrix). The idea is therefore to
employ a tensor factorization model to ease the matrix-vector
multiplication of the diagonal matrix and the interim solution
vector. This multiplication can be computed by summing the
elements of the pointwise (Hadamard) product of the two
vectors – the vectorized tensor on the matrix diagonal and
the interim solution vector. This product can be efficiently
computed for various tensor models, but the pointwise multi-
plication of two tensors of rank Rgenerally has rank up to R2,
necessitating conversion back to a rank-Rtensor. The natural
way to do this is via rank-Rleast-squares approximation of
the higher-rank product. This is a hard and generally ill-posed
computational problem, because the best approximation may
not even exist. Thus the method cannot be used with a tensor
rank decomposition model (known as Canonical Polyadic
Decomposition or CPD). The pointwise multiplication and
least-squares approximation are easier with a so-called tensor
train (TT) model. The Hadamard product of two TT models
is another TT model whose wagon (factor) matrices are the
Kronecker products of the respective factor matrices of the two
TT models. This implies that pointwise multiplication of two
TT models has complexity of order NIR4for an INtensor of
TT rank Rfor each wagon. Moreover, rank reduction for TT
models is SVD-based and has complexity order NIR3. Hence,
reducing the rank of the pointwise multiplication (which is R2
due to the Kronecker product) induces a complexity of order
NI(R2)3=NIR6, and as a result, only small TT ranks can
be supported in practice. Another limitation of [5] is that it
is hard to pin down if and when the iteration will converge,
because of the rank reduction step. The numerical experiments
in [5] are interesting but limited in terms of exploring the
accuracy-speedup trade-offs that can be achieved.
Unlike [5], our emphasis is on the tensor rank decom-
position (CPD) model, in good part because many of the
problems that we consider herein admit closed-form tensor
rank decomposition formulations, thus bypassing the need for
computational model fitting and/or rank reduction. Another
issue is that, as we show for the first time in this paper,
the core problem is NP-hard even with a TT model, so
there is no intrinsic computational benefit to using one tensor
decomposition model over the other.
B. Contributions
Our contributions relative to the prior art are as follows:
•We focus on the higher-order tensor version of the
problem, and we analyze its computational complexity.
We show that the problem is easy when the tensor rank
is equal to one, but NP-hard otherwise – even if the rank
is as small as two.
•We consider optimization-based (as opposed to algebraic
or randomized) algorithms that can be used to compute
good approximate solutions in the high-order tensor case.
We provide an equivalent “fluid” (continuous-variable)
reformulation of what is inherently a multi-way selection
problem, and a suite of relatively lightweight gradient-
based approximation algorithms. The continuous refor-
mulation is derived by showing that a certain continuous
relaxation involving a probability distribution instead of
hard selection for each mode is actually tight. This is true
for any rank, i.e., even for full-rank (unstructured) ten-
sor models. The proposed algorithms take advantage of
various optimization frameworks, from projected gradient
descent to Franke-Wolfe and various ways of explicitly
parametrizing the probability simplex constraints. For
low-enough ranks, the associated gradient computations
are computationally lightweight, even for very high-order
tensors.
•We show that our main results remain valid no
matter what kind of polyadic tensor model is used
to represent the tensor of interest, including Tucker,
HOSVD/MLSVD, tensor train, or tensor ring.
•We explore the “span” of the core problem considered,
i.e., the class of problems that can be posed as special in-
stances of computing the minimum or maximum element
of a tensor from its rank decomposition. We show that
this class includes the partition problem (and thus all NP-
complete problems via polynomial-time transformation),
as well as integer least squares, integer linear program-
ming, integer quadratic programming, sign retrieval (a
special kind of mixed integer programming / restricted
version of phase retrieval), maximum likelihood decoding
of parity check codes, and finding the least inconsistent
solution of an overdetermined system of linear equations
over a Galois field.
•Finally, we demonstrate promising experimental results
on a number of hard problems, including better than state-
of-art performance in decoding both low density parity
check codes and general parity check codes, and sign
retrieval.
II. PRELIMINARIES
Consider a matrix M=A1AT
2, where Mis I1×I2,A1
is I1×Rand A2is I2×R, with R≤min(I1, I2)(R <
min(I1, I2)if Mis not full-rank). Three other ways to write
the same relation are M=PR
r=1 A1(:, r)A2(:, r)T, where
A(:, r)stands for the r-th column of A;M=PR
r=1 A1(:
, r)◦A2(:, r), where ◦stands for the vector outer product;
and M(i1, i2) = PR
r=1 A1(i1, r)A2(i2, r), where M(i1, i2)
denotes an element of matrix M, with obvious notation. From
the latter, notice that element (i1, i2)of matrix Mis the inner
product of two row vectors, namely row A1(i1,:) of matrix
A1, and row A2(i2,:) of matrix A2.
We are interested in finding the smallest or largest element
of matrix Mfrom the factors A1and A2. From the latter
relation, we see that seeking the smallest (or largest) element
of Mcorresponds to seeking a pair of row vectors, one drawn
from A1and the other from A2, which have the smallest
(or largest) possible inner product. One obvious way to do
this is to generate all inner products, i.e., the elements of the
product A1AT
2, and find the smallest or largest. This entails
complexity I1I2R, which can be high, even when Ris small
– e.g., consider the case where I1and I2are in the order
of 105or more. Is there a way to avoid performing order
I1I2computations? When R= min(I1, I2), such complexity
is unavoidable, because then the elements of Mcan be