
Minasyan, Galstyan, Hunanyan, Dalalyan
algorithms, based on fast approximate methods (see e.g.
Malkov and Yashunin [2020], Wang et al. [2018], Harwood
and Drummond [2016], Jiang et al. [2016]). Another di-
rection is to improve the matching quality by considering
alternative local descriptors [Rublee et al.,2011,Chen et al.,
2010,Calonder et al.,2010] for given keypoints. The choice
of keypoints is considered in Tian et al. [2020], Bai et al.
[2020].
The minimum cost flow problem was first studied in the
context of the Hungarian algorithm [Kuhn,2012] and the
assignment problem, which is a special case of minimum
cost flow on bipartite graphs with all edges having unit ca-
pacity. Generalization of Hungarian algorithm for graphs
with arbitrary edge costs guarantees
O((n+F)m)
time
complexity, where
n
is the number of nodes in the graph,
m
is the number of edges and
F
is the total flow sent through
the graph. There have also been other algorithms with sim-
ilar complexity guarantees [Fulkerson,1961,Ahuja et al.,
1992]. Since then many algorithms have been proposed for
solving minimum cost flow problems in strongly polynomial
time [Orlin et al.,1993,Orlin,1993,1996,Goldberg and
Tarjan,1989,Galil and Tardos,1988] with the fastest run-
time of around
O(nm)
. Recent advances for solving MCF
problems have been proposed in Goldberg et al. [2015] and
Chen et al. [2022a]. The latter proposes an algorithm with
an almost-linear computational time.
Permutation estimation and related problems have been re-
cently investigated in different contexts such as statistical
seriation [Flammarion et al.,2019,Giraud et al.,2021,Cai
and Ma,2022], noisy sorting [Mao et al.,2018], regression
with shuffled data [Pananjady et al.,2017,Slawski and Ben-
David,2019], isotonic regression and matrices [Mao et al.,
2020,Pananjady and Samworth,2020,Ma et al.,2020],
crowd labeling [Shah et al.,2021], recovery of general dis-
crete structure [Gao and Zhang,2019], and multitarget track-
ing [Chertkov et al.,2010,Kunisky and Niles-Weed,2022].
3 MAIN THEORETICAL RESULT
This section contains the main theoretical contribution of
the present work. In order to be able to recover
S∗
and the
matching map
π∗
, the key ingredient we use is the max-
imization of the profile likelihood. This corresponds to
looking for the least sum of squares (LSS) of errors over
all injective mappings defined on a subset of
[n]
of size
k
.
Formally, if we define
Pk:= π:S→[m]such that S⊂[n],|S|=k,
πis injective
to be the set of all
k
-matching maps, we can define the
procedure k-LSS as a solution to the optimization problem
bπLSS
k∈arg min
π∈PkX
i∈SπkXi−X#
π(i)k2
2,(3)
where
Sπ
denotes the support of function
π
. In the particular
case of
k∗=n
, the optimization above is conducted over
all the injective mappings from
[n]
to
[m]
. This coincides
with the LSS method from [Galstyan et al.,2022].
Let b
Φ(k)be the error of bπLSS
k, that is
b
Φ(k) = min
π∈PkXi∈SπkXi−X#
π(i)k2
2.
For some values of tuning parameters
λ > 0
and
γ > 0
, as
well as for some kmin ∈[n], initialize k←kmin and
1. Compute b
Φ(k)and b
Φ(k+ 1).
2. Set ¯σ2
k=b
Φ(k)/(kd).
3. If k=nor b
Φ(k+ 1) −b
Φ(k)>d+λ
1−γ¯σ2
k,
then output (k, ¯σk,bπLSS
k).
4. Otherwise, increase k←k+ 1 and go to Step 1.
In the sequel, we denote by
(b
k, ¯σb
k,bπLSS
b
k)
the output of this
procedure. Notice that we start with the value of k=kmin,
which in the absence of any information on the number of
inliers might be set to
k= 1
. However, using a higher
value of
kmin
might considerably speed up the procedure
and improve its quality.
For appropriately chosen values of
γ
and
λ
, as stated in the
next theorem, the described procedure outputs the correct
values of k∗and π∗with high probability.
Theorem 1.
Let
α∈(0,1)
and
λn,m,d,α
be defined by
(2)
. If
¯κall >(5
/4)λn,m,d,α
, then the output
(b
k, bπLSS
b
k)
of the model selection algorithm with parameters
λ=
(1
/4)λ2
n,m,d,α, γ =λ
/dsatisfies P(bπLSS
b
k=π∗)≥1−α.
Since the condition on the separation distance
¯κall
compared
to the case of known
k∗
is different by only a slightly larger
constant, from the perspective of statistical accuracy, the
case of unknown
k∗
is not more challenging than that of the
known k∗.
In the sequel, without much loss of generality, we assume
that the sizes of
X
and
X#
are equal, i.e.,
n=m
. Indeed,
in the case,
m>n
one can add
m−n
points arbitrarily far
from the rest of the points to the smaller set
X
obtaining
equal size sets X+and X#.
Notice that in the optimization problem
(3)
the domain of
π
is a finite set of injective functions. For a given value
of
k
, the number of such functions is
k!n
k2
making thus
an exhaustive search computationally infeasible. Instead,
we show in Section 5that the optimization problem formu-
lated in
(3)
can indeed be solved efficiently with complexity
e
O(√k n2)
, where the notation
e
O
hides polylogarithmic fac-
tors, i.e., up to polylogarithmic factors, the computational
cost is of order √kn2.