
Shortest Edit Path Crossover
To meet the above challenges, this paper first proposes a new
crossover operator based on shortest edit path (SEP) in the
original graph space. The SEP crossover does not impose
any constraints on other algorithmic components or appli-
cation scope, thereby forming a simple and generalizable
solution to the permutation problem. Second, a theory is
derived for analyzing mutation, standard crossover, RL, and
the proposed SEP crossover in the NAS domain. The SEP
crossover is shown to have the best expected improvement in
terms of graph edit distance (GED) between the found archi-
tecture and the global optimum. Third, empirical results on
SOTA NAS benchmarks further verify the theoretical analy-
sis, demonstrating that the SEP approach is effective. It thus
allows taking full advantage of population-based search, and
serves as a theoretical foundation for further research on
methods for NAS and similar problems. All source codes
for reproducing the experimental results are provided at:
(https://github.com/cognizant-ai-labs/sepx-paper).
2. Related Work
NAS NAS approaches can generally be categorized into
two groups: one-shot methods and black-box methods
(Mehta et al.,2022). In one-shot approaches (Liu et al.,
2019;Dong & Yang,2019;Chen et al.,2021), a supernet
is trained to represent the entire search space. The overall
training cost is reduced significantly; however, these ap-
proaches can only be run on small cell-based search spaces
with a complete graph (Mehta et al.,2022;Zela et al.,2020)
and the search objectives must be differentiable. In contrast,
although computationally more expensive, black-box meth-
ods have no restrictions on the search space or objectives,
making them a more general solution to NAS. Thus, this
paper will focus on black-box NAS.
Black-box NAS Black-box NAS methods, also called
zeroth-order methods, iteratively generate architectures for
evaluation, and then use the outcome to update the search
strategy. There are four main types of search strategies in
black-box NAS approaches: random search (Li & Talwalkar,
2020;Yu et al.,2020), RL (Zoph & Le,2017;Zoph et al.,
2018), evolutionary search (Real et al.,2017;2019), and
local search (White et al.,2021a;b). Local search, whether
used alone or together with neural predictors (e.g., Bayesian
models), is based on operations that are essentially the same
as mutation (although different terminology may be used)
(White et al.,2021a;b). They can therefore be seen as equiv-
alent to mutation-only evolutionary search with a population
size of one. The one search strategy that is significantly dif-
ferent from evolutionary methods is RL, and will thus be
included in the theoretical analysis in this paper. The the-
ory developed in this paper thus covers most of the search
strategies in Black-box NAS.
RL-based black-box NAS RL-based methods work by iter-
atively sampling architectures using a RL agent, then col-
lecting the prediction accuracies as the reward for updating
the policy. Zoph & Le (2017) successfully generated well-
performing convolutional networks and recurrent cells using
a specially designed recurrent neural network as the agent.
Zoph et al. (2018) further showed that the approach finds
architectures that transfer well between different datasets.
In a recent empirical study (Ying et al.,2019), a simple RL
agent based on multinomial probability distributions was
found to perform significantly better on NAS-bench-101
than previous RL-based NAS methods. This RL controller
is analyzed in this paper as well.
Evolutionary Black-box NAS Evolutionary NAS methods
work by improving a population of architectures over time
(Liu et al.,2021). To generate new offspring architectures,
two operators can be used: a random edge/node mutation ap-
plied to an existing architecture, and crossover to recombine
two existing architectures. Architectures that do not perform
well are removed periodically from the population, and the
best-performing architecture returned in the end. While
crossover is a powerful operator, most existing methods rely
on mutation only because of the permutation problem. It is
this problem that this work aims to solve, in order to take
full advantage of the evolutionary approch in NAS.
The permutation problem and existing solutions The
permutation problem has been discussed in the Neuroevolu-
tion community for many years. One simple but common
solution is simply to get rid of crossover completely dur-
ing evolution (Angeline et al.,1994;Yao & Liu,1998).
Indeed, almost all newly developed evolutionary NAS meth-
ods avoid using a crossover operator (Real et al.,2017;
Fernando et al.,2017;Liu et al.,2018;Elsken et al.,2019;
Real et al.,2019;So et al.,2021;Co-Reyes et al.,2021;
Gao et al.,2022). For instance, Real et al. (2017) reported
that crossover operators were included in their initial experi-
ments, but no performance improvement was observed, and
therefore only mutation was deployed in their final Amoe-
baNet algorithm.
A number of principled solutions have been proposed to
overcome the permutation problem as well. Many of them
require that the network topologies are fixed. For instance,
Thierens (1996) proposed a non-redundant encoding for
matching neurons during crossover, Uriot & Izzo (2020) de-
veloped a safe crossover through a neural alignment mecha-
nism, and Gangwani & Peng (2018) used genetic distillation
to improve crossover. Further, Dragoni et al. (2014) pro-
posed a generalization where the population can include
different topologies, but only parents with a similar topol-
ogy can be crossed over.
Other solutions have been developed for special cases, mak-
ing them non-applicable to arbitrary architectures. For in-
stance, the unit-alignment method (Sun et al.,2020) utilizes
2