is observed during training, as measured along fixed final-policy trajectories. Therefore, the com-
plexity of a final learned policy does not come principally from increased density on-and-around
the optimal trajectories, which is a potentially surprising result. In contrast, as measured along the
evolving current-policy trajectories, a decrease in region density is observed during training. Across
all settings, we also observe that the region-transition count, as observed during fixed time-duration
episodes, grows during training before converging to a plateau. However, the trajectory length, as
measured in the input space, also grows towards a plateau, although not at the same rate, and this
leads to variations in the mean region densities as observed along current trajectories during training.
2 Related Work
Understanding the expressivity of a neural network is fundamental to better understanding its opera-
tion. Several works study exprressivity of deep neural networks with piecewise linear activations by
counting their linear regions [Arora et al., 2016, Bianchini and Scarselli, 2014]. On the theoretical
side, Pascanu et al. [2013] show that in the asymptotic limit of many hidden layers, deep ReLU net-
works are capable of separating their input space into exponentially more linear regions compared
with shallow networks, despite using the same number of computational units. Following this, Mon-
túfar et al. [2014] also explore the complexity of functions computable by deep feed-forward neural
networks with ReLU and maxout activations, and provide tighter upper and lower bounds for the
maximal number of linear regions. They show that the number of linear regions is polynomial in the
network width and exponential with respect to the network depth. Furthermore, Raghu et al. [2017]
improve the upper bound for ReLU networks by introducing a new set of expressivity measures
and showing that expressivity grows exponentially with network depth for ReLU and tanh activated
neural networks. Serra et al. [2018] generalizes these results by providing even tighter upper and
lower bounds for the number of regions for ReLU networks and show that the maximal number of
regions grows exponentially with depth when input dimension is sufficiently large.
More recent works that touch on expressivity of depth show that the effect of depth on the expres-
sivity of neural networks is likely far below that of the theoretical maximum proposed by prior liter-
ature. Hanin and Rolnick [2019b] study the importance of depth on expressivity of neural networks
in practice. They show that the average number of linear regions for ReLU networks at initialization
is bounded by the number of neurons raised to the input dimension, and is independent of network
depth. They also empirically show that this bound remains tight during training. Similarly, Hanin
and Rolnick [2019a] find that the average distance to the boundary of the linear regions depends only
on the number of neurons and not on the network depth – both at initialization and during training
for supervised-learning tasks on ReLU networks. This strongly suggests that deeper networks do
not necessarily learn more complex functions in comparison to shallow networks. Prior to this, a
number of works have shown that the strength of deep learning may arise in part from a good match
between deep architectures and current training procedures [Mhaskar and Poggio, 2016, Mhaskar
et al., 2016, Zhang et al., 2021]. Notably, Ba and Caruana [2014] show that, once deep networks are
trained to perform a task successfully, their behavior can often be replicated by shallow networks,
suggesting that the advantages of depth may be linked to easier learning.
Another line of work studies function complexity in terms of robustness to perturbations to the
input. Sokoli´
c et al. [2017] theoretically studies the input-output Jacobian, which is a measure of
robustness and also relates to generalization. Similarly, Zahavy et al. [2016b] propose a sensitivity
measure in terms of adversarial robustness, and provides theoretical and experimental insights on
how it relates to generalization. Novak et al. [2018] also study robustness and sensitivity using the
input-output Jacobian and number of transitions along trajectories in the input space as measures of
robustness. They show that neural networks trained for image classification tasks are more robust to
input perturbations in the vicinity of the training data manifold, due to training points lying in regions
of lower density. Several other recent works have also focused on proposing tight generalization
bounds for neural networks [Bartlett et al., 2017, Dziugaite and Roy, 2017, Neyshabur et al., 2017].
There are a number of works that touch on understanding deep neural networks by finding general
principles and patterns during training. Arpit et al. [2017] empirically show that deep networks prior-
itize learning simple patterns of the data during training. Xu et al. [2019] find a similar phenomenon
in the case of 2-layer networks with Sigmoid activations. Rahaman et al. [2019] study deep ReLU
activated networks through the lens of Fourier analysis and show that while deep neural networks
can approximate arbitrary functions, they favour low frequency ones and thus, they exhibit a bias
3