recurrent network transforms the activity vector based on the movement velocity of the agent, so
that the transformation is a representation of self-motion, when considered from the perspective of
representational learning. The vector and the transformation together form a representation of the 2D
Euclidean group, which is an abelian additive Lie group.
In a recent paper, [
21
] studied the group representation property and the isotropic scaling or conformal
isometry property for the general transformation model. In the context of linear transformation models,
they connected this property to the hexagon periodic patterns of the grid cell response maps. With the
conformal isometry property of the transformation of the recurrent neural network, the change of the
activity vector in the neural space is proportional to the input velocity of the self-motion in the 2D
physical space. [
21
] justified this condition in terms of robustness to errors or noises in the neurons.
Although [
21
] studied general transformation model theoretically, they focused on a prototype model
of linear recurrent network numerically, which has an explicit algebraic and geometric structure in
the form of a matrix group of rotations.
In this paper, we study conformal isometry in the context of the non-linear recurrent model that
underlies the hand-crafted continuous attractor neural network (CANN) [
6
,
7
,
34
,
1
]. In particular, we
will focus on the vanilla version of the recurrent network that is linear in the vector representation of
self-position and is additive in the input velocity, followed by an element-wise non-linear rectification
(such as ReLU non-linearity). This model has the simplicity that it is additive in input velocity before
rectification. We also explore more complex variants for non-linear recurrent networks, such as
the long short-term memory network (LSTM) [
26
]. Such models have been studied in recent work
[9, 3, 37, 8].
Our numerical experiments show that our conformal isometry condition is able to learn highly
structured multi-scale hexagon grid code, consistent with the properties of experimentally observed
grid cells of rodents. In addition, our learned model is capable of accurate path integration over a
long distance. Our results generalize previous results of linear network models in [
21
] to an important
class of non-linear neural network models in theoretical neuroscience that are more physiologically
realistic.
The main contributions of our paper are as follows. (1) We investigate the algebraic, geometric, and
topological properties of the transformation models of grid cells. (2) We study a simple non-linear
recurrent network that underlies the hand-crafted continuous attractor networks for grid cells, and our
numerical experiments suggest that conformal isometry is linked to hexagonal periodic patterns of
grid cells.
2 Lie group representation and conformal isometry
2.1 Representations of self-position and self-motion
We start by introducing the basic components of our model.
x= (x1,x2)∈R2
denotes the position
of the agent. Let
∆x= (∆x1,∆x2)
be the input velocity of the self-motion, i.e., displacement of the
agent within a unit time, so that the agent moves from xto x+∆xafter the unit time.
We assume
v(x) = (vi(x),i=1,...,D)
to be the vector representation of self-position
x
, where
each element
vi(x)
can be interpreted as the activity of a grid cell when the agent is at position
x
.
(vi(x),∀x)
corresponds to the response map of grid cell
i
.
D
is the dimensionality of
v
, i.e., the
number of grid cells. We refer to the space of
v
as the “neural space”. We normalize
kv(x)k=1
in
our experiments.
The set
(v(x),x∈R2)
forms a 2D manifold, or an embedding of
R2
, in the
D
-dimensional neural
space. We will refer to (v(x),x∈R2)as the “coding manifold”.
With self-motion
∆x
, the vector representation
v(x)
is transformed to
v(x+∆x)
by a general
transformation model:
v(x+∆x) = F(v(x),∆x) = F∆x(v(x)),(1)
where by simplifying
F(·,∆x)
as
F∆x(·)
in notation, we emphasize that the transformation
F
is
dependent on
∆x
. While
v(x)
is a representation of
x
,
F∆x
is a representation of
∆x
.
(v(x),∀x)
and
(F∆x(·),∀∆x)
together form a representation of the 2D additive Euclidean group
R2
, which is
an abelian Lie group. Specifically, we have the following group representation condition for the
transformation model:
2