Reference [17] emphasizes that the model can predict the
dynamics of a new robot(i.e., where its data has not been
used for training) by learning the physical relations between
links and joints. Notably, GNNs can even be applied to zero-
shot transfer learning for modular robots [18], [19].
However, asking models to learn all types of robot physics
is unrealistic. It would require tremendous training data; it
is not applicable to robots with different types of sensors
and actuators; and sensors produced by different manufac-
turers can have different performances. Instead, we focus
on reducing required sample complexity for system identi-
fication using physical priors. In this paper, we introduce
new data representations for symmetrical-legged robots and
how GNNs and other group-equivariant neural networks
can improve sample efficiency based on the permutation
equivariance in legged-robot dynamics.
A. Contribution
This paper proposes sample efficient dynamics learning for
symmetrical-legged robots. By using grouped-by-leg struc-
tured data, we can define group invariance and equivariance
followed by geometric symmetries of a robot based on the
data representation. Then, by training a model based on
the networks designed to preserve the group equivariance,
we can learn the dynamics that incorporate physics priors,
providing sample efficiency and generalization capability.
II. PRELIMINARIES
Using a priori representational and computational assump-
tions to achieve generalization is actively studied in the
deep learning. This can be realized by representing prior
knowledge as equivariance or invariance under group actions.
Below we introduce the relevant concepts and prior works.
A. Invariance, Equivariance, and Symmetries
In geometry, we say that an object has symmetry if the
object has an invariance under the group action. In group
theory, group action, invariance, and equivariance can be
defined as follows:
Definition 1 (Group Action) A group G is said to act
on a set Xif there is a map ψ:G×X→Xcalled
the Group Action such that ψ(e, x) = xfor all x∈X,
and ψ(g, ψ(h, x)) = ψ(gh, x)for all g, h ∈G, x ∈X.
For compactness we follow the practice of denoting a group
action by a ‘◦’ e.g. g◦x.
Definition 2 (Invariance) Let Gbe a group which acts
on the sets X. We say that the function f:X→Yis
G-invariant if f(g◦x) = f(x)for all x∈X, g ∈G.
Definition 3 (Equivariance) Let Gbe a group which acts
on the sets Xand Y. We say that the function f:X→Yis
G-equivariant if f(g◦x) = g◦(f(x)) for all x∈X, g ∈G.
B. Equivariance and Invariance in Neural Networks
Many works in deep learning leverage equivariance and
invariance for constructing the neural network architecture.
Data augmentation is a straightforward way to learn group-
equivariant representation [20]; however, it is computation-
ally inefficient due to the increased data volume to train the
model.
Convolutional Neural Networks One well-known example
to consider equivariance and invariance is convolutional neu-
ral networks (CNNs) [21], [22], which utilize convolutional
layer for preserving object identity over the translation group.
Group-Equivariant Neural Networks A group convolu-
tional layer is proposed to consider other symmetries like
rotation and reflection [23],[24]. For more general structures,
deep symmetry networks [25] capture a wide variety of
invariances defined over arbitrary symmetry groups using
kernel-based interpolation and symmetric space pooling.
For a set structure like point clouds, simple permutation-
equivariant layers [26] are shown to be beneficial.
Graph Neural Networks Graph neural networks [15], [27]
is a type of neural networks directly operating on graph struc-
tures where a system of objects and relations is represented
through nodes and edges. GNNs can capture all permutation
invariance, and equivariance [28] in nodes and edges. Graph
Networks (GNs), proposed by [10], provides a general form
of graph structure for learning, G= (u,{ni},{ej, sj, rj})
that includes a graph embedding vector called global features
u, a set of node features {ni}i=1,..,Nnand a set of edges
{ej, sj, rj}j=1,..,Ne, where ejis a vector representing edge
features and sj,rjare the indices of the sender and receiver
nodes ranging from 1to Nn. This data structure allows
GNNs to capture the permutation equivariance of nodes and
edges. Then graph networks are defined as a ”graph2graph”
module, which takes an input graph and returns an output
graph with the updated features.
III. STRUCTURED REPRESENTATION FOR SYMMETRICAL
LEGGED ROBOT DYNAMICS
This section describes the equivariance and invariance in
inverse dynamics of symmetrical-legged robots. Then, we
suggest data representation that can capture those properties
in neural networks.
A. Dynamics of Floating-base Legged Robots
The dynamics of legged robots are commonly described as
a function of the generalized coordinates q= [q>
b,q>
j]>∈
Rnq, where qb∈R6,qj∈Rnjrepresent the configuration of
the floating base and joints, respectively. As such, the rigid-
body dynamics of a floating base robot can be obtained using
Euler-Lagrange formalism yielding:
M(q)¨
q+b(q,˙
q) = S>
aτa+Jc(q)>Fc(1)
where M(q),b(q,˙
q)represent the mass/inertia matrix and
the Coriolis/centrifugal force plus the gravitational force.
Sa∈Rna×nqdenotes the selection matrix corresponding to
the index set of actuated joints, which maps τa(the actuated
joint torques) into the generalized forces and, Jc(q)denotes
the stacked contact Jacobian matrix that maps contact wrench
vector Fcinto the generalized forces. Finally, from the
dynamics equation, we can derive the forward model ffwd
and inverse model finv as:
¨
q=ffwd(q,˙
q,τ),τ=finv(q,˙
q,¨
q)(2)