
Tensor-reduced atomic density representations
James P. Darby∗,1, 2 Dávid P. Kovács∗,2Ilyes Batatia,2, 3 Miguel A. Caro,4
Gus L. W. Hart,5Christoph Ortner,6and Gábor Csányi2
1Warwick Centre for Predictive Modelling, School of Engineering,
University of Warwick, Coventry, CV4 7AL, UK
2Engineering Laboratory, University of Cambridge, Cambridge, CB2 1PZ UK
3ENS Paris-Saclay, Université Paris-Saclay, 91190 Gif-sur-Yvette, France
4Department of Electrical Engineering and Automation, Aalto University, FIN-02150 Espoo, Finland
5Department of Physics and Astronomy, Brigham Young University, Provo, Utah, 84602, USA
6Department of Mathematics, University of British Columbia,
1984 Mathematics Road, Vancouver, BC, Canada V6T 1Z2
(Dated: December 7, 2022)
Density based representations of atomic environments that are invariant under Euclidean sym-
metries have become a widely used tool in the machine learning of interatomic potentials, broader
data-driven atomistic modelling and the visualisation and analysis of materials datasets. The stan-
dard mechanism used to incorporate chemical element information is to create separate densities
for each element and form tensor products between them. This leads to a steep scaling in the size
of the representation as the number of elements increases. Graph neural networks, which do not
explicitly use density representations, escape this scaling by mapping the chemical element infor-
mation into a fixed dimensional space in a learnable way. By exploiting symmetry, we recast this
approach as tensor factorisation of the standard neighbour density based descriptors and, using a
new notation, identify connections to existing compression algorithms. In doing so, we form com-
pact tensor-reduced representation of the local atomic environment whose size does not depend on
the number of chemical elements, is systematically convergable and therefore remains applicable to
a wide range of data analysis and regression tasks.
Over the past decade, machine learning methods
for studying atomistic systems have become widely
adopted [1–3]. Most of these methods utilise representa-
tions of local atomic environments that are invariant un-
der relevant symmetries; typically rotations, reflections,
translations and permutations of equivalent atoms [4].
Enforcing these symmetries allows for greater data effi-
ciency during model training and ensures that predictions
are made in a physically consistent manner. There are
many different ways of constructing such representations
which are broadly split into two categories: (i) descrip-
tors based on internal coordinates, such as the Behler-
Parrinello Atom-Centered Symmetry Functions [5], and
(ii) density-based descriptors such as Smooth Overlap of
Atomic Positions (SOAP) [6] or the bispectrum [7, 8],
which employ a symmetrised expansion of ν-correlations
of the atomic neighbourhood density (ν= 2 for SOAP
and ν= 3 for the bispectrum). A major drawback of all
these representations is that their size increases dramati-
cally with the number of chemical elements Sin the sys-
tem. For instance, the number of features in the linearly
complete Atomic Cluster Expansion (ACE) [9, 10] de-
scriptor which unifies, extends and generalises the afore-
mentioned representations, scales as Sνfor terms with
correlation order ν(i.e. a body order of ν+ 1). This
poor scaling severely restricts the use of these represen-
tations in many applications. For example, in the case of
*These authors contributed equally.
machine learned interatomic potentials for systems with
many (e.g. more than 5) different chemical elements,
the large size of the models results in memory limita-
tions being reached during parameter estimation as well
as significantly reducing evaluation speed.
Multiple strategies to tackle this scaling problem have
been proposed including element weighting [11, 12] or
embedding the elements into a fixed small dimensional
space [13, 14], directly reducing the element-sensitive cor-
relation order [15], low-rank tensor-train approximations
for lattice models [16] and data-driven approaches for se-
lecting the most relevant subset or combination of the
original features for a given dataset [17–19]. A rather
different class of machine learning methods are Message
Passing Neural Networks (MPNNs) [20, 21]. Instead of
constructing full tensor products, these models also em-
bed chemical element information in a fixed size latent
space using a learnable transformation RS→RKwhere
Kis the dimension of the latent space, and thus avoid
the poor scaling with the number of chemical elements.
Recently these methods have achieved very high accu-
racy [22–24], strongly suggesting that the true complex-
ity of the relevant chemical element space does not grow
as Sν.
In this paper we introduce a general approach for sig-
nificantly reducing the scaling of density-based represen-
tations like SOAP and ACE. We show that by exploiting
the tensor structures of the descriptors and applying low-
rank approximations we can derive new tensor-reduced
descriptors which are systematically convergeable to the
arXiv:2210.01705v2 [physics.chem-ph] 6 Dec 2022