
General framework for E(3)-equivariant neural network representation of density
functional theory Hamiltonian
Xiaoxun Gong,1, 2 He Li,1, 5 Nianlong Zou,1Runzhang Xu,1Wenhui Duan,1, 3, 4, 5, 6, ∗and Yong Xu1, 3, 4, 7, †
1State Key Laboratory of Low Dimensional Quantum Physics and
Department of Physics, Tsinghua University, Beijing, 100084, China
2School of Physics, Peking University, Beijing 100871, China
3Tencent Quantum Laboratory, Tencent, Shenzhen, Guangdong 518057, China
4Frontier Science Center for Quantum Information, Beijing, China
5Institute for Advanced Study, Tsinghua University, Beijing 100084, China
6Beijing Academy of Quantum Information Sciences, Beijing 100193, China
7RIKEN Center for Emergent Matter Science (CEMS), Wako, Saitama 351-0198, Japan
Combination of deep learning and ab initio calculation has shown great promise in revolutionizing
future scientific research, but how to design neural network models incorporating a priori knowledge
and symmetry requirements is a key challenging subject. Here we propose an E(3)-equivariant
deep-learning framework to represent density functional theory (DFT) Hamiltonian as a function
of material structure, which can naturally preserve the Euclidean symmetry even in the presence
of spin-orbit coupling. Our DeepH-E3 method enables very efficient electronic-structure calculation
at ab initio accuracy by learning from DFT data of small-sized structures, making routine study of
large-scale supercells (>104atoms) feasible. Remarkably, the method can reach sub-meV prediction
accuracy at high training efficiency, showing state-of-the-art performance in our experiments. The
work is not only of general significance to deep-learning method development, but also creates new
opportunities for materials research, such as building Moir´e-twisted material database.
I. INTRODUCTION
It has been well recognized that deep learning methods
could offer a potential solution to the accuracy-efficiency
dilemma of ab initio material calculations. Deep-learning
potential [1,2] and a series of other neural network
models [3–7] are capable of predicting the total energies
and atomic forces of given material structures, enabling
molecular dynamics simulation at large length and time
scales. The paradigm has been used for deep-learning
research of various kinds of physical and chemical prop-
erties [8–19]. Remarkably, a deep neural network rep-
resentation of density functional theory (DFT) Hamil-
tonian (named DeepH) was developed by employing the
locality of electronic matter, localized basis, and local
coordinate transformation [20]. By the DeepH approach
the computationally demanding self-consistent field it-
erations could be bypassed and all the electron-related
physical quantities in the single-particle picture can in
principle be derived very efficiently. This opens oppor-
tunities for the electronic-structure calculation of large-
scale material systems.
Introducing physical insights and a priori knowledge
into neural networks is of crucial importance to the deep-
learning approaches. Specifically, the deep-learning po-
tential takes advantage of the invariance of the total en-
ergy under rotation, translation and spatial inversion as
well as permutation of atoms. For DeepH, the property
that the Hamiltonian matrix changes covariantly (i.e.
∗duanw@tsinghua.edu.cn
†yongxu@mail.tsinghua.edu.cn
equivariantly) under rotation or gauge transformations
should be preserved by the neural network model for effi-
cient learning and accurate prediction (Fig. 1). A strat-
egy is developed to apply local coordinate transformation
which changes the rotation covariant problem into an in-
variant one and thus the transformed Hamiltonian matri-
ces can be learned flexibly via rotation-invariant neural
networks [20]. Nevertheless, the large amount of local
coordinate information seriously increases the computa-
tional load, and the model performance depends critically
on a proper selection of local coordinates, which relies
on human intuition and is not easy to optimize. Alter-
natively, one may get rid of the local coordinate trans-
formation by applying the equivariant neural network
(ENN) [21–24]. The key innovation of ENN is that all
the internal features transform under the same symmery
group with the input, thus the symmetry requirements
are explicitly treated and exactly satisfied, as shown by
a series of neural network models for various material
properties [6,7,13–15], including PhiSNet [25] for pre-
dicting the Hamiltonian of molecules with fixed system
size. However, the key capability of DeepH that learns
from DFT results on small-sized material systems and
predicts the electronic structures of much larger ones has
not been demonstrated by ENN models. More critically,
the existing ENN models have neglected the equivari-
ance in the spin degrees of freedom, although the elec-
tronic spin and spin-orbit coupling (SOC) play a key role
in modern condensed matter physics and materials sci-
ence. With SOC, one should take care of the spin-orbital
Hamiltonian, whose spin and orbital degrees of freedom
are coupled and transform together under change of co-
ordinate system or basis set, as illustrated in Fig. 1. This
would raise critical difficulties in designing ENN models
arXiv:2210.13955v1 [physics.comp-ph] 25 Oct 2022