
Mining Causality from Continuous-time Dynamics Models:
An Application to Tsunami Forecasting
Fan Wu∗Sanghyun Hong†Donsub Rim‡Noseong Park§Kookjin Lee¶
Abstract
Continuous-time dynamics models, such as neural ordinary
differential equations, have enabled the modeling of under-
lying dynamics in time-series data and accurate forecast-
ing. However, parameterization of dynamics using a neu-
ral network makes it difficult for humans to identify causal
structures in the data. In consequence, this opaqueness hin-
ders the use of these models in the domains where capturing
causal relationships carries the same importance as accurate
predictions, e.g., tsunami forecasting. In this paper, we ad-
dress this challenge by proposing a mechanism for mining
causal structures from continuous-time models. We train
models to capture the causal structure by enforcing sparsity
in the weights of the input layers of the dynamics models.
We first verify the effectiveness of our method in the sce-
nario where the exact causal-structures of time-series are
known as a priori. We next apply our method to a real-
world problem, namely tsunami forecasting, where the exact
causal-structures are difficult to characterize. Experimental
results show that the proposed method is effective in learn-
ing physically-consistent causal relationships while achieving
high forecasting accuracy.
Keywords: Causality mining, Neural ordinary differential
equations, Tsunami forecasting
1 Introduction.
An emerging paradigm for modeling the underlying dy-
namics in time-series data is to use continuous-time
dynamics models, such as neural ordinary differen-
tial equations (NODEs) [2, 28, 5]. Widely known as
a continuous-depth extension of residual networks [8],
NODEs have a great fit for data-driven dynamics mod-
eling as they construct models in the form of systems of
ODEs. This new paradigm has enabled breakthroughs
in many applications, such as patient status/human
activity prediction [28], computational physics prob-
lems [12, 13, 14], or climate modeling [11, 24].
A promising approach to improve the performance
of those models is to increase the expressivity of the
neural network they use to parameterize a system,
e.g., by employing convolutional neural networks or
∗Arizona State University, fanwu8@asu.edu
†Oregon State University, sanghyun.hong@oregonstate.edu
‡Washington University in St. Louis, rim@wustl.edu
§Yonsei University, noseong@yonsei.ac.kr
¶Arizona State University, kookjin.lee@asu.edu
by augmenting extra dimension in the state space [5].
However, the more complex neural networks are, the
more challenging humans are to interpret the learned
models. It will be even more problematic for tasks that
require interpretable causality,e.g., a high-consequence
event prediction, as in tsunami forecasting.
In this work, we study a mechanism for mining
the causality in time-series data from continuous-time
neural networks trained on it. Specifically, we ask:
How can we identify causal structures from a
continuous-time model? How can we promote
the model to learn causality in time-series?
This work focuses on Granger causality [6], a common
framework that are used to quantify the impact of a
past event observed on the future evolution of data.
Our contributions. First, we propose a mechanism
for extracting the causal structures from the parameters
of a continuous-time neural network. We adapt the con-
cept of component-wise neural networks, studied in [33],
for continuous-time models. We enforce column-wise
sparsity to the weights of the input layers so that the
impact of the input elements of less contributions (less
causal) can be small. This is achieved by using a train-
ing algorithm, which minimizes data-matching loss and
sparsity-promoting loss and prunes the columns whose
weights are smaller than a threshold. At the end of the
training, the norms of columns corresponding to the
important input elements will have high magnitudes,
which will make interpretation on causality easier.
Second, we test our approach if the learned causal
structures match with the known ground-truth causal
structures. To evaluate, we train two continuous-time
dynamics models, NODEs [2] and neural delay differen-
tial equations (NDDEs) [37], on the data sampled from
the Lorenz-96 and Mackey–Grass systems, respectively.
The results demonstrate that our mechanism precisely
extracts the causal structures from the data.
Third, we further evaluate our approach in tsunami
forecasting at the Strait of Juan de Fuca [22], where
we do not know the exact causal relationships in the
data. We train NDDEs on the tsunami dataset [20]
arXiv:2210.04958v2 [cs.LG] 13 Oct 2022