
Quantifying Complexity:
An Object-Relations Approach to Complex Systems
Stephen Casey1
NASA Langley Research Center
(*Electronic mail: Stephen.Casey@nasa.gov)
(Dated: 10 November 2022)
The best way to model, understand, and quantify the information contained in complex systems is an open question in
physics, mathematics, and computer science. The uncertain relationship between entropy and complexity further com-
plicates this question. With ideas drawn from the object-relations theory of psychology, this paper develops an object-
relations model of complex systems which generalizes to systems of all types, including mathematical operations,
machines, biological organisms, and social structures. The resulting Complex Information Entropy (CIE) equation is a
robust method to quantify complexity across various contexts. The paper also describes algorithms to iteratively update
and improve approximate solutions to the CIE equation, to recursively infer the composition of complex systems, and
to discover the connections among objects across different lengthscales and timescales. Applications are discussed in
the fields of engineering design, atomic and molecular physics, chemistry, materials science, neuroscience, psychology,
sociology, ecology, economics, and medicine.
How can one observe the behavior of a system and deter-
mine its inner workings? It is usually impossible to solve
this problem exactly, since most real-world models con-
tain uncertainty in relation to the systems they represent.
There is a tradeoff between model accuracy, model size,
and computational cost. There are often also hidden vari-
ables and elements of stochasticity.
I. INTRODUCTION
Minimum Description Length (MDL) is often used as a pa-
rameter in model selection to fit a particular dataset, whereby
the shortest description of the data, or the model with the best
compression ratio, is assumed to be the best model1. MDL
is a mathematical implementation of Occam’s razor, which
dictates that among competing models, the model with the
fewest assumptions is to be preferred. If an MDL descrip-
tion can contain all of the operations in a Turing-complete
programming language, the description length then becomes
equivalent to the Kolmogorov complexity, which is the length
of the shortest computer program that produces the dataset as
output2. For a given dataset, a program is Pareto-optimal if
there is no shorter program that produces a more accurate out-
put. The graph of Kolmogorov program length vs. accuracy
is the Pareto frontier of a dataset.
If one combines Occam’s razor with the Epicurean Prin-
ciple of Multiple Explanations, the result is Solomonoff’s
theory of inductive inference34. According to the Principle
of Multiple Explanations, if more than one theory is consis-
tent with the observations, all such theories should be kept.
Solomonoff’s induction considers all computable theories that
may have generated an observed dataset, while assigning a
higher Bayesian posterior probability to shorter computable
theories. The theory of inductive inference uses a computa-
tional formalization of Bayesian statistics to consider multiple
competing programs simultaneously in accordance with the
Principle of Multiple Explanations, while prioritizing shorter
programs in accordance with Occam’s razor.
A characteristic feature of complex systems is their activ-
ity across a wide range of lengthscales and timescales. As
a consequence, they can often be divided into a hierarchy of
sub-components. In order to reduce the computational dif-
ficulty of programs that can be divided into sub-programs,
the dynamic programming method, originally developed by
Richard Bellman5, simplifies a complicated problem by re-
cursively breaking it into simpler sub-problems. The divide-
and-conquer technique used by dynamic programming can
applied to both computer programming and mathematical op-
timization. A problem is said to have optimal substructure
if it can be solved optimally by recursively breaking into
sub-problems and finding the optimal solution to each sub-
problem.
In contrast to the divide-and-conquer procedure, a unifi-
cation procedure finds a single underlying theory that can ex-
plain two or more separate higher-level theories. Causal learn-
ing algorithms or structural learning algorithms are examples
of unification procedures, which determine cause-and-effect
among variables in an observed dataset. These variables may
be hidden or unknown, in which case they must be inferred via
hidden Markov models, Bayesian inference models, or other
methods.
These four principles – Occam’s razor, the Principle of
Multiple Explanations, divide-and-conquer, and unification –
have been used in various combinations to analyze complex
systems. Some researchers have used them all together, such
as Wu, Udrescu, and Tegmark in their endeavor to create an
AI physicist789. In order to fully characterize the informa-
tion contained in complex systems, a fifth item should be
added to this list: the principle of overinterpretation. This
principle states there can be multiple correct interpretations
of the same dataset, and the validity of any particular lens
of interpretation is determined by whether or not it produces
useful insights. This is distinct from the Principle of Multi-
ple Explanations, which states that no plausible theory can be
ruled out, but which still assumes there is one correct theory
arXiv:2210.12347v2 [cs.LG] 9 Nov 2022