Scaling Knowledge Graphs for Automating AI of Digital Twins 3
for doing this with semantic knowledge graphs that also may integrate external
data. Pan et al. [25] presents a survey of semantic data management systems and
benchmarks. The authors classify the systems using a taxonomy that includes
native RDF stores, RDBMS-based and NoSql-based data management systems.
Besta et al. [6] provide a classification based on the database technology. Their
analysis shows that the different design have various pros and cons. Some of the
widely-used generic triple-stores such as OpenLink Virtuoso [13], Apache Jena,
Blazegraph, GraphDB excel on managing RDF data, but, do not scale well in
integrating non RDF data. General purpose property graphs like Neo4J or Janus-
Graph lack intrinsic understanding of semantic models. Multi-modal databases
like ArgangoDB or Redis combine a no-sql database with a graph database that
allows to manage documents alongside the graph. But, they also suffer from a
good understanding of semantic [30]. Entris [12] and Schmidt [31] extend this
idea and use semantic models to manage additional data in a data lake. In Sec-
tion 3 we will discuss some unique requirements that create challenges in scaling
such knowledge graphs. We derive a reference architecture that separation the
semantic graph layer from the data layer to scale better to large volumes of data
and have federated access to address the Semantic Digital Threads requirements.
As shown by our experiments, such design seems to provide better scalability
for our use case compared to the other semantic data management approaches.
Benchmarks for Semantic Data: To validate that the requirements in modelling
Digital Twins are unique and evaluate different knowledge graph technologies, we
created a new Digital Twin Benchmark Model (DTBM). We compare it against
some established benchmarks. The Berlin SPARQL Benchmark (BSBM) [7] and
Lehigh University Benchmark (LUBM) [14] are generic RDF Benchmarks that
run a variant of queries on generated datasets. SP2Bench [33] is based on DBLP
library dataset and reflects the social network characteristics of semantic web
data. DBpedia SPARQL benchmark [23] uses real queries that were performed
by humans and applications against on DBpedia. Additional work reflects the
requirements and characteristics of certain domains. PODiGG [35] and GTFS-
Madrid-Bench [9] are examples of benchmarks for public transport domain fo-
cused on use cases and requirements on route planning on gespatial and temporal
transport data. LSLOD [15] contains datasets from the life sciences domain and
the Linked Data cloud and 10 simple and 10 complex queries that need query
federation. Fedbench suite [32] evaluates efficiency and effectiveness of federated
SPARQL queries over three different datasets: cross-domain, life science, and
SPBenc. We will use BSBM and LUBM in the evaluation in Section 6 as they
are very well established and tested for many knowledge graphs technologies and
address themselves different RDF characteristics. In addition, we will propose a
new benchmark focused on our use case.
3 Requirements for Semantic Digital Threads
A Digital Thread is linking data from different life cycle stages of a Digital
Twin. This starts from design documents such as textual requirements, test